Efficient training of generative models from multireference simulations and its application to the design of Dy complexes with large magnetic anisotropy
Abstract
Semi-supervised chemically-inspired training-by-proxy enables efficient generation of complex coordination compounds with targeted magnetic properties using reduced computational resources.
Generative machine learning models can potentially provide direct access to novel and relevant portions of the full chemical space, overcoming the cost of systematic sampling. However, the training of these models generally requires a large amount of data, often precluding the use of expensive high-level ab initio simulations for this task. The generation of coordination compounds of Dy with large magnetic anisotropy represents a topical example, where multireference simulations of large molecules are necessary to perform reliable predictions. Here, we show that a semi-supervised chemically-inspired training-by-proxy of generative variational autoencoders can reduce the cost associated with building a training set from multireference simulations by two orders of magnitude. We illustrate the power of this approach by generating 100s of new organic ligands for Dy(III) pentagonal bipyramidal complexes exhibiting record values of magnetic anisotropy, while starting from datasets as small as 1k multireference calculations. This work thus paves the way to the computational generation of molecules as complex coordination compounds with target electronic and magnetic properties.
Models citing this paper 0
No model linking this paper
Datasets citing this paper 0
No dataset linking this paper
Spaces citing this paper 0
No Space linking this paper
Collections including this paper 0
No Collection including this paper