Amer El-Samman

Physics-Backed ML Models in Chemistry

Compressing First-Principles Physics with Machine Learning

From 118 elements, quantum mechanics provides the rules that shape our universe. These rules determine how we treat disease, store energy, design materials, and process information. But first principles physics is enormous in scale. Even a modest molecule like caffeine requires an exact quantum calculation so large that no existing computer could store the full solution. Modern machine learning, trained on quantum mechanical data, offers a practical way through this. It compresses memory heavy quantum information into high quality statistical representations that remain faithful to structure, reflect the language of reactions, and can be transferred to new tasks with only small datasets.

The Hidden Language of AI in Chemistry: Unifying Graph and Language Models of Chemistry

The decision making frameworks of graph neural network models contain a hidden map of reaction formulae. Much like how natural language models capture relationships between words through vector arithmetic, where “King” minus “Man” plus “Woman” yields “Queen,” graph based models appear to learn an analogous language for chemistry. In this representation space, the model encodes reactions such as “Hydrogen” plus “Oxygen” resulting in “Water,” revealing that fundamental chemical transformations emerge naturally within the geometry of the learned graph embeddings.

Learn more

Understanding the Internal Model Learned by Graph Neural Networks on Quantum Chemistry

Graph neural networks are among the most sophisticated probabilistic models used in chemistry, yet the principles guiding their internal decision making often remain obscure. Understanding their learned representations is essential for revealing the hidden logic within these complex GNN models. This work explores how to access and interpret GNN’s deeply embedded decision making frameworks.

Learn more

Highlighted Projects

How Does Message-Passing Work in Graph Neural Networks Trained on Quantum Data?

How does message passing in graph neural networks constructs a chemical representation from local atomic environments? By learning how each atom’s contribution propagates through the molecular graph, we learn how graph model develop a latent space that captures electronic structure, functional groups, and reaction relevant patterns. This structure explains why these models can transfer learning to new tasks, enabling accurate prediction of diverse chemical properties and even recovery of reaction formulas.

Learn More

Transfer Learning From Latent Space of Graph Neural Networks Trained on Quantum Data

Does a graph neural network trained on quantum chemical data compress a representation that extends beyond its original task. By probing transfer to properties such as electron occupancy, NMR, pKa, and solubility, we assess whether its compressed latent space can serve as a general purpose engine for chemical prediction and design.

Learn More