Articles

Filter By:

Article Type
  • Machine learning can improve scoring methods to evaluate protein–ligand interactions, but achieving good generalization is an outstanding challenge. Cao et al. introduce EquiScore, which is based on a graph neural network that integrates physical knowledge and is shown to have robust capabilities when applied to unseen protein targets.

    • Duanhua Cao
    • Geng Chen
    • Mingyue Zheng
    Article
  • The credit assignment problem involves assigning credit to synapses in a neural network so that weights are updated appropriately and the circuit learns. Max et al. developed an efficient solution to the weight transport problem in networks of biophysical neurons. The method exploits noise as an information carrier and enables networks to learn to solve a task efficiently.

    • Kevin Max
    • Laura Kriener
    • Mihai A. Petrovici
    Article
  • Bolstering the broad and deep applicability of graph neural networks, Heydaribeni et al. introduce HypOp, a framework that uses hypergraph neural networks to solve general constrained combinatorial optimization problems. The presented method scales and generalizes well, improves accuracy and outperforms existing solvers on various benchmarking examples.

    • Nasimeh Heydaribeni
    • Xinrui Zhan
    • Farinaz Koushanfar
    Article
  • Deep learning has led to great advances in predicting protein structure from sequences. Ren and colleagues present here a method for the inverse problem of finding a sequence that results in a desired protein structure, which is inspired by various components of AlphaFold combined with Markov random fields to decode sequences more efficiently.

    • Milong Ren
    • Chungong Yu
    • Haicang Zhang
    Article
  • Achieving the promised advantages of quantum computing relies on translating quantum operations into physical realizations. Fürrutter and colleagues use diffusion models to create quantum circuits that are based on user specifications and tailored to experimental constraints.

    • Florian Fürrutter
    • Gorka Muñoz-Gil
    • Hans J. Briegel
    Article
  • Machine learning-based surrogate models are important to model complex systems at a reduced computational cost; however, they must often be re-evaluated and adapted for validity on future data. Diaw and colleagues propose an online training method leveraging optimizer-directed sampling to produce surrogate models that can be applied to any future data and demonstrate the approach on a dense nuclear-matter equation of state containing a phase transition.

    • A. Diaw
    • M. McKerns
    • M. S. Murillo
    Article
  • Despite the existence of various pretrained language models for nucleotide sequence analysis, achieving good performance on a broad range of downstream tasks using a single model is challenging. Wang and colleagues develop a pretrained language model specifically optimized for RNA sequence analysis and show that it can outperform state-of-the-art methods in a diverse set of downstream tasks.

    • Ning Wang
    • Jiang Bian
    • Haoyi Xiong
    ArticleOpen Access
  • Large language models can be queried to perform chain-of-thought reasoning on text descriptions of data or computational tools, which can enable flexible and autonomous workflows. Bran et al. developed ChemCrow, a GPT-4-based agent that has access to computational chemistry tools and a robotic chemistry platform, which can autonomously solve tasks for designing or synthesizing chemicals such as drugs or materials.

    • Andres M. Bran
    • Sam Cox
    • Philippe Schwaller
    ArticleOpen Access
  • Methods for predicting molecular structure predictions have so far focused on only the most probable conformation, but molecular structures are dynamic and can change when performing their biological functions, for example. Zheng et al. use a graph transformer approach to learn the equilibrium distribution of molecular systems and show that this can be helpful for a number of downstream tasks, including protein structure prediction, ligand docking and molecular design.

    • Shuxin Zheng
    • Jiyan He
    • Tie-Yan Liu
    ArticleOpen Access
  • The central assumption in machine learning that data are independent and identically distributed does not hold in many reinforcement learning settings, as experiences of reinforcement learning agents are sequential and intrinsically correlated in time. Berrueta and colleagues use the mathematical theory of ergodic processes to develop a reinforcement framework that can decorrelate agent experiences and is capable of learning in single-shot deployments.

    • Thomas A. Berrueta
    • Allison Pinosky
    • Todd D. Murphey
    Article
  • Current limb-driven methods often result in suboptimal prosthetic motions. Kühn and colleagues develop a framework called synergy complement control (SCC) that advances prosthetics by learning ‘cyborg’ limb-driven control, ensuring natural coordination. Validated in diverse trials, SCC offers reliable and intuitive enhancement for limb functionality.

    • Johannes Kühn
    • Tingli Hu
    • Sami Haddadin
    ArticleOpen Access
  • Modelling the statistical and geometrical properties of particle trajectories in turbulent flows is key to many scientific and technological applications. Li and colleagues introduce a data-driven diffusion model that can generate high-Reynolds-number Lagrangian turbulence trajectories with statistical properties consistent with those of the training set and even generalize to rare, intense events unseen during training.

    • T. Li
    • L. Biferale
    • M. Buzzicotti
    ArticleOpen Access
  • Fragment-based molecular design uses chemical motifs and combines them into bio-active compounds. While this approach has grown in capability, molecular linker methods are restricted to linking fragments one by one, which makes the search for effective combinations harder. Igashov and colleagues use a conditional diffusion model to link multiple fragments in a one-shot generative process.

    • Ilia Igashov
    • Hannes Stärk
    • Bruno Correia
    ArticleOpen Access
  • Identifying compounds in tandem mass spectrometry requires extensive databases of known compounds or computational methods to simulate spectra for samples not found in databases. Simulating tandem mass spectra is still challenging, and long-range connections in particular are difficult to model for graph neural networks. Young and colleagues use a graph transformer model to learn patterns of long-distance relations between atoms and molecules.

    • Adamo Young
    • Hannes Röst
    • Bo Wang
    Article
  • The 5′ untranslated region is a critical regulatory region of mRNA, influencing gene expression regulation and translation. Chu, Yu and colleagues develop a language model for analysing untranslated regions of mRNA. The model, pretrained on data from diverse species, enhances the prediction of mRNA translation activities and has implications for new vaccine design.

    • Yanyi Chu
    • Dan Yu
    • Mengdi Wang
    Article
  • Using machine learning methods to model interatomic potentials enables molecular dynamics simulations with ab initio level accuracy at a relatively low computational cost, but requires a large number of labelled training data obtained through expensive ab initio computations. Cui and colleagues propose a geometric learning framework that leverages self-supervised learning pretraining to enhance existing machine learning based interatomic potential models at a negligible additional computational cost.

    • Taoyong Cui
    • Chenyu Tang
    • Wanli Ouyang
    Article
  • Generative models for chemical structures are often trained to create output in the common SMILES notation. Michael Skinnider shows that training models with the goal of avoiding the generation of incorrect SMILES strings is detrimental to learning other chemical properties and that allowing models to generate incorrect molecules, which can be easily removed post hoc, leads to better performing models.

    • Michael A. Skinnider
    ArticleOpen Access