iMechanica - Comments for "Journal Club for November 2021: Machine Learning Potential for Atomistic Simulation" //m.limpotrade.com/node/25544 Comments for "Journal Club for November 2021: Machine Learning Potential for Atomistic Simulation" en Dear Ying, //m.limpotrade.com/comment/30706#comment-30706 <一个id ="comment-30706">

In reply to Timely topic!

Dear Ying,

Thank you for the informative comments! As you mentioned, DeepMD is certainly another useful tool. PINN is also an interesting direction which I consider as a combintation between classical empirical potential and ML potential. Thanks for sharing your review article on ML potential for CG systems.

Best,

Wei

Wed, 24 Nov 2021 22:29:31 +0000 Wei Gao comment 30706 at //m.limpotrade.com
Timely topic! //m.limpotrade.com/comment/30705#comment-30705 <一个id = "评论- 30705 " > < / > < p > < em >回复< href="//m.limpotrade.com/node/25544">Journal Club for November 2021: Machine Learning Potential for Atomistic Simulation

Hi Wei,

Many thanks for this timely topic and for leading the great discussion.

Added to your summary, there are a few related works that might be useful:

1) The ML potential tool, such as the DeepMD is another excellent package, linked with LAMMPS to use. The DeepMD package can reproduce the temperature-pressure phase diagram of water molecules in the recent PRL paper, which is an intriguing demonstration for many mechanics problems under extreme environments. It also won 2020 ACM Gordon Bell Prize at SC20!

2) Physics-informed ML potential could be particularly useful since we want to avoid the unphysical interactions (energy, force, or stress) during the mechanical deformation. Imagining that we only train the ML potential from equilibrium conditions, it could not be used for large deformation simulations:) The recent physically informed artificial neural networks for atomistic modeling of materials can nicely address this issue, opening another avenue for ML potential to be applicable for large deformations, fractures, etc.

3) ML potential is not only useful for all-atom molecular simulations, but also for coarse-grained molecular simulations. When the ML potential is trained based on DFT data, it can reproduce the quantum accuracy, with efficiency as typical all-atom molecular simulations. Similarly, the ML coarse-grained model can archive the all-atom accuracy, with much less computational cost. It will open another door for us to model many large-scale mechanics problems with less computational time. We have a recent review article on this topic, Machine Learning of Coarse-Grained Models for Organic Molecules and Polymers: Progress, Opportunities, and Challenges.

Again, these are just my two cents.

I look forward to your fascinating works in this area and more discussions :)

Best, Ying

Tue, 23 Nov 2021 16:58:40 +0000 Ying Li comment 30705 at //m.limpotrade.com
Dear Haoran, //m.limpotrade.com/comment/30704#comment-30704 <一个id = "评论- 30704 " > < / > < p > < em >回复< href="//m.limpotrade.com/comment/30703#comment-30703">Thanks for the excellent review!

Dear Haoran,

Thanks for your interests and questions.

(1) Like those reactive potentials, there is no need to define fixed bonds in the ML potentials that are trained with DFT data, so ML potentials are able to capture bond breaking/formation. There are some good examples, such as the references [1-4] listed in the text.

(2) Molecules can be more conveniently descripted by a graph, so they are first studied with end-to-end ML potentials. However, descriptor-based ML potentials also have been used for molecules. The performance of the potentials is not only dependent on the ML model, but also many other factors, such as the quality of dataset, the choice of descriptors (or the quality of feature learning), as well as the rigor of training and validation process. Therefore, I have not seen a rigorous comparison between descriptor-based and end-to-end ML potentials in terms of prediction accuracy. If someone wants the machine to learn as much as possible from data (maybe better than human designed descriptors), then end-to-end model wins. This motivation is driving the development of new methods along with the rapid development of AI technology. However, at the moment, descriptor-based ML potentials are better connected to large scale simulator such as LAMMPS, so it may be a good choice to start with descriptor-based potentials if the final target is for large scale MD simulation.

Best,

Wei Gao

Sun, 14 Nov 2021 16:21:15 +0000 Wei Gao comment 30704 at //m.limpotrade.com
Thanks for the excellent review! //m.limpotrade.com/comment/30703#comment-30703 <一个id = "评论- 30703 " > < / > < p > < em >回复< href="//m.limpotrade.com/node/25544">Journal Club for November 2021: Machine Learning Potential for Atomistic Simulation

Dear Wei,

Thank you for summarizing the recent developments in ML potentials for MD simulations. I'm not working in this specific field. Reading your review satisfies lots of my curiosities.

I have 2 questions to ask here.

(1) In MD simulations, some interatomic potentials can capture bond breaking/formation, like Reaxff; some potentials cannot do so. And you mentioned that the ML potentials are learned from DFT. So are the existing ML potentials capable of capturing the bond breaking/formation? Are there any successful examples?

(2) You mentioned that the early development of end-to-end graph-based ML potentials was focused on molecules. Does it mean the end-to-end graph-based ML potentials are a better choice for polymer systems?

Thanks,

Haoran

Sat, 13 Nov 2021 19:35:20 +0000 Haoran Wang comment 30703 at //m.limpotrade.com
Hi Ajit, //m.limpotrade.com/comment/30702#comment-30702 <一个id ="comment-30702">

In reply to Impressive!

Hi Ajit,

Thanks for your interests and questions. To your first question, yes, the polarization effect could be described by ML potential as long as the potential is trained with the information of charges. There is a recent work (Nature communications 12.1 (2021): 1-11) specifically targeting this type of problem. Your second question is about the gain of ML potential. The main advantage of ML potentials as compared to classical potentials is that they can be much more accurate, although they are still generally slower than classical potentials.

Wei Gao

Sat, 13 Nov 2021 15:10:01 +0000 Wei Gao comment 30702 at //m.limpotrade.com
令人印象深刻的! //m.limpotrade.com/comment/30699#comment-30699 <一个id = "评论- 30699 " > < / > < p > < em >回复< href="//m.limpotrade.com/node/25544">Journal Club for November 2021: Machine Learning Potential for Atomistic Simulation

Dear Wei,

Thanks for a very informative edition of the journal club.

Though I have been in the ML field for some time, when it comes to this particular area (ML Potentials for Atomistic Simulations) I am an absolute newbie. In fact, this journal club edition is my first ever contact with this research area. ... It does look like there has been a great deal of activity in this area in the recent times, and people seem to have approached the problems with a lot of creativity too. All in all, very impressive! [Even if you set everything else aside, anything like "accelerating the computational time of the high-throughput search by a factor of 130" just has to be impressive!]

OK, now, allow me a couple of newbie questions...

How do these approaches work for systems / phenomena involving polarization? Any work or notable results in this direction?

How precisely does the gain in the high-throughput search come about? ... If I understand it right, these potentials are just going to get incorporated into the MD / atomistic packages like LAMMPS, right? So, speaking simple-mindedly, the run-time computational complexity should also stay more or less the same, right? If so, how come there still is a gain?

Best,

--Ajit

Wed, 10 Nov 2021 09:00:19 +0000 Ajit R. Jadhav comment 30699 at //m.limpotrade.com
manbetx体育|万博manbetx平台|万博体育平台 //m.limpotrade.com/comment/30698#comment-30698 <一个id ="comment-30698">

In reply to Journal Club for November 2021: Machine Learning Potential for Atomistic Simulation

Dear Rui,

Thank you so much for the kind and encouraging words. To your questions:

(1) You made an important point: the performance of ML potential is highly dependent on the training data. Only high quality data that covers the essential physics of interests can produce reliable ML potential. Therefore, developing ML potential starts from the data generation, usually by using DFT calculations. First, a variety of atomic structures have to be carefully prepared. For example, the atomic structures could come from the random perturbation of atom positions and lattice constants based on a perfect crystal structure. In addition, typical defect structures can be built into the data. Recently, we also use atomic structures coming from Nudged elastic band and Dimer calculations to inform the data with phase transition information and to better sample the potential energy surface. Nowadays, more and more researchers share materials dataset publicly, so that one could re-use those already available datasets and enrich them with specific physics of interests if needed. After atomic structures are determined, one just need to run DFT calculations to get the outputs of interests that will be used for training, such as energy, force, and stress.

The inputs to a ML model are the well-prepared atomic structures. These structures are converted to descriptors when one uses a descriptor-based methods. This conversion can be done automatically within the packages described in section 5 without interference. The outputs to a ML model depend on the application. Most of the time, the potential energy, atomic force and stress are used as outputs. ML model (e.g. neural network) can be conveniently built using some machine learning platforms such as Tensorflow and Pytorch, which provide the library functions for training the model. There are many hyper parameters that can be tuned in order to achieve a good convergence when the loss settles to within an error range. After training is done, the ML model can be saved as the ML potential, which can be used later like classical potentials.

(2) The tools described in section 5 can be used as black box, and the products (ML potentials) can be directly used for MD or MS simulations. Those tools except Schnet can be all connected to LAMMPS and used just like classical potentials.

Wei

Sun, 07 Nov 2021 17:07:57 +0000 Wei Gao comment 30698 at //m.limpotrade.com
very nice! //m.limpotrade.com/comment/30697#comment-30697 <一个id ="comment-30697">

In reply to Journal Club for November 2021: Machine Learning Potential for Atomistic Simulation

Dear Wei,

Thank you for this nice summary on machine learning potentials. I enjoyed reading it and learning the state of the art. I have two questions in mind:

(1) How to train a machine learning model? As you noted, the performance of a ML potential depends on both the choice of descriptor and ML model. In addition, I think it also depends on the training and the data used for training. Can you elaborate on the steps taken to train a ML model?

(2) How to use one of the ML tools (Section 5) as a blackbox for atomistic simulations? I assume that these tools have been trained one way or another and thus are ready to be used directly in place of the standard empirical potentials (e.g. in LAMMPS). Is it as simple as that?

Again, I am impressed by how much you have done in this area, and congratulations on your recent CAREER award!

Rui

坐,2021年11月06 17:36:00 + 0000 Rui Huang comment 30697 at //m.limpotrade.com
Hi Zheng, //m.limpotrade.com/comment/30696#comment-30696 <一个id = "评论- 30696 " > < / > < p > < em >回复< href="//m.limpotrade.com/comment/30695#comment-30695">Hi Wei, thanks so much for

Hi Zheng,

Thanks for your intetests! Most of the descriptor-based packages, as listed in section 5, only ask for atom coordinates as inputs, which are transformed to descritpors internally through descritpor functions (such as ACSF, SOAP, as listed in section 3). Some packages (such as n2p2) implicitly run descritpor generation so that it's not convenient (or impossible) for users to see the descriptors. In our code (AtomDNN), we compute the descritpors (ACSF and SOAP are currently supported) through customized LAMMPS Compute commands (can be found here with an example) and make the descriptor generation explict through a well defined data pipeline. In this way, descritpors can be computed without training the network. There are other tools, such as DScribe, which is only used for computing many types of descriptors. However, we found it is slow for computing the derivatives of descriptors (which are needed for compute atomic force and stress). For SOAP descriptors, you can also use QUIP to get the descriptors directly.

Wei

Tue, 02 Nov 2021 13:50:41 +0000 Wei Gao comment 30696 at //m.limpotrade.com
Hi Wei, thanks so much for //m.limpotrade.com/comment/30695#comment-30695 <一个id ="comment-30695">

In reply to Journal Club for November 2021: Machine Learning Potential for Atomistic Simulation

Hi Wei, thanks so much for offering such an informative tutorial. Just one quick question: For descriptor-based ML potentials, can we directly use the coordinates of atoms as the input for the neural network? If we have to convert the coordinates into descriptors as the inputs, what is the standard procedure to do so? Many thanks!

Tue, 02 Nov 2021 04:53:27 +0000 Zheng Jia comment 30695 at //m.limpotrade.com