Skip to content

Torch artificial neural network function for plumed2

License

Notifications You must be signed in to change notification settings

zwpku/torchann-plumed

Folders and files

NameName
Last commit message
Last commit date

Latest commit

 

History

16 Commits
 
 
 
 
 
 
 
 
 
 
 
 

Repository files navigation

Torch artificial neural networks (TorchANNs) for PLUMED

This is a plugin for PLUMED that implements the TorchFunc class as a subclass of Function class and the TorchColVar class as a subclass of ColVar class. It allows one to define functions and collective variables (ColVars) in PLUMED that are represented by artificial neural networks. In contrast to the ANN function module that supports fully-connected feedforward neural networks, this plugin allows the use of general neural network architectures.

Prerequiste

Installation

This plugin uses CMake as its build system. Once you are in the directory containing files of this package, follow these steps to install it:

  1. Create the directory in which to build the plugin: mkdir ./build
  2. In ./build directory run ccmake ../, and set the following variables.
  • set Torch_DIR to point to the directory where LibTorch is installed (it is typically /path-to-libtorch/share/cmake/Torch);
  • set PlUMED_INC_DIR to point to the directory that contains the header files of PLUMED;
  • set PLUMED to point to the directory that contains the library files of PlUMED, i.e. libplumedKernel.so;
  • set CMAKE_INSTALL_PREFIX to point to the directory where the plugin (i.e. the library file libTorchANNPlumed.so) will be installed.
  1. Build: make
  2. Install: make install

Usage

To use this plugin, we need to first load the dynamic library libTorchANNPlumed.so in the PLUMED script.

A file that contains the torch computational graph (i.e. the neural network) is required.

Similar to other plumed functions, a TorchANN function object requires the following keywords:

  • ARG (string array): input variable names of the function.
  • MODULE_FILE (string): filename of a saved computational graph.
  • NUM_OUTPUT(int): number of output components.

The output components can be accessed using output-0, output-1, and so on.

A TorchColVar object is assumed to be a function of coordinates of all atoms. It requires the following keywords:

  • MODULE_FILE (string): filename of a saved computational graph.
  • NUM_OUTPUT(int): number of output components.

The output components can be accessed using output-0, output-1, and so on.

Example

Below is an example that shows how to use this plugin with the OpenMM package. Suppose that both OpenMM and its plugin to interface with PLUMED are installed.

Step 1: Create the computational graphs

import torch
import numpy as np

# The user-defined neural network function
class MyFunc(torch.nn.Module):
    def forward(self, positions):
        return torch.cat((torch.sum(positions**2).reshape((1)), torch.mean(positions).reshape((1))), 0)

# The user-defined collective variables
class MyColVar(torch.nn.Module):
    def forward(self, positions):
        return torch.cat((positions[0].reshape((3)), torch.mean(positions).reshape((1))), 0)

# Render the compute graph to a TorchScript module
module = torch.jit.script(MyFunc())

# Serialize the compute graph to a file
module.save('myfunc.pt')

# Render the compute graph to a TorchScript module
module = torch.jit.script(MyColVar())

# Serialize the compute graph to a file
module.save('mycolvar.pt')

Step 2: Use this plugin in OpenMM code

plumed_script = """
  LOAD FILE=/path-to-library/libTorchANNPlumed.so
  t1: DISTANCE ATOMS=1,2
  t2: DISTANCE ATOMS=3,4
  ann: TORCHFUNC ARG=t1,t2 MODULE_FILE=myfunc.pt NUM_OUTPUT=2
  cv: TORCHCOLVAR MODULE_FILE=mycolvar.pt NUM_OUTPUT=4
  PRINT ARG=ann.output-0,cv.output-0 FILE=colvar"""
system.addForce(PlumedForce(plumed_script))

See the use of the openmm-plumed plugin.

Authors

Wei Zhang

About

Torch artificial neural network function for plumed2

Topics

Resources

License

Stars

Watchers

Forks

Releases

No releases published

Packages

No packages published