Task 8 - Molecular Dynamics Simulations
In this section we will simulate the wildtype protein and two interesting mutants with MD, e.g. the gromacs package. For this we will use an automatic pipeline. As the final simulations will take a while, we will post the analysis part at a later point. The pipeline is available as a git repository. All the work needs to be done on the LRZ now.
The slides of the task: File:MD talk.pdf
- Login to the LRZ:
ssh -XY firstname.lastname@example.org
ssh -XY email@example.com
- In order to use git you have to load the software module first. http://www.lrz.de/services/compute/supermuc/software/
- Go to a designated directory and clone the repository from https://github.com/offmarc/AGroS
- Include all the scripts in the PATH environment variable
- Get a license for SCRWL4 and install it into the same dir where the scripts are: http://dunbrack.fccc.edu/scwrl4/
- Finally copy the WT and two mutants to the LRZ (scp)
- IMPORTANT: Before you continue you should have a look at the scripts and check what they do!
Prepare Job Scripts
General info about preparing the Job Scripts can be found at http://www.lrz.de/services/compute/linux-cluster/batch_parallel/
Submission can only be done from lxia4-1, lxia4-2.
For each of the three structures you will have to create a separate job script.
Here is an example that together with the info on the above stated LRZ page should give you an idea how to do it.
#!/bin/bash #SBATCH -o /home/hpc/pr32fi/lu32xul/test/info.out #SBATCH -D /home/hpc/pr32fi/lu32xul/test/ #SBATCH -J 1whz_MD #SBATCH --partition=mpp1_inter #SBATCH --get-user-env #SBATCH --ntasks=32 #SBATCH --mail-type=end #SBATCH --firstname.lastname@example.org #SBATCH --export=NONE #SBATCH --time=02:00:00 source /etc/profile.d/modules.sh module load gromacs export PATH="$HOME/test/AGroS:$PATH" export PATH="$HOME/apps/bin/:$PATH" AGroS 1whz_new.pdb -dir /home/hpc/pr32fi/lu32xul/test -threads 32
In this script we do not use the standard cluster
--clusters=mpp1 but a test queue to get a quicker answer whether the simulation works at all.
Submission is done using the following command
If the test simulation fails due to a gromacs problem try to use only 16 cores and change that also for the commandline call of AGroS.
In the real script you choose the standard cluster and instead of only 2 hours (limit) you set something like 16-32 hours depending on the size of your protein.
The state of the job and whether it really sits in the queue can be checked with the command
squeue -u <username> <queue> where the queue can either be
Once this all worked you have to wait and write a bit about the different steps of the simulation etc.
We also want you to look at the intermediate PDB files created in the workflow, visualize them and explain what is special, different about them and why we need them.