![]() |
|
||
|
|||
Using CHARMM on Biowulf
Cover scriptsA couple cover scripts are now available in /usr/local/bin for general use, setup to run several CHARMM versions for multiple sizes, with appropriate MPI libraries for ethernet and Myrinet. The commands are mpicharmm, which is designed to be used in scripts submitted via the PBS qsub command, and qcharmm, which invokes qsub and passes arguments to a secondary script (named runcharmm). The syntax of both is nearly identical, except that qcharmm has an additional required argument, which is the input file prefix; for that, the .inp extension is assumed, and the output file is created with the .out extension.Both cover scripts have required arguments; simply typing the name of either w/o args prints brief syntax messages. There is also a provision for the use of private CHARMM versions. The scripts may change over time as the Biowulf system and CHARMM evolve; changes are documented in the Revision Notes section.
qcharmmFor submitting single CHARMM jobs, qcharmm consolidates qsub, mpirun, and the chosen CHARMM executable into a single command interface. Examples:To submit and run the input min-prot.inp with the default version (c27b3) on ethernet with 2 procs, creating the log file min-prot.out qcharmm proc=2 prfx=min-prot As above, but c29b1 for 8 procs, Myrinet 2000, and the large size qcharmm proc=8 prfx=min-prot vrsn=c29b1 comm=myr2k size=large To submit an analysis run (bkbnrms.inp), passing the CHARMM variable ("parameter") N set to 50 to the input script (a reference to @N in the script is replaced by the number 50) qcharmm proc=single prfx=bkbnrms N:50 Finally, it should be noted that CHARMM does not perform well in parallel using TCP/IP over ethernet on the nodes running the 2.4.x Linux kernel (m2048, large memory; p1400, Athlon 1.4 MHz), so both qcharmm (and mpicharmm) will balk at running more than 2 parallel processes on these nodes. Nodes running the older kernel can be requested by using the optional node=linux22 arg of qcharmm, e.g.: qcharmm proc=4 prfx=min-prot size=large node=linux22 See the Biowulf Users Guide and the output of the shnodes command for more information on node memory sizes and usage.
mpicharmmThe other command, mpicharmm, is designed for use in scripts, and requires the user to know how to use PBS qsub. Like qcharmm, it allows CHARMM variables ("parameters") to be set on the command line via the colon syntax (e.g. N:50). A simple example is the following, called run.csh:#!/bin/csh #PBS -N rundyn #PBS -j oe #PBS -m ae # CHANGE TO SUBMISSION SUBDIR cd $PBS_O_WORKDIR # INVOKE CHARMM mpicharmm proc=2 < dyn.inp >& dyn.out The above script is run on Biowulf via the PBS command qsub -l nodes=1 run.csh Likewise, for 8 procs on Myrinet: #!/bin/csh #PBS -N rundyn #PBS -j oe #PBS -m ae # CHANGE TO SUBMISSION SUBDIR cd $PBS_O_WORKDIR # INVOKE CHARMM mpicharmm proc=8 comm=myr2k < dyn.inp >& dyn.out The above script is run on Biowulf via the PBS command qsub -l nodes=4:myr2k run.cshNote that you CANNOT pass arguments to run.csh via the 'qsub' command line; it is possible to pass variables, however, via the -v option to 'qsub'. For a detailed variable passing example, see the qcharmm and runcharmm scripts themselves (in /usr/local/bin). The CHARMM X11 graphics commands can also be used via mpicharmm; a pre-requisite is to login to biowulf using the ssh command. From some systems, the -X option to ssh is needed to setup use of X11, e.g. ssh -X biowulf The next step is to request an "interactive batch" node, via qsub -l nodes=1 -I -V Finally, start CHARMM via mpicharmm proc=single At this point, it's useful to have one or two CHARMM stream files that do all the system setup, e.g. read in RTF, PARAM, PSF and COOR file(s). Then one can simply type
* a title Graphics features include trajectory file viewing, and output of coordinate data as a file of POV-Ray objects.
Command syntax listingsThe following listings show the result of typing either command w/o any of the required arguments:
biobos ~ [7] qcharmm (qcharmm) ERROR: Arg prfx= is REQUIRED qcharmm prfx=STR [proc=N] [vrsn=STR] [size=STR] [comm=STR] [fort=STR] ARGS, DEFAULT INDICATED WITH *: prfx=STR CHARMM input file prefix; required proc=N N = integer no. of processors; defaults to single vrsn=STR version of CHARMM; c27b3, c28b1, c28b2, c29b1*, c29b2, c30b1, c31a1 fort=STR version of Fortran; g77, pgi* size=STR medium*, large, xlarge ---------- Node Properties ------------- comm=STR eth*, myr2k; communications interface type mmry=STR m512, m1024, m2048 node=STR node options, no default; p1400 p1800 linux22 (1) N = integer ; STR = character string (2) Equals sign required, case sensitive, no spaces Set the env var CHMEXE to use a private executable. Be sure to specify comm=myr2k for Myrinet 2000. Use proc=single or proc=0 for a single-threaded (non-MPI) version; default. Not all combinations are valid; see http://biowulf.nih.gov/charmm.html biobos ~ [8] mpicharmm (mpicharmm) ERROR: Arg proc= REQUIRED mpicharmm proc=N [vrsn=STR] [size=STR] [comm=STR] [fort=STR] ARGS, DEFAULT INDICATED WITH *: proc=N N = no. of processors; required vrsn=STR version of CHARMM; c27b3, c28b1, c28b2, c29b1*, c29b2, c30b1, c31a1 fort=STR version of Fortran; g77, pgi* size=STR medium*, large, xlarge comm=STR eth, myr2k; defaults to node type; obsolete (1) N = integer ; STR = character string (2) Equals sign required, case sensitive, no spaces Set the env var CHMEXE to use a private executable. Use proc=single or proc=0 for a single-threaded (non-MPI) version. Not all combinations are valid; see http://biowulf.nih.gov/charmm.html
Supported VersionsThe following listing indicates the currently available CHARMM versions, sizes with atom limits, and parallel communciations interfaces supported; c.f. the syntax listings above. The CHARMM versions are:
Locations of executable programs and support filesAcademic CHARMM executables with different features and sizes are available via /usr/local/charmm, as well as the doc, toppar, test, and support subdirs, e.g.:biobos /<2>charmm [13] ( cd /usr/local/charmm/c28b1 ; ls -F ) doc/ g77-lrg* g77-lrg.m2k* g77-lrg.one* g77-med* g77-med.m2k* g77-med.one* g77-xlg* g77-xlg.m2k* g77-xlg.one* pgi-lrg* pgi-lrg.m2k* pgi-lrg.one* pgi-med* pgi-med.m2k* pgi-med.one* pgi-xlg* pgi-xlg.m2k* pgi-xlg.one* support/ test/ toppar/The executables (indicated with *) are named according to the pattern COMPILER-SIZE for the g77 and PGI compilers, with med, lrg, and xlg for medium (25K atoms), large (60K atoms), and xlarge (240K atoms), respectively. The pgi-* versions are recommended, the code runs 20-30% faster; the g77 versions were kept for troubleshooting purposes. The .m2k extensions indicates the Myrinet MPI libraries were used instead of the standard TCP/IP MPI libraries used for ethernet. The .m2k versions should only be used for jobs submitted to Biowulf via the -l nodes=N:myr2k options to the PBS 'qsub' command (where N is some integer).
Running extended simulations via "daisy-chaining"Here is a more detailed usage example for mpicharmm illustrating the use of paired scripts to keep a simulation running for an extended period w/o intervention. Several points need to made about these scripts:
First, there's the initiating script, used to start the process, which we'll call biowulf.com: #!/bin/csh -f qsub -l nodes=4:myr2k dyn.cshMore important (and more complicated) is the dyn.csh script, which runs CHARMM, checks the output files to test for a successful run, renames each output file by adding a sequence number, and submits the next run to the queue. For a failed run, the output file is renamed to dyn.err.TS, where TS is a timestamp. This script has been extensively tested and used on many systems with appropriate modifcations, including LoBoS, Galaxy, and Biowulf. #! /bin/csh -f #PBS -N RunDyn #PBS -j oe # ASSUMPTION (1): input files are named dynXX.inp where XX is optional # useful for startup phase; for example, to start using dynstrt.inp, # dyn.csh strt # would be used; for production dynamics runs, simply dyn.csh # ASSUMPTION (2): output files are named dyn.res dyn.trj dyn.out # ASSUMPTION (3): previous restart file read as dyn.rea # cd $PBS_O_WORKDIR set chm = "mpicharmm proc=8 size=large" if ( -e next.seqno ) then $chm < dyn.inp >& dyn.out else $chm < dynstrt.inp >& dyn.out endif set okay = true # TEST FOR EXISTENCE, THEN NONZERO LENGTH OF OUTPUT FILES if ( -e dyn.res && -e dyn.trj ) then @ res = `wc dyn.res | awk '{print $1}'` @ tsz = `ls -s dyn.trj | awk '{print $1}'` @ nrm = `grep ' NORMAL TERMINATION ' dyn.out | wc -l` if ( $res > 100 && $tsz > 0 && $nrm == 1 ) then # SUCCESSFUL RUN; COPY RESTART FILE cp dyn.res dyn.rea # DETERMINE RUN NUMBER if ( -e next.seqno ) then @ i = `cat next.seqno` else @ i = 1 endif # ENABLE/DISABLE ANALYSIS EVERY 2ND RUN # if ( 0 == $i % 2 ) qsub -l nodes=1 chk.csh # NUMBER THE OUTPUT FILES; CHANGE EXTENSIONS TO SUIT APPLICATION foreach fil ( out res trj ) mv dyn.$fil dyn$i.$fil end gzip dyn$i.out dyn$i.res # CONDITIONAL END CHECK if ( -e last.seqno ) then @ l = `cat last.seqno` if ( $i == $l ) goto endrun endif # SUBMIT THE NEXT JOB ./biowulf.com endrun: @ i += 1 echo $i > next.seqno else # ZERO LENGTH FILE(S) set okay = false endif else # FILE DOESN'T EXIST set okay = false endif # TEST FOR CHARMM RUN FAILED; CREATE .ERR FILE WITH TIMESTAMP if ( $okay == true ) then else set ts = `date +%m%d.%H%M` mv dyn.out dyn.err.$ts exit(201) endifTo complete the setup, the CHARMM input files need to use the generic names; first, there is dynstrt.inp used only for the first run: * start NPAT melittin in dopc bilayer sim at high hydration * bomlev -1 stream inim2p1.str ! SETUP SHAKE shake fast bonh param open unit 41 write file name dyn.trj open unit 51 write card name dyn.res prnlev 3 NODE 0 dyna cpt start nstep 50000 timestep 0.001 - pcons pint pref 1.0 pmzz 2500. pmxx 0.0 pmyy 0.0 - hoover reft 303.15 tmass 20000. - inbfrq 20 atom vatom cutnb 14.0 ctofnb 11. cdie eps 1. - ctonnb 7. vswitch cutim 14.0 imgfrq 20 wmin 1.0 - ewald pmew fftx 48 ffty 48 fftz 64 kappa .34 spline order 6 - iprfrq 5000 ihtfrq 0 ieqfrq 0 ntrfrq 1000 - iuncrd 41 iunrea -1 iunwri 51 kunit -1 - nprint 100 nsavc 500 nsavv 0 ihbfrq 0 ilbfrq 0 - firstt 303.15 finalt 303.15 teminc 10.0 tstruct 303.15 tbath 303.15 - iasors 1 iasvel 1 iscvel 0 ichecw 0 twindh 5.0 twindl -5.0 stopFinally, there is the truly generic dyn.inp file, which is run repeatedly until the desired simulation time duration is achieved: * melittin in dopc lipid bilayer NPAT at high hydration * bomlev -1 stream inim2p1.str ! SETUP SHAKE shake fast bonh param open unit 31 read card name dyn.rea open unit 41 write file name dyn.trj open unit 51 write card name dyn.res prnlev 3 node 0 dyna cpt restart nstep 50000 timestep 0.001 - pcons pint pref 1.0 pmzz 2500.0 pmxx 0.0 pmyy 0.0 - inbfrq 20 atom vatom cutnb 14. ctofnb 11. cdie eps 1. - ctonnb 7. vswitch cutim 14.0 imgfrq 20 wmin 1.0 - ewald pmew fftx 48 ffty 48 fftz 64 kappa .34 spline order 6 - hoover reft 303.15 tmass 20000. - iprfrq 5000 ihtfrq 0 ieqfrq 0 ntrfrq 1000 - iuncrd 41 iunrea 31 iunwri 51 kunit -1 - nprint 100 nsavc 500 nsavv 0 ihbfrq 0 ilbfrq 0 - firstt 303.15 finalt 303.15 teminc 10.0 tstruct 303.15 - iasors 1 iasvel 1 iscvel 0 ichecw 0 twindh 5.0 twindl -5.0 stopA key thing to note is that dyn.rea is always the restart file created from the preceding run (see dyn.csh). An added feature is the use of gzip to compress output and restart files, saving considerable disk space. Compressed files may be browsed easily via e.g. zcat dyn2.out.gz | less
Using private executablesThere is also a provision to use both qcharmm and mpicharmm with a private CHARMM executable; vrsn=, size=, and fort= args are ignored in this case, but comm=myr2k is needed for Myrinet based executables run via qcharmm. For both, the env var CHMEXE must be set; for qcharmm, it must be set in ~/.cshrc, because the job will be run in a new shell when started by the PBS queueing system. It can be set in a script such as the above run.csh for mpicharmm. Note that the executable must be compiled using the proper MPI include and library files in order to use the high-speed Myrinet communications.
Revision NotesMarch 2001Problems related to the use of the Linux 2.4.0 kernel on new Biowulf nodes with 1 or 2 GB of RAM (m1024 or m2048) have lead to the following changes in the cover scripts:
qsub -l nodes=4:m512 run.csh April 2001Kernel changes, a new CHARMM version (c27b4), and the final installation of Myrinet 2000 resulted in the following changes to the cover scripts and documentation:
October 2001Changes to cover scripts, available versions, and this documentation:
June 2002Changes to cover scripts, available versions, and this documentation:
October 2002Changes to cover scripts, available versions, and this documentation:
August 2003Changes to cover scripts, available versions, and this documentation:
|
|||
Biowulf
Home Page: http://biowulf.nih.gov CHARMM .doc Page: http://biowulf.nih.gov/apps/charmm/charmmdoc/ This document prepared by Rick Venable, FDA/CBER Biophysics Lab 23 Aug 2003 (Rv) |