Compare commits

...

98 Commits

Author SHA1 Message Date
2106dce2b3 new fix halt command, other sundry small bug fixes 2016-10-18 13:11:40 -06:00
713b2af067 Merge pull request #221 from akohlmey/collected-bugfixes
Collected bugfixes and small changes
2016-10-18 12:58:25 -06:00
a752966eef Merge branch 'collected-bugfixes' of github.com:akohlmey/lammps into collected-bugfixes 2016-10-18 14:07:05 -04:00
1f7693faf2 fix typo 2016-10-18 14:06:43 -04:00
2c5ea9fc61 reapply changes to the timer class that got lost somehow 2016-10-18 14:05:08 -04:00
3e88fb5355 Merge branch 'master' into collected-bugfixes 2016-10-18 13:54:15 -04:00
6eadd45c45 Merge branch 'halt2' 2016-10-18 11:16:44 -06:00
003581c6a8 final changes to fix halt 2016-10-18 11:16:28 -06:00
1f1c87235a add support for "error continue" option, which resets the timer timeout setting after a soft exit 2016-10-18 12:21:27 -04:00
394c3bb504 add Timer::reset_timeout() API, that allows restoring the previous timeout setting after a forced timeout 2016-10-18 12:20:49 -04:00
954d536826 added bondmax method to fix halt 2016-10-18 09:27:01 -06:00
e4ff8128f1 fix typo
(cherry picked from commit c65fb0e311)
2016-10-16 12:29:53 -04:00
e7825ba21a import fix halt prototype code with corrections for soft exit 2016-10-15 07:30:07 -04:00
e77e1f6b8e replace complex solution to implement a soft exit with simpler one 2016-10-14 19:40:51 -04:00
909ec2c096 remove unused class members 2016-10-14 12:04:32 -04:00
f67975fd8a apply proper initialization and remove redundant variable declaration 2016-10-14 07:37:06 -04:00
deceb9d5c6 remove assignment without effect 2016-10-14 07:31:52 -04:00
c9c66ca0bd replace unsafe code 2016-10-14 07:28:25 -04:00
d07703efff re-apply bugfix from two years ago 2016-10-14 07:27:40 -04:00
411ecca8df plug memory leak 2016-10-14 07:27:14 -04:00
d11363c7eb Merge pull request #220 from rbberger/fix-doc-makefile
Allow building non-html doc targets without Python3 and virtualenv
2016-10-13 17:00:23 -06:00
5aefb2a882 Merge pull request #219 from akohlmey/python-no-double-load
do not require the LAMMPS shared library when loading the python wrapper from inside LAMMPS
2016-10-13 16:58:35 -06:00
40f2310a2a Merge pull request #218 from timattox/USER-DPD_fix_rx_init_bugfix
USER-DPD fix_rx initialization bugfix
2016-10-13 16:56:19 -06:00
2c8a7a318a bug fix for fix GCMC w/ fix shake, enhance of fix wall/gran/region with restarting 2016-10-13 16:55:53 -06:00
95cca1bd9f Allow building other targets without Python3 and virtualenv 2016-10-13 11:40:44 -04:00
0b426dadc1 do not require the LAMMPS shared library when loading the python wrapper from inside LAMMPS
Thanks to Giacomo Fiorin for figuring this out with NAMD/Colvars.
This requires linking with -Xlinker -export-dynamic or equivalent,
which is the default when using python-config to provide linker flags.
We will fall back to loading the DSO in case the initial load fails.
2016-10-12 18:36:38 -04:00
fcb5271026 USER-DPD: Initialize the dpdThetaLocal array consistently in fix_rx 2016-10-12 15:56:45 -04:00
4958e114ba USER-DPD bugfix: Properly initialize the local temperature averaging array. 2016-10-12 15:46:46 -04:00
63e71cd45b patch to add DPD-VV 2016-10-12 07:35:47 -06:00
4a5d9eaae2 Merge pull request #217 from akohlmey/small-fixes
Collected small changes and bugfixes
2016-10-12 07:32:08 -06:00
4e3a55047f Merge pull request #215 from timattox/USER-DPD_bugfix_for_dtsqrt
USER-DPD Bugfix: reset_dt() is not called when I thought it should be called.
2016-10-12 07:30:10 -06:00
f8a26dd158 update Timer::force_timeout() to trigger at next loop iteration 2016-10-12 07:26:03 -04:00
c24bf512f3 update #include statements for system includes 2016-10-12 00:00:53 -04:00
6b4ab0a390 update .gitignore 2016-10-12 00:00:21 -04:00
adc98e07df whitespace cleanup in USER-DPD 2016-10-11 23:58:36 -04:00
39a22039e9 correct broken link 2016-10-11 23:57:40 -04:00
b75860048b updates for recent changes to the manual 2016-10-11 23:50:45 -04:00
0eb7fbf34d tweaks to new USER-DPD docs 2016-10-11 15:43:59 -06:00
2f07a627a2 Forgot to remove my call to reset_dt() 2016-10-11 16:30:41 -04:00
559637f4bc USER-DPD Bugfix: reset_dt() is not called when I thought it should be called.
Move the calculation of dtsqrt inside FixShardlow::initial_integrate()
2016-10-11 16:11:29 -04:00
fbf7df14b5 Merge pull request #212 from timattox/USER-DPD_fix_eos_atom_style_checks
USER-DPD: Add atom_style compatibility checks in fix_eos_*.cpp files.
2016-10-11 13:40:00 -06:00
6f1162927a Merge pull request #207 from timattox/USER-DPD_new_VV_for_DPD
USER-DPD: add support for using VV with DPD
2016-10-11 13:39:25 -06:00
803dc57bfa Merge pull request #214 from akohlmey/make-no-lib-no-mpiio
make no-lib should also remove MPIIO and USER-LB packages
2016-10-11 12:42:53 -06:00
3e8e2911cc Merge pull request #213 from akohlmey/improper-virial-bugfixes
Improper virial bugfixes
2016-10-11 12:42:08 -06:00
04f5eadcf1 added LAST option to dump_modify thresh, more restart info printed out to screen 2016-10-11 12:39:52 -06:00
b00b40bccd make no-lib should also remove MPIIO and USER-LB packages 2016-10-11 08:03:59 -04:00
ef079ae4eb bugfix for AngleAngle term in CLASS2 impropers by Ivan A. Strelnikov, ICP RAS
this closes #56
2016-10-10 23:56:36 -04:00
bb0bfd508b Merge branch 'master' into improper-virial-bugfixes 2016-10-10 23:55:36 -04:00
e70d530c46 Merge pull request #203 from rbberger/txt2rst-external-link-fix
txt2rst external link fix
2016-10-10 13:59:27 -06:00
ed8cc82713 Merge pull request #211 from akohlmey/add-respa-to-fix-flow-gauss
Add respa support to fix flow/gauss
2016-10-10 13:59:01 -06:00
27dac02466 Merge pull request #209 from akohlmey/static-double-deallocation-workaround
workaround for double free issue when using USER-COLVARS with with lammps python wrapper and python package
2016-10-10 13:58:16 -06:00
467bcad0a0 Merge pull request #204 from rbberger/fix-user-omp
Migrate changes from GRANULAR to USER-OMP
2016-10-10 13:57:37 -06:00
a2b0840064 USER-DPD: Add atom_style compatibility checks in fix_eos_*.cpp files. 2016-10-10 13:40:33 -04:00
144e6a8091 whitespace cleanup 2016-10-10 09:40:09 -04:00
72ac073412 edited documentation
(cherry picked from commit eff14c74b0)
2016-10-10 09:38:54 -04:00
49c45ab03b edited documentation
(cherry picked from commit fd560889c3)
2016-10-10 09:38:53 -04:00
c2cd439944 first draft of documentation for respa
(cherry picked from commit d7dcbcfbd9)
2016-10-10 09:38:53 -04:00
e96ebb29bc adjusted default respa level to be outermost
(cherry picked from commit 7fc4d46a41)
2016-10-10 09:38:53 -04:00
3ce178d43f now understand how respa works in lammps
(cherry picked from commit c829027e83)
2016-10-10 09:38:52 -04:00
23781d6ec9 added respa to fix_flow_gauss, not fully understood yet
(cherry picked from commit 8d9737b04d)
2016-10-10 09:38:52 -04:00
fca6d721c0 completed synchronization with non-threaded version 2016-10-10 09:16:21 -04:00
dd192ca7ea whitespace cleanup 2016-10-10 09:15:42 -04:00
683689c808 revert to previous style conventions for size_t constants 2016-10-08 11:00:23 -04:00
e01e90eb96 workaround for double free issue when using USER-COLVARS with lammps code loaded as shared library into a standalone executable 2016-10-08 10:45:22 -04:00
9507a786f0 USER-DPD: whitespace and indentation fixes 2016-10-07 15:59:47 -04:00
9789f047d7 USER-DPD: update the USER/dpd examples and their reference outputs 2016-10-07 15:55:35 -04:00
e27ed6c94a USER-DPD: Added support to use VV integrator with USER-DPD if desired.
Includes documentation and examples.
NOTE: VV requires very small timesteps under isoenergetic conditions.
Consider using fix_shardlow instead, since this VV support is
primarily for comparison purposes.
2016-10-07 15:03:30 -04:00
615a2da044 Migrate changes from GRANULAR to USER-OMP 2016-10-06 21:48:06 -04:00
7f3a7c5cbe Fix broken link 2016-10-06 20:33:24 -04:00
e78b4267b7 Fix issue with external links containing anchors 2016-10-06 20:29:07 -04:00
e9fed80928 Merge pull request #202 from akohlmey/doc-formatting-fixes
collected documentation updates and corrections from LAMMPS-ICMS
2016-10-06 15:49:44 -06:00
54fc194e5b Merge pull request #199 from akohlmey/small-changes
Collected small changes and bugfixes
2016-10-06 15:49:24 -06:00
b3d2fb91bb new fix wall/gran/region command, REBO bug fix, new example log files 2016-10-06 15:47:41 -06:00
19984c9bd1 Revert "bugfix for AngleAngle term in CLASS2 impropers by Ivan A. Strelnikov, ICP RAS"
This reverts commit 83bcdb6a50.
2016-10-06 17:23:10 -04:00
f92618a33b Revert "bugfix for virial tally for improper style umbrella from Steven Vandenbrande (U Gent)"
This reverts commit 4921dc18a0.
2016-10-06 17:21:38 -04:00
887981cfaa bugfix for virial tally for improper style umbrella from Steven Vandenbrande (U Gent)
this closes #182

(cherry picked from commit 4921dc18a0)
2016-10-06 17:20:22 -04:00
0b5d71537a collected documentation updates and corrections from LAMMPS-ICMS
fixes formatting issues due to tabs, permission issues and
a few typos and badly worded text.
2016-10-06 15:48:18 -04:00
c213457550 Merge pull request #197 from giacomofiorin/colvars_2016-10-05
Colvars 2016-10-05
2016-10-06 13:02:52 -06:00
0f45cd61a5 Merge pull request #196 from akohlmey/charmm-cmap-updates
Some more cmap-related updates for ch2lmp
2016-10-06 13:02:27 -06:00
493873fb93 clean up doc src 2016-10-06 13:00:46 -06:00
60a031ebac Merge branch 'USER-DPD_pair_exp6_rx_mathfix' of https://github.com/timattox/lammps_USER-DPD into small-changes
This closes #201
2016-10-06 14:28:08 -04:00
27e76a70b9 Merge branch 'USER-DPD_hybrid_atom_bugfix' of https://github.com/timattox/lammps_USER-DPD into small-changes
This closes #200
2016-10-06 14:27:27 -04:00
e1e9a5c126 USER-DPD: math corrections in pair_exp6_rx.cpp (by Jim Larentzos) 2016-10-06 13:49:47 -04:00
d31121b18c USER-DPD: bugfix in unpack_comm_hybrid(); now works with hybrid atom style 2016-10-06 13:21:27 -04:00
0853cdbe6f update reference data files for updated/corrected clayff parameters 2016-10-06 11:47:08 -04:00
83bcdb6a50 bugfix for AngleAngle term in CLASS2 impropers by Ivan A. Strelnikov, ICP RAS
this closes #56
2016-10-06 11:27:18 -04:00
22ce671804 improved whitespace handling in msi2lmp for force fields and topologies 2016-10-06 11:16:59 -04:00
4921dc18a0 bugfix for virial tally for improper style umbrella from Steven Vandenbrande (U Gent)
this closes #182
2016-10-06 10:47:08 -04:00
d133167bf6 Merge branch 'master' of https://github.com/albapa/lammps into small-changes
USER-QUIP related improvements from github user albapa. This closes #198
2016-10-06 09:32:50 -04:00
8ea063378e add NETCDF libs (as defined in QUIP) to the linking line if QUIP was built with NETCDF support 2016-10-06 12:16:25 +01:00
fd16118cbb removed dump_modify command 2016-10-06 12:02:41 +01:00
f9f955d5b5 update include statement format 2016-10-05 22:34:44 -04:00
d7d321a512 some more updates to the README file to reflect the inclusion of the CMAP example and renamed file names 2016-10-05 18:41:45 -04:00
8809a603fb Colvars update: issue a warning that cannot be ignored regarding total forces 2016-10-05 18:26:21 -04:00
969d3cf4b0 Colvars update: make ABF check that the colvar isn't using already subtractAppliedForce 2016-10-05 18:25:40 -04:00
326fdf2cf1 added 1GB1 example from Robert Latour and update 1AC7 example files 2016-10-05 18:20:09 -04:00
f32819dd10 added tweak to write out the command line used for the conversion to the beginning of the LAMMPS input 2016-10-05 18:13:46 -04:00
c07a01c661 import updated README file for charmm2lammps.pl with CMAP support 2016-10-05 18:11:52 -04:00
746 changed files with 220033 additions and 136047 deletions

View File

@ -1,4 +1,4 @@
LAMMPS (15 Feb 2016)
LAMMPS (6 Oct 2016)
# FENE beadspring benchmark
units lj
@ -43,25 +43,25 @@ Neighbor list info ...
master list distance cutoff = 1.52
ghost atom cutoff = 1.52
binsize = 0.76 -> bins = 45 45 45
Memory usage per processor = 11.5189 Mbytes
Memory usage per processor = 12.0423 Mbytes
Step Temp E_pair E_mol TotEng Press
0 0.97029772 0.44484087 20.494523 22.394765 4.6721833
100 0.9729966 0.4361122 20.507698 22.40326 4.6548819
Loop time of 0.978585 on 1 procs for 100 steps with 32000 atoms
Loop time of 0.977647 on 1 procs for 100 steps with 32000 atoms
Performance: 105948.895 tau/day, 102.188 timesteps/s
100.0% CPU use with 1 MPI tasks x no OpenMP threads
Performance: 106050.541 tau/day, 102.286 timesteps/s
99.9% CPU use with 1 MPI tasks x no OpenMP threads
MPI task timing breakdown:
Section | min time | avg time | max time |%varavg| %total
---------------------------------------------------------------
Pair | 0.19562 | 0.19562 | 0.19562 | 0.0 | 19.99
Bond | 0.087475 | 0.087475 | 0.087475 | 0.0 | 8.94
Neigh | 0.44861 | 0.44861 | 0.44861 | 0.0 | 45.84
Comm | 0.032932 | 0.032932 | 0.032932 | 0.0 | 3.37
Output | 0.00010395 | 0.00010395 | 0.00010395 | 0.0 | 0.01
Modify | 0.19413 | 0.19413 | 0.19413 | 0.0 | 19.84
Other | | 0.01972 | | | 2.02
Pair | 0.19421 | 0.19421 | 0.19421 | 0.0 | 19.86
Bond | 0.08741 | 0.08741 | 0.08741 | 0.0 | 8.94
Neigh | 0.45791 | 0.45791 | 0.45791 | 0.0 | 46.84
Comm | 0.032649 | 0.032649 | 0.032649 | 0.0 | 3.34
Output | 0.00012207 | 0.00012207 | 0.00012207 | 0.0 | 0.01
Modify | 0.18071 | 0.18071 | 0.18071 | 0.0 | 18.48
Other | | 0.02464 | | | 2.52
Nlocal: 32000 ave 32000 max 32000 min
Histogram: 1 0 0 0 0 0 0 0 0 0

View File

@ -1,4 +1,4 @@
LAMMPS (15 Feb 2016)
LAMMPS (6 Oct 2016)
# FENE beadspring benchmark
units lj
@ -43,25 +43,25 @@ Neighbor list info ...
master list distance cutoff = 1.52
ghost atom cutoff = 1.52
binsize = 0.76 -> bins = 45 45 45
Memory usage per processor = 3.91518 Mbytes
Memory usage per processor = 4.14663 Mbytes
Step Temp E_pair E_mol TotEng Press
0 0.97029772 0.44484087 20.494523 22.394765 4.6721833
100 0.97145835 0.43803883 20.502691 22.397872 4.626988
Loop time of 0.271187 on 4 procs for 100 steps with 32000 atoms
Loop time of 0.269205 on 4 procs for 100 steps with 32000 atoms
Performance: 382319.453 tau/day, 368.749 timesteps/s
99.6% CPU use with 4 MPI tasks x no OpenMP threads
Performance: 385133.446 tau/day, 371.464 timesteps/s
99.8% CPU use with 4 MPI tasks x no OpenMP threads
MPI task timing breakdown:
Section | min time | avg time | max time |%varavg| %total
---------------------------------------------------------------
Pair | 0.048621 | 0.050076 | 0.051229 | 0.4 | 18.47
Bond | 0.022254 | 0.022942 | 0.023567 | 0.3 | 8.46
Neigh | 0.11873 | 0.11881 | 0.11887 | 0.0 | 43.81
Comm | 0.019066 | 0.021357 | 0.024297 | 1.3 | 7.88
Output | 5.0068e-05 | 5.5015e-05 | 6.1035e-05 | 0.1 | 0.02
Modify | 0.048737 | 0.050198 | 0.051231 | 0.4 | 18.51
Other | | 0.007751 | | | 2.86
Pair | 0.049383 | 0.049756 | 0.049988 | 0.1 | 18.48
Bond | 0.022701 | 0.022813 | 0.022872 | 0.0 | 8.47
Neigh | 0.11982 | 0.12002 | 0.12018 | 0.0 | 44.58
Comm | 0.020274 | 0.021077 | 0.022348 | 0.5 | 7.83
Output | 5.3167e-05 | 5.6148e-05 | 6.3181e-05 | 0.1 | 0.02
Modify | 0.046276 | 0.046809 | 0.047016 | 0.1 | 17.39
Other | | 0.008669 | | | 3.22
Nlocal: 8000 ave 8030 max 7974 min
Histogram: 1 0 0 1 0 1 0 0 0 1

View File

@ -1,4 +1,4 @@
LAMMPS (15 Feb 2016)
LAMMPS (6 Oct 2016)
# FENE beadspring benchmark
variable x index 1
@ -59,25 +59,25 @@ Neighbor list info ...
master list distance cutoff = 1.52
ghost atom cutoff = 1.52
binsize = 0.76 -> bins = 89 89 45
Memory usage per processor = 12.8735 Mbytes
Memory usage per processor = 13.2993 Mbytes
Step Temp E_pair E_mol TotEng Press
0 0.97027498 0.44484087 20.494523 22.394765 4.6721833
100 0.97682955 0.44239968 20.500229 22.407862 4.6527025
Loop time of 1.20889 on 4 procs for 100 steps with 128000 atoms
Loop time of 1.14845 on 4 procs for 100 steps with 128000 atoms
Performance: 85764.410 tau/day, 82.720 timesteps/s
99.8% CPU use with 4 MPI tasks x no OpenMP threads
Performance: 90277.919 tau/day, 87.074 timesteps/s
99.9% CPU use with 4 MPI tasks x no OpenMP threads
MPI task timing breakdown:
Section | min time | avg time | max time |%varavg| %total
---------------------------------------------------------------
Pair | 0.21738 | 0.23306 | 0.23926 | 1.9 | 19.28
Bond | 0.094536 | 0.10196 | 0.10534 | 1.4 | 8.43
Neigh | 0.52311 | 0.52392 | 0.52519 | 0.1 | 43.34
Comm | 0.090161 | 0.10022 | 0.12557 | 4.7 | 8.29
Output | 0.00012207 | 0.00017327 | 0.00019598 | 0.2 | 0.01
Modify | 0.19662 | 0.20262 | 0.20672 | 0.8 | 16.76
Other | | 0.04694 | | | 3.88
Pair | 0.2203 | 0.22207 | 0.22386 | 0.3 | 19.34
Bond | 0.094861 | 0.095302 | 0.095988 | 0.1 | 8.30
Neigh | 0.52127 | 0.5216 | 0.52189 | 0.0 | 45.42
Comm | 0.079585 | 0.082159 | 0.084366 | 0.7 | 7.15
Output | 0.00013304 | 0.00015306 | 0.00018501 | 0.2 | 0.01
Modify | 0.18351 | 0.18419 | 0.1856 | 0.2 | 16.04
Other | | 0.04298 | | | 3.74
Nlocal: 32000 ave 32015 max 31983 min
Histogram: 1 0 1 0 0 0 0 0 1 1

View File

@ -1,4 +1,4 @@
LAMMPS (15 Feb 2016)
LAMMPS (6 Oct 2016)
# LAMMPS benchmark of granular flow
# chute flow of 32000 atoms with frozen base at 26 degrees
@ -47,24 +47,24 @@ Neighbor list info ...
master list distance cutoff = 1.1
ghost atom cutoff = 1.1
binsize = 0.55 -> bins = 73 37 68
Memory usage per processor = 15.567 Mbytes
Step Atoms KinEng 1 Volume
Memory usage per processor = 16.0904 Mbytes
Step Atoms KinEng c_1 Volume
0 32000 784139.13 1601.1263 29833.783
100 32000 784292.08 1571.0968 29834.707
Loop time of 0.550482 on 1 procs for 100 steps with 32000 atoms
Loop time of 0.534174 on 1 procs for 100 steps with 32000 atoms
Performance: 1569.534 tau/day, 181.659 timesteps/s
100.1% CPU use with 1 MPI tasks x no OpenMP threads
Performance: 1617.451 tau/day, 187.205 timesteps/s
99.8% CPU use with 1 MPI tasks x no OpenMP threads
MPI task timing breakdown:
Section | min time | avg time | max time |%varavg| %total
---------------------------------------------------------------
Pair | 0.33849 | 0.33849 | 0.33849 | 0.0 | 61.49
Neigh | 0.040353 | 0.040353 | 0.040353 | 0.0 | 7.33
Comm | 0.018023 | 0.018023 | 0.018023 | 0.0 | 3.27
Output | 0.00020385 | 0.00020385 | 0.00020385 | 0.0 | 0.04
Modify | 0.13155 | 0.13155 | 0.13155 | 0.0 | 23.90
Other | | 0.02186 | | | 3.97
Pair | 0.33346 | 0.33346 | 0.33346 | 0.0 | 62.43
Neigh | 0.043902 | 0.043902 | 0.043902 | 0.0 | 8.22
Comm | 0.018391 | 0.018391 | 0.018391 | 0.0 | 3.44
Output | 0.00022411 | 0.00022411 | 0.00022411 | 0.0 | 0.04
Modify | 0.11666 | 0.11666 | 0.11666 | 0.0 | 21.84
Other | | 0.02153 | | | 4.03
Nlocal: 32000 ave 32000 max 32000 min
Histogram: 1 0 0 0 0 0 0 0 0 0

View File

@ -1,4 +1,4 @@
LAMMPS (15 Feb 2016)
LAMMPS (6 Oct 2016)
# LAMMPS benchmark of granular flow
# chute flow of 32000 atoms with frozen base at 26 degrees
@ -47,24 +47,24 @@ Neighbor list info ...
master list distance cutoff = 1.1
ghost atom cutoff = 1.1
binsize = 0.55 -> bins = 73 37 68
Memory usage per processor = 6.81783 Mbytes
Step Atoms KinEng 1 Volume
Memory usage per processor = 7.04927 Mbytes
Step Atoms KinEng c_1 Volume
0 32000 784139.13 1601.1263 29833.783
100 32000 784292.08 1571.0968 29834.707
Loop time of 0.13141 on 4 procs for 100 steps with 32000 atoms
Loop time of 0.171815 on 4 procs for 100 steps with 32000 atoms
Performance: 6574.833 tau/day, 760.976 timesteps/s
99.3% CPU use with 4 MPI tasks x no OpenMP threads
Performance: 5028.653 tau/day, 582.020 timesteps/s
99.7% CPU use with 4 MPI tasks x no OpenMP threads
MPI task timing breakdown:
Section | min time | avg time | max time |%varavg| %total
---------------------------------------------------------------
Pair | 0.062505 | 0.067 | 0.07152 | 1.5 | 50.99
Neigh | 0.010041 | 0.0101 | 0.010178 | 0.1 | 7.69
Comm | 0.012347 | 0.012895 | 0.013444 | 0.5 | 9.81
Output | 6.3896e-05 | 0.00010294 | 0.00014091 | 0.3 | 0.08
Modify | 0.031802 | 0.032348 | 0.032897 | 0.3 | 24.62
Other | | 0.008965 | | | 6.82
Pair | 0.093691 | 0.096898 | 0.10005 | 0.8 | 56.40
Neigh | 0.011976 | 0.012059 | 0.012146 | 0.1 | 7.02
Comm | 0.016384 | 0.017418 | 0.018465 | 0.8 | 10.14
Output | 7.7963e-05 | 0.00010747 | 0.00013304 | 0.2 | 0.06
Modify | 0.031744 | 0.031943 | 0.032167 | 0.1 | 18.59
Other | | 0.01339 | | | 7.79
Nlocal: 8000 ave 8008 max 7992 min
Histogram: 2 0 0 0 0 0 0 0 0 2

View File

@ -1,4 +1,4 @@
LAMMPS (15 Feb 2016)
LAMMPS (6 Oct 2016)
# LAMMPS benchmark of granular flow
# chute flow of 32000 atoms with frozen base at 26 degrees
@ -57,24 +57,24 @@ Neighbor list info ...
master list distance cutoff = 1.1
ghost atom cutoff = 1.1
binsize = 0.55 -> bins = 146 73 68
Memory usage per processor = 15.7007 Mbytes
Step Atoms KinEng 1 Volume
Memory usage per processor = 16.1265 Mbytes
Step Atoms KinEng c_1 Volume
0 128000 3136556.5 6404.5051 119335.13
100 128000 3137168.3 6284.3873 119338.83
Loop time of 0.906913 on 4 procs for 100 steps with 128000 atoms
Loop time of 0.832365 on 4 procs for 100 steps with 128000 atoms
Performance: 952.683 tau/day, 110.264 timesteps/s
99.7% CPU use with 4 MPI tasks x no OpenMP threads
Performance: 1038.006 tau/day, 120.140 timesteps/s
99.8% CPU use with 4 MPI tasks x no OpenMP threads
MPI task timing breakdown:
Section | min time | avg time | max time |%varavg| %total
---------------------------------------------------------------
Pair | 0.51454 | 0.53094 | 0.55381 | 2.0 | 58.54
Neigh | 0.042597 | 0.043726 | 0.045801 | 0.6 | 4.82
Comm | 0.063027 | 0.064657 | 0.067367 | 0.7 | 7.13
Output | 0.00024891 | 0.00059718 | 0.00086498 | 1.0 | 0.07
Modify | 0.16508 | 0.17656 | 0.1925 | 2.6 | 19.47
Other | | 0.09043 | | | 9.97
Pair | 0.5178 | 0.52208 | 0.52793 | 0.5 | 62.72
Neigh | 0.047003 | 0.047113 | 0.047224 | 0.0 | 5.66
Comm | 0.05233 | 0.052988 | 0.053722 | 0.2 | 6.37
Output | 0.00024986 | 0.00032717 | 0.00036693 | 0.3 | 0.04
Modify | 0.15517 | 0.15627 | 0.15808 | 0.3 | 18.77
Other | | 0.0536 | | | 6.44
Nlocal: 32000 ave 32000 max 32000 min
Histogram: 4 0 0 0 0 0 0 0 0 0
@ -87,4 +87,4 @@ Total # of neighbors = 460532
Ave neighs/atom = 3.59791
Neighbor list builds = 2
Dangerous builds = 0
Total wall time: 0:00:01
Total wall time: 0:00:00

View File

@ -1,4 +1,4 @@
LAMMPS (15 Feb 2016)
LAMMPS (6 Oct 2016)
# bulk Cu lattice
variable x index 1
@ -49,25 +49,25 @@ Neighbor list info ...
master list distance cutoff = 5.95
ghost atom cutoff = 5.95
binsize = 2.975 -> bins = 25 25 25
Memory usage per processor = 10.2238 Mbytes
Memory usage per processor = 11.2238 Mbytes
Step Temp E_pair E_mol TotEng Press
0 1600 -113280 0 -106662.09 18703.573
50 781.69049 -109873.35 0 -106640.13 52273.088
100 801.832 -109957.3 0 -106640.77 51322.821
Loop time of 5.90097 on 1 procs for 100 steps with 32000 atoms
Loop time of 5.96529 on 1 procs for 100 steps with 32000 atoms
Performance: 7.321 ns/day, 3.278 hours/ns, 16.946 timesteps/s
Performance: 7.242 ns/day, 3.314 hours/ns, 16.764 timesteps/s
99.9% CPU use with 1 MPI tasks x no OpenMP threads
MPI task timing breakdown:
Section | min time | avg time | max time |%varavg| %total
---------------------------------------------------------------
Pair | 5.2121 | 5.2121 | 5.2121 | 0.0 | 88.33
Neigh | 0.58212 | 0.58212 | 0.58212 | 0.0 | 9.86
Comm | 0.030392 | 0.030392 | 0.030392 | 0.0 | 0.52
Output | 0.00023389 | 0.00023389 | 0.00023389 | 0.0 | 0.00
Modify | 0.060871 | 0.060871 | 0.060871 | 0.0 | 1.03
Other | | 0.01527 | | | 0.26
Pair | 5.2743 | 5.2743 | 5.2743 | 0.0 | 88.42
Neigh | 0.59212 | 0.59212 | 0.59212 | 0.0 | 9.93
Comm | 0.030399 | 0.030399 | 0.030399 | 0.0 | 0.51
Output | 0.00026202 | 0.00026202 | 0.00026202 | 0.0 | 0.00
Modify | 0.050487 | 0.050487 | 0.050487 | 0.0 | 0.85
Other | | 0.01776 | | | 0.30
Nlocal: 32000 ave 32000 max 32000 min
Histogram: 1 0 0 0 0 0 0 0 0 0

View File

@ -1,4 +1,4 @@
LAMMPS (15 Feb 2016)
LAMMPS (6 Oct 2016)
# bulk Cu lattice
variable x index 1
@ -49,25 +49,25 @@ Neighbor list info ...
master list distance cutoff = 5.95
ghost atom cutoff = 5.95
binsize = 2.975 -> bins = 25 25 25
Memory usage per processor = 5.09629 Mbytes
Memory usage per processor = 5.59629 Mbytes
Step Temp E_pair E_mol TotEng Press
0 1600 -113280 0 -106662.09 18703.573
50 781.69049 -109873.35 0 -106640.13 52273.088
100 801.832 -109957.3 0 -106640.77 51322.821
Loop time of 1.58019 on 4 procs for 100 steps with 32000 atoms
Loop time of 1.64562 on 4 procs for 100 steps with 32000 atoms
Performance: 27.338 ns/day, 0.878 hours/ns, 63.284 timesteps/s
Performance: 26.252 ns/day, 0.914 hours/ns, 60.767 timesteps/s
99.8% CPU use with 4 MPI tasks x no OpenMP threads
MPI task timing breakdown:
Section | min time | avg time | max time |%varavg| %total
---------------------------------------------------------------
Pair | 1.3617 | 1.366 | 1.3723 | 0.4 | 86.45
Neigh | 0.15123 | 0.15232 | 0.15374 | 0.2 | 9.64
Comm | 0.033429 | 0.041275 | 0.047066 | 2.7 | 2.61
Output | 0.00011301 | 0.0001573 | 0.000211 | 0.3 | 0.01
Modify | 0.014694 | 0.015085 | 0.015421 | 0.2 | 0.95
Other | | 0.005342 | | | 0.34
Pair | 1.408 | 1.4175 | 1.4341 | 0.9 | 86.14
Neigh | 0.15512 | 0.15722 | 0.16112 | 0.6 | 9.55
Comm | 0.029105 | 0.049986 | 0.061822 | 5.8 | 3.04
Output | 0.00010991 | 0.00011539 | 0.00012302 | 0.0 | 0.01
Modify | 0.013383 | 0.013573 | 0.013883 | 0.2 | 0.82
Other | | 0.007264 | | | 0.44
Nlocal: 8000 ave 8008 max 7993 min
Histogram: 2 0 0 0 0 0 0 0 1 1

View File

@ -1,4 +1,4 @@
LAMMPS (15 Feb 2016)
LAMMPS (6 Oct 2016)
# bulk Cu lattice
variable x index 1
@ -49,25 +49,25 @@ Neighbor list info ...
master list distance cutoff = 5.95
ghost atom cutoff = 5.95
binsize = 2.975 -> bins = 49 49 25
Memory usage per processor = 10.1402 Mbytes
Memory usage per processor = 11.1402 Mbytes
Step Temp E_pair E_mol TotEng Press
0 1600 -453120 0 -426647.73 18704.012
50 779.50001 -439457.02 0 -426560.06 52355.276
100 797.97828 -439764.76 0 -426562.07 51474.74
Loop time of 6.46849 on 4 procs for 100 steps with 128000 atoms
Loop time of 6.60121 on 4 procs for 100 steps with 128000 atoms
Performance: 6.679 ns/day, 3.594 hours/ns, 15.460 timesteps/s
Performance: 6.544 ns/day, 3.667 hours/ns, 15.149 timesteps/s
99.9% CPU use with 4 MPI tasks x no OpenMP threads
MPI task timing breakdown:
Section | min time | avg time | max time |%varavg| %total
---------------------------------------------------------------
Pair | 5.581 | 5.5997 | 5.6265 | 0.8 | 86.57
Neigh | 0.65287 | 0.658 | 0.66374 | 0.5 | 10.17
Comm | 0.075706 | 0.11015 | 0.13655 | 7.2 | 1.70
Output | 0.00026488 | 0.00028312 | 0.00029302 | 0.1 | 0.00
Modify | 0.069607 | 0.072407 | 0.074555 | 0.7 | 1.12
Other | | 0.02794 | | | 0.43
Pair | 5.6676 | 5.7011 | 5.7469 | 1.3 | 86.36
Neigh | 0.66423 | 0.67119 | 0.68082 | 0.7 | 10.17
Comm | 0.079367 | 0.13668 | 0.1791 | 10.5 | 2.07
Output | 0.00026989 | 0.00028622 | 0.00031209 | 0.1 | 0.00
Modify | 0.060046 | 0.062203 | 0.065009 | 0.9 | 0.94
Other | | 0.02974 | | | 0.45
Nlocal: 32000 ave 32092 max 31914 min
Histogram: 1 0 0 1 0 1 0 0 0 1

View File

@ -1,4 +1,4 @@
LAMMPS (15 Feb 2016)
LAMMPS (6 Oct 2016)
# 3d Lennard-Jones melt
variable x index 1
@ -50,20 +50,20 @@ Memory usage per processor = 8.21387 Mbytes
Step Temp E_pair E_mol TotEng Press
0 1.44 -6.7733681 0 -4.6134356 -5.0197073
100 0.7574531 -5.7585055 0 -4.6223613 0.20726105
Loop time of 2.26309 on 1 procs for 100 steps with 32000 atoms
Loop time of 2.26185 on 1 procs for 100 steps with 32000 atoms
Performance: 19088.920 tau/day, 44.187 timesteps/s
Performance: 19099.377 tau/day, 44.212 timesteps/s
99.9% CPU use with 1 MPI tasks x no OpenMP threads
MPI task timing breakdown:
Section | min time | avg time | max time |%varavg| %total
---------------------------------------------------------------
Pair | 1.9341 | 1.9341 | 1.9341 | 0.0 | 85.46
Neigh | 0.2442 | 0.2442 | 0.2442 | 0.0 | 10.79
Comm | 0.024158 | 0.024158 | 0.024158 | 0.0 | 1.07
Output | 0.00011611 | 0.00011611 | 0.00011611 | 0.0 | 0.01
Modify | 0.053222 | 0.053222 | 0.053222 | 0.0 | 2.35
Other | | 0.007258 | | | 0.32
Pair | 1.9328 | 1.9328 | 1.9328 | 0.0 | 85.45
Neigh | 0.2558 | 0.2558 | 0.2558 | 0.0 | 11.31
Comm | 0.024061 | 0.024061 | 0.024061 | 0.0 | 1.06
Output | 0.00012612 | 0.00012612 | 0.00012612 | 0.0 | 0.01
Modify | 0.040887 | 0.040887 | 0.040887 | 0.0 | 1.81
Other | | 0.008214 | | | 0.36
Nlocal: 32000 ave 32000 max 32000 min
Histogram: 1 0 0 0 0 0 0 0 0 0

View File

@ -1,4 +1,4 @@
LAMMPS (15 Feb 2016)
LAMMPS (6 Oct 2016)
# 3d Lennard-Jones melt
variable x index 1
@ -50,20 +50,20 @@ Memory usage per processor = 4.09506 Mbytes
Step Temp E_pair E_mol TotEng Press
0 1.44 -6.7733681 0 -4.6134356 -5.0197073
100 0.7574531 -5.7585055 0 -4.6223613 0.20726105
Loop time of 0.640733 on 4 procs for 100 steps with 32000 atoms
Loop time of 0.635957 on 4 procs for 100 steps with 32000 atoms
Performance: 67422.779 tau/day, 156.071 timesteps/s
99.7% CPU use with 4 MPI tasks x no OpenMP threads
Performance: 67929.172 tau/day, 157.243 timesteps/s
99.9% CPU use with 4 MPI tasks x no OpenMP threads
MPI task timing breakdown:
Section | min time | avg time | max time |%varavg| %total
---------------------------------------------------------------
Pair | 0.49487 | 0.51733 | 0.5322 | 1.9 | 80.74
Neigh | 0.061131 | 0.063685 | 0.065433 | 0.6 | 9.94
Comm | 0.02457 | 0.042349 | 0.069598 | 8.1 | 6.61
Output | 5.9843e-05 | 6.3181e-05 | 6.6996e-05 | 0.0 | 0.01
Modify | 0.012961 | 0.013863 | 0.014491 | 0.5 | 2.16
Other | | 0.003448 | | | 0.54
Pair | 0.51335 | 0.51822 | 0.52569 | 0.7 | 81.49
Neigh | 0.063695 | 0.064309 | 0.065397 | 0.3 | 10.11
Comm | 0.027525 | 0.03629 | 0.041959 | 3.1 | 5.71
Output | 6.3896e-05 | 6.6698e-05 | 7.081e-05 | 0.0 | 0.01
Modify | 0.012472 | 0.01254 | 0.012618 | 0.1 | 1.97
Other | | 0.004529 | | | 0.71
Nlocal: 8000 ave 8037 max 7964 min
Histogram: 2 0 0 0 0 0 0 0 1 1

View File

@ -1,4 +1,4 @@
LAMMPS (15 Feb 2016)
LAMMPS (6 Oct 2016)
# 3d Lennard-Jones melt
variable x index 1
@ -50,20 +50,20 @@ Memory usage per processor = 8.13678 Mbytes
Step Temp E_pair E_mol TotEng Press
0 1.44 -6.7733681 0 -4.6133849 -5.0196788
100 0.75841891 -5.759957 0 -4.6223375 0.20008866
Loop time of 2.57914 on 4 procs for 100 steps with 128000 atoms
Loop time of 2.55762 on 4 procs for 100 steps with 128000 atoms
Performance: 16749.768 tau/day, 38.773 timesteps/s
Performance: 16890.677 tau/day, 39.099 timesteps/s
99.8% CPU use with 4 MPI tasks x no OpenMP threads
MPI task timing breakdown:
Section | min time | avg time | max time |%varavg| %total
---------------------------------------------------------------
Pair | 2.042 | 2.1092 | 2.1668 | 3.1 | 81.78
Neigh | 0.23982 | 0.24551 | 0.25233 | 1.0 | 9.52
Comm | 0.067088 | 0.13887 | 0.22681 | 15.7 | 5.38
Output | 0.00013185 | 0.00021666 | 0.00027108 | 0.4 | 0.01
Modify | 0.060348 | 0.071269 | 0.077063 | 2.5 | 2.76
Other | | 0.01403 | | | 0.54
Pair | 2.0583 | 2.0988 | 2.1594 | 2.6 | 82.06
Neigh | 0.24411 | 0.24838 | 0.25585 | 0.9 | 9.71
Comm | 0.066397 | 0.13872 | 0.1863 | 11.9 | 5.42
Output | 0.00012994 | 0.00021023 | 0.00025702 | 0.3 | 0.01
Modify | 0.055533 | 0.058343 | 0.061791 | 1.2 | 2.28
Other | | 0.0132 | | | 0.52
Nlocal: 32000 ave 32060 max 31939 min
Histogram: 1 0 1 0 0 0 0 1 0 1

View File

@ -1,4 +1,4 @@
LAMMPS (15 Feb 2016)
LAMMPS (6 Oct 2016)
# Rhodopsin model
units real
@ -56,6 +56,7 @@ timestep 2.0
run 100
PPPM initialization ...
WARNING: Using 12-bit tables for long-range coulomb (../kspace.cpp:316)
G vector (1/distance) = 0.248835
grid = 25 32 32
stencil order = 5
@ -70,41 +71,41 @@ Neighbor list info ...
master list distance cutoff = 12
ghost atom cutoff = 12
binsize = 6 -> bins = 10 13 13
Memory usage per processor = 91.7487 Mbytes
Memory usage per processor = 93.2721 Mbytes
---------------- Step 0 ----- CPU = 0.0000 (sec) ----------------
TotEng = -25356.2064 KinEng = 21444.8313 Temp = 299.0397
PotEng = -46801.0377 E_bond = 2537.9940 E_angle = 10921.3742
E_dihed = 5211.7865 E_impro = 213.5116 E_vdwl = -2307.8634
E_coul = 207025.8927 E_long = -270403.7333 Press = -142.6035
E_coul = 207025.8927 E_long = -270403.7333 Press = -149.3301
Volume = 307995.0335
---------------- Step 50 ----- CPU = 17.6362 (sec) ----------------
TotEng = -25330.0828 KinEng = 21501.0029 Temp = 299.8230
PotEng = -46831.0857 E_bond = 2471.7004 E_angle = 10836.4975
E_dihed = 5239.6299 E_impro = 227.1218 E_vdwl = -1993.2754
E_coul = 206797.6331 E_long = -270410.3930 Press = 237.6701
Volume = 308031.5639
---------------- Step 100 ----- CPU = 35.9089 (sec) ----------------
TotEng = -25290.7593 KinEng = 21592.0117 Temp = 301.0920
PotEng = -46882.7709 E_bond = 2567.9807 E_angle = 10781.9408
E_dihed = 5198.7432 E_impro = 216.7834 E_vdwl = -1902.4783
E_coul = 206659.2326 E_long = -270404.9733 Press = 6.9960
Volume = 308133.9888
Loop time of 35.9089 on 1 procs for 100 steps with 32000 atoms
---------------- Step 50 ----- CPU = 17.2007 (sec) ----------------
TotEng = -25330.0321 KinEng = 21501.0036 Temp = 299.8230
PotEng = -46831.0357 E_bond = 2471.7033 E_angle = 10836.5108
E_dihed = 5239.6316 E_impro = 227.1219 E_vdwl = -1993.2763
E_coul = 206797.6655 E_long = -270410.3927 Press = 237.6866
Volume = 308031.5640
---------------- Step 100 ----- CPU = 35.0315 (sec) ----------------
TotEng = -25290.7387 KinEng = 21591.9096 Temp = 301.0906
PotEng = -46882.6484 E_bond = 2567.9789 E_angle = 10781.9556
E_dihed = 5198.7493 E_impro = 216.7863 E_vdwl = -1902.6458
E_coul = 206659.5006 E_long = -270404.9733 Press = 6.7898
Volume = 308133.9933
Loop time of 35.0316 on 1 procs for 100 steps with 32000 atoms
Performance: 0.481 ns/day, 49.874 hours/ns, 2.785 timesteps/s
Performance: 0.493 ns/day, 48.655 hours/ns, 2.855 timesteps/s
99.9% CPU use with 1 MPI tasks x no OpenMP threads
MPI task timing breakdown:
Section | min time | avg time | max time |%varavg| %total
---------------------------------------------------------------
Pair | 25.731 | 25.731 | 25.731 | 0.0 | 71.66
Bond | 1.2771 | 1.2771 | 1.2771 | 0.0 | 3.56
Kspace | 3.2094 | 3.2094 | 3.2094 | 0.0 | 8.94
Neigh | 4.4538 | 4.4538 | 4.4538 | 0.0 | 12.40
Comm | 0.068507 | 0.068507 | 0.068507 | 0.0 | 0.19
Output | 0.00025916 | 0.00025916 | 0.00025916 | 0.0 | 0.00
Modify | 1.1417 | 1.1417 | 1.1417 | 0.0 | 3.18
Other | | 0.027 | | | 0.08
Pair | 25.021 | 25.021 | 25.021 | 0.0 | 71.42
Bond | 1.2834 | 1.2834 | 1.2834 | 0.0 | 3.66
Kspace | 3.2116 | 3.2116 | 3.2116 | 0.0 | 9.17
Neigh | 4.2767 | 4.2767 | 4.2767 | 0.0 | 12.21
Comm | 0.069283 | 0.069283 | 0.069283 | 0.0 | 0.20
Output | 0.00028205 | 0.00028205 | 0.00028205 | 0.0 | 0.00
Modify | 1.14 | 1.14 | 1.14 | 0.0 | 3.25
Other | | 0.02938 | | | 0.08
Nlocal: 32000 ave 32000 max 32000 min
Histogram: 1 0 0 0 0 0 0 0 0 0
@ -113,9 +114,9 @@ Histogram: 1 0 0 0 0 0 0 0 0 0
Neighs: 1.20281e+07 ave 1.20281e+07 max 1.20281e+07 min
Histogram: 1 0 0 0 0 0 0 0 0 0
Total # of neighbors = 12028107
Total # of neighbors = 12028098
Ave neighs/atom = 375.878
Ave special neighs/atom = 7.43187
Neighbor list builds = 11
Dangerous builds = 0
Total wall time: 0:00:37
Total wall time: 0:00:36

View File

@ -1,4 +1,4 @@
LAMMPS (15 Feb 2016)
LAMMPS (6 Oct 2016)
# Rhodopsin model
units real
@ -56,6 +56,7 @@ timestep 2.0
run 100
PPPM initialization ...
WARNING: Using 12-bit tables for long-range coulomb (../kspace.cpp:316)
G vector (1/distance) = 0.248835
grid = 25 32 32
stencil order = 5
@ -70,52 +71,52 @@ Neighbor list info ...
master list distance cutoff = 12
ghost atom cutoff = 12
binsize = 6 -> bins = 10 13 13
Memory usage per processor = 36.629 Mbytes
Memory usage per processor = 37.3604 Mbytes
---------------- Step 0 ----- CPU = 0.0000 (sec) ----------------
TotEng = -25356.2064 KinEng = 21444.8313 Temp = 299.0397
PotEng = -46801.0377 E_bond = 2537.9940 E_angle = 10921.3742
E_dihed = 5211.7865 E_impro = 213.5116 E_vdwl = -2307.8634
E_coul = 207025.8927 E_long = -270403.7333 Press = -142.6035
E_coul = 207025.8927 E_long = -270403.7333 Press = -149.3301
Volume = 307995.0335
---------------- Step 50 ----- CPU = 4.7461 (sec) ----------------
TotEng = -25330.0828 KinEng = 21501.0029 Temp = 299.8230
PotEng = -46831.0857 E_bond = 2471.7004 E_angle = 10836.4975
E_dihed = 5239.6299 E_impro = 227.1218 E_vdwl = -1993.2754
E_coul = 206797.6331 E_long = -270410.3930 Press = 237.6701
Volume = 308031.5639
---------------- Step 100 ----- CPU = 9.6332 (sec) ----------------
TotEng = -25290.7591 KinEng = 21592.0117 Temp = 301.0920
PotEng = -46882.7708 E_bond = 2567.9807 E_angle = 10781.9408
E_dihed = 5198.7432 E_impro = 216.7834 E_vdwl = -1902.4783
E_coul = 206659.2327 E_long = -270404.9733 Press = 6.9960
Volume = 308133.9888
Loop time of 9.63322 on 4 procs for 100 steps with 32000 atoms
---------------- Step 50 ----- CPU = 4.6056 (sec) ----------------
TotEng = -25330.0321 KinEng = 21501.0036 Temp = 299.8230
PotEng = -46831.0357 E_bond = 2471.7033 E_angle = 10836.5108
E_dihed = 5239.6316 E_impro = 227.1219 E_vdwl = -1993.2763
E_coul = 206797.6655 E_long = -270410.3927 Press = 237.6866
Volume = 308031.5640
---------------- Step 100 ----- CPU = 9.3910 (sec) ----------------
TotEng = -25290.7386 KinEng = 21591.9096 Temp = 301.0906
PotEng = -46882.6482 E_bond = 2567.9789 E_angle = 10781.9556
E_dihed = 5198.7493 E_impro = 216.7863 E_vdwl = -1902.6458
E_coul = 206659.5007 E_long = -270404.9733 Press = 6.7898
Volume = 308133.9933
Loop time of 9.39107 on 4 procs for 100 steps with 32000 atoms
Performance: 1.794 ns/day, 13.379 hours/ns, 10.381 timesteps/s
99.9% CPU use with 4 MPI tasks x no OpenMP threads
Performance: 1.840 ns/day, 13.043 hours/ns, 10.648 timesteps/s
99.8% CPU use with 4 MPI tasks x no OpenMP threads
MPI task timing breakdown:
Section | min time | avg time | max time |%varavg| %total
---------------------------------------------------------------
Pair | 6.4364 | 6.5993 | 6.7208 | 4.7 | 68.51
Bond | 0.30755 | 0.32435 | 0.35704 | 3.4 | 3.37
Kspace | 0.92248 | 1.0782 | 1.2597 | 13.0 | 11.19
Neigh | 1.1669 | 1.1672 | 1.1675 | 0.0 | 12.12
Comm | 0.094674 | 0.098065 | 0.10543 | 1.4 | 1.02
Output | 0.00015521 | 0.00016224 | 0.00018215 | 0.1 | 0.00
Modify | 0.32982 | 0.34654 | 0.35365 | 1.6 | 3.60
Other | | 0.01943 | | | 0.20
Pair | 6.2189 | 6.3266 | 6.6072 | 6.5 | 67.37
Bond | 0.30793 | 0.32122 | 0.3414 | 2.4 | 3.42
Kspace | 0.87994 | 1.1644 | 1.2855 | 15.3 | 12.40
Neigh | 1.1358 | 1.136 | 1.1362 | 0.0 | 12.10
Comm | 0.08292 | 0.084935 | 0.087077 | 0.5 | 0.90
Output | 0.00015712 | 0.00016558 | 0.00018501 | 0.1 | 0.00
Modify | 0.33717 | 0.34246 | 0.34794 | 0.7 | 3.65
Other | | 0.01526 | | | 0.16
Nlocal: 8000 ave 8143 max 7933 min
Histogram: 1 2 0 0 0 0 0 0 0 1
Nghost: 22733.5 ave 22769 max 22693 min
Histogram: 1 0 0 0 0 2 0 0 0 1
Neighs: 3.00703e+06 ave 3.0975e+06 max 2.96493e+06 min
Neighs: 3.00702e+06 ave 3.0975e+06 max 2.96492e+06 min
Histogram: 1 2 0 0 0 0 0 0 0 1
Total # of neighbors = 12028107
Total # of neighbors = 12028098
Ave neighs/atom = 375.878
Ave special neighs/atom = 7.43187
Neighbor list builds = 11
Dangerous builds = 0
Total wall time: 0:00:10
Total wall time: 0:00:09

View File

@ -1,4 +1,4 @@
LAMMPS (15 Feb 2016)
LAMMPS (6 Oct 2016)
# Rhodopsin model
variable x index 1
@ -77,6 +77,7 @@ timestep 2.0
run 100
PPPM initialization ...
WARNING: Using 12-bit tables for long-range coulomb (../kspace.cpp:316)
G vector (1/distance) = 0.248593
grid = 48 60 36
stencil order = 5
@ -91,52 +92,52 @@ Neighbor list info ...
master list distance cutoff = 12
ghost atom cutoff = 12
binsize = 6 -> bins = 19 26 13
Memory usage per processor = 95.5339 Mbytes
Memory usage per processor = 96.9597 Mbytes
---------------- Step 0 ----- CPU = 0.0000 (sec) ----------------
TotEng = -101425.4887 KinEng = 85779.3251 Temp = 299.0304
PotEng = -187204.8138 E_bond = 10151.9760 E_angle = 43685.4968
E_dihed = 20847.1460 E_impro = 854.0463 E_vdwl = -9231.4537
E_coul = 827053.5824 E_long = -1080565.6077 Press = -142.3092
E_coul = 827053.5824 E_long = -1080565.6077 Press = -149.0358
Volume = 1231980.1340
---------------- Step 50 ----- CPU = 18.7806 (sec) ----------------
TotEng = -101320.2677 KinEng = 86003.4837 Temp = 299.8118
PotEng = -187323.7514 E_bond = 9887.1072 E_angle = 43346.7922
E_dihed = 20958.7032 E_impro = 908.4715 E_vdwl = -7973.4457
E_coul = 826141.3831 E_long = -1080592.7629 Press = 238.0161
Volume = 1232126.1855
---------------- Step 100 ----- CPU = 38.3684 (sec) ----------------
TotEng = -101158.1849 KinEng = 86355.6149 Temp = 301.0393
PotEng = -187513.7998 E_bond = 10272.0693 E_angle = 43128.6454
E_dihed = 20793.9759 E_impro = 867.0826 E_vdwl = -7586.7186
E_coul = 825583.7122 E_long = -1080572.5667 Press = 15.2151
Volume = 1232535.8423
Loop time of 38.3684 on 4 procs for 100 steps with 128000 atoms
---------------- Step 50 ----- CPU = 18.1689 (sec) ----------------
TotEng = -101320.0211 KinEng = 86003.4933 Temp = 299.8118
PotEng = -187323.5144 E_bond = 9887.1189 E_angle = 43346.8448
E_dihed = 20958.7108 E_impro = 908.4721 E_vdwl = -7973.4486
E_coul = 826141.5493 E_long = -1080592.7617 Press = 238.0404
Volume = 1232126.1814
---------------- Step 100 ----- CPU = 37.2027 (sec) ----------------
TotEng = -101157.9546 KinEng = 86355.7413 Temp = 301.0398
PotEng = -187513.6959 E_bond = 10272.0456 E_angle = 43128.7018
E_dihed = 20794.0107 E_impro = 867.0928 E_vdwl = -7587.2409
E_coul = 825584.2416 E_long = -1080572.5474 Press = 15.1729
Volume = 1232535.8440
Loop time of 37.2028 on 4 procs for 100 steps with 128000 atoms
Performance: 0.450 ns/day, 53.289 hours/ns, 2.606 timesteps/s
Performance: 0.464 ns/day, 51.671 hours/ns, 2.688 timesteps/s
99.9% CPU use with 4 MPI tasks x no OpenMP threads
MPI task timing breakdown:
Section | min time | avg time | max time |%varavg| %total
---------------------------------------------------------------
Pair | 26.205 | 26.538 | 26.911 | 5.0 | 69.17
Bond | 1.298 | 1.3125 | 1.3277 | 1.0 | 3.42
Kspace | 3.7099 | 4.0992 | 4.4422 | 13.3 | 10.68
Neigh | 4.6137 | 4.6144 | 4.615 | 0.0 | 12.03
Comm | 0.21398 | 0.21992 | 0.22886 | 1.2 | 0.57
Output | 0.00030518 | 0.00031543 | 0.00033307 | 0.1 | 0.00
Modify | 1.5066 | 1.5232 | 1.5388 | 1.0 | 3.97
Other | | 0.06051 | | | 0.16
Pair | 25.431 | 25.738 | 25.984 | 4.0 | 69.18
Bond | 1.2966 | 1.3131 | 1.3226 | 0.9 | 3.53
Kspace | 3.7563 | 4.0123 | 4.3127 | 10.0 | 10.79
Neigh | 4.3778 | 4.378 | 4.3782 | 0.0 | 11.77
Comm | 0.1903 | 0.19549 | 0.20485 | 1.3 | 0.53
Output | 0.00031805 | 0.00037521 | 0.00039601 | 0.2 | 0.00
Modify | 1.4861 | 1.5051 | 1.5122 | 0.9 | 4.05
Other | | 0.05992 | | | 0.16
Nlocal: 32000 ave 32000 max 32000 min
Histogram: 4 0 0 0 0 0 0 0 0 0
Nghost: 47957 ave 47957 max 47957 min
Histogram: 4 0 0 0 0 0 0 0 0 0
Neighs: 1.20281e+07 ave 1.20572e+07 max 1.1999e+07 min
Neighs: 1.20281e+07 ave 1.20572e+07 max 1.19991e+07 min
Histogram: 2 0 0 0 0 0 0 0 0 2
Total # of neighbors = 48112472
Total # of neighbors = 48112540
Ave neighs/atom = 375.879
Ave special neighs/atom = 7.43187
Neighbor list builds = 11
Dangerous builds = 0
Total wall time: 0:00:39
Total wall time: 0:00:38

View File

@ -8,13 +8,15 @@ VENV = $(BUILDDIR)/docenv
TXT2RST = $(VENV)/bin/txt2rst
PYTHON = $(shell which python3)
HAS_PYTHON3 = NO
HAS_VIRTUALENV = NO
ifeq ($(shell which python3 >/dev/null 2>&1; echo $$?), 1)
$(error Python3 was not found! Please check README.md for further instructions)
ifeq ($(shell which python3 >/dev/null 2>&1; echo $$?), 0)
HAS_PYTHON3 = YES
endif
ifeq ($(shell which virtualenv >/dev/null 2>&1; echo $$?), 1)
$(error virtualenv was not found! Please check README.md for further instructions)
ifeq ($(shell which virtualenv >/dev/null 2>&1; echo $$?), 0)
HAS_VIRTUALENV = YES
endif
SOURCES=$(wildcard src/*.txt)
@ -109,6 +111,8 @@ $(RSTDIR)/%.rst : src/%.txt $(TXT2RST)
)
$(VENV):
@if [ "$(HAS_PYTHON3)" == "NO" ] ; then echo "Python3 was not found! Please check README.md for further instructions" 1>&2; exit 1; fi
@if [ "$(HAS_VIRTUALENV)" == "NO" ] ; then echo "virtualenv was not found! Please check README.md for further instructions" 1>&2; exit 1; fi
@( \
virtualenv -p $(PYTHON) $(VENV); \
. $(VENV)/bin/activate; \

View File

@ -91,6 +91,3 @@ This will install virtualenv from the Python Package Index.
----------------
Installing prerequisites for PDF build

Binary file not shown.

Before

Width:  |  Height:  |  Size: 1.6 KiB

View File

@ -1,9 +0,0 @@
\documentclass[12pt]{article}
\pagestyle{empty}
\begin{document}
$$
F^C = A \omega_{ij} \qquad \qquad r_{ij} < r_c
$$
\end{document}

Binary file not shown.

After

Width:  |  Height:  |  Size: 25 KiB

View File

@ -0,0 +1,12 @@
\documentclass[12pt]{article}
\pagestyle{empty}
\begin{document}
\begin{eqnarray*}
du_{i}^{cond} & = & \kappa_{ij}(\frac{1}{\theta_{i}}-\frac{1}{\theta_{j}})\omega_{ij}^{2} + \alpha_{ij}\omega_{ij}\zeta_{ij}^{q}(\Delta{t})^{-1/2} \\
du_{i}^{mech} & = & -\frac{1}{2}\gamma_{ij}\omega_{ij}^{2}(\frac{\vec{r_{ij}}}{r_{ij}}\bullet\vec{v_{ij}})^{2} -
\frac{\sigma^{2}_{ij}}{4}(\frac{1}{m_{i}}+\frac{1}{m_{j}})\omega_{ij}^{2} -
\frac{1}{2}\sigma_{ij}\omega_{ij}(\frac{\vec{r_{ij}}}{r_{ij}}\bullet\vec{v_{ij}})\zeta_{ij}(\Delta{t})^{-1/2} \\
\end{eqnarray*}
\end{document}

Binary file not shown.

After

Width:  |  Height:  |  Size: 8.7 KiB

View File

@ -0,0 +1,11 @@
\documentclass[12pt]{article}
\pagestyle{empty}
\begin{document}
\begin{eqnarray*}
\alpha_{ij}^{2} & = & 2k_{B}\kappa_{ij} \\
\sigma^{2}_{ij} & = & 2\gamma_{ij}k_{B}\Theta_{ij} \\
\Theta_{ij}^{-1} & = & \frac{1}{2}(\frac{1}{\theta_{i}}+\frac{1}{\theta_{j}}) \\
\end{eqnarray*}
\end{document}

BIN
doc/src/JPG/gran_funnel.png Normal file

Binary file not shown.

After

Width:  |  Height:  |  Size: 117 KiB

Binary file not shown.

After

Width:  |  Height:  |  Size: 2.2 KiB

BIN
doc/src/JPG/gran_mixer.png Normal file

Binary file not shown.

After

Width:  |  Height:  |  Size: 224 KiB

Binary file not shown.

After

Width:  |  Height:  |  Size: 3.0 KiB

View File

@ -1,7 +1,7 @@
<!-- HTML_ONLY -->
<HEAD>
<TITLE>LAMMPS Users Manual</TITLE>
<META NAME="docnumber" CONTENT="5 Oct 2016 version">
<META NAME="docnumber" CONTENT="18 Oct 2016 version">
<META NAME="author" CONTENT="http://lammps.sandia.gov - Sandia National Laboratories">
<META NAME="copyright" CONTENT="Copyright (2003) Sandia Corporation. This software and manual is distributed under the GNU General Public License.">
</HEAD>
@ -21,7 +21,7 @@
<H1></H1>
LAMMPS Documentation :c,h3
5 Oct 2016 version :c,h4
18 Oct 2016 version :c,h4
Version info: :h4
@ -109,7 +109,7 @@ it gives quick access to documentation for all LAMMPS commands.
:caption: User Documentation
:name: userdoc
:includehidden:
Section_intro
Section_start
Section_commands
@ -144,7 +144,7 @@ Indices and tables
* :ref:`genindex`
* :ref:`search`
END_RST -->
<!-- HTML_ONLY -->

View File

@ -117,7 +117,7 @@ PPPM. However, 2-FFT PPPM also requires a slightly larger mesh size to
achieve the same accuracy as 4-FFT PPPM. For problems where the FFT
cost is the performance bottleneck (typically large problems running
on many processors), 2-FFT PPPM may be faster than 4-FFT PPPM.
Staggered PPPM performs calculations using two different meshes, one
shifted slightly with respect to the other. This can reduce force
aliasing errors and increase the accuracy of the method, but also

View File

@ -37,14 +37,14 @@ simulation with all the settings. Rather, the input script is read
one line at a time and each command takes effect when it is read.
Thus this sequence of commands:
timestep 0.5
run 100
timestep 0.5
run 100
run 100 :pre
does something different than this sequence:
run 100
timestep 0.5
run 100
timestep 0.5
run 100 :pre
In the first case, the specified timestep (0.5 fmsec) is used for two
@ -97,7 +97,7 @@ single leading "#" will comment out the entire command.
(3) The line is searched repeatedly for $ characters, which indicate
variables that are replaced with a text string. See an exception in
(6).
(6).
If the $ is followed by curly brackets, then the variable name is the
text inside the curly brackets. If no curly brackets follow the $,
@ -123,7 +123,7 @@ variable X equal (xlo+xhi)/2+sqrt(v_area)
region 1 block $X 2 INF INF EDGE EDGE
variable X delete :pre
can be replaced by
can be replaced by
region 1 block $((xlo+xhi)/2+sqrt(v_area)) 2 INF INF EDGE EDGE :pre
@ -282,78 +282,135 @@ the "minimize"_minimize.html command. A parallel tempering
3.4 Commands listed by category :link(cmd_4),h4
This section lists all LAMMPS commands, grouped by category. The
"next section"_#cmd_5 lists the same commands alphabetically. Note
that some style options for some commands are part of specific LAMMPS
packages, which means they cannot be used unless the package was
included when LAMMPS was built. Not all packages are included in a
default LAMMPS build. These dependencies are listed as Restrictions
in the command's documentation.
"next section"_#cmd_5 lists the same commands alphabetically. The
next section also includes (long) lists of style options for entries
that appear in the following categories as a single command (fix,
compute, pair, etc). Commands that are added by user packages are not
included in these categories, but they are in the next section.
Initialization:
"atom_modify"_atom_modify.html, "atom_style"_atom_style.html,
"boundary"_boundary.html, "dimension"_dimension.html,
"newton"_newton.html, "processors"_processors.html, "units"_units.html
"newton"_newton.html,
"package"_package.html,
"processors"_processors.html,
"suffix"_suffix.html,
"units"_units.html
Atom definition:
Setup simulation box:
"create_atoms"_create_atoms.html, "create_box"_create_box.html,
"lattice"_lattice.html, "read_data"_read_data.html,
"read_dump"_read_dump.html, "read_restart"_read_restart.html,
"region"_region.html, "replicate"_replicate.html
"boundary"_boundary.html,
"box"_box.html,
"change_box"_change_box.html,
"create_box"_create_box.html,
"dimension"_dimension.html,
"lattice"_lattice.html,
"region"_region.html
Setup atoms:
"atom_modify"_atom_modify.html,
"atom_style"_atom_style.html,
"balance"_balance.html,
"create_atoms"_create_atoms.html,
"create_bonds"_create_bonds.html,
"delete_atoms"_delete_atoms.html,
"delete_bonds"_delete_bonds.html,
"displace_atoms"_displace_atoms.html,
"group"_group.html,
"mass"_mass.html,
"molecule"_molecule.html,
"read_data"_read_data.html,
"read_dump"_read_dump.html,
"read_restart"_read_restart.html,
"replicate"_replicate.html,
"set"_set.html,
"velocity"_velocity.html
Force fields:
"angle_coeff"_angle_coeff.html, "angle_style"_angle_style.html,
"bond_coeff"_bond_coeff.html, "bond_style"_bond_style.html,
"dielectric"_dielectric.html, "dihedral_coeff"_dihedral_coeff.html,
"angle_coeff"_angle_coeff.html,
"angle_style"_angle_style.html,
"bond_coeff"_bond_coeff.html,
"bond_style"_bond_style.html,
"bond_write"_bond_write.html,
"dielectric"_dielectric.html,
"dihedral_coeff"_dihedral_coeff.html,
"dihedral_style"_dihedral_style.html,
"improper_coeff"_improper_coeff.html,
"improper_style"_improper_style.html,
"kspace_modify"_kspace_modify.html, "kspace_style"_kspace_style.html,
"pair_coeff"_pair_coeff.html, "pair_modify"_pair_modify.html,
"pair_style"_pair_style.html, "pair_write"_pair_write.html,
"kspace_modify"_kspace_modify.html,
"kspace_style"_kspace_style.html,
"pair_coeff"_pair_coeff.html,
"pair_modify"_pair_modify.html,
"pair_style"_pair_style.html,
"pair_write"_pair_write.html,
"special_bonds"_special_bonds.html
Settings:
"comm_style"_comm_style.html, "group"_group.html, "mass"_mass.html,
"min_modify"_min_modify.html, "min_style"_min_style.html,
"neigh_modify"_neigh_modify.html, "neighbor"_neighbor.html,
"reset_timestep"_reset_timestep.html, "run_style"_run_style.html,
"set"_set.html, "timestep"_timestep.html, "velocity"_velocity.html
"comm_modify"_comm_modify.html,
"comm_style"_comm_style.html,
"info"_info.html,
"min_modify"_min_modify.html,
"min_style"_min_style.html,
"neigh_modify"_neigh_modify.html,
"neighbor"_neighbor.html,
"partition"_partition.html,
"reset_timestep"_reset_timestep.html,
"run_style"_run_style.html,
"timer"_timer.html,
"timestep"_timestep.html
Fixes:
Operations within timestepping (fixes) and diagnositics (computes):
"fix"_fix.html, "fix_modify"_fix_modify.html, "unfix"_unfix.html
Computes:
"compute"_compute.html, "compute_modify"_compute_modify.html,
"uncompute"_uncompute.html
"compute"_compute.html,
"compute_modify"_compute_modify.html,
"fix"_fix.html,
"fix_modify"_fix_modify.html,
"uncompute"_uncompute.html,
"unfix"_unfix.html
Output:
"dump"_dump.html, "dump image"_dump_image.html,
"dump_modify"_dump_modify.html, "dump movie"_dump_image.html,
"restart"_restart.html, "thermo"_thermo.html,
"thermo_modify"_thermo_modify.html, "thermo_style"_thermo_style.html,
"undump"_undump.html, "write_data"_write_data.html,
"write_dump"_write_dump.html, "write_restart"_write_restart.html
"dump image"_dump_image.html,
"dump movie"_dump_image.html,
"dump"_dump.html,
"dump_modify"_dump_modify.html,
"restart"_restart.html,
"thermo"_thermo.html,
"thermo_modify"_thermo_modify.html,
"thermo_style"_thermo_style.html,
"undump"_undump.html,
"write_coeff"_write_coeff.html,
"write_data"_write_data.html,
"write_dump"_write_dump.html,
"write_restart"_write_restart.html
Actions:
"delete_atoms"_delete_atoms.html, "delete_bonds"_delete_bonds.html,
"displace_atoms"_displace_atoms.html, "change_box"_change_box.html,
"minimize"_minimize.html, "neb"_neb.html "prd"_prd.html,
"rerun"_rerun.html, "run"_run.html, "temper"_temper.html
"minimize"_minimize.html,
"neb"_neb.html,
"prd"_prd.html,
"rerun"_rerun.html,
"run"_run.html,
"tad"_tad.html,
"temper"_temper.html
Miscellaneous:
Input script control:
"clear"_clear.html, "echo"_echo.html, "if"_if.html,
"include"_include.html, "jump"_jump.html, "label"_label.html,
"log"_log.html, "next"_next.html, "print"_print.html,
"shell"_shell.html, "variable"_variable.html
"clear"_clear.html,
"echo"_echo.html,
"if"_if.html,
"include"_include.html,
"jump"_jump.html,
"label"_label.html,
"log"_log.html,
"next"_next.html,
"print"_print.html,
"python"_python.html,
"quit"_quit.html,
"shell"_shell.html,
"variable"_variable.html
:line
@ -516,6 +573,7 @@ USER-INTEL, k = KOKKOS, o = USER-OMP, t = OPT.
"gcmc"_fix_gcmc.html,
"gld"_fix_gld.html,
"gravity (o)"_fix_gravity.html,
"halt"_fix_halt.html,
"heat"_fix_heat.html,
"indent"_fix_indent.html,
"langevin (k)"_fix_langevin.html,
@ -599,6 +657,7 @@ USER-INTEL, k = KOKKOS, o = USER-OMP, t = OPT.
"viscous"_fix_viscous.html,
"wall/colloid"_fix_wall.html,
"wall/gran"_fix_wall_gran.html,
"wall/gran/region"_fix_wall_gran_region.html,
"wall/harmonic"_fix_wall.html,
"wall/lj1043"_fix_wall.html,
"wall/lj126"_fix_wall.html,
@ -617,6 +676,7 @@ package"_Section_start.html#start_3.
"atc"_fix_atc.html,
"ave/correlate/long"_fix_ave_correlate_long.html,
"colvars"_fix_colvars.html,
"dpd/energy"_fix_dpd_energy.html,
"drude"_fix_drude.html,
"drude/transform/direct"_fix_drude_transform.html,
"drude/transform/reverse"_fix_drude_transform.html,
@ -922,6 +982,7 @@ KOKKOS, o = USER-OMP, t = OPT.
"tip4p/long (o)"_pair_coul.html,
"tri/lj"_pair_tri_lj.html,
"vashishta (o)"_pair_vashishta.html,
"vashishta/table (o)"_pair_vashishta.html,
"yukawa (go)"_pair_yukawa.html,
"yukawa/colloid (go)"_pair_yukawa_colloid.html,
"zbl (go)"_pair_zbl.html :tb(c=4,ea=c)

View File

@ -159,7 +159,7 @@ As a last resort, you can send an email directly to the
These are two alphabetic lists of the "ERROR"_#error and
"WARNING"_#warn messages LAMMPS prints out and the reason why. If the
explanation here is not sufficient, the documentation for the
offending command may help.
offending command may help.
Error and warning messages also list the source file and line number
where the error was generated. For example, this message

View File

@ -54,30 +54,30 @@ accelerate: run with various acceleration options (OpenMP, GPU, Phi)
balance: dynamic load balancing, 2d system
body: body particles, 2d system
colloid: big colloid particles in a small particle solvent, 2d system
comb: models using the COMB potential
comb: models using the COMB potential
coreshell: core/shell model using CORESHELL package
crack: crack propagation in a 2d solid
crack: crack propagation in a 2d solid
deposit: deposit atoms and molecules on a surface
dipole: point dipolar particles, 2d system
dreiding: methanol via Dreiding FF
eim: NaCl using the EIM potential
ellipse: ellipsoidal particles in spherical solvent, 2d system
flow: Couette and Poiseuille flow in a 2d channel
flow: Couette and Poiseuille flow in a 2d channel
friction: frictional contact of spherical asperities between 2d surfaces
hugoniostat: Hugoniostat shock dynamics
indent: spherical indenter into a 2d solid
indent: spherical indenter into a 2d solid
kim: use of potentials in Knowledge Base for Interatomic Models (KIM)
meam: MEAM test for SiC and shear (same as shear examples)
melt: rapid melt of 3d LJ system
meam: MEAM test for SiC and shear (same as shear examples)
melt: rapid melt of 3d LJ system
micelle: self-assembly of small lipid-like molecules into 2d bilayers
min: energy minimization of 2d LJ melt
msst: MSST shock dynamics
min: energy minimization of 2d LJ melt
msst: MSST shock dynamics
nb3b: use of nonbonded 3-body harmonic pair style
neb: nudged elastic band (NEB) calculation for barrier finding
nemd: non-equilibrium MD of 2d sheared system
neb: nudged elastic band (NEB) calculation for barrier finding
nemd: non-equilibrium MD of 2d sheared system
obstacle: flow around two voids in a 2d channel
peptide: dynamics of a small solvated peptide chain (5-mer)
peri: Peridynamic model of cylinder impacted by indenter
peri: Peridynamic model of cylinder impacted by indenter
pour: pouring of granular particles into a 3d box, then chute flow
prd: parallel replica dynamics of vacancy diffusion in bulk Si
python: using embedded Python in a LAMMPS input script
@ -120,7 +120,7 @@ browser.
Uppercase directories :h4
ASPHERE: various aspherical particle models, using ellipsoids, rigid bodies, line/triangle particles, etc
COUPLE: examples of how to use LAMMPS as a library
COUPLE: examples of how to use LAMMPS as a library
DIFFUSE: compute diffusion coefficients via several methods
ELASTIC: compute elastic constants at zero temperature
ELASTIC_T: compute elastic constants at finite temperature

View File

@ -37,7 +37,7 @@ pitfalls or alternatives.
Please see some of the closed issues for examples of how to
suggest code enhancements, submit proposed changes, or report
elated issues and how they are resoved.
possible bugs and how they are resoved.
As an alternative to using GitHub, you may e-mail the
"core developers"_http://lammps.sandia.gov/authors.html or send
@ -71,7 +71,7 @@ a parallel framework similar to LAMMPS. Most notably, these have
included many-body potentials - Stillinger-Weber, Tersoff, ReaxFF -
and the associated charge-equilibration routines needed for ReaxFF.
The "History link"_http://lammps.sandia.gov/history.html on the
The "History link"_http://lammps.sandia.gov/history.html on the
LAMMPS WWW page gives a timeline of features added to the
C++ open-source version of LAMMPS over the last several years.
@ -80,7 +80,7 @@ site"_lws, except for Warp & GranFlow which were primarily used
internally. A brief listing of their features is given here.
LAMMPS 2001
F90 + MPI
dynamic memory
spatial-decomposition parallelism
@ -96,7 +96,7 @@ LAMMPS 2001
user-defined diagnostics :ul
LAMMPS 99
F77 + MPI
static memory allocation
spatial-decomposition parallelism

View File

@ -4,7 +4,7 @@
:link(ld,Manual.html)
:link(lc,Section_commands.html#comm)
:line
:line
6. How-to discussions :h3
@ -68,7 +68,7 @@ Look at the {in.chain} input script provided in the {bench} directory
of the LAMMPS distribution to see the original script that these 2
scripts are based on. If that script had the line
restart 50 tmp.restart :pre
restart 50 tmp.restart :pre
added to it, it would produce 2 binary restart files (tmp.restart.50
and tmp.restart.100) as it ran.
@ -76,17 +76,17 @@ and tmp.restart.100) as it ran.
This script could be used to read the 1st restart file and re-run the
last 50 timesteps:
read_restart tmp.restart.50 :pre
read_restart tmp.restart.50 :pre
neighbor 0.4 bin
neigh_modify every 1 delay 1 :pre
neighbor 0.4 bin
neigh_modify every 1 delay 1 :pre
fix 1 all nve
fix 2 all langevin 1.0 1.0 10.0 904297 :pre
fix 1 all nve
fix 2 all langevin 1.0 1.0 10.0 904297 :pre
timestep 0.012 :pre
timestep 0.012 :pre
run 50 :pre
run 50 :pre
Note that the following commands do not need to be repeated because
their settings are included in the restart file: {units, atom_style,
@ -107,25 +107,25 @@ lmp_g++ -r tmp.restart.50 tmp.restart.data :pre
Then, this script could be used to re-run the last 50 steps:
units lj
atom_style bond
pair_style lj/cut 1.12
pair_modify shift yes
bond_style fene
units lj
atom_style bond
pair_style lj/cut 1.12
pair_modify shift yes
bond_style fene
special_bonds 0.0 1.0 1.0 :pre
read_data tmp.restart.data :pre
read_data tmp.restart.data :pre
neighbor 0.4 bin
neigh_modify every 1 delay 1 :pre
neighbor 0.4 bin
neigh_modify every 1 delay 1 :pre
fix 1 all nve
fix 2 all langevin 1.0 1.0 10.0 904297 :pre
fix 1 all nve
fix 2 all langevin 1.0 1.0 10.0 904297 :pre
timestep 0.012 :pre
timestep 0.012 :pre
reset_timestep 50
run 50 :pre
reset_timestep 50
run 50 :pre
Note that nearly all the settings specified in the original {in.chain}
script must be repeated, except the {pair_coeff} and {bond_coeff}
@ -522,7 +522,7 @@ H mass = 1.008
O charge = -1.040
H charge = 0.520
r0 of OH bond = 0.9572
theta of HOH angle = 104.52
theta of HOH angle = 104.52
OM distance = 0.15
LJ epsilon of O-O = 0.1550
LJ sigma of O-O = 3.1536
@ -629,7 +629,7 @@ the SPC and SPC/E models.
Wikipedia also has a nice article on "water
models"_http://en.wikipedia.org/wiki/Water_model.
:line
:line
6.10 Coupling LAMMPS to other codes :link(howto_10),h4
@ -729,7 +729,7 @@ LAMMPS and half to the other code and run both codes simultaneously
before syncing them up periodically. Or it might instantiate multiple
instances of LAMMPS to perform different calculations.
:line
:line
6.11 Visualizing LAMMPS snapshots :link(howto_11),h4
@ -832,7 +832,7 @@ rotation of [A], [B], and [C] and can be computed as follows:
where A = | [A] | indicates the scalar length of [A]. The hat symbol (^)
indicates the corresponding unit vector. {beta} and {gamma} are angles
between the vectors described below. Note that by construction,
between the vectors described below. Note that by construction,
[a], [b], and [c] have strictly positive x, y, and z components, respectively.
If it should happen that
[A], [B], and [C] form a left-handed basis, then the above equations
@ -841,17 +841,17 @@ to first apply an inversion. This can be achieved
by interchanging two basis vectors or by changing the sign of one of them.
For consistency, the same rotation/inversion applied to the basis vectors
must also be applied to atom positions, velocities,
must also be applied to atom positions, velocities,
and any other vector quantities.
This can be conveniently achieved by first converting to
This can be conveniently achieved by first converting to
fractional coordinates in the
old basis and then converting to distance coordinates in the new basis.
The transformation is given by the following equation:
:c,image(Eqs/rotate.jpg)
where {V} is the volume of the box, [X] is the original vector quantity and
[x] is the vector in the LAMMPS basis.
where {V} is the volume of the box, [X] is the original vector quantity and
[x] is the vector in the LAMMPS basis.
There is no requirement that a triclinic box be periodic in any
dimension, though it typically should be in at least the 2nd dimension
@ -938,17 +938,17 @@ defined above. The relationship between these 6 quantities
(a,b,c,alpha,beta,gamma) and the LAMMPS box sizes (lx,ly,lz) =
(xhi-xlo,yhi-ylo,zhi-zlo) and tilt factors (xy,xz,yz) is as follows:
:c,image(Eqs/box.jpg)
:c,image(Eqs/box.jpg)
The inverse relationship can be written as follows:
:c,image(Eqs/box_inverse.jpg)
:c,image(Eqs/box_inverse.jpg)
The values of {a}, {b}, {c} , {alpha}, {beta} , and {gamma} can be printed
out or accessed by computes using the
"thermo_style custom"_thermo_style.html keywords
The values of {a}, {b}, {c} , {alpha}, {beta} , and {gamma} can be printed
out or accessed by computes using the
"thermo_style custom"_thermo_style.html keywords
{cella}, {cellb}, {cellc}, {cellalpha}, {cellbeta}, {cellgamma},
respectively.
respectively.
As discussed on the "dump"_dump.html command doc page, when the BOX
BOUNDS for a snapshot is written to a dump file for a triclinic box,
@ -2092,11 +2092,11 @@ lattice fcc 5.376 orient x 1 0 0 orient y 0 1 0 orient z 0 0 1
region box block 0 4 0 4 0 4
create_box 1 box
create_atoms 1 box
mass 1 39.948
mass 1 39.948
pair_style lj/cut 13.0
pair_coeff * * 0.2381 3.405
timestep $\{dt\}
thermo $d :pre
thermo $d :pre
# equilibration and thermalization :pre
@ -2130,7 +2130,7 @@ but uses the Einstein formulation, analogous to the Einstein
mean-square-displacement formulation for self-diffusivity. The
time-integrated momentum fluxes play the role of Cartesian
coordinates, whose mean-square displacement increases linearly
with time at sufficiently long times.
with time at sufficiently long times.
:line
@ -2510,8 +2510,8 @@ the electrostatic environment inducing polarizability.
Technically, shells are attached to the cores by a spring force f =
k*r where k is a parametrized spring constant and r is the distance
between the core and the shell. The charges of the core and the shell
add up to the ion charge, thus q(ion) = q(core) + q(shell). This
setup introduces the ion polarizability (alpha) given by
add up to the ion charge, thus q(ion) = q(core) + q(shell). This
setup introduces the ion polarizability (alpha) given by
alpha = q(shell)^2 / k. In a
similar fashion the mass of the ion is distributed on the core and the
shell with the core having the larger mass.
@ -2526,7 +2526,7 @@ for NaCl, as found in examples/coreshell, has this format:
432 atoms # core and shell atoms
216 bonds # number of core/shell springs :pre
4 atom types # 2 cores and 2 shells for Na and Cl
4 atom types # 2 cores and 2 shells for Na and Cl
2 bond types :pre
0.0 24.09597 xlo xhi
@ -2545,19 +2545,19 @@ Atoms :pre
1 1 2 1.5005 0.00000000 0.00000000 0.00000000 # core of core/shell pair 1
2 1 4 -2.5005 0.00000000 0.00000000 0.00000000 # shell of core/shell pair 1
3 2 1 1.5056 4.01599500 4.01599500 4.01599500 # core of core/shell pair 2
4 2 3 -0.5056 4.01599500 4.01599500 4.01599500 # shell of core/shell pair 2
4 2 3 -0.5056 4.01599500 4.01599500 4.01599500 # shell of core/shell pair 2
(...) :pre
Bonds # Bond topology for spring forces :pre
1 2 1 2 # spring for core/shell pair 1
2 2 3 4 # spring for core/shell pair 2
2 2 3 4 # spring for core/shell pair 2
(...) :pre
Non-Coulombic (e.g. Lennard-Jones) pairwise interactions are only
defined between the shells. Coulombic interactions are defined
between all cores and shells. If desired, additional bonds can be
specified between cores.
specified between cores.
The "special_bonds"_special_bonds.html command should be used to
turn-off the Coulombic interaction within core/shell pairs, since that
@ -2620,7 +2620,7 @@ Note that to perform thermostatting using this definition of
temperature, the "fix modify temp"_fix_modify.html command should be
used to assign the compute to the thermostat fix. Likewise the
"thermo_modify temp"_thermo_modify.html command can be used to make
this temperature be output for the overall system.
this temperature be output for the overall system.
For the NaCl example, this can be done as follows:
@ -2632,13 +2632,13 @@ fix thermostatequ all nve # integrator as needed f
fix_modify thermoberendsen temp CSequ
thermo_modify temp CSequ # output of center-of-mass derived temperature :pre
If "compute temp/cs"_compute_temp_cs.html is used, the decoupled
relative motion of the core and the shell should in theory be
If "compute temp/cs"_compute_temp_cs.html is used, the decoupled
relative motion of the core and the shell should in theory be
stable. However numerical fluctuation can introduce a small
momentum to the system, which is noticable over long trajectories.
Therefore it is recomendable to use the "fix
momentum"_fix_momentum.html command in combination with "compute
temp/cs"_compute_temp_cs.html when equilibrating the system to
Therefore it is recomendable to use the "fix
momentum"_fix_momentum.html command in combination with "compute
temp/cs"_compute_temp_cs.html when equilibrating the system to
prevent any drift.
When intializing the velocities of a system with core/shell pairs, it
@ -2661,17 +2661,17 @@ to the electrostatic environment. This fast movement also limits the
timestep size that can be used.
The primary literature of the adiabatic core/shell model suggests that
the fast relative motion of the core/shell pairs only allows negligible
the fast relative motion of the core/shell pairs only allows negligible
energy transfer to the environment. Therefore it is not intended to
decouple the core/shell degree of freedom from the physical system
during production runs. In other words, the "compute
temp/cs"_compute_temp_cs.html command should not be used during
production runs and is only required during equilibration. This way one
is consistent with literature (based on the code packages DL_POLY or
production runs and is only required during equilibration. This way one
is consistent with literature (based on the code packages DL_POLY or
GULP for instance).
The mentioned energy transfer will typically lead to a a small drift
in total energy over time. This internal energy can be monitored
The mentioned energy transfer will typically lead to a a small drift
in total energy over time. This internal energy can be monitored
using the "compute chunk/atom"_compute_chunk_atom.html and "compute
temp/chunk"_compute_temp_chunk.html commands. The internal kinetic
energies of each core/shell pair can then be summed using the sum()
@ -2702,14 +2702,14 @@ The additional section in the date file would be formatted like this:
CS-Info # header of additional section :pre
1 1 # column 1 = atom ID, column 2 = core/shell ID
2 1
3 2
4 2
5 3
6 3
7 4
8 4
1 1 # column 1 = atom ID, column 2 = core/shell ID
2 1
3 2
4 2
5 3
6 3
7 4
8 4
(...) :pre
:line

View File

@ -181,7 +181,7 @@ Atom creation :h5
displace atoms :ul
Ensembles, constraints, and boundary conditions :h5
("fix"_fix.html command)
("fix"_fix.html command)
2d or 3d systems
orthogonal or non-orthogonal (triclinic symmetry) simulation domains
@ -199,7 +199,7 @@ Ensembles, constraints, and boundary conditions :h5
variety of additional boundary conditions and constraints :ul
Integrators :h5
("run"_run.html, "run_style"_run_style.html, "minimize"_minimize.html commands)
("run"_run.html, "run_style"_run_style.html, "minimize"_minimize.html commands)
velocity-Verlet integrator
Brownian dynamics
@ -213,7 +213,7 @@ Diagnostics :h5
see the various flavors of the "fix"_fix.html and "compute"_compute.html commands :ul
Output :h5
("dump"_dump.html, "restart"_restart.html commands)
("dump"_dump.html, "restart"_restart.html commands)
log file of thermodynamic info
text dump files of atom coords, velocities, other per-atom quantities

View File

@ -182,7 +182,7 @@ Supporting info: "atom_style body"_atom_style.html, "body"_body.html,
"pair_style body"_pair_body.html, examples/body
:line
CLASS2 package :link(CLASS2),h5
Contents: Bond, angle, dihedral, improper, and pair styles for the
@ -206,9 +206,9 @@ Supporting info: "bond_style class2"_bond_class2.html, "angle_style
class2"_angle_class2.html, "dihedral_style
class2"_dihedral_class2.html, "improper_style
class2"_improper_class2.html, "pair_style lj/class2"_pair_class2.html
:line
COLLOID package :link(COLLOID),h5
Contents: Support for coarse-grained colloidal particles. Wall fix
@ -239,9 +239,9 @@ lubricate"_pair_lubricate.html, "pair_style
lubricateU"_pair_lubricateU.html, examples/colloid, examples/srd
:line
COMPRESS package :link(COMPRESS),h5
Contents: Support for compressed output of dump files via the zlib
compression library, using dump styles with a "gz" in their style
name.
@ -271,7 +271,7 @@ atom/gz"_dump.html, "dump cfg/gz"_dump.html, "dump
custom/gz"_dump.html, "dump xyz/gz"_dump.html
:line
CORESHELL package :link(CORESHELL),h5
Contents: Compute and pair styles that implement the adiabatic
@ -302,7 +302,7 @@ buck/coul/long/cs"_pair_cs.html, pair_style
lj/cut/coul/long/cs"_pair_lj.html, examples/coreshell
:line
DIPOLE package :link(DIPOLE),h5
Contents: An atom style and several pair styles to support point
@ -326,9 +326,9 @@ Supporting info: "atom_style dipole"_atom_style.html, "pair_style
lj/cut/dipole/cut"_pair_dipole.html, "pair_style
lj/cut/dipole/long"_pair_dipole.html, "pair_style
lj/long/dipole/long"_pair_dipole.html, examples/dipole
:line
GPU package :link(GPU),h5
Contents: Dozens of pair styles and a version of the PPPM long-range
@ -385,9 +385,9 @@ Pair Styles section of "Section 3.5"_Section_commands.html#cmd_5
for any pair style listed with a (g),
"kspace_style"_kspace_style.html, "package gpu"_package.html,
examples/accelerate, bench/FERMI, bench/KEPLER
:line
GRANULAR package :link(GRANULAR),h5
Contents: Fixes and pair styles that support models of finite-size
@ -412,9 +412,9 @@ Supporting info: "Section 6.6"_Section_howto.html#howto_6, "fix
pour"_fix_pour.html, "fix wall/gran"_fix_wall_gran.html, "pair_style
gran/hooke"_pair_gran.html, "pair_style
gran/hertz/history"_pair_gran.html, examples/pour, bench/in.chute
:line
KIM package :link(KIM),h5
Contents: A pair style that interfaces to the Knowledge Base for
@ -443,9 +443,9 @@ Make.py -p ^kim -a machine :pre
Supporting info: src/KIM/README, lib/kim/README, "pair_style
kim"_pair_kim.html, examples/kim
:line
KOKKOS package :link(KOKKOS),h5
Contents: Dozens of atom, pair, bond, angle, dihedral, improper styles
@ -501,7 +501,7 @@ for any pair style listed with a (k), "package kokkos"_package.html,
examples/accelerate, bench/FERMI, bench/KEPLER
:line
KSPACE package :link(KSPACE),h5
Contents: A variety of long-range Coulombic solvers, and pair styles
@ -543,7 +543,7 @@ which have "long" or "msm" in their style name,
examples/peptide, bench/in.rhodo
:line
MANYBODY package :link(MANYBODY),h5
Contents: A variety of many-body and bond-order potentials. These
@ -565,14 +565,14 @@ make machine :pre
Make.py -p ^manybody -a machine :pre
Supporting info:
Supporting info:
Examples: Pair Styles section of "Section
3.5"_Section_commands.html#cmd_5, examples/comb, examples/eim,
examples/nb3d, examples/vashishta
:line
MC package :link(MC),h5
Contents: Several fixes and a pair style that have Monte Carlo (MC) or
@ -598,9 +598,9 @@ Supporting info: "fix atom/swap"_fix_atom_swap.html, "fix
bond/break"_fix_bond_break.html, "fix
bond/create"_fix_bond_create.html, "fix bond/swap"_fix_bond_swap.html,
"fix gcmc"_fix_gcmc.html, "pair_style dsmc"_pair_dsmc.html
:line
MEAM package :link(MEAM),h5
Contents: A pair style for the modified embedded atom (MEAM)
@ -644,9 +644,9 @@ Make.py -p ^meam -a machine :pre
Supporting info: lib/meam/README, "pair_style meam"_pair_meam.html,
examples/meam
:line
MISC package :link(MISC),h5
Contents: A variety of computes, fixes, and pair styles that are not
@ -670,9 +670,9 @@ Make.py -p ^misc -a machine :pre
Supporting info: "compute ti"_compute_ti.html, "fix
evaporate"_fix_evaporate.html, "fix tmm"_fix_ttm.html, "fix
viscosity"_fix_viscosity.html, examples/misc
:line
MOLECULE package :link(MOLECULE),h5
Contents: A large number of atom, pair, bond, angle, dihedral,
@ -704,7 +704,7 @@ lj/charmm/coul/charmm"_pair_charmm.html,
examples/micelle, examples/peptide, bench/in.chain, bench/in.rhodo
:line
MPIIO package :link(MPIIO),h5
Contents: Support for parallel output/input of dump and restart files
@ -729,9 +729,9 @@ Make.py -p ^mpiio -a machine :pre
Supporting info: "dump"_dump.html, "restart"_restart.html,
"write_restart"_write_restart.html, "read_restart"_read_restart.html
:line
OPT package :link(OPT),h5
Contents: A handful of pair styles with an "opt" in their style name
@ -768,7 +768,7 @@ Supporting info: "Section 5.3"_Section_accelerate.html#acc_3,
listed with an (t), examples/accelerate, bench/KEPLER
:line
PERI package :link(PERI),h5
Contents: Support for the Peridynamics method, a particle-based
@ -796,9 +796,9 @@ Supporting info:
"doc/PDF/PDLammps_VES.pdf"_PDF/PDLammps_VES.pdf, "atom_style
peri"_atom_style.html, "compute damage/atom"_compute_damage_atom.html,
"pair_style peri/pmb"_pair_peri.html, examples/peri
:line
POEMS package :link(POEMS),h5
Contents: A fix that wraps the Parallelizable Open source Efficient
@ -839,7 +839,7 @@ Supporting info: src/POEMS/README, lib/poems/README,
"fix poems"_fix_poems.html, examples/rigid
:line
PYTHON package :link(PYTHON),h5
Contents: A "python"_python.html command which allow you to execute
@ -873,9 +873,9 @@ make machine :pre
Make.py -p ^python -a machine :pre
Supporting info: examples/python
:line
QEQ package :link(QEQ),h5
Contents: Several fixes for performing charge equilibration (QEq) via
@ -897,9 +897,9 @@ make machine :pre
Make.py -p ^qeq -a machine :pre
Supporting info: "fix qeq/*"_fix_qeq.html, examples/qeq
:line
REAX package :link(REAX),h5
Contents: A pair style for the ReaxFF potential, a universal reactive
@ -941,9 +941,9 @@ Make.py -p ^reax -a machine :pre
Supporting info: lib/reax/README, "pair_style reax"_pair_reax.html,
"fix reax/bonds"_fix_reax_bonds.html, examples/reax
:line
REPLICA package :link(REPLICA),h5
Contents: A collection of multi-replica methods that are used by
@ -978,7 +978,7 @@ Supporting info: "Section 6.5"_Section_howto.html#howto_5,
examples/tad
:line
RIGID package :link(RIGID),h5
Contents: A collection of computes and fixes which enforce rigid
@ -1005,7 +1005,7 @@ Supporting info: "compute erotate/rigid"_compute_erotate_rigid.html,
rigid/*"_fix_rigid.html, examples/ASPHERE, examples/rigid
:line
SHOCK package :link(SHOCK),h5
Contents: A small number of fixes useful for running impact
@ -1028,15 +1028,15 @@ Make.py -p ^shock -a machine :pre
Supporting info: "fix append/atoms"_fix_append_atoms.html, "fix
msst"_fix_msst.html, "fix nphug"_fix_nphug.html, "fix
wall/piston"_fix_wall_piston.html, examples/hugoniostat, examples/msst
:line
SNAP package :link(SNAP),h5
Contents: A pair style for the spectral neighbor analysis potential
(SNAP), which is an empirical potential which can be quantum accurate
when fit to an archive of DFT data. Computes useful for analyzing
properties of the potential are also included.
when fit to an archive of DFT data. Computes useful for analyzing
properties of the potential are also included.
To install via make or Make.py:
@ -1055,9 +1055,9 @@ Make.py -p ^snap -a machine :pre
Supporting info: "pair snap"_pair_snap.html, "compute
sna/atom"_compute_sna_atom.html, "compute snad/atom"_compute_sna_atom.html,
"compute snav/atom"_compute_sna_atom.html, examples/snap
:line
SRD package :link(SRD),h5
Contents: Two fixes which implement the Stochastic Rotation Dynamics
@ -1080,9 +1080,9 @@ Make.py -p ^srd -a machine :pre
Supporting info: "fix srd"_fix_srd.html, "fix
wall/srd"_fix_wall_srd.html, examples/srd, examples/ASPHERE
:line
VORONOI package :link(VORONOI),h5
Contents: A "compute voronoi/atom"_compute_voronoi_atom.html command
@ -1129,9 +1129,9 @@ Make.py -p ^voronoi -a machine :pre
Supporting info: src/VORONOI/README, lib/voronoi/README, "compute
voronoi/atom"_compute_voronoi_atom.html, examples/voronoi
:line
4.2 User packages :h4,link(pkg_2)
The current list of user-contributed packages is as follows:
@ -1302,7 +1302,7 @@ fix. The COLVARS library itself is written and maintained by Giacomo
Fiorin (ICMS, Temple University, Philadelphia, PA, USA) and Jerome
Henin (LISM, CNRS, Marseille, France). Contact them directly if you
have questions.
:line
USER-DIFFRACTION package :link(USER-DIFFRACTION),h5
@ -1380,7 +1380,7 @@ in 2007. See src/USER-EFF/README for more details. There are
auxiliary tools for using this package in tools/eff; see its README
file.
Supporting info:
Supporting info:
Author: Andres Jaramillo-Botero at CalTech (ajaramil at
wag.caltech.edu). Contact him directly if you have questions.
@ -1456,21 +1456,21 @@ LINKFLAGS: add -fopenmp :ul
For Phi mode add the following in addition to the CPU mode flags:
CCFLAGS: add -DLMP_INTEL_OFFLOAD and
CCFLAGS: add -DLMP_INTEL_OFFLOAD and
LINKFLAGS: add -offload :ul
And also add this to CCFLAGS:
-offload-option,mic,compiler,"-fp-model fast=2 -mGLOB_default_function_attrs=\"gather_scatter_loop_unroll=4\"" :pre
Examples:
Examples:
:line
USER-LB package :link(USER-LB),h5
Supporting info:
Supporting info:
This package contains a LAMMPS implementation of a background
Lattice-Boltzmann fluid, which can be used to model MD particles
influenced by hydrodynamic forces.
@ -1489,8 +1489,8 @@ Examples: examples/USER/lb
USER-MGPT package :link(USER-MGPT),h5
Supporting info:
Supporting info:
This package contains a fast implementation for LAMMPS of
quantum-based MGPT multi-ion potentials. The MGPT or model GPT method
derives from first-principles DFT-based generalized pseudopotential
@ -1521,8 +1521,8 @@ Examples: examples/USER/mgpt
USER-MISC package :link(USER-MISC),h5
Supporting info:
Supporting info:
The files in this package are a potpourri of (mostly) unrelated
features contributed to LAMMPS by users. Each feature is a single
pair of files (*.cpp and *.h).
@ -1548,8 +1548,8 @@ Examples: examples/USER/misc
USER-MANIFOLD package :link(USER-MANIFOLD),h5
Supporting info:
Supporting info:
This package contains a dump molfile command which uses molfile
plugins that are bundled with the
"VMD"_http://www.ks.uiuc.edu/Research/vmd molecular visualization and
@ -1574,8 +1574,8 @@ Contact him directly if you have questions.
USER-MOLFILE package :link(USER-MOLFILE),h5
Supporting info:
Supporting info:
This package contains a dump molfile command which uses molfile
plugins that are bundled with the
"VMD"_http://www.ks.uiuc.edu/Research/vmd molecular visualization and
@ -1600,12 +1600,12 @@ The person who created this package is Axel Kohlmeyer at Temple U
USER-OMP package :link(USER-OMP),h5
Supporting info:
Supporting info:
This package provides OpenMP multi-threading support and
other optimizations of various LAMMPS pair styles, dihedral
styles, and fix styles.
See this section of the manual to get started:
"Section 5.3"_Section_accelerate.html#acc_3
@ -1643,8 +1643,8 @@ Examples: examples/USER/phonon
USER-QMMM package :link(USER-QMMM),h5
Supporting info:
Supporting info:
This package provides a fix qmmm command which allows LAMMPS to be
used in a QM/MM simulation, currently only in combination with pw.x
code from the "Quantum ESPRESSO"_espresso package.
@ -1667,11 +1667,11 @@ The person who created this package is Axel Kohlmeyer at Temple U
(akohlmey at gmail.com). Contact him directly if you have questions.
:line
USER-QTB package :link(USER-QTB),h5
Supporting info:
Supporting info:
This package provides a self-consistent quantum treatment of the
vibrational modes in a classical molecular dynamics simulation. By
coupling the MD simulation to a colored thermostat, it introduces zero
@ -1701,16 +1701,16 @@ Examples: examples/USER/qtb
USER-QUIP package :link(USER-QUIP),h5
Supporting info:
Supporting info:
Examples: examples/USER/quip
:line
USER-REAXC package :link(USER-REAXC),h5
Supporting info:
Supporting info:
This package contains a implementation for LAMMPS of the ReaxFF force
field. ReaxFF uses distance-dependent bond-order functions to
represent the contributions of chemical bonding to the potential
@ -1748,24 +1748,24 @@ Examples: examples/reax
USER-SMD package :link(USER-SMD),h5
Supporting info:
Supporting info:
This package implements smoothed Mach dynamics (SMD) in
LAMMPS. Currently, the package has the following features:
* Does liquids via traditional Smooth Particle Hydrodynamics (SPH)
* Also solves solids mechanics problems via a state of the art
* Also solves solids mechanics problems via a state of the art
stabilized meshless method with hourglass control.
* Can specify hydrostatic interactions independently from material
* Can specify hydrostatic interactions independently from material
strength models, i.e. pressure and deviatoric stresses are separated.
* Many material models available (Johnson-Cook, plasticity with
hardening, Mie-Grueneisen, Polynomial EOS). Easy to add new
* Many material models available (Johnson-Cook, plasticity with
hardening, Mie-Grueneisen, Polynomial EOS). Easy to add new
material models.
* Rigid boundary conditions (walls) can be loaded as surface geometries
* Rigid boundary conditions (walls) can be loaded as surface geometries
from *.STL files.
See the file doc/PDF/SMD_LAMMPS_userguide.pdf to get started.
@ -1783,8 +1783,8 @@ Examples: examples/USER/smd
USER-SMTBQ package :link(USER-SMTBQ),h5
Supporting info:
Supporting info:
This package implements the Second Moment Tight Binding - QEq (SMTB-Q)
potential for the description of ionocovalent bonds in oxides.
@ -1806,22 +1806,22 @@ Examples: examples/USER/smtbq
USER-SPH package :link(USER-SPH),h5
Supporting info:
Supporting info:
This package implements smoothed particle hydrodynamics (SPH) in
LAMMPS. Currently, the package has the following features:
* Tait, ideal gas, Lennard-Jones equation of states, full support for
* Tait, ideal gas, Lennard-Jones equation of states, full support for
complete (i.e. internal-energy dependent) equations of state
* Plain or Monaghans XSPH integration of the equations of motion
* Density continuity or density summation to propagate the density field
* Commands to set internal energy and density of particles from the
* Commands to set internal energy and density of particles from the
input script
* Output commands to access internal energy and density for dumping and
* Output commands to access internal energy and density for dumping and
thermo output
See the file doc/PDF/SPH_LAMMPS_userguide.pdf to get started.
@ -1839,7 +1839,7 @@ Examples: examples/USER/sph
USER-TALLY package :link(USER-TALLY),h5
Supporting info:
Supporting info:
Examples: examples/USER/tally

View File

@ -23,7 +23,7 @@ In Python lingo, this is "embedding" Python in LAMMPS.
This section describes how to do both.
11.1 "Overview of running LAMMPS from Python"_#py_1
11.2 "Overview of using Python from a LAMMPS script"_#py_2
11.2 "Overview of using Python from a LAMMPS script"_#py_2
11.3 "Building LAMMPS as a shared library"_#py_3
11.4 "Installing the Python wrapper into Python"_#py_4
11.5 "Extending Python with MPI to run in parallel"_#py_5
@ -503,7 +503,7 @@ one of several ways:
The last command requires that the first line of the script be
something like this:
#!/usr/local/bin/python
#!/usr/local/bin/python
#!/usr/local/bin/python -i :pre
where the path points to where you have Python installed, and that you
@ -552,32 +552,32 @@ lmp.command(cmd) # invoke a single LAMMPS command, cmd = "run 100" :pre
xlo = lmp.extract_global(name,type) # extract a global quantity
# name = "boxxlo", "nlocal", etc
# type = 0 = int
# 1 = double :pre
# type = 0 = int
# 1 = double :pre
coords = lmp.extract_atom(name,type) # extract a per-atom quantity
# name = "x", "type", etc
# type = 0 = vector of ints
# 1 = array of ints
# 2 = vector of doubles
# 3 = array of doubles :pre
# type = 0 = vector of ints
# 1 = array of ints
# 2 = vector of doubles
# 3 = array of doubles :pre
eng = lmp.extract_compute(id,style,type) # extract value(s) from a compute
v3 = lmp.extract_fix(id,style,type,i,j) # extract value(s) from a fix
# id = ID of compute or fix
# style = 0 = global data
# 1 = per-atom data
# 2 = local data
# type = 0 = scalar
# 1 = vector
# 2 = array
# i,j = indices of value in global vector or array :pre
# style = 0 = global data
# 1 = per-atom data
# 2 = local data
# type = 0 = scalar
# 1 = vector
# 2 = array
# i,j = indices of value in global vector or array :pre
var = lmp.extract_variable(name,group,flag) # extract value(s) from a variable
# name = name of variable
# group = group ID (ignored for equal-style variables)
# flag = 0 = equal-style variable
# 1 = atom-style variable :pre
# name = name of variable
# group = group ID (ignored for equal-style variables)
# flag = 0 = equal-style variable
# 1 = atom-style variable :pre
flag = lmp.set_variable(name,value) # set existing named string-style variable to value, flag = 0 if successful
natoms = lmp.get_natoms() # total # of atoms as int
@ -724,7 +724,7 @@ lmp.scatter_coords("x",1,3,x) :pre
Alternatively, you can just change values in the vector returned by
gather_atoms("x",1,3), since it is a ctypes vector of doubles.
:line
:line
As noted above, these Python class methods correspond one-to-one with
the functions in the LAMMPS library interface in src/library.cpp and
@ -767,7 +767,7 @@ vizplotgui_tool.py, combination of viz_tool.py and plot.py and gui.py :tb(c=2)
For the viz_tool.py and vizplotgui_tool.py commands, replace "tool"
with "gl" or "atomeye" or "pymol" or "vmd", depending on what
visualization package you have installed.
visualization package you have installed.
Note that for GL, you need to be able to run the Pizza.py GL tool,
which is included in the pizza sub-directory. See the "Pizza.py doc

View File

@ -33,7 +33,7 @@ tar -xzvf lammps*.tar.gz :pre
This will create a LAMMPS directory containing two files and several
sub-directories:
README: text file
LICENSE: the GNU General Public License (GPL)
bench: benchmark problems
@ -600,10 +600,10 @@ LAMMPS will generate a run-time error. As far as we know, the
settings defined in src/lmptype.h are portable and work on every
current system.
In all cases, the size of problem that can be run on a per-processor
basis is limited by 4-byte integer storage to 2^31 atoms per processor
(about 2 billion). This should not normally be a limitation since such
a problem would have a huge per-processor memory footprint due to
In all cases, the size of problem that can be run on a per-processor
basis is limited by 4-byte integer storage to 2^31 atoms per processor
(about 2 billion). This should not normally be a limitation since such
a problem would have a huge per-processor memory footprint due to
neighbor lists and would run very slowly in terms of CPU secs/timestep.
:line
@ -841,7 +841,7 @@ libpackage.a
Makefile.lammps :pre
The Makefile.lammps file will typically be a copy of one of the
Makefile.lammps.* files in the library directory.
Makefile.lammps.* files in the library directory.
Note that you must insure that the settings in Makefile.lammps are
appropriate for your system. If they are not, the LAMMPS build may
@ -883,7 +883,7 @@ A few packages require specific settings in Makefile.machine, to
either build or use the package effectively. These are the
USER-INTEL, KOKKOS, USER-OMP, and OPT packages, used for accelerating
code performance on CPUs or other hardware, as discussed in "Section
5.3"_Section_accelerate.html#acc_3.
5.3"_Section_accelerate.html#acc_3.
A summary of what Makefile.machine changes are needed for each of
these packages is given in "Section 4"_Section_packages.html.
@ -1199,7 +1199,7 @@ installer package from "here"_http://rpm.lammps.org/windows.html
For running the non-MPI executable, follow these steps:
Get a command prompt by going to Start->Run... ,
Get a command prompt by going to Start->Run... ,
then typing "cmd". :ulb,l
Move to the directory where you have your input, e.g. a copy of
@ -1209,7 +1209,7 @@ At the command prompt, type "lmp_serial -in in.lj", replacing [in.lj]
with the name of your LAMMPS input script. :l
:ule
For the MPI version, which allows you to run LAMMPS under Windows on
For the MPI version, which allows you to run LAMMPS under Windows on
multiple processors, follow these steps:
Download and install
@ -1224,7 +1224,7 @@ For this you need to start a Command Prompt in {Administrator Mode}
installation directory, then into the subdirectory [bin] and execute
[smpd.exe -install]. Exit the command window.
Get a new, regular command prompt by going to Start->Run... ,
Get a new, regular command prompt by going to Start->Run... ,
then typing "cmd". :l
Move to the directory where you have your input file
@ -1488,7 +1488,7 @@ of the manual. World- and universe-style "variables"_variable.html
are useful in this context.
-plog file :pre
Specify the base name for the partition log files, so partition N
writes log information to file.N. If file is none, then no partition
log files are created. This overrides the filename specified in the
@ -1499,7 +1499,7 @@ replica_files/log.lammps) If this option is not used the log file for
partition N is log.lammps.N or whatever is specified by the -log
command-line option.
-pscreen file :pre
-pscreen file :pre
Specify the base name for the partition screen file, so partition N
writes screen information to file.N. If file is none, then no
@ -1511,7 +1511,7 @@ sub-directory (-pscreen replica_files/screen). If this option is not
used the screen file for partition N is screen.N or whatever is
specified by the -screen command-line option.
-restart restartfile {remap} datafile keyword value ... :pre
-restart restartfile {remap} datafile keyword value ... :pre
Convert the restart file into a data file and immediately exit. This
is the same operation as if the following 2-line input script were
@ -1572,7 +1572,7 @@ to
so that the processors in each partition will be
0 1 2 4 5 6 8 9 10
0 1 2 4 5 6 8 9 10
3 7 11 :pre
See the "processors" command for how to insure processors from each
@ -1663,12 +1663,12 @@ invokes the default USER-INTEL settings, as if the command "package
intel 1" were used at the top of your input script. These settings
can be changed by using the "-package intel" command-line switch or
the "package intel"_package.html command in your script. If the
USER-OMP package is also installed, the hybrid style with "intel omp"
arguments can be used to make the omp suffix a second choice, if a
requested style is not available in the USER-INTEL package. It will
also invoke the default USER-OMP settings, as if the command "package
omp 0" were used at the top of your input script. These settings can
be changed by using the "-package omp" command-line switch or the
USER-OMP package is also installed, the hybrid style with "intel omp"
arguments can be used to make the omp suffix a second choice, if a
requested style is not available in the USER-INTEL package. It will
also invoke the default USER-OMP settings, as if the command "package
omp 0" were used at the top of your input script. These settings can
be changed by using the "-package omp" command-line switch or the
"package omp"_package.html command in your script.
For the KOKKOS package, using this command-line switch also invokes
@ -1833,7 +1833,7 @@ e.g.
Minimization stats:
Stopping criterion = linesearch alpha is zero
Energy initial, next-to-last, final =
Energy initial, next-to-last, final =
-6372.3765206 -8328.46998942 -8328.46998942
Force two-norm initial, final = 1059.36 5.36874
Force max component initial, final = 58.6026 1.46872

View File

@ -104,7 +104,7 @@ since binary files are not compatible across all platforms.
ch2lmp tool :h4,link(charmm)
The ch2lmp sub-directory contains tools for converting files
back-and-forth between the CHARMM MD code and LAMMPS.
back-and-forth between the CHARMM MD code and LAMMPS.
They are intended to make it easy to use CHARMM as a builder and as a
post-processor for LAMMPS. Using charmm2lammps.pl, you can convert a

View File

@ -29,80 +29,80 @@ Bond Styles: fene, harmonic :l
Dihedral Styles: charmm, harmonic, opls :l
Fixes: nve, npt, nvt, nvt/sllod :l
Improper Styles: cvff, harmonic :l
Pair Styles: buck/coul/cut, buck/coul/long, buck, gayberne,
Pair Styles: buck/coul/cut, buck/coul/long, buck, gayberne,
charmm/coul/long, lj/cut, lj/cut/coul/long, sw, tersoff :l
K-Space Styles: pppm :l
:ule
[Speed-ups to expect:]
The speedups will depend on your simulation, the hardware, which
styles are used, the number of atoms, and the floating-point
precision mode. Performance improvements are shown compared to
LAMMPS {without using other acceleration packages} as these are
under active development (and subject to performance changes). The
The speedups will depend on your simulation, the hardware, which
styles are used, the number of atoms, and the floating-point
precision mode. Performance improvements are shown compared to
LAMMPS {without using other acceleration packages} as these are
under active development (and subject to performance changes). The
measurements were performed using the input files available in
the src/USER-INTEL/TEST directory. These are scalable in size; the
results given are with 512K particles (524K for Liquid Crystal).
the src/USER-INTEL/TEST directory. These are scalable in size; the
results given are with 512K particles (524K for Liquid Crystal).
Most of the simulations are standard LAMMPS benchmarks (indicated
by the filename extension in parenthesis) with modifications to the
run length and to add a warmup run (for use with offload
benchmarks).
by the filename extension in parenthesis) with modifications to the
run length and to add a warmup run (for use with offload
benchmarks).
:c,image(JPG/user_intel.png)
Results are speedups obtained on Intel Xeon E5-2697v4 processors
(code-named Broadwell) and Intel Xeon Phi 7250 processors
Results are speedups obtained on Intel Xeon E5-2697v4 processors
(code-named Broadwell) and Intel Xeon Phi 7250 processors
(code-named Knights Landing) with "18 Jun 2016" LAMMPS built with
Intel Parallel Studio 2016 update 3. Results are with 1 MPI task
per physical core. See {src/USER-INTEL/TEST/README} for the raw
Intel Parallel Studio 2016 update 3. Results are with 1 MPI task
per physical core. See {src/USER-INTEL/TEST/README} for the raw
simulation rates and instructions to reproduce.
:line
[Quick Start for Experienced Users:]
LAMMPS should be built with the USER-INTEL package installed.
LAMMPS should be built with the USER-INTEL package installed.
Simulations should be run with 1 MPI task per physical {core},
not {hardware thread}.
For Intel Xeon CPUs:
Edit src/MAKE/OPTIONS/Makefile.intel_cpu_intelmpi as necessary. :ulb,l
If using {kspace_style pppm} in the input script, add "neigh_modify binsize 3" and "kspace_modify diff ad" to the input script for better
If using {kspace_style pppm} in the input script, add "neigh_modify binsize 3" and "kspace_modify diff ad" to the input script for better
performance. :l
"-pk intel 0 omp 2 -sf intel" added to LAMMPS command-line :l
:ule
For Intel Xeon Phi CPUs for simulations without {kspace_style
For Intel Xeon Phi CPUs for simulations without {kspace_style
pppm} in the input script :
Edit src/MAKE/OPTIONS/Makefile.knl as necessary. :ulb,l
Runs should be performed using MCDRAM. :l
"-pk intel 0 omp 2 -sf intel" {or} "-pk intel 0 omp 4 -sf intel"
should be added to the LAMMPS command-line. Choice for best
"-pk intel 0 omp 2 -sf intel" {or} "-pk intel 0 omp 4 -sf intel"
should be added to the LAMMPS command-line. Choice for best
performance will depend on the simulation. :l
:ule
For Intel Xeon Phi CPUs for simulations with {kspace_style
For Intel Xeon Phi CPUs for simulations with {kspace_style
pppm} in the input script:
Edit src/MAKE/OPTIONS/Makefile.knl as necessary. :ulb,l
Runs should be performed using MCDRAM. :l
Add "neigh_modify binsize 3" to the input script for better
Add "neigh_modify binsize 3" to the input script for better
performance. :l
Add "kspace_modify diff ad" to the input script for better
Add "kspace_modify diff ad" to the input script for better
performance. :l
export KMP_AFFINITY=none :l
"-pk intel 0 omp 3 lrt yes -sf intel" or "-pk intel 0 omp 1 lrt yes
-sf intel" added to LAMMPS command-line. Choice for best performance
-sf intel" added to LAMMPS command-line. Choice for best performance
will depend on the simulation. :l
:ule
For Intel Xeon Phi coprocessors (Offload):
For Intel Xeon Phi coprocessors (Offload):
Edit src/MAKE/OPTIONS/Makefile.intel_coprocessor as necessary :ulb,l
"-pk intel N omp 1" added to command-line where N is the number of
"-pk intel N omp 1" added to command-line where N is the number of
coprocessors per node. :l
:ule
@ -111,7 +111,7 @@ coprocessors per node. :l
[Required hardware/software:]
In order to use offload to coprocessors, an Intel Xeon Phi
coprocessor and an Intel compiler are required. For this, the
coprocessor and an Intel compiler are required. For this, the
recommended version of the Intel compiler is 14.0.1.106 or
versions 15.0.2.044 and higher.
@ -133,7 +133,7 @@ slightly lower.
[Notes about Simultaneous Multithreading:]
Modern CPUs often support Simultaneous Multithreading (SMT). On
Modern CPUs often support Simultaneous Multithreading (SMT). On
Intel processors, this is called Hyper-Threading (HT) technology.
SMT is hardware support for running multiple threads efficiently on
a single core. {Hardware threads} or {logical cores} are often used
@ -141,8 +141,8 @@ to refer to the number of threads that are supported in hardware.
For example, the Intel Xeon E5-2697v4 processor is described
as having 36 cores and 72 threads. This means that 36 MPI processes
or OpenMP threads can run simultaneously on separate cores, but that
up to 72 MPI processes or OpenMP threads can be running on the CPU
without costly operating system context switches.
up to 72 MPI processes or OpenMP threads can be running on the CPU
without costly operating system context switches.
Molecular dynamics simulations will often run faster when making use
of SMT. If a thread becomes stalled, for example because it is
@ -150,7 +150,7 @@ waiting on data that has not yet arrived from memory, another thread
can start running so that the CPU pipeline is still being used
efficiently. Although benefits can be seen by launching a MPI task
for every hardware thread, for multinode simulations, we recommend
that OpenMP threads are used for SMT instead, either with the
that OpenMP threads are used for SMT instead, either with the
USER-INTEL package, "USER-OMP package"_accelerate_omp.html", or
"KOKKOS package"_accelerate_kokkos.html. In the example above, up
to 36X speedups can be observed by using all 36 physical cores with
@ -158,10 +158,10 @@ LAMMPS. By using all 72 hardware threads, an additional 10-30%
performance gain can be achieved.
The BIOS on many platforms allows SMT to be disabled, however, we do
not recommend this on modern processors as there is little to no
not recommend this on modern processors as there is little to no
benefit for any software package in most cases. The operating system
will report every hardware thread as a separate core allowing one to
determine the number of hardware threads available. On Linux systems,
will report every hardware thread as a separate core allowing one to
determine the number of hardware threads available. On Linux systems,
this information can normally be obtained with:
cat /proc/cpuinfo :pre
@ -182,21 +182,21 @@ Makefile.intel_cpu_openpmi # Intel Compiler, OpenMPI, No Offload
Makefile.intel_coprocessor # Intel Compiler, Intel MPI, Offload :pre
Makefile.knl is identical to Makefile.intel_cpu_intelmpi except that
it explicitly specifies that vectorization should be for Intel
Xeon Phi x200 processors making it easier to cross-compile. For
users with recent installations of Intel Parallel Studio, the
it explicitly specifies that vectorization should be for Intel
Xeon Phi x200 processors making it easier to cross-compile. For
users with recent installations of Intel Parallel Studio, the
process can be as simple as:
make yes-user-intel
source /opt/intel/parallel_studio_xe_2016.3.067/psxevars.sh
source /opt/intel/parallel_studio_xe_2016.3.067/psxevars.sh
# or psxevars.csh for C-shell
make intel_cpu_intelmpi :pre
Alternatively, the build can be accomplished with the src/Make.py
script, described in "Section 2.4"_Section_start.html#start_4 of the
Alternatively, the build can be accomplished with the src/Make.py
script, described in "Section 2.4"_Section_start.html#start_4 of the
manual. Type "Make.py -h" for help. For an example:
Make.py -v -p intel omp -intel cpu -a file intel_cpu_intelmpi :pre
Make.py -v -p intel omp -intel cpu -a file intel_cpu_intelmpi :pre
Note that if you build with support for a Phi coprocessor, the same
binary can be used on nodes with or without coprocessors installed.
@ -205,26 +205,26 @@ without offload support will produce a smaller binary.
The general requirements for Makefiles with the USER-INTEL package
are as follows. "-DLAMMPS_MEMALIGN=64" is required for CCFLAGS. When
using Intel compilers, "-restrict" is required and "-qopenmp" is
highly recommended for CCFLAGS and LINKFLAGS. LIB should include
using Intel compilers, "-restrict" is required and "-qopenmp" is
highly recommended for CCFLAGS and LINKFLAGS. LIB should include
"-ltbbmalloc". For builds supporting offload, "-DLMP_INTEL_OFFLOAD"
is required for CCFLAGS and "-qoffload" is required for LINKFLAGS.
Other recommended CCFLAG options for best performance are
"-O2 -fno-alias -ansi-alias -qoverride-limits fp-model fast=2
-no-prec-div". The Make.py command will add all of these
Other recommended CCFLAG options for best performance are
"-O2 -fno-alias -ansi-alias -qoverride-limits fp-model fast=2
-no-prec-div". The Make.py command will add all of these
automatically.
NOTE: The vectorization and math capabilities can differ depending on
the CPU. For Intel compilers, the "-x" flag specifies the type of
processor for which to optimize. "-xHost" specifies that the compiler
should build for the processor used for compiling. For Intel Xeon Phi
should build for the processor used for compiling. For Intel Xeon Phi
x200 series processors, this option is "-xMIC-AVX512". For fourth
generation Intel Xeon (v4/Broadwell) processors, "-xCORE-AVX2" should
generation Intel Xeon (v4/Broadwell) processors, "-xCORE-AVX2" should
be used. For older Intel Xeon processors, "-xAVX" will perform best
in general for the different simulations in LAMMPS. The default
in most of the example Makefiles is to use "-xHost", however this
should not be used when cross-compiling.
[Running LAMMPS with the USER-INTEL package:]
Running LAMMPS with the USER-INTEL package is similar to normal use
@ -232,7 +232,7 @@ with the exceptions that one should 1) specify that LAMMPS should use
the USER-INTEL package, 2) specify the number of OpenMP threads, and
3) optionally specify the specific LAMMPS styles that should use the
USER-INTEL package. 1) and 2) can be performed from the command-line
or by editing the input script. 3) requires editing the input script.
or by editing the input script. 3) requires editing the input script.
Advanced performance tuning options are also described below to get
the best performance.
@ -241,14 +241,14 @@ coprocessor), best performance is normally obtained by using 1 MPI
task per physical core and additional OpenMP threads with SMT. For
Intel Xeon processors, 2 OpenMP threads should be used for SMT.
For Intel Xeon Phi CPUs, 2 or 4 OpenMP threads should be used
(best choice depends on the simulation). In cases where the user
specifies that LRT mode is used (described below), 1 or 3 OpenMP
(best choice depends on the simulation). In cases where the user
specifies that LRT mode is used (described below), 1 or 3 OpenMP
threads should be used. For multi-node runs, using 1 MPI task per
physical core will often perform best, however, depending on the
machine and scale, users might get better performance by decreasing
the number of MPI tasks and using more OpenMP threads. For
performance, the product of the number of MPI tasks and OpenMP
threads should not exceed the number of available hardware threads in
the number of MPI tasks and using more OpenMP threads. For
performance, the product of the number of MPI tasks and OpenMP
threads should not exceed the number of available hardware threads in
almost all cases.
NOTE: Setting core affinity is often used to pin MPI tasks and OpenMP
@ -257,21 +257,21 @@ uniform. Unless disabled at build time, affinity for MPI tasks and
OpenMP threads on the host (CPU) will be set by default on the host
{when using offload to a coprocessor}. In this case, it is unnecessary
to use other methods to control affinity (e.g. taskset, numactl,
I_MPI_PIN_DOMAIN, etc.). This can be disabled with the {no_affinity}
option to the "package intel"_package.html command or by disabling the
option at build time (by adding -DINTEL_OFFLOAD_NOAFFINITY to the
CCFLAGS line of your Makefile). Disabling this option is not
recommended, especially when running on a machine with Intel
I_MPI_PIN_DOMAIN, etc.). This can be disabled with the {no_affinity}
option to the "package intel"_package.html command or by disabling the
option at build time (by adding -DINTEL_OFFLOAD_NOAFFINITY to the
CCFLAGS line of your Makefile). Disabling this option is not
recommended, especially when running on a machine with Intel
Hyper-Threading technology disabled.
[Run with the USER-INTEL package from the command line:]
To enable USER-INTEL optimizations for all available styles used in
the input script, the "-sf intel"
To enable USER-INTEL optimizations for all available styles used in
the input script, the "-sf intel"
"command-line switch"_Section_start.html#start_7 can be used without
any requirement for editing the input script. This switch will
automatically append "intel" to styles that support it. It also
invokes a default command: "package intel 1"_package.html. This
automatically append "intel" to styles that support it. It also
invokes a default command: "package intel 1"_package.html. This
package command is used to set options for the USER-INTEL package.
The default package command will specify that USER-INTEL calculations
are performed in mixed precision, that the number of OpenMP threads
@ -281,16 +281,16 @@ support, that 1 coprocessor per node will be used with automatic
balancing of work between the CPU and the coprocessor.
You can specify different options for the USER-INTEL package by using
the "-pk intel Nphi" "command-line switch"_Section_start.html#start_7
the "-pk intel Nphi" "command-line switch"_Section_start.html#start_7
with keyword/value pairs as specified in the documentation. Here,
Nphi = # of Xeon Phi coprocessors/node (ignored without offload
support). Common options to the USER-INTEL package include {omp} to
override any OMP_NUM_THREADS setting and specify the number of OpenMP
threads, {mode} to set the floating-point precision mode, and
{lrt} to enable Long-Range Thread mode as described below. See the
"package intel"_package.html command for details, including the
default values used for all its options if not specified, and how to
set the number of OpenMP threads via the OMP_NUM_THREADS environment
{lrt} to enable Long-Range Thread mode as described below. See the
"package intel"_package.html command for details, including the
default values used for all its options if not specified, and how to
set the number of OpenMP threads via the OMP_NUM_THREADS environment
variable if desired.
Examples (see documentation for your MPI/Machine for differences in
@ -303,7 +303,7 @@ mpirun -np 72 -ppn 36 lmp_machine -sf intel -in in.script -pk intel 0 omp 2 mode
As an alternative to adding command-line arguments, the input script
can be edited to enable the USER-INTEL package. This requires adding
the "package intel"_package.html command to the top of the input
the "package intel"_package.html command to the top of the input
script. For the second example above, this would be:
package intel 0 omp 2 mode double :pre
@ -314,46 +314,46 @@ add an "intel" suffix to the individual style, e.g.:
pair_style lj/cut/intel 2.5 :pre
Alternatively, the "suffix intel"_suffix.html command can be added to
the input script to enable USER-INTEL styles for the commands that
the input script to enable USER-INTEL styles for the commands that
follow in the input script.
[Tuning for Performance:]
NOTE: The USER-INTEL package will perform better with modifications
to the input script when "PPPM"_kspace_style.html is used:
"kspace_modify diff ad"_kspace_modify.html and "neigh_modify binsize
NOTE: The USER-INTEL package will perform better with modifications
to the input script when "PPPM"_kspace_style.html is used:
"kspace_modify diff ad"_kspace_modify.html and "neigh_modify binsize
3"_neigh_modify.html should be added to the input script.
Long-Range Thread (LRT) mode is an option to the "package
Long-Range Thread (LRT) mode is an option to the "package
intel"_package.html command that can improve performance when using
"PPPM"_kspace_style.html for long-range electrostatics on processors
with SMT. It generates an extra pthread for each MPI task. The thread
is dedicated to performing some of the PPPM calculations and MPI
with SMT. It generates an extra pthread for each MPI task. The thread
is dedicated to performing some of the PPPM calculations and MPI
communications. On Intel Xeon Phi x200 series CPUs, this will likely
always improve performance, even on a single node. On Intel Xeon
processors, using this mode might result in better performance when
using multiple nodes, depending on the machine. To use this mode,
specify that the number of OpenMP threads is one less than would
specify that the number of OpenMP threads is one less than would
normally be used for the run and add the "lrt yes" option to the "-pk"
command-line suffix or "package intel" command. For example, if a run
would normally perform best with "-pk intel 0 omp 4", instead use
"-pk intel 0 omp 3 lrt yes". When using LRT, you should set the
environment variable "KMP_AFFINITY=none". LRT mode is not supported
"-pk intel 0 omp 3 lrt yes". When using LRT, you should set the
environment variable "KMP_AFFINITY=none". LRT mode is not supported
when using offload.
Not all styles are supported in the USER-INTEL package. You can mix
the USER-INTEL package with styles from the "OPT"_accelerate_opt.html
package or the "USER-OMP package"_accelerate_omp.html". Of course,
the USER-INTEL package with styles from the "OPT"_accelerate_opt.html
package or the "USER-OMP package"_accelerate_omp.html". Of course,
this requires that these packages were installed at build time. This
can performed automatically by using "-sf hybrid intel opt" or
"-sf hybrid intel omp" command-line options. Alternatively, the "opt"
and "omp" suffixes can be appended manually in the input script. For
the latter, the "package omp"_package.html command must be in the
input script or the "-pk omp Nt" "command-line
switch"_Section_start.html#start_7 must be used where Nt is the
input script or the "-pk omp Nt" "command-line
switch"_Section_start.html#start_7 must be used where Nt is the
number of OpenMP threads. The number of OpenMP threads should not be
set differently for the different packages. Note that the "suffix
hybrid intel omp"_suffix.html command can also be used within the
set differently for the different packages. Note that the "suffix
hybrid intel omp"_suffix.html command can also be used within the
input script to automatically append the "omp" suffix to styles when
USER-INTEL styles are not available.
@ -374,33 +374,33 @@ that MPI runs are performed in MCDRAM.
[Tuning for Offload Performance:]
The default settings for offload should give good performance.
The default settings for offload should give good performance.
When using LAMMPS with offload to Intel coprocessors, best performance
will typically be achieved with concurrent calculations performed on
both the CPU and the coprocessor. This is achieved by offloading only
a fraction of the neighbor and pair computations to the coprocessor or
using "hybrid"_pair_hybrid.html pair styles where only one style uses
the "intel" suffix. For simulations with long-range electrostatics or
bond, angle, dihedral, improper calculations, computation and data
transfer to the coprocessor will run concurrently with computations
the "intel" suffix. For simulations with long-range electrostatics or
bond, angle, dihedral, improper calculations, computation and data
transfer to the coprocessor will run concurrently with computations
and MPI communications for these calculations on the host CPU. This
is illustrated in the figure below for the rhodopsin protein benchmark
running on E5-2697v2 processors with a Intel Xeon Phi 7120p
running on E5-2697v2 processors with a Intel Xeon Phi 7120p
coprocessor. In this plot, the vertical access is time and routines
running at the same time are running concurrently on both the host and
the coprocessor.
:c,image(JPG/offload_knc.png)
The fraction of the offloaded work is controlled by the {balance}
keyword in the "package intel"_package.html command. A balance of 0
runs all calculations on the CPU. A balance of 1 runs all
supported calculations on the coprocessor. A balance of 0.5 runs half
of the calculations on the coprocessor. Setting the balance to -1
(the default) will enable dynamic load balancing that continously
adjusts the fraction of offloaded work throughout the simulation.
Because data transfer cannot be timed, this option typically produces
The fraction of the offloaded work is controlled by the {balance}
keyword in the "package intel"_package.html command. A balance of 0
runs all calculations on the CPU. A balance of 1 runs all
supported calculations on the coprocessor. A balance of 0.5 runs half
of the calculations on the coprocessor. Setting the balance to -1
(the default) will enable dynamic load balancing that continously
adjusts the fraction of offloaded work throughout the simulation.
Because data transfer cannot be timed, this option typically produces
results within 5 to 10 percent of the optimal fixed balance.
If running short benchmark runs with dynamic load balancing, adding a
@ -418,15 +418,15 @@ with 60 cores available for offload and 4 hardware threads per core
each MPI task to use a subset of 10 threads on the coprocessor. Fine
tuning of the number of threads to use per MPI task or the number of
threads to use per core can be accomplished with keyword settings of
the "package intel"_package.html command.
the "package intel"_package.html command.
The USER-INTEL package has two modes for deciding which atoms will be
handled by the coprocessor. This choice is controlled with the {ghost}
keyword of the "package intel"_package.html command. When set to 0,
ghost atoms (atoms at the borders between MPI tasks) are not offloaded
to the card. This allows for overlap of MPI communication of forces
with computation on the coprocessor when the "newton"_newton.html
setting is "on". The default is dependent on the style being used,
The USER-INTEL package has two modes for deciding which atoms will be
handled by the coprocessor. This choice is controlled with the {ghost}
keyword of the "package intel"_package.html command. When set to 0,
ghost atoms (atoms at the borders between MPI tasks) are not offloaded
to the card. This allows for overlap of MPI communication of forces
with computation on the coprocessor when the "newton"_newton.html
setting is "on". The default is dependent on the style being used,
however, better performance may be achieved by setting this option
explictly.
@ -442,10 +442,10 @@ mode is being used and indicating the number of coprocessor threads
per MPI task. Additionally, an offload timing summary is printed at
the end of each run. When offloading, the frequency for "atom
sorting"_atom_modify.html is changed to 1 so that the per-atom data is
effectively sorted at every rebuild of the neighbor lists. All the
available coprocessor threads on each Phi will be divided among MPI
tasks, unless the {tptask} option of the "-pk intel" "command-line
switch"_Section_start.html#start_7 is used to limit the coprocessor
effectively sorted at every rebuild of the neighbor lists. All the
available coprocessor threads on each Phi will be divided among MPI
tasks, unless the {tptask} option of the "-pk intel" "command-line
switch"_Section_start.html#start_7 is used to limit the coprocessor
threads per MPI task.
[Restrictions:]

View File

@ -65,7 +65,7 @@ Make.py -v -p kokkos -kokkos omp -o mpi -a file mpi # or one-line build via Ma
mpirun -np 16 lmp_mpi -k on -sf kk -in in.lj # 1 node, 16 MPI tasks/node, no threads
mpirun -np 2 -ppn 1 lmp_mpi -k on t 16 -sf kk -in in.lj # 2 nodes, 1 MPI task/node, 16 threads/task
mpirun -np 2 lmp_mpi -k on t 8 -sf kk -in in.lj # 1 node, 2 MPI tasks/node, 8 threads/task
mpirun -np 2 lmp_mpi -k on t 8 -sf kk -in in.lj # 1 node, 2 MPI tasks/node, 8 threads/task
mpirun -np 32 -ppn 4 lmp_mpi -k on t 4 -sf kk -in in.lj # 8 nodes, 4 MPI tasks/node, 4 threads/task :pre
specify variables and settings in your Makefile.machine that enable OpenMP, GPU, or Phi support
@ -178,7 +178,7 @@ make kokkos_cuda_mpich :pre
These examples set the KOKKOS-specific OMP, MIC, CUDA variables on the
make command line which requires a GNU-compatible make command. Try
"gmake" if your system's standard make complains.
"gmake" if your system's standard make complains.
NOTE: If you build using make line variables and re-build LAMMPS twice
with different KOKKOS options and the *same* target, e.g. g++ in the
@ -394,7 +394,7 @@ additional parallelism (beyond MPI) will be invoked on the host
CPU(s).
You can compare the performance running in different modes:
run with 1 MPI task/node and N threads/task
run with N MPI tasks/node and 1 thread/task
run with settings in between these extremes :ul
@ -427,7 +427,7 @@ e.g. src/MAKE/Makefile.cuda, is correct for your GPU hardware/software
details).
The -np setting of the mpirun command should set the number of MPI
tasks/node to be equal to the # of physical GPUs on the node.
tasks/node to be equal to the # of physical GPUs on the node.
Use the "-k" "command-line switch"_Section_commands.html#start_7 to
specify the number of GPUs per node, and the number of threads per MPI

View File

@ -96,7 +96,7 @@ variable.
Depending on which styles are accelerated, you should look for a
reduction in the "Pair time", "Bond time", "KSpace time", and "Loop
time" values printed at the end of a run.
time" values printed at the end of a run.
You may see a small performance advantage (5 to 20%) when running a
USER-OMP style (in serial or parallel) with a single thread per MPI

View File

@ -21,11 +21,11 @@ angle_coeff 6 2.1 180.0 :pre
[Description:]
The {dipole} angle style is used to control the orientation of a dipolar
atom within a molecule "(Orsi)"_#Orsi. Specifically, the {dipole} angle
style restrains the orientation of a point dipole mu_j (embedded in atom
'j') with respect to a reference (bond) vector r_ij = r_i - r_j, where 'i'
is another atom of the same molecule (typically, 'i' and 'j' are also
covalently bonded).
atom within a molecule "(Orsi)"_#Orsi. Specifically, the {dipole} angle
style restrains the orientation of a point dipole mu_j (embedded in atom
'j') with respect to a reference (bond) vector r_ij = r_i - r_j, where 'i'
is another atom of the same molecule (typically, 'i' and 'j' are also
covalently bonded).
It is convenient to define an angle gamma between the 'free' vector mu_j
and the reference (bond) vector r_ij:
@ -37,21 +37,21 @@ The {dipole} angle style uses the potential:
:c,image(Eqs/angle_dipole_potential.jpg)
where K is a rigidity constant and gamma0 is an equilibrium (reference)
angle.
angle.
The torque on the dipole can be obtained by differentiating the
potential using the 'chain rule' as in appendix C.3 of
The torque on the dipole can be obtained by differentiating the
potential using the 'chain rule' as in appendix C.3 of
"(Allen)"_#Allen:
:c,image(Eqs/angle_dipole_torque.jpg)
Example: if gamma0 is set to 0 degrees, the torque generated by
the potential will tend to align the dipole along the reference
the potential will tend to align the dipole along the reference
direction defined by the (bond) vector r_ij (in other words, mu_j is
restrained to point towards atom 'i').
The dipolar torque T_j must be counterbalanced in order to conserve
the local angular momentum. This is achieved via an additional force
The dipolar torque T_j must be counterbalanced in order to conserve
the local angular momentum. This is achieved via an additional force
couple generating a torque equivalent to the opposite of T_j:
:c,image(Eqs/angle_dipole_couple.jpg)
@ -118,7 +118,7 @@ This angle style should not be used with SHAKE.
:line
:link(Orsi)
[(Orsi)] Orsi & Essex, The ELBA force field for coarse-grain modeling of
[(Orsi)] Orsi & Essex, The ELBA force field for coarse-grain modeling of
lipid membranes, PloS ONE 6(12): e28637, 2011.
:link(Allen)

View File

@ -62,7 +62,7 @@ more instructions on how to use the accelerated styles effectively.
[Restrictions:]
This angle style can only be used if LAMMPS was built with the
USER_MISC package. See the "Making LAMMPS"_Section_start.html#start_3
USER_MISC package. See the "Making LAMMPS"_Section_start.html#start_3
section for more info on packages.
[Related commands:]

View File

@ -61,7 +61,7 @@ more instructions on how to use the accelerated styles effectively.
[Restrictions:]
This angle style can only be used if LAMMPS was built with the
USER_MISC package. See the "Making LAMMPS"_Section_start.html#start_3
USER_MISC package. See the "Making LAMMPS"_Section_start.html#start_3
section for more info on packages.
[Related commands:]

View File

@ -68,7 +68,7 @@ more instructions on how to use the accelerated styles effectively.
[Restrictions:]
This angle style can only be used if LAMMPS was built with the
USER_MISC package. See the "Making LAMMPS"_Section_start.html#start_3
USER_MISC package. See the "Making LAMMPS"_Section_start.html#start_3
section for more info on packages.
[Related commands:]

View File

@ -43,7 +43,7 @@ internally; hence the units of K are in energy/radian^2.
The also required {lj/sdk} parameters will be extracted automatically
from the pair_style.
[Restrictions:]
[Restrictions:]
This angle style can only be used if LAMMPS was built with the
USER-CG-CMM package. See the "Making

View File

@ -1,4 +1,4 @@
"LAMMPS WWW Site"_lws - "LAMMPS Documentation"_ld - "LAMMPS
"LAMMPS WWW Site"_lws - "LAMMPS Documentation"_ld - "LAMMPS
Commands"_lc :c
:link(lws,http://lammps.sandia.gov)
@ -156,12 +156,12 @@ used with a group-ID that is not "all".
[Default:]
By default, {id} is yes. By default, atomic systems (no bond topology
info) do not use a map. For molecular systems (with bond topology
info), a map is used. The default map style is array if no atom ID is
larger than 1 million, otherwise the default is hash. By default, a
"first" group is not defined. By default, sorting is enabled with a
frequency of 1000 and a binsize of 0.0, which means the neighbor
By default, {id} is yes. By default, atomic systems (no bond topology
info) do not use a map. For molecular systems (with bond topology
info), a map is used. The default map style is array if no atom ID is
larger than 1 million, otherwise the default is hash. By default, a
"first" group is not defined. By default, sorting is enabled with a
frequency of 1000 and a binsize of 0.0, which means the neighbor
cutoff will be used to set the bin size.
:line

View File

@ -14,7 +14,7 @@ atom_style style args :pre
style = {angle} or {atomic} or {body} or {bond} or {charge} or {dipole} or \
{dpd} or {electron} or {ellipsoid} or {full} or {line} or {meso} or \
{molecular} or {peri} or {smd} or {sphere} or {tri} or \
{molecular} or {peri} or {smd} or {sphere} or {tri} or \
{template} or {hybrid} :ulb,l
args = none for any style except the following
{body} args = bstyle bstyle-args
@ -193,7 +193,7 @@ For the {body} style, the particles are arbitrary bodies with internal
attributes defined by the "style" of the bodies, which is specified by
the {bstyle} argument. Body particles can represent complex entities,
such as surface meshes of discrete points, collections of
sub-particles, deformable objects, etc.
sub-particles, deformable objects, etc.
The "body"_body.html doc page descibes the body styles LAMMPS
currently supports, and provides more details as to the kind of body
@ -269,7 +269,7 @@ The {line} and {tri} styles are part of the ASPHERE package.
The {body} style is part of the BODY package.
The {dipole} style is part of the DIPOLE package.
The {dipole} style is part of the DIPOLE package.
The {peri} style is part of the PERI package for Peridynamics.

View File

@ -490,7 +490,7 @@ per processor. Note that the 4 sub-domains share vertices, so there
will be duplicate nodes in the list.
The "SQUARES" section lists the node IDs of the 4 vertices in a
rectangle for each processor (1 to 4).
rectangle for each processor (1 to 4).
For a 3d problem, the syntax is similar with 8 vertices listed for
each processor, instead of 4, and "SQUARES" replaced by "CUBES".

View File

@ -125,7 +125,7 @@ in the {Bodies} section of the data file:
atom-ID 1 M
N
ixx iyy izz ixy ixz iyz
ixx iyy izz ixy ixz iyz
x1 y1 z1
...
xN yN zN :pre
@ -198,11 +198,11 @@ in the {Bodies} section of the data file:
atom-ID 1 M
N
ixx iyy izz ixy ixz iyz
ixx iyy izz ixy ixz iyz
x1 y1 z1
...
xN yN zN
i j j k k ...
i j j k k ...
radius :pre
N is the number of vertices in the body particle. M = 6 + 3*N + 2*N +
@ -230,11 +230,11 @@ particles whose edge length is sqrt(2):
3 1 27
4
1 1 4 0 0 0
-0.7071 -0.7071 0
-0.7071 0.7071 0
0.7071 0.7071 0
0.7071 -0.7071 0
1 1 4 0 0 0
-0.7071 -0.7071 0
-0.7071 0.7071 0
0.7071 0.7071 0
0.7071 -0.7071 0
0 1 1 2 2 3 3 0
1.0 :pre

View File

@ -173,7 +173,7 @@ change_box all x scale 1.1 y volume z volume :pre
The {volume} style changes the associated dimension so that the
overall box volume is unchanged relative to its value before the
preceding keyword was invoked.
preceding keyword was invoked.
If the following command is used, then the z box length will shrink by
the same 1.1 factor the x box length was increased by:

View File

@ -135,7 +135,7 @@ and angular momentum of a particle. If the {vel} option is set to
{yes}, then ghost atoms store these quantities; if {no} then they do
not. The {yes} setting is needed by some pair styles which require
the velocity state of both the I and J particles to compute a pairwise
I,J interaction.
I,J interaction, as well as by some compute and fix commands.
Note that if the "fix deform"_fix_deform.html command is being used
with its "remap v" option enabled, then the velocities for ghost atoms

View File

@ -13,7 +13,7 @@ compute cna/atom command :h3
compute ID group-ID cna/atom cutoff :pre
ID, group-ID are documented in "compute"_compute.html command
cna/atom = style name of this compute command
cna/atom = style name of this compute command
cutoff = cutoff distance for nearest neighbors (distance units) :ul
[Examples:]

View File

@ -63,4 +63,4 @@ LAMMPS"_Section_start.html#start_3 section for more info.
"compute damage/atom"_compute_damage_atom.html,
"compute plasticity/atom"_compute_plasticity_atom.html
[Default:] none
[Default:] none

View File

@ -19,7 +19,7 @@ charge-correction = {mass} or {geometry}, use COM or geometric center for charge
[Examples:]
compute 1 fluid dipole/chunk molchunk
compute 1 fluid dipole/chunk molchunk
compute dw water dipole/chunk 1 geometry :pre
[Description:]

View File

@ -46,7 +46,7 @@ output options.
The vector values will be in energy and temperature "units"_units.html.
[Restrictions:]
[Restrictions:]
This command is part of the USER-DPD package. It is only enabled if
LAMMPS was built with that package. See the "Making
@ -64,7 +64,7 @@ command.
:line
:link(Larentzos)
:link(Larentzos)
[(Larentzos)] J.P. Larentzos, J.K. Brennan, J.D. Moore, and
W.D. Mattson, "LAMMPS Implementation of Constant Energy Dissipative
Particle Dynamics (DPD-E)", ARL-TR-6863, U.S. Army Research

View File

@ -22,7 +22,7 @@ compute 1 all dpd/atom
[Description:]
Define a computation that accesses the per-particle internal
conductive energy (u_cond), internal mechanical energy (u_mech),
conductive energy (u_cond), internal mechanical energy (u_mech),
internal chemical energy (u_chem) and
internal temperatures (dpdTheta) for each particle in a group. See
the "compute dpd"_compute_dpd.html command if you want the total
@ -39,10 +39,10 @@ that uses per-particle values from a compute as input. See
"Section 6.15"_Section_howto.html#howto_15 for an overview of
LAMMPS output options.
The per-particle array values will be in energy (u_cond, u_mech, u_chem)
The per-particle array values will be in energy (u_cond, u_mech, u_chem)
and temperature (dpdTheta) "units"_units.html.
[Restrictions:]
[Restrictions:]
This command is part of the USER-DPD package. It is only enabled if
LAMMPS was built with that package. See the "Making

View File

@ -26,7 +26,7 @@ Define a computation that flags an "event" if any particle in the
group has moved a distance greater than the specified threshold
distance when compared to a previously stored reference state
(i.e. the previous event). This compute is typically used in
conjunction with the "prd"_prd.html and "tad"_tad.html commands,
conjunction with the "prd"_prd.html and "tad"_tad.html commands,
to detect if a transition
to a new minimum energy basin has occurred.
@ -34,8 +34,8 @@ This value calculated by the compute is equal to 0 if no particle has
moved far enough, and equal to 1 if one or more particles have moved
further than the threshold distance.
NOTE: If the system is undergoing significant center-of-mass motion,
due to thermal motion, an external force, or an initial net momentum,
NOTE: If the system is undergoing significant center-of-mass motion,
due to thermal motion, an external force, or an initial net momentum,
then this compute will not be able to distinguish that motion from
local atom displacements and may generate "false postives."

View File

@ -64,7 +64,7 @@ these atoms:
A coupling parameter \(\lambda\) varying from 0 to 1 connects the
reference and perturbed systems:
:c,image(Eqs/compute_fep_lambda.jpg)
:c,image(Eqs/compute_fep_lambda.jpg)
It is possible but not necessary that the coupling parameter (or a
function thereof) appears as a multiplication factor of the potential

View File

@ -28,7 +28,7 @@ compute 2 molecule gyration/chunk molchunk tensor :pre
[Description:]
Define a computation that calculates the radius of gyration Rg for
multiple chunks of atoms.
multiple chunks of atoms.
In LAMMPS, chunks are collections of atoms defined by a "compute
chunk/atom"_compute_chunk_atom.html command, which assigns each atom

View File

@ -20,7 +20,7 @@ stress-ID = ID of a compute that calculates per-atom stress :ul
[Examples:]
compute myFlux all heat/flux myKE myPE myStress :pre
compute myFlux all heat/flux myKE myPE myStress :pre
[Description:]
@ -38,7 +38,7 @@ subtracted to a group of atoms.
The compute takes three arguments which are IDs of other
"computes"_compute.html. One calculates per-atom kinetic energy
({ke-ID}), one calculates per-atom potential energy ({pe-ID)}, and the
third calcualtes per-atom stress ({stress-ID}).
third calcualtes per-atom stress ({stress-ID}).
NOTE: These other computes should provide values for all the atoms in
the group this compute specifies. That means the other computes could
@ -152,11 +152,11 @@ lattice fcc 5.376 orient x 1 0 0 orient y 0 1 0 orient z 0 0 1
region box block 0 4 0 4 0 4
create_box 1 box
create_atoms 1 box
mass 1 39.948
mass 1 39.948
pair_style lj/cut 13.0
pair_coeff * * 0.2381 3.405
timestep $\{dt\}
thermo $d :pre
thermo $d :pre
# equilibration and thermalization :pre

View File

@ -15,7 +15,7 @@ compute ID group-ID hexorder/atom keyword values ... :pre
ID, group-ID are documented in "compute"_compute.html command :ulb,l
hexorder/atom = style name of this compute command :l
one or more keyword/value pairs may be appended :l
keyword = {degree} or {nnn} or {cutoff}
keyword = {degree} or {nnn} or {cutoff}
{cutoff} value = distance cutoff
{nnn} value = number of nearest neighbors
{degree} value = degree {n} of order parameter :pre
@ -24,27 +24,27 @@ keyword = {degree} or {nnn} or {cutoff}
[Examples:]
compute 1 all hexorder/atom
compute 1 all hexorder/atom
compute 1 all hexorder/atom degree 4 nnn 4 cutoff 1.2 :pre
[Description:]
Define a computation that calculates {qn} the bond-orientational
order parameter for each atom in a group. The hexatic ({n} = 6) order
Define a computation that calculates {qn} the bond-orientational
order parameter for each atom in a group. The hexatic ({n} = 6) order
parameter was introduced by "Nelson and Halperin"_#Nelson as a way to detect
hexagonal symmetry in two-dimensional systems. For each atom, {qn}
hexagonal symmetry in two-dimensional systems. For each atom, {qn}
is a complex number (stored as two real numbers) defined as follows:
:c,image(Eqs/hexorder.jpg)
where the sum is over the {nnn} nearest neighbors
where the sum is over the {nnn} nearest neighbors
of the central atom. The angle theta
is formed by the bond vector rij and the {x} axis. theta is calculated
only using the {x} and {y} components, whereas the distance from the
central atom is calculated using all three
central atom is calculated using all three
{x}, {y}, and {z} components of the bond vector.
Neighbor atoms not in the group
are included in the order parameter of atoms in the group.
Neighbor atoms not in the group
are included in the order parameter of atoms in the group.
The optional keyword {cutoff} defines the distance cutoff
used when searching for neighbors. The default value, also
@ -53,22 +53,22 @@ by the pair style.
The optional keyword {nnn} defines the number of nearest
neighbors used to calculate {qn}. The default value is 6.
If the value is NULL, then all neighbors up to the
If the value is NULL, then all neighbors up to the
distance cutoff are used.
The optional keyword {degree} sets the degree {n} of the order parameter.
The default value is 6. For a perfect hexagonal lattice with
The optional keyword {degree} sets the degree {n} of the order parameter.
The default value is 6. For a perfect hexagonal lattice with
{nnn} = 6,
{q}6 = exp(6 i phi) for all atoms, where the constant 0 < phi < pi/3
depends only on the orientation of the lattice relative to the {x} axis.
In an isotropic liquid, local neighborhoods may still exhibit
{q}6 = exp(6 i phi) for all atoms, where the constant 0 < phi < pi/3
depends only on the orientation of the lattice relative to the {x} axis.
In an isotropic liquid, local neighborhoods may still exhibit
weak hexagonal symmetry, but because the orientational correlation
decays quickly with distance, the value of phi will be different for
different atoms, and so when {q}6 is averaged over all the atoms
different atoms, and so when {q}6 is averaged over all the atoms
in the system, \|<{q}6>\| << 1.
The value of {qn} is set to zero for atoms not in the
specified compute group, as well as for atoms that have less than
specified compute group, as well as for atoms that have less than
{nnn} neighbors within the distance cutoff.
The neighbor list needed to compute this quantity is constructed each
@ -92,7 +92,7 @@ the neighbor list.
[Output info:]
This compute calculates a per-atom array with 2 columns, giving the
real and imaginary parts {qn}, a complex number restricted to the
real and imaginary parts {qn}, a complex number restricted to the
unit disk of the complex plane i.e. Re({qn})^2 + Im({qn})^2 <= 1 .
These values can be accessed by any command that uses
@ -106,7 +106,7 @@ options.
"compute orientorder/atom"_compute_orientorder_atom.html, "compute coord/atom"_compute_coord_atom.html, "compute centro/atom"_compute_centro_atom.html
[Default:]
[Default:]
The option defaults are {cutoff} = pair style cutoff, {nnn} = 6, {degree} = 6

View File

@ -23,7 +23,7 @@ compute 1 fluid inertia/chunk molchunk :pre
[Description:]
Define a computation that calculates the inertia tensor for multiple
chunks of atoms.
chunks of atoms.
In LAMMPS, chunks are collections of atoms defined by a "compute
chunk/atom"_compute_chunk_atom.html command, which assigns each atom

View File

@ -48,9 +48,9 @@ thermodynamic output by using the "thermo_modify"_thermo_modify.html
command, as shown in the following example:
compute effTemp all temp/eff
thermo_style custom step etotal pe ke temp press
thermo_style custom step etotal pe ke temp press
thermo_modify temp effTemp :pre
The value of the kinetic energy will be 0.0 for atoms (nuclei or
electrons) not in the specified compute group.

View File

@ -52,9 +52,9 @@ printed with thermodynamic output by using the
example:
compute effTemp all temp/eff
thermo_style custom step etotal pe ke temp press
thermo_style custom step etotal pe ke temp press
thermo_modify temp effTemp :pre
See "compute temp/eff"_compute_temp_eff.html.
[Output info:]

View File

@ -61,4 +61,4 @@ the temperature is correctly normalized.
[Default:]
The option defaults are extra = 2 or 3 for 2d or 3d systems and
dynamic = no.
dynamic = no.

View File

@ -44,7 +44,7 @@ proportional to the diffusion coefficient of the diffusing atoms.
The displacement of an atom is from its reference position. This is
normally the original position at the time
the compute command was issued, unless the {average} keyword is set to {yes}.
the compute command was issued, unless the {average} keyword is set to {yes}.
The value of the displacement will be
0.0 for atoms not in the specified compute group.

View File

@ -15,7 +15,7 @@ compute ID group-ID orientorder/atom keyword values ... :pre
ID, group-ID are documented in "compute"_compute.html command :ulb,l
orientorder/atom = style name of this compute command :l
one or more keyword/value pairs may be appended :l
keyword = {cutoff} or {nnn} or {ql}
keyword = {cutoff} or {nnn} or {ql}
{cutoff} value = distance cutoff
{nnn} value = number of nearest neighbors
{degrees} values = nlvalues, l1, l2,... :pre
@ -24,30 +24,30 @@ keyword = {cutoff} or {nnn} or {ql}
[Examples:]
compute 1 all orientorder/atom
compute 1 all orientorder/atom
compute 1 all orientorder/atom degrees 5 4 6 8 10 12 nnn NULL cutoff 1.5 :pre
[Description:]
Define a computation that calculates a set of bond-orientational
Define a computation that calculates a set of bond-orientational
order parameters {Ql} for each atom in a group. These order parameters
were introduced by "Steinhardt et al."_#Steinhardt as a way to
characterize the local orientational order in atomic structures.
characterize the local orientational order in atomic structures.
For each atom, {Ql} is a real number defined as follows:
:c,image(Eqs/orientorder.jpg)
The first equation defines the spherical harmonic order parameters.
These are complex number components of the 3D analog of the 2D order
parameter {qn}, which is implemented as LAMMPS compute
"hexorder/atom"_compute_hexorder_atom.html.
The summation is over the {nnn} nearest
neighbors of the central atom.
The angles theta and phi are the standard spherical polar angles
The first equation defines the spherical harmonic order parameters.
These are complex number components of the 3D analog of the 2D order
parameter {qn}, which is implemented as LAMMPS compute
"hexorder/atom"_compute_hexorder_atom.html.
The summation is over the {nnn} nearest
neighbors of the central atom.
The angles theta and phi are the standard spherical polar angles
defining the direction of the bond vector {rij}.
The second equation defines {Ql}, which is a
rotationally invariant scalar quantity obtained by summing
over all the components of degree {l}.
rotationally invariant scalar quantity obtained by summing
over all the components of degree {l}.
The optional keyword {cutoff} defines the distance cutoff
used when searching for neighbors. The default value, also
@ -56,23 +56,23 @@ by the pair style.
The optional keyword {nnn} defines the number of nearest
neighbors used to calculate {Ql}. The default value is 12.
If the value is NULL, then all neighbors up to the
If the value is NULL, then all neighbors up to the
specified distance cutoff are used.
The optional keyword {degrees} defines the list of order parameters to
be computed. The first argument {nlvalues} is the number of order
be computed. The first argument {nlvalues} is the number of order
parameters. This is followed by that number of integers giving the
degree of each order parameter. Because {Q}2 and all odd-degree
order parameters are zero for atoms in cubic crystals
degree of each order parameter. Because {Q}2 and all odd-degree
order parameters are zero for atoms in cubic crystals
(see "Steinhardt"_#Steinhardt), the default order parameters
are {Q}4, {Q}6, {Q}8, {Q}10, and {Q}12. For the
FCC crystal with {nnn}=12, {Q}4 = sqrt(7/3)/8 = 0.19094....
The numerical values of all order parameters up to {Q}12
for a range of commonly encountered high-symmetry structures are given
for a range of commonly encountered high-symmetry structures are given
in Table I of "Mickel et al."_#Mickel.
The value of {Ql} is set to zero for atoms not in the
specified compute group, as well as for atoms that have less than
specified compute group, as well as for atoms that have less than
{nnn} neighbors within the distance cutoff.
The neighbor list needed to compute this quantity is constructed each
@ -109,9 +109,9 @@ options.
"compute coord/atom"_compute_coord_atom.html, "compute centro/atom"_compute_centro_atom.html, "compute hexorder/atom"_compute_hexorder_atom.html
[Default:]
[Default:]
The option defaults are {cutoff} = pair style cutoff, {nnn} = 12, {degrees} = 5 4 6 8 9 10 12 i.e. {Q}4, {Q}6, {Q}8, {Q}10, and {Q}12.
The option defaults are {cutoff} = pair style cutoff, {nnn} = 12, {degrees} = 5 4 6 8 9 10 12 i.e. {Q}4, {Q}6, {Q}8, {Q}10, and {Q}12.
:line

View File

@ -52,7 +52,7 @@ The KSpace contribution is calculated using the method in
"(Heyes)"_#Heyes for the Ewald method and a related method for PPPM,
as specified by the "kspace_style pppm"_kspace_style.html command.
For PPPM, the calcluation requires 1 extra FFT each timestep that
per-atom energy is calculated. Thie "document"_PDF/kspace.pdf
per-atom energy is calculated. This "document"_PDF/kspace.pdf
describes how the long-range per-atom energy calculation is performed.
Various fixes can contribute to the per-atom potential energy of the
@ -68,13 +68,14 @@ As an example of per-atom potential energy compared to total potential
energy, these lines in an input script should yield the same result
in the last 2 columns of thermo output:
compute peratom all pe/atom
compute pe all reduce sum c_peratom
thermo_style custom step temp etotal press pe c_pe :pre
compute peratom all pe/atom
compute pe all reduce sum c_peratom
thermo_style custom step temp etotal press pe c_pe :pre
NOTE: The per-atom energy does not any Lennard-Jones tail corrections
invoked by the "pair_modify tail yes"_pair_modify.html command, since
those are global contributions to the system energy.
NOTE: The per-atom energy does not include any Lennard-Jones tail
corrections to the energy added by the "pair_modify tail
yes"_pair_modify.html command, since those are contributions to the
global system energy.
[Output info:]

View File

@ -30,7 +30,7 @@ The plasticity for a Peridynamic particle is the so-called consistency
parameter (lambda). For elastic deformation lambda = 0, otherwise
lambda > 0 for plastic deformation. For details, see
"(Mitchell)"_#Mitchell and the PDF doc included in the LAMMPS
distro in "doc/PDF/PDLammps_EPS.pdf"_PDF/PDLammps_EPS.pdf.
distro in "doc/PDF/PDLammps_EPS.pdf"_PDF/PDLammps_EPS.pdf.
This command can be invoked for one of the Peridynamic "pair
styles"_pair_peri.html: peri/eps.
@ -57,7 +57,7 @@ LAMMPS"_Section_start.html#start_3 section for more info.
"compute damage/atom"_compute_damage_atom.html,
"compute dilatation/atom"_compute_dilatation_atom.html
[Default:] none
[Default:] none
:line

View File

@ -50,7 +50,7 @@ ordered xx, yy, zz, xy, xz, yz. The equation for the I,J components
(where I and J = x,y,z) is similar to the above formula, except that
the first term uses components of the kinetic energy tensor and the
second term uses components of the virial tensor:
:c,image(Eqs/pressure_tensor.jpg)
If no extra keywords are listed, the entire equations above are

View File

@ -16,20 +16,20 @@ ID, group-ID are documented in "compute"_compute.html command :ulb,l
property/atom = style name of this compute command :l
input = one or more atom attributes :l
possible attributes = id, mol, proc, type, mass,
x, y, z, xs, ys, zs, xu, yu, zu, ix, iy, iz,
vx, vy, vz, fx, fy, fz,
x, y, z, xs, ys, zs, xu, yu, zu, ix, iy, iz,
vx, vy, vz, fx, fy, fz,
q, mux, muy, muz, mu,
radius, diameter, omegax, omegay, omegaz,
angmomx, angmomy, angmomz,
shapex,shapey, shapez,
quatw, quati, quatj, quatk, tqx, tqy, tqz,
end1x, end1y, end1z, end2x, end2y, end2z,
corner1x, corner1y, corner1z,
corner2x, corner2y, corner2z,
corner3x, corner3y, corner3z,
nbonds,
angmomx, angmomy, angmomz,
shapex,shapey, shapez,
quatw, quati, quatj, quatk, tqx, tqy, tqz,
end1x, end1y, end1z, end2x, end2y, end2z,
corner1x, corner1y, corner1z,
corner2x, corner2y, corner2z,
corner3x, corner3y, corner3z,
nbonds,
vfrac, s0,
spin, eradius, ervel, erforce,
spin, eradius, ervel, erforce,
rho, drho, e, de, cv,
i_name, d_name :pre
id = atom ID
@ -80,7 +80,7 @@ input = one or more atom attributes :l
[Examples:]
compute 1 all property/atom xs vx fx mux
compute 1 all property/atom xs vx fx mux
compute 2 all property/atom type
compute 1 all property/atom ix iy iz :pre

View File

@ -16,7 +16,7 @@ ID, group-ID are documented in "compute"_compute.html command :ulb,l
property/chunk = style name of this compute command :l
input = one or more attributes :l
attributes = count, id, coord1, coord2, coord3
count = # of atoms in chunk
count = # of atoms in chunk
id = original chunk IDs before compression by "compute chunk/atom"_compute_chunk_atom.html
coord123 = coordinates for spatial bins calculated by "compute chunk/atom"_compute_chunk_atom.html :pre
:ule

View File

@ -15,12 +15,12 @@ compute ID group-ID property/local attribute1 attribute2 ... keyword args ... :p
ID, group-ID are documented in "compute"_compute.html command :ulb,l
property/local = style name of this compute command :l
one or more attributes may be appended :l
possible attributes = natom1 natom2 ntype1 ntype2
patom1 patom2 ptype1 ptype2
batom1 batom2 btype
aatom1 aatom2 aatom3 atype
datom1 datom2 datom3 dtype
iatom1 iatom2 iatom3 itype :pre
possible attributes = natom1 natom2 ntype1 ntype2
patom1 patom2 ptype1 ptype2
batom1 batom2 btype
aatom1 aatom2 aatom3 atype
datom1 datom2 datom3 dtype
iatom1 iatom2 iatom3 itype :pre
natom1, natom2 = IDs of 2 atoms in each pair (within neighbor cutoff)
ntype1, ntype2 = type of 2 atoms in each pair (within neighbor cutoff)
@ -129,8 +129,6 @@ The attributes that start with "a", "d", "i", refer to similar values
for "angles"_angle_style.html, "dihedrals"_dihedral_style.html, and
"impropers"_improper_style.html.
The optional {cutoff} keyword
[Output info:]
This compute calculates a local vector or local array depending on the

View File

@ -155,8 +155,8 @@ Thus, for example, if you wish to use this compute to find the bond
with maximum stretch, you can do it as follows:
compute 1 all property/local batom1 batom2
compute 2 all bond/local dist
compute 3 all reduce max c_1\[1\] c_1\[2\] c_2 replace 1 3 replace 2 3
compute 2 all bond/local dist
compute 3 all reduce max c_1\[1\] c_1\[2\] c_2 replace 1 3 replace 2 3
thermo_style custom step temp c_3\[1\] c_3\[2\] c_3\[3\] :pre
The first two input values in the compute reduce command are vectors

View File

@ -17,11 +17,11 @@ rigid/local = style name of this compute command :l
rigidID = ID of fix rigid/small command or one of its variants :l
input = one or more rigid body attributes :l
possible attributes = id, mol, mass,
x, y, z, xu, yu, zu, ix, iy, iz
vx, vy, vz, fx, fy, fz,
x, y, z, xu, yu, zu, ix, iy, iz
vx, vy, vz, fx, fy, fz,
omegax, omegay, omegaz,
angmomx, angmomy, angmomz,
quatw, quati, quatj, quatk,
angmomx, angmomy, angmomz,
quatw, quati, quatj, quatk,
tqx, tqy, tqz,
inertiax, inertiay, inertiaz
id = atom ID of atom within body which owns body properties
@ -29,7 +29,7 @@ input = one or more rigid body attributes :l
mass = total mass of body
x,y,z = center of mass coords of body
xu,yu,zu = unwrapped center of mass coords of body
ix,iy,iz = box image that the center of mass is in
ix,iy,iz = box image that the center of mass is in
vx,vy,vz = center of mass velocities
fx,fy,fz = force of center of mass
omegax,omegay,omegaz = angular velocity of body
@ -71,7 +71,7 @@ it is skipped (only one atom per body is so assigned). If it is the
assigned atom, then the info for that body is output. This means that
information for N bodies is generated. N may be less than the # of
bodies defined by the fix rigid command, if the atoms in some bodies
are not in the {group-ID}.
are not in the {group-ID}.
NOTE: Which atom in a body owns the body info is determined internal
to LAMMPS; it's the one nearest the geometric center of the body.
@ -109,7 +109,7 @@ sigma, etc). Use {xu}, {yu}, {zu} if you want the COM "unwrapped" by
the image flags for each atobody. Unwrapped means that if the body
COM has passed thru a periodic boundary one or more times, the value
is generated what the COM coordinate would be if it had not been
wrapped back into the periodic box.
wrapped back into the periodic box.
The image flags for the body can be generated directly using the {ix},
{iy}, {iz} attributes. For periodic dimensions, they specify which

View File

@ -19,18 +19,18 @@ type1 type2 ... typeN = chemical symbol of each atom type (see valid options bel
zero or more keyword/value pairs may be appended :l
keyword = {Kmax} or {Zone} or {dR_Ewald} or {c} or {manual} or {echo} :l
{Kmax} value = Maximum distance explored from reciprocal space origin
{Kmax} value = Maximum distance explored from reciprocal space origin
(inverse length units)
{Zone} values = z1 z2 z3
z1,z2,z3 = Zone axis of incident radiation. If z1=z2=z3=0 all
z1,z2,z3 = Zone axis of incident radiation. If z1=z2=z3=0 all
reciprocal space will be meshed up to {Kmax}
{dR_Ewald} value = Thickness of Ewald sphere slice intercepting
{dR_Ewald} value = Thickness of Ewald sphere slice intercepting
reciprocal space (inverse length units)
{c} values = c1 c2 c3
c1,c2,c3 = parameters to adjust the spacing of the reciprocal
c1,c2,c3 = parameters to adjust the spacing of the reciprocal
lattice nodes in the h, k, and l directions respectively
{manual} = flag to use manual spacing of reciprocal lattice points
based on the values of the {c} parameters
{manual} = flag to use manual spacing of reciprocal lattice points
based on the values of the {c} parameters
{echo} = flag to provide extra output for debugging purposes :pre
:ule
@ -44,22 +44,22 @@ fix saed/vtk 1 1 1 c_2 file Ni_000.saed :pre
[Description:]
Define a computation that calculates electron diffraction intensity as
described in "(Coleman)"_#saed-Coleman on a mesh of reciprocal lattice nodes
defined by the entire simulation domain (or manually) using simulated
radiation of wavelength lambda.
Define a computation that calculates electron diffraction intensity as
described in "(Coleman)"_#saed-Coleman on a mesh of reciprocal lattice nodes
defined by the entire simulation domain (or manually) using simulated
radiation of wavelength lambda.
The electron diffraction intensity I at each reciprocal lattice point
The electron diffraction intensity I at each reciprocal lattice point
is computed from the structure factor F using the equations:
:c,image(Eqs/compute_saed1.jpg)
:c,image(Eqs/compute_saed1.jpg)
:c,image(Eqs/compute_saed2.jpg)
Here, K is the location of the reciprocal lattice node, rj is the
Here, K is the location of the reciprocal lattice node, rj is the
position of each atom, fj are atomic scattering factors.
Diffraction intensities are calculated on a three-dimensional mesh of
reciprocal lattice nodes. The mesh spacing is defined either (a) by
Diffraction intensities are calculated on a three-dimensional mesh of
reciprocal lattice nodes. The mesh spacing is defined either (a) by
the entire simulation domain or (b) manually using selected values as
shown in the 2D diagram below.
@ -74,12 +74,12 @@ average of the (inversed) box lengths with periodic boundary conditions.
Meshes defined by the simulation domain must contain at least one periodic
boundary.
If the {manual} flag is included, the mesh of reciprocal lattice nodes
will defined using the {c} values for the spacing along each reciprocal
lattice axis. Note that manual mapping of the reciprocal space mesh is
good for comparing diffraction results from multiple simulations; however
it can reduce the likelihood that Bragg reflections will be satisfied
unless small spacing parameters <0.05 Angstrom^(-1) are implemented.
If the {manual} flag is included, the mesh of reciprocal lattice nodes
will defined using the {c} values for the spacing along each reciprocal
lattice axis. Note that manual mapping of the reciprocal space mesh is
good for comparing diffraction results from multiple simulations; however
it can reduce the likelihood that Bragg reflections will be satisfied
unless small spacing parameters <0.05 Angstrom^(-1) are implemented.
Meshes with manual spacing do not require a periodic boundary.
The limits of the reciprocal lattice mesh are determined by the use of
@ -98,17 +98,17 @@ in the below image.
:c,image(JPG/saed_ewald_intersect_small.jpg,JPG/saed_ewald_intersect.jpg)
The atomic scattering factors, fj, accounts for the reduction in
diffraction intensity due to Compton scattering. Compute saed uses
analytical approximations of the atomic scattering factors that vary
for each atom type (type1 type2 ... typeN) and angle of diffraction.
The atomic scattering factors, fj, accounts for the reduction in
diffraction intensity due to Compton scattering. Compute saed uses
analytical approximations of the atomic scattering factors that vary
for each atom type (type1 type2 ... typeN) and angle of diffraction.
The analytic approximation is computed using the formula
"(Brown)"_#Brown:
:c,image(Eqs/compute_saed3.jpg)
Coefficients parameterized by "(Fox)"_#Fox are assigned for each
atom type designating the chemical symbol and charge of each atom
Coefficients parameterized by "(Fox)"_#Fox are assigned for each
atom type designating the chemical symbol and charge of each atom
type. Valid chemical symbols for compute saed are:
H: He: Li: Be: B:
@ -133,14 +133,14 @@ type. Valid chemical symbols for compute saed are:
Cm: Bk: Cf:tb(c=5,s=:)
If the {echo} keyword is specified, compute saed will provide extra
reporting information to the screen.
If the {echo} keyword is specified, compute saed will provide extra
reporting information to the screen.
[Output info:]
This compute calculates a global vector. The length of the vector is
the number of reciprocal lattice nodes that are explored by the mesh.
The entries of the global vector are the computed diffraction
This compute calculates a global vector. The length of the vector is
the number of reciprocal lattice nodes that are explored by the mesh.
The entries of the global vector are the computed diffraction
intensities as described above.
The vector can be accessed by any command that uses global values
@ -148,21 +148,21 @@ from a compute as input. See "this
section"_Section_howto.html#howto_15 for an overview of LAMMPS output
options.
All array values calculated by this compute are "intensive".
All array values calculated by this compute are "intensive".
[Restrictions:]
[Restrictions:]
This compute is part of the USER-DIFFRACTION package. It is only
enabled if LAMMPS was built with that package. See the "Making
LAMMPS"_Section_start.html#start_3 section for more info.
The compute_saed command does not work for triclinic cells.
The compute_saed command does not work for triclinic cells.
[Related commands:]
[Related commands:]
"fix saed_vtk"_fix_saed_vtk.html, "compute xrd"_compute_xrd.html
[Default:]
[Default:]
The option defaults are Kmax = 1.70, Zone 1 0 0, c 1 1 1, dR_Ewald =
0.01.
@ -174,7 +174,7 @@ The option defaults are Kmax = 1.70, Zone 1 0 0, c 1 1 1, dR_Ewald =
(2013).
:link(Brown)
[(Brown)] Brown et al. International Tables for Crystallography
[(Brown)] Brown et al. International Tables for Crystallography
Volume C: Mathematical and Chemical Tables, 554-95 (2004).
:link(Fox)

View File

@ -22,7 +22,7 @@ compute 1 all smd/damage :pre
[Description:]
Define a computation that calculates the damage status of SPH particles
according to the damage model which is defined via the SMD SPH pair styles, e.g., the maximum plastic strain failure criterion.
according to the damage model which is defined via the SMD SPH pair styles, e.g., the maximum plastic strain failure criterion.
See "this PDF guide"_USER/smd/SMD_LAMMPS_userguide.pdf to use Smooth Mach Dynamics in LAMMPS.

View File

@ -31,7 +31,7 @@ keyword = {diagonal} or {rmin0} or {switchflag} :l
{2} = subset satisfying j1 == j2 == j3
{3} = subset satisfying j2 <= j1 <= j
{rmin0} value = parameter in distance to angle conversion (distance units)
{switchflag} value = {0} or {1}
{switchflag} value = {0} or {1}
{0} = do not use switching function
{1} = use switching function :pre
:ule
@ -60,12 +60,12 @@ onto the 3-sphere, the surface of the unit ball in a four-dimensional
space. The radial distance {r} within {R_ii'} is mapped on to a third
polar angle {theta0} defined by,
:c,image(Eqs/compute_sna_atom1.jpg)
:c,image(Eqs/compute_sna_atom1.jpg)
In this way, all possible neighbor positions are mapped on to a subset
of the 3-sphere. Points south of the latitude {theta0max=rfac0*Pi}
are excluded.
The natural basis for functions on the 3-sphere is formed by the 4D
hyperspherical harmonics {U^j_m,m'(theta, phi, theta0).} These
functions are better known as {D^j_m,m',} the elements of the Wigner
@ -78,7 +78,7 @@ radial distance. Expanding this density function as a generalized
Fourier series in the basis functions, we can write each Fourier
coefficient as
:c,image(Eqs/compute_sna_atom2.jpg)
:c,image(Eqs/compute_sna_atom2.jpg)
The {w_i'} neighbor weights are dimensionless numbers that are chosen
to distinguish atoms of different types, while the central atom is
@ -86,7 +86,7 @@ arbitrarily assigned a unit weight. The function {fc(r)} ensures that
the contribution of each neighbor atom goes smoothly to zero at
{R_ii'}:
:c,image(Eqs/compute_sna_atom4.jpg)
:c,image(Eqs/compute_sna_atom4.jpg)
The expansion coefficients {u^j_m,m'} are complex-valued and they are
not directly useful as descriptors, because they are not invariant
@ -94,7 +94,7 @@ under rotation of the polar coordinate frame. However, the following
scalar triple products of expansion coefficients can be shown to be
real-valued and invariant under rotation "(Bartok)"_#Bartok2010.
:c,image(Eqs/compute_sna_atom3.jpg)
:c,image(Eqs/compute_sna_atom3.jpg)
The constants {H^jmm'_j1m1m1'_j2m2m2'} are coupling coefficients,
analogous to Clebsch-Gordan coefficients for rotations on the
@ -112,17 +112,17 @@ atom.
Compute {snad/atom} calculates the derivative of the bispectrum components
summed separately for each atom type:
:c,image(Eqs/compute_sna_atom5.jpg)
:c,image(Eqs/compute_sna_atom5.jpg)
The sum is over all atoms {i'} of atom type {I}. For each atom {i},
this compute evaluates the above expression for each direction, each
atom type, and each bispectrum component. See section below on output
for a detailed explanation.
Compute {snav/atom} calculates the virial contribution due to the
derivatives:
:c,image(Eqs/compute_sna_atom6.jpg)
:c,image(Eqs/compute_sna_atom6.jpg)
Again, the sum is over all atoms {i'} of atom type {I}. For each atom
{i}, this compute evaluates the above expression for each of the six
@ -140,7 +140,7 @@ too frequently.
The argument {rcutfac} is a scale factor that controls the ratio of
atomic radius to radial cutoff distance.
The argument {rfac0} and the optional keyword {rmin0} define the
linear mapping from radial distance to polar angle {theta0} on the
3-sphere.
@ -176,18 +176,18 @@ each column depend on the values of {twojmax} and {diagonal}, as
described by the following piece of python code:
for j1 in range(0,twojmax+1):
if(diagonal==2):
if(diagonal==2):
print j1/2.,j1/2.,j1/2.
elif(diagonal==1):
for j in range(0,min(twojmax,2*j1)+1,2):
for j in range(0,min(twojmax,2*j1)+1,2):
print j1/2.,j1/2.,j/2.
elif(diagonal==0):
for j2 in range(0,j1+1):
for j in range(j1-j2,min(twojmax,j1+j2)+1,2):
for j in range(j1-j2,min(twojmax,j1+j2)+1,2):
print j1/2.,j2/2.,j/2.
elif(diagonal==3):
for j2 in range(0,j1+1):
for j in range(j1-j2,min(twojmax,j1+j2)+1,2):
for j in range(j1-j2,min(twojmax,j1+j2)+1,2):
if (j>=j1): print j1/2.,j2/2.,j/2. :pre
Compute {snad/atom} evaluates a per-atom array. The columns are
@ -227,7 +227,7 @@ The optional keyword defaults are {diagonal} = 0, {rmin0} = 0,
:line
:link(Thompson2014)
[(Thompson)] Thompson, Swiler, Trott, Foiles, Tucker, under review, preprint
[(Thompson)] Thompson, Swiler, Trott, Foiles, Tucker, under review, preprint
available at "arXiv:1409.3880"_http://arxiv.org/abs/1409.3880
:link(Bartok2010)
@ -235,7 +235,7 @@ available at "arXiv:1409.3880"_http://arxiv.org/abs/1409.3880
:link(Meremianin2006)
[(Meremianin)] Meremianin, J. Phys. A, 39, 3099 (2006).
:link(Varshalovich1987)
[(Varshalovich)] Varshalovich, Moskalev, Khersonskii, Quantum Theory
of Angular Momentum, World Scientific, Singapore (1987).

View File

@ -128,10 +128,15 @@ d = dimension and V is the volume of the system, the result should be
These lines in an input script for a 3d system should yield that
result. I.e. the last 2 columns of thermo output will be the same:
compute peratom all stress/atom NULL
compute p all reduce sum c_peratom\[1\] c_peratom\[2\] c_peratom\[3\]
variable press equal -(c_p\[1\]+c_p\[2\]+c_p\[3\])/(3*vol)
thermo_style custom step temp etotal press v_press :pre
compute peratom all stress/atom NULL
compute p all reduce sum c_peratom\[1\] c_peratom\[2\] c_peratom\[3\]
variable press equal -(c_p\[1\]+c_p\[2\]+c_p\[3\])/(3*vol)
thermo_style custom step temp etotal press v_press :pre
NOTE: The per-atom stress does not include any Lennard-Jones tail
corrections to the pressure added by the "pair_modify tail
yes"_pair_modify.html command, since those are contributions to the
global system pressure.
[Output info:]

View File

@ -88,7 +88,7 @@ potentials only include the pair potential portion of the EAM
interaction when used by this compute, not the embedding term. Also
bonded or Kspace interactions do not contribute to this compute.
[Related commands:]
[Related commands:]
{compute group/group}_compute_group_group.html, {compute
heat/flux}_compute_heat_flux.html

0
doc/src/compute_temp_asphere.txt Executable file → Normal file
View File

0
doc/src/compute_temp_body.txt Executable file → Normal file
View File

View File

@ -16,7 +16,7 @@ ID, group-ID are documented in "compute"_compute.html command
temp/cs = style name of this compute command
group1 = group-ID of either cores or shells
group2 = group-ID of either shells or cores :ul
[Examples:]
compute oxygen_c-s all temp/cs O_core O_shell
@ -64,7 +64,7 @@ calculated by this compute for use in the computation of a pressure
tensor. The formula for the components of the tensor is the same as
the above formula, except that v^2 is replaced by vx*vy for the xy
component, etc. The 6 components of the vector are ordered xx, yy,
zz, xy, xz, yz. In contrast to the temperature, the velocity of
zz, xy, xz, yz. In contrast to the temperature, the velocity of
each core or shell atom is taken individually.
The change this fix makes to core/shell atom velocities is essentially

View File

@ -14,7 +14,7 @@ compute ID group-ID temp/drude :pre
ID, group-ID are documented in "compute"_compute.html command
temp/drude = style name of this compute command :ul
[Examples:]
compute TDRUDE all temp/drude :pre
@ -68,7 +68,7 @@ are "extensive".
[Restrictions:]
The number of degrees of freedom contributing to the temperature is
assumed to be constant for the duration of the run unless the
assumed to be constant for the duration of the run unless the
{fix_modify} command sets the option {dynamic yes}.
[Related commands:]

View File

@ -49,9 +49,9 @@ reported by LAMMPS in the thermodynamic quantities reported via the
example:
compute effTemp all temp/eff
thermo_style custom step etotal pe ke temp press
thermo_style custom step etotal pe ke temp press
thermo_modify temp effTemp :pre
A 6-component kinetic energy tensor is also calculated by this compute
for use in the computation of a pressure tensor. The formula for the
components of the tensor is the same as the above formula, except that
@ -80,7 +80,7 @@ is independent of the number of atoms in the simulation. The vector
values are "extensive", meaning they scale with the number of atoms in
the simulation.
[Restrictions:]
[Restrictions:]
This compute is part of the USER-EFF package. It is only enabled if
LAMMPS was built with that package. See the "Making

View File

@ -68,7 +68,7 @@ temp/berendsen"_fix_temp_berendsen.html, and "fix
langevin"_fix_langevin.html. This means that when this compute
is used to calculate the temperature for any of the thermostatting
fixes via the "fix modify temp"_fix_modify.html command, the thermostat
will operate only on atoms that are currently in the geometric
will operate only on atoms that are currently in the geometric
region.
Unlike other compute styles that calculate temperature, this compute

2
doc/src/compute_temp_sphere.txt Executable file → Normal file
View File

@ -122,7 +122,7 @@ vector values will be in energy "units"_units.html.
This fix requires that atoms store torque and angular velocity (omega)
and a radius as defined by the "atom_style sphere"_atom_style.html
command.
command.
All particles in the group must be finite-size spheres, or point
particles with radius = 0.0.

View File

@ -6,7 +6,7 @@
:line
compute ti command :h3
compute ti command :h3
[Syntax:]
@ -35,7 +35,7 @@ keyword = pair style (lj/cut, gauss, born, etc) or {tail} or {kspace} :l
compute 1 all ti lj/cut 1 v_lj v_dlj coul/long 2 v_c v_dc kspace 1 v_ks v_dks
compute 1 all ti lj/cut 1*3 v_lj v_dlj coul/long * v_c v_dc kspace * v_ks v_dks :pre
[Description:]
[Description:]
Define a computation that calculates the derivative of the interaction
potential with respect to {lambda}, the coupling parameter used in a
@ -107,7 +107,7 @@ du/dl can be found in the paper by "Eike"_#Eike.
:line
[Output info:]
[Output info:]
This compute calculates a global scalar, namely dUs/dlambda. This
value can be used by any command that uses a global scalar value from

View File

@ -15,7 +15,7 @@ compute ID group-ID voronoi/atom keyword arg ... :pre
ID, group-ID are documented in "compute"_compute.html command :ulb,l
voronoi/atom = style name of this compute command :l
zero or more keyword/value pairs may be appended :l
keyword = {only_group} or {surface} or {radius} or {edge_histo} or {edge_threshold}
keyword = {only_group} or {surface} or {radius} or {edge_histo} or {edge_threshold}
or {face_threshold} or {neighbors} or {peratom} :l
{only_group} = no arg
{occupation} = no arg
@ -25,7 +25,7 @@ or {face_threshold} or {neighbors} or {peratom} :l
{radius} arg = v_r
v_r = radius atom style variable for a poly-disperse Voronoi tessellation
{edge_histo} arg = maxedge
maxedge = maximum number of Voronoi cell edges to be accounted in the histogram
maxedge = maximum number of Voronoi cell edges to be accounted in the histogram
{edge_threshold} arg = minlength
minlength = minimum length for an edge to be counted
{face_threshold} arg = minarea
@ -38,7 +38,7 @@ or {face_threshold} or {neighbors} or {peratom} :l
compute 1 all voronoi/atom
compute 2 precipitate voronoi/atom surface matrix
compute 3b precipitate voronoi/atom radius v_r
compute 3b precipitate voronoi/atom radius v_r
compute 4 solute voronoi/atom only_group
compute 5 defects voronoi/atom occupation
compute 6 all voronoi/atom neighbors yes :pre
@ -53,11 +53,11 @@ in the group.
By default two per-atom quantities are calculated by this compute.
The first is the volume of the Voronoi cell around each atom. Any
point in an atom's Voronoi cell is closer to that atom than any other.
The second is the number of faces of the Voronoi cell. This is
The second is the number of faces of the Voronoi cell. This is
equal to the number of nearest neighbors of the central atom,
plus any exterior faces (see note below). If the {peratom} keyword
is set to "no", the per-atom quantities are still calculated,
but they are not accessible.
plus any exterior faces (see note below). If the {peratom} keyword
is set to "no", the per-atom quantities are still calculated,
but they are not accessible.
:line
@ -122,23 +122,23 @@ to locate vacancies (the coordinates are given by the atom coordinates
at the time step when the compute was first invoked), while column two
data can be used to identify interstitial atoms.
If the {neighbors} value is set to yes, then
If the {neighbors} value is set to yes, then
this compute creates a local array with 3 columns. There
is one row for each face of each Voronoi cell. The
3 columns are the atom ID of the atom that owns the cell,
the atom ID of the atom in the neighboring cell
(or zero if the face is external), and the area of the face.
is one row for each face of each Voronoi cell. The
3 columns are the atom ID of the atom that owns the cell,
the atom ID of the atom in the neighboring cell
(or zero if the face is external), and the area of the face.
The array can be accessed by any command that
uses local values from a compute as input. See "this
section"_Section_howto.html#howto_15 for an overview of LAMMPS output
options. More specifically, the array can be accessed by a
options. More specifically, the array can be accessed by a
"dump local"_dump.html command to write a file containing
all the Voronoi neighbors in a system:
compute 6 all voronoi/atom neighbors yes
dump d2 all local 1 dump.neighbors index c_6\[1\] c_6\[2\] c_6\[3\] :pre
If the {face_threshold} keyword is used, then only faces
If the {face_threshold} keyword is used, then only faces
with areas greater than the threshold are stored.
:line
@ -190,7 +190,7 @@ per-atom values from a compute as input. See "Section
6.15"_Section_howto.html#howto_15 for an overview of LAMMPS output
options. If the {peratom} keyword is set to "no", the per-atom array
is still created, but it is not accessible.
If the {edge_histo} keyword is used, then this compute generates a
global vector of length {maxedge}+1, containing a histogram of the
number of edges per face.

View File

@ -20,21 +20,21 @@ type1 type2 ... typeN = chemical symbol of each atom type (see valid options bel
zero or more keyword/value pairs may be appended :l
keyword = {2Theta} or {c} or {LP} or {manual} or {echo} :l
{2Theta} values = Min2Theta Max2Theta
Min2Theta,Max2Theta = minimum and maximum 2 theta range to explore
Min2Theta,Max2Theta = minimum and maximum 2 theta range to explore
(radians or degrees)
{c} values = c1 c2 c3
c1,c2,c3 = parameters to adjust the spacing of the reciprocal
c1,c2,c3 = parameters to adjust the spacing of the reciprocal
lattice nodes in the h, k, and l directions respectively
{LP} value = switch to apply Lorentz-polarization factor
0/1 = off/on
{manual} = flag to use manual spacing of reciprocal lattice points
based on the values of the {c} parameters
{manual} = flag to use manual spacing of reciprocal lattice points
based on the values of the {c} parameters
{echo} = flag to provide extra output for debugging purposes :pre
:ule
[Examples:]
compute 1 all xrd 1.541838 Al O 2Theta 0.087 0.87 c 1 1 1 LP 1 echo
compute 1 all xrd 1.541838 Al O 2Theta 0.087 0.87 c 1 1 1 LP 1 echo
compute 2 all xrd 1.541838 Al O 2Theta 10 100 c 0.05 0.05 0.05 LP 1 manual :pre
fix 1 all ave/histo/weight 1 1 1 0.087 0.87 250 c_1\[1\] c_1\[2\] mode vector file Rad2Theta.xrd
@ -43,11 +43,11 @@ fix 2 all ave/histo/weight 1 1 1 10 100 250 c_2\[1\] c_2\[2\] mode vector file D
[Description:]
Define a computation that calculates x-ray diffraction intensity as described
in "(Coleman)"_#xrd-Coleman on a mesh of reciprocal lattice nodes defined
in "(Coleman)"_#xrd-Coleman on a mesh of reciprocal lattice nodes defined
by the entire simulation domain (or manually) using a simulated radiation
of wavelength lambda.
of wavelength lambda.
The x-ray diffraction intensity, I, at each reciprocal lattice point, k,
The x-ray diffraction intensity, I, at each reciprocal lattice point, k,
is computed from the structure factor, F, using the equations:
:c,image(Eqs/compute_xrd1.jpg)
@ -55,14 +55,14 @@ is computed from the structure factor, F, using the equations:
:c,image(Eqs/compute_xrd3.jpg)
:c,image(Eqs/compute_xrd4.jpg)
Here, K is the location of the reciprocal lattice node, rj is the
position of each atom, fj are atomic scattering factors, LP is the
Lorentz-polarization factor, and theta is the scattering angle of
diffraction. The Lorentz-polarization factor can be turned off using
Here, K is the location of the reciprocal lattice node, rj is the
position of each atom, fj are atomic scattering factors, LP is the
Lorentz-polarization factor, and theta is the scattering angle of
diffraction. The Lorentz-polarization factor can be turned off using
the optional {LP} keyword.
Diffraction intensities are calculated on a three-dimensional mesh of
reciprocal lattice nodes. The mesh spacing is defined either (a)
Diffraction intensities are calculated on a three-dimensional mesh of
reciprocal lattice nodes. The mesh spacing is defined either (a)
by the entire simulation domain or (b) manually using selected values as
shown in the 2D diagram below.
@ -101,8 +101,8 @@ The analytic approximation is computed using the formula
:c,image(Eqs/compute_xrd5.jpg)
Coefficients parameterized by "(Peng)"_#Peng are assigned for each
atom type designating the chemical symbol and charge of each atom
Coefficients parameterized by "(Peng)"_#Peng are assigned for each
atom type designating the chemical symbol and charge of each atom
type. Valid chemical symbols for compute xrd are:
H| He1-| He| Li| Li1+|
@ -148,39 +148,39 @@ type. Valid chemical symbols for compute xrd are:
Np4+| Np6+| Pu| Pu3+| Pu4+|
Pu6+| Am| Cm| Bk| Cf :tb(c=5,s=|)
If the {echo} keyword is specified, compute xrd will provide extra
reporting information to the screen.
If the {echo} keyword is specified, compute xrd will provide extra
reporting information to the screen.
[Output info:]
This compute calculates a global array. The number of rows in the
array is the number of reciprocal lattice nodes that are explored
which by the mesh. The global array has 2 columns.
This compute calculates a global array. The number of rows in the
array is the number of reciprocal lattice nodes that are explored
which by the mesh. The global array has 2 columns.
The first column contains the diffraction angle in the units (radians
or degrees) provided with the {2Theta} values. The second column contains
or degrees) provided with the {2Theta} values. The second column contains
the computed diffraction intensities as described above.
The array can be accessed by any command that uses global values from
a compute as input. See "this section"_Section_howto.html#howto_15
for an overview of LAMMPS output options.
All array values calculated by this compute are "intensive".
All array values calculated by this compute are "intensive".
[Restrictions:]
[Restrictions:]
This compute is part of the USER-DIFFRACTION package. It is only
enabled if LAMMPS was built with that package. See the "Making
LAMMPS"_Section_start.html#start_3 section for more info.
The compute_xrd command does not work for triclinic cells.
The compute_xrd command does not work for triclinic cells.
[Related commands:]
[Related commands:]
"fix ave/histo"_fix_ave_histo.html,
"compute saed"_compute_saed.html
[Default:]
[Default:]
The option defaults are 2Theta = 1 179 (degrees), c = 1 1 1, LP = 1,
no manual flag, no echo flag.
@ -192,7 +192,7 @@ no manual flag, no echo flag.
(2013).
:link(Colliex)
[(Colliex)] Colliex et al. International Tables for Crystallography
[(Colliex)] Colliex et al. International Tables for Crystallography
Volume C: Mathematical and Chemical Tables, 249-429 (2004).
:link(Peng)

View File

@ -48,7 +48,7 @@ keyword = {mol} or {basis} or {remap} or {var} or {set} or {units} :l
create_atoms 1 box
create_atoms 3 region regsphere basis 2 3
create_atoms 3 single 0 0 5
create_atoms 3 single 0 0 5
create_atoms 1 box var v set x xpos set y ypos :pre
[Description:]
@ -218,14 +218,14 @@ larger version.
variable x equal 100
variable y equal 25
lattice hex 0.8442
region box block 0 $x 0 $y -0.5 0.5
create_box 1 box :pre
lattice hex 0.8442
region box block 0 $x 0 $y -0.5 0.5
create_box 1 box :pre
variable xx equal 0.0
variable yy equal 0.0
variable v equal "(0.2*v_y*ylat * cos(v_xx/xlat * 2.0*PI*4.0/v_x) + 0.5*v_y*ylat - v_yy) > 0.0"
create_atoms 1 box var v set x xx set y yy :pre
create_atoms 1 box var v set x xx set y yy :pre
:c,image(JPG/sinusoid_small.jpg,JPG/sinusoid.jpg)
@ -245,7 +245,7 @@ style. A {box} value selects standard distance units as defined by
the "units"_units.html command, e.g. Angstroms for units = real or
metal. A {lattice} value means the distance units are in lattice
spacings.
:line
Atom IDs are assigned to created atoms in the following way. The

View File

@ -121,7 +121,7 @@ The {special} keyword is invoked at the end of the delete_bonds
operation, after (optional) removal. It re-computes the pairwise 1-2,
1-3, 1-4 weighting list. The weighting list computation treats
turned-off bonds the same as turned-on. Thus, turned-off bonds must
be removed if you wish to change the weighting list.
be removed if you wish to change the weighting list.
Note that the choice of {remove} and {special} options affects how
1-2, 1-3, 1-4 pairwise interactions will be computed across bonds that

View File

@ -6,7 +6,7 @@
:line
dielectric command :h3
dielectric command :h3
[Syntax:]

View File

@ -17,10 +17,10 @@ dihedral_style class2 :pre
dihedral_style class2
dihedral_coeff 1 100 75 100 70 80 60
dihedral_coeff * mbt 3.5945 0.1704 -0.5490 1.5228
dihedral_coeff * mbt 3.5945 0.1704 -0.5490 1.5228
dihedral_coeff * ebt 0.3417 0.3264 -0.9036 0.1368 0.0 -0.8080 1.0119 1.1010
dihedral_coeff 2 at 0.0 -0.1850 -0.7963 -2.0220 0.0 -0.3991 110.2453 105.1270
dihedral_coeff * aat -13.5271 110.2453 105.1270
dihedral_coeff 2 at 0.0 -0.1850 -0.7963 -2.0220 0.0 -0.3991 110.2453 105.1270
dihedral_coeff * aat -13.5271 110.2453 105.1270
dihedral_coeff * bb13 0.0 1.0119 1.1010 :pre
[Description:]

View File

@ -66,7 +66,7 @@ more instructions on how to use the accelerated styles effectively.
[Restrictions:]
This angle style can only be used if LAMMPS was built with the
USER_MISC package. See the "Making LAMMPS"_Section_start.html#start_3
USER_MISC package. See the "Making LAMMPS"_Section_start.html#start_3
section for more info on packages.
[Related commands:]

View File

@ -63,7 +63,7 @@ more instructions on how to use the accelerated styles effectively.
[Restrictions:]
This angle style can only be used if LAMMPS was built with the
USER_MISC package. See the "Making LAMMPS"_Section_start.html#start_3
USER_MISC package. See the "Making LAMMPS"_Section_start.html#start_3
section for more info on packages.
[Related commands:]

View File

@ -33,7 +33,7 @@ above, or in the data file or restart files read by the
"read_data"_read_data.html or "read_restart"_read_restart.html
commands:
K (energy/radian^2)
K (energy/radian^2)
phi0 (degrees) :ul
:line
@ -64,7 +64,7 @@ more instructions on how to use the accelerated styles effectively.
[Restrictions:]
This angle style can only be used if LAMMPS was built with the
USER_MISC package. See the "Making LAMMPS"_Section_start.html#start_3
USER_MISC package. See the "Making LAMMPS"_Section_start.html#start_3
section for more info on packages.
[Related commands:]

View File

@ -14,7 +14,7 @@ dihedral_style spherical :pre
[Examples:]
dihedral_coeff 1 1 286.1 1 124 1 1 90.0 0 1 90.0 0
dihedral_coeff 1 1 286.1 1 124 1 1 90.0 0 1 90.0 0
dihedral_coeff 1 3 286.1 1 114 1 1 90 0 1 90.0 0 &
17.3 0 0.0 0 1 158 1 0 0.0 0 &
15.1 0 0.0 0 0 0.0 0 1 167.3 1 :pre
@ -76,7 +76,7 @@ wn (typically 0.0 or 1.0) :ul
[Restrictions:]
This dihedral style can only be used if LAMMPS was built with the
USER_MISC package. See the "Making LAMMPS"_Section_start.html#start_3
USER_MISC package. See the "Making LAMMPS"_Section_start.html#start_3
section for more info on packages.
[Related commands:]

Some files were not shown because too many files have changed in this diff Show More