Compare commits

...

317 Commits

Author SHA1 Message Date
04f5eadcf1 added LAST option to dump_modify thresh, more restart info printed out to screen 2016-10-11 12:39:52 -06:00
e70d530c46 Merge pull request #203 from rbberger/txt2rst-external-link-fix
txt2rst external link fix
2016-10-10 13:59:27 -06:00
ed8cc82713 Merge pull request #211 from akohlmey/add-respa-to-fix-flow-gauss
Add respa support to fix flow/gauss
2016-10-10 13:59:01 -06:00
27dac02466 Merge pull request #209 from akohlmey/static-double-deallocation-workaround
workaround for double free issue when using USER-COLVARS with with lammps python wrapper and python package
2016-10-10 13:58:16 -06:00
467bcad0a0 Merge pull request #204 from rbberger/fix-user-omp
Migrate changes from GRANULAR to USER-OMP
2016-10-10 13:57:37 -06:00
144e6a8091 whitespace cleanup 2016-10-10 09:40:09 -04:00
72ac073412 edited documentation
(cherry picked from commit eff14c74b0)
2016-10-10 09:38:54 -04:00
49c45ab03b edited documentation
(cherry picked from commit fd560889c3)
2016-10-10 09:38:53 -04:00
c2cd439944 first draft of documentation for respa
(cherry picked from commit d7dcbcfbd9)
2016-10-10 09:38:53 -04:00
e96ebb29bc adjusted default respa level to be outermost
(cherry picked from commit 7fc4d46a41)
2016-10-10 09:38:53 -04:00
3ce178d43f now understand how respa works in lammps
(cherry picked from commit c829027e83)
2016-10-10 09:38:52 -04:00
23781d6ec9 added respa to fix_flow_gauss, not fully understood yet
(cherry picked from commit 8d9737b04d)
2016-10-10 09:38:52 -04:00
fca6d721c0 completed synchronization with non-threaded version 2016-10-10 09:16:21 -04:00
dd192ca7ea whitespace cleanup 2016-10-10 09:15:42 -04:00
683689c808 revert to previous style conventions for size_t constants 2016-10-08 11:00:23 -04:00
e01e90eb96 workaround for double free issue when using USER-COLVARS with lammps code loaded as shared library into a standalone executable 2016-10-08 10:45:22 -04:00
615a2da044 Migrate changes from GRANULAR to USER-OMP 2016-10-06 21:48:06 -04:00
7f3a7c5cbe Fix broken link 2016-10-06 20:33:24 -04:00
e78b4267b7 Fix issue with external links containing anchors 2016-10-06 20:29:07 -04:00
e9fed80928 Merge pull request #202 from akohlmey/doc-formatting-fixes
collected documentation updates and corrections from LAMMPS-ICMS
2016-10-06 15:49:44 -06:00
54fc194e5b Merge pull request #199 from akohlmey/small-changes
Collected small changes and bugfixes
2016-10-06 15:49:24 -06:00
b3d2fb91bb new fix wall/gran/region command, REBO bug fix, new example log files 2016-10-06 15:47:41 -06:00
19984c9bd1 Revert "bugfix for AngleAngle term in CLASS2 impropers by Ivan A. Strelnikov, ICP RAS"
This reverts commit 83bcdb6a50.
2016-10-06 17:23:10 -04:00
f92618a33b Revert "bugfix for virial tally for improper style umbrella from Steven Vandenbrande (U Gent)"
This reverts commit 4921dc18a0.
2016-10-06 17:21:38 -04:00
0b5d71537a collected documentation updates and corrections from LAMMPS-ICMS
fixes formatting issues due to tabs, permission issues and
a few typos and badly worded text.
2016-10-06 15:48:18 -04:00
c213457550 Merge pull request #197 from giacomofiorin/colvars_2016-10-05
Colvars 2016-10-05
2016-10-06 13:02:52 -06:00
0f45cd61a5 Merge pull request #196 from akohlmey/charmm-cmap-updates
Some more cmap-related updates for ch2lmp
2016-10-06 13:02:27 -06:00
493873fb93 clean up doc src 2016-10-06 13:00:46 -06:00
60a031ebac Merge branch 'USER-DPD_pair_exp6_rx_mathfix' of https://github.com/timattox/lammps_USER-DPD into small-changes
This closes #201
2016-10-06 14:28:08 -04:00
27e76a70b9 Merge branch 'USER-DPD_hybrid_atom_bugfix' of https://github.com/timattox/lammps_USER-DPD into small-changes
This closes #200
2016-10-06 14:27:27 -04:00
e1e9a5c126 USER-DPD: math corrections in pair_exp6_rx.cpp (by Jim Larentzos) 2016-10-06 13:49:47 -04:00
d31121b18c USER-DPD: bugfix in unpack_comm_hybrid(); now works with hybrid atom style 2016-10-06 13:21:27 -04:00
0853cdbe6f update reference data files for updated/corrected clayff parameters 2016-10-06 11:47:08 -04:00
83bcdb6a50 bugfix for AngleAngle term in CLASS2 impropers by Ivan A. Strelnikov, ICP RAS
this closes #56
2016-10-06 11:27:18 -04:00
22ce671804 improved whitespace handling in msi2lmp for force fields and topologies 2016-10-06 11:16:59 -04:00
4921dc18a0 bugfix for virial tally for improper style umbrella from Steven Vandenbrande (U Gent)
this closes #182
2016-10-06 10:47:08 -04:00
d133167bf6 Merge branch 'master' of https://github.com/albapa/lammps into small-changes
USER-QUIP related improvements from github user albapa. This closes #198
2016-10-06 09:32:50 -04:00
8ea063378e add NETCDF libs (as defined in QUIP) to the linking line if QUIP was built with NETCDF support 2016-10-06 12:16:25 +01:00
fd16118cbb removed dump_modify command 2016-10-06 12:02:41 +01:00
f9f955d5b5 update include statement format 2016-10-05 22:34:44 -04:00
d7d321a512 some more updates to the README file to reflect the inclusion of the CMAP example and renamed file names 2016-10-05 18:41:45 -04:00
8809a603fb Colvars update: issue a warning that cannot be ignored regarding total forces 2016-10-05 18:26:21 -04:00
969d3cf4b0 Colvars update: make ABF check that the colvar isn't using already subtractAppliedForce 2016-10-05 18:25:40 -04:00
326fdf2cf1 added 1GB1 example from Robert Latour and update 1AC7 example files 2016-10-05 18:20:09 -04:00
f32819dd10 added tweak to write out the command line used for the conversion to the beginning of the LAMMPS input 2016-10-05 18:13:46 -04:00
c07a01c661 import updated README file for charmm2lammps.pl with CMAP support 2016-10-05 18:11:52 -04:00
02bfa898ee adjustments to balancing weights and factors, also XOR op for formulas, if, dump_modify thresh 2016-10-05 15:46:20 -06:00
030df745bc Merge pull request #193 from akohlmey/eam-bugfix
bugfix for eam/alloy/omp and eam/fs/omp
2016-10-05 10:54:36 -06:00
6a97211932 Merge pull request #192 from rbberger/python-interface-bugfix
Revert type checking commit from July
2016-10-05 10:54:08 -06:00
c46be7db62 changes to imbalance weight factors 2016-10-05 10:33:39 -06:00
4381db846b correct the bug discovered by stan due to uninitialized scale factors for eam/alloy/omp and eam/fs/omp 2016-10-04 14:33:26 -04:00
e2caf5c105 Fix code path which allows passing a C++ ptr to PyLammps 2016-10-04 13:57:21 -04:00
11c2892e54 Merge branch 'restrict-weights-and-weight-factors' of https://github.com/akohlmey/lammps 2016-10-04 09:49:09 -06:00
91be47a0d0 Revert type checking commit from July
0aebb2eabe
2016-10-04 11:43:12 -04:00
ab92529b19 Merge pull request #191 from akohlmey/updated-charmm2lammps
Updated charmm2lammps
2016-10-03 17:59:21 -06:00
e079362776 Merge pull request #190 from akohlmey/small-bufixes-and-enhancements
Small bufixes and enhancements
2016-10-03 17:58:36 -06:00
c3ff8812b3 added XOR operator to variable command 2016-10-03 17:57:33 -06:00
03766dbda7 apply bugfix for MEAM provided by Wolfgang Verestek on lammps-users
this closes lammps/#188
2016-10-03 16:28:59 -04:00
6e719f2d94 remove trailing whitespace 2016-10-03 07:07:28 -04:00
45d2cc2895 permission update for ch2lmp tool folder 2016-10-03 07:03:42 -04:00
690f91300b rebuild charmm2lammps example output files with updated tools 2016-10-03 06:58:51 -04:00
3b94627dfe properly handle -nohints flag, make -cmap flag take version as option. step version number 2016-10-03 06:52:30 -04:00
c2e11dffa2 import updated charmm2lammps.pl script from Rober Latour 2016-10-02 20:33:20 -04:00
1985db4fb1 correct designation of meam supporting USER-OMP and meam/spline not 2016-10-01 23:05:05 -04:00
a3e05a2bac permission cleanup 2016-10-01 06:34:45 -04:00
035279de87 correct logic bug in bufix for fix tmd
(cherry picked from commit 267c1ec957)
2016-10-01 06:26:52 -04:00
e2c7acabac Merge pull request #187 from akohlmey/colvars-update-2016-09-30
update colvars library to version 2016-09-30
2016-09-30 09:21:00 -06:00
91edee2530 Merge pull request #186 from akohlmey/small-bugfixes
Collected small bugfixes and enhancements
2016-09-30 09:20:25 -06:00
b9d0f96a19 change purge target in Makefile, also fixed one issue with Make.py 2016-09-30 09:17:55 -06:00
d45e333f7c restrict choice of weight factors and guarantee that weights are >= 0.001 2016-09-30 11:11:32 -04:00
5bb85b482d remove unused variable 2016-09-30 09:38:50 -04:00
d4b074d85b enable dynamic groups for fix dt/reset 2016-09-30 09:09:44 -04:00
6d200061ca update colvars library to version 2016-09-30 2016-09-30 08:15:44 -04:00
cb7bd2799e flag header as C++ to emacs 2016-09-30 07:39:45 -04:00
4337f2c240 include charmm22 and charmm36 cmap files and include date added signature 2016-09-30 07:39:12 -04:00
0eeb240730 whitespace cleanup, fix bug in looking for empty strings, improve read performance and handling of comments 2016-09-30 07:22:47 -04:00
c88acc9613 make reader for target geometry file more resilient 2016-09-29 22:59:46 -04:00
f7b5afee82 Merge pull request #184 from akohlmey/dynamic-groups-for-respa
Dynamic groups for respa
2016-09-29 15:51:34 -06:00
a315dcda9b remove dead code
(cherry picked from commit 7f0994aac0)
2016-09-29 15:13:46 -04:00
f6c77c3aba support dynamic groups with run style respa
(cherry picked from commit b7baa1680d)
2016-09-29 15:13:46 -04:00
5b2becd09b Merge branch 'integration' into new-master 2016-09-29 10:37:09 -04:00
78a22be93f sync Make.py and fix addforce change with GHub
git-svn-id: svn://svn.icms.temple.edu/lammps-ro/trunk@15675 f3b2605a-c512-4ea7-a41b-209d697bcdaa
2016-09-28 22:36:54 +00:00
596b260f5d Merge pull request #45 from akohlmey/small-bugfixes
Small bugfixes
2016-09-28 16:36:04 -06:00
446e7e7369 patch for allowing prd command to work with sorted atoms 2016-09-28 16:33:30 -06:00
189825489c git-svn-id: svn://svn.icms.temple.edu/lammps-ro/trunk@15673 f3b2605a-c512-4ea7-a41b-209d697bcdaa 2016-09-28 22:32:14 +00:00
bdd0f665ca git-svn-id: svn://svn.icms.temple.edu/lammps-ro/trunk@15672 f3b2605a-c512-4ea7-a41b-209d697bcdaa 2016-09-28 22:32:12 +00:00
6897cc803f git-svn-id: svn://svn.icms.temple.edu/lammps-ro/trunk@15671 f3b2605a-c512-4ea7-a41b-209d697bcdaa 2016-09-28 22:29:06 +00:00
f511c177c6 git-svn-id: svn://svn.icms.temple.edu/lammps-ro/trunk@15670 f3b2605a-c512-4ea7-a41b-209d697bcdaa 2016-09-28 14:37:45 +00:00
1ec3987b31 git-svn-id: svn://svn.icms.temple.edu/lammps-ro/trunk@15669 f3b2605a-c512-4ea7-a41b-209d697bcdaa 2016-09-28 14:36:30 +00:00
8c1d0031c9 correct typo in Make.py 2016-09-27 18:20:06 -04:00
45e50b46c3 sync with GH
git-svn-id: svn://svn.icms.temple.edu/lammps-ro/trunk@15668 f3b2605a-c512-4ea7-a41b-209d697bcdaa
2016-09-27 21:37:17 +00:00
829d11e88b Merge pull request #44 from rbberger/doc-fixes
Some documentation fixes and IPython updates
2016-09-27 15:36:10 -06:00
1adf3858a9 correct bug and synchronize fix addforce respa level init with other fixes 2016-09-27 17:36:02 -04:00
96f31d6dad Merge pull request #43 from akohlmey/doc-fixes
Documentation fixes
2016-09-27 15:35:41 -06:00
35705217f4 enable multi-processor NEB replicas 2016-09-27 15:34:08 -06:00
9a2f738673 sync with SVN 2016-09-27 15:32:57 -06:00
f82e0c53b6 git-svn-id: svn://svn.icms.temple.edu/lammps-ro/trunk@15666 f3b2605a-c512-4ea7-a41b-209d697bcdaa 2016-09-27 21:31:04 +00:00
1fbddc97d1 git-svn-id: svn://svn.icms.temple.edu/lammps-ro/trunk@15665 f3b2605a-c512-4ea7-a41b-209d697bcdaa 2016-09-27 21:31:02 +00:00
1cfa49f03d git-svn-id: svn://svn.icms.temple.edu/lammps-ro/trunk@15664 f3b2605a-c512-4ea7-a41b-209d697bcdaa 2016-09-27 21:28:06 +00:00
3486b7d503 git-svn-id: svn://svn.icms.temple.edu/lammps-ro/trunk@15663 f3b2605a-c512-4ea7-a41b-209d697bcdaa 2016-09-27 21:24:10 +00:00
6fedf8d899 git-svn-id: svn://svn.icms.temple.edu/lammps-ro/trunk@15662 f3b2605a-c512-4ea7-a41b-209d697bcdaa 2016-09-27 21:18:32 +00:00
56b0856e2f git-svn-id: svn://svn.icms.temple.edu/lammps-ro/trunk@15661 f3b2605a-c512-4ea7-a41b-209d697bcdaa 2016-09-27 21:16:33 +00:00
f9c2049724 need to ignore new fix cmap sources 2016-09-27 17:12:17 -04:00
e1c6b6b7d1 correctly handle exceptions raised from subprocess module 2016-09-27 17:01:45 -04:00
3333e4b475 Put snap before zbl to get more helpful error message
git-svn-id: svn://svn.icms.temple.edu/lammps-ro/trunk@15660 f3b2605a-c512-4ea7-a41b-209d697bcdaa
2016-09-27 17:21:42 +00:00
a3a3af691c Merge branch 'balance' into integration 2016-09-27 10:53:56 -06:00
f9677e6d7b released version of weighted balancing 2016-09-27 10:52:27 -06:00
2ae966c26f git-svn-id: svn://svn.icms.temple.edu/lammps-ro/trunk@15657 f3b2605a-c512-4ea7-a41b-209d697bcdaa 2016-09-27 16:49:51 +00:00
d1b8ffd924 git-svn-id: svn://svn.icms.temple.edu/lammps-ro/trunk@15656 f3b2605a-c512-4ea7-a41b-209d697bcdaa 2016-09-27 16:49:48 +00:00
b66039b8bb git-svn-id: svn://svn.icms.temple.edu/lammps-ro/trunk@15653 f3b2605a-c512-4ea7-a41b-209d697bcdaa 2016-09-27 16:43:18 +00:00
995ecea5ed git-svn-id: svn://svn.icms.temple.edu/lammps-ro/trunk@15652 f3b2605a-c512-4ea7-a41b-209d697bcdaa 2016-09-27 16:02:08 +00:00
43633180eb git-svn-id: svn://svn.icms.temple.edu/lammps-ro/trunk@15651 f3b2605a-c512-4ea7-a41b-209d697bcdaa 2016-09-27 15:08:34 +00:00
b68e954761 git-svn-id: svn://svn.icms.temple.edu/lammps-ro/trunk@15650 f3b2605a-c512-4ea7-a41b-209d697bcdaa 2016-09-27 15:06:58 +00:00
2b88050a1f git-svn-id: svn://svn.icms.temple.edu/lammps-ro/trunk@15649 f3b2605a-c512-4ea7-a41b-209d697bcdaa 2016-09-27 15:06:14 +00:00
063307c71c git-svn-id: svn://svn.icms.temple.edu/lammps-ro/trunk@15648 f3b2605a-c512-4ea7-a41b-209d697bcdaa 2016-09-27 15:05:29 +00:00
f280bd32a6 git-svn-id: svn://svn.icms.temple.edu/lammps-ro/trunk@15647 f3b2605a-c512-4ea7-a41b-209d697bcdaa 2016-09-26 23:34:26 +00:00
53eac4431d git-svn-id: svn://svn.icms.temple.edu/lammps-ro/trunk@15646 f3b2605a-c512-4ea7-a41b-209d697bcdaa 2016-09-26 23:32:16 +00:00
a3277117e2 Add filter which merges preformatted sections 2016-09-26 18:52:43 -04:00
67d4c07689 Do not escape underscore inside preformat blocks 2016-09-26 18:52:31 -04:00
877a504933 Fix typo in Section_howto.txt 2016-09-26 18:44:25 -04:00
8a951f9d79 fix typo 2016-09-26 18:43:03 -04:00
69a8842ecb update load balance weights documentation for fix balance and balance 2016-09-26 18:33:50 -04:00
2af5c75f42 correct issue from merge 2016-09-26 18:32:01 -04:00
158599fca2 Merge branch 'balance2' into weighted-balancing 2016-09-26 18:25:36 -04:00
7732548b3c correct issues related to the addition of fix cmap 2016-09-26 18:14:32 -04:00
2c5f6e1a99 fix a broken link that slipped through the cracks in the previous cleanup 2016-09-26 18:13:18 -04:00
d0aa13b543 Fix broken link in Section_packages.txt 2016-09-26 16:53:18 -04:00
c31b026797 Merge branch 'integration' into weighted-balancing 2016-09-26 15:20:22 -04:00
47b52ed2dd Merge branch 'integration' into balance2 2016-09-26 15:19:48 -04:00
fb64ae612f git-svn-id: svn://svn.icms.temple.edu/lammps-ro/trunk@15645 f3b2605a-c512-4ea7-a41b-209d697bcdaa 2016-09-26 16:23:53 +00:00
c87f9aeb9f Merge remote-tracking branch 'akohlmey/integration' into ipython-update-and-cleanup 2016-09-26 11:59:30 -04:00
5769c10189 git-svn-id: svn://svn.icms.temple.edu/lammps-ro/trunk@15643 f3b2605a-c512-4ea7-a41b-209d697bcdaa 2016-09-26 14:39:43 +00:00
7453a4f55f git-svn-id: svn://svn.icms.temple.edu/lammps-ro/trunk@15642 f3b2605a-c512-4ea7-a41b-209d697bcdaa 2016-09-26 14:39:40 +00:00
50d59454d2 git-svn-id: svn://svn.icms.temple.edu/lammps-ro/trunk@15640 f3b2605a-c512-4ea7-a41b-209d697bcdaa 2016-09-23 23:06:49 +00:00
24ff008a0f git-svn-id: svn://svn.icms.temple.edu/lammps-ro/trunk@15639 f3b2605a-c512-4ea7-a41b-209d697bcdaa 2016-09-23 23:06:44 +00:00
da480bd4d4 git-svn-id: svn://svn.icms.temple.edu/lammps-ro/trunk@15638 f3b2605a-c512-4ea7-a41b-209d697bcdaa 2016-09-23 23:00:00 +00:00
8a6e5ed3ce git-svn-id: svn://svn.icms.temple.edu/lammps-ro/trunk@15637 f3b2605a-c512-4ea7-a41b-209d697bcdaa 2016-09-23 22:59:43 +00:00
756cac0f60 git-svn-id: svn://svn.icms.temple.edu/lammps-ro/trunk@15636 f3b2605a-c512-4ea7-a41b-209d697bcdaa 2016-09-23 22:59:35 +00:00
8662662afe fix ti/spring
git-svn-id: svn://svn.icms.temple.edu/lammps-ro/trunk@15635 f3b2605a-c512-4ea7-a41b-209d697bcdaa
2016-09-23 21:14:00 +00:00
f718c54430 sync with GH
git-svn-id: svn://svn.icms.temple.edu/lammps-ro/trunk@15634 f3b2605a-c512-4ea7-a41b-209d697bcdaa
2016-09-23 21:04:56 +00:00
08f0bf9025 new verion of balance weighting 2016-09-23 14:37:53 -06:00
2a30b76277 git-svn-id: svn://svn.icms.temple.edu/lammps-ro/trunk@15633 f3b2605a-c512-4ea7-a41b-209d697bcdaa 2016-09-23 16:49:40 +00:00
31e41707e0 git-svn-id: svn://svn.icms.temple.edu/lammps-ro/trunk@15632 f3b2605a-c512-4ea7-a41b-209d697bcdaa 2016-09-22 15:46:03 +00:00
32cec47ffb git-svn-id: svn://svn.icms.temple.edu/lammps-ro/trunk@15631 f3b2605a-c512-4ea7-a41b-209d697bcdaa 2016-09-22 15:45:27 +00:00
c22df8db57 git-svn-id: svn://svn.icms.temple.edu/lammps-ro/trunk@15630 f3b2605a-c512-4ea7-a41b-209d697bcdaa 2016-09-22 14:05:57 +00:00
d0bbf3fb97 git-svn-id: svn://svn.icms.temple.edu/lammps-ro/trunk@15629 f3b2605a-c512-4ea7-a41b-209d697bcdaa 2016-09-22 02:22:08 +00:00
32872a7b35 git-svn-id: svn://svn.icms.temple.edu/lammps-ro/trunk@15628 f3b2605a-c512-4ea7-a41b-209d697bcdaa 2016-09-22 02:22:05 +00:00
6dd4480482 git-svn-id: svn://svn.icms.temple.edu/lammps-ro/trunk@15626 f3b2605a-c512-4ea7-a41b-209d697bcdaa 2016-09-21 22:31:49 +00:00
26e16ed968 git-svn-id: svn://svn.icms.temple.edu/lammps-ro/trunk@15625 f3b2605a-c512-4ea7-a41b-209d697bcdaa 2016-09-21 22:31:45 +00:00
ca5ad04b01 git-svn-id: svn://svn.icms.temple.edu/lammps-ro/trunk@15624 f3b2605a-c512-4ea7-a41b-209d697bcdaa 2016-09-21 22:15:17 +00:00
0329aaaf72 git-svn-id: svn://svn.icms.temple.edu/lammps-ro/trunk@15623 f3b2605a-c512-4ea7-a41b-209d697bcdaa 2016-09-21 22:14:06 +00:00
fc434b36b3 git-svn-id: svn://svn.icms.temple.edu/lammps-ro/trunk@15622 f3b2605a-c512-4ea7-a41b-209d697bcdaa 2016-09-21 21:29:19 +00:00
a1364adce1 git-svn-id: svn://svn.icms.temple.edu/lammps-ro/trunk@15621 f3b2605a-c512-4ea7-a41b-209d697bcdaa 2016-09-21 21:26:00 +00:00
c382759406 git-svn-id: svn://svn.icms.temple.edu/lammps-ro/trunk@15620 f3b2605a-c512-4ea7-a41b-209d697bcdaa 2016-09-21 21:25:55 +00:00
e7fb82a645 git-svn-id: svn://svn.icms.temple.edu/lammps-ro/trunk@15619 f3b2605a-c512-4ea7-a41b-209d697bcdaa 2016-09-21 21:22:57 +00:00
03c5ce601b git-svn-id: svn://svn.icms.temple.edu/lammps-ro/trunk@15618 f3b2605a-c512-4ea7-a41b-209d697bcdaa 2016-09-21 21:22:32 +00:00
d7c6f57fe4 git-svn-id: svn://svn.icms.temple.edu/lammps-ro/trunk@15617 f3b2605a-c512-4ea7-a41b-209d697bcdaa 2016-09-21 20:50:23 +00:00
0bcd90195d git-svn-id: svn://svn.icms.temple.edu/lammps-ro/trunk@15616 f3b2605a-c512-4ea7-a41b-209d697bcdaa 2016-09-21 20:38:57 +00:00
d3406df6a0 Updated instructions in IPython notebooks
Make.py is now used to enable exceptions support
2016-09-21 12:07:59 -04:00
72c5792230 git-svn-id: svn://svn.icms.temple.edu/lammps-ro/trunk@15615 f3b2605a-c512-4ea7-a41b-209d697bcdaa 2016-09-21 15:40:42 +00:00
a4c8c9b1f9 Strip IPython notebooks of output 2016-09-21 11:35:00 -04:00
f1183cb97c Remove old copies of IPython notebooks 2016-09-21 11:28:15 -04:00
71f7dde12a git-svn-id: svn://svn.icms.temple.edu/lammps-ro/trunk@15614 f3b2605a-c512-4ea7-a41b-209d697bcdaa 2016-09-21 15:26:31 +00:00
f8c8434c44 git-svn-id: svn://svn.icms.temple.edu/lammps-ro/trunk@15613 f3b2605a-c512-4ea7-a41b-209d697bcdaa 2016-09-20 22:39:20 +00:00
3eee584956 git-svn-id: svn://svn.icms.temple.edu/lammps-ro/trunk@15612 f3b2605a-c512-4ea7-a41b-209d697bcdaa 2016-09-20 18:08:38 +00:00
26b9b955a9 git-svn-id: svn://svn.icms.temple.edu/lammps-ro/trunk@15611 f3b2605a-c512-4ea7-a41b-209d697bcdaa 2016-09-20 18:04:18 +00:00
fe73c3e4e3 git-svn-id: svn://svn.icms.temple.edu/lammps-ro/trunk@15610 f3b2605a-c512-4ea7-a41b-209d697bcdaa 2016-09-20 17:25:49 +00:00
8944d48bd1 git-svn-id: svn://svn.icms.temple.edu/lammps-ro/trunk@15608 f3b2605a-c512-4ea7-a41b-209d697bcdaa 2016-09-20 16:36:15 +00:00
f86bd1fceb git-svn-id: svn://svn.icms.temple.edu/lammps-ro/trunk@15607 f3b2605a-c512-4ea7-a41b-209d697bcdaa 2016-09-20 16:35:30 +00:00
f1d3637b03 git-svn-id: svn://svn.icms.temple.edu/lammps-ro/trunk@15605 f3b2605a-c512-4ea7-a41b-209d697bcdaa 2016-09-20 16:26:57 +00:00
ce3676677e git-svn-id: svn://svn.icms.temple.edu/lammps-ro/trunk@15604 f3b2605a-c512-4ea7-a41b-209d697bcdaa 2016-09-20 16:21:39 +00:00
f81f0da734 git-svn-id: svn://svn.icms.temple.edu/lammps-ro/trunk@15603 f3b2605a-c512-4ea7-a41b-209d697bcdaa 2016-09-20 16:20:45 +00:00
ed9f13663b git-svn-id: svn://svn.icms.temple.edu/lammps-ro/trunk@15602 f3b2605a-c512-4ea7-a41b-209d697bcdaa 2016-09-20 16:20:28 +00:00
4f941abdfd git-svn-id: svn://svn.icms.temple.edu/lammps-ro/trunk@15601 f3b2605a-c512-4ea7-a41b-209d697bcdaa 2016-09-20 16:19:25 +00:00
af4a42345f git-svn-id: svn://svn.icms.temple.edu/lammps-ro/trunk@15600 f3b2605a-c512-4ea7-a41b-209d697bcdaa 2016-09-20 16:14:15 +00:00
df0ed58bbd git-svn-id: svn://svn.icms.temple.edu/lammps-ro/trunk@15599 f3b2605a-c512-4ea7-a41b-209d697bcdaa 2016-09-20 16:12:56 +00:00
8b80d0cf9a git-svn-id: svn://svn.icms.temple.edu/lammps-ro/trunk@15598 f3b2605a-c512-4ea7-a41b-209d697bcdaa 2016-09-20 16:09:43 +00:00
558303072d sync with GH
git-svn-id: svn://svn.icms.temple.edu/lammps-ro/trunk@15597 f3b2605a-c512-4ea7-a41b-209d697bcdaa
2016-09-20 14:52:43 +00:00
900c83960e git-svn-id: svn://svn.icms.temple.edu/lammps-ro/trunk@15595 f3b2605a-c512-4ea7-a41b-209d697bcdaa 2016-09-18 00:06:34 +00:00
484122b8b6 sync with GH
git-svn-id: svn://svn.icms.temple.edu/lammps-ro/trunk@15592 f3b2605a-c512-4ea7-a41b-209d697bcdaa
2016-09-16 19:21:34 +00:00
ed532358ad git-svn-id: svn://svn.icms.temple.edu/lammps-ro/trunk@15591 f3b2605a-c512-4ea7-a41b-209d697bcdaa 2016-09-16 16:29:55 +00:00
5336ec0735 git-svn-id: svn://svn.icms.temple.edu/lammps-ro/trunk@15590 f3b2605a-c512-4ea7-a41b-209d697bcdaa 2016-09-16 16:26:52 +00:00
7d77aea42d git-svn-id: svn://svn.icms.temple.edu/lammps-ro/trunk@15589 f3b2605a-c512-4ea7-a41b-209d697bcdaa 2016-09-16 16:24:05 +00:00
6fd60f50ad git-svn-id: svn://svn.icms.temple.edu/lammps-ro/trunk@15588 f3b2605a-c512-4ea7-a41b-209d697bcdaa 2016-09-16 16:20:06 +00:00
54b2f3c970 git-svn-id: svn://svn.icms.temple.edu/lammps-ro/trunk@15583 f3b2605a-c512-4ea7-a41b-209d697bcdaa 2016-09-15 21:02:02 +00:00
e14eab610e git-svn-id: svn://svn.icms.temple.edu/lammps-ro/trunk@15582 f3b2605a-c512-4ea7-a41b-209d697bcdaa 2016-09-15 21:01:16 +00:00
2049fa7380 git-svn-id: svn://svn.icms.temple.edu/lammps-ro/trunk@15581 f3b2605a-c512-4ea7-a41b-209d697bcdaa 2016-09-15 17:18:05 +00:00
cf33c0e7fb git-svn-id: svn://svn.icms.temple.edu/lammps-ro/trunk@15580 f3b2605a-c512-4ea7-a41b-209d697bcdaa 2016-09-15 16:59:28 +00:00
b23e9f0d54 git-svn-id: svn://svn.icms.temple.edu/lammps-ro/trunk@15579 f3b2605a-c512-4ea7-a41b-209d697bcdaa 2016-09-14 19:29:22 +00:00
40b68820d9 update html docs to be used with latest version of converter tools 2016-09-14 14:06:25 -04:00
90e22a7909 Merge branch 'integration' into weighted-balancing 2016-09-14 14:04:02 -04:00
b29782d5ab git-svn-id: svn://svn.icms.temple.edu/lammps-ro/trunk@15577 f3b2605a-c512-4ea7-a41b-209d697bcdaa 2016-09-14 15:33:37 +00:00
0f6d21acda sync with Git
git-svn-id: svn://svn.icms.temple.edu/lammps-ro/trunk@15576 f3b2605a-c512-4ea7-a41b-209d697bcdaa
2016-09-14 15:27:51 +00:00
206f4e18a6 git-svn-id: svn://svn.icms.temple.edu/lammps-ro/trunk@15573 f3b2605a-c512-4ea7-a41b-209d697bcdaa 2016-09-13 23:06:23 +00:00
b3fa20718f git-svn-id: svn://svn.icms.temple.edu/lammps-ro/trunk@15572 f3b2605a-c512-4ea7-a41b-209d697bcdaa 2016-09-13 23:05:03 +00:00
9d0e853925 git-svn-id: svn://svn.icms.temple.edu/lammps-ro/trunk@15571 f3b2605a-c512-4ea7-a41b-209d697bcdaa 2016-09-13 22:58:33 +00:00
babaa839b0 git-svn-id: svn://svn.icms.temple.edu/lammps-ro/trunk@15570 f3b2605a-c512-4ea7-a41b-209d697bcdaa 2016-09-13 22:55:40 +00:00
9f3118341a git-svn-id: svn://svn.icms.temple.edu/lammps-ro/trunk@15569 f3b2605a-c512-4ea7-a41b-209d697bcdaa 2016-09-13 21:00:30 +00:00
342421babb git-svn-id: svn://svn.icms.temple.edu/lammps-ro/trunk@15568 f3b2605a-c512-4ea7-a41b-209d697bcdaa 2016-09-13 20:43:30 +00:00
423052134b git-svn-id: svn://svn.icms.temple.edu/lammps-ro/trunk@15567 f3b2605a-c512-4ea7-a41b-209d697bcdaa 2016-09-13 20:43:17 +00:00
fd5363fb6e git-svn-id: svn://svn.icms.temple.edu/lammps-ro/trunk@15566 f3b2605a-c512-4ea7-a41b-209d697bcdaa 2016-09-13 20:42:47 +00:00
d913f5e094 Fixing Kokkos bugs
git-svn-id: svn://svn.icms.temple.edu/lammps-ro/trunk@15565 f3b2605a-c512-4ea7-a41b-209d697bcdaa
2016-09-12 21:09:35 +00:00
a8d7ca367d git-svn-id: svn://svn.icms.temple.edu/lammps-ro/trunk@15564 f3b2605a-c512-4ea7-a41b-209d697bcdaa 2016-09-10 20:19:52 +00:00
99d5bf89bc git-svn-id: svn://svn.icms.temple.edu/lammps-ro/trunk@15563 f3b2605a-c512-4ea7-a41b-209d697bcdaa 2016-09-10 19:48:18 +00:00
1dd7a13d82 sync with GH
git-svn-id: svn://svn.icms.temple.edu/lammps-ro/trunk@15562 f3b2605a-c512-4ea7-a41b-209d697bcdaa
2016-09-08 20:37:31 +00:00
b190abea39 sync with GH
git-svn-id: svn://svn.icms.temple.edu/lammps-ro/trunk@15561 f3b2605a-c512-4ea7-a41b-209d697bcdaa
2016-09-08 20:20:32 +00:00
06b7d56e16 git-svn-id: svn://svn.icms.temple.edu/lammps-ro/trunk@15560 f3b2605a-c512-4ea7-a41b-209d697bcdaa 2016-09-07 17:17:53 +00:00
ee4a1f0452 git-svn-id: svn://svn.icms.temple.edu/lammps-ro/trunk@15559 f3b2605a-c512-4ea7-a41b-209d697bcdaa 2016-09-07 16:12:51 +00:00
d3694613fd git-svn-id: svn://svn.icms.temple.edu/lammps-ro/trunk@15558 f3b2605a-c512-4ea7-a41b-209d697bcdaa 2016-09-07 15:49:34 +00:00
bf0c18a0f2 git-svn-id: svn://svn.icms.temple.edu/lammps-ro/trunk@15557 f3b2605a-c512-4ea7-a41b-209d697bcdaa 2016-09-06 23:19:15 +00:00
39be4185c4 Updating Kokkos lib
git-svn-id: svn://svn.icms.temple.edu/lammps-ro/trunk@15556 f3b2605a-c512-4ea7-a41b-209d697bcdaa
2016-09-06 23:06:32 +00:00
1ad033ec0c Updating Kokkos lib
git-svn-id: svn://svn.icms.temple.edu/lammps-ro/trunk@15555 f3b2605a-c512-4ea7-a41b-209d697bcdaa
2016-09-06 23:02:50 +00:00
f67a9722ea git-svn-id: svn://svn.icms.temple.edu/lammps-ro/trunk@15554 f3b2605a-c512-4ea7-a41b-209d697bcdaa 2016-09-06 23:01:25 +00:00
06bac161ae git-svn-id: svn://svn.icms.temple.edu/lammps-ro/trunk@15553 f3b2605a-c512-4ea7-a41b-209d697bcdaa 2016-09-06 22:58:43 +00:00
5277242cfe GH changes to doc pages
git-svn-id: svn://svn.icms.temple.edu/lammps-ro/trunk@15552 f3b2605a-c512-4ea7-a41b-209d697bcdaa
2016-09-06 22:56:36 +00:00
83f139642e Reverting optimizations that hurt performance on some compilers
git-svn-id: svn://svn.icms.temple.edu/lammps-ro/trunk@15551 f3b2605a-c512-4ea7-a41b-209d697bcdaa
2016-09-06 22:09:41 +00:00
5568320bd6 git-svn-id: svn://svn.icms.temple.edu/lammps-ro/trunk@15549 f3b2605a-c512-4ea7-a41b-209d697bcdaa 2016-09-06 22:05:53 +00:00
74d0bc4df6 git-svn-id: svn://svn.icms.temple.edu/lammps-ro/trunk@15548 f3b2605a-c512-4ea7-a41b-209d697bcdaa 2016-09-06 22:05:14 +00:00
56945a56aa git-svn-id: svn://svn.icms.temple.edu/lammps-ro/trunk@15547 f3b2605a-c512-4ea7-a41b-209d697bcdaa 2016-09-06 21:55:39 +00:00
f9c106897f git-svn-id: svn://svn.icms.temple.edu/lammps-ro/trunk@15545 f3b2605a-c512-4ea7-a41b-209d697bcdaa 2016-09-06 16:53:15 +00:00
626ae8d85c git-svn-id: svn://svn.icms.temple.edu/lammps-ro/trunk@15544 f3b2605a-c512-4ea7-a41b-209d697bcdaa 2016-09-06 16:52:36 +00:00
4282107e5d git-svn-id: svn://svn.icms.temple.edu/lammps-ro/trunk@15543 f3b2605a-c512-4ea7-a41b-209d697bcdaa 2016-09-06 16:39:57 +00:00
1e11d2d923 git-svn-id: svn://svn.icms.temple.edu/lammps-ro/trunk@15541 f3b2605a-c512-4ea7-a41b-209d697bcdaa 2016-09-06 16:38:58 +00:00
c21cf0364f git-svn-id: svn://svn.icms.temple.edu/lammps-ro/trunk@15540 f3b2605a-c512-4ea7-a41b-209d697bcdaa 2016-09-06 16:33:48 +00:00
688b1f1efc Fixing bug in Kokkos ReaxFF
git-svn-id: svn://svn.icms.temple.edu/lammps-ro/trunk@15539 f3b2605a-c512-4ea7-a41b-209d697bcdaa
2016-09-06 14:06:59 +00:00
fc80281fd9 Fixing bugs in per-atom
git-svn-id: svn://svn.icms.temple.edu/lammps-ro/trunk@15538 f3b2605a-c512-4ea7-a41b-209d697bcdaa
2016-09-02 22:45:29 +00:00
519a3ee242 Adding Kokkos version of PPPM
git-svn-id: svn://svn.icms.temple.edu/lammps-ro/trunk@15537 f3b2605a-c512-4ea7-a41b-209d697bcdaa
2016-09-01 21:45:00 +00:00
a4914bc9d8 Adding Kokkos version of PPPM
git-svn-id: svn://svn.icms.temple.edu/lammps-ro/trunk@15536 f3b2605a-c512-4ea7-a41b-209d697bcdaa
2016-09-01 21:01:23 +00:00
b4785cd038 Adding Kokkos version of PPPM
git-svn-id: svn://svn.icms.temple.edu/lammps-ro/trunk@15535 f3b2605a-c512-4ea7-a41b-209d697bcdaa
2016-09-01 20:53:40 +00:00
0f7873c0b8 Merge branch 'integration' into weighted-balancing 2016-09-01 08:26:08 -04:00
3769f9077f chunk doc pages
git-svn-id: svn://svn.icms.temple.edu/lammps-ro/trunk@15534 f3b2605a-c512-4ea7-a41b-209d697bcdaa
2016-09-01 01:58:35 +00:00
159d722cc2 removing searchindex.js
git-svn-id: svn://svn.icms.temple.edu/lammps-ro/trunk@15533 f3b2605a-c512-4ea7-a41b-209d697bcdaa
2016-09-01 01:55:31 +00:00
f94bbc0de0 git-svn-id: svn://svn.icms.temple.edu/lammps-ro/trunk@15532 f3b2605a-c512-4ea7-a41b-209d697bcdaa 2016-08-31 22:21:11 +00:00
fab2f01a58 git-svn-id: svn://svn.icms.temple.edu/lammps-ro/trunk@15531 f3b2605a-c512-4ea7-a41b-209d697bcdaa 2016-08-31 22:20:28 +00:00
ae458497bf git-svn-id: svn://svn.icms.temple.edu/lammps-ro/trunk@15530 f3b2605a-c512-4ea7-a41b-209d697bcdaa 2016-08-31 21:11:34 +00:00
bcb2e6dd38 git-svn-id: svn://svn.icms.temple.edu/lammps-ro/trunk@15529 f3b2605a-c512-4ea7-a41b-209d697bcdaa 2016-08-31 21:10:51 +00:00
93c6c26b83 sync with Git
git-svn-id: svn://svn.icms.temple.edu/lammps-ro/trunk@15528 f3b2605a-c512-4ea7-a41b-209d697bcdaa
2016-08-31 21:08:32 +00:00
eac7217720 Merge remote-tracking branch 'lammps-rw/integration' into weighted-balancing 2016-08-31 16:34:51 -04:00
083ff54c0c small bug fixes
git-svn-id: svn://svn.icms.temple.edu/lammps-ro/trunk@15527 f3b2605a-c512-4ea7-a41b-209d697bcdaa
2016-08-31 20:26:15 +00:00
e3d0a32272 git-svn-id: svn://svn.icms.temple.edu/lammps-ro/trunk@15526 f3b2605a-c512-4ea7-a41b-209d697bcdaa 2016-08-31 20:10:32 +00:00
93401a83c6 reintroduce pointer nullification for fix balance 2016-08-31 15:58:33 -04:00
82859c4e25 Merge branch 'integration' into weighted-balancing 2016-08-31 15:57:02 -04:00
8f6439843d git-svn-id: svn://svn.icms.temple.edu/lammps-ro/trunk@15525 f3b2605a-c512-4ea7-a41b-209d697bcdaa 2016-08-31 19:25:40 +00:00
9d8027c900 git-svn-id: svn://svn.icms.temple.edu/lammps-ro/trunk@15524 f3b2605a-c512-4ea7-a41b-209d697bcdaa 2016-08-31 19:25:08 +00:00
10edfa297b Merge branch 'integration' into weighted-balancing 2016-08-31 06:42:00 -04:00
76acb8caf1 Fixing Kokkos memory issue
git-svn-id: svn://svn.icms.temple.edu/lammps-ro/trunk@15523 f3b2605a-c512-4ea7-a41b-209d697bcdaa
2016-08-30 23:18:07 +00:00
ba444a4c6b git-svn-id: svn://svn.icms.temple.edu/lammps-ro/trunk@15522 f3b2605a-c512-4ea7-a41b-209d697bcdaa 2016-08-30 19:33:56 +00:00
dbaaf4dbbd Removing aggressive_vectorization flag due to safety issue
git-svn-id: svn://svn.icms.temple.edu/lammps-ro/trunk@15521 f3b2605a-c512-4ea7-a41b-209d697bcdaa
2016-08-30 17:52:49 +00:00
958e3e6a80 sync with Git
git-svn-id: svn://svn.icms.temple.edu/lammps-ro/trunk@15520 f3b2605a-c512-4ea7-a41b-209d697bcdaa
2016-08-29 23:49:20 +00:00
2993aec312 sync with Git
git-svn-id: svn://svn.icms.temple.edu/lammps-ro/trunk@15519 f3b2605a-c512-4ea7-a41b-209d697bcdaa
2016-08-29 22:52:03 +00:00
236241b100 sync with Git
git-svn-id: svn://svn.icms.temple.edu/lammps-ro/trunk@15518 f3b2605a-c512-4ea7-a41b-209d697bcdaa
2016-08-27 23:09:15 +00:00
a62bae7d33 git-svn-id: svn://svn.icms.temple.edu/lammps-ro/trunk@15517 f3b2605a-c512-4ea7-a41b-209d697bcdaa 2016-08-27 23:07:38 +00:00
57b24b5668 updated USER-MANIFOLD doc pages
git-svn-id: svn://svn.icms.temple.edu/lammps-ro/trunk@15516 f3b2605a-c512-4ea7-a41b-209d697bcdaa
2016-08-27 23:07:03 +00:00
fc4e63130c git-svn-id: svn://svn.icms.temple.edu/lammps-ro/trunk@15514 f3b2605a-c512-4ea7-a41b-209d697bcdaa 2016-08-27 22:41:46 +00:00
0ec104088f git-svn-id: svn://svn.icms.temple.edu/lammps-ro/trunk@15513 f3b2605a-c512-4ea7-a41b-209d697bcdaa 2016-08-27 22:41:05 +00:00
4f49acf903 git-svn-id: svn://svn.icms.temple.edu/lammps-ro/trunk@15511 f3b2605a-c512-4ea7-a41b-209d697bcdaa 2016-08-27 22:40:37 +00:00
5714890627 git-svn-id: svn://svn.icms.temple.edu/lammps-ro/trunk@15510 f3b2605a-c512-4ea7-a41b-209d697bcdaa 2016-08-27 22:40:11 +00:00
18d05e04a2 git-svn-id: svn://svn.icms.temple.edu/lammps-ro/trunk@15509 f3b2605a-c512-4ea7-a41b-209d697bcdaa 2016-08-27 22:37:35 +00:00
90e6032f97 new fix flow/gauss command
git-svn-id: svn://svn.icms.temple.edu/lammps-ro/trunk@15508 f3b2605a-c512-4ea7-a41b-209d697bcdaa
2016-08-27 22:18:45 +00:00
646d5bb1b9 Added check for undefined hbonds
git-svn-id: svn://svn.icms.temple.edu/lammps-ro/trunk@15507 f3b2605a-c512-4ea7-a41b-209d697bcdaa
2016-08-26 20:03:55 +00:00
5348c1c70f Adding Kokkos warning
git-svn-id: svn://svn.icms.temple.edu/lammps-ro/trunk@15506 f3b2605a-c512-4ea7-a41b-209d697bcdaa
2016-08-26 18:37:44 +00:00
56628fe2b6 Adding Kokkos warning
git-svn-id: svn://svn.icms.temple.edu/lammps-ro/trunk@15505 f3b2605a-c512-4ea7-a41b-209d697bcdaa
2016-08-26 18:17:16 +00:00
8a7fecbd91 Cleaning up code
git-svn-id: svn://svn.icms.temple.edu/lammps-ro/trunk@15504 f3b2605a-c512-4ea7-a41b-209d697bcdaa
2016-08-26 16:32:11 +00:00
cc4b2dd6ed Changing default
git-svn-id: svn://svn.icms.temple.edu/lammps-ro/trunk@15503 f3b2605a-c512-4ea7-a41b-209d697bcdaa
2016-08-26 15:50:25 +00:00
3366136493 Fixing Kokkos memory issue
git-svn-id: svn://svn.icms.temple.edu/lammps-ro/trunk@15502 f3b2605a-c512-4ea7-a41b-209d697bcdaa
2016-08-26 15:43:13 +00:00
b2470fd80d git-svn-id: svn://svn.icms.temple.edu/lammps-ro/trunk@15501 f3b2605a-c512-4ea7-a41b-209d697bcdaa 2016-08-25 17:19:46 +00:00
484e726c78 git-svn-id: svn://svn.icms.temple.edu/lammps-ro/trunk@15500 f3b2605a-c512-4ea7-a41b-209d697bcdaa 2016-08-25 17:15:22 +00:00
67958a8bfa git-svn-id: svn://svn.icms.temple.edu/lammps-ro/trunk@15499 f3b2605a-c512-4ea7-a41b-209d697bcdaa 2016-08-25 17:03:56 +00:00
bfb01b84e6 Fixing compiler warning
git-svn-id: svn://svn.icms.temple.edu/lammps-ro/trunk@15498 f3b2605a-c512-4ea7-a41b-209d697bcdaa
2016-08-25 16:59:45 +00:00
e96ac8eb59 git-svn-id: svn://svn.icms.temple.edu/lammps-ro/trunk@15497 f3b2605a-c512-4ea7-a41b-209d697bcdaa 2016-08-25 16:55:30 +00:00
29d04c1fbb git-svn-id: svn://svn.icms.temple.edu/lammps-ro/trunk@15496 f3b2605a-c512-4ea7-a41b-209d697bcdaa 2016-08-24 20:31:41 +00:00
a411023a75 git-svn-id: svn://svn.icms.temple.edu/lammps-ro/trunk@15495 f3b2605a-c512-4ea7-a41b-209d697bcdaa 2016-08-24 20:25:54 +00:00
647ffab74f git-svn-id: svn://svn.icms.temple.edu/lammps-ro/trunk@15493 f3b2605a-c512-4ea7-a41b-209d697bcdaa 2016-08-23 22:45:54 +00:00
662335db13 git-svn-id: svn://svn.icms.temple.edu/lammps-ro/trunk@15492 f3b2605a-c512-4ea7-a41b-209d697bcdaa 2016-08-23 22:44:48 +00:00
1e1f68c30d git-svn-id: svn://svn.icms.temple.edu/lammps-ro/trunk@15491 f3b2605a-c512-4ea7-a41b-209d697bcdaa 2016-08-23 22:41:41 +00:00
7646321bfb git-svn-id: svn://svn.icms.temple.edu/lammps-ro/trunk@15490 f3b2605a-c512-4ea7-a41b-209d697bcdaa 2016-08-23 22:21:04 +00:00
7bf1d9b40f git-svn-id: svn://svn.icms.temple.edu/lammps-ro/trunk@15489 f3b2605a-c512-4ea7-a41b-209d697bcdaa 2016-08-23 22:17:44 +00:00
d007faca51 Fixing Kokkos output for number of OpenMP threads
git-svn-id: svn://svn.icms.temple.edu/lammps-ro/trunk@15488 f3b2605a-c512-4ea7-a41b-209d697bcdaa
2016-08-23 16:07:26 +00:00
89fc866ba7 Fixing bug on Macs
git-svn-id: svn://svn.icms.temple.edu/lammps-ro/trunk@15487 f3b2605a-c512-4ea7-a41b-209d697bcdaa
2016-08-23 15:45:00 +00:00
74b1caf2e6 undo changes that belong to a different branch or are redundant 2016-08-22 15:46:01 -04:00
243137d552 undo obsoleted changes to group command by iain bethune 2016-08-22 15:23:16 -04:00
40fd97bd4c silence warnings about cases, that cannot happen
(cherry picked from commit 60bf26bad9)
2016-08-22 15:12:24 -04:00
8492212c4b fix bug found by coverity scan
(cherry picked from commit 63b41cb139)
2016-08-22 15:12:24 -04:00
1976314f40 improve the weight assignment algorithm for compute time based balancing
(cherry picked from commit 2b052c2a9c)
2016-08-22 15:12:23 -04:00
17c1d3a941 Fix typo
(cherry picked from commit 3b8ecd5c06)
2016-08-22 15:12:23 -04:00
fec59ee3b9 update documentation for refactored load-balancing
(cherry picked from commit 7abc061bf7)
2016-08-22 15:12:23 -04:00
33a98d79fe remove upper limit for weigh factor on neighbor list and time weights
(cherry picked from commit 797c6dc2dd)
2016-08-22 15:12:23 -04:00
0902b600fb add new imbalance module store, which allows to store weights in an atom property
(cherry picked from commit 5405622f3b)
2016-08-22 15:12:23 -04:00
7f20afe122 convert from using fix property/atom to using fix store
(cherry picked from commit 280aef55d2)
2016-08-22 15:12:22 -04:00
7e0dc7a74d whitespace cleanup
(cherry picked from commit b3bd35c7be)
2016-08-22 15:12:22 -04:00
b954283ec2 properly handle the case of neighbor lists never been computed before
(cherry picked from commit fcba14a0aa)
2016-08-22 15:12:22 -04:00
ecc136b6dc plug small memory leak
(cherry picked from commit c00aa3c600)
2016-08-22 15:12:22 -04:00
4a536d71eb simplify and correct logic to pass weight to balancer algorithms
(cherry picked from commit 529417f86c)
2016-08-22 15:12:22 -04:00
460bc14822 correct string hanlding with building custom property label
(cherry picked from commit 6a519e5eef)
2016-08-22 15:12:21 -04:00
bb40f63a34 we cannot add a fix while creating a fix. move fix addintion to Fix::init()
(cherry picked from commit 4c26534245)
2016-08-22 15:12:21 -04:00
c6699e19e6 rewrote balancing to use per-atom data stored via fix property/atom
(cherry picked from commit 1da862b440)
2016-08-22 15:12:21 -04:00
2574891160 fix optional argument scanning bug
(cherry picked from commit 2a90afe7e9)
2016-08-22 15:12:21 -04:00
332d6821ca remove unused class member
(cherry picked from commit f884bb2c92)
2016-08-22 15:12:20 -04:00
b20108bddb incorporate refactored weighting into fix balance
(cherry picked from commit 71ef6fb4d9)
2016-08-22 15:12:20 -04:00
8d38db07c7 convert weight array from class member to local pointer to temporary storage
(cherry picked from commit ecbbdc2e7f)
2016-08-22 15:12:20 -04:00
4114bafc28 proof-of-concept implementation for neighbor list based balancing with yet unsolved problems
(cherry picked from commit d40de42af8)
2016-08-22 15:12:20 -04:00
23a48916d7 re-factored balance command now works with group and time weights
(cherry picked from commit 3f674e5062)
2016-08-22 15:12:20 -04:00
34b34d8410 complete implementation for group based imbalance class
(cherry picked from commit 8ff0085cba)
2016-08-22 15:12:19 -04:00
a5d38c0875 prototype implementation for extensible imbalance scheme
(cherry picked from commit 362a26a3de)
2016-08-22 15:12:19 -04:00
eb273ab9ea fix elusive uninitialized data bug reported by valgrind
(cherry picked from commit b44492ee05)
2016-08-22 15:12:19 -04:00
3cf6715d40 be a bit more paranoid about initializing data structures
(cherry picked from commit bda51f2bac)
2016-08-22 15:12:19 -04:00
0b0db201d1 make it so that dynamic load balancing only uses the timing since the last balancing
(cherry picked from commit f758a4f4d0)
2016-08-22 15:12:18 -04:00
f76f2c881b minor tweaks and comment fixes
(cherry picked from commit f14e9cee83)
2016-08-22 15:12:18 -04:00
7d08d9991e improve c++-11 compliance. replace variable size stack allocation.
(cherry picked from commit af224028a9)
2016-08-22 15:12:18 -04:00
85cafde77c whitespace cleanup
(cherry picked from commit 2e0b9cae29)
2016-08-22 15:12:18 -04:00
db734c3003 disable debug output and include bond cost as well
(cherry picked from commit 9ea86965c5)
2016-08-22 15:12:18 -04:00
cc77679851 implement wall clock based load balancing cost function support
(cherry picked from commit 2a57dc6db4)
2016-08-22 15:12:17 -04:00
b8ae885de8 update documentation according to the modified implementation based on iain bethune's contributed code
(cherry picked from commit 76b8bbca8e)
2016-08-22 15:12:17 -04:00
66b4c9b847 implement modified version of balance and fix balance according to steve's suggestions and requirements
(cherry picked from commit 5a81288329)
2016-08-22 15:12:17 -04:00
85f58624a7 Comments
(cherry picked from commit 638fb5c119)
2016-08-22 15:12:17 -04:00
fc6270e590 Docs for load balance changes
(cherry picked from commit fc7afc2242)
2016-08-22 15:12:17 -04:00
f784f07b87 Set up branch with load balancing code from master
(cherry picked from commit fd8794f52a)
2016-08-22 15:12:16 -04:00
819 changed files with 86520 additions and 26518 deletions

View File

@ -1,4 +1,4 @@
LAMMPS (15 Feb 2016)
LAMMPS (6 Oct 2016)
# FENE beadspring benchmark
units lj
@ -43,25 +43,25 @@ Neighbor list info ...
master list distance cutoff = 1.52
ghost atom cutoff = 1.52
binsize = 0.76 -> bins = 45 45 45
Memory usage per processor = 11.5189 Mbytes
Memory usage per processor = 12.0423 Mbytes
Step Temp E_pair E_mol TotEng Press
0 0.97029772 0.44484087 20.494523 22.394765 4.6721833
100 0.9729966 0.4361122 20.507698 22.40326 4.6548819
Loop time of 0.978585 on 1 procs for 100 steps with 32000 atoms
Loop time of 0.977647 on 1 procs for 100 steps with 32000 atoms
Performance: 105948.895 tau/day, 102.188 timesteps/s
100.0% CPU use with 1 MPI tasks x no OpenMP threads
Performance: 106050.541 tau/day, 102.286 timesteps/s
99.9% CPU use with 1 MPI tasks x no OpenMP threads
MPI task timing breakdown:
Section | min time | avg time | max time |%varavg| %total
---------------------------------------------------------------
Pair | 0.19562 | 0.19562 | 0.19562 | 0.0 | 19.99
Bond | 0.087475 | 0.087475 | 0.087475 | 0.0 | 8.94
Neigh | 0.44861 | 0.44861 | 0.44861 | 0.0 | 45.84
Comm | 0.032932 | 0.032932 | 0.032932 | 0.0 | 3.37
Output | 0.00010395 | 0.00010395 | 0.00010395 | 0.0 | 0.01
Modify | 0.19413 | 0.19413 | 0.19413 | 0.0 | 19.84
Other | | 0.01972 | | | 2.02
Pair | 0.19421 | 0.19421 | 0.19421 | 0.0 | 19.86
Bond | 0.08741 | 0.08741 | 0.08741 | 0.0 | 8.94
Neigh | 0.45791 | 0.45791 | 0.45791 | 0.0 | 46.84
Comm | 0.032649 | 0.032649 | 0.032649 | 0.0 | 3.34
Output | 0.00012207 | 0.00012207 | 0.00012207 | 0.0 | 0.01
Modify | 0.18071 | 0.18071 | 0.18071 | 0.0 | 18.48
Other | | 0.02464 | | | 2.52
Nlocal: 32000 ave 32000 max 32000 min
Histogram: 1 0 0 0 0 0 0 0 0 0

View File

@ -1,4 +1,4 @@
LAMMPS (15 Feb 2016)
LAMMPS (6 Oct 2016)
# FENE beadspring benchmark
units lj
@ -43,25 +43,25 @@ Neighbor list info ...
master list distance cutoff = 1.52
ghost atom cutoff = 1.52
binsize = 0.76 -> bins = 45 45 45
Memory usage per processor = 3.91518 Mbytes
Memory usage per processor = 4.14663 Mbytes
Step Temp E_pair E_mol TotEng Press
0 0.97029772 0.44484087 20.494523 22.394765 4.6721833
100 0.97145835 0.43803883 20.502691 22.397872 4.626988
Loop time of 0.271187 on 4 procs for 100 steps with 32000 atoms
Loop time of 0.269205 on 4 procs for 100 steps with 32000 atoms
Performance: 382319.453 tau/day, 368.749 timesteps/s
99.6% CPU use with 4 MPI tasks x no OpenMP threads
Performance: 385133.446 tau/day, 371.464 timesteps/s
99.8% CPU use with 4 MPI tasks x no OpenMP threads
MPI task timing breakdown:
Section | min time | avg time | max time |%varavg| %total
---------------------------------------------------------------
Pair | 0.048621 | 0.050076 | 0.051229 | 0.4 | 18.47
Bond | 0.022254 | 0.022942 | 0.023567 | 0.3 | 8.46
Neigh | 0.11873 | 0.11881 | 0.11887 | 0.0 | 43.81
Comm | 0.019066 | 0.021357 | 0.024297 | 1.3 | 7.88
Output | 5.0068e-05 | 5.5015e-05 | 6.1035e-05 | 0.1 | 0.02
Modify | 0.048737 | 0.050198 | 0.051231 | 0.4 | 18.51
Other | | 0.007751 | | | 2.86
Pair | 0.049383 | 0.049756 | 0.049988 | 0.1 | 18.48
Bond | 0.022701 | 0.022813 | 0.022872 | 0.0 | 8.47
Neigh | 0.11982 | 0.12002 | 0.12018 | 0.0 | 44.58
Comm | 0.020274 | 0.021077 | 0.022348 | 0.5 | 7.83
Output | 5.3167e-05 | 5.6148e-05 | 6.3181e-05 | 0.1 | 0.02
Modify | 0.046276 | 0.046809 | 0.047016 | 0.1 | 17.39
Other | | 0.008669 | | | 3.22
Nlocal: 8000 ave 8030 max 7974 min
Histogram: 1 0 0 1 0 1 0 0 0 1

View File

@ -1,4 +1,4 @@
LAMMPS (15 Feb 2016)
LAMMPS (6 Oct 2016)
# FENE beadspring benchmark
variable x index 1
@ -59,25 +59,25 @@ Neighbor list info ...
master list distance cutoff = 1.52
ghost atom cutoff = 1.52
binsize = 0.76 -> bins = 89 89 45
Memory usage per processor = 12.8735 Mbytes
Memory usage per processor = 13.2993 Mbytes
Step Temp E_pair E_mol TotEng Press
0 0.97027498 0.44484087 20.494523 22.394765 4.6721833
100 0.97682955 0.44239968 20.500229 22.407862 4.6527025
Loop time of 1.20889 on 4 procs for 100 steps with 128000 atoms
Loop time of 1.14845 on 4 procs for 100 steps with 128000 atoms
Performance: 85764.410 tau/day, 82.720 timesteps/s
99.8% CPU use with 4 MPI tasks x no OpenMP threads
Performance: 90277.919 tau/day, 87.074 timesteps/s
99.9% CPU use with 4 MPI tasks x no OpenMP threads
MPI task timing breakdown:
Section | min time | avg time | max time |%varavg| %total
---------------------------------------------------------------
Pair | 0.21738 | 0.23306 | 0.23926 | 1.9 | 19.28
Bond | 0.094536 | 0.10196 | 0.10534 | 1.4 | 8.43
Neigh | 0.52311 | 0.52392 | 0.52519 | 0.1 | 43.34
Comm | 0.090161 | 0.10022 | 0.12557 | 4.7 | 8.29
Output | 0.00012207 | 0.00017327 | 0.00019598 | 0.2 | 0.01
Modify | 0.19662 | 0.20262 | 0.20672 | 0.8 | 16.76
Other | | 0.04694 | | | 3.88
Pair | 0.2203 | 0.22207 | 0.22386 | 0.3 | 19.34
Bond | 0.094861 | 0.095302 | 0.095988 | 0.1 | 8.30
Neigh | 0.52127 | 0.5216 | 0.52189 | 0.0 | 45.42
Comm | 0.079585 | 0.082159 | 0.084366 | 0.7 | 7.15
Output | 0.00013304 | 0.00015306 | 0.00018501 | 0.2 | 0.01
Modify | 0.18351 | 0.18419 | 0.1856 | 0.2 | 16.04
Other | | 0.04298 | | | 3.74
Nlocal: 32000 ave 32015 max 31983 min
Histogram: 1 0 1 0 0 0 0 0 1 1

View File

@ -1,4 +1,4 @@
LAMMPS (15 Feb 2016)
LAMMPS (6 Oct 2016)
# LAMMPS benchmark of granular flow
# chute flow of 32000 atoms with frozen base at 26 degrees
@ -47,24 +47,24 @@ Neighbor list info ...
master list distance cutoff = 1.1
ghost atom cutoff = 1.1
binsize = 0.55 -> bins = 73 37 68
Memory usage per processor = 15.567 Mbytes
Step Atoms KinEng 1 Volume
Memory usage per processor = 16.0904 Mbytes
Step Atoms KinEng c_1 Volume
0 32000 784139.13 1601.1263 29833.783
100 32000 784292.08 1571.0968 29834.707
Loop time of 0.550482 on 1 procs for 100 steps with 32000 atoms
Loop time of 0.534174 on 1 procs for 100 steps with 32000 atoms
Performance: 1569.534 tau/day, 181.659 timesteps/s
100.1% CPU use with 1 MPI tasks x no OpenMP threads
Performance: 1617.451 tau/day, 187.205 timesteps/s
99.8% CPU use with 1 MPI tasks x no OpenMP threads
MPI task timing breakdown:
Section | min time | avg time | max time |%varavg| %total
---------------------------------------------------------------
Pair | 0.33849 | 0.33849 | 0.33849 | 0.0 | 61.49
Neigh | 0.040353 | 0.040353 | 0.040353 | 0.0 | 7.33
Comm | 0.018023 | 0.018023 | 0.018023 | 0.0 | 3.27
Output | 0.00020385 | 0.00020385 | 0.00020385 | 0.0 | 0.04
Modify | 0.13155 | 0.13155 | 0.13155 | 0.0 | 23.90
Other | | 0.02186 | | | 3.97
Pair | 0.33346 | 0.33346 | 0.33346 | 0.0 | 62.43
Neigh | 0.043902 | 0.043902 | 0.043902 | 0.0 | 8.22
Comm | 0.018391 | 0.018391 | 0.018391 | 0.0 | 3.44
Output | 0.00022411 | 0.00022411 | 0.00022411 | 0.0 | 0.04
Modify | 0.11666 | 0.11666 | 0.11666 | 0.0 | 21.84
Other | | 0.02153 | | | 4.03
Nlocal: 32000 ave 32000 max 32000 min
Histogram: 1 0 0 0 0 0 0 0 0 0

View File

@ -1,4 +1,4 @@
LAMMPS (15 Feb 2016)
LAMMPS (6 Oct 2016)
# LAMMPS benchmark of granular flow
# chute flow of 32000 atoms with frozen base at 26 degrees
@ -47,24 +47,24 @@ Neighbor list info ...
master list distance cutoff = 1.1
ghost atom cutoff = 1.1
binsize = 0.55 -> bins = 73 37 68
Memory usage per processor = 6.81783 Mbytes
Step Atoms KinEng 1 Volume
Memory usage per processor = 7.04927 Mbytes
Step Atoms KinEng c_1 Volume
0 32000 784139.13 1601.1263 29833.783
100 32000 784292.08 1571.0968 29834.707
Loop time of 0.13141 on 4 procs for 100 steps with 32000 atoms
Loop time of 0.171815 on 4 procs for 100 steps with 32000 atoms
Performance: 6574.833 tau/day, 760.976 timesteps/s
99.3% CPU use with 4 MPI tasks x no OpenMP threads
Performance: 5028.653 tau/day, 582.020 timesteps/s
99.7% CPU use with 4 MPI tasks x no OpenMP threads
MPI task timing breakdown:
Section | min time | avg time | max time |%varavg| %total
---------------------------------------------------------------
Pair | 0.062505 | 0.067 | 0.07152 | 1.5 | 50.99
Neigh | 0.010041 | 0.0101 | 0.010178 | 0.1 | 7.69
Comm | 0.012347 | 0.012895 | 0.013444 | 0.5 | 9.81
Output | 6.3896e-05 | 0.00010294 | 0.00014091 | 0.3 | 0.08
Modify | 0.031802 | 0.032348 | 0.032897 | 0.3 | 24.62
Other | | 0.008965 | | | 6.82
Pair | 0.093691 | 0.096898 | 0.10005 | 0.8 | 56.40
Neigh | 0.011976 | 0.012059 | 0.012146 | 0.1 | 7.02
Comm | 0.016384 | 0.017418 | 0.018465 | 0.8 | 10.14
Output | 7.7963e-05 | 0.00010747 | 0.00013304 | 0.2 | 0.06
Modify | 0.031744 | 0.031943 | 0.032167 | 0.1 | 18.59
Other | | 0.01339 | | | 7.79
Nlocal: 8000 ave 8008 max 7992 min
Histogram: 2 0 0 0 0 0 0 0 0 2

View File

@ -1,4 +1,4 @@
LAMMPS (15 Feb 2016)
LAMMPS (6 Oct 2016)
# LAMMPS benchmark of granular flow
# chute flow of 32000 atoms with frozen base at 26 degrees
@ -57,24 +57,24 @@ Neighbor list info ...
master list distance cutoff = 1.1
ghost atom cutoff = 1.1
binsize = 0.55 -> bins = 146 73 68
Memory usage per processor = 15.7007 Mbytes
Step Atoms KinEng 1 Volume
Memory usage per processor = 16.1265 Mbytes
Step Atoms KinEng c_1 Volume
0 128000 3136556.5 6404.5051 119335.13
100 128000 3137168.3 6284.3873 119338.83
Loop time of 0.906913 on 4 procs for 100 steps with 128000 atoms
Loop time of 0.832365 on 4 procs for 100 steps with 128000 atoms
Performance: 952.683 tau/day, 110.264 timesteps/s
99.7% CPU use with 4 MPI tasks x no OpenMP threads
Performance: 1038.006 tau/day, 120.140 timesteps/s
99.8% CPU use with 4 MPI tasks x no OpenMP threads
MPI task timing breakdown:
Section | min time | avg time | max time |%varavg| %total
---------------------------------------------------------------
Pair | 0.51454 | 0.53094 | 0.55381 | 2.0 | 58.54
Neigh | 0.042597 | 0.043726 | 0.045801 | 0.6 | 4.82
Comm | 0.063027 | 0.064657 | 0.067367 | 0.7 | 7.13
Output | 0.00024891 | 0.00059718 | 0.00086498 | 1.0 | 0.07
Modify | 0.16508 | 0.17656 | 0.1925 | 2.6 | 19.47
Other | | 0.09043 | | | 9.97
Pair | 0.5178 | 0.52208 | 0.52793 | 0.5 | 62.72
Neigh | 0.047003 | 0.047113 | 0.047224 | 0.0 | 5.66
Comm | 0.05233 | 0.052988 | 0.053722 | 0.2 | 6.37
Output | 0.00024986 | 0.00032717 | 0.00036693 | 0.3 | 0.04
Modify | 0.15517 | 0.15627 | 0.15808 | 0.3 | 18.77
Other | | 0.0536 | | | 6.44
Nlocal: 32000 ave 32000 max 32000 min
Histogram: 4 0 0 0 0 0 0 0 0 0
@ -87,4 +87,4 @@ Total # of neighbors = 460532
Ave neighs/atom = 3.59791
Neighbor list builds = 2
Dangerous builds = 0
Total wall time: 0:00:01
Total wall time: 0:00:00

View File

@ -1,4 +1,4 @@
LAMMPS (15 Feb 2016)
LAMMPS (6 Oct 2016)
# bulk Cu lattice
variable x index 1
@ -49,25 +49,25 @@ Neighbor list info ...
master list distance cutoff = 5.95
ghost atom cutoff = 5.95
binsize = 2.975 -> bins = 25 25 25
Memory usage per processor = 10.2238 Mbytes
Memory usage per processor = 11.2238 Mbytes
Step Temp E_pair E_mol TotEng Press
0 1600 -113280 0 -106662.09 18703.573
50 781.69049 -109873.35 0 -106640.13 52273.088
100 801.832 -109957.3 0 -106640.77 51322.821
Loop time of 5.90097 on 1 procs for 100 steps with 32000 atoms
Loop time of 5.96529 on 1 procs for 100 steps with 32000 atoms
Performance: 7.321 ns/day, 3.278 hours/ns, 16.946 timesteps/s
Performance: 7.242 ns/day, 3.314 hours/ns, 16.764 timesteps/s
99.9% CPU use with 1 MPI tasks x no OpenMP threads
MPI task timing breakdown:
Section | min time | avg time | max time |%varavg| %total
---------------------------------------------------------------
Pair | 5.2121 | 5.2121 | 5.2121 | 0.0 | 88.33
Neigh | 0.58212 | 0.58212 | 0.58212 | 0.0 | 9.86
Comm | 0.030392 | 0.030392 | 0.030392 | 0.0 | 0.52
Output | 0.00023389 | 0.00023389 | 0.00023389 | 0.0 | 0.00
Modify | 0.060871 | 0.060871 | 0.060871 | 0.0 | 1.03
Other | | 0.01527 | | | 0.26
Pair | 5.2743 | 5.2743 | 5.2743 | 0.0 | 88.42
Neigh | 0.59212 | 0.59212 | 0.59212 | 0.0 | 9.93
Comm | 0.030399 | 0.030399 | 0.030399 | 0.0 | 0.51
Output | 0.00026202 | 0.00026202 | 0.00026202 | 0.0 | 0.00
Modify | 0.050487 | 0.050487 | 0.050487 | 0.0 | 0.85
Other | | 0.01776 | | | 0.30
Nlocal: 32000 ave 32000 max 32000 min
Histogram: 1 0 0 0 0 0 0 0 0 0

View File

@ -1,4 +1,4 @@
LAMMPS (15 Feb 2016)
LAMMPS (6 Oct 2016)
# bulk Cu lattice
variable x index 1
@ -49,25 +49,25 @@ Neighbor list info ...
master list distance cutoff = 5.95
ghost atom cutoff = 5.95
binsize = 2.975 -> bins = 25 25 25
Memory usage per processor = 5.09629 Mbytes
Memory usage per processor = 5.59629 Mbytes
Step Temp E_pair E_mol TotEng Press
0 1600 -113280 0 -106662.09 18703.573
50 781.69049 -109873.35 0 -106640.13 52273.088
100 801.832 -109957.3 0 -106640.77 51322.821
Loop time of 1.58019 on 4 procs for 100 steps with 32000 atoms
Loop time of 1.64562 on 4 procs for 100 steps with 32000 atoms
Performance: 27.338 ns/day, 0.878 hours/ns, 63.284 timesteps/s
Performance: 26.252 ns/day, 0.914 hours/ns, 60.767 timesteps/s
99.8% CPU use with 4 MPI tasks x no OpenMP threads
MPI task timing breakdown:
Section | min time | avg time | max time |%varavg| %total
---------------------------------------------------------------
Pair | 1.3617 | 1.366 | 1.3723 | 0.4 | 86.45
Neigh | 0.15123 | 0.15232 | 0.15374 | 0.2 | 9.64
Comm | 0.033429 | 0.041275 | 0.047066 | 2.7 | 2.61
Output | 0.00011301 | 0.0001573 | 0.000211 | 0.3 | 0.01
Modify | 0.014694 | 0.015085 | 0.015421 | 0.2 | 0.95
Other | | 0.005342 | | | 0.34
Pair | 1.408 | 1.4175 | 1.4341 | 0.9 | 86.14
Neigh | 0.15512 | 0.15722 | 0.16112 | 0.6 | 9.55
Comm | 0.029105 | 0.049986 | 0.061822 | 5.8 | 3.04
Output | 0.00010991 | 0.00011539 | 0.00012302 | 0.0 | 0.01
Modify | 0.013383 | 0.013573 | 0.013883 | 0.2 | 0.82
Other | | 0.007264 | | | 0.44
Nlocal: 8000 ave 8008 max 7993 min
Histogram: 2 0 0 0 0 0 0 0 1 1

View File

@ -1,4 +1,4 @@
LAMMPS (15 Feb 2016)
LAMMPS (6 Oct 2016)
# bulk Cu lattice
variable x index 1
@ -49,25 +49,25 @@ Neighbor list info ...
master list distance cutoff = 5.95
ghost atom cutoff = 5.95
binsize = 2.975 -> bins = 49 49 25
Memory usage per processor = 10.1402 Mbytes
Memory usage per processor = 11.1402 Mbytes
Step Temp E_pair E_mol TotEng Press
0 1600 -453120 0 -426647.73 18704.012
50 779.50001 -439457.02 0 -426560.06 52355.276
100 797.97828 -439764.76 0 -426562.07 51474.74
Loop time of 6.46849 on 4 procs for 100 steps with 128000 atoms
Loop time of 6.60121 on 4 procs for 100 steps with 128000 atoms
Performance: 6.679 ns/day, 3.594 hours/ns, 15.460 timesteps/s
Performance: 6.544 ns/day, 3.667 hours/ns, 15.149 timesteps/s
99.9% CPU use with 4 MPI tasks x no OpenMP threads
MPI task timing breakdown:
Section | min time | avg time | max time |%varavg| %total
---------------------------------------------------------------
Pair | 5.581 | 5.5997 | 5.6265 | 0.8 | 86.57
Neigh | 0.65287 | 0.658 | 0.66374 | 0.5 | 10.17
Comm | 0.075706 | 0.11015 | 0.13655 | 7.2 | 1.70
Output | 0.00026488 | 0.00028312 | 0.00029302 | 0.1 | 0.00
Modify | 0.069607 | 0.072407 | 0.074555 | 0.7 | 1.12
Other | | 0.02794 | | | 0.43
Pair | 5.6676 | 5.7011 | 5.7469 | 1.3 | 86.36
Neigh | 0.66423 | 0.67119 | 0.68082 | 0.7 | 10.17
Comm | 0.079367 | 0.13668 | 0.1791 | 10.5 | 2.07
Output | 0.00026989 | 0.00028622 | 0.00031209 | 0.1 | 0.00
Modify | 0.060046 | 0.062203 | 0.065009 | 0.9 | 0.94
Other | | 0.02974 | | | 0.45
Nlocal: 32000 ave 32092 max 31914 min
Histogram: 1 0 0 1 0 1 0 0 0 1

View File

@ -1,4 +1,4 @@
LAMMPS (15 Feb 2016)
LAMMPS (6 Oct 2016)
# 3d Lennard-Jones melt
variable x index 1
@ -50,20 +50,20 @@ Memory usage per processor = 8.21387 Mbytes
Step Temp E_pair E_mol TotEng Press
0 1.44 -6.7733681 0 -4.6134356 -5.0197073
100 0.7574531 -5.7585055 0 -4.6223613 0.20726105
Loop time of 2.26309 on 1 procs for 100 steps with 32000 atoms
Loop time of 2.26185 on 1 procs for 100 steps with 32000 atoms
Performance: 19088.920 tau/day, 44.187 timesteps/s
Performance: 19099.377 tau/day, 44.212 timesteps/s
99.9% CPU use with 1 MPI tasks x no OpenMP threads
MPI task timing breakdown:
Section | min time | avg time | max time |%varavg| %total
---------------------------------------------------------------
Pair | 1.9341 | 1.9341 | 1.9341 | 0.0 | 85.46
Neigh | 0.2442 | 0.2442 | 0.2442 | 0.0 | 10.79
Comm | 0.024158 | 0.024158 | 0.024158 | 0.0 | 1.07
Output | 0.00011611 | 0.00011611 | 0.00011611 | 0.0 | 0.01
Modify | 0.053222 | 0.053222 | 0.053222 | 0.0 | 2.35
Other | | 0.007258 | | | 0.32
Pair | 1.9328 | 1.9328 | 1.9328 | 0.0 | 85.45
Neigh | 0.2558 | 0.2558 | 0.2558 | 0.0 | 11.31
Comm | 0.024061 | 0.024061 | 0.024061 | 0.0 | 1.06
Output | 0.00012612 | 0.00012612 | 0.00012612 | 0.0 | 0.01
Modify | 0.040887 | 0.040887 | 0.040887 | 0.0 | 1.81
Other | | 0.008214 | | | 0.36
Nlocal: 32000 ave 32000 max 32000 min
Histogram: 1 0 0 0 0 0 0 0 0 0

View File

@ -1,4 +1,4 @@
LAMMPS (15 Feb 2016)
LAMMPS (6 Oct 2016)
# 3d Lennard-Jones melt
variable x index 1
@ -50,20 +50,20 @@ Memory usage per processor = 4.09506 Mbytes
Step Temp E_pair E_mol TotEng Press
0 1.44 -6.7733681 0 -4.6134356 -5.0197073
100 0.7574531 -5.7585055 0 -4.6223613 0.20726105
Loop time of 0.640733 on 4 procs for 100 steps with 32000 atoms
Loop time of 0.635957 on 4 procs for 100 steps with 32000 atoms
Performance: 67422.779 tau/day, 156.071 timesteps/s
99.7% CPU use with 4 MPI tasks x no OpenMP threads
Performance: 67929.172 tau/day, 157.243 timesteps/s
99.9% CPU use with 4 MPI tasks x no OpenMP threads
MPI task timing breakdown:
Section | min time | avg time | max time |%varavg| %total
---------------------------------------------------------------
Pair | 0.49487 | 0.51733 | 0.5322 | 1.9 | 80.74
Neigh | 0.061131 | 0.063685 | 0.065433 | 0.6 | 9.94
Comm | 0.02457 | 0.042349 | 0.069598 | 8.1 | 6.61
Output | 5.9843e-05 | 6.3181e-05 | 6.6996e-05 | 0.0 | 0.01
Modify | 0.012961 | 0.013863 | 0.014491 | 0.5 | 2.16
Other | | 0.003448 | | | 0.54
Pair | 0.51335 | 0.51822 | 0.52569 | 0.7 | 81.49
Neigh | 0.063695 | 0.064309 | 0.065397 | 0.3 | 10.11
Comm | 0.027525 | 0.03629 | 0.041959 | 3.1 | 5.71
Output | 6.3896e-05 | 6.6698e-05 | 7.081e-05 | 0.0 | 0.01
Modify | 0.012472 | 0.01254 | 0.012618 | 0.1 | 1.97
Other | | 0.004529 | | | 0.71
Nlocal: 8000 ave 8037 max 7964 min
Histogram: 2 0 0 0 0 0 0 0 1 1

View File

@ -1,4 +1,4 @@
LAMMPS (15 Feb 2016)
LAMMPS (6 Oct 2016)
# 3d Lennard-Jones melt
variable x index 1
@ -50,20 +50,20 @@ Memory usage per processor = 8.13678 Mbytes
Step Temp E_pair E_mol TotEng Press
0 1.44 -6.7733681 0 -4.6133849 -5.0196788
100 0.75841891 -5.759957 0 -4.6223375 0.20008866
Loop time of 2.57914 on 4 procs for 100 steps with 128000 atoms
Loop time of 2.55762 on 4 procs for 100 steps with 128000 atoms
Performance: 16749.768 tau/day, 38.773 timesteps/s
Performance: 16890.677 tau/day, 39.099 timesteps/s
99.8% CPU use with 4 MPI tasks x no OpenMP threads
MPI task timing breakdown:
Section | min time | avg time | max time |%varavg| %total
---------------------------------------------------------------
Pair | 2.042 | 2.1092 | 2.1668 | 3.1 | 81.78
Neigh | 0.23982 | 0.24551 | 0.25233 | 1.0 | 9.52
Comm | 0.067088 | 0.13887 | 0.22681 | 15.7 | 5.38
Output | 0.00013185 | 0.00021666 | 0.00027108 | 0.4 | 0.01
Modify | 0.060348 | 0.071269 | 0.077063 | 2.5 | 2.76
Other | | 0.01403 | | | 0.54
Pair | 2.0583 | 2.0988 | 2.1594 | 2.6 | 82.06
Neigh | 0.24411 | 0.24838 | 0.25585 | 0.9 | 9.71
Comm | 0.066397 | 0.13872 | 0.1863 | 11.9 | 5.42
Output | 0.00012994 | 0.00021023 | 0.00025702 | 0.3 | 0.01
Modify | 0.055533 | 0.058343 | 0.061791 | 1.2 | 2.28
Other | | 0.0132 | | | 0.52
Nlocal: 32000 ave 32060 max 31939 min
Histogram: 1 0 1 0 0 0 0 1 0 1

View File

@ -1,4 +1,4 @@
LAMMPS (15 Feb 2016)
LAMMPS (6 Oct 2016)
# Rhodopsin model
units real
@ -56,6 +56,7 @@ timestep 2.0
run 100
PPPM initialization ...
WARNING: Using 12-bit tables for long-range coulomb (../kspace.cpp:316)
G vector (1/distance) = 0.248835
grid = 25 32 32
stencil order = 5
@ -70,41 +71,41 @@ Neighbor list info ...
master list distance cutoff = 12
ghost atom cutoff = 12
binsize = 6 -> bins = 10 13 13
Memory usage per processor = 91.7487 Mbytes
Memory usage per processor = 93.2721 Mbytes
---------------- Step 0 ----- CPU = 0.0000 (sec) ----------------
TotEng = -25356.2064 KinEng = 21444.8313 Temp = 299.0397
PotEng = -46801.0377 E_bond = 2537.9940 E_angle = 10921.3742
E_dihed = 5211.7865 E_impro = 213.5116 E_vdwl = -2307.8634
E_coul = 207025.8927 E_long = -270403.7333 Press = -142.6035
E_coul = 207025.8927 E_long = -270403.7333 Press = -149.3301
Volume = 307995.0335
---------------- Step 50 ----- CPU = 17.6362 (sec) ----------------
TotEng = -25330.0828 KinEng = 21501.0029 Temp = 299.8230
PotEng = -46831.0857 E_bond = 2471.7004 E_angle = 10836.4975
E_dihed = 5239.6299 E_impro = 227.1218 E_vdwl = -1993.2754
E_coul = 206797.6331 E_long = -270410.3930 Press = 237.6701
Volume = 308031.5639
---------------- Step 100 ----- CPU = 35.9089 (sec) ----------------
TotEng = -25290.7593 KinEng = 21592.0117 Temp = 301.0920
PotEng = -46882.7709 E_bond = 2567.9807 E_angle = 10781.9408
E_dihed = 5198.7432 E_impro = 216.7834 E_vdwl = -1902.4783
E_coul = 206659.2326 E_long = -270404.9733 Press = 6.9960
Volume = 308133.9888
Loop time of 35.9089 on 1 procs for 100 steps with 32000 atoms
---------------- Step 50 ----- CPU = 17.2007 (sec) ----------------
TotEng = -25330.0321 KinEng = 21501.0036 Temp = 299.8230
PotEng = -46831.0357 E_bond = 2471.7033 E_angle = 10836.5108
E_dihed = 5239.6316 E_impro = 227.1219 E_vdwl = -1993.2763
E_coul = 206797.6655 E_long = -270410.3927 Press = 237.6866
Volume = 308031.5640
---------------- Step 100 ----- CPU = 35.0315 (sec) ----------------
TotEng = -25290.7387 KinEng = 21591.9096 Temp = 301.0906
PotEng = -46882.6484 E_bond = 2567.9789 E_angle = 10781.9556
E_dihed = 5198.7493 E_impro = 216.7863 E_vdwl = -1902.6458
E_coul = 206659.5006 E_long = -270404.9733 Press = 6.7898
Volume = 308133.9933
Loop time of 35.0316 on 1 procs for 100 steps with 32000 atoms
Performance: 0.481 ns/day, 49.874 hours/ns, 2.785 timesteps/s
Performance: 0.493 ns/day, 48.655 hours/ns, 2.855 timesteps/s
99.9% CPU use with 1 MPI tasks x no OpenMP threads
MPI task timing breakdown:
Section | min time | avg time | max time |%varavg| %total
---------------------------------------------------------------
Pair | 25.731 | 25.731 | 25.731 | 0.0 | 71.66
Bond | 1.2771 | 1.2771 | 1.2771 | 0.0 | 3.56
Kspace | 3.2094 | 3.2094 | 3.2094 | 0.0 | 8.94
Neigh | 4.4538 | 4.4538 | 4.4538 | 0.0 | 12.40
Comm | 0.068507 | 0.068507 | 0.068507 | 0.0 | 0.19
Output | 0.00025916 | 0.00025916 | 0.00025916 | 0.0 | 0.00
Modify | 1.1417 | 1.1417 | 1.1417 | 0.0 | 3.18
Other | | 0.027 | | | 0.08
Pair | 25.021 | 25.021 | 25.021 | 0.0 | 71.42
Bond | 1.2834 | 1.2834 | 1.2834 | 0.0 | 3.66
Kspace | 3.2116 | 3.2116 | 3.2116 | 0.0 | 9.17
Neigh | 4.2767 | 4.2767 | 4.2767 | 0.0 | 12.21
Comm | 0.069283 | 0.069283 | 0.069283 | 0.0 | 0.20
Output | 0.00028205 | 0.00028205 | 0.00028205 | 0.0 | 0.00
Modify | 1.14 | 1.14 | 1.14 | 0.0 | 3.25
Other | | 0.02938 | | | 0.08
Nlocal: 32000 ave 32000 max 32000 min
Histogram: 1 0 0 0 0 0 0 0 0 0
@ -113,9 +114,9 @@ Histogram: 1 0 0 0 0 0 0 0 0 0
Neighs: 1.20281e+07 ave 1.20281e+07 max 1.20281e+07 min
Histogram: 1 0 0 0 0 0 0 0 0 0
Total # of neighbors = 12028107
Total # of neighbors = 12028098
Ave neighs/atom = 375.878
Ave special neighs/atom = 7.43187
Neighbor list builds = 11
Dangerous builds = 0
Total wall time: 0:00:37
Total wall time: 0:00:36

View File

@ -1,4 +1,4 @@
LAMMPS (15 Feb 2016)
LAMMPS (6 Oct 2016)
# Rhodopsin model
units real
@ -56,6 +56,7 @@ timestep 2.0
run 100
PPPM initialization ...
WARNING: Using 12-bit tables for long-range coulomb (../kspace.cpp:316)
G vector (1/distance) = 0.248835
grid = 25 32 32
stencil order = 5
@ -70,52 +71,52 @@ Neighbor list info ...
master list distance cutoff = 12
ghost atom cutoff = 12
binsize = 6 -> bins = 10 13 13
Memory usage per processor = 36.629 Mbytes
Memory usage per processor = 37.3604 Mbytes
---------------- Step 0 ----- CPU = 0.0000 (sec) ----------------
TotEng = -25356.2064 KinEng = 21444.8313 Temp = 299.0397
PotEng = -46801.0377 E_bond = 2537.9940 E_angle = 10921.3742
E_dihed = 5211.7865 E_impro = 213.5116 E_vdwl = -2307.8634
E_coul = 207025.8927 E_long = -270403.7333 Press = -142.6035
E_coul = 207025.8927 E_long = -270403.7333 Press = -149.3301
Volume = 307995.0335
---------------- Step 50 ----- CPU = 4.7461 (sec) ----------------
TotEng = -25330.0828 KinEng = 21501.0029 Temp = 299.8230
PotEng = -46831.0857 E_bond = 2471.7004 E_angle = 10836.4975
E_dihed = 5239.6299 E_impro = 227.1218 E_vdwl = -1993.2754
E_coul = 206797.6331 E_long = -270410.3930 Press = 237.6701
Volume = 308031.5639
---------------- Step 100 ----- CPU = 9.6332 (sec) ----------------
TotEng = -25290.7591 KinEng = 21592.0117 Temp = 301.0920
PotEng = -46882.7708 E_bond = 2567.9807 E_angle = 10781.9408
E_dihed = 5198.7432 E_impro = 216.7834 E_vdwl = -1902.4783
E_coul = 206659.2327 E_long = -270404.9733 Press = 6.9960
Volume = 308133.9888
Loop time of 9.63322 on 4 procs for 100 steps with 32000 atoms
---------------- Step 50 ----- CPU = 4.6056 (sec) ----------------
TotEng = -25330.0321 KinEng = 21501.0036 Temp = 299.8230
PotEng = -46831.0357 E_bond = 2471.7033 E_angle = 10836.5108
E_dihed = 5239.6316 E_impro = 227.1219 E_vdwl = -1993.2763
E_coul = 206797.6655 E_long = -270410.3927 Press = 237.6866
Volume = 308031.5640
---------------- Step 100 ----- CPU = 9.3910 (sec) ----------------
TotEng = -25290.7386 KinEng = 21591.9096 Temp = 301.0906
PotEng = -46882.6482 E_bond = 2567.9789 E_angle = 10781.9556
E_dihed = 5198.7493 E_impro = 216.7863 E_vdwl = -1902.6458
E_coul = 206659.5007 E_long = -270404.9733 Press = 6.7898
Volume = 308133.9933
Loop time of 9.39107 on 4 procs for 100 steps with 32000 atoms
Performance: 1.794 ns/day, 13.379 hours/ns, 10.381 timesteps/s
99.9% CPU use with 4 MPI tasks x no OpenMP threads
Performance: 1.840 ns/day, 13.043 hours/ns, 10.648 timesteps/s
99.8% CPU use with 4 MPI tasks x no OpenMP threads
MPI task timing breakdown:
Section | min time | avg time | max time |%varavg| %total
---------------------------------------------------------------
Pair | 6.4364 | 6.5993 | 6.7208 | 4.7 | 68.51
Bond | 0.30755 | 0.32435 | 0.35704 | 3.4 | 3.37
Kspace | 0.92248 | 1.0782 | 1.2597 | 13.0 | 11.19
Neigh | 1.1669 | 1.1672 | 1.1675 | 0.0 | 12.12
Comm | 0.094674 | 0.098065 | 0.10543 | 1.4 | 1.02
Output | 0.00015521 | 0.00016224 | 0.00018215 | 0.1 | 0.00
Modify | 0.32982 | 0.34654 | 0.35365 | 1.6 | 3.60
Other | | 0.01943 | | | 0.20
Pair | 6.2189 | 6.3266 | 6.6072 | 6.5 | 67.37
Bond | 0.30793 | 0.32122 | 0.3414 | 2.4 | 3.42
Kspace | 0.87994 | 1.1644 | 1.2855 | 15.3 | 12.40
Neigh | 1.1358 | 1.136 | 1.1362 | 0.0 | 12.10
Comm | 0.08292 | 0.084935 | 0.087077 | 0.5 | 0.90
Output | 0.00015712 | 0.00016558 | 0.00018501 | 0.1 | 0.00
Modify | 0.33717 | 0.34246 | 0.34794 | 0.7 | 3.65
Other | | 0.01526 | | | 0.16
Nlocal: 8000 ave 8143 max 7933 min
Histogram: 1 2 0 0 0 0 0 0 0 1
Nghost: 22733.5 ave 22769 max 22693 min
Histogram: 1 0 0 0 0 2 0 0 0 1
Neighs: 3.00703e+06 ave 3.0975e+06 max 2.96493e+06 min
Neighs: 3.00702e+06 ave 3.0975e+06 max 2.96492e+06 min
Histogram: 1 2 0 0 0 0 0 0 0 1
Total # of neighbors = 12028107
Total # of neighbors = 12028098
Ave neighs/atom = 375.878
Ave special neighs/atom = 7.43187
Neighbor list builds = 11
Dangerous builds = 0
Total wall time: 0:00:10
Total wall time: 0:00:09

View File

@ -1,4 +1,4 @@
LAMMPS (15 Feb 2016)
LAMMPS (6 Oct 2016)
# Rhodopsin model
variable x index 1
@ -77,6 +77,7 @@ timestep 2.0
run 100
PPPM initialization ...
WARNING: Using 12-bit tables for long-range coulomb (../kspace.cpp:316)
G vector (1/distance) = 0.248593
grid = 48 60 36
stencil order = 5
@ -91,52 +92,52 @@ Neighbor list info ...
master list distance cutoff = 12
ghost atom cutoff = 12
binsize = 6 -> bins = 19 26 13
Memory usage per processor = 95.5339 Mbytes
Memory usage per processor = 96.9597 Mbytes
---------------- Step 0 ----- CPU = 0.0000 (sec) ----------------
TotEng = -101425.4887 KinEng = 85779.3251 Temp = 299.0304
PotEng = -187204.8138 E_bond = 10151.9760 E_angle = 43685.4968
E_dihed = 20847.1460 E_impro = 854.0463 E_vdwl = -9231.4537
E_coul = 827053.5824 E_long = -1080565.6077 Press = -142.3092
E_coul = 827053.5824 E_long = -1080565.6077 Press = -149.0358
Volume = 1231980.1340
---------------- Step 50 ----- CPU = 18.7806 (sec) ----------------
TotEng = -101320.2677 KinEng = 86003.4837 Temp = 299.8118
PotEng = -187323.7514 E_bond = 9887.1072 E_angle = 43346.7922
E_dihed = 20958.7032 E_impro = 908.4715 E_vdwl = -7973.4457
E_coul = 826141.3831 E_long = -1080592.7629 Press = 238.0161
Volume = 1232126.1855
---------------- Step 100 ----- CPU = 38.3684 (sec) ----------------
TotEng = -101158.1849 KinEng = 86355.6149 Temp = 301.0393
PotEng = -187513.7998 E_bond = 10272.0693 E_angle = 43128.6454
E_dihed = 20793.9759 E_impro = 867.0826 E_vdwl = -7586.7186
E_coul = 825583.7122 E_long = -1080572.5667 Press = 15.2151
Volume = 1232535.8423
Loop time of 38.3684 on 4 procs for 100 steps with 128000 atoms
---------------- Step 50 ----- CPU = 18.1689 (sec) ----------------
TotEng = -101320.0211 KinEng = 86003.4933 Temp = 299.8118
PotEng = -187323.5144 E_bond = 9887.1189 E_angle = 43346.8448
E_dihed = 20958.7108 E_impro = 908.4721 E_vdwl = -7973.4486
E_coul = 826141.5493 E_long = -1080592.7617 Press = 238.0404
Volume = 1232126.1814
---------------- Step 100 ----- CPU = 37.2027 (sec) ----------------
TotEng = -101157.9546 KinEng = 86355.7413 Temp = 301.0398
PotEng = -187513.6959 E_bond = 10272.0456 E_angle = 43128.7018
E_dihed = 20794.0107 E_impro = 867.0928 E_vdwl = -7587.2409
E_coul = 825584.2416 E_long = -1080572.5474 Press = 15.1729
Volume = 1232535.8440
Loop time of 37.2028 on 4 procs for 100 steps with 128000 atoms
Performance: 0.450 ns/day, 53.289 hours/ns, 2.606 timesteps/s
Performance: 0.464 ns/day, 51.671 hours/ns, 2.688 timesteps/s
99.9% CPU use with 4 MPI tasks x no OpenMP threads
MPI task timing breakdown:
Section | min time | avg time | max time |%varavg| %total
---------------------------------------------------------------
Pair | 26.205 | 26.538 | 26.911 | 5.0 | 69.17
Bond | 1.298 | 1.3125 | 1.3277 | 1.0 | 3.42
Kspace | 3.7099 | 4.0992 | 4.4422 | 13.3 | 10.68
Neigh | 4.6137 | 4.6144 | 4.615 | 0.0 | 12.03
Comm | 0.21398 | 0.21992 | 0.22886 | 1.2 | 0.57
Output | 0.00030518 | 0.00031543 | 0.00033307 | 0.1 | 0.00
Modify | 1.5066 | 1.5232 | 1.5388 | 1.0 | 3.97
Other | | 0.06051 | | | 0.16
Pair | 25.431 | 25.738 | 25.984 | 4.0 | 69.18
Bond | 1.2966 | 1.3131 | 1.3226 | 0.9 | 3.53
Kspace | 3.7563 | 4.0123 | 4.3127 | 10.0 | 10.79
Neigh | 4.3778 | 4.378 | 4.3782 | 0.0 | 11.77
Comm | 0.1903 | 0.19549 | 0.20485 | 1.3 | 0.53
Output | 0.00031805 | 0.00037521 | 0.00039601 | 0.2 | 0.00
Modify | 1.4861 | 1.5051 | 1.5122 | 0.9 | 4.05
Other | | 0.05992 | | | 0.16
Nlocal: 32000 ave 32000 max 32000 min
Histogram: 4 0 0 0 0 0 0 0 0 0
Nghost: 47957 ave 47957 max 47957 min
Histogram: 4 0 0 0 0 0 0 0 0 0
Neighs: 1.20281e+07 ave 1.20572e+07 max 1.1999e+07 min
Neighs: 1.20281e+07 ave 1.20572e+07 max 1.19991e+07 min
Histogram: 2 0 0 0 0 0 0 0 0 2
Total # of neighbors = 48112472
Total # of neighbors = 48112540
Ave neighs/atom = 375.879
Ave special neighs/atom = 7.43187
Neighbor list builds = 11
Dangerous builds = 0
Total wall time: 0:00:39
Total wall time: 0:00:38

View File

@ -1,10 +1,46 @@
Generation of LAMMPS Documentation
LAMMPS Documentation
Depending on how you obtained LAMMPS, this directory has 2 or 3
sub-directories and optionally 2 PDF files:
src content files for LAMMPS documentation
html HTML version of the LAMMPS manual (see html/Manual.html)
tools tools and settings for building the documentation
Manual.pdf large PDF version of entire manual
Developer.pdf small PDF with info about how LAMMPS is structured
If you downloaded LAMMPS as a tarball from the web site, all these
directories and files should be included.
If you downloaded LAMMPS from the public SVN or Git repositories, then
the HTML and PDF files are not included. Instead you need to create
them, in one of three ways:
(a) You can "fetch" the current HTML and PDF files from the LAMMPS web
site. Just type "make fetch". This should create a html_www dir and
Manual_www.pdf/Developer_www.pdf files. Note that if new LAMMPS
features have been added more recently than the date of your version,
the fetched documentation will include those changes (but your source
code will not, unless you update your local repository).
(b) You can build the HTML and PDF files yourself, by typing "make
html" followed by "make pdf". Note that the PDF make requires the
HTML files already exist. This requires various tools including
Sphinx, which the build process will attempt to download and install
on your system, if not already available. See more details below.
(c) You can genererate an older, simpler, less-fancy style of HTML
documentation by typing "make old". This will create an "old"
directory. This can be useful if (b) does not work on your box for
some reason, or you want to quickly view the HTML version of a doc
page you have created or edited yourself within the src directory.
E.g. if you are planning to submit a new feature to LAMMPS.
----------------
The generation of all documentation is managed by the Makefile in this
dir.
----------------
Options:
make html # generate HTML in html dir using Sphinx
@ -51,3 +87,10 @@ Once Python 3 is installed, open a Terminal and type
pip3 install virtualenv
This will install virtualenv from the Python Package Index.
----------------
Installing prerequisites for PDF build

BIN
doc/src/JPG/gran_funnel.png Normal file

Binary file not shown.

After

Width:  |  Height:  |  Size: 117 KiB

Binary file not shown.

After

Width:  |  Height:  |  Size: 2.2 KiB

BIN
doc/src/JPG/gran_mixer.png Normal file

Binary file not shown.

After

Width:  |  Height:  |  Size: 224 KiB

Binary file not shown.

After

Width:  |  Height:  |  Size: 3.0 KiB

View File

@ -1,7 +1,7 @@
<!-- HTML_ONLY -->
<HEAD>
<TITLE>LAMMPS Users Manual</TITLE>
<META NAME="docnumber" CONTENT="26 Sep 2016 version">
<META NAME="docnumber" CONTENT="11 Oct 2016 version">
<META NAME="author" CONTENT="http://lammps.sandia.gov - Sandia National Laboratories">
<META NAME="copyright" CONTENT="Copyright (2003) Sandia Corporation. This software and manual is distributed under the GNU General Public License.">
</HEAD>
@ -21,7 +21,7 @@
<H1></H1>
LAMMPS Documentation :c,h3
26 Sep 2016 version :c,h4
11 Oct 2016 version :c,h4
Version info: :h4
@ -109,7 +109,7 @@ it gives quick access to documentation for all LAMMPS commands.
:caption: User Documentation
:name: userdoc
:includehidden:
Section_intro
Section_start
Section_commands
@ -144,7 +144,7 @@ Indices and tables
* :ref:`genindex`
* :ref:`search`
END_RST -->
<!-- HTML_ONLY -->

View File

@ -117,7 +117,7 @@ PPPM. However, 2-FFT PPPM also requires a slightly larger mesh size to
achieve the same accuracy as 4-FFT PPPM. For problems where the FFT
cost is the performance bottleneck (typically large problems running
on many processors), 2-FFT PPPM may be faster than 4-FFT PPPM.
Staggered PPPM performs calculations using two different meshes, one
shifted slightly with respect to the other. This can reduce force
aliasing errors and increase the accuracy of the method, but also

View File

@ -37,14 +37,14 @@ simulation with all the settings. Rather, the input script is read
one line at a time and each command takes effect when it is read.
Thus this sequence of commands:
timestep 0.5
run 100
timestep 0.5
run 100
run 100 :pre
does something different than this sequence:
run 100
timestep 0.5
run 100
timestep 0.5
run 100 :pre
In the first case, the specified timestep (0.5 fmsec) is used for two
@ -97,7 +97,7 @@ single leading "#" will comment out the entire command.
(3) The line is searched repeatedly for $ characters, which indicate
variables that are replaced with a text string. See an exception in
(6).
(6).
If the $ is followed by curly brackets, then the variable name is the
text inside the curly brackets. If no curly brackets follow the $,
@ -123,7 +123,7 @@ variable X equal (xlo+xhi)/2+sqrt(v_area)
region 1 block $X 2 INF INF EDGE EDGE
variable X delete :pre
can be replaced by
can be replaced by
region 1 block $((xlo+xhi)/2+sqrt(v_area)) 2 INF INF EDGE EDGE :pre
@ -599,6 +599,7 @@ USER-INTEL, k = KOKKOS, o = USER-OMP, t = OPT.
"viscous"_fix_viscous.html,
"wall/colloid"_fix_wall.html,
"wall/gran"_fix_wall_gran.html,
"wall/gran/region"_fix_wall_gran_region.html,
"wall/harmonic"_fix_wall.html,
"wall/lj1043"_fix_wall.html,
"wall/lj126"_fix_wall.html,
@ -896,7 +897,7 @@ KOKKOS, o = USER-OMP, t = OPT.
"lubricate/poly (o)"_pair_lubricate.html,
"lubricateU"_pair_lubricateU.html,
"lubricateU/poly"_pair_lubricateU.html,
"meam (o)"_pair_meam.html,
"meam"_pair_meam.html,
"mie/cut (o)"_pair_mie.html,
"morse (got)"_pair_morse.html,
"nb3b/harmonic (o)"_pair_nb3b_harmonic.html,
@ -956,7 +957,7 @@ package"_Section_start.html#start_3.
"lj/sdk/coul/long (go)"_pair_sdk.html,
"lj/sdk/coul/msm (o)"_pair_sdk.html,
"lj/sf (o)"_pair_lj_sf.html,
"meam/spline"_pair_meam_spline.html,
"meam/spline (o)"_pair_meam_spline.html,
"meam/sw/spline"_pair_meam_sw_spline.html,
"mgpt"_pair_mgpt.html,
"morse/smooth/linear"_pair_morse.html,

View File

@ -159,7 +159,7 @@ As a last resort, you can send an email directly to the
These are two alphabetic lists of the "ERROR"_#error and
"WARNING"_#warn messages LAMMPS prints out and the reason why. If the
explanation here is not sufficient, the documentation for the
offending command may help.
offending command may help.
Error and warning messages also list the source file and line number
where the error was generated. For example, this message

View File

@ -54,30 +54,30 @@ accelerate: run with various acceleration options (OpenMP, GPU, Phi)
balance: dynamic load balancing, 2d system
body: body particles, 2d system
colloid: big colloid particles in a small particle solvent, 2d system
comb: models using the COMB potential
comb: models using the COMB potential
coreshell: core/shell model using CORESHELL package
crack: crack propagation in a 2d solid
crack: crack propagation in a 2d solid
deposit: deposit atoms and molecules on a surface
dipole: point dipolar particles, 2d system
dreiding: methanol via Dreiding FF
eim: NaCl using the EIM potential
ellipse: ellipsoidal particles in spherical solvent, 2d system
flow: Couette and Poiseuille flow in a 2d channel
flow: Couette and Poiseuille flow in a 2d channel
friction: frictional contact of spherical asperities between 2d surfaces
hugoniostat: Hugoniostat shock dynamics
indent: spherical indenter into a 2d solid
indent: spherical indenter into a 2d solid
kim: use of potentials in Knowledge Base for Interatomic Models (KIM)
meam: MEAM test for SiC and shear (same as shear examples)
melt: rapid melt of 3d LJ system
meam: MEAM test for SiC and shear (same as shear examples)
melt: rapid melt of 3d LJ system
micelle: self-assembly of small lipid-like molecules into 2d bilayers
min: energy minimization of 2d LJ melt
msst: MSST shock dynamics
min: energy minimization of 2d LJ melt
msst: MSST shock dynamics
nb3b: use of nonbonded 3-body harmonic pair style
neb: nudged elastic band (NEB) calculation for barrier finding
nemd: non-equilibrium MD of 2d sheared system
neb: nudged elastic band (NEB) calculation for barrier finding
nemd: non-equilibrium MD of 2d sheared system
obstacle: flow around two voids in a 2d channel
peptide: dynamics of a small solvated peptide chain (5-mer)
peri: Peridynamic model of cylinder impacted by indenter
peri: Peridynamic model of cylinder impacted by indenter
pour: pouring of granular particles into a 3d box, then chute flow
prd: parallel replica dynamics of vacancy diffusion in bulk Si
python: using embedded Python in a LAMMPS input script
@ -120,7 +120,7 @@ browser.
Uppercase directories :h4
ASPHERE: various aspherical particle models, using ellipsoids, rigid bodies, line/triangle particles, etc
COUPLE: examples of how to use LAMMPS as a library
COUPLE: examples of how to use LAMMPS as a library
DIFFUSE: compute diffusion coefficients via several methods
ELASTIC: compute elastic constants at zero temperature
ELASTIC_T: compute elastic constants at finite temperature

View File

@ -37,7 +37,7 @@ pitfalls or alternatives.
Please see some of the closed issues for examples of how to
suggest code enhancements, submit proposed changes, or report
elated issues and how they are resoved.
possible bugs and how they are resoved.
As an alternative to using GitHub, you may e-mail the
"core developers"_http://lammps.sandia.gov/authors.html or send
@ -71,7 +71,7 @@ a parallel framework similar to LAMMPS. Most notably, these have
included many-body potentials - Stillinger-Weber, Tersoff, ReaxFF -
and the associated charge-equilibration routines needed for ReaxFF.
The "History link"_http://lammps.sandia.gov/history.html on the
The "History link"_http://lammps.sandia.gov/history.html on the
LAMMPS WWW page gives a timeline of features added to the
C++ open-source version of LAMMPS over the last several years.
@ -80,7 +80,7 @@ site"_lws, except for Warp & GranFlow which were primarily used
internally. A brief listing of their features is given here.
LAMMPS 2001
F90 + MPI
dynamic memory
spatial-decomposition parallelism
@ -96,7 +96,7 @@ LAMMPS 2001
user-defined diagnostics :ul
LAMMPS 99
F77 + MPI
static memory allocation
spatial-decomposition parallelism

View File

@ -4,7 +4,7 @@
:link(ld,Manual.html)
:link(lc,Section_commands.html#comm)
:line
:line
6. How-to discussions :h3
@ -68,7 +68,7 @@ Look at the {in.chain} input script provided in the {bench} directory
of the LAMMPS distribution to see the original script that these 2
scripts are based on. If that script had the line
restart 50 tmp.restart :pre
restart 50 tmp.restart :pre
added to it, it would produce 2 binary restart files (tmp.restart.50
and tmp.restart.100) as it ran.
@ -76,17 +76,17 @@ and tmp.restart.100) as it ran.
This script could be used to read the 1st restart file and re-run the
last 50 timesteps:
read_restart tmp.restart.50 :pre
read_restart tmp.restart.50 :pre
neighbor 0.4 bin
neigh_modify every 1 delay 1 :pre
neighbor 0.4 bin
neigh_modify every 1 delay 1 :pre
fix 1 all nve
fix 2 all langevin 1.0 1.0 10.0 904297 :pre
fix 1 all nve
fix 2 all langevin 1.0 1.0 10.0 904297 :pre
timestep 0.012 :pre
timestep 0.012 :pre
run 50 :pre
run 50 :pre
Note that the following commands do not need to be repeated because
their settings are included in the restart file: {units, atom_style,
@ -107,25 +107,25 @@ lmp_g++ -r tmp.restart.50 tmp.restart.data :pre
Then, this script could be used to re-run the last 50 steps:
units lj
atom_style bond
pair_style lj/cut 1.12
pair_modify shift yes
bond_style fene
units lj
atom_style bond
pair_style lj/cut 1.12
pair_modify shift yes
bond_style fene
special_bonds 0.0 1.0 1.0 :pre
read_data tmp.restart.data :pre
read_data tmp.restart.data :pre
neighbor 0.4 bin
neigh_modify every 1 delay 1 :pre
neighbor 0.4 bin
neigh_modify every 1 delay 1 :pre
fix 1 all nve
fix 2 all langevin 1.0 1.0 10.0 904297 :pre
fix 1 all nve
fix 2 all langevin 1.0 1.0 10.0 904297 :pre
timestep 0.012 :pre
timestep 0.012 :pre
reset_timestep 50
run 50 :pre
reset_timestep 50
run 50 :pre
Note that nearly all the settings specified in the original {in.chain}
script must be repeated, except the {pair_coeff} and {bond_coeff}
@ -522,7 +522,7 @@ H mass = 1.008
O charge = -1.040
H charge = 0.520
r0 of OH bond = 0.9572
theta of HOH angle = 104.52
theta of HOH angle = 104.52
OM distance = 0.15
LJ epsilon of O-O = 0.1550
LJ sigma of O-O = 3.1536
@ -629,7 +629,7 @@ the SPC and SPC/E models.
Wikipedia also has a nice article on "water
models"_http://en.wikipedia.org/wiki/Water_model.
:line
:line
6.10 Coupling LAMMPS to other codes :link(howto_10),h4
@ -729,7 +729,7 @@ LAMMPS and half to the other code and run both codes simultaneously
before syncing them up periodically. Or it might instantiate multiple
instances of LAMMPS to perform different calculations.
:line
:line
6.11 Visualizing LAMMPS snapshots :link(howto_11),h4
@ -832,7 +832,7 @@ rotation of [A], [B], and [C] and can be computed as follows:
where A = | [A] | indicates the scalar length of [A]. The hat symbol (^)
indicates the corresponding unit vector. {beta} and {gamma} are angles
between the vectors described below. Note that by construction,
between the vectors described below. Note that by construction,
[a], [b], and [c] have strictly positive x, y, and z components, respectively.
If it should happen that
[A], [B], and [C] form a left-handed basis, then the above equations
@ -841,17 +841,17 @@ to first apply an inversion. This can be achieved
by interchanging two basis vectors or by changing the sign of one of them.
For consistency, the same rotation/inversion applied to the basis vectors
must also be applied to atom positions, velocities,
must also be applied to atom positions, velocities,
and any other vector quantities.
This can be conveniently achieved by first converting to
This can be conveniently achieved by first converting to
fractional coordinates in the
old basis and then converting to distance coordinates in the new basis.
The transformation is given by the following equation:
:c,image(Eqs/rotate.jpg)
where {V} is the volume of the box, [X] is the original vector quantity and
[x] is the vector in the LAMMPS basis.
where {V} is the volume of the box, [X] is the original vector quantity and
[x] is the vector in the LAMMPS basis.
There is no requirement that a triclinic box be periodic in any
dimension, though it typically should be in at least the 2nd dimension
@ -938,17 +938,17 @@ defined above. The relationship between these 6 quantities
(a,b,c,alpha,beta,gamma) and the LAMMPS box sizes (lx,ly,lz) =
(xhi-xlo,yhi-ylo,zhi-zlo) and tilt factors (xy,xz,yz) is as follows:
:c,image(Eqs/box.jpg)
:c,image(Eqs/box.jpg)
The inverse relationship can be written as follows:
:c,image(Eqs/box_inverse.jpg)
:c,image(Eqs/box_inverse.jpg)
The values of {a}, {b}, {c} , {alpha}, {beta} , and {gamma} can be printed
out or accessed by computes using the
"thermo_style custom"_thermo_style.html keywords
The values of {a}, {b}, {c} , {alpha}, {beta} , and {gamma} can be printed
out or accessed by computes using the
"thermo_style custom"_thermo_style.html keywords
{cella}, {cellb}, {cellc}, {cellalpha}, {cellbeta}, {cellgamma},
respectively.
respectively.
As discussed on the "dump"_dump.html command doc page, when the BOX
BOUNDS for a snapshot is written to a dump file for a triclinic box,
@ -2092,11 +2092,11 @@ lattice fcc 5.376 orient x 1 0 0 orient y 0 1 0 orient z 0 0 1
region box block 0 4 0 4 0 4
create_box 1 box
create_atoms 1 box
mass 1 39.948
mass 1 39.948
pair_style lj/cut 13.0
pair_coeff * * 0.2381 3.405
timestep $\{dt\}
thermo $d :pre
thermo $d :pre
# equilibration and thermalization :pre
@ -2123,14 +2123,14 @@ thermo_style custom step temp press v_pxy v_pxz v_pyz v_v11 v_v22 v_v33
run 100000
variable v equal (v_v11+v_v22+v_v33)/3.0
variable ndens equal count(all)/vol
print "average viscosity: $v \[Pa.s/] @ $T K, $\{ndens\} /A^3" :pre
print "average viscosity: $v \[Pa.s\] @ $T K, $\{ndens\} /A^3" :pre
The fifth method is related to the above Green-Kubo method,
but uses the Einstein formulation, analogous to the Einstein
mean-square-displacement formulation for self-diffusivity. The
time-integrated momentum fluxes play the role of Cartesian
coordinates, whose mean-square displacement increases linearly
with time at sufficiently long times.
with time at sufficiently long times.
:line
@ -2510,8 +2510,8 @@ the electrostatic environment inducing polarizability.
Technically, shells are attached to the cores by a spring force f =
k*r where k is a parametrized spring constant and r is the distance
between the core and the shell. The charges of the core and the shell
add up to the ion charge, thus q(ion) = q(core) + q(shell). This
setup introduces the ion polarizability (alpha) given by
add up to the ion charge, thus q(ion) = q(core) + q(shell). This
setup introduces the ion polarizability (alpha) given by
alpha = q(shell)^2 / k. In a
similar fashion the mass of the ion is distributed on the core and the
shell with the core having the larger mass.
@ -2526,7 +2526,7 @@ for NaCl, as found in examples/coreshell, has this format:
432 atoms # core and shell atoms
216 bonds # number of core/shell springs :pre
4 atom types # 2 cores and 2 shells for Na and Cl
4 atom types # 2 cores and 2 shells for Na and Cl
2 bond types :pre
0.0 24.09597 xlo xhi
@ -2545,19 +2545,19 @@ Atoms :pre
1 1 2 1.5005 0.00000000 0.00000000 0.00000000 # core of core/shell pair 1
2 1 4 -2.5005 0.00000000 0.00000000 0.00000000 # shell of core/shell pair 1
3 2 1 1.5056 4.01599500 4.01599500 4.01599500 # core of core/shell pair 2
4 2 3 -0.5056 4.01599500 4.01599500 4.01599500 # shell of core/shell pair 2
4 2 3 -0.5056 4.01599500 4.01599500 4.01599500 # shell of core/shell pair 2
(...) :pre
Bonds # Bond topology for spring forces :pre
1 2 1 2 # spring for core/shell pair 1
2 2 3 4 # spring for core/shell pair 2
2 2 3 4 # spring for core/shell pair 2
(...) :pre
Non-Coulombic (e.g. Lennard-Jones) pairwise interactions are only
defined between the shells. Coulombic interactions are defined
between all cores and shells. If desired, additional bonds can be
specified between cores.
specified between cores.
The "special_bonds"_special_bonds.html command should be used to
turn-off the Coulombic interaction within core/shell pairs, since that
@ -2620,7 +2620,7 @@ Note that to perform thermostatting using this definition of
temperature, the "fix modify temp"_fix_modify.html command should be
used to assign the compute to the thermostat fix. Likewise the
"thermo_modify temp"_thermo_modify.html command can be used to make
this temperature be output for the overall system.
this temperature be output for the overall system.
For the NaCl example, this can be done as follows:
@ -2632,13 +2632,13 @@ fix thermostatequ all nve # integrator as needed f
fix_modify thermoberendsen temp CSequ
thermo_modify temp CSequ # output of center-of-mass derived temperature :pre
If "compute temp/cs"_compute_temp_cs.html is used, the decoupled
relative motion of the core and the shell should in theory be
If "compute temp/cs"_compute_temp_cs.html is used, the decoupled
relative motion of the core and the shell should in theory be
stable. However numerical fluctuation can introduce a small
momentum to the system, which is noticable over long trajectories.
Therefore it is recomendable to use the "fix
momentum"_fix_momentum.html command in combination with "compute
temp/cs"_compute_temp_cs.html when equilibrating the system to
Therefore it is recomendable to use the "fix
momentum"_fix_momentum.html command in combination with "compute
temp/cs"_compute_temp_cs.html when equilibrating the system to
prevent any drift.
When intializing the velocities of a system with core/shell pairs, it
@ -2661,17 +2661,17 @@ to the electrostatic environment. This fast movement also limits the
timestep size that can be used.
The primary literature of the adiabatic core/shell model suggests that
the fast relative motion of the core/shell pairs only allows negligible
the fast relative motion of the core/shell pairs only allows negligible
energy transfer to the environment. Therefore it is not intended to
decouple the core/shell degree of freedom from the physical system
during production runs. In other words, the "compute
temp/cs"_compute_temp_cs.html command should not be used during
production runs and is only required during equilibration. This way one
is consistent with literature (based on the code packages DL_POLY or
production runs and is only required during equilibration. This way one
is consistent with literature (based on the code packages DL_POLY or
GULP for instance).
The mentioned energy transfer will typically lead to a a small drift
in total energy over time. This internal energy can be monitored
The mentioned energy transfer will typically lead to a a small drift
in total energy over time. This internal energy can be monitored
using the "compute chunk/atom"_compute_chunk_atom.html and "compute
temp/chunk"_compute_temp_chunk.html commands. The internal kinetic
energies of each core/shell pair can then be summed using the sum()
@ -2702,14 +2702,14 @@ The additional section in the date file would be formatted like this:
CS-Info # header of additional section :pre
1 1 # column 1 = atom ID, column 2 = core/shell ID
2 1
3 2
4 2
5 3
6 3
7 4
8 4
1 1 # column 1 = atom ID, column 2 = core/shell ID
2 1
3 2
4 2
5 3
6 3
7 4
8 4
(...) :pre
:line

View File

@ -181,7 +181,7 @@ Atom creation :h5
displace atoms :ul
Ensembles, constraints, and boundary conditions :h5
("fix"_fix.html command)
("fix"_fix.html command)
2d or 3d systems
orthogonal or non-orthogonal (triclinic symmetry) simulation domains
@ -199,7 +199,7 @@ Ensembles, constraints, and boundary conditions :h5
variety of additional boundary conditions and constraints :ul
Integrators :h5
("run"_run.html, "run_style"_run_style.html, "minimize"_minimize.html commands)
("run"_run.html, "run_style"_run_style.html, "minimize"_minimize.html commands)
velocity-Verlet integrator
Brownian dynamics
@ -213,7 +213,7 @@ Diagnostics :h5
see the various flavors of the "fix"_fix.html and "compute"_compute.html commands :ul
Output :h5
("dump"_dump.html, "restart"_restart.html commands)
("dump"_dump.html, "restart"_restart.html commands)
log file of thermodynamic info
text dump files of atom coords, velocities, other per-atom quantities

View File

@ -182,7 +182,7 @@ Supporting info: "atom_style body"_atom_style.html, "body"_body.html,
"pair_style body"_pair_body.html, examples/body
:line
CLASS2 package :link(CLASS2),h5
Contents: Bond, angle, dihedral, improper, and pair styles for the
@ -206,9 +206,9 @@ Supporting info: "bond_style class2"_bond_class2.html, "angle_style
class2"_angle_class2.html, "dihedral_style
class2"_dihedral_class2.html, "improper_style
class2"_improper_class2.html, "pair_style lj/class2"_pair_class2.html
:line
COLLOID package :link(COLLOID),h5
Contents: Support for coarse-grained colloidal particles. Wall fix
@ -239,9 +239,9 @@ lubricate"_pair_lubricate.html, "pair_style
lubricateU"_pair_lubricateU.html, examples/colloid, examples/srd
:line
COMPRESS package :link(COMPRESS),h5
Contents: Support for compressed output of dump files via the zlib
compression library, using dump styles with a "gz" in their style
name.
@ -271,7 +271,7 @@ atom/gz"_dump.html, "dump cfg/gz"_dump.html, "dump
custom/gz"_dump.html, "dump xyz/gz"_dump.html
:line
CORESHELL package :link(CORESHELL),h5
Contents: Compute and pair styles that implement the adiabatic
@ -302,7 +302,7 @@ buck/coul/long/cs"_pair_cs.html, pair_style
lj/cut/coul/long/cs"_pair_lj.html, examples/coreshell
:line
DIPOLE package :link(DIPOLE),h5
Contents: An atom style and several pair styles to support point
@ -326,9 +326,9 @@ Supporting info: "atom_style dipole"_atom_style.html, "pair_style
lj/cut/dipole/cut"_pair_dipole.html, "pair_style
lj/cut/dipole/long"_pair_dipole.html, "pair_style
lj/long/dipole/long"_pair_dipole.html, examples/dipole
:line
GPU package :link(GPU),h5
Contents: Dozens of pair styles and a version of the PPPM long-range
@ -385,9 +385,9 @@ Pair Styles section of "Section 3.5"_Section_commands.html#cmd_5
for any pair style listed with a (g),
"kspace_style"_kspace_style.html, "package gpu"_package.html,
examples/accelerate, bench/FERMI, bench/KEPLER
:line
GRANULAR package :link(GRANULAR),h5
Contents: Fixes and pair styles that support models of finite-size
@ -412,9 +412,9 @@ Supporting info: "Section 6.6"_Section_howto.html#howto_6, "fix
pour"_fix_pour.html, "fix wall/gran"_fix_wall_gran.html, "pair_style
gran/hooke"_pair_gran.html, "pair_style
gran/hertz/history"_pair_gran.html, examples/pour, bench/in.chute
:line
KIM package :link(KIM),h5
Contents: A pair style that interfaces to the Knowledge Base for
@ -443,9 +443,9 @@ Make.py -p ^kim -a machine :pre
Supporting info: src/KIM/README, lib/kim/README, "pair_style
kim"_pair_kim.html, examples/kim
:line
KOKKOS package :link(KOKKOS),h5
Contents: Dozens of atom, pair, bond, angle, dihedral, improper styles
@ -501,7 +501,7 @@ for any pair style listed with a (k), "package kokkos"_package.html,
examples/accelerate, bench/FERMI, bench/KEPLER
:line
KSPACE package :link(KSPACE),h5
Contents: A variety of long-range Coulombic solvers, and pair styles
@ -543,7 +543,7 @@ which have "long" or "msm" in their style name,
examples/peptide, bench/in.rhodo
:line
MANYBODY package :link(MANYBODY),h5
Contents: A variety of many-body and bond-order potentials. These
@ -565,14 +565,14 @@ make machine :pre
Make.py -p ^manybody -a machine :pre
Supporting info:
Supporting info:
Examples: Pair Styles section of "Section
3.5"_Section_commands.html#cmd_5, examples/comb, examples/eim,
examples/nb3d, examples/vashishta
:line
MC package :link(MC),h5
Contents: Several fixes and a pair style that have Monte Carlo (MC) or
@ -598,9 +598,9 @@ Supporting info: "fix atom/swap"_fix_atom_swap.html, "fix
bond/break"_fix_bond_break.html, "fix
bond/create"_fix_bond_create.html, "fix bond/swap"_fix_bond_swap.html,
"fix gcmc"_fix_gcmc.html, "pair_style dsmc"_pair_dsmc.html
:line
MEAM package :link(MEAM),h5
Contents: A pair style for the modified embedded atom (MEAM)
@ -644,9 +644,9 @@ Make.py -p ^meam -a machine :pre
Supporting info: lib/meam/README, "pair_style meam"_pair_meam.html,
examples/meam
:line
MISC package :link(MISC),h5
Contents: A variety of computes, fixes, and pair styles that are not
@ -670,9 +670,9 @@ Make.py -p ^misc -a machine :pre
Supporting info: "compute ti"_compute_ti.html, "fix
evaporate"_fix_evaporate.html, "fix tmm"_fix_ttm.html, "fix
viscosity"_fix_viscosity.html, examples/misc
:line
MOLECULE package :link(MOLECULE),h5
Contents: A large number of atom, pair, bond, angle, dihedral,
@ -704,7 +704,7 @@ lj/charmm/coul/charmm"_pair_charmm.html,
examples/micelle, examples/peptide, bench/in.chain, bench/in.rhodo
:line
MPIIO package :link(MPIIO),h5
Contents: Support for parallel output/input of dump and restart files
@ -729,9 +729,9 @@ Make.py -p ^mpiio -a machine :pre
Supporting info: "dump"_dump.html, "restart"_restart.html,
"write_restart"_write_restart.html, "read_restart"_read_restart.html
:line
OPT package :link(OPT),h5
Contents: A handful of pair styles with an "opt" in their style name
@ -768,7 +768,7 @@ Supporting info: "Section 5.3"_Section_accelerate.html#acc_3,
listed with an (t), examples/accelerate, bench/KEPLER
:line
PERI package :link(PERI),h5
Contents: Support for the Peridynamics method, a particle-based
@ -796,9 +796,9 @@ Supporting info:
"doc/PDF/PDLammps_VES.pdf"_PDF/PDLammps_VES.pdf, "atom_style
peri"_atom_style.html, "compute damage/atom"_compute_damage_atom.html,
"pair_style peri/pmb"_pair_peri.html, examples/peri
:line
POEMS package :link(POEMS),h5
Contents: A fix that wraps the Parallelizable Open source Efficient
@ -839,13 +839,13 @@ Supporting info: src/POEMS/README, lib/poems/README,
"fix poems"_fix_poems.html, examples/rigid
:line
PYTHON package :link(PYTHON),h5
Contents: A "python"_python.html command which allow you to execute
Python code from a LAMMPS input script. The code can be in a separate
file or embedded in the input script itself. See "Section
11.2"_Section_python.html#py-2 for an overview of using Python from
11.2"_Section_python.html#py_2 for an overview of using Python from
LAMMPS and for other ways to use LAMMPS and Python together.
Building with the PYTHON package assumes you have a Python shared
@ -873,9 +873,9 @@ make machine :pre
Make.py -p ^python -a machine :pre
Supporting info: examples/python
:line
QEQ package :link(QEQ),h5
Contents: Several fixes for performing charge equilibration (QEq) via
@ -897,9 +897,9 @@ make machine :pre
Make.py -p ^qeq -a machine :pre
Supporting info: "fix qeq/*"_fix_qeq.html, examples/qeq
:line
REAX package :link(REAX),h5
Contents: A pair style for the ReaxFF potential, a universal reactive
@ -941,9 +941,9 @@ Make.py -p ^reax -a machine :pre
Supporting info: lib/reax/README, "pair_style reax"_pair_reax.html,
"fix reax/bonds"_fix_reax_bonds.html, examples/reax
:line
REPLICA package :link(REPLICA),h5
Contents: A collection of multi-replica methods that are used by
@ -978,7 +978,7 @@ Supporting info: "Section 6.5"_Section_howto.html#howto_5,
examples/tad
:line
RIGID package :link(RIGID),h5
Contents: A collection of computes and fixes which enforce rigid
@ -1005,7 +1005,7 @@ Supporting info: "compute erotate/rigid"_compute_erotate_rigid.html,
rigid/*"_fix_rigid.html, examples/ASPHERE, examples/rigid
:line
SHOCK package :link(SHOCK),h5
Contents: A small number of fixes useful for running impact
@ -1028,15 +1028,15 @@ Make.py -p ^shock -a machine :pre
Supporting info: "fix append/atoms"_fix_append_atoms.html, "fix
msst"_fix_msst.html, "fix nphug"_fix_nphug.html, "fix
wall/piston"_fix_wall_piston.html, examples/hugoniostat, examples/msst
:line
SNAP package :link(SNAP),h5
Contents: A pair style for the spectral neighbor analysis potential
(SNAP), which is an empirical potential which can be quantum accurate
when fit to an archive of DFT data. Computes useful for analyzing
properties of the potential are also included.
when fit to an archive of DFT data. Computes useful for analyzing
properties of the potential are also included.
To install via make or Make.py:
@ -1055,9 +1055,9 @@ Make.py -p ^snap -a machine :pre
Supporting info: "pair snap"_pair_snap.html, "compute
sna/atom"_compute_sna_atom.html, "compute snad/atom"_compute_sna_atom.html,
"compute snav/atom"_compute_sna_atom.html, examples/snap
:line
SRD package :link(SRD),h5
Contents: Two fixes which implement the Stochastic Rotation Dynamics
@ -1080,9 +1080,9 @@ Make.py -p ^srd -a machine :pre
Supporting info: "fix srd"_fix_srd.html, "fix
wall/srd"_fix_wall_srd.html, examples/srd, examples/ASPHERE
:line
VORONOI package :link(VORONOI),h5
Contents: A "compute voronoi/atom"_compute_voronoi_atom.html command
@ -1129,9 +1129,9 @@ Make.py -p ^voronoi -a machine :pre
Supporting info: src/VORONOI/README, lib/voronoi/README, "compute
voronoi/atom"_compute_voronoi_atom.html, examples/voronoi
:line
4.2 User packages :h4,link(pkg_2)
The current list of user-contributed packages is as follows:
@ -1302,7 +1302,7 @@ fix. The COLVARS library itself is written and maintained by Giacomo
Fiorin (ICMS, Temple University, Philadelphia, PA, USA) and Jerome
Henin (LISM, CNRS, Marseille, France). Contact them directly if you
have questions.
:line
USER-DIFFRACTION package :link(USER-DIFFRACTION),h5
@ -1380,7 +1380,7 @@ in 2007. See src/USER-EFF/README for more details. There are
auxiliary tools for using this package in tools/eff; see its README
file.
Supporting info:
Supporting info:
Author: Andres Jaramillo-Botero at CalTech (ajaramil at
wag.caltech.edu). Contact him directly if you have questions.
@ -1456,21 +1456,21 @@ LINKFLAGS: add -fopenmp :ul
For Phi mode add the following in addition to the CPU mode flags:
CCFLAGS: add -DLMP_INTEL_OFFLOAD and
CCFLAGS: add -DLMP_INTEL_OFFLOAD and
LINKFLAGS: add -offload :ul
And also add this to CCFLAGS:
-offload-option,mic,compiler,"-fp-model fast=2 -mGLOB_default_function_attrs=\"gather_scatter_loop_unroll=4\"" :pre
Examples:
Examples:
:line
USER-LB package :link(USER-LB),h5
Supporting info:
Supporting info:
This package contains a LAMMPS implementation of a background
Lattice-Boltzmann fluid, which can be used to model MD particles
influenced by hydrodynamic forces.
@ -1489,8 +1489,8 @@ Examples: examples/USER/lb
USER-MGPT package :link(USER-MGPT),h5
Supporting info:
Supporting info:
This package contains a fast implementation for LAMMPS of
quantum-based MGPT multi-ion potentials. The MGPT or model GPT method
derives from first-principles DFT-based generalized pseudopotential
@ -1521,8 +1521,8 @@ Examples: examples/USER/mgpt
USER-MISC package :link(USER-MISC),h5
Supporting info:
Supporting info:
The files in this package are a potpourri of (mostly) unrelated
features contributed to LAMMPS by users. Each feature is a single
pair of files (*.cpp and *.h).
@ -1548,8 +1548,8 @@ Examples: examples/USER/misc
USER-MANIFOLD package :link(USER-MANIFOLD),h5
Supporting info:
Supporting info:
This package contains a dump molfile command which uses molfile
plugins that are bundled with the
"VMD"_http://www.ks.uiuc.edu/Research/vmd molecular visualization and
@ -1574,8 +1574,8 @@ Contact him directly if you have questions.
USER-MOLFILE package :link(USER-MOLFILE),h5
Supporting info:
Supporting info:
This package contains a dump molfile command which uses molfile
plugins that are bundled with the
"VMD"_http://www.ks.uiuc.edu/Research/vmd molecular visualization and
@ -1600,12 +1600,12 @@ The person who created this package is Axel Kohlmeyer at Temple U
USER-OMP package :link(USER-OMP),h5
Supporting info:
Supporting info:
This package provides OpenMP multi-threading support and
other optimizations of various LAMMPS pair styles, dihedral
styles, and fix styles.
See this section of the manual to get started:
"Section 5.3"_Section_accelerate.html#acc_3
@ -1643,8 +1643,8 @@ Examples: examples/USER/phonon
USER-QMMM package :link(USER-QMMM),h5
Supporting info:
Supporting info:
This package provides a fix qmmm command which allows LAMMPS to be
used in a QM/MM simulation, currently only in combination with pw.x
code from the "Quantum ESPRESSO"_espresso package.
@ -1667,11 +1667,11 @@ The person who created this package is Axel Kohlmeyer at Temple U
(akohlmey at gmail.com). Contact him directly if you have questions.
:line
USER-QTB package :link(USER-QTB),h5
Supporting info:
Supporting info:
This package provides a self-consistent quantum treatment of the
vibrational modes in a classical molecular dynamics simulation. By
coupling the MD simulation to a colored thermostat, it introduces zero
@ -1701,16 +1701,16 @@ Examples: examples/USER/qtb
USER-QUIP package :link(USER-QUIP),h5
Supporting info:
Supporting info:
Examples: examples/USER/quip
:line
USER-REAXC package :link(USER-REAXC),h5
Supporting info:
Supporting info:
This package contains a implementation for LAMMPS of the ReaxFF force
field. ReaxFF uses distance-dependent bond-order functions to
represent the contributions of chemical bonding to the potential
@ -1748,24 +1748,24 @@ Examples: examples/reax
USER-SMD package :link(USER-SMD),h5
Supporting info:
Supporting info:
This package implements smoothed Mach dynamics (SMD) in
LAMMPS. Currently, the package has the following features:
* Does liquids via traditional Smooth Particle Hydrodynamics (SPH)
* Also solves solids mechanics problems via a state of the art
* Also solves solids mechanics problems via a state of the art
stabilized meshless method with hourglass control.
* Can specify hydrostatic interactions independently from material
* Can specify hydrostatic interactions independently from material
strength models, i.e. pressure and deviatoric stresses are separated.
* Many material models available (Johnson-Cook, plasticity with
hardening, Mie-Grueneisen, Polynomial EOS). Easy to add new
* Many material models available (Johnson-Cook, plasticity with
hardening, Mie-Grueneisen, Polynomial EOS). Easy to add new
material models.
* Rigid boundary conditions (walls) can be loaded as surface geometries
* Rigid boundary conditions (walls) can be loaded as surface geometries
from *.STL files.
See the file doc/PDF/SMD_LAMMPS_userguide.pdf to get started.
@ -1783,8 +1783,8 @@ Examples: examples/USER/smd
USER-SMTBQ package :link(USER-SMTBQ),h5
Supporting info:
Supporting info:
This package implements the Second Moment Tight Binding - QEq (SMTB-Q)
potential for the description of ionocovalent bonds in oxides.
@ -1806,22 +1806,22 @@ Examples: examples/USER/smtbq
USER-SPH package :link(USER-SPH),h5
Supporting info:
Supporting info:
This package implements smoothed particle hydrodynamics (SPH) in
LAMMPS. Currently, the package has the following features:
* Tait, ideal gas, Lennard-Jones equation of states, full support for
* Tait, ideal gas, Lennard-Jones equation of states, full support for
complete (i.e. internal-energy dependent) equations of state
* Plain or Monaghans XSPH integration of the equations of motion
* Density continuity or density summation to propagate the density field
* Commands to set internal energy and density of particles from the
* Commands to set internal energy and density of particles from the
input script
* Output commands to access internal energy and density for dumping and
* Output commands to access internal energy and density for dumping and
thermo output
See the file doc/PDF/SPH_LAMMPS_userguide.pdf to get started.
@ -1839,7 +1839,7 @@ Examples: examples/USER/sph
USER-TALLY package :link(USER-TALLY),h5
Supporting info:
Supporting info:
Examples: examples/USER/tally

View File

@ -23,7 +23,7 @@ In Python lingo, this is "embedding" Python in LAMMPS.
This section describes how to do both.
11.1 "Overview of running LAMMPS from Python"_#py_1
11.2 "Overview of using Python from a LAMMPS script"_#py_2
11.2 "Overview of using Python from a LAMMPS script"_#py_2
11.3 "Building LAMMPS as a shared library"_#py_3
11.4 "Installing the Python wrapper into Python"_#py_4
11.5 "Extending Python with MPI to run in parallel"_#py_5
@ -503,7 +503,7 @@ one of several ways:
The last command requires that the first line of the script be
something like this:
#!/usr/local/bin/python
#!/usr/local/bin/python
#!/usr/local/bin/python -i :pre
where the path points to where you have Python installed, and that you
@ -552,32 +552,32 @@ lmp.command(cmd) # invoke a single LAMMPS command, cmd = "run 100" :pre
xlo = lmp.extract_global(name,type) # extract a global quantity
# name = "boxxlo", "nlocal", etc
# type = 0 = int
# 1 = double :pre
# type = 0 = int
# 1 = double :pre
coords = lmp.extract_atom(name,type) # extract a per-atom quantity
# name = "x", "type", etc
# type = 0 = vector of ints
# 1 = array of ints
# 2 = vector of doubles
# 3 = array of doubles :pre
# type = 0 = vector of ints
# 1 = array of ints
# 2 = vector of doubles
# 3 = array of doubles :pre
eng = lmp.extract_compute(id,style,type) # extract value(s) from a compute
v3 = lmp.extract_fix(id,style,type,i,j) # extract value(s) from a fix
# id = ID of compute or fix
# style = 0 = global data
# 1 = per-atom data
# 2 = local data
# type = 0 = scalar
# 1 = vector
# 2 = array
# i,j = indices of value in global vector or array :pre
# style = 0 = global data
# 1 = per-atom data
# 2 = local data
# type = 0 = scalar
# 1 = vector
# 2 = array
# i,j = indices of value in global vector or array :pre
var = lmp.extract_variable(name,group,flag) # extract value(s) from a variable
# name = name of variable
# group = group ID (ignored for equal-style variables)
# flag = 0 = equal-style variable
# 1 = atom-style variable :pre
# name = name of variable
# group = group ID (ignored for equal-style variables)
# flag = 0 = equal-style variable
# 1 = atom-style variable :pre
flag = lmp.set_variable(name,value) # set existing named string-style variable to value, flag = 0 if successful
natoms = lmp.get_natoms() # total # of atoms as int
@ -724,7 +724,7 @@ lmp.scatter_coords("x",1,3,x) :pre
Alternatively, you can just change values in the vector returned by
gather_atoms("x",1,3), since it is a ctypes vector of doubles.
:line
:line
As noted above, these Python class methods correspond one-to-one with
the functions in the LAMMPS library interface in src/library.cpp and
@ -767,7 +767,7 @@ vizplotgui_tool.py, combination of viz_tool.py and plot.py and gui.py :tb(c=2)
For the viz_tool.py and vizplotgui_tool.py commands, replace "tool"
with "gl" or "atomeye" or "pymol" or "vmd", depending on what
visualization package you have installed.
visualization package you have installed.
Note that for GL, you need to be able to run the Pizza.py GL tool,
which is included in the pizza sub-directory. See the "Pizza.py doc

View File

@ -33,7 +33,7 @@ tar -xzvf lammps*.tar.gz :pre
This will create a LAMMPS directory containing two files and several
sub-directories:
README: text file
LICENSE: the GNU General Public License (GPL)
bench: benchmark problems
@ -600,10 +600,10 @@ LAMMPS will generate a run-time error. As far as we know, the
settings defined in src/lmptype.h are portable and work on every
current system.
In all cases, the size of problem that can be run on a per-processor
basis is limited by 4-byte integer storage to 2^31 atoms per processor
(about 2 billion). This should not normally be a limitation since such
a problem would have a huge per-processor memory footprint due to
In all cases, the size of problem that can be run on a per-processor
basis is limited by 4-byte integer storage to 2^31 atoms per processor
(about 2 billion). This should not normally be a limitation since such
a problem would have a huge per-processor memory footprint due to
neighbor lists and would run very slowly in terms of CPU secs/timestep.
:line
@ -841,7 +841,7 @@ libpackage.a
Makefile.lammps :pre
The Makefile.lammps file will typically be a copy of one of the
Makefile.lammps.* files in the library directory.
Makefile.lammps.* files in the library directory.
Note that you must insure that the settings in Makefile.lammps are
appropriate for your system. If they are not, the LAMMPS build may
@ -883,7 +883,7 @@ A few packages require specific settings in Makefile.machine, to
either build or use the package effectively. These are the
USER-INTEL, KOKKOS, USER-OMP, and OPT packages, used for accelerating
code performance on CPUs or other hardware, as discussed in "Section
5.3"_Section_accelerate.html#acc_3.
5.3"_Section_accelerate.html#acc_3.
A summary of what Makefile.machine changes are needed for each of
these packages is given in "Section 4"_Section_packages.html.
@ -1199,7 +1199,7 @@ installer package from "here"_http://rpm.lammps.org/windows.html
For running the non-MPI executable, follow these steps:
Get a command prompt by going to Start->Run... ,
Get a command prompt by going to Start->Run... ,
then typing "cmd". :ulb,l
Move to the directory where you have your input, e.g. a copy of
@ -1209,7 +1209,7 @@ At the command prompt, type "lmp_serial -in in.lj", replacing [in.lj]
with the name of your LAMMPS input script. :l
:ule
For the MPI version, which allows you to run LAMMPS under Windows on
For the MPI version, which allows you to run LAMMPS under Windows on
multiple processors, follow these steps:
Download and install
@ -1224,7 +1224,7 @@ For this you need to start a Command Prompt in {Administrator Mode}
installation directory, then into the subdirectory [bin] and execute
[smpd.exe -install]. Exit the command window.
Get a new, regular command prompt by going to Start->Run... ,
Get a new, regular command prompt by going to Start->Run... ,
then typing "cmd". :l
Move to the directory where you have your input file
@ -1488,7 +1488,7 @@ of the manual. World- and universe-style "variables"_variable.html
are useful in this context.
-plog file :pre
Specify the base name for the partition log files, so partition N
writes log information to file.N. If file is none, then no partition
log files are created. This overrides the filename specified in the
@ -1499,7 +1499,7 @@ replica_files/log.lammps) If this option is not used the log file for
partition N is log.lammps.N or whatever is specified by the -log
command-line option.
-pscreen file :pre
-pscreen file :pre
Specify the base name for the partition screen file, so partition N
writes screen information to file.N. If file is none, then no
@ -1511,7 +1511,7 @@ sub-directory (-pscreen replica_files/screen). If this option is not
used the screen file for partition N is screen.N or whatever is
specified by the -screen command-line option.
-restart restartfile {remap} datafile keyword value ... :pre
-restart restartfile {remap} datafile keyword value ... :pre
Convert the restart file into a data file and immediately exit. This
is the same operation as if the following 2-line input script were
@ -1572,7 +1572,7 @@ to
so that the processors in each partition will be
0 1 2 4 5 6 8 9 10
0 1 2 4 5 6 8 9 10
3 7 11 :pre
See the "processors" command for how to insure processors from each
@ -1663,12 +1663,12 @@ invokes the default USER-INTEL settings, as if the command "package
intel 1" were used at the top of your input script. These settings
can be changed by using the "-package intel" command-line switch or
the "package intel"_package.html command in your script. If the
USER-OMP package is also installed, the hybrid style with "intel omp"
arguments can be used to make the omp suffix a second choice, if a
requested style is not available in the USER-INTEL package. It will
also invoke the default USER-OMP settings, as if the command "package
omp 0" were used at the top of your input script. These settings can
be changed by using the "-package omp" command-line switch or the
USER-OMP package is also installed, the hybrid style with "intel omp"
arguments can be used to make the omp suffix a second choice, if a
requested style is not available in the USER-INTEL package. It will
also invoke the default USER-OMP settings, as if the command "package
omp 0" were used at the top of your input script. These settings can
be changed by using the "-package omp" command-line switch or the
"package omp"_package.html command in your script.
For the KOKKOS package, using this command-line switch also invokes
@ -1833,7 +1833,7 @@ e.g.
Minimization stats:
Stopping criterion = linesearch alpha is zero
Energy initial, next-to-last, final =
Energy initial, next-to-last, final =
-6372.3765206 -8328.46998942 -8328.46998942
Force two-norm initial, final = 1059.36 5.36874
Force max component initial, final = 58.6026 1.46872

View File

@ -104,7 +104,7 @@ since binary files are not compatible across all platforms.
ch2lmp tool :h4,link(charmm)
The ch2lmp sub-directory contains tools for converting files
back-and-forth between the CHARMM MD code and LAMMPS.
back-and-forth between the CHARMM MD code and LAMMPS.
They are intended to make it easy to use CHARMM as a builder and as a
post-processor for LAMMPS. Using charmm2lammps.pl, you can convert a

View File

@ -29,80 +29,80 @@ Bond Styles: fene, harmonic :l
Dihedral Styles: charmm, harmonic, opls :l
Fixes: nve, npt, nvt, nvt/sllod :l
Improper Styles: cvff, harmonic :l
Pair Styles: buck/coul/cut, buck/coul/long, buck, gayberne,
Pair Styles: buck/coul/cut, buck/coul/long, buck, gayberne,
charmm/coul/long, lj/cut, lj/cut/coul/long, sw, tersoff :l
K-Space Styles: pppm :l
:ule
[Speed-ups to expect:]
The speedups will depend on your simulation, the hardware, which
styles are used, the number of atoms, and the floating-point
precision mode. Performance improvements are shown compared to
LAMMPS {without using other acceleration packages} as these are
under active development (and subject to performance changes). The
The speedups will depend on your simulation, the hardware, which
styles are used, the number of atoms, and the floating-point
precision mode. Performance improvements are shown compared to
LAMMPS {without using other acceleration packages} as these are
under active development (and subject to performance changes). The
measurements were performed using the input files available in
the src/USER-INTEL/TEST directory. These are scalable in size; the
results given are with 512K particles (524K for Liquid Crystal).
the src/USER-INTEL/TEST directory. These are scalable in size; the
results given are with 512K particles (524K for Liquid Crystal).
Most of the simulations are standard LAMMPS benchmarks (indicated
by the filename extension in parenthesis) with modifications to the
run length and to add a warmup run (for use with offload
benchmarks).
by the filename extension in parenthesis) with modifications to the
run length and to add a warmup run (for use with offload
benchmarks).
:c,image(JPG/user_intel.png)
Results are speedups obtained on Intel Xeon E5-2697v4 processors
(code-named Broadwell) and Intel Xeon Phi 7250 processors
Results are speedups obtained on Intel Xeon E5-2697v4 processors
(code-named Broadwell) and Intel Xeon Phi 7250 processors
(code-named Knights Landing) with "18 Jun 2016" LAMMPS built with
Intel Parallel Studio 2016 update 3. Results are with 1 MPI task
per physical core. See {src/USER-INTEL/TEST/README} for the raw
Intel Parallel Studio 2016 update 3. Results are with 1 MPI task
per physical core. See {src/USER-INTEL/TEST/README} for the raw
simulation rates and instructions to reproduce.
:line
[Quick Start for Experienced Users:]
LAMMPS should be built with the USER-INTEL package installed.
LAMMPS should be built with the USER-INTEL package installed.
Simulations should be run with 1 MPI task per physical {core},
not {hardware thread}.
For Intel Xeon CPUs:
Edit src/MAKE/OPTIONS/Makefile.intel_cpu_intelmpi as necessary. :ulb,l
If using {kspace_style pppm} in the input script, add "neigh_modify binsize 3" and "kspace_modify diff ad" to the input script for better
If using {kspace_style pppm} in the input script, add "neigh_modify binsize 3" and "kspace_modify diff ad" to the input script for better
performance. :l
"-pk intel 0 omp 2 -sf intel" added to LAMMPS command-line :l
:ule
For Intel Xeon Phi CPUs for simulations without {kspace_style
For Intel Xeon Phi CPUs for simulations without {kspace_style
pppm} in the input script :
Edit src/MAKE/OPTIONS/Makefile.knl as necessary. :ulb,l
Runs should be performed using MCDRAM. :l
"-pk intel 0 omp 2 -sf intel" {or} "-pk intel 0 omp 4 -sf intel"
should be added to the LAMMPS command-line. Choice for best
"-pk intel 0 omp 2 -sf intel" {or} "-pk intel 0 omp 4 -sf intel"
should be added to the LAMMPS command-line. Choice for best
performance will depend on the simulation. :l
:ule
For Intel Xeon Phi CPUs for simulations with {kspace_style
For Intel Xeon Phi CPUs for simulations with {kspace_style
pppm} in the input script:
Edit src/MAKE/OPTIONS/Makefile.knl as necessary. :ulb,l
Runs should be performed using MCDRAM. :l
Add "neigh_modify binsize 3" to the input script for better
Add "neigh_modify binsize 3" to the input script for better
performance. :l
Add "kspace_modify diff ad" to the input script for better
Add "kspace_modify diff ad" to the input script for better
performance. :l
export KMP_AFFINITY=none :l
"-pk intel 0 omp 3 lrt yes -sf intel" or "-pk intel 0 omp 1 lrt yes
-sf intel" added to LAMMPS command-line. Choice for best performance
-sf intel" added to LAMMPS command-line. Choice for best performance
will depend on the simulation. :l
:ule
For Intel Xeon Phi coprocessors (Offload):
For Intel Xeon Phi coprocessors (Offload):
Edit src/MAKE/OPTIONS/Makefile.intel_coprocessor as necessary :ulb,l
"-pk intel N omp 1" added to command-line where N is the number of
"-pk intel N omp 1" added to command-line where N is the number of
coprocessors per node. :l
:ule
@ -111,7 +111,7 @@ coprocessors per node. :l
[Required hardware/software:]
In order to use offload to coprocessors, an Intel Xeon Phi
coprocessor and an Intel compiler are required. For this, the
coprocessor and an Intel compiler are required. For this, the
recommended version of the Intel compiler is 14.0.1.106 or
versions 15.0.2.044 and higher.
@ -133,7 +133,7 @@ slightly lower.
[Notes about Simultaneous Multithreading:]
Modern CPUs often support Simultaneous Multithreading (SMT). On
Modern CPUs often support Simultaneous Multithreading (SMT). On
Intel processors, this is called Hyper-Threading (HT) technology.
SMT is hardware support for running multiple threads efficiently on
a single core. {Hardware threads} or {logical cores} are often used
@ -141,8 +141,8 @@ to refer to the number of threads that are supported in hardware.
For example, the Intel Xeon E5-2697v4 processor is described
as having 36 cores and 72 threads. This means that 36 MPI processes
or OpenMP threads can run simultaneously on separate cores, but that
up to 72 MPI processes or OpenMP threads can be running on the CPU
without costly operating system context switches.
up to 72 MPI processes or OpenMP threads can be running on the CPU
without costly operating system context switches.
Molecular dynamics simulations will often run faster when making use
of SMT. If a thread becomes stalled, for example because it is
@ -150,7 +150,7 @@ waiting on data that has not yet arrived from memory, another thread
can start running so that the CPU pipeline is still being used
efficiently. Although benefits can be seen by launching a MPI task
for every hardware thread, for multinode simulations, we recommend
that OpenMP threads are used for SMT instead, either with the
that OpenMP threads are used for SMT instead, either with the
USER-INTEL package, "USER-OMP package"_accelerate_omp.html", or
"KOKKOS package"_accelerate_kokkos.html. In the example above, up
to 36X speedups can be observed by using all 36 physical cores with
@ -158,10 +158,10 @@ LAMMPS. By using all 72 hardware threads, an additional 10-30%
performance gain can be achieved.
The BIOS on many platforms allows SMT to be disabled, however, we do
not recommend this on modern processors as there is little to no
not recommend this on modern processors as there is little to no
benefit for any software package in most cases. The operating system
will report every hardware thread as a separate core allowing one to
determine the number of hardware threads available. On Linux systems,
will report every hardware thread as a separate core allowing one to
determine the number of hardware threads available. On Linux systems,
this information can normally be obtained with:
cat /proc/cpuinfo :pre
@ -182,21 +182,21 @@ Makefile.intel_cpu_openpmi # Intel Compiler, OpenMPI, No Offload
Makefile.intel_coprocessor # Intel Compiler, Intel MPI, Offload :pre
Makefile.knl is identical to Makefile.intel_cpu_intelmpi except that
it explicitly specifies that vectorization should be for Intel
Xeon Phi x200 processors making it easier to cross-compile. For
users with recent installations of Intel Parallel Studio, the
it explicitly specifies that vectorization should be for Intel
Xeon Phi x200 processors making it easier to cross-compile. For
users with recent installations of Intel Parallel Studio, the
process can be as simple as:
make yes-user-intel
source /opt/intel/parallel_studio_xe_2016.3.067/psxevars.sh
source /opt/intel/parallel_studio_xe_2016.3.067/psxevars.sh
# or psxevars.csh for C-shell
make intel_cpu_intelmpi :pre
Alternatively, the build can be accomplished with the src/Make.py
script, described in "Section 2.4"_Section_start.html#start_4 of the
Alternatively, the build can be accomplished with the src/Make.py
script, described in "Section 2.4"_Section_start.html#start_4 of the
manual. Type "Make.py -h" for help. For an example:
Make.py -v -p intel omp -intel cpu -a file intel_cpu_intelmpi :pre
Make.py -v -p intel omp -intel cpu -a file intel_cpu_intelmpi :pre
Note that if you build with support for a Phi coprocessor, the same
binary can be used on nodes with or without coprocessors installed.
@ -205,26 +205,26 @@ without offload support will produce a smaller binary.
The general requirements for Makefiles with the USER-INTEL package
are as follows. "-DLAMMPS_MEMALIGN=64" is required for CCFLAGS. When
using Intel compilers, "-restrict" is required and "-qopenmp" is
highly recommended for CCFLAGS and LINKFLAGS. LIB should include
using Intel compilers, "-restrict" is required and "-qopenmp" is
highly recommended for CCFLAGS and LINKFLAGS. LIB should include
"-ltbbmalloc". For builds supporting offload, "-DLMP_INTEL_OFFLOAD"
is required for CCFLAGS and "-qoffload" is required for LINKFLAGS.
Other recommended CCFLAG options for best performance are
"-O2 -fno-alias -ansi-alias -qoverride-limits fp-model fast=2
-no-prec-div". The Make.py command will add all of these
Other recommended CCFLAG options for best performance are
"-O2 -fno-alias -ansi-alias -qoverride-limits fp-model fast=2
-no-prec-div". The Make.py command will add all of these
automatically.
NOTE: The vectorization and math capabilities can differ depending on
the CPU. For Intel compilers, the "-x" flag specifies the type of
processor for which to optimize. "-xHost" specifies that the compiler
should build for the processor used for compiling. For Intel Xeon Phi
should build for the processor used for compiling. For Intel Xeon Phi
x200 series processors, this option is "-xMIC-AVX512". For fourth
generation Intel Xeon (v4/Broadwell) processors, "-xCORE-AVX2" should
generation Intel Xeon (v4/Broadwell) processors, "-xCORE-AVX2" should
be used. For older Intel Xeon processors, "-xAVX" will perform best
in general for the different simulations in LAMMPS. The default
in most of the example Makefiles is to use "-xHost", however this
should not be used when cross-compiling.
[Running LAMMPS with the USER-INTEL package:]
Running LAMMPS with the USER-INTEL package is similar to normal use
@ -232,7 +232,7 @@ with the exceptions that one should 1) specify that LAMMPS should use
the USER-INTEL package, 2) specify the number of OpenMP threads, and
3) optionally specify the specific LAMMPS styles that should use the
USER-INTEL package. 1) and 2) can be performed from the command-line
or by editing the input script. 3) requires editing the input script.
or by editing the input script. 3) requires editing the input script.
Advanced performance tuning options are also described below to get
the best performance.
@ -241,14 +241,14 @@ coprocessor), best performance is normally obtained by using 1 MPI
task per physical core and additional OpenMP threads with SMT. For
Intel Xeon processors, 2 OpenMP threads should be used for SMT.
For Intel Xeon Phi CPUs, 2 or 4 OpenMP threads should be used
(best choice depends on the simulation). In cases where the user
specifies that LRT mode is used (described below), 1 or 3 OpenMP
(best choice depends on the simulation). In cases where the user
specifies that LRT mode is used (described below), 1 or 3 OpenMP
threads should be used. For multi-node runs, using 1 MPI task per
physical core will often perform best, however, depending on the
machine and scale, users might get better performance by decreasing
the number of MPI tasks and using more OpenMP threads. For
performance, the product of the number of MPI tasks and OpenMP
threads should not exceed the number of available hardware threads in
the number of MPI tasks and using more OpenMP threads. For
performance, the product of the number of MPI tasks and OpenMP
threads should not exceed the number of available hardware threads in
almost all cases.
NOTE: Setting core affinity is often used to pin MPI tasks and OpenMP
@ -257,21 +257,21 @@ uniform. Unless disabled at build time, affinity for MPI tasks and
OpenMP threads on the host (CPU) will be set by default on the host
{when using offload to a coprocessor}. In this case, it is unnecessary
to use other methods to control affinity (e.g. taskset, numactl,
I_MPI_PIN_DOMAIN, etc.). This can be disabled with the {no_affinity}
option to the "package intel"_package.html command or by disabling the
option at build time (by adding -DINTEL_OFFLOAD_NOAFFINITY to the
CCFLAGS line of your Makefile). Disabling this option is not
recommended, especially when running on a machine with Intel
I_MPI_PIN_DOMAIN, etc.). This can be disabled with the {no_affinity}
option to the "package intel"_package.html command or by disabling the
option at build time (by adding -DINTEL_OFFLOAD_NOAFFINITY to the
CCFLAGS line of your Makefile). Disabling this option is not
recommended, especially when running on a machine with Intel
Hyper-Threading technology disabled.
[Run with the USER-INTEL package from the command line:]
To enable USER-INTEL optimizations for all available styles used in
the input script, the "-sf intel"
To enable USER-INTEL optimizations for all available styles used in
the input script, the "-sf intel"
"command-line switch"_Section_start.html#start_7 can be used without
any requirement for editing the input script. This switch will
automatically append "intel" to styles that support it. It also
invokes a default command: "package intel 1"_package.html. This
automatically append "intel" to styles that support it. It also
invokes a default command: "package intel 1"_package.html. This
package command is used to set options for the USER-INTEL package.
The default package command will specify that USER-INTEL calculations
are performed in mixed precision, that the number of OpenMP threads
@ -281,16 +281,16 @@ support, that 1 coprocessor per node will be used with automatic
balancing of work between the CPU and the coprocessor.
You can specify different options for the USER-INTEL package by using
the "-pk intel Nphi" "command-line switch"_Section_start.html#start_7
the "-pk intel Nphi" "command-line switch"_Section_start.html#start_7
with keyword/value pairs as specified in the documentation. Here,
Nphi = # of Xeon Phi coprocessors/node (ignored without offload
support). Common options to the USER-INTEL package include {omp} to
override any OMP_NUM_THREADS setting and specify the number of OpenMP
threads, {mode} to set the floating-point precision mode, and
{lrt} to enable Long-Range Thread mode as described below. See the
"package intel"_package.html command for details, including the
default values used for all its options if not specified, and how to
set the number of OpenMP threads via the OMP_NUM_THREADS environment
{lrt} to enable Long-Range Thread mode as described below. See the
"package intel"_package.html command for details, including the
default values used for all its options if not specified, and how to
set the number of OpenMP threads via the OMP_NUM_THREADS environment
variable if desired.
Examples (see documentation for your MPI/Machine for differences in
@ -303,7 +303,7 @@ mpirun -np 72 -ppn 36 lmp_machine -sf intel -in in.script -pk intel 0 omp 2 mode
As an alternative to adding command-line arguments, the input script
can be edited to enable the USER-INTEL package. This requires adding
the "package intel"_package.html command to the top of the input
the "package intel"_package.html command to the top of the input
script. For the second example above, this would be:
package intel 0 omp 2 mode double :pre
@ -314,46 +314,46 @@ add an "intel" suffix to the individual style, e.g.:
pair_style lj/cut/intel 2.5 :pre
Alternatively, the "suffix intel"_suffix.html command can be added to
the input script to enable USER-INTEL styles for the commands that
the input script to enable USER-INTEL styles for the commands that
follow in the input script.
[Tuning for Performance:]
NOTE: The USER-INTEL package will perform better with modifications
to the input script when "PPPM"_kspace_style.html is used:
"kspace_modify diff ad"_kspace_modify.html and "neigh_modify binsize
NOTE: The USER-INTEL package will perform better with modifications
to the input script when "PPPM"_kspace_style.html is used:
"kspace_modify diff ad"_kspace_modify.html and "neigh_modify binsize
3"_neigh_modify.html should be added to the input script.
Long-Range Thread (LRT) mode is an option to the "package
Long-Range Thread (LRT) mode is an option to the "package
intel"_package.html command that can improve performance when using
"PPPM"_kspace_style.html for long-range electrostatics on processors
with SMT. It generates an extra pthread for each MPI task. The thread
is dedicated to performing some of the PPPM calculations and MPI
with SMT. It generates an extra pthread for each MPI task. The thread
is dedicated to performing some of the PPPM calculations and MPI
communications. On Intel Xeon Phi x200 series CPUs, this will likely
always improve performance, even on a single node. On Intel Xeon
processors, using this mode might result in better performance when
using multiple nodes, depending on the machine. To use this mode,
specify that the number of OpenMP threads is one less than would
specify that the number of OpenMP threads is one less than would
normally be used for the run and add the "lrt yes" option to the "-pk"
command-line suffix or "package intel" command. For example, if a run
would normally perform best with "-pk intel 0 omp 4", instead use
"-pk intel 0 omp 3 lrt yes". When using LRT, you should set the
environment variable "KMP_AFFINITY=none". LRT mode is not supported
"-pk intel 0 omp 3 lrt yes". When using LRT, you should set the
environment variable "KMP_AFFINITY=none". LRT mode is not supported
when using offload.
Not all styles are supported in the USER-INTEL package. You can mix
the USER-INTEL package with styles from the "OPT"_accelerate_opt.html
package or the "USER-OMP package"_accelerate_omp.html". Of course,
the USER-INTEL package with styles from the "OPT"_accelerate_opt.html
package or the "USER-OMP package"_accelerate_omp.html". Of course,
this requires that these packages were installed at build time. This
can performed automatically by using "-sf hybrid intel opt" or
"-sf hybrid intel omp" command-line options. Alternatively, the "opt"
and "omp" suffixes can be appended manually in the input script. For
the latter, the "package omp"_package.html command must be in the
input script or the "-pk omp Nt" "command-line
switch"_Section_start.html#start_7 must be used where Nt is the
input script or the "-pk omp Nt" "command-line
switch"_Section_start.html#start_7 must be used where Nt is the
number of OpenMP threads. The number of OpenMP threads should not be
set differently for the different packages. Note that the "suffix
hybrid intel omp"_suffix.html command can also be used within the
set differently for the different packages. Note that the "suffix
hybrid intel omp"_suffix.html command can also be used within the
input script to automatically append the "omp" suffix to styles when
USER-INTEL styles are not available.
@ -374,33 +374,33 @@ that MPI runs are performed in MCDRAM.
[Tuning for Offload Performance:]
The default settings for offload should give good performance.
The default settings for offload should give good performance.
When using LAMMPS with offload to Intel coprocessors, best performance
will typically be achieved with concurrent calculations performed on
both the CPU and the coprocessor. This is achieved by offloading only
a fraction of the neighbor and pair computations to the coprocessor or
using "hybrid"_pair_hybrid.html pair styles where only one style uses
the "intel" suffix. For simulations with long-range electrostatics or
bond, angle, dihedral, improper calculations, computation and data
transfer to the coprocessor will run concurrently with computations
the "intel" suffix. For simulations with long-range electrostatics or
bond, angle, dihedral, improper calculations, computation and data
transfer to the coprocessor will run concurrently with computations
and MPI communications for these calculations on the host CPU. This
is illustrated in the figure below for the rhodopsin protein benchmark
running on E5-2697v2 processors with a Intel Xeon Phi 7120p
running on E5-2697v2 processors with a Intel Xeon Phi 7120p
coprocessor. In this plot, the vertical access is time and routines
running at the same time are running concurrently on both the host and
the coprocessor.
:c,image(JPG/offload_knc.png)
The fraction of the offloaded work is controlled by the {balance}
keyword in the "package intel"_package.html command. A balance of 0
runs all calculations on the CPU. A balance of 1 runs all
supported calculations on the coprocessor. A balance of 0.5 runs half
of the calculations on the coprocessor. Setting the balance to -1
(the default) will enable dynamic load balancing that continously
adjusts the fraction of offloaded work throughout the simulation.
Because data transfer cannot be timed, this option typically produces
The fraction of the offloaded work is controlled by the {balance}
keyword in the "package intel"_package.html command. A balance of 0
runs all calculations on the CPU. A balance of 1 runs all
supported calculations on the coprocessor. A balance of 0.5 runs half
of the calculations on the coprocessor. Setting the balance to -1
(the default) will enable dynamic load balancing that continously
adjusts the fraction of offloaded work throughout the simulation.
Because data transfer cannot be timed, this option typically produces
results within 5 to 10 percent of the optimal fixed balance.
If running short benchmark runs with dynamic load balancing, adding a
@ -418,15 +418,15 @@ with 60 cores available for offload and 4 hardware threads per core
each MPI task to use a subset of 10 threads on the coprocessor. Fine
tuning of the number of threads to use per MPI task or the number of
threads to use per core can be accomplished with keyword settings of
the "package intel"_package.html command.
the "package intel"_package.html command.
The USER-INTEL package has two modes for deciding which atoms will be
handled by the coprocessor. This choice is controlled with the {ghost}
keyword of the "package intel"_package.html command. When set to 0,
ghost atoms (atoms at the borders between MPI tasks) are not offloaded
to the card. This allows for overlap of MPI communication of forces
with computation on the coprocessor when the "newton"_newton.html
setting is "on". The default is dependent on the style being used,
The USER-INTEL package has two modes for deciding which atoms will be
handled by the coprocessor. This choice is controlled with the {ghost}
keyword of the "package intel"_package.html command. When set to 0,
ghost atoms (atoms at the borders between MPI tasks) are not offloaded
to the card. This allows for overlap of MPI communication of forces
with computation on the coprocessor when the "newton"_newton.html
setting is "on". The default is dependent on the style being used,
however, better performance may be achieved by setting this option
explictly.
@ -442,10 +442,10 @@ mode is being used and indicating the number of coprocessor threads
per MPI task. Additionally, an offload timing summary is printed at
the end of each run. When offloading, the frequency for "atom
sorting"_atom_modify.html is changed to 1 so that the per-atom data is
effectively sorted at every rebuild of the neighbor lists. All the
available coprocessor threads on each Phi will be divided among MPI
tasks, unless the {tptask} option of the "-pk intel" "command-line
switch"_Section_start.html#start_7 is used to limit the coprocessor
effectively sorted at every rebuild of the neighbor lists. All the
available coprocessor threads on each Phi will be divided among MPI
tasks, unless the {tptask} option of the "-pk intel" "command-line
switch"_Section_start.html#start_7 is used to limit the coprocessor
threads per MPI task.
[Restrictions:]

View File

@ -65,7 +65,7 @@ Make.py -v -p kokkos -kokkos omp -o mpi -a file mpi # or one-line build via Ma
mpirun -np 16 lmp_mpi -k on -sf kk -in in.lj # 1 node, 16 MPI tasks/node, no threads
mpirun -np 2 -ppn 1 lmp_mpi -k on t 16 -sf kk -in in.lj # 2 nodes, 1 MPI task/node, 16 threads/task
mpirun -np 2 lmp_mpi -k on t 8 -sf kk -in in.lj # 1 node, 2 MPI tasks/node, 8 threads/task
mpirun -np 2 lmp_mpi -k on t 8 -sf kk -in in.lj # 1 node, 2 MPI tasks/node, 8 threads/task
mpirun -np 32 -ppn 4 lmp_mpi -k on t 4 -sf kk -in in.lj # 8 nodes, 4 MPI tasks/node, 4 threads/task :pre
specify variables and settings in your Makefile.machine that enable OpenMP, GPU, or Phi support
@ -178,7 +178,7 @@ make kokkos_cuda_mpich :pre
These examples set the KOKKOS-specific OMP, MIC, CUDA variables on the
make command line which requires a GNU-compatible make command. Try
"gmake" if your system's standard make complains.
"gmake" if your system's standard make complains.
NOTE: If you build using make line variables and re-build LAMMPS twice
with different KOKKOS options and the *same* target, e.g. g++ in the
@ -394,7 +394,7 @@ additional parallelism (beyond MPI) will be invoked on the host
CPU(s).
You can compare the performance running in different modes:
run with 1 MPI task/node and N threads/task
run with N MPI tasks/node and 1 thread/task
run with settings in between these extremes :ul
@ -427,7 +427,7 @@ e.g. src/MAKE/Makefile.cuda, is correct for your GPU hardware/software
details).
The -np setting of the mpirun command should set the number of MPI
tasks/node to be equal to the # of physical GPUs on the node.
tasks/node to be equal to the # of physical GPUs on the node.
Use the "-k" "command-line switch"_Section_commands.html#start_7 to
specify the number of GPUs per node, and the number of threads per MPI

View File

@ -96,7 +96,7 @@ variable.
Depending on which styles are accelerated, you should look for a
reduction in the "Pair time", "Bond time", "KSpace time", and "Loop
time" values printed at the end of a run.
time" values printed at the end of a run.
You may see a small performance advantage (5 to 20%) when running a
USER-OMP style (in serial or parallel) with a single thread per MPI

View File

@ -21,11 +21,11 @@ angle_coeff 6 2.1 180.0 :pre
[Description:]
The {dipole} angle style is used to control the orientation of a dipolar
atom within a molecule "(Orsi)"_#Orsi. Specifically, the {dipole} angle
style restrains the orientation of a point dipole mu_j (embedded in atom
'j') with respect to a reference (bond) vector r_ij = r_i - r_j, where 'i'
is another atom of the same molecule (typically, 'i' and 'j' are also
covalently bonded).
atom within a molecule "(Orsi)"_#Orsi. Specifically, the {dipole} angle
style restrains the orientation of a point dipole mu_j (embedded in atom
'j') with respect to a reference (bond) vector r_ij = r_i - r_j, where 'i'
is another atom of the same molecule (typically, 'i' and 'j' are also
covalently bonded).
It is convenient to define an angle gamma between the 'free' vector mu_j
and the reference (bond) vector r_ij:
@ -37,21 +37,21 @@ The {dipole} angle style uses the potential:
:c,image(Eqs/angle_dipole_potential.jpg)
where K is a rigidity constant and gamma0 is an equilibrium (reference)
angle.
angle.
The torque on the dipole can be obtained by differentiating the
potential using the 'chain rule' as in appendix C.3 of
The torque on the dipole can be obtained by differentiating the
potential using the 'chain rule' as in appendix C.3 of
"(Allen)"_#Allen:
:c,image(Eqs/angle_dipole_torque.jpg)
Example: if gamma0 is set to 0 degrees, the torque generated by
the potential will tend to align the dipole along the reference
the potential will tend to align the dipole along the reference
direction defined by the (bond) vector r_ij (in other words, mu_j is
restrained to point towards atom 'i').
The dipolar torque T_j must be counterbalanced in order to conserve
the local angular momentum. This is achieved via an additional force
The dipolar torque T_j must be counterbalanced in order to conserve
the local angular momentum. This is achieved via an additional force
couple generating a torque equivalent to the opposite of T_j:
:c,image(Eqs/angle_dipole_couple.jpg)
@ -118,7 +118,7 @@ This angle style should not be used with SHAKE.
:line
:link(Orsi)
[(Orsi)] Orsi & Essex, The ELBA force field for coarse-grain modeling of
[(Orsi)] Orsi & Essex, The ELBA force field for coarse-grain modeling of
lipid membranes, PloS ONE 6(12): e28637, 2011.
:link(Allen)

View File

@ -62,7 +62,7 @@ more instructions on how to use the accelerated styles effectively.
[Restrictions:]
This angle style can only be used if LAMMPS was built with the
USER_MISC package. See the "Making LAMMPS"_Section_start.html#start_3
USER_MISC package. See the "Making LAMMPS"_Section_start.html#start_3
section for more info on packages.
[Related commands:]

View File

@ -61,7 +61,7 @@ more instructions on how to use the accelerated styles effectively.
[Restrictions:]
This angle style can only be used if LAMMPS was built with the
USER_MISC package. See the "Making LAMMPS"_Section_start.html#start_3
USER_MISC package. See the "Making LAMMPS"_Section_start.html#start_3
section for more info on packages.
[Related commands:]

View File

@ -68,7 +68,7 @@ more instructions on how to use the accelerated styles effectively.
[Restrictions:]
This angle style can only be used if LAMMPS was built with the
USER_MISC package. See the "Making LAMMPS"_Section_start.html#start_3
USER_MISC package. See the "Making LAMMPS"_Section_start.html#start_3
section for more info on packages.
[Related commands:]

View File

@ -43,7 +43,7 @@ internally; hence the units of K are in energy/radian^2.
The also required {lj/sdk} parameters will be extracted automatically
from the pair_style.
[Restrictions:]
[Restrictions:]
This angle style can only be used if LAMMPS was built with the
USER-CG-CMM package. See the "Making

View File

@ -1,4 +1,4 @@
"LAMMPS WWW Site"_lws - "LAMMPS Documentation"_ld - "LAMMPS
"LAMMPS WWW Site"_lws - "LAMMPS Documentation"_ld - "LAMMPS
Commands"_lc :c
:link(lws,http://lammps.sandia.gov)
@ -156,12 +156,12 @@ used with a group-ID that is not "all".
[Default:]
By default, {id} is yes. By default, atomic systems (no bond topology
info) do not use a map. For molecular systems (with bond topology
info), a map is used. The default map style is array if no atom ID is
larger than 1 million, otherwise the default is hash. By default, a
"first" group is not defined. By default, sorting is enabled with a
frequency of 1000 and a binsize of 0.0, which means the neighbor
By default, {id} is yes. By default, atomic systems (no bond topology
info) do not use a map. For molecular systems (with bond topology
info), a map is used. The default map style is array if no atom ID is
larger than 1 million, otherwise the default is hash. By default, a
"first" group is not defined. By default, sorting is enabled with a
frequency of 1000 and a binsize of 0.0, which means the neighbor
cutoff will be used to set the bin size.
:line

View File

@ -14,7 +14,7 @@ atom_style style args :pre
style = {angle} or {atomic} or {body} or {bond} or {charge} or {dipole} or \
{dpd} or {electron} or {ellipsoid} or {full} or {line} or {meso} or \
{molecular} or {peri} or {smd} or {sphere} or {tri} or \
{molecular} or {peri} or {smd} or {sphere} or {tri} or \
{template} or {hybrid} :ulb,l
args = none for any style except the following
{body} args = bstyle bstyle-args
@ -193,7 +193,7 @@ For the {body} style, the particles are arbitrary bodies with internal
attributes defined by the "style" of the bodies, which is specified by
the {bstyle} argument. Body particles can represent complex entities,
such as surface meshes of discrete points, collections of
sub-particles, deformable objects, etc.
sub-particles, deformable objects, etc.
The "body"_body.html doc page descibes the body styles LAMMPS
currently supports, and provides more details as to the kind of body
@ -269,7 +269,7 @@ The {line} and {tri} styles are part of the ASPHERE package.
The {body} style is part of the BODY package.
The {dipole} style is part of the DIPOLE package.
The {dipole} style is part of the DIPOLE package.
The {peri} style is part of the PERI package for Peridynamics.

View File

@ -10,7 +10,7 @@ balance command :h3
[Syntax:]
balance thresh style args ... keyword value ... :pre
balance thresh style args ... keyword args ... :pre
thresh = imbalance threshhold that must be exceeded to perform a re-balance :ulb,l
one style/arg pair can be used (or multiple for {x},{y},{z}) :l
@ -32,9 +32,23 @@ style = {x} or {y} or {z} or {shift} or {rcb} :l
Niter = # of times to iterate within each dimension of dimstr sequence
stopthresh = stop balancing when this imbalance threshhold is reached
{rcb} args = none :pre
zero or more keyword/value pairs may be appended :l
keyword = {out} :l
{out} value = filename
zero or more keyword/arg pairs may be appended :l
keyword = {weight} or {out} :l
{weight} style args = use weighted particle counts for the balancing
{style} = {group} or {neigh} or {time} or {var} or {store}
{group} args = Ngroup group1 weight1 group2 weight2 ...
Ngroup = number of groups with assigned weights
group1, group2, ... = group IDs
weight1, weight2, ... = corresponding weight factors
{neigh} factor = compute weight based on number of neighbors
factor = scaling factor (> 0)
{time} factor = compute weight based on time spend computing
factor = scaling factor (> 0)
{var} name = take weight from atom-style variable
name = name of the atom-style variable
{store} name = store weight in custom atom property defined by "fix property/atom"_fix_property_atom.html command
name = atom property name (without d_ prefix)
{out} arg = filename
filename = write each processor's sub-domain to a file :pre
:ule
@ -44,28 +58,42 @@ balance 0.9 x uniform y 0.4 0.5 0.6
balance 1.2 shift xz 5 1.1
balance 1.0 shift xz 5 1.1
balance 1.1 rcb
balance 1.0 shift x 10 1.1 weight group 2 fast 0.5 slow 2.0
balance 1.0 shift x 10 1.1 weight time 0.8 weight neigh 0.5 weight store balance
balance 1.0 shift x 20 1.0 out tmp.balance :pre
[Description:]
This command adjusts the size and shape of processor sub-domains
within the simulation box, to attempt to balance the number of
particles and thus the computational cost (load) evenly across
processors. The load balancing is "static" in the sense that this
command performs the balancing once, before or between simulations.
The processor sub-domains will then remain static during the
subsequent run. To perform "dynamic" balancing, see the "fix
within the simulation box, to attempt to balance the number of atoms
or particles and thus indirectly the computational cost (load) more
evenly across processors. The load balancing is "static" in the sense
that this command performs the balancing once, before or between
simulations. The processor sub-domains will then remain static during
the subsequent run. To perform "dynamic" balancing, see the "fix
balance"_fix_balance.html command, which can adjust processor
sub-domain sizes and shapes on-the-fly during a "run"_run.html.
Load-balancing is typically only useful if the particles in the
simulation box have a spatially-varying density distribution. E.g. a
model of a vapor/liquid interface, or a solid with an irregular-shaped
geometry containing void regions. In this case, the LAMMPS default of
Load-balancing is typically most useful if the particles in the
simulation box have a spatially-varying density distribution or when
the computational cost varies signficantly between different
particles. E.g. a model of a vapor/liquid interface, or a solid with
an irregular-shaped geometry containing void regions, or "hybrid pair
style simulations"_pair_hybrid.html which combine pair styles with
different computational cost. In these cases, the LAMMPS default of
dividing the simulation box volume into a regular-spaced grid of 3d
bricks, with one equal-volume sub-domain per procesor, may assign very
different numbers of particles per processor. This can lead to poor
performance when the simulation is run in parallel.
bricks, with one equal-volume sub-domain per procesor, may assign
numbers of particles per processor in a way that the computational
effort varies significantly. This can lead to poor performance when
the simulation is run in parallel.
The balancing can be performed with or without per-particle weighting.
With no weighting, the balancing attempts to assign an equal number of
particles to each processor. With weighting, the balancing attempts
to assign an equal aggregate computational weight to each processor,
which typically inducces a diffrent number of atoms assigned to each
processor. Details on the various weighting options and examples for
how they can be used are "given below"_#weighted_balance.
Note that the "processors"_processors.html command allows some control
over how the box volume is split across processors. Specifically, for
@ -78,9 +106,9 @@ sub-domains will still have the same shape and same volume.
The requested load-balancing operation is only performed if the
current "imbalance factor" in particles owned by each processor
exceeds the specified {thresh} parameter. The imbalance factor is
defined as the maximum number of particles owned by any processor,
divided by the average number of particles per processor. Thus an
imbalance factor of 1.0 is perfect balance.
defined as the maximum number of particles (or weight) owned by any
processor, divided by the average number of particles (or weight) per
processor. Thus an imbalance factor of 1.0 is perfect balance.
As an example, for 10000 particles running on 10 processors, if the
most heavily loaded processor has 1200 particles, then the factor is
@ -108,7 +136,7 @@ defined above. But depending on the method a perfect balance (1.0)
may not be achieved. For example, "grid" methods (defined below) that
create a logical 3d grid cannot achieve perfect balance for many
irregular distributions of particles. Likewise, if a portion of the
system is a perfect lattice, e.g. the intiial system is generated by
system is a perfect lattice, e.g. the initial system is generated by
the "create_atoms"_create_atoms.html command, then "grid" methods may
be unable to achieve exact balance. This is because entire lattice
planes will be owned or not owned by a single processor.
@ -134,11 +162,11 @@ The {x}, {y}, {z}, and {shift} styles are "grid" methods which produce
a logical 3d grid of processors. They operate by changing the cutting
planes (or lines) between processors in 3d (or 2d), to adjust the
volume (area in 2d) assigned to each processor, as in the following 2d
diagram where processor sub-domains are shown and atoms are colored by
the processor that owns them. The leftmost diagram is the default
partitioning of the simulation box across processors (one sub-box for
each of 16 processors); the middle diagram is after a "grid" method
has been applied.
diagram where processor sub-domains are shown and particles are
colored by the processor that owns them. The leftmost diagram is the
default partitioning of the simulation box across processors (one
sub-box for each of 16 processors); the middle diagram is after a
"grid" method has been applied.
:image(JPG/balance_uniform_small.jpg,JPG/balance_uniform.jpg),image(JPG/balance_nonuniform_small.jpg,JPG/balance_nonuniform.jpg),image(JPG/balance_rcb_small.jpg,JPG/balance_rcb.jpg)
:c
@ -146,8 +174,8 @@ has been applied.
The {rcb} style is a "tiling" method which does not produce a logical
3d grid of processors. Rather it tiles the simulation domain with
rectangular sub-boxes of varying size and shape in an irregular
fashion so as to have equal numbers of particles in each sub-box, as
in the rightmost diagram above.
fashion so as to have equal numbers of particles (or weight) in each
sub-box, as in the rightmost diagram above.
The "grid" methods can be used with either of the
"comm_style"_comm_style.html command options, {brick} or {tiled}. The
@ -230,7 +258,7 @@ counts do not match the target value for the plane, the position of
the cut is adjusted to be halfway between a low and high bound. The
low and high bounds are adjusted on each iteration, using new count
information, so that they become closer together over time. Thus as
the recustion progresses, the count of particles on either side of the
the recursion progresses, the count of particles on either side of the
plane gets closer to the target value.
Once the rebalancing is complete and final processor sub-domains
@ -262,21 +290,155 @@ the longest dimension, leaving one new box on either side of the cut.
All the processors are also partitioned into 2 groups, half assigned
to the box on the lower side of the cut, and half to the box on the
upper side. (If the processor count is odd, one side gets an extra
processor.) The cut is positioned so that the number of atoms in the
lower box is exactly the number that the processors assigned to that
box should own for load balance to be perfect. This also makes load
balance for the upper box perfect. The positioning is done
iteratively, by a bisectioning method. Note that counting atoms on
either side of the cut requires communication between all processors
at each iteration.
processor.) The cut is positioned so that the number of particles in
the lower box is exactly the number that the processors assigned to
that box should own for load balance to be perfect. This also makes
load balance for the upper box perfect. The positioning is done
iteratively, by a bisectioning method. Note that counting particles
on either side of the cut requires communication between all
processors at each iteration.
That is the procedure for the first cut. Subsequent cuts are made
recursively, in exactly the same manner. The subset of processors
assigned to each box make a new cut in the longest dimension of that
box, splitting the box, the subset of processsors, and the atoms in
the box in two. The recursion continues until every processor is
assigned a sub-box of the entire simulation domain, and owns the atoms
in that sub-box.
box, splitting the box, the subset of processsors, and the particles
in the box in two. The recursion continues until every processor is
assigned a sub-box of the entire simulation domain, and owns the
particles in that sub-box.
:line
This sub-section describes how to perform weighted load balancing
using the {weight} keyword. :link(weighted_balance)
By default, all particles have a weight of 1.0, which means each
particle is assumed to require the same amount of computation during a
timestep. There are, however, scenarios where this is not a good
assumption. Measuring the computational cost for each particle
accurately would be impractical and slow down the computation.
Instead the {weight} keyword implements several ways to influence the
per-particle weights empirically by properties readily available or
using the user's knowledge of the system. Note that the absolute
value of the weights are not important; only their relative ratios
affect which particle is assigned to which processor. A particle with
a weight of 2.5 is assumed to require 5x more computational than a
particle with a weight of 0.5. For all the options below the weight
assigned to a particle must be a positive value; an error will be be
generated if a weight is <= 0.0.
Below is a list of possible weight options with a short description of
their usage and some example scenarios where they might be applicable.
It is possible to apply multiple weight flags and the weightings they
induce will be combined through multiplication. Most of the time,
however, it is sufficient to use just one method.
The {group} weight style assigns weight factors to specified
"groups"_group.html of particles. The {group} style keyword is
followed by the number of groups, then pairs of group IDs and the
corresponding weight factor. If a particle belongs to none of the
specified groups, its weight is not changed. If it belongs to
multiple groups, its weight is the product of the weight factors.
This weight style is useful in combination with pair style
"hybrid"_pair_hybrid.html, e.g. when combining a more costly manybody
potential with a fast pair-wise potential. It is also useful when
using "run_style respa"_run_style.html where some portions of the
system have many bonded interactions and others none. It assumes that
the computational cost for each group remains constant over time.
This is a purely empirical weighting, so a series test runs to tune
the assigned weight factors for optimal performance is recommended.
The {neigh} weight style assigns the same weight to each particle
owned by a processor based on the total count of neighbors in the
neighbor list owned by that processor. The motivation is that more
neighbors means a higher computational cost. The style does not use
neighbors per atom to assign a unique weight to each atom, because
that value can vary depending on how the neighbor list is built.
The {factor} setting is applied as an overall scale factor to the
{neigh} weights which allows adjustment of their impact on the
balancing operation. The specified {factor} value must be positive.
A value > 1.0 will increase the weights so that the ratio of max
weight to min weight increases by {factor}. A value < 1.0 will
decrease the weights so that the ratio of max weight to min weight
decreases by {factor}. In both cases the intermediate weight values
increase/decrease proportionally as well. A value = 1.0 has no effect
on the {neigh} weights. As a rule of thumb, we have found a {factor}
of about 0.8 often results in the best performance, since the number
of neighbors is likely to overestimate the ideal weight.
This weight style is useful for systems where there are different
cutoffs used for different pairs of interations, or the density
fluctuates, or a large number of particles are in the vicinity of a
wall, or a combination of these effects. If a simulation uses
multiple neighbor lists, this weight style will use the first suitable
neighbor list it finds. It will not request or compute a new list. A
warning will be issued if there is no suitable neighbor list available
or if it is not current, e.g. if the balance command is used before a
"run"_run.html or "minimize"_minimize.html command is used, in which
case the neighbor list may not yet have been built. In this case no
weights are computed. Inserting a "run 0 post no"_run.html command
before issuing the {balance} command, may be a workaround for this
case, as it will induce the neighbor list to be built.
The {time} weight style uses "timer data"_timer.html to estimate
weights. It assigns the same weight to each particle owned by a
processor based on the total computational time spent by that
processor. See details below on what time window is used. It uses
the same timing information as is used for the "MPI task timing
breakdown"_Section_start.html#start_8, namely, for sections {Pair},
{Bond}, {Kspace}, and {Neigh}. The time spent in those portions of
the timestep are measured for each MPI rank, summed, then divided by
the number of particles owned by that processor. I.e. the weight is
an effective CPU time/particle averaged over the particles on that
processor.
The {factor} setting is applied as an overall scale factor to the
{time} weights which allows adjustment of their impact on the
balancing operation. The specified {factor} value must be positive.
A value > 1.0 will increase the weights so that the ratio of max
weight to min weight increases by {factor}. A value < 1.0 will
decrease the weights so that the ratio of max weight to min weight
decreases by {factor}. In both cases the intermediate weight values
increase/decrease proportionally as well. A value = 1.0 has no effect
on the {time} weights. As a rule of thumb, effective values to use
are typicall between 0.5 and 1.2. Note that the timer quantities
mentioned above can be affected by communication which occurs in the
middle of the operations, e.g. pair styles with intermediate exchange
of data witin the force computation, and likewise for KSpace solves.
When using the {time} weight style with the {balance} command, the
timing data is taken from the preceding run command, i.e. the timings
are for the entire previous run. For the {fix balance} command the
timing data is for only the timesteps since the last balancing
operation was performed. If timing information for the required
sections is not available, e.g. at the beginning of a run, or when the
"timer"_timer.html command is set to either {loop} or {off}, a warning
is issued. In this case no weights are computed.
NOTE: The {time} weight style is the most generic option, and should
be tried first, unless the {group} style is easily applicable.
However, since the computed cost function is averaged over all
particles on a processor, the weights may not be highly accurate.
This style can also be effective as a secondary weight in combination
with either {group} or {neigh} to offset some of inaccuracies in
either of those heuristics.
The {var} weight style assigns per-particle weights by evaluating an
"atom-style variable"_variable.html specified by {name}. This is
provided as a more flexible alternative to the {group} weight style,
allowing definition of a more complex heuristics based on information
(global and per atom) available inside of LAMMPS. For example,
atom-style variables can reference the position of a particle, its
velocity, the volume of its Voronoi cell, etc.
The {store} weight style does not compute a weight factor. Instead it
stores the current accumulated weights in a custom per-atom property
specified by {name}. This must be a property defined as {d_name} via
the "fix property/atom"_fix_property_atom.html command. Note that
these custom per-atom properties can be output in a "dump"_dump.html
file, so this is a way to examine, debug, or visualize the
per-particle weights computed during the load-balancing operation.
:line
@ -328,7 +490,7 @@ per processor. Note that the 4 sub-domains share vertices, so there
will be duplicate nodes in the list.
The "SQUARES" section lists the node IDs of the 4 vertices in a
rectangle for each processor (1 to 4).
rectangle for each processor (1 to 4).
For a 3d problem, the syntax is similar with 8 vertices listed for
each processor, instead of 4, and "SQUARES" replaced by "CUBES".
@ -342,6 +504,7 @@ appear in {dimstr} for the {shift} style.
[Related commands:]
"processors"_processors.html, "fix balance"_fix_balance.html
"group"_group.html, "processors"_processors.html,
"fix balance"_fix_balance.html
[Default:] none

View File

@ -125,7 +125,7 @@ in the {Bodies} section of the data file:
atom-ID 1 M
N
ixx iyy izz ixy ixz iyz
ixx iyy izz ixy ixz iyz
x1 y1 z1
...
xN yN zN :pre
@ -198,11 +198,11 @@ in the {Bodies} section of the data file:
atom-ID 1 M
N
ixx iyy izz ixy ixz iyz
ixx iyy izz ixy ixz iyz
x1 y1 z1
...
xN yN zN
i j j k k ...
i j j k k ...
radius :pre
N is the number of vertices in the body particle. M = 6 + 3*N + 2*N +
@ -230,11 +230,11 @@ particles whose edge length is sqrt(2):
3 1 27
4
1 1 4 0 0 0
-0.7071 -0.7071 0
-0.7071 0.7071 0
0.7071 0.7071 0
0.7071 -0.7071 0
1 1 4 0 0 0
-0.7071 -0.7071 0
-0.7071 0.7071 0
0.7071 0.7071 0
0.7071 -0.7071 0
0 1 1 2 2 3 3 0
1.0 :pre

View File

@ -173,7 +173,7 @@ change_box all x scale 1.1 y volume z volume :pre
The {volume} style changes the associated dimension so that the
overall box volume is unchanged relative to its value before the
preceding keyword was invoked.
preceding keyword was invoked.
If the following command is used, then the z box length will shrink by
the same 1.1 factor the x box length was increased by:

View File

@ -135,7 +135,7 @@ and angular momentum of a particle. If the {vel} option is set to
{yes}, then ghost atoms store these quantities; if {no} then they do
not. The {yes} setting is needed by some pair styles which require
the velocity state of both the I and J particles to compute a pairwise
I,J interaction.
I,J interaction, as well as by some compute and fix commands.
Note that if the "fix deform"_fix_deform.html command is being used
with its "remap v" option enabled, then the velocities for ghost atoms

View File

@ -13,7 +13,7 @@ compute cna/atom command :h3
compute ID group-ID cna/atom cutoff :pre
ID, group-ID are documented in "compute"_compute.html command
cna/atom = style name of this compute command
cna/atom = style name of this compute command
cutoff = cutoff distance for nearest neighbors (distance units) :ul
[Examples:]

View File

@ -63,4 +63,4 @@ LAMMPS"_Section_start.html#start_3 section for more info.
"compute damage/atom"_compute_damage_atom.html,
"compute plasticity/atom"_compute_plasticity_atom.html
[Default:] none
[Default:] none

View File

@ -19,7 +19,7 @@ charge-correction = {mass} or {geometry}, use COM or geometric center for charge
[Examples:]
compute 1 fluid dipole/chunk molchunk
compute 1 fluid dipole/chunk molchunk
compute dw water dipole/chunk 1 geometry :pre
[Description:]

View File

@ -46,7 +46,7 @@ output options.
The vector values will be in energy and temperature "units"_units.html.
[Restrictions:]
[Restrictions:]
This command is part of the USER-DPD package. It is only enabled if
LAMMPS was built with that package. See the "Making
@ -64,7 +64,7 @@ command.
:line
:link(Larentzos)
:link(Larentzos)
[(Larentzos)] J.P. Larentzos, J.K. Brennan, J.D. Moore, and
W.D. Mattson, "LAMMPS Implementation of Constant Energy Dissipative
Particle Dynamics (DPD-E)", ARL-TR-6863, U.S. Army Research

View File

@ -22,7 +22,7 @@ compute 1 all dpd/atom
[Description:]
Define a computation that accesses the per-particle internal
conductive energy (u_cond), internal mechanical energy (u_mech),
conductive energy (u_cond), internal mechanical energy (u_mech),
internal chemical energy (u_chem) and
internal temperatures (dpdTheta) for each particle in a group. See
the "compute dpd"_compute_dpd.html command if you want the total
@ -39,10 +39,10 @@ that uses per-particle values from a compute as input. See
"Section 6.15"_Section_howto.html#howto_15 for an overview of
LAMMPS output options.
The per-particle array values will be in energy (u_cond, u_mech, u_chem)
The per-particle array values will be in energy (u_cond, u_mech, u_chem)
and temperature (dpdTheta) "units"_units.html.
[Restrictions:]
[Restrictions:]
This command is part of the USER-DPD package. It is only enabled if
LAMMPS was built with that package. See the "Making

View File

@ -26,7 +26,7 @@ Define a computation that flags an "event" if any particle in the
group has moved a distance greater than the specified threshold
distance when compared to a previously stored reference state
(i.e. the previous event). This compute is typically used in
conjunction with the "prd"_prd.html and "tad"_tad.html commands,
conjunction with the "prd"_prd.html and "tad"_tad.html commands,
to detect if a transition
to a new minimum energy basin has occurred.
@ -34,8 +34,8 @@ This value calculated by the compute is equal to 0 if no particle has
moved far enough, and equal to 1 if one or more particles have moved
further than the threshold distance.
NOTE: If the system is undergoing significant center-of-mass motion,
due to thermal motion, an external force, or an initial net momentum,
NOTE: If the system is undergoing significant center-of-mass motion,
due to thermal motion, an external force, or an initial net momentum,
then this compute will not be able to distinguish that motion from
local atom displacements and may generate "false postives."

View File

@ -64,7 +64,7 @@ these atoms:
A coupling parameter \(\lambda\) varying from 0 to 1 connects the
reference and perturbed systems:
:c,image(Eqs/compute_fep_lambda.jpg)
:c,image(Eqs/compute_fep_lambda.jpg)
It is possible but not necessary that the coupling parameter (or a
function thereof) appears as a multiplication factor of the potential

View File

@ -28,7 +28,7 @@ compute 2 molecule gyration/chunk molchunk tensor :pre
[Description:]
Define a computation that calculates the radius of gyration Rg for
multiple chunks of atoms.
multiple chunks of atoms.
In LAMMPS, chunks are collections of atoms defined by a "compute
chunk/atom"_compute_chunk_atom.html command, which assigns each atom

View File

@ -20,7 +20,7 @@ stress-ID = ID of a compute that calculates per-atom stress :ul
[Examples:]
compute myFlux all heat/flux myKE myPE myStress :pre
compute myFlux all heat/flux myKE myPE myStress :pre
[Description:]
@ -38,7 +38,7 @@ subtracted to a group of atoms.
The compute takes three arguments which are IDs of other
"computes"_compute.html. One calculates per-atom kinetic energy
({ke-ID}), one calculates per-atom potential energy ({pe-ID)}, and the
third calcualtes per-atom stress ({stress-ID}).
third calcualtes per-atom stress ({stress-ID}).
NOTE: These other computes should provide values for all the atoms in
the group this compute specifies. That means the other computes could
@ -152,11 +152,11 @@ lattice fcc 5.376 orient x 1 0 0 orient y 0 1 0 orient z 0 0 1
region box block 0 4 0 4 0 4
create_box 1 box
create_atoms 1 box
mass 1 39.948
mass 1 39.948
pair_style lj/cut 13.0
pair_coeff * * 0.2381 3.405
timestep $\{dt\}
thermo $d :pre
thermo $d :pre
# equilibration and thermalization :pre

View File

@ -15,7 +15,7 @@ compute ID group-ID hexorder/atom keyword values ... :pre
ID, group-ID are documented in "compute"_compute.html command :ulb,l
hexorder/atom = style name of this compute command :l
one or more keyword/value pairs may be appended :l
keyword = {degree} or {nnn} or {cutoff}
keyword = {degree} or {nnn} or {cutoff}
{cutoff} value = distance cutoff
{nnn} value = number of nearest neighbors
{degree} value = degree {n} of order parameter :pre
@ -24,27 +24,27 @@ keyword = {degree} or {nnn} or {cutoff}
[Examples:]
compute 1 all hexorder/atom
compute 1 all hexorder/atom
compute 1 all hexorder/atom degree 4 nnn 4 cutoff 1.2 :pre
[Description:]
Define a computation that calculates {qn} the bond-orientational
order parameter for each atom in a group. The hexatic ({n} = 6) order
Define a computation that calculates {qn} the bond-orientational
order parameter for each atom in a group. The hexatic ({n} = 6) order
parameter was introduced by "Nelson and Halperin"_#Nelson as a way to detect
hexagonal symmetry in two-dimensional systems. For each atom, {qn}
hexagonal symmetry in two-dimensional systems. For each atom, {qn}
is a complex number (stored as two real numbers) defined as follows:
:c,image(Eqs/hexorder.jpg)
where the sum is over the {nnn} nearest neighbors
where the sum is over the {nnn} nearest neighbors
of the central atom. The angle theta
is formed by the bond vector rij and the {x} axis. theta is calculated
only using the {x} and {y} components, whereas the distance from the
central atom is calculated using all three
central atom is calculated using all three
{x}, {y}, and {z} components of the bond vector.
Neighbor atoms not in the group
are included in the order parameter of atoms in the group.
Neighbor atoms not in the group
are included in the order parameter of atoms in the group.
The optional keyword {cutoff} defines the distance cutoff
used when searching for neighbors. The default value, also
@ -53,22 +53,22 @@ by the pair style.
The optional keyword {nnn} defines the number of nearest
neighbors used to calculate {qn}. The default value is 6.
If the value is NULL, then all neighbors up to the
If the value is NULL, then all neighbors up to the
distance cutoff are used.
The optional keyword {degree} sets the degree {n} of the order parameter.
The default value is 6. For a perfect hexagonal lattice with
The optional keyword {degree} sets the degree {n} of the order parameter.
The default value is 6. For a perfect hexagonal lattice with
{nnn} = 6,
{q}6 = exp(6 i phi) for all atoms, where the constant 0 < phi < pi/3
depends only on the orientation of the lattice relative to the {x} axis.
In an isotropic liquid, local neighborhoods may still exhibit
{q}6 = exp(6 i phi) for all atoms, where the constant 0 < phi < pi/3
depends only on the orientation of the lattice relative to the {x} axis.
In an isotropic liquid, local neighborhoods may still exhibit
weak hexagonal symmetry, but because the orientational correlation
decays quickly with distance, the value of phi will be different for
different atoms, and so when {q}6 is averaged over all the atoms
different atoms, and so when {q}6 is averaged over all the atoms
in the system, \|<{q}6>\| << 1.
The value of {qn} is set to zero for atoms not in the
specified compute group, as well as for atoms that have less than
specified compute group, as well as for atoms that have less than
{nnn} neighbors within the distance cutoff.
The neighbor list needed to compute this quantity is constructed each
@ -92,7 +92,7 @@ the neighbor list.
[Output info:]
This compute calculates a per-atom array with 2 columns, giving the
real and imaginary parts {qn}, a complex number restricted to the
real and imaginary parts {qn}, a complex number restricted to the
unit disk of the complex plane i.e. Re({qn})^2 + Im({qn})^2 <= 1 .
These values can be accessed by any command that uses
@ -106,7 +106,7 @@ options.
"compute orientorder/atom"_compute_orientorder_atom.html, "compute coord/atom"_compute_coord_atom.html, "compute centro/atom"_compute_centro_atom.html
[Default:]
[Default:]
The option defaults are {cutoff} = pair style cutoff, {nnn} = 6, {degree} = 6

View File

@ -23,7 +23,7 @@ compute 1 fluid inertia/chunk molchunk :pre
[Description:]
Define a computation that calculates the inertia tensor for multiple
chunks of atoms.
chunks of atoms.
In LAMMPS, chunks are collections of atoms defined by a "compute
chunk/atom"_compute_chunk_atom.html command, which assigns each atom

View File

@ -48,9 +48,9 @@ thermodynamic output by using the "thermo_modify"_thermo_modify.html
command, as shown in the following example:
compute effTemp all temp/eff
thermo_style custom step etotal pe ke temp press
thermo_style custom step etotal pe ke temp press
thermo_modify temp effTemp :pre
The value of the kinetic energy will be 0.0 for atoms (nuclei or
electrons) not in the specified compute group.

View File

@ -52,9 +52,9 @@ printed with thermodynamic output by using the
example:
compute effTemp all temp/eff
thermo_style custom step etotal pe ke temp press
thermo_style custom step etotal pe ke temp press
thermo_modify temp effTemp :pre
See "compute temp/eff"_compute_temp_eff.html.
[Output info:]

View File

@ -61,4 +61,4 @@ the temperature is correctly normalized.
[Default:]
The option defaults are extra = 2 or 3 for 2d or 3d systems and
dynamic = no.
dynamic = no.

View File

@ -44,7 +44,7 @@ proportional to the diffusion coefficient of the diffusing atoms.
The displacement of an atom is from its reference position. This is
normally the original position at the time
the compute command was issued, unless the {average} keyword is set to {yes}.
the compute command was issued, unless the {average} keyword is set to {yes}.
The value of the displacement will be
0.0 for atoms not in the specified compute group.

View File

@ -15,7 +15,7 @@ compute ID group-ID orientorder/atom keyword values ... :pre
ID, group-ID are documented in "compute"_compute.html command :ulb,l
orientorder/atom = style name of this compute command :l
one or more keyword/value pairs may be appended :l
keyword = {cutoff} or {nnn} or {ql}
keyword = {cutoff} or {nnn} or {ql}
{cutoff} value = distance cutoff
{nnn} value = number of nearest neighbors
{degrees} values = nlvalues, l1, l2,... :pre
@ -24,30 +24,30 @@ keyword = {cutoff} or {nnn} or {ql}
[Examples:]
compute 1 all orientorder/atom
compute 1 all orientorder/atom
compute 1 all orientorder/atom degrees 5 4 6 8 10 12 nnn NULL cutoff 1.5 :pre
[Description:]
Define a computation that calculates a set of bond-orientational
Define a computation that calculates a set of bond-orientational
order parameters {Ql} for each atom in a group. These order parameters
were introduced by "Steinhardt et al."_#Steinhardt as a way to
characterize the local orientational order in atomic structures.
characterize the local orientational order in atomic structures.
For each atom, {Ql} is a real number defined as follows:
:c,image(Eqs/orientorder.jpg)
The first equation defines the spherical harmonic order parameters.
These are complex number components of the 3D analog of the 2D order
parameter {qn}, which is implemented as LAMMPS compute
"hexorder/atom"_compute_hexorder_atom.html.
The summation is over the {nnn} nearest
neighbors of the central atom.
The angles theta and phi are the standard spherical polar angles
The first equation defines the spherical harmonic order parameters.
These are complex number components of the 3D analog of the 2D order
parameter {qn}, which is implemented as LAMMPS compute
"hexorder/atom"_compute_hexorder_atom.html.
The summation is over the {nnn} nearest
neighbors of the central atom.
The angles theta and phi are the standard spherical polar angles
defining the direction of the bond vector {rij}.
The second equation defines {Ql}, which is a
rotationally invariant scalar quantity obtained by summing
over all the components of degree {l}.
rotationally invariant scalar quantity obtained by summing
over all the components of degree {l}.
The optional keyword {cutoff} defines the distance cutoff
used when searching for neighbors. The default value, also
@ -56,23 +56,23 @@ by the pair style.
The optional keyword {nnn} defines the number of nearest
neighbors used to calculate {Ql}. The default value is 12.
If the value is NULL, then all neighbors up to the
If the value is NULL, then all neighbors up to the
specified distance cutoff are used.
The optional keyword {degrees} defines the list of order parameters to
be computed. The first argument {nlvalues} is the number of order
be computed. The first argument {nlvalues} is the number of order
parameters. This is followed by that number of integers giving the
degree of each order parameter. Because {Q}2 and all odd-degree
order parameters are zero for atoms in cubic crystals
degree of each order parameter. Because {Q}2 and all odd-degree
order parameters are zero for atoms in cubic crystals
(see "Steinhardt"_#Steinhardt), the default order parameters
are {Q}4, {Q}6, {Q}8, {Q}10, and {Q}12. For the
FCC crystal with {nnn}=12, {Q}4 = sqrt(7/3)/8 = 0.19094....
The numerical values of all order parameters up to {Q}12
for a range of commonly encountered high-symmetry structures are given
for a range of commonly encountered high-symmetry structures are given
in Table I of "Mickel et al."_#Mickel.
The value of {Ql} is set to zero for atoms not in the
specified compute group, as well as for atoms that have less than
specified compute group, as well as for atoms that have less than
{nnn} neighbors within the distance cutoff.
The neighbor list needed to compute this quantity is constructed each
@ -109,9 +109,9 @@ options.
"compute coord/atom"_compute_coord_atom.html, "compute centro/atom"_compute_centro_atom.html, "compute hexorder/atom"_compute_hexorder_atom.html
[Default:]
[Default:]
The option defaults are {cutoff} = pair style cutoff, {nnn} = 12, {degrees} = 5 4 6 8 9 10 12 i.e. {Q}4, {Q}6, {Q}8, {Q}10, and {Q}12.
The option defaults are {cutoff} = pair style cutoff, {nnn} = 12, {degrees} = 5 4 6 8 9 10 12 i.e. {Q}4, {Q}6, {Q}8, {Q}10, and {Q}12.
:line

View File

@ -52,7 +52,7 @@ The KSpace contribution is calculated using the method in
"(Heyes)"_#Heyes for the Ewald method and a related method for PPPM,
as specified by the "kspace_style pppm"_kspace_style.html command.
For PPPM, the calcluation requires 1 extra FFT each timestep that
per-atom energy is calculated. Thie "document"_PDF/kspace.pdf
per-atom energy is calculated. This "document"_PDF/kspace.pdf
describes how the long-range per-atom energy calculation is performed.
Various fixes can contribute to the per-atom potential energy of the
@ -68,9 +68,9 @@ As an example of per-atom potential energy compared to total potential
energy, these lines in an input script should yield the same result
in the last 2 columns of thermo output:
compute peratom all pe/atom
compute pe all reduce sum c_peratom
thermo_style custom step temp etotal press pe c_pe :pre
compute peratom all pe/atom
compute pe all reduce sum c_peratom
thermo_style custom step temp etotal press pe c_pe :pre
NOTE: The per-atom energy does not any Lennard-Jones tail corrections
invoked by the "pair_modify tail yes"_pair_modify.html command, since

View File

@ -30,7 +30,7 @@ The plasticity for a Peridynamic particle is the so-called consistency
parameter (lambda). For elastic deformation lambda = 0, otherwise
lambda > 0 for plastic deformation. For details, see
"(Mitchell)"_#Mitchell and the PDF doc included in the LAMMPS
distro in "doc/PDF/PDLammps_EPS.pdf"_PDF/PDLammps_EPS.pdf.
distro in "doc/PDF/PDLammps_EPS.pdf"_PDF/PDLammps_EPS.pdf.
This command can be invoked for one of the Peridynamic "pair
styles"_pair_peri.html: peri/eps.
@ -57,7 +57,7 @@ LAMMPS"_Section_start.html#start_3 section for more info.
"compute damage/atom"_compute_damage_atom.html,
"compute dilatation/atom"_compute_dilatation_atom.html
[Default:] none
[Default:] none
:line

View File

@ -50,7 +50,7 @@ ordered xx, yy, zz, xy, xz, yz. The equation for the I,J components
(where I and J = x,y,z) is similar to the above formula, except that
the first term uses components of the kinetic energy tensor and the
second term uses components of the virial tensor:
:c,image(Eqs/pressure_tensor.jpg)
If no extra keywords are listed, the entire equations above are

View File

@ -16,20 +16,20 @@ ID, group-ID are documented in "compute"_compute.html command :ulb,l
property/atom = style name of this compute command :l
input = one or more atom attributes :l
possible attributes = id, mol, proc, type, mass,
x, y, z, xs, ys, zs, xu, yu, zu, ix, iy, iz,
vx, vy, vz, fx, fy, fz,
x, y, z, xs, ys, zs, xu, yu, zu, ix, iy, iz,
vx, vy, vz, fx, fy, fz,
q, mux, muy, muz, mu,
radius, diameter, omegax, omegay, omegaz,
angmomx, angmomy, angmomz,
shapex,shapey, shapez,
quatw, quati, quatj, quatk, tqx, tqy, tqz,
end1x, end1y, end1z, end2x, end2y, end2z,
corner1x, corner1y, corner1z,
corner2x, corner2y, corner2z,
corner3x, corner3y, corner3z,
nbonds,
angmomx, angmomy, angmomz,
shapex,shapey, shapez,
quatw, quati, quatj, quatk, tqx, tqy, tqz,
end1x, end1y, end1z, end2x, end2y, end2z,
corner1x, corner1y, corner1z,
corner2x, corner2y, corner2z,
corner3x, corner3y, corner3z,
nbonds,
vfrac, s0,
spin, eradius, ervel, erforce,
spin, eradius, ervel, erforce,
rho, drho, e, de, cv,
i_name, d_name :pre
id = atom ID
@ -80,7 +80,7 @@ input = one or more atom attributes :l
[Examples:]
compute 1 all property/atom xs vx fx mux
compute 1 all property/atom xs vx fx mux
compute 2 all property/atom type
compute 1 all property/atom ix iy iz :pre

View File

@ -16,7 +16,7 @@ ID, group-ID are documented in "compute"_compute.html command :ulb,l
property/chunk = style name of this compute command :l
input = one or more attributes :l
attributes = count, id, coord1, coord2, coord3
count = # of atoms in chunk
count = # of atoms in chunk
id = original chunk IDs before compression by "compute chunk/atom"_compute_chunk_atom.html
coord123 = coordinates for spatial bins calculated by "compute chunk/atom"_compute_chunk_atom.html :pre
:ule

View File

@ -15,12 +15,12 @@ compute ID group-ID property/local attribute1 attribute2 ... keyword args ... :p
ID, group-ID are documented in "compute"_compute.html command :ulb,l
property/local = style name of this compute command :l
one or more attributes may be appended :l
possible attributes = natom1 natom2 ntype1 ntype2
patom1 patom2 ptype1 ptype2
batom1 batom2 btype
aatom1 aatom2 aatom3 atype
datom1 datom2 datom3 dtype
iatom1 iatom2 iatom3 itype :pre
possible attributes = natom1 natom2 ntype1 ntype2
patom1 patom2 ptype1 ptype2
batom1 batom2 btype
aatom1 aatom2 aatom3 atype
datom1 datom2 datom3 dtype
iatom1 iatom2 iatom3 itype :pre
natom1, natom2 = IDs of 2 atoms in each pair (within neighbor cutoff)
ntype1, ntype2 = type of 2 atoms in each pair (within neighbor cutoff)
@ -129,8 +129,6 @@ The attributes that start with "a", "d", "i", refer to similar values
for "angles"_angle_style.html, "dihedrals"_dihedral_style.html, and
"impropers"_improper_style.html.
The optional {cutoff} keyword
[Output info:]
This compute calculates a local vector or local array depending on the

View File

@ -155,8 +155,8 @@ Thus, for example, if you wish to use this compute to find the bond
with maximum stretch, you can do it as follows:
compute 1 all property/local batom1 batom2
compute 2 all bond/local dist
compute 3 all reduce max c_1\[1\] c_1\[2\] c_2 replace 1 3 replace 2 3
compute 2 all bond/local dist
compute 3 all reduce max c_1\[1\] c_1\[2\] c_2 replace 1 3 replace 2 3
thermo_style custom step temp c_3\[1\] c_3\[2\] c_3\[3\] :pre
The first two input values in the compute reduce command are vectors

View File

@ -17,11 +17,11 @@ rigid/local = style name of this compute command :l
rigidID = ID of fix rigid/small command or one of its variants :l
input = one or more rigid body attributes :l
possible attributes = id, mol, mass,
x, y, z, xu, yu, zu, ix, iy, iz
vx, vy, vz, fx, fy, fz,
x, y, z, xu, yu, zu, ix, iy, iz
vx, vy, vz, fx, fy, fz,
omegax, omegay, omegaz,
angmomx, angmomy, angmomz,
quatw, quati, quatj, quatk,
angmomx, angmomy, angmomz,
quatw, quati, quatj, quatk,
tqx, tqy, tqz,
inertiax, inertiay, inertiaz
id = atom ID of atom within body which owns body properties
@ -29,7 +29,7 @@ input = one or more rigid body attributes :l
mass = total mass of body
x,y,z = center of mass coords of body
xu,yu,zu = unwrapped center of mass coords of body
ix,iy,iz = box image that the center of mass is in
ix,iy,iz = box image that the center of mass is in
vx,vy,vz = center of mass velocities
fx,fy,fz = force of center of mass
omegax,omegay,omegaz = angular velocity of body
@ -71,7 +71,7 @@ it is skipped (only one atom per body is so assigned). If it is the
assigned atom, then the info for that body is output. This means that
information for N bodies is generated. N may be less than the # of
bodies defined by the fix rigid command, if the atoms in some bodies
are not in the {group-ID}.
are not in the {group-ID}.
NOTE: Which atom in a body owns the body info is determined internal
to LAMMPS; it's the one nearest the geometric center of the body.
@ -109,7 +109,7 @@ sigma, etc). Use {xu}, {yu}, {zu} if you want the COM "unwrapped" by
the image flags for each atobody. Unwrapped means that if the body
COM has passed thru a periodic boundary one or more times, the value
is generated what the COM coordinate would be if it had not been
wrapped back into the periodic box.
wrapped back into the periodic box.
The image flags for the body can be generated directly using the {ix},
{iy}, {iz} attributes. For periodic dimensions, they specify which

View File

@ -19,18 +19,18 @@ type1 type2 ... typeN = chemical symbol of each atom type (see valid options bel
zero or more keyword/value pairs may be appended :l
keyword = {Kmax} or {Zone} or {dR_Ewald} or {c} or {manual} or {echo} :l
{Kmax} value = Maximum distance explored from reciprocal space origin
{Kmax} value = Maximum distance explored from reciprocal space origin
(inverse length units)
{Zone} values = z1 z2 z3
z1,z2,z3 = Zone axis of incident radiation. If z1=z2=z3=0 all
z1,z2,z3 = Zone axis of incident radiation. If z1=z2=z3=0 all
reciprocal space will be meshed up to {Kmax}
{dR_Ewald} value = Thickness of Ewald sphere slice intercepting
{dR_Ewald} value = Thickness of Ewald sphere slice intercepting
reciprocal space (inverse length units)
{c} values = c1 c2 c3
c1,c2,c3 = parameters to adjust the spacing of the reciprocal
c1,c2,c3 = parameters to adjust the spacing of the reciprocal
lattice nodes in the h, k, and l directions respectively
{manual} = flag to use manual spacing of reciprocal lattice points
based on the values of the {c} parameters
{manual} = flag to use manual spacing of reciprocal lattice points
based on the values of the {c} parameters
{echo} = flag to provide extra output for debugging purposes :pre
:ule
@ -44,22 +44,22 @@ fix saed/vtk 1 1 1 c_2 file Ni_000.saed :pre
[Description:]
Define a computation that calculates electron diffraction intensity as
described in "(Coleman)"_#saed-Coleman on a mesh of reciprocal lattice nodes
defined by the entire simulation domain (or manually) using simulated
radiation of wavelength lambda.
Define a computation that calculates electron diffraction intensity as
described in "(Coleman)"_#saed-Coleman on a mesh of reciprocal lattice nodes
defined by the entire simulation domain (or manually) using simulated
radiation of wavelength lambda.
The electron diffraction intensity I at each reciprocal lattice point
The electron diffraction intensity I at each reciprocal lattice point
is computed from the structure factor F using the equations:
:c,image(Eqs/compute_saed1.jpg)
:c,image(Eqs/compute_saed1.jpg)
:c,image(Eqs/compute_saed2.jpg)
Here, K is the location of the reciprocal lattice node, rj is the
Here, K is the location of the reciprocal lattice node, rj is the
position of each atom, fj are atomic scattering factors.
Diffraction intensities are calculated on a three-dimensional mesh of
reciprocal lattice nodes. The mesh spacing is defined either (a) by
Diffraction intensities are calculated on a three-dimensional mesh of
reciprocal lattice nodes. The mesh spacing is defined either (a) by
the entire simulation domain or (b) manually using selected values as
shown in the 2D diagram below.
@ -74,12 +74,12 @@ average of the (inversed) box lengths with periodic boundary conditions.
Meshes defined by the simulation domain must contain at least one periodic
boundary.
If the {manual} flag is included, the mesh of reciprocal lattice nodes
will defined using the {c} values for the spacing along each reciprocal
lattice axis. Note that manual mapping of the reciprocal space mesh is
good for comparing diffraction results from multiple simulations; however
it can reduce the likelihood that Bragg reflections will be satisfied
unless small spacing parameters <0.05 Angstrom^(-1) are implemented.
If the {manual} flag is included, the mesh of reciprocal lattice nodes
will defined using the {c} values for the spacing along each reciprocal
lattice axis. Note that manual mapping of the reciprocal space mesh is
good for comparing diffraction results from multiple simulations; however
it can reduce the likelihood that Bragg reflections will be satisfied
unless small spacing parameters <0.05 Angstrom^(-1) are implemented.
Meshes with manual spacing do not require a periodic boundary.
The limits of the reciprocal lattice mesh are determined by the use of
@ -98,17 +98,17 @@ in the below image.
:c,image(JPG/saed_ewald_intersect_small.jpg,JPG/saed_ewald_intersect.jpg)
The atomic scattering factors, fj, accounts for the reduction in
diffraction intensity due to Compton scattering. Compute saed uses
analytical approximations of the atomic scattering factors that vary
for each atom type (type1 type2 ... typeN) and angle of diffraction.
The atomic scattering factors, fj, accounts for the reduction in
diffraction intensity due to Compton scattering. Compute saed uses
analytical approximations of the atomic scattering factors that vary
for each atom type (type1 type2 ... typeN) and angle of diffraction.
The analytic approximation is computed using the formula
"(Brown)"_#Brown:
:c,image(Eqs/compute_saed3.jpg)
Coefficients parameterized by "(Fox)"_#Fox are assigned for each
atom type designating the chemical symbol and charge of each atom
Coefficients parameterized by "(Fox)"_#Fox are assigned for each
atom type designating the chemical symbol and charge of each atom
type. Valid chemical symbols for compute saed are:
H: He: Li: Be: B:
@ -133,14 +133,14 @@ type. Valid chemical symbols for compute saed are:
Cm: Bk: Cf:tb(c=5,s=:)
If the {echo} keyword is specified, compute saed will provide extra
reporting information to the screen.
If the {echo} keyword is specified, compute saed will provide extra
reporting information to the screen.
[Output info:]
This compute calculates a global vector. The length of the vector is
the number of reciprocal lattice nodes that are explored by the mesh.
The entries of the global vector are the computed diffraction
This compute calculates a global vector. The length of the vector is
the number of reciprocal lattice nodes that are explored by the mesh.
The entries of the global vector are the computed diffraction
intensities as described above.
The vector can be accessed by any command that uses global values
@ -148,21 +148,21 @@ from a compute as input. See "this
section"_Section_howto.html#howto_15 for an overview of LAMMPS output
options.
All array values calculated by this compute are "intensive".
All array values calculated by this compute are "intensive".
[Restrictions:]
[Restrictions:]
This compute is part of the USER-DIFFRACTION package. It is only
enabled if LAMMPS was built with that package. See the "Making
LAMMPS"_Section_start.html#start_3 section for more info.
The compute_saed command does not work for triclinic cells.
The compute_saed command does not work for triclinic cells.
[Related commands:]
[Related commands:]
"fix saed_vtk"_fix_saed_vtk.html, "compute xrd"_compute_xrd.html
[Default:]
[Default:]
The option defaults are Kmax = 1.70, Zone 1 0 0, c 1 1 1, dR_Ewald =
0.01.
@ -174,7 +174,7 @@ The option defaults are Kmax = 1.70, Zone 1 0 0, c 1 1 1, dR_Ewald =
(2013).
:link(Brown)
[(Brown)] Brown et al. International Tables for Crystallography
[(Brown)] Brown et al. International Tables for Crystallography
Volume C: Mathematical and Chemical Tables, 554-95 (2004).
:link(Fox)

View File

@ -22,7 +22,7 @@ compute 1 all smd/damage :pre
[Description:]
Define a computation that calculates the damage status of SPH particles
according to the damage model which is defined via the SMD SPH pair styles, e.g., the maximum plastic strain failure criterion.
according to the damage model which is defined via the SMD SPH pair styles, e.g., the maximum plastic strain failure criterion.
See "this PDF guide"_USER/smd/SMD_LAMMPS_userguide.pdf to use Smooth Mach Dynamics in LAMMPS.

View File

@ -31,7 +31,7 @@ keyword = {diagonal} or {rmin0} or {switchflag} :l
{2} = subset satisfying j1 == j2 == j3
{3} = subset satisfying j2 <= j1 <= j
{rmin0} value = parameter in distance to angle conversion (distance units)
{switchflag} value = {0} or {1}
{switchflag} value = {0} or {1}
{0} = do not use switching function
{1} = use switching function :pre
:ule
@ -60,12 +60,12 @@ onto the 3-sphere, the surface of the unit ball in a four-dimensional
space. The radial distance {r} within {R_ii'} is mapped on to a third
polar angle {theta0} defined by,
:c,image(Eqs/compute_sna_atom1.jpg)
:c,image(Eqs/compute_sna_atom1.jpg)
In this way, all possible neighbor positions are mapped on to a subset
of the 3-sphere. Points south of the latitude {theta0max=rfac0*Pi}
are excluded.
The natural basis for functions on the 3-sphere is formed by the 4D
hyperspherical harmonics {U^j_m,m'(theta, phi, theta0).} These
functions are better known as {D^j_m,m',} the elements of the Wigner
@ -78,7 +78,7 @@ radial distance. Expanding this density function as a generalized
Fourier series in the basis functions, we can write each Fourier
coefficient as
:c,image(Eqs/compute_sna_atom2.jpg)
:c,image(Eqs/compute_sna_atom2.jpg)
The {w_i'} neighbor weights are dimensionless numbers that are chosen
to distinguish atoms of different types, while the central atom is
@ -86,7 +86,7 @@ arbitrarily assigned a unit weight. The function {fc(r)} ensures that
the contribution of each neighbor atom goes smoothly to zero at
{R_ii'}:
:c,image(Eqs/compute_sna_atom4.jpg)
:c,image(Eqs/compute_sna_atom4.jpg)
The expansion coefficients {u^j_m,m'} are complex-valued and they are
not directly useful as descriptors, because they are not invariant
@ -94,7 +94,7 @@ under rotation of the polar coordinate frame. However, the following
scalar triple products of expansion coefficients can be shown to be
real-valued and invariant under rotation "(Bartok)"_#Bartok2010.
:c,image(Eqs/compute_sna_atom3.jpg)
:c,image(Eqs/compute_sna_atom3.jpg)
The constants {H^jmm'_j1m1m1'_j2m2m2'} are coupling coefficients,
analogous to Clebsch-Gordan coefficients for rotations on the
@ -112,17 +112,17 @@ atom.
Compute {snad/atom} calculates the derivative of the bispectrum components
summed separately for each atom type:
:c,image(Eqs/compute_sna_atom5.jpg)
:c,image(Eqs/compute_sna_atom5.jpg)
The sum is over all atoms {i'} of atom type {I}. For each atom {i},
this compute evaluates the above expression for each direction, each
atom type, and each bispectrum component. See section below on output
for a detailed explanation.
Compute {snav/atom} calculates the virial contribution due to the
derivatives:
:c,image(Eqs/compute_sna_atom6.jpg)
:c,image(Eqs/compute_sna_atom6.jpg)
Again, the sum is over all atoms {i'} of atom type {I}. For each atom
{i}, this compute evaluates the above expression for each of the six
@ -140,7 +140,7 @@ too frequently.
The argument {rcutfac} is a scale factor that controls the ratio of
atomic radius to radial cutoff distance.
The argument {rfac0} and the optional keyword {rmin0} define the
linear mapping from radial distance to polar angle {theta0} on the
3-sphere.
@ -176,18 +176,18 @@ each column depend on the values of {twojmax} and {diagonal}, as
described by the following piece of python code:
for j1 in range(0,twojmax+1):
if(diagonal==2):
if(diagonal==2):
print j1/2.,j1/2.,j1/2.
elif(diagonal==1):
for j in range(0,min(twojmax,2*j1)+1,2):
for j in range(0,min(twojmax,2*j1)+1,2):
print j1/2.,j1/2.,j/2.
elif(diagonal==0):
for j2 in range(0,j1+1):
for j in range(j1-j2,min(twojmax,j1+j2)+1,2):
for j in range(j1-j2,min(twojmax,j1+j2)+1,2):
print j1/2.,j2/2.,j/2.
elif(diagonal==3):
for j2 in range(0,j1+1):
for j in range(j1-j2,min(twojmax,j1+j2)+1,2):
for j in range(j1-j2,min(twojmax,j1+j2)+1,2):
if (j>=j1): print j1/2.,j2/2.,j/2. :pre
Compute {snad/atom} evaluates a per-atom array. The columns are
@ -227,7 +227,7 @@ The optional keyword defaults are {diagonal} = 0, {rmin0} = 0,
:line
:link(Thompson2014)
[(Thompson)] Thompson, Swiler, Trott, Foiles, Tucker, under review, preprint
[(Thompson)] Thompson, Swiler, Trott, Foiles, Tucker, under review, preprint
available at "arXiv:1409.3880"_http://arxiv.org/abs/1409.3880
:link(Bartok2010)
@ -235,7 +235,7 @@ available at "arXiv:1409.3880"_http://arxiv.org/abs/1409.3880
:link(Meremianin2006)
[(Meremianin)] Meremianin, J. Phys. A, 39, 3099 (2006).
:link(Varshalovich1987)
[(Varshalovich)] Varshalovich, Moskalev, Khersonskii, Quantum Theory
of Angular Momentum, World Scientific, Singapore (1987).

View File

@ -128,10 +128,10 @@ d = dimension and V is the volume of the system, the result should be
These lines in an input script for a 3d system should yield that
result. I.e. the last 2 columns of thermo output will be the same:
compute peratom all stress/atom NULL
compute p all reduce sum c_peratom\[1\] c_peratom\[2\] c_peratom\[3\]
variable press equal -(c_p\[1\]+c_p\[2\]+c_p\[3\])/(3*vol)
thermo_style custom step temp etotal press v_press :pre
compute peratom all stress/atom NULL
compute p all reduce sum c_peratom\[1\] c_peratom\[2\] c_peratom\[3\]
variable press equal -(c_p\[1\]+c_p\[2\]+c_p\[3\])/(3*vol)
thermo_style custom step temp etotal press v_press :pre
[Output info:]

View File

@ -88,7 +88,7 @@ potentials only include the pair potential portion of the EAM
interaction when used by this compute, not the embedding term. Also
bonded or Kspace interactions do not contribute to this compute.
[Related commands:]
[Related commands:]
{compute group/group}_compute_group_group.html, {compute
heat/flux}_compute_heat_flux.html

0
doc/src/compute_temp_asphere.txt Executable file → Normal file
View File

0
doc/src/compute_temp_body.txt Executable file → Normal file
View File

View File

@ -16,7 +16,7 @@ ID, group-ID are documented in "compute"_compute.html command
temp/cs = style name of this compute command
group1 = group-ID of either cores or shells
group2 = group-ID of either shells or cores :ul
[Examples:]
compute oxygen_c-s all temp/cs O_core O_shell
@ -64,7 +64,7 @@ calculated by this compute for use in the computation of a pressure
tensor. The formula for the components of the tensor is the same as
the above formula, except that v^2 is replaced by vx*vy for the xy
component, etc. The 6 components of the vector are ordered xx, yy,
zz, xy, xz, yz. In contrast to the temperature, the velocity of
zz, xy, xz, yz. In contrast to the temperature, the velocity of
each core or shell atom is taken individually.
The change this fix makes to core/shell atom velocities is essentially

View File

@ -14,7 +14,7 @@ compute ID group-ID temp/drude :pre
ID, group-ID are documented in "compute"_compute.html command
temp/drude = style name of this compute command :ul
[Examples:]
compute TDRUDE all temp/drude :pre
@ -68,7 +68,7 @@ are "extensive".
[Restrictions:]
The number of degrees of freedom contributing to the temperature is
assumed to be constant for the duration of the run unless the
assumed to be constant for the duration of the run unless the
{fix_modify} command sets the option {dynamic yes}.
[Related commands:]

View File

@ -49,9 +49,9 @@ reported by LAMMPS in the thermodynamic quantities reported via the
example:
compute effTemp all temp/eff
thermo_style custom step etotal pe ke temp press
thermo_style custom step etotal pe ke temp press
thermo_modify temp effTemp :pre
A 6-component kinetic energy tensor is also calculated by this compute
for use in the computation of a pressure tensor. The formula for the
components of the tensor is the same as the above formula, except that
@ -80,7 +80,7 @@ is independent of the number of atoms in the simulation. The vector
values are "extensive", meaning they scale with the number of atoms in
the simulation.
[Restrictions:]
[Restrictions:]
This compute is part of the USER-EFF package. It is only enabled if
LAMMPS was built with that package. See the "Making

View File

@ -68,7 +68,7 @@ temp/berendsen"_fix_temp_berendsen.html, and "fix
langevin"_fix_langevin.html. This means that when this compute
is used to calculate the temperature for any of the thermostatting
fixes via the "fix modify temp"_fix_modify.html command, the thermostat
will operate only on atoms that are currently in the geometric
will operate only on atoms that are currently in the geometric
region.
Unlike other compute styles that calculate temperature, this compute

2
doc/src/compute_temp_sphere.txt Executable file → Normal file
View File

@ -122,7 +122,7 @@ vector values will be in energy "units"_units.html.
This fix requires that atoms store torque and angular velocity (omega)
and a radius as defined by the "atom_style sphere"_atom_style.html
command.
command.
All particles in the group must be finite-size spheres, or point
particles with radius = 0.0.

View File

@ -6,7 +6,7 @@
:line
compute ti command :h3
compute ti command :h3
[Syntax:]
@ -35,7 +35,7 @@ keyword = pair style (lj/cut, gauss, born, etc) or {tail} or {kspace} :l
compute 1 all ti lj/cut 1 v_lj v_dlj coul/long 2 v_c v_dc kspace 1 v_ks v_dks
compute 1 all ti lj/cut 1*3 v_lj v_dlj coul/long * v_c v_dc kspace * v_ks v_dks :pre
[Description:]
[Description:]
Define a computation that calculates the derivative of the interaction
potential with respect to {lambda}, the coupling parameter used in a
@ -107,7 +107,7 @@ du/dl can be found in the paper by "Eike"_#Eike.
:line
[Output info:]
[Output info:]
This compute calculates a global scalar, namely dUs/dlambda. This
value can be used by any command that uses a global scalar value from

View File

@ -15,7 +15,7 @@ compute ID group-ID voronoi/atom keyword arg ... :pre
ID, group-ID are documented in "compute"_compute.html command :ulb,l
voronoi/atom = style name of this compute command :l
zero or more keyword/value pairs may be appended :l
keyword = {only_group} or {surface} or {radius} or {edge_histo} or {edge_threshold}
keyword = {only_group} or {surface} or {radius} or {edge_histo} or {edge_threshold}
or {face_threshold} or {neighbors} or {peratom} :l
{only_group} = no arg
{occupation} = no arg
@ -25,7 +25,7 @@ or {face_threshold} or {neighbors} or {peratom} :l
{radius} arg = v_r
v_r = radius atom style variable for a poly-disperse Voronoi tessellation
{edge_histo} arg = maxedge
maxedge = maximum number of Voronoi cell edges to be accounted in the histogram
maxedge = maximum number of Voronoi cell edges to be accounted in the histogram
{edge_threshold} arg = minlength
minlength = minimum length for an edge to be counted
{face_threshold} arg = minarea
@ -38,7 +38,7 @@ or {face_threshold} or {neighbors} or {peratom} :l
compute 1 all voronoi/atom
compute 2 precipitate voronoi/atom surface matrix
compute 3b precipitate voronoi/atom radius v_r
compute 3b precipitate voronoi/atom radius v_r
compute 4 solute voronoi/atom only_group
compute 5 defects voronoi/atom occupation
compute 6 all voronoi/atom neighbors yes :pre
@ -53,11 +53,11 @@ in the group.
By default two per-atom quantities are calculated by this compute.
The first is the volume of the Voronoi cell around each atom. Any
point in an atom's Voronoi cell is closer to that atom than any other.
The second is the number of faces of the Voronoi cell. This is
The second is the number of faces of the Voronoi cell. This is
equal to the number of nearest neighbors of the central atom,
plus any exterior faces (see note below). If the {peratom} keyword
is set to "no", the per-atom quantities are still calculated,
but they are not accessible.
plus any exterior faces (see note below). If the {peratom} keyword
is set to "no", the per-atom quantities are still calculated,
but they are not accessible.
:line
@ -122,23 +122,23 @@ to locate vacancies (the coordinates are given by the atom coordinates
at the time step when the compute was first invoked), while column two
data can be used to identify interstitial atoms.
If the {neighbors} value is set to yes, then
If the {neighbors} value is set to yes, then
this compute creates a local array with 3 columns. There
is one row for each face of each Voronoi cell. The
3 columns are the atom ID of the atom that owns the cell,
the atom ID of the atom in the neighboring cell
(or zero if the face is external), and the area of the face.
is one row for each face of each Voronoi cell. The
3 columns are the atom ID of the atom that owns the cell,
the atom ID of the atom in the neighboring cell
(or zero if the face is external), and the area of the face.
The array can be accessed by any command that
uses local values from a compute as input. See "this
section"_Section_howto.html#howto_15 for an overview of LAMMPS output
options. More specifically, the array can be accessed by a
options. More specifically, the array can be accessed by a
"dump local"_dump.html command to write a file containing
all the Voronoi neighbors in a system:
compute 6 all voronoi/atom neighbors yes
dump d2 all local 1 dump.neighbors index c_6\[1\] c_6\[2\] c_6\[3\] :pre
If the {face_threshold} keyword is used, then only faces
If the {face_threshold} keyword is used, then only faces
with areas greater than the threshold are stored.
:line
@ -190,7 +190,7 @@ per-atom values from a compute as input. See "Section
6.15"_Section_howto.html#howto_15 for an overview of LAMMPS output
options. If the {peratom} keyword is set to "no", the per-atom array
is still created, but it is not accessible.
If the {edge_histo} keyword is used, then this compute generates a
global vector of length {maxedge}+1, containing a histogram of the
number of edges per face.

View File

@ -20,21 +20,21 @@ type1 type2 ... typeN = chemical symbol of each atom type (see valid options bel
zero or more keyword/value pairs may be appended :l
keyword = {2Theta} or {c} or {LP} or {manual} or {echo} :l
{2Theta} values = Min2Theta Max2Theta
Min2Theta,Max2Theta = minimum and maximum 2 theta range to explore
Min2Theta,Max2Theta = minimum and maximum 2 theta range to explore
(radians or degrees)
{c} values = c1 c2 c3
c1,c2,c3 = parameters to adjust the spacing of the reciprocal
c1,c2,c3 = parameters to adjust the spacing of the reciprocal
lattice nodes in the h, k, and l directions respectively
{LP} value = switch to apply Lorentz-polarization factor
0/1 = off/on
{manual} = flag to use manual spacing of reciprocal lattice points
based on the values of the {c} parameters
{manual} = flag to use manual spacing of reciprocal lattice points
based on the values of the {c} parameters
{echo} = flag to provide extra output for debugging purposes :pre
:ule
[Examples:]
compute 1 all xrd 1.541838 Al O 2Theta 0.087 0.87 c 1 1 1 LP 1 echo
compute 1 all xrd 1.541838 Al O 2Theta 0.087 0.87 c 1 1 1 LP 1 echo
compute 2 all xrd 1.541838 Al O 2Theta 10 100 c 0.05 0.05 0.05 LP 1 manual :pre
fix 1 all ave/histo/weight 1 1 1 0.087 0.87 250 c_1\[1\] c_1\[2\] mode vector file Rad2Theta.xrd
@ -43,11 +43,11 @@ fix 2 all ave/histo/weight 1 1 1 10 100 250 c_2\[1\] c_2\[2\] mode vector file D
[Description:]
Define a computation that calculates x-ray diffraction intensity as described
in "(Coleman)"_#xrd-Coleman on a mesh of reciprocal lattice nodes defined
in "(Coleman)"_#xrd-Coleman on a mesh of reciprocal lattice nodes defined
by the entire simulation domain (or manually) using a simulated radiation
of wavelength lambda.
of wavelength lambda.
The x-ray diffraction intensity, I, at each reciprocal lattice point, k,
The x-ray diffraction intensity, I, at each reciprocal lattice point, k,
is computed from the structure factor, F, using the equations:
:c,image(Eqs/compute_xrd1.jpg)
@ -55,14 +55,14 @@ is computed from the structure factor, F, using the equations:
:c,image(Eqs/compute_xrd3.jpg)
:c,image(Eqs/compute_xrd4.jpg)
Here, K is the location of the reciprocal lattice node, rj is the
position of each atom, fj are atomic scattering factors, LP is the
Lorentz-polarization factor, and theta is the scattering angle of
diffraction. The Lorentz-polarization factor can be turned off using
Here, K is the location of the reciprocal lattice node, rj is the
position of each atom, fj are atomic scattering factors, LP is the
Lorentz-polarization factor, and theta is the scattering angle of
diffraction. The Lorentz-polarization factor can be turned off using
the optional {LP} keyword.
Diffraction intensities are calculated on a three-dimensional mesh of
reciprocal lattice nodes. The mesh spacing is defined either (a)
Diffraction intensities are calculated on a three-dimensional mesh of
reciprocal lattice nodes. The mesh spacing is defined either (a)
by the entire simulation domain or (b) manually using selected values as
shown in the 2D diagram below.
@ -101,8 +101,8 @@ The analytic approximation is computed using the formula
:c,image(Eqs/compute_xrd5.jpg)
Coefficients parameterized by "(Peng)"_#Peng are assigned for each
atom type designating the chemical symbol and charge of each atom
Coefficients parameterized by "(Peng)"_#Peng are assigned for each
atom type designating the chemical symbol and charge of each atom
type. Valid chemical symbols for compute xrd are:
H| He1-| He| Li| Li1+|
@ -148,39 +148,39 @@ type. Valid chemical symbols for compute xrd are:
Np4+| Np6+| Pu| Pu3+| Pu4+|
Pu6+| Am| Cm| Bk| Cf :tb(c=5,s=|)
If the {echo} keyword is specified, compute xrd will provide extra
reporting information to the screen.
If the {echo} keyword is specified, compute xrd will provide extra
reporting information to the screen.
[Output info:]
This compute calculates a global array. The number of rows in the
array is the number of reciprocal lattice nodes that are explored
which by the mesh. The global array has 2 columns.
This compute calculates a global array. The number of rows in the
array is the number of reciprocal lattice nodes that are explored
which by the mesh. The global array has 2 columns.
The first column contains the diffraction angle in the units (radians
or degrees) provided with the {2Theta} values. The second column contains
or degrees) provided with the {2Theta} values. The second column contains
the computed diffraction intensities as described above.
The array can be accessed by any command that uses global values from
a compute as input. See "this section"_Section_howto.html#howto_15
for an overview of LAMMPS output options.
All array values calculated by this compute are "intensive".
All array values calculated by this compute are "intensive".
[Restrictions:]
[Restrictions:]
This compute is part of the USER-DIFFRACTION package. It is only
enabled if LAMMPS was built with that package. See the "Making
LAMMPS"_Section_start.html#start_3 section for more info.
The compute_xrd command does not work for triclinic cells.
The compute_xrd command does not work for triclinic cells.
[Related commands:]
[Related commands:]
"fix ave/histo"_fix_ave_histo.html,
"compute saed"_compute_saed.html
[Default:]
[Default:]
The option defaults are 2Theta = 1 179 (degrees), c = 1 1 1, LP = 1,
no manual flag, no echo flag.
@ -192,7 +192,7 @@ no manual flag, no echo flag.
(2013).
:link(Colliex)
[(Colliex)] Colliex et al. International Tables for Crystallography
[(Colliex)] Colliex et al. International Tables for Crystallography
Volume C: Mathematical and Chemical Tables, 249-429 (2004).
:link(Peng)

View File

@ -48,7 +48,7 @@ keyword = {mol} or {basis} or {remap} or {var} or {set} or {units} :l
create_atoms 1 box
create_atoms 3 region regsphere basis 2 3
create_atoms 3 single 0 0 5
create_atoms 3 single 0 0 5
create_atoms 1 box var v set x xpos set y ypos :pre
[Description:]
@ -218,14 +218,14 @@ larger version.
variable x equal 100
variable y equal 25
lattice hex 0.8442
region box block 0 $x 0 $y -0.5 0.5
create_box 1 box :pre
lattice hex 0.8442
region box block 0 $x 0 $y -0.5 0.5
create_box 1 box :pre
variable xx equal 0.0
variable yy equal 0.0
variable v equal "(0.2*v_y*ylat * cos(v_xx/xlat * 2.0*PI*4.0/v_x) + 0.5*v_y*ylat - v_yy) > 0.0"
create_atoms 1 box var v set x xx set y yy :pre
create_atoms 1 box var v set x xx set y yy :pre
:c,image(JPG/sinusoid_small.jpg,JPG/sinusoid.jpg)
@ -245,7 +245,7 @@ style. A {box} value selects standard distance units as defined by
the "units"_units.html command, e.g. Angstroms for units = real or
metal. A {lattice} value means the distance units are in lattice
spacings.
:line
Atom IDs are assigned to created atoms in the following way. The

View File

@ -121,7 +121,7 @@ The {special} keyword is invoked at the end of the delete_bonds
operation, after (optional) removal. It re-computes the pairwise 1-2,
1-3, 1-4 weighting list. The weighting list computation treats
turned-off bonds the same as turned-on. Thus, turned-off bonds must
be removed if you wish to change the weighting list.
be removed if you wish to change the weighting list.
Note that the choice of {remove} and {special} options affects how
1-2, 1-3, 1-4 pairwise interactions will be computed across bonds that

View File

@ -6,7 +6,7 @@
:line
dielectric command :h3
dielectric command :h3
[Syntax:]

View File

@ -17,10 +17,10 @@ dihedral_style class2 :pre
dihedral_style class2
dihedral_coeff 1 100 75 100 70 80 60
dihedral_coeff * mbt 3.5945 0.1704 -0.5490 1.5228
dihedral_coeff * mbt 3.5945 0.1704 -0.5490 1.5228
dihedral_coeff * ebt 0.3417 0.3264 -0.9036 0.1368 0.0 -0.8080 1.0119 1.1010
dihedral_coeff 2 at 0.0 -0.1850 -0.7963 -2.0220 0.0 -0.3991 110.2453 105.1270
dihedral_coeff * aat -13.5271 110.2453 105.1270
dihedral_coeff 2 at 0.0 -0.1850 -0.7963 -2.0220 0.0 -0.3991 110.2453 105.1270
dihedral_coeff * aat -13.5271 110.2453 105.1270
dihedral_coeff * bb13 0.0 1.0119 1.1010 :pre
[Description:]

View File

@ -66,7 +66,7 @@ more instructions on how to use the accelerated styles effectively.
[Restrictions:]
This angle style can only be used if LAMMPS was built with the
USER_MISC package. See the "Making LAMMPS"_Section_start.html#start_3
USER_MISC package. See the "Making LAMMPS"_Section_start.html#start_3
section for more info on packages.
[Related commands:]

View File

@ -63,7 +63,7 @@ more instructions on how to use the accelerated styles effectively.
[Restrictions:]
This angle style can only be used if LAMMPS was built with the
USER_MISC package. See the "Making LAMMPS"_Section_start.html#start_3
USER_MISC package. See the "Making LAMMPS"_Section_start.html#start_3
section for more info on packages.
[Related commands:]

View File

@ -33,7 +33,7 @@ above, or in the data file or restart files read by the
"read_data"_read_data.html or "read_restart"_read_restart.html
commands:
K (energy/radian^2)
K (energy/radian^2)
phi0 (degrees) :ul
:line
@ -64,7 +64,7 @@ more instructions on how to use the accelerated styles effectively.
[Restrictions:]
This angle style can only be used if LAMMPS was built with the
USER_MISC package. See the "Making LAMMPS"_Section_start.html#start_3
USER_MISC package. See the "Making LAMMPS"_Section_start.html#start_3
section for more info on packages.
[Related commands:]

View File

@ -14,7 +14,7 @@ dihedral_style spherical :pre
[Examples:]
dihedral_coeff 1 1 286.1 1 124 1 1 90.0 0 1 90.0 0
dihedral_coeff 1 1 286.1 1 124 1 1 90.0 0 1 90.0 0
dihedral_coeff 1 3 286.1 1 114 1 1 90 0 1 90.0 0 &
17.3 0 0.0 0 1 158 1 0 0.0 0 &
15.1 0 0.0 0 0 0.0 0 1 167.3 1 :pre
@ -76,7 +76,7 @@ wn (typically 0.0 or 1.0) :ul
[Restrictions:]
This dihedral style can only be used if LAMMPS was built with the
USER_MISC package. See the "Making LAMMPS"_Section_start.html#start_3
USER_MISC package. See the "Making LAMMPS"_Section_start.html#start_3
section for more info on packages.
[Related commands:]

View File

@ -5,7 +5,7 @@
:link(lc,Section_commands.html#comm)
:line
dump command :h3
"dump custom/vtk"_dump_custom_vtk.html command :h3
"dump h5md"_dump_h5md.html command :h3
@ -55,13 +55,13 @@ args = list of arguments for a particular style :l
{custom} or {custom/gz} or {custom/mpiio} args = list of atom attributes
possible attributes = id, mol, proc, procp1, type, element, mass,
x, y, z, xs, ys, zs, xu, yu, zu,
xsu, ysu, zsu, ix, iy, iz,
vx, vy, vz, fx, fy, fz,
x, y, z, xs, ys, zs, xu, yu, zu,
xsu, ysu, zsu, ix, iy, iz,
vx, vy, vz, fx, fy, fz,
q, mux, muy, muz, mu,
radius, diameter, omegax, omegay, omegaz,
angmomx, angmomy, angmomz, tqx, tqy, tqz,
c_ID, c_ID\[N\], f_ID, f_ID\[N\], v_name :pre
angmomx, angmomy, angmomz, tqx, tqy, tqz,
c_ID, c_ID\[N\], f_ID, f_ID\[N\], v_name :pre
id = atom ID
mol = molecule ID
@ -211,7 +211,7 @@ bounding box which encloses the triclinic simulation box is output,
along with the 3 tilt factors (xy, xz, yz) of the triclinic box,
formatted as follows:
ITEM: BOX BOUNDS xy xz yz xx yy zz
ITEM: BOX BOUNDS xy xz yz xx yy zz
xlo_bound xhi_bound xy
ylo_bound yhi_bound xz
zlo_bound zhi_bound yz :pre
@ -305,8 +305,8 @@ by the GROMACS molecular dynamics package, and described
The precision used in XTC files can be adjusted via the
"dump_modify"_dump_modify.html command. The default value of 1000
means that coordinates are stored to 1/1000 nanometer accuracy. XTC
files are portable binary files written in the NFS XDR data format,
so that any machine which supports XDR should be able to read them.
files are portable binary files written in the NFS XDR data format,
so that any machine which supports XDR should be able to read them.
The number of atoms per snapshot cannot change with the {xtc} style.
The {unwrap} option of the "dump_modify"_dump_modify.html command allows
XTC coordinates to be written "unwrapped" by the image flags for each
@ -499,7 +499,7 @@ values.
Here is an example of how to dump bond info for a system, including
the distance and energy of each bond:
compute 1 all property/local batom1 batom2 btype
compute 1 all property/local batom1 batom2 btype
compute 2 all bond/local dist eng
dump 1 all local 1000 tmp.dump index c_1\[1\] c_1\[2\] c_1\[3\] c_2\[1\] c_2\[2\] :pre

View File

@ -5,7 +5,7 @@
:link(lc,Section_commands.html#comm)
:line
dump custom/vtk command :h3
[Syntax:]
@ -20,14 +20,14 @@ file = name of file to write dump info to :l
args = list of arguments for a particular style :l
{custom/vtk} args = list of atom attributes
possible attributes = id, mol, proc, procp1, type, element, mass,
x, y, z, xs, ys, zs, xu, yu, zu,
xsu, ysu, zsu, ix, iy, iz,
vx, vy, vz, fx, fy, fz,
x, y, z, xs, ys, zs, xu, yu, zu,
xsu, ysu, zsu, ix, iy, iz,
vx, vy, vz, fx, fy, fz,
q, mux, muy, muz, mu,
radius, diameter, omegax, omegay, omegaz,
angmomx, angmomy, angmomz, tqx, tqy, tqz,
spin, eradius, ervel, erforce,
c_ID, c_ID\[N\], f_ID, f_ID\[N\], v_name :pre
angmomx, angmomy, angmomz, tqx, tqy, tqz,
spin, eradius, ervel, erforce,
c_ID, c_ID\[N\], f_ID, f_ID\[N\], v_name :pre
id = atom ID
mol = molecule ID

View File

@ -587,7 +587,7 @@ b) Use the freely available mplayer or ffplay tool to view a
movie. Both are available for multiple OSes and support a large
variety of file formats and decoders. :l
% mplayer foo.mpg
% mplayer foo.mpg
% ffplay bar.avi :pre
c) Use the "Pizza.py"_http://www.sandia.gov/~sjplimp/pizza.html
@ -631,9 +631,9 @@ Note that since FFmpeg is run as an external program via a pipe,
LAMMPS has limited control over its execution and no knowledge about
errors and warnings printed by it. Those warnings and error messages
will be printed to the screen only. Due to the way image data is
communicated to FFmpeg, it will often print the message
communicated to FFmpeg, it will often print the message
pipe:: Input/output error :pre
pipe:: Input/output error :pre
which can be safely ignored. Other warnings
and errors have to be addressed according to the FFmpeg documentation.

View File

@ -49,12 +49,12 @@ keyword = {append} or {buffer} or {element} or {every} or {fileper} or {first} o
-N = sort per-atom lines in descending order by the Nth column
{thresh} args = attribute operation value
attribute = same attributes (x,fy,etotal,sxx,etc) used by dump custom style
operation = "<" or "<=" or ">" or ">=" or "==" or "!="
value = numeric value to compare to
operation = "<" or "<=" or ">" or ">=" or "==" or "!=" or "|^"
value = numeric value to compare to, or LAST
these 3 args can be replaced by the word "none" to turn off thresholding
{unwrap} arg = {yes} or {no} :pre
these keywords apply only to the {image} and {movie} "styles"_dump_image.html :l
keyword = {acolor} or {adiam} or {amap} or {backcolor} or {bcolor} or {bdiam} or {boxcolor} or {color} or {bitrate} or {framerate} :l
keyword = {acolor} or {adiam} or {amap} or {backcolor} or {bcolor} or {bdiam} or {boxcolor} or {color} or {bitrate} or {framerate} :l
{acolor} args = type color
type = atom type or range of types (see below)
color = name of color or color1/color2/...
@ -215,17 +215,17 @@ to the dump file. The {every} keyword cannot be used with the dump
For example, the following commands will
write snapshots at timesteps 0,10,20,30,100,200,300,1000,2000,etc:
variable s equal logfreq(10,3,10)
dump 1 all atom 100 tmp.dump
dump_modify 1 every v_s first yes :pre
variable s equal logfreq(10,3,10)
dump 1 all atom 100 tmp.dump
dump_modify 1 every v_s first yes :pre
The following commands would write snapshots at the timesteps listed
in file tmp.times:
variable f file tmp.times
variable s equal next(f)
dump 1 all atom 100 tmp.dump
dump_modify 1 every v_s :pre
variable s equal next(f)
dump 1 all atom 100 tmp.dump
dump_modify 1 every v_s :pre
NOTE: When using a file-style variable with the {every} keyword, the
file of timesteps must list a first timestep that is beyond the
@ -458,16 +458,56 @@ as well as memory, versus unsorted output.
The {thresh} keyword only applies to the dump {custom}, {cfg},
{image}, and {movie} styles. Multiple thresholds can be specified.
Specifying "none" turns off all threshold criteria. If thresholds are
Specifying {none} turns off all threshold criteria. If thresholds are
specified, only atoms whose attributes meet all the threshold criteria
are written to the dump file or included in the image. The possible
attributes that can be tested for are the same as those that can be
specified in the "dump custom"_dump.html command, with the exception
of the {element} attribute, since it is not a numeric value. Note
that different attributes can be output by the dump custom command
than are used as threshold criteria by the dump_modify command.
E.g. you can output the coordinates and stress of atoms whose energy
is above some threshold.
that a different attributes can be used than those output by the "dump
custom"_dump.html command. E.g. you can output the coordinates and
stress of atoms whose energy is above some threshold.
If an atom-style variable is used as the attribute, then it can
produce continuous numeric values or effective Boolean 0/1 values
which may be useful for the comparision operation. Boolean values can
be generated by variable formulas that use comparison or Boolean math
operators or special functions like gmask() and rmask() and grmask().
See the "variable"_variable.html command doc page for details.
The specified value must be a simple numeric value or the word LAST.
If LAST is used, it refers to the value of the attribute the last time
the dump command was invoked to produce a snapshot. This is a way to
only dump atoms whose attribute has changed (or not changed).
Three examples follow.
dump_modify ... thresh ix != LAST :pre
This will dump atoms which have crossed the periodic x boundary of the
simulation box since the last dump. (Note that atoms that crossed
once and then crossed back between the two dump timesteps would not be
included.)
region foo sphere 10 20 10 15
variable inregion atom rmask(foo)
dump_modify ... thresh v_inregion |^ LAST
This will dump atoms which crossed the boundary of the spherical
region since the last dump.
variable charge atom "(q > 0.5) || (q < -0.5)"
dump_modify ... thresh v_charge |^ LAST
This will dump atoms whose charge has changed from an absolute value
less than 1/2 to greater than 1/2 (or vice versa) since the last dump.
E.g. due to reactions and subsequent charge equilibration in a
reactive force field.
The choice of operations are the usual comparison operators. The XOR
operation (exclusive or) is also included as "|^". In this context,
XOR means that if either the attribute or value is 0.0 and the other
is non-zero, then the result is "true" and the threshold criterion is
met. Otherwise it is not met.
:line
@ -643,10 +683,10 @@ this is used.
variable colors string &
"red green blue yellow white &
purple pink orange lime gray"
variable mol atom mol%10
dump 1 all image 250 image.*.jpg v_mol type &
zoom 1.6 adiam 1.5
dump_modify 1 pad 5 amap 0 10 sa 1 10 $\{colors\} :pre
variable mol atom mol%10
dump 1 all image 250 image.*.jpg v_mol type &
zoom 1.6 adiam 1.5
dump_modify 1 pad 5 amap 0 10 sa 1 10 $\{colors\} :pre
In this case, 10 colors are defined, and molecule IDs are
mapped to one of the colors, even if there are 1000s of molecules.

View File

@ -17,17 +17,17 @@ group = name of group fix is to be applied :l
type = {thermal} or {two_temperature} or {hardy} or {field} :l
{thermal} = thermal coupling with fields: temperature
{two_temperature} = electron-phonon coupling with field: temperature and electron_temperature
{hardy} = on-the-fly post-processing using kernel localization functions (see "related" section for possible fields)
{field} = on-the-fly post-processing using mesh-based localization functions (see "related" section for possible fields) :pre
{hardy} = on-the-fly post-processing using kernel localization functions (see "related" section for possible fields)
{field} = on-the-fly post-processing using mesh-based localization functions (see "related" section for possible fields) :pre
parameter_file = name of the file with material parameters. Note: Neither hardy nor field requires a parameter file :l
:ule
:ule
[Examples:]
fix AtC internal atc thermal Ar_thermal.dat
fix AtC internal atc thermal Ar_thermal.dat
fix AtC internal atc two_temperature Ar_ttm.mat
fix AtC internal atc hardy
fix AtC internal atc field :pre
fix AtC internal atc field :pre
[Description:]
@ -38,67 +38,67 @@ This fix is the beginning to creating a coupled FE/MD simulation and/or an on-th
The following coupling example is typical, but non-exhaustive:
# ... commands to create and initialize the MD system :pre
# initial fix to designate coupling type and group to apply it to
# tag group physics material_file
# initial fix to designate coupling type and group to apply it to
# tag group physics material_file
fix AtC internal atc thermal Ar_thermal.mat :pre
# create a uniform 12 x 2 x 2 mesh that covers region contain the group
# nx ny nz region periodicity
# create a uniform 12 x 2 x 2 mesh that covers region contain the group
# nx ny nz region periodicity
fix_modify AtC mesh create 12 2 2 mdRegion f p p :pre
# specify the control method for the type of coupling
# physics control_type
# specify the control method for the type of coupling
# physics control_type
fix_modify AtC thermal control flux :pre
# specify the initial values for the empirical field "temperature"
# field node_group value
# specify the initial values for the empirical field "temperature"
# field node_group value
fix_modify AtC initial temperature all 30 :pre
# create an output stream for nodal fields
# filename output_frequency
# create an output stream for nodal fields
# filename output_frequency
fix_modify AtC output atc_fe_output 100 :pre
run 1000 :pre
likewise for this post-processing example:
run 1000 :pre
likewise for this post-processing example:
# ... commands to create and initialize the MD system :pre
# initial fix to designate post-processing and the group to apply it to
# no material file is allowed nor required
# initial fix to designate post-processing and the group to apply it to
# no material file is allowed nor required
fix AtC internal atc hardy :pre
# for hardy fix, specific kernel function (function type and range) to # be used as a localization function
fix AtC kernel quartic_sphere 10.0 :pre
# create a uniform 1 x 1 x 1 mesh that covers region contain the group
# with periodicity this effectively creats a system average
fix_modify AtC mesh create 1 1 1 box p p p :pre
# change from default lagrangian map to eulerian
# refreshed every 100 steps
# for hardy fix, specific kernel function (function type and range) to # be used as a localization function
fix AtC kernel quartic_sphere 10.0 :pre
# create a uniform 1 x 1 x 1 mesh that covers region contain the group
# with periodicity this effectively creats a system average
fix_modify AtC mesh create 1 1 1 box p p p :pre
# change from default lagrangian map to eulerian
# refreshed every 100 steps
fix_modify AtC atom_element_map eulerian 100 :pre
# start with no field defined
# add mass density, potential energy density, stress and temperature
# start with no field defined
# add mass density, potential energy density, stress and temperature
fix_modify AtC fields add density energy stress temperature :pre
# create an output stream for nodal fields
# filename output_frequency
# create an output stream for nodal fields
# filename output_frequency
fix_modify AtC output nvtFE 100 text :pre
run 1000 :pre
the mesh's linear interpolation functions can be used as the localization function
by using the field option:
fix AtC internal atc field
fix_modify AtC mesh create 1 1 1 box p p p
the mesh's linear interpolation functions can be used as the localization function
by using the field option:
...
fix AtC internal atc field
Note coupling and post-processing can be combined in the same simulations using separate fixes.
fix_modify AtC mesh create 1 1 1 box p p p
...
Note coupling and post-processing can be combined in the same simulations using separate fixes.
:line
@ -108,17 +108,17 @@ No information about this fix is written to "binary restart files"_restart.html.
[Restrictions:]
Thermal and two_temperature (coupling) types use a Verlet time-integration algorithm. The hardy type does not contain its own time-integrator and must be used with a separate fix that does contain one, e.g. nve, nvt, etc.
Thermal and two_temperature (coupling) types use a Verlet time-integration algorithm. The hardy type does not contain its own time-integrator and must be used with a separate fix that does contain one, e.g. nve, nvt, etc.
Currently,
- the coupling is restricted to thermal physics
Currently,
- the coupling is restricted to thermal physics
- the FE computations are done in serial on each processor. :ul
[Related commands:]
After specifying this fix in your input script, several other "fix_modify"_fix_modify.html commands are used to setup the problem, e.g. define the finite element mesh and prescribe initial and boundary conditions.
fix_modify commands for setup:
fix_modify commands for setup:
"fix_modify AtC mesh create"_USER/atc/man_mesh_create.html
"fix_modify AtC mesh quadrature"_USER/atc/man_mesh_quadrature.html
@ -149,7 +149,7 @@ fix_modify commands for boundary and initial conditions:
"fix_modify AtC source"_USER/atc/man_source.html
"fix_modify AtC remove_source"_USER/atc/man_remove_source.html :ul
fix_modify commands for control and filtering:
fix_modify commands for control and filtering:
"fix_modify AtC control"_USER/atc/man_control.html
"fix_modify AtC control thermal"_USER/atc/man_control_thermal.html
@ -165,7 +165,7 @@ fix_modify commands for control and filtering:
"fix_modify AtC extrinsic exchange"_USER/atc/man_extrinsic_exchange.html
"fix_modify AtC poisson_solver"_USER/atc/man_poisson_solver.html :ul
fix_modify commands for output:
fix_modify commands for output:
"fix_modify AtC output"_USER/atc/man_output.html
"fix_modify AtC output nodeset"_USER/atc/man_output_nodeset.html
@ -176,7 +176,7 @@ fix_modify commands for output:
"fix_modify AtC write_restart"_USER/atc/man_write_restart.html
"fix_modify AtC read_restart"_USER/atc/man_read_restart.html :ul
fix_modify commands for post-processing:
fix_modify commands for post-processing:
"fix_modify AtC kernel"_USER/atc/man_hardy_kernel.html
"fix_modify AtC fields"_USER/atc/man_hardy_fields.html
@ -188,7 +188,7 @@ fix_modify commands for post-processing:
"fix_modify AtC sample_frequency"_USER/atc/man_sample_frequency.html
"fix_modify AtC set"_USER/atc/man_set.html :ul
miscellaneous fix_modify commands:
miscellaneous fix_modify commands:
"fix_modify AtC atom_element_map"_USER/atc/man_atom_element_map.html
"fix_modify AtC atom_weight"_USER/atc/man_atom_weight.html
@ -210,35 +210,35 @@ miscellaneous fix_modify commands:
"fix_modify AtC remove_species"_USER/atc/man_remove_species.html
"fix_modify AtC remove_molecule"_USER/atc/man_remove_molecule.html :ul
Note: a set of example input files with the attendant material files are included with this package
Note: a set of example input files with the attendant material files are included with this package
[Default:]
None
None
:line
For detailed exposition of the theory and algorithms please see:
:link(Wagner)
[(Wagner)] Wagner, GJ; Jones, RE; Templeton, JA; Parks, MA, "An atomistic-to-continuum coupling method for heat transfer in solids." Special Issue of Computer Methods and Applied Mechanics (2008) 197:3351.
[(Wagner)] Wagner, GJ; Jones, RE; Templeton, JA; Parks, MA, "An atomistic-to-continuum coupling method for heat transfer in solids." Special Issue of Computer Methods and Applied Mechanics (2008) 197:3351.
:link(Zimmeman2004)
[(Zimmerman2004)] Zimmerman, JA; Webb, EB; Hoyt, JJ;. Jones, RE; Klein, PA; Bammann, DJ, "Calculation of stress in atomistic simulation." Special Issue of Modelling and Simulation in Materials Science and Engineering (2004), 12:S319.
[(Zimmerman2004)] Zimmerman, JA; Webb, EB; Hoyt, JJ;. Jones, RE; Klein, PA; Bammann, DJ, "Calculation of stress in atomistic simulation." Special Issue of Modelling and Simulation in Materials Science and Engineering (2004), 12:S319.
:link(Zimmerman2010)
[(Zimmerman2010)] Zimmerman, JA; Jones, RE; Templeton, JA, "A material frame approach for evaluating continuum variables in atomistic simulations." Journal of Computational Physics (2010), 229:2364.
[(Zimmerman2010)] Zimmerman, JA; Jones, RE; Templeton, JA, "A material frame approach for evaluating continuum variables in atomistic simulations." Journal of Computational Physics (2010), 229:2364.
:link(Templeton2010)
[(Templeton2010)] Templeton, JA; Jones, RE; Wagner, GJ, "Application of a field-based method to spatially varying thermal transport problems in molecular dynamics." Modelling and Simulation in Materials Science and Engineering (2010), 18:085007.
[(Templeton2010)] Templeton, JA; Jones, RE; Wagner, GJ, "Application of a field-based method to spatially varying thermal transport problems in molecular dynamics." Modelling and Simulation in Materials Science and Engineering (2010), 18:085007.
:link(Jones)
[(Jones)] Jones, RE; Templeton, JA; Wagner, GJ; Olmsted, D; Modine, JA, "Electron transport enhanced molecular dynamics for metals and semi-metals." International Journal for Numerical Methods in Engineering (2010), 83:940.
[(Jones)] Jones, RE; Templeton, JA; Wagner, GJ; Olmsted, D; Modine, JA, "Electron transport enhanced molecular dynamics for metals and semi-metals." International Journal for Numerical Methods in Engineering (2010), 83:940.
:link(Templeton2011)
[(Templeton2011)] Templeton, JA; Jones, RE; Lee, JW; Zimmerman, JA; Wong, BM, "A long-range electric field solver for molecular dynamics based on atomistic-to-continuum modeling." Journal of Chemical Theory and Computation (2011), 7:1736.
[(Templeton2011)] Templeton, JA; Jones, RE; Lee, JW; Zimmerman, JA; Wong, BM, "A long-range electric field solver for molecular dynamics based on atomistic-to-continuum modeling." Journal of Chemical Theory and Computation (2011), 7:1736.
:link(Mandadapu)
[(Mandadapu)] Mandadapu, KK; Templeton, JA; Lee, JW, "Polarization as a field variable from molecular dynamics simulations." Journal of Chemical Physics (2013), 139:054115.
[(Mandadapu)] Mandadapu, KK; Templeton, JA; Lee, JW, "Polarization as a field variable from molecular dynamics simulations." Journal of Chemical Physics (2013), 139:054115.
Please refer to the standard finite element (FE) texts, e.g. T.J.R Hughes " The finite element method ", Dover 2003, for the basics of FE simulation.

View File

@ -35,7 +35,7 @@ keyword = {types} or {mu} or {ke} or {semi-grand} or {region} :l
[Examples:]
fix 2 all atom/swap 1 1 29494 300.0 ke no types 1 2
fix myFix all atom/swap 100 1 12345 298.0 region my_swap_region types 5 6
fix myFix all atom/swap 100 1 12345 298.0 region my_swap_region types 5 6
fix SGMC all atom/swap 1 100 345 1.0 semi-grand yes types 1 2 3 mu 0.0 4.3 -5.0 :pre
[Description:]
@ -78,15 +78,15 @@ atoms remains constant. Whereas in the semi-grand canonical ensemble,
the composition of the system can change. Note that when using
{semi-grand}, atoms in the fix group whose type is not listed
in the {types} keyword are ineligible for attempted
conversion. An attempt is made to switch
the selected atom (if eligible) to one of the other listed types
conversion. An attempt is made to switch
the selected atom (if eligible) to one of the other listed types
with equal probability. Acceptance of each attempt depends upon the Metropolis criterion.
The {mu} keyword allows users to specify chemical
potentials. This is required and allowed only when using {semi-grand}.
All chemical potentials are absolute, so there is one for
each swap type listed following the {types} keyword.
In semi-grand canonical ensemble simulations the chemical composition
potentials. This is required and allowed only when using {semi-grand}.
All chemical potentials are absolute, so there is one for
each swap type listed following the {types} keyword.
In semi-grand canonical ensemble simulations the chemical composition
of the system is controlled by the difference in these values. So
shifting all values by a constant amount will have no effect
on the simulation.
@ -103,17 +103,17 @@ LAMMPS will warn you if any of the atoms eligible for swapping have a
non-zero molecule ID, but does not check for this at the time of
swapping.
If not using {semi-grand} this fix checks to ensure all atoms of the
given types have the same atomic charge. LAMMPS doesn't enforce this
in general, but it is needed for this fix to simplify the
swapping procedure. Successful swaps will swap the atom type and charge
If not using {semi-grand} this fix checks to ensure all atoms of the
given types have the same atomic charge. LAMMPS doesn't enforce this
in general, but it is needed for this fix to simplify the
swapping procedure. Successful swaps will swap the atom type and charge
of the swapped atoms. Conversely, when using {semi-grand}, it is assumed that all the atom
types involved in switches have the same charge. Otherwise, charge
would not be conserved. As a consequence, no checks on atomic charges are
performed, and successful switches update the atom type but not the
atom charge. While it is possible to use {semi-grand} with groups of
atom charge. While it is possible to use {semi-grand} with groups of
atoms that have different charges, these charges will not be changed when the
atom types change.
atom types change.
Since this fix computes total potential energies before and after
proposed swaps, so even complicated potential energy calculations are
@ -123,15 +123,15 @@ OK, including the following:
many body pair styles
hybrid pair styles
eam pair styles
triclinic systems
triclinic systems
need to include potential energy contributions from other fixes :ul
Some fixes have an associated potential energy. Examples of such fixes
include: "efield"_fix_efield.html, "gravity"_fix_gravity.html,
"addforce"_fix_addforce.html, "langevin"_fix_langevin.html,
"restrain"_fix_restrain.html, "temp/berendsen"_fix_temp_berendsen.html,
"temp/rescale"_fix_temp_rescale.html, and "wall fixes"_fix_wall.html.
For that energy to be included in the total potential energy of the
include: "efield"_fix_efield.html, "gravity"_fix_gravity.html,
"addforce"_fix_addforce.html, "langevin"_fix_langevin.html,
"restrain"_fix_restrain.html, "temp/berendsen"_fix_temp_berendsen.html,
"temp/rescale"_fix_temp_rescale.html, and "wall fixes"_fix_wall.html.
For that energy to be included in the total potential energy of the
system (the quantity used when performing GCMC moves),
you MUST enable the "fix_modify"_fix_modify.html {energy} option for
that fix. The doc pages for individual "fix"_fix.html commands
@ -170,18 +170,18 @@ LAMMPS"_Section_start.html#start_3 section for more info.
[Related commands:]
"fix nvt"_fix_nh.html, "neighbor"_neighbor.html,
"fix nvt"_fix_nh.html, "neighbor"_neighbor.html,
"fix deposit"_fix_deposit.html, "fix evaporate"_fix_evaporate.html,
"delete_atoms"_delete_atoms.html, "fix gcmc"_fix_gcmc.html
[Default:]
The option defaults are ke = yes, semi-grand = no, mu = 0.0 for
The option defaults are ke = yes, semi-grand = no, mu = 0.0 for
all atom types.
:line
:link(Sadigh)
[(Sadigh)] B Sadigh, P Erhart, A Stukowski, A Caro, E Martinez, and
[(Sadigh)] B Sadigh, P Erhart, A Stukowski, A Caro, E Martinez, and
L Zepeda-Ruiz, Phys. Rev. B, 85, 184203 (2012).

Some files were not shown because too many files have changed in this diff Show More