Compare commits

..

1271 Commits

Author SHA1 Message Date
b9b044e180 new fix ti/spring command, remove fix ti/rs 2016-09-20 10:44:12 -06:00
7dc8746f9e Merge pull request #39 from akohlmey/small-bugfixes
Collected small changes and bugfixes
2016-09-20 08:51:42 -06:00
5d89493a10 Merge pull request #38 from akohlmey/consolidate-pycache
consolidate ignoring python bytecompiled cache files
2016-09-20 08:49:38 -06:00
7bb880f0a8 remove debug code
(cherry picked from commit 1dbd52a30c)
2016-09-20 09:06:03 -04:00
849ff25d92 demote OMP_NUM_THREAD "warning" to "message" in the hope to be less confusing for inexperienced users 2016-09-20 08:57:04 -04:00
faa0b401aa fix argument indexing bug in fix ave/atom
(cherry picked from commit bc11518960)
2016-09-19 23:24:37 -04:00
aa9fe38c5c consolidate ignoring python bytecompiled cache files 2016-09-17 09:50:49 -04:00
d17421eb7c Merge pull request #36 from akohlmey/small_bugfixes_and_cleanups
Small bugfixes and cleanups
2016-09-16 13:19:39 -06:00
60dfdbc063 Merge pull request #35 from akohlmey/lj_sf_dipole_sf_correction
corrections for pair style lj/sf/dipole/sf and its /omp variant
2016-09-16 13:19:23 -06:00
e4bd63759b Merge pull request #34 from rbberger/make_py_fixes
Update Make.py
2016-09-16 13:19:05 -06:00
ca558f6712 Merge pull request #33 from rbberger/pylammps_additions
PyLammps additions
2016-09-16 13:18:17 -06:00
abf05eed61 comm tiled bug fix 2016-09-16 10:31:52 -06:00
72ce8ff89f Merge pull request #31 from akohlmey/doc-cleanups-and-updates
More documentation cleanups and updates
2016-09-16 10:05:53 -06:00
9637a5b530 colvars bugfix. updates colvars library to version 2016-09-14 2016-09-15 19:44:26 -04:00
4149413057 reduce global name space pollution in pair style meam 2016-09-15 16:32:41 -04:00
400ef87c05 make certain, that -log() doesn't overflow and assign a suitable maximum number instead. 2016-09-15 14:55:32 -04:00
e9e9790d6e corrections for pair style lj/sf/dipole/sf and its /omp variant
(cherry picked from commit f0c8b2af28a58485f6795cf85c7d88eaafa3e52a)
2016-09-15 09:30:40 -04:00
319b160752 Add exceptions setting to help message of Make.py 2016-09-14 19:26:03 -04:00
cddc1dbb11 Added exceptions flag to Make.py 2016-09-14 19:21:52 -04:00
2831f50790 Merge branch 'integration' into doc-cleanups-and-updates 2016-09-14 19:02:45 -04:00
62bf307d3c Fix bug introduced into Make.py by latest changes 2016-09-14 18:15:50 -04:00
5cdc48dd0c Update ipython example to show interface usage 2016-09-14 16:26:19 -04:00
0ec8fa02e0 Make thermo output accessable through PyLammps
Thermo data of the last run is now accessable through the `last_run.thermo`
property. This is a dictionary containing the data columns of thermo output.
All run data is kept as list and can be found in the `runs` property.
See issue #144
2016-09-14 16:14:41 -04:00
2fb666dc69 Merge branch 'doc' into integration 2016-09-14 13:32:39 -06:00
6e3705f380 Allow writing input script from Python session
This implements the requested feature in issue #145. The `write_script`
method now gives you a way of dumping out all used commands into a
LAMMPS input script file.

Note: this also dumps all commands which are indirectly issued by PyLammps
2016-09-14 14:24:19 -04:00
2f298951cf resolved conflicts 2016-09-14 10:22:46 -06:00
717e719b83 HTML update 2016-09-14 09:35:03 -06:00
523c70e0be regenerate html files one more time 2016-09-14 10:01:13 -04:00
77e0a84877 fix an inconsistency and a typo 2016-09-14 09:59:36 -04:00
9779911cea regenerate html files from scratch after merging from upstream and including richard's last updates to the converter tools 2016-09-14 09:54:06 -04:00
1ad15b8711 Merge branch 'integration' into doc-cleanups-and-updates 2016-09-14 09:48:46 -04:00
7025a3f5d1 Merge pull request #32 from rbberger/doc-cleanups-and-updates
Various updates to tools and some minor documentation fixes
2016-09-14 09:46:05 -04:00
df304f8ca1 Updated HTML after latest tool changes and corrections 2016-09-13 20:22:17 -04:00
3c88fa1436 Fix bug introduced in latest changes 2016-09-13 20:18:36 -04:00
b7ddc860c7 Fix error in Manual.txt 2016-09-13 20:00:17 -04:00
c61d5a1a29 Raise exception and output error if ulb,ule and olb,ole are unbalanced 2016-09-13 19:53:57 -04:00
10b4411d5b Merge branch 'integration' of github.com:akohlmey/lammps into integration 2016-09-13 17:22:14 -06:00
c744b23c4c doc updates 2016-09-13 17:15:15 -06:00
a69e059be3 Merge pull request #27 from akohlmey/small-bugfixes
Collected small changes and bugfixes
2016-09-13 17:11:23 -06:00
dbc548dd88 Merge pull request #26 from rbberger/is_available_feature
Extend is_available() function to query optional features
2016-09-13 17:10:10 -06:00
1dc19eceb2 Update unit tests to reflect new behavior 2016-09-13 19:03:57 -04:00
ae6b540d3c Add indentation to multi-paragraph entries 2016-09-13 19:03:46 -04:00
25e518a4f4 added options to compute bond/local 2016-09-13 17:03:37 -06:00
fe2fca4e9b clean-up of example files 2016-09-13 14:58:02 -06:00
ed52f9ea5e pair vashishta/table 2016-09-13 14:54:12 -06:00
944289b018 final version of pair vashishta/table 2016-09-13 14:45:53 -06:00
80c5b01bfc flag 'allocated' must be initialized in constructor 2016-09-13 14:33:00 -04:00
51e4a568c9 Merge branch 'small-bugfixes' of github.com:akohlmey/lammps into small-bugfixes 2016-09-13 14:32:27 -04:00
300d1ef52e Remove references to former empty file 2016-09-13 14:17:27 -04:00
633840c876 remove empty file 2016-09-13 13:46:45 -04:00
c44228b0cc Created more complete TOC and indices
This adds index sections on the main navigation bar and finally silences the
missing TOC warnings.
2016-09-13 13:40:18 -04:00
90f6395ddc Filter out more header numbers 2016-09-13 13:31:30 -04:00
a8081d4507 fix qeq Kokkos bug fix by Stan 2016-09-13 10:54:04 -06:00
14bed44743 ignore Makefile.lammps in lib tree 2016-09-13 09:36:48 -04:00
18cacb8e1d ignore *.mod generated by fortran compilers 2016-09-13 09:36:19 -04:00
546582ea02 Escape pipe characters 2016-09-12 23:18:38 -04:00
b76a42d3e0 making a few more in-page links unique. some more small corrections and clenups 2016-09-12 19:57:37 -04:00
54d5a14fe3 more doc formatting issues triggering errors/warnings in sphinx 2016-09-12 14:48:47 -04:00
f6efde3730 some more smaller doc file cleanups. almost there... 2016-09-12 13:37:08 -04:00
4c399fc553 more typesetting corrections in docs indicated by sphinx warnings 2016-09-12 12:23:53 -04:00
328b7abeaa correct various formatting issues flagged by sphinx 2016-09-11 23:24:32 -04:00
c3de3c142f correct references to Howto subsections 2016-09-11 23:01:16 -04:00
80f94c7d02 avoid bad typesetting of include images 2016-09-11 20:09:46 -04:00
e11bfcf117 updates in "Getting Started" for running on Windows. 2016-09-11 20:09:25 -04:00
be1cf40f2b update html files with recent changes 2016-09-11 19:34:04 -04:00
555a02786d correct and improve typesetting for various sub-sub-subsections in "Getting started" 2016-09-11 19:28:56 -04:00
cf6f504977 update description of Windows support in "Getting started" Section 2016-09-11 19:07:45 -04:00
b698f389bc correct and improve links to PyMol and MATLAB 2016-09-11 18:46:06 -04:00
e53862ca4a update html file updated boundary.txt file 2016-09-11 14:55:13 -04:00
a64eb330e3 document how using shrink-wrap boundaries can cause lost atoms in parallel 2016-09-11 14:48:37 -04:00
e96a8a4677 regenerate html pages with the updated content from the corrections 2016-09-10 21:15:44 -04:00
f8d5488409 correct section link descriptions 2016-09-10 21:03:38 -04:00
4d298ccf2f Give credit to Richard for being the lammps.org and senior jenkins wrangler 2016-09-10 20:18:05 -04:00
cb3044091c place ellipses in square brackets to avoid sphinx choking on them 2016-09-10 20:05:50 -04:00
d70e051ecd update docs for lj/sf/dipole/sf 2016-09-09 10:00:15 -04:00
37833b537b restore c++ marker in header comment 2016-09-09 09:43:21 -04:00
5fcbfa8248 simplify phonon code some more
(cherry picked from commit 1c54dc77ea)
2016-09-09 09:40:52 -04:00
c437195928 simplify code
(cherry picked from commit 3575913379)
2016-09-09 09:40:52 -04:00
8b1ef1c686 import updated pair lj/sf/dipole/sf files from sam genheden
(cherry picked from commit ae691ab786)
2016-09-09 09:37:17 -04:00
c3e8cb2f30 fix typo in comments
(cherry picked from commit 7875009218)
2016-09-09 09:34:17 -04:00
365707704c update error message to be consistent with other styles 2016-09-09 09:19:15 -04:00
16323ba391 require an atom map for compute voronoi/atom occupation
(cherry picked from commit 2b53f80392)
2016-09-09 07:39:09 -04:00
e27869daf6 Add updated HTML documentation about new is_available feature category 2016-09-09 05:56:24 -04:00
dc0c0ab214 Add documentation about new is_available feature category 2016-09-09 05:55:28 -04:00
4b22443b25 Add feature category to is_available function
This allows checking if the LAMMPS binary/library was compiled with PNG, JPEG,
FFMPEG, GZIP, or exceptions support.

Usage:
```
is_available(feature,gzip)
is_available(feature,png)
is_available(feature,jpeg)
is_available(feature,ffmpeg)
is_available(feature,exceptions)
```
2016-09-09 05:09:45 -04:00
956af8cebb update html file 2016-09-08 17:41:46 -04:00
5c927ca839 Merge remote-tracking branch 'lammps-rw/integration' into vashishta-tabulation 2016-09-08 17:22:21 -04:00
4bb42be3cc implement vashishta/table/omp in USER-OMP 2016-09-08 17:22:03 -04:00
7de5143050 update vashishta pair style example 2016-09-08 17:21:14 -04:00
71eed1d612 update documentation for vashishta pair style 2016-09-08 17:20:52 -04:00
dd34feb2bd Merge branch 'small' into integration 2016-09-08 14:35:36 -06:00
2524c5b526 Merge branch 'small-bugfixes' of https://github.com/akohlmey/lammps into small 2016-09-08 14:31:27 -06:00
fe581e8ced Merge remote-tracking branch 'lammps-rw/integration' into vashishta-tabulation 2016-09-08 16:21:05 -04:00
b866e0663b Merge pull request #24 from rbberger/info_command_styles
Info command extensions
2016-09-08 14:16:41 -06:00
5d0da95a0b Merge pull request #23 from akohlmey/doc-corrections-and-updates
Doc corrections and updates
2016-09-08 14:15:56 -06:00
07e55ef61e Merge pull request #22 from akohlmey/colvars-update
accumulated colvars library update and lib "make clean" improvement
2016-09-08 14:14:49 -06:00
236ebf7fab Kokkos lib update 2016-09-08 13:56:18 -06:00
a6df1e53b4 ignore all variants of the vashishta pair styles in src/ 2016-09-08 14:46:09 -04:00
9b2d5ff3e7 remove executable permissions on non-executables and non-scripts 2016-09-08 14:45:07 -04:00
c33e1049d8 delete duplicate fclose() in fix ttm/mod constructor 2016-09-08 11:51:22 -04:00
1f901c9b2d Update generated HTML docs of info command 2016-09-08 09:53:57 -04:00
79b8f6320d Update docs of info command 2016-09-08 09:53:14 -04:00
2dcfb51d18 Change info command styles syntax
Change to `info styles [all|atom|pair|compute|...]`.
By default, `info styles` will print all styles.
2016-09-08 09:38:52 -04:00
ba2b523bf4 Use factory for region style creation 2016-09-08 09:38:51 -04:00
fd2b886422 Use factory for dump style creation 2016-09-08 09:38:51 -04:00
9952d8a210 Use factory for minimize style creation 2016-09-08 09:38:51 -04:00
85c132943e Use factory for integrate style creation 2016-09-08 09:38:48 -04:00
55260ad53e Add missing delete in destructor 2016-09-08 09:32:48 -04:00
a1e5fc0fca Use factory for AtomVec style creation 2016-09-08 09:32:48 -04:00
88e10b401d Cleanup force.h after refactoring 2016-09-08 09:32:48 -04:00
1d03913aa3 Use factory for kspace style creation 2016-09-08 09:32:48 -04:00
0745a9f33f Use factory for improper style creation 2016-09-08 09:32:48 -04:00
906c50223a Use factory for dihedral style creation 2016-09-08 09:32:47 -04:00
35bdeb63e2 Use factory for angle style creation 2016-09-08 09:32:47 -04:00
69c58ef0d5 Use factory for bond style creation 2016-09-08 09:32:47 -04:00
95ee6440ad Use pair_map to list pair styles 2016-09-08 09:32:47 -04:00
00b08bb5e1 Use compute_map and fix_map to list compute and fix styles 2016-09-08 09:32:47 -04:00
e483cb9ef9 Use command_map to list command styles 2016-09-08 09:32:47 -04:00
06e3a11c2d Add styles output to info command
Adds the ability to list all available styles in LAMMPS with:

```
info styles
```

Each style can also be printed separately using one of the following:

```
info atom_styles
info integrate_styles
info minimize_styles
info pair_styles
info bond_styles
info angle_styles
info dihedral_styles
info improper_styles
info kspace_styles
info fix_styles
info compute_styles
info region_styles
info dump_styles
```
2016-09-08 09:32:46 -04:00
7e8440cbab undo changes to vashishita/omp 2016-09-07 12:02:38 -06:00
43b05a60c7 created vashishta/table variant 2016-09-07 11:56:38 -06:00
0fe7d1d361 correct internal links for Modify LAMMPS section 2016-09-07 10:13:34 -04:00
346ff42498 correct reference in modify section 2016-09-07 10:10:12 -04:00
5feedbd829 regenerate html pages with new txt2rst tool 2016-09-07 10:09:53 -04:00
44ce6fac4b replace backquote with regular quote and `` + '' with double quote. 2016-09-07 08:29:16 -04:00
70d6718aa3 Update discussion on parallel python packages. There seem to be only two left. 2016-09-07 07:42:06 -04:00
348b677148 Make formatting and links more consistent and unique 2016-09-07 07:41:17 -04:00
4c783ea3b7 Enforce l,ule or l,ole command order for RST
(cherry picked from commit 79e867c213)
2016-09-07 02:16:55 -04:00
9e8256aeb0 Refactored code and escape RST special character '_'
(cherry picked from commit 4629a464f7)
2016-09-07 02:16:55 -04:00
925f1bfb6f Escape RST special character '^' and take care of special math cases
(cherry picked from commit 453521c8e6)
2016-09-07 02:16:55 -04:00
3f312244a0 Escape RST special character '*' in final output
(cherry picked from commit 7cb39811d4)
2016-09-07 02:16:54 -04:00
55022d1263 replace :l,ule and :l,ole with :l :ule or :l :ole 2016-09-07 02:11:59 -04:00
0d491d483c avoid Inline substitution_reference start-string without end-string 2016-09-07 02:10:49 -04:00
a31c507370 correct error message text, sync with source 2016-09-06 22:15:49 -04:00
3a74ccffa2 update colvars library to version 2016-09-03
(cherry picked from commit 4181f5ac9d)
2016-09-06 21:28:59 -04:00
c8cfd53c1b adjust makefiles in lib tree, so they don't return an error exit code on "make clean" targets.
(cherry picked from commit 4117218ca3)
2016-09-06 21:28:55 -04:00
16607a0132 update colvars library to version 2016-08-19
(cherry picked from commit c67326be6d)
2016-09-06 21:27:46 -04:00
3b476d914f update colvars to version 2016-08-10
(cherry picked from commit eba3ad9abb)
2016-09-06 21:27:46 -04:00
977b9e542f update colvars to another set of changes from the upstream repo
(cherry picked from commit cb816f8cba)
2016-09-06 21:27:45 -04:00
1b33d00785 update colvars makefile dependencies
(cherry picked from commit df99a85930)
2016-09-06 21:27:45 -04:00
3d2e5d0a50 suppress compiler warning
(cherry picked from commit 6f227e194e)
2016-09-06 21:27:45 -04:00
ec2a6b9f0d update colvars to version 2016-08-10
(cherry picked from commit f2ddf828e4)
2016-09-06 21:27:41 -04:00
77620106a4 update colvars library to version 2016-08-05
(cherry picked from commit 459db2eb6b)

# Conflicts:
#	doc/src/PDF/colvars-refman-lammps.pdf
2016-09-06 21:26:50 -04:00
f56c41eec0 update with correct error description. sync with manual. 2016-09-06 21:21:28 -04:00
fc2d878305 update manual introduction to represent current status. correct some markup to better pass through sphinx 2016-09-06 21:20:47 -04:00
1c17b98500 Merge branch 'integration' into doc-corrections-and-updates 2016-09-06 20:19:02 -04:00
9138152563 Merge branch 'integration' into small-bugfixes 2016-09-06 20:17:07 -04:00
ace5dc3c7c Merge remote-tracking branch 'lammps-rw/clean-up-docs-for-sphinx' into clean-up-docs-for-sphinx 2016-09-06 19:54:35 -04:00
0252347d43 sync with 7Sep16 patch 2016-09-06 17:00:32 -06:00
c9455c90de Merge pull request #20 from akohlmey/lammps-contribute
Update description of submitting contributions to LAMMPS
2016-09-06 16:55:16 -06:00
1e4d6fee93 Merge pull request #19 from akohlmey/clean-up-docs-for-sphinx
Clean up docs for sphinx (redo)
2016-09-06 16:54:39 -06:00
42db93e198 one more small correction to document recent fix ave/time changes 2016-09-06 16:10:16 -04:00
906bd24543 fix typos 2016-09-06 16:03:16 -04:00
4f88c75401 update the discussion on how to submit modifications 2016-09-06 16:02:59 -04:00
4314299be9 harden fix shear/history against use-after-delete and from incorrect use of reverse communication 2016-09-06 14:50:04 -04:00
1a7b04e8a6 generate updated html files for corrected .txt files
(cherry picked from commit 7de57ffd94)
2016-09-06 13:40:11 -04:00
fbc955e549 correct link targets
(cherry picked from commit ff75cf51bb)
2016-09-06 13:40:11 -04:00
3bb3c1a45c add pppm/kk and fix reax/c issues 2016-09-06 10:46:51 -06:00
c543cba95c Merge pull request #18 from akohlmey/clean-up-docs-for-sphinx
Clean up incorrect link targets in documentation
2016-09-06 10:31:33 -06:00
b12ad2cecf Merge branch 'integration' into vashishta-tabulation 2016-09-01 08:25:31 -04:00
431d1a6dae mention stable releases 2016-09-01 02:37:55 -04:00
ab84acc2cd synchronize LAMMPS developer list with webpage and move ray to "past" 2016-09-01 02:26:49 -04:00
fc093a0aab replace :ule,l and :ole,l with :l<br>:ule and :l<br>:ole 2016-09-01 02:07:16 -04:00
5e6dff36e4 add new commands to lammps.book 2016-09-01 02:01:47 -04:00
7de57ffd94 generate updated html files for corrected .txt files 2016-09-01 00:02:40 -04:00
ff75cf51bb correct link targets 2016-09-01 00:01:26 -04:00
fb2c18ee88 Merge pull request #17 from akohlmey/small-doc-fixes
corrections for various compute something/chunk examples
2016-08-31 19:57:36 -06:00
b5c758f22c Merge branch 'goo' into integration 2016-08-31 19:53:50 -06:00
de0036fafd removing searchindex.js 2016-08-31 19:52:28 -06:00
c3c9788dc7 include autogenerated html code 2016-08-31 21:49:06 -04:00
2abd5ad28a Merge branch 'integration' into small-doc-fixes 2016-08-31 21:42:22 -04:00
1c3302d1db Merge pull request #15 from rbberger/pylammps-improvements
PyLammps improvements
2016-08-31 16:34:59 -06:00
24409b6178 Merge branch 'error' into integration 2016-08-31 16:16:47 -06:00
de21cb2cd5 small changes to doc page 2016-08-31 16:16:43 -06:00
639ab0fd3e Merge branch 'core/cpp_exceptions' of https://github.com/rbberger/lammps into error 2016-08-31 16:11:08 -06:00
6c65af710c bad file? 2016-08-31 16:10:20 -06:00
29e480ad66 corrections for various compute something/chunk examples 2016-08-31 17:31:14 -04:00
7c01ef57ee sync with SVN 2016-08-31 15:17:00 -06:00
0316bb579b Merge pull request #3 from akohlmey/fix-flow-gauss
Fix flow/gauss for USER-MISC package
2016-08-31 15:03:27 -06:00
f89448d73c Merge pull request #6 from akohlmey/manifold-doc-update
Small update for USER-MANIFOLD docs.
2016-08-31 14:53:16 -06:00
ad879d97db sync w/ SVN 2016-08-31 14:29:44 -06:00
4051aedf2c Merge branch 'small' into integration
Conflicts:
	src/compute_omega_chunk.cpp
2016-08-31 13:58:12 -06:00
ec8b9e21db sync with SVN 2016-08-31 13:28:26 -06:00
1986eda4d5 Remove no longer needed generated files in doc
If we use the Google Custom Search API, we do not need to keep the
generated searchindex.js file anymore. We also can safely remove
the _sources directory for good.

Since these get generated during each Sphinx build, additional
steps have been added to the Makefile to get rid of them. They
are also added to .gitignore to avoid commiting them by accident.
2016-08-31 00:36:56 -04:00
e71fafdd25 Use Google Custom Search API for searching in docs 2016-08-31 00:28:27 -04:00
6cbdad7a97 Add utility function to IPyLammps for embedding videos 2016-08-30 11:34:16 -04:00
a08cf7a4b6 Add verbose option in PyLammps methods
This option allows forcing LAMMPS output to be printed to the console.
2016-08-30 11:34:16 -04:00
691de01b33 Allow setting the position of atoms 2016-08-30 11:34:16 -04:00
33a87a470a Return 2D vectors in 2D cases 2016-08-30 11:34:15 -04:00
59dc83eadb Fix resource leak 2016-08-30 11:34:15 -04:00
a2ea263652 remove vector/array class members shadowing the base class 2016-08-30 07:04:54 -04:00
493613b495 avoid NaNs in MEAM 2016-08-30 06:54:36 -04:00
021ade199a null-ing of pointers and a couple select integers for fix ave/chunk 2016-08-30 06:41:50 -04:00
b7749ab212 individual computes do not need to set vector/array to NULL. reordering fixes to silence compiler warnings 2016-08-30 06:41:02 -04:00
554ac7dd12 Use MPI_Allreduce values instead of each processor's values
(cherry picked from commit f30232b41e)
2016-08-30 06:38:15 -04:00
ef86d11729 Merge branch 'integration' into small-bugfixes 2016-08-30 06:29:41 -04:00
62b7b69a87 Merge branch 'init' into integration 2016-08-29 17:45:56 -06:00
1c1c9c3101 use correct order when initializing arrays 2016-08-28 22:23:22 -04:00
48ba812f0a Fixed initialization of arrays in computes/fixes in MC package 2016-08-28 22:23:22 -04:00
f9a21ae654 Fixed initialization of arrays in computes/fixes in MANYBODY package 2016-08-28 22:23:22 -04:00
d6b9d0b9b6 Fixed initialization of arrays in fixes 2016-08-28 22:23:22 -04:00
36e085e393 Fixed initialization of arrays in fixes 2016-08-28 22:23:22 -04:00
425142ba2e Fixed initialization of arrays in fixes 2016-08-28 22:23:22 -04:00
07eb1d443b Fixed initialization of arrays in fixes 2016-08-28 22:23:22 -04:00
265cc14125 Fixed initialization of arrays in fixes 2016-08-28 22:23:22 -04:00
fd05a1325e Fixed initialization of arrays in fixes 2016-08-28 22:23:22 -04:00
b5a562788b Fixed initialization of arrays in fixes 2016-08-28 22:23:22 -04:00
2c7241bfe2 Fixed initialization of arrays in fixes 2016-08-28 22:23:22 -04:00
ee2f6ded29 Fixed initialization of arrays in fixes 2016-08-28 22:23:22 -04:00
db077ef186 Fixed initialization of arrays in fixes 2016-08-28 22:23:22 -04:00
fc5db8a737 fix off-by one bug when looking for bonds. 2016-08-28 14:20:43 -04:00
56d0ab9474 updated doc build readme for current Fedora and RHEL/CentOS 2016-08-28 14:20:23 -04:00
f8d6b979ec Merge branch 'integration' into small-bugfixes 2016-08-28 14:16:14 -04:00
4e03df2d19 Merge branch 'integration' into fix-flow-gauss 2016-08-28 06:55:12 -04:00
e1045851c0 incorporate bugfix and cleanup from lammps-icms and upstream 2016-08-28 06:52:03 -04:00
cdf06646ef Update documentation for tabulation in vashista pair style 2016-08-27 23:05:17 -04:00
490b3402a7 optimize twobody term by passing a const reference instead of a pointer 2016-08-27 23:05:17 -04:00
ebce76c7f0 updated and slightly refactored tabulation for vashishta pair style
- tables are now dimensioned by nelements instead of ntypes
- tables are only created if used
- correctly identify max size of table
- add test for illegal cutoff for tabulation
- allocated memory for tables is accounted for
- add example input using 16-bit tables
2016-08-27 23:05:17 -04:00
bf59c976f8 Added curly brackets and spaces for better code readability 2016-08-27 23:05:17 -04:00
06cc38e16c Fixed so tabulated pair_vashishta uses same pair_modify command style as other pair styles 2016-08-27 23:05:17 -04:00
10ec14f0fd Remembering to clean up memory with new arrays in vashishta. 2016-08-27 23:05:17 -04:00
82d9f5f5e6 Added 3-body neighbor list building for faster short range 3 body forces. 2016-08-27 23:05:17 -04:00
944ebdcf44 Added tabulated version of vashishta potential 2016-08-27 23:05:17 -04:00
f5a50c3cd1 Added documentation about -DLAMMPS_EXCEPTIONS flag 2016-08-27 22:13:36 -04:00
0192d2e359 Merge pull request #5 from akohlmey/initialize-pointers
enforce initializing pointers in constructors to NULL
2016-08-27 17:14:27 -06:00
3a1397dc7c sync with SVN 2016-08-27 17:11:16 -06:00
bb721db8de Merge branch 'integration' into fix-flow-gauss 2016-08-27 19:02:33 -04:00
0c2e643062 Merge pull request #9 from rbberger/remove_sha1sum_dependency
Remove sha1sum dependency for doc generation
2016-08-27 17:02:17 -06:00
ef69bf8695 Merge pull request #11 from akohlmey/python3-for-make-py
Python 3 support for Make.py with Python 2.7 compatibility
2016-08-27 17:01:29 -06:00
6a4633af0a Merge pull request #12 from akohlmey/shell-script-paranoia
Disable custom locale and grep options in shell scripts
2016-08-27 16:59:43 -06:00
c80dad0028 new fix flow/gauss command 2016-08-27 16:25:01 -06:00
1c13b30a70 small cleanup and generalization of fix flow/gauss
- remove unused or unneeded class members
- make the code compatible with per-atom masses
- test for and abend in case of an invalid group mass

(cherry picked from commit e017b33898)
2016-08-27 17:33:46 -04:00
c570bf26e0 enforce C or POSIX locale in all shell scripts and turn of enforced global grep options 2016-08-27 01:57:45 -04:00
742c853775 need to ignore src/Make.py.last as well 2016-08-26 17:48:59 -04:00
9932b73227 Merge branch 'integration' into python3-for-make-py 2016-08-26 17:48:06 -04:00
90272f6c71 some more tweaks needed to improve python2/3 compatibility 2016-08-26 17:46:57 -04:00
8dd42789f8 correct for futurize not being able to fully convert functionality of the commands module 2016-08-26 17:15:39 -04:00
a0592d1b64 Merge branch 'gitignore-improvements' into integration 2016-08-26 13:54:29 -06:00
9be235d872 improved .gitignore files to show only changes in relevant folders and files 2016-08-26 14:34:42 -04:00
2beecd1e73 removal of doc/html/_sources and minor sync with SVN 2016-08-26 11:34:37 -06:00
95aabdf51a Add MacOS X instructions to doc generation README 2016-08-26 10:08:22 -04:00
ea368919f3 Remove sha1sum dependency for doc generation
On MacOS X there is no sha1sum. So to simplify doc generation on those systems
use a Python script instead to generate a unique string from the repository
path.
2016-08-26 00:01:34 -04:00
74516b571e port Make.py so it is compatible with python 2.7 *and* python 3.x 2016-08-25 06:30:53 -04:00
b06fa5670a find functions in atom.cpp checks for id/name being NULL
(cherry picked from commit f7e741f344)
2016-08-25 06:00:58 -04:00
a635c70a26 Fixed so we test for compute,fix and variable id's being NULL in find_ functions
(cherry picked from commit 12e1857e30)
2016-08-25 06:00:58 -04:00
b8e7f53017 Small update for USER-MANIFOLD docs.
(cherry picked from commit d188bdbd86)
2016-08-24 23:36:04 -04:00
849cec3400 fix memory leak in compute omega/chunk
(cherry picked from commit e8d0342503)
2016-08-24 23:15:08 -04:00
a692398b6c fix bug in reserving sufficient space for special atoms
assigning atom->maxspecial will not work, since it will be reset, e.g. when reading from a data file that doesn't have any special neighbors.
instead we need to set force->special_extra so this is going to be preserved.
2016-08-24 23:08:03 -04:00
ff541e9a84 fix typo in compute omega/chunk usage example 2016-08-24 23:04:27 -04:00
7d43f349e6 Fixed initialization of arrays in computes 2016-08-24 17:26:04 -04:00
5e811f16e8 Fixed initialization of arrays in computes 2016-08-24 17:26:03 -04:00
fcd54f02e6 Fixed initialization of arrays in computes 2016-08-24 17:26:03 -04:00
1f3ef8e0ee Fixed initialization of arrays in computes 2016-08-24 17:26:03 -04:00
3e793d6eb7 Fixed initialization of arrays in computes 2016-08-24 17:26:03 -04:00
95dde5c041 Fixed initialization of arrays in computes 2016-08-24 17:26:02 -04:00
d09a85733b Fixed initialization of arrays in computes 2016-08-24 17:26:02 -04:00
0e7ce194eb Fixed initialization of arrays in computes 2016-08-24 17:26:02 -04:00
e5c37bc7cb Implemented optional C++ exceptions in Error class
These can be activated using the -DLAMMPS_EXCEPTIONS compiler flag.
It has no effect for regular execution. However, while using
it as a library, any issued command will capture the exception
and save its error message. This can be queried using the
lammps_has_error() and lammps_get_last_error_message() methods.

The Python wrapper checks these in order to rethrow these errors
as Python exceptions. See issue #146.

(cherry picked from commit 6c154bb0b67a13d38968bc42d31013b97f87db75)
2016-08-24 15:31:30 -04:00
e27196e91c doc files 2016-08-23 16:28:36 -06:00
268fdab71b Merge branch 'small-bugfixes' into integration 2016-08-23 15:39:27 -06:00
8750515cc4 changed dump.cpp back to the way it was 2016-08-23 15:38:38 -06:00
270b07b035 Merge branch 'integration' into small-bugfixes 2016-08-23 15:35:31 -06:00
abc5a32c8a gpu lib sync 2016-08-23 15:30:01 -06:00
0a3464eb30 test 2016-08-23 15:27:02 -06:00
1ab3891caf Merge branch 'merge-pull-153' into lammps-icms
Submitted by Steven E. Strong via github
Contributing authors: Steven E. Strong and Joel D. Eaves   Joel.Eaves@Colorado.edu

This branch implements Gaussian dynamics (GD), which is a method to do
nonequilibrium molecular dynamics simulations of steady-state flow. See
http://dx.doi.org/10.1021/acs.jpclett.6b00748. It is simple to implement
and derives rigorously from Gauss's principle of least constraint.

(cherry picked from commit 75929ee01b)
2016-08-23 15:33:16 -04:00
55fe1f6b29 fixed integer division problem in python code snippet
git-svn-id: svn://svn.icms.temple.edu/lammps-ro/trunk@15486 f3b2605a-c512-4ea7-a41b-209d697bcdaa
2016-08-22 22:35:23 +00:00
50a82bb345 address uninitialized variable issues pointed out by valgrind/coverity 2016-08-22 15:49:33 -04:00
5909bd5429 correct bug in tracking atom->nlocal vs. atom->nmax when allocating pbc enforcement buffers
(cherry picked from commit 45a2dd36d0)
2016-08-22 14:36:47 -04:00
1383684048 fix bug in recent dump changes for -DLAMMPS_BIGBIG
(cherry picked from commit a507936878)
2016-08-22 14:36:47 -04:00
587bafdf2d Remove memory leak and unnecessary allocation
(cherry picked from commit c998f7b81f)
2016-08-22 14:36:46 -04:00
c8fe3799ed Add missing initialization
(cherry picked from commit 054256cf0a)
2016-08-22 14:36:46 -04:00
9babb7a4c2 fix indexing bugs in accessing compute and fix labels in fix ave/histo
(cherry picked from commit 579c527718)
2016-08-22 14:35:32 -04:00
c88e9b46cf thread timing summare needs to be marked as preformatted
(cherry picked from commit b745636a67)
2016-08-22 14:35:11 -04:00
730e3cb4ac correct small (but harmless) logic error.
(cherry picked from commit ac6f4f8a56)
2016-08-22 14:34:43 -04:00
2a6561e52a add run 0 to USER-TALLY examples to enforce shake constraints on step 0
(cherry picked from commit 433741564d)
2016-08-22 14:32:47 -04:00
2fff78a78e git-svn-id: svn://svn.icms.temple.edu/lammps-ro/trunk@15484 f3b2605a-c512-4ea7-a41b-209d697bcdaa 2016-08-20 22:43:04 +00:00
d4891754c8 git-svn-id: svn://svn.icms.temple.edu/lammps-ro/trunk@15483 f3b2605a-c512-4ea7-a41b-209d697bcdaa 2016-08-20 22:42:24 +00:00
ccf8cf20b3 git-svn-id: svn://svn.icms.temple.edu/lammps-ro/trunk@15482 f3b2605a-c512-4ea7-a41b-209d697bcdaa 2016-08-20 22:42:17 +00:00
6ccf4b4525 git-svn-id: svn://svn.icms.temple.edu/lammps-ro/trunk@15481 f3b2605a-c512-4ea7-a41b-209d697bcdaa 2016-08-20 22:41:29 +00:00
bfba361f65 git-svn-id: svn://svn.icms.temple.edu/lammps-ro/trunk@15480 f3b2605a-c512-4ea7-a41b-209d697bcdaa 2016-08-20 22:22:00 +00:00
b1829c107c git-svn-id: svn://svn.icms.temple.edu/lammps-ro/trunk@15479 f3b2605a-c512-4ea7-a41b-209d697bcdaa 2016-08-13 15:03:16 +00:00
832f6a9f11 git-svn-id: svn://svn.icms.temple.edu/lammps-ro/trunk@15475 f3b2605a-c512-4ea7-a41b-209d697bcdaa 2016-08-13 14:52:15 +00:00
6503590875 git-svn-id: svn://svn.icms.temple.edu/lammps-ro/trunk@15474 f3b2605a-c512-4ea7-a41b-209d697bcdaa 2016-08-13 14:51:00 +00:00
2c9eef57d0 git-svn-id: svn://svn.icms.temple.edu/lammps-ro/trunk@15473 f3b2605a-c512-4ea7-a41b-209d697bcdaa 2016-08-13 14:49:57 +00:00
c03252f08d git-svn-id: svn://svn.icms.temple.edu/lammps-ro/trunk@15471 f3b2605a-c512-4ea7-a41b-209d697bcdaa 2016-08-12 14:34:16 +00:00
ff933712b7 git-svn-id: svn://svn.icms.temple.edu/lammps-ro/trunk@15470 f3b2605a-c512-4ea7-a41b-209d697bcdaa 2016-08-12 14:33:36 +00:00
112c98159b git-svn-id: svn://svn.icms.temple.edu/lammps-ro/trunk@15469 f3b2605a-c512-4ea7-a41b-209d697bcdaa 2016-08-12 13:49:50 +00:00
639ea6c396 git-svn-id: svn://svn.icms.temple.edu/lammps-ro/trunk@15468 f3b2605a-c512-4ea7-a41b-209d697bcdaa 2016-08-10 14:59:04 +00:00
b3364ae943 git-svn-id: svn://svn.icms.temple.edu/lammps-ro/trunk@15467 f3b2605a-c512-4ea7-a41b-209d697bcdaa 2016-08-10 14:46:16 +00:00
3349f7e143 git-svn-id: svn://svn.icms.temple.edu/lammps-ro/trunk@15466 f3b2605a-c512-4ea7-a41b-209d697bcdaa 2016-08-09 23:25:06 +00:00
de2b6cf6bd git-svn-id: svn://svn.icms.temple.edu/lammps-ro/trunk@15465 f3b2605a-c512-4ea7-a41b-209d697bcdaa 2016-08-09 20:45:02 +00:00
6c8d3ed4e8 git-svn-id: svn://svn.icms.temple.edu/lammps-ro/trunk@15463 f3b2605a-c512-4ea7-a41b-209d697bcdaa 2016-08-09 14:10:55 +00:00
550be1d512 git-svn-id: svn://svn.icms.temple.edu/lammps-ro/trunk@15462 f3b2605a-c512-4ea7-a41b-209d697bcdaa 2016-08-08 19:41:43 +00:00
563e069971 git-svn-id: svn://svn.icms.temple.edu/lammps-ro/trunk@15461 f3b2605a-c512-4ea7-a41b-209d697bcdaa 2016-08-08 13:50:31 +00:00
17e8e9a9c9 git-svn-id: svn://svn.icms.temple.edu/lammps-ro/trunk@15460 f3b2605a-c512-4ea7-a41b-209d697bcdaa 2016-08-08 13:48:22 +00:00
056ff192c6 git-svn-id: svn://svn.icms.temple.edu/lammps-ro/trunk@15459 f3b2605a-c512-4ea7-a41b-209d697bcdaa 2016-08-08 13:47:55 +00:00
0029040f11 git-svn-id: svn://svn.icms.temple.edu/lammps-ro/trunk@15456 f3b2605a-c512-4ea7-a41b-209d697bcdaa 2016-08-05 22:12:27 +00:00
decd072117 git-svn-id: svn://svn.icms.temple.edu/lammps-ro/trunk@15455 f3b2605a-c512-4ea7-a41b-209d697bcdaa 2016-08-05 20:58:38 +00:00
30d6007565 git-svn-id: svn://svn.icms.temple.edu/lammps-ro/trunk@15454 f3b2605a-c512-4ea7-a41b-209d697bcdaa 2016-08-05 20:45:14 +00:00
390b492d08 git-svn-id: svn://svn.icms.temple.edu/lammps-ro/trunk@15453 f3b2605a-c512-4ea7-a41b-209d697bcdaa 2016-08-05 18:28:23 +00:00
9a3b988b2b git-svn-id: svn://svn.icms.temple.edu/lammps-ro/trunk@15452 f3b2605a-c512-4ea7-a41b-209d697bcdaa 2016-08-05 18:28:02 +00:00
36a0c644a9 Attempted to straighten out italics
git-svn-id: svn://svn.icms.temple.edu/lammps-ro/trunk@15451 f3b2605a-c512-4ea7-a41b-209d697bcdaa
2016-08-05 18:21:12 +00:00
96a73d51cd git-svn-id: svn://svn.icms.temple.edu/lammps-ro/trunk@15450 f3b2605a-c512-4ea7-a41b-209d697bcdaa 2016-08-05 18:16:10 +00:00
1780c86b86 Fixed error in ave/time version
git-svn-id: svn://svn.icms.temple.edu/lammps-ro/trunk@15449 f3b2605a-c512-4ea7-a41b-209d697bcdaa
2016-08-05 17:53:52 +00:00
32c8c9d63e clarified energy contribution to minimization energy
git-svn-id: svn://svn.icms.temple.edu/lammps-ro/trunk@15448 f3b2605a-c512-4ea7-a41b-209d697bcdaa
2016-08-05 16:27:48 +00:00
7541033246 git-svn-id: svn://svn.icms.temple.edu/lammps-ro/trunk@15445 f3b2605a-c512-4ea7-a41b-209d697bcdaa 2016-08-05 15:50:41 +00:00
2d694f934e git-svn-id: svn://svn.icms.temple.edu/lammps-ro/trunk@15444 f3b2605a-c512-4ea7-a41b-209d697bcdaa 2016-08-05 15:50:00 +00:00
ce7581b869 git-svn-id: svn://svn.icms.temple.edu/lammps-ro/trunk@15442 f3b2605a-c512-4ea7-a41b-209d697bcdaa 2016-08-05 15:38:38 +00:00
42cc69b843 git-svn-id: svn://svn.icms.temple.edu/lammps-ro/trunk@15441 f3b2605a-c512-4ea7-a41b-209d697bcdaa 2016-08-05 15:17:36 +00:00
e09caf843d git-svn-id: svn://svn.icms.temple.edu/lammps-ro/trunk@15440 f3b2605a-c512-4ea7-a41b-209d697bcdaa 2016-08-05 15:17:23 +00:00
82a29a8adb git-svn-id: svn://svn.icms.temple.edu/lammps-ro/trunk@15439 f3b2605a-c512-4ea7-a41b-209d697bcdaa 2016-08-05 15:14:30 +00:00
091fb71a93 git-svn-id: svn://svn.icms.temple.edu/lammps-ro/trunk@15438 f3b2605a-c512-4ea7-a41b-209d697bcdaa 2016-08-05 15:14:06 +00:00
3fe162a42d git-svn-id: svn://svn.icms.temple.edu/lammps-ro/trunk@15437 f3b2605a-c512-4ea7-a41b-209d697bcdaa 2016-08-05 15:11:09 +00:00
9a9897cd44 git-svn-id: svn://svn.icms.temple.edu/lammps-ro/trunk@15436 f3b2605a-c512-4ea7-a41b-209d697bcdaa 2016-08-05 15:07:34 +00:00
0bffc1711e git-svn-id: svn://svn.icms.temple.edu/lammps-ro/trunk@15435 f3b2605a-c512-4ea7-a41b-209d697bcdaa 2016-08-04 16:39:15 +00:00
677da2ea52 git-svn-id: svn://svn.icms.temple.edu/lammps-ro/trunk@15432 f3b2605a-c512-4ea7-a41b-209d697bcdaa 2016-08-03 19:40:39 +00:00
a93e6e83ed git-svn-id: svn://svn.icms.temple.edu/lammps-ro/trunk@15431 f3b2605a-c512-4ea7-a41b-209d697bcdaa 2016-08-03 19:39:58 +00:00
efaa84a4ea git-svn-id: svn://svn.icms.temple.edu/lammps-ro/trunk@15430 f3b2605a-c512-4ea7-a41b-209d697bcdaa 2016-08-03 19:32:03 +00:00
1f6518400e git-svn-id: svn://svn.icms.temple.edu/lammps-ro/trunk@15429 f3b2605a-c512-4ea7-a41b-209d697bcdaa 2016-08-03 19:31:52 +00:00
be6c086cb4 git-svn-id: svn://svn.icms.temple.edu/lammps-ro/trunk@15428 f3b2605a-c512-4ea7-a41b-209d697bcdaa 2016-08-03 16:29:35 +00:00
a6ade15e60 git-svn-id: svn://svn.icms.temple.edu/lammps-ro/trunk@15427 f3b2605a-c512-4ea7-a41b-209d697bcdaa 2016-08-03 16:29:25 +00:00
38226b8086 git-svn-id: svn://svn.icms.temple.edu/lammps-ro/trunk@15426 f3b2605a-c512-4ea7-a41b-209d697bcdaa 2016-08-03 16:09:38 +00:00
56502c7c09 git-svn-id: svn://svn.icms.temple.edu/lammps-ro/trunk@15423 f3b2605a-c512-4ea7-a41b-209d697bcdaa 2016-08-01 22:57:27 +00:00
7326f88521 git-svn-id: svn://svn.icms.temple.edu/lammps-ro/trunk@15422 f3b2605a-c512-4ea7-a41b-209d697bcdaa 2016-08-01 22:56:43 +00:00
1cdd1fd4f8 git-svn-id: svn://svn.icms.temple.edu/lammps-ro/trunk@15420 f3b2605a-c512-4ea7-a41b-209d697bcdaa 2016-08-01 22:43:21 +00:00
69513a29ab git-svn-id: svn://svn.icms.temple.edu/lammps-ro/trunk@15419 f3b2605a-c512-4ea7-a41b-209d697bcdaa 2016-08-01 22:40:49 +00:00
98aee05152 git-svn-id: svn://svn.icms.temple.edu/lammps-ro/trunk@15418 f3b2605a-c512-4ea7-a41b-209d697bcdaa 2016-08-01 21:30:11 +00:00
e6c5898e53 git-svn-id: svn://svn.icms.temple.edu/lammps-ro/trunk@15417 f3b2605a-c512-4ea7-a41b-209d697bcdaa 2016-08-01 21:11:50 +00:00
a9a8a599d0 git-svn-id: svn://svn.icms.temple.edu/lammps-ro/trunk@15416 f3b2605a-c512-4ea7-a41b-209d697bcdaa 2016-08-01 21:01:23 +00:00
0995ab4b65 git-svn-id: svn://svn.icms.temple.edu/lammps-ro/trunk@15415 f3b2605a-c512-4ea7-a41b-209d697bcdaa 2016-08-01 20:55:29 +00:00
ce5f174281 git-svn-id: svn://svn.icms.temple.edu/lammps-ro/trunk@15414 f3b2605a-c512-4ea7-a41b-209d697bcdaa 2016-08-01 20:35:57 +00:00
6ff87be849 git-svn-id: svn://svn.icms.temple.edu/lammps-ro/trunk@15413 f3b2605a-c512-4ea7-a41b-209d697bcdaa 2016-07-30 21:24:55 +00:00
26622f2826 git-svn-id: svn://svn.icms.temple.edu/lammps-ro/trunk@15412 f3b2605a-c512-4ea7-a41b-209d697bcdaa 2016-07-30 18:40:03 +00:00
7ec3c90b86 git-svn-id: svn://svn.icms.temple.edu/lammps-ro/trunk@15411 f3b2605a-c512-4ea7-a41b-209d697bcdaa 2016-07-30 16:09:32 +00:00
fd05571ee0 git-svn-id: svn://svn.icms.temple.edu/lammps-ro/trunk@15410 f3b2605a-c512-4ea7-a41b-209d697bcdaa 2016-07-30 16:09:11 +00:00
5b5f839265 git-svn-id: svn://svn.icms.temple.edu/lammps-ro/trunk@15409 f3b2605a-c512-4ea7-a41b-209d697bcdaa 2016-07-30 16:08:42 +00:00
d3e47eef1c git-svn-id: svn://svn.icms.temple.edu/lammps-ro/trunk@15407 f3b2605a-c512-4ea7-a41b-209d697bcdaa 2016-07-30 15:17:32 +00:00
f9c6ec768b git-svn-id: svn://svn.icms.temple.edu/lammps-ro/trunk@15406 f3b2605a-c512-4ea7-a41b-209d697bcdaa 2016-07-30 15:11:41 +00:00
fb13763ef9 git-svn-id: svn://svn.icms.temple.edu/lammps-ro/trunk@15404 f3b2605a-c512-4ea7-a41b-209d697bcdaa 2016-07-30 14:52:58 +00:00
22fe3fd178 git-svn-id: svn://svn.icms.temple.edu/lammps-ro/trunk@15403 f3b2605a-c512-4ea7-a41b-209d697bcdaa 2016-07-30 14:52:18 +00:00
58be921a09 git-svn-id: svn://svn.icms.temple.edu/lammps-ro/trunk@15402 f3b2605a-c512-4ea7-a41b-209d697bcdaa 2016-07-30 14:51:20 +00:00
b5836fa491 git-svn-id: svn://svn.icms.temple.edu/lammps-ro/trunk@15401 f3b2605a-c512-4ea7-a41b-209d697bcdaa 2016-07-30 14:47:01 +00:00
4eb6664ba8 git-svn-id: svn://svn.icms.temple.edu/lammps-ro/trunk@15400 f3b2605a-c512-4ea7-a41b-209d697bcdaa 2016-07-30 14:45:47 +00:00
ee86bd2463 git-svn-id: svn://svn.icms.temple.edu/lammps-ro/trunk@15399 f3b2605a-c512-4ea7-a41b-209d697bcdaa 2016-07-30 14:45:26 +00:00
f609827de8 git-svn-id: svn://svn.icms.temple.edu/lammps-ro/trunk@15397 f3b2605a-c512-4ea7-a41b-209d697bcdaa 2016-07-29 14:35:56 +00:00
2ac26ab8ac git-svn-id: svn://svn.icms.temple.edu/lammps-ro/trunk@15396 f3b2605a-c512-4ea7-a41b-209d697bcdaa 2016-07-29 14:35:16 +00:00
3297fbeb26 git-svn-id: svn://svn.icms.temple.edu/lammps-ro/trunk@15395 f3b2605a-c512-4ea7-a41b-209d697bcdaa 2016-07-29 14:34:36 +00:00
642c01a150 git-svn-id: svn://svn.icms.temple.edu/lammps-ro/trunk@15394 f3b2605a-c512-4ea7-a41b-209d697bcdaa 2016-07-29 14:32:49 +00:00
b936d69f12 git-svn-id: svn://svn.icms.temple.edu/lammps-ro/trunk@15393 f3b2605a-c512-4ea7-a41b-209d697bcdaa 2016-07-29 14:28:58 +00:00
f486709150 git-svn-id: svn://svn.icms.temple.edu/lammps-ro/trunk@15392 f3b2605a-c512-4ea7-a41b-209d697bcdaa 2016-07-29 14:28:54 +00:00
b8bf3ae5b0 git-svn-id: svn://svn.icms.temple.edu/lammps-ro/trunk@15391 f3b2605a-c512-4ea7-a41b-209d697bcdaa 2016-07-29 14:24:21 +00:00
cf2aa231f5 git-svn-id: svn://svn.icms.temple.edu/lammps-ro/trunk@15390 f3b2605a-c512-4ea7-a41b-209d697bcdaa 2016-07-28 18:09:57 +00:00
87b6626358 git-svn-id: svn://svn.icms.temple.edu/lammps-ro/trunk@15389 f3b2605a-c512-4ea7-a41b-209d697bcdaa 2016-07-28 13:53:23 +00:00
70bfc1c097 git-svn-id: svn://svn.icms.temple.edu/lammps-ro/trunk@15388 f3b2605a-c512-4ea7-a41b-209d697bcdaa 2016-07-28 13:52:57 +00:00
c95ec24d1a git-svn-id: svn://svn.icms.temple.edu/lammps-ro/trunk@15387 f3b2605a-c512-4ea7-a41b-209d697bcdaa 2016-07-28 13:51:40 +00:00
f2f9fe0a65 git-svn-id: svn://svn.icms.temple.edu/lammps-ro/trunk@15386 f3b2605a-c512-4ea7-a41b-209d697bcdaa 2016-07-28 13:43:23 +00:00
e561aa0529 git-svn-id: svn://svn.icms.temple.edu/lammps-ro/trunk@15382 f3b2605a-c512-4ea7-a41b-209d697bcdaa 2016-07-27 23:34:59 +00:00
a01058e7d2 git-svn-id: svn://svn.icms.temple.edu/lammps-ro/trunk@15381 f3b2605a-c512-4ea7-a41b-209d697bcdaa 2016-07-27 23:34:21 +00:00
308461125e git-svn-id: svn://svn.icms.temple.edu/lammps-ro/trunk@15379 f3b2605a-c512-4ea7-a41b-209d697bcdaa 2016-07-27 23:27:21 +00:00
a91dae3f2b git-svn-id: svn://svn.icms.temple.edu/lammps-ro/trunk@15378 f3b2605a-c512-4ea7-a41b-209d697bcdaa 2016-07-27 23:26:29 +00:00
e8c5280ad9 Added axes keyword
git-svn-id: svn://svn.icms.temple.edu/lammps-ro/trunk@15377 f3b2605a-c512-4ea7-a41b-209d697bcdaa
2016-07-27 23:06:18 +00:00
5241c0326e Added axes keyword
git-svn-id: svn://svn.icms.temple.edu/lammps-ro/trunk@15376 f3b2605a-c512-4ea7-a41b-209d697bcdaa
2016-07-27 23:03:59 +00:00
ae255c847b git-svn-id: svn://svn.icms.temple.edu/lammps-ro/trunk@15375 f3b2605a-c512-4ea7-a41b-209d697bcdaa 2016-07-27 22:51:18 +00:00
2479d8031c Fixed memory leak from rlist
git-svn-id: svn://svn.icms.temple.edu/lammps-ro/trunk@15374 f3b2605a-c512-4ea7-a41b-209d697bcdaa
2016-07-27 21:16:09 +00:00
6b79bbfaf3 Fixing Kokkos memory issue
git-svn-id: svn://svn.icms.temple.edu/lammps-ro/trunk@15373 f3b2605a-c512-4ea7-a41b-209d697bcdaa
2016-07-27 15:48:50 +00:00
ef6ff80366 git-svn-id: svn://svn.icms.temple.edu/lammps-ro/trunk@15372 f3b2605a-c512-4ea7-a41b-209d697bcdaa 2016-07-27 14:13:02 +00:00
ee24be38cb git-svn-id: svn://svn.icms.temple.edu/lammps-ro/trunk@15371 f3b2605a-c512-4ea7-a41b-209d697bcdaa 2016-07-27 14:09:42 +00:00
c1fc5aef12 git-svn-id: svn://svn.icms.temple.edu/lammps-ro/trunk@15370 f3b2605a-c512-4ea7-a41b-209d697bcdaa 2016-07-27 14:08:32 +00:00
ddd85f006c git-svn-id: svn://svn.icms.temple.edu/lammps-ro/trunk@15369 f3b2605a-c512-4ea7-a41b-209d697bcdaa 2016-07-27 14:04:56 +00:00
8c04540e8a git-svn-id: svn://svn.icms.temple.edu/lammps-ro/trunk@15368 f3b2605a-c512-4ea7-a41b-209d697bcdaa 2016-07-25 19:55:01 +00:00
751786364d git-svn-id: svn://svn.icms.temple.edu/lammps-ro/trunk@15367 f3b2605a-c512-4ea7-a41b-209d697bcdaa 2016-07-23 16:11:53 +00:00
08ac3d54dd git-svn-id: svn://svn.icms.temple.edu/lammps-ro/trunk@15364 f3b2605a-c512-4ea7-a41b-209d697bcdaa 2016-07-23 00:24:54 +00:00
caf9483a0a git-svn-id: svn://svn.icms.temple.edu/lammps-ro/trunk@15362 f3b2605a-c512-4ea7-a41b-209d697bcdaa 2016-07-22 23:05:05 +00:00
c04c775631 git-svn-id: svn://svn.icms.temple.edu/lammps-ro/trunk@15361 f3b2605a-c512-4ea7-a41b-209d697bcdaa 2016-07-22 23:04:25 +00:00
7dc6873c51 git-svn-id: svn://svn.icms.temple.edu/lammps-ro/trunk@15360 f3b2605a-c512-4ea7-a41b-209d697bcdaa 2016-07-22 22:59:39 +00:00
952a0d1fb7 git-svn-id: svn://svn.icms.temple.edu/lammps-ro/trunk@15359 f3b2605a-c512-4ea7-a41b-209d697bcdaa 2016-07-22 22:58:42 +00:00
778f4d338c git-svn-id: svn://svn.icms.temple.edu/lammps-ro/trunk@15358 f3b2605a-c512-4ea7-a41b-209d697bcdaa 2016-07-22 22:57:54 +00:00
2f936d5e56 git-svn-id: svn://svn.icms.temple.edu/lammps-ro/trunk@15357 f3b2605a-c512-4ea7-a41b-209d697bcdaa 2016-07-22 22:57:39 +00:00
87a6c1368f git-svn-id: svn://svn.icms.temple.edu/lammps-ro/trunk@15356 f3b2605a-c512-4ea7-a41b-209d697bcdaa 2016-07-22 22:56:08 +00:00
cff65b956a git-svn-id: svn://svn.icms.temple.edu/lammps-ro/trunk@15355 f3b2605a-c512-4ea7-a41b-209d697bcdaa 2016-07-22 22:52:03 +00:00
87bfe1941b git-svn-id: svn://svn.icms.temple.edu/lammps-ro/trunk@15354 f3b2605a-c512-4ea7-a41b-209d697bcdaa 2016-07-22 22:51:37 +00:00
e519cfb1dc git-svn-id: svn://svn.icms.temple.edu/lammps-ro/trunk@15353 f3b2605a-c512-4ea7-a41b-209d697bcdaa 2016-07-22 22:50:41 +00:00
5abac962ed git-svn-id: svn://svn.icms.temple.edu/lammps-ro/trunk@15352 f3b2605a-c512-4ea7-a41b-209d697bcdaa 2016-07-22 22:46:14 +00:00
0659a1fc5a git-svn-id: svn://svn.icms.temple.edu/lammps-ro/trunk@15351 f3b2605a-c512-4ea7-a41b-209d697bcdaa 2016-07-22 22:10:50 +00:00
ff10357ef2 git-svn-id: svn://svn.icms.temple.edu/lammps-ro/trunk@15350 f3b2605a-c512-4ea7-a41b-209d697bcdaa 2016-07-22 22:10:46 +00:00
f9a4b697a4 git-svn-id: svn://svn.icms.temple.edu/lammps-ro/trunk@15349 f3b2605a-c512-4ea7-a41b-209d697bcdaa 2016-07-22 21:20:38 +00:00
0f71c66b6f git-svn-id: svn://svn.icms.temple.edu/lammps-ro/trunk@15348 f3b2605a-c512-4ea7-a41b-209d697bcdaa 2016-07-22 21:19:08 +00:00
f7c2fdf335 git-svn-id: svn://svn.icms.temple.edu/lammps-ro/trunk@15347 f3b2605a-c512-4ea7-a41b-209d697bcdaa 2016-07-22 19:41:45 +00:00
137f19ba9f git-svn-id: svn://svn.icms.temple.edu/lammps-ro/trunk@15346 f3b2605a-c512-4ea7-a41b-209d697bcdaa 2016-07-22 19:41:42 +00:00
efa508d905 git-svn-id: svn://svn.icms.temple.edu/lammps-ro/trunk@15345 f3b2605a-c512-4ea7-a41b-209d697bcdaa 2016-07-21 23:18:48 +00:00
3ceaf974f8 git-svn-id: svn://svn.icms.temple.edu/lammps-ro/trunk@15344 f3b2605a-c512-4ea7-a41b-209d697bcdaa 2016-07-21 23:18:44 +00:00
667983a09d Fixed typo
git-svn-id: svn://svn.icms.temple.edu/lammps-ro/trunk@15343 f3b2605a-c512-4ea7-a41b-209d697bcdaa
2016-07-21 21:35:28 +00:00
64516e4314 git-svn-id: svn://svn.icms.temple.edu/lammps-ro/trunk@15342 f3b2605a-c512-4ea7-a41b-209d697bcdaa 2016-07-18 22:08:53 +00:00
1e3b96c486 git-svn-id: svn://svn.icms.temple.edu/lammps-ro/trunk@15341 f3b2605a-c512-4ea7-a41b-209d697bcdaa 2016-07-18 17:02:53 +00:00
6e2a74eaf9 git-svn-id: svn://svn.icms.temple.edu/lammps-ro/trunk@15340 f3b2605a-c512-4ea7-a41b-209d697bcdaa 2016-07-18 16:59:02 +00:00
0ec028d9f2 git-svn-id: svn://svn.icms.temple.edu/lammps-ro/trunk@15339 f3b2605a-c512-4ea7-a41b-209d697bcdaa 2016-07-16 22:35:08 +00:00
778ae37e6e git-svn-id: svn://svn.icms.temple.edu/lammps-ro/trunk@15338 f3b2605a-c512-4ea7-a41b-209d697bcdaa 2016-07-16 22:27:01 +00:00
1111c8ba93 git-svn-id: svn://svn.icms.temple.edu/lammps-ro/trunk@15337 f3b2605a-c512-4ea7-a41b-209d697bcdaa 2016-07-16 22:19:46 +00:00
a6f62209af git-svn-id: svn://svn.icms.temple.edu/lammps-ro/trunk@15336 f3b2605a-c512-4ea7-a41b-209d697bcdaa 2016-07-16 22:14:50 +00:00
89b839056c git-svn-id: svn://svn.icms.temple.edu/lammps-ro/trunk@15335 f3b2605a-c512-4ea7-a41b-209d697bcdaa 2016-07-16 22:14:24 +00:00
596d6840a5 git-svn-id: svn://svn.icms.temple.edu/lammps-ro/trunk@15334 f3b2605a-c512-4ea7-a41b-209d697bcdaa 2016-07-16 22:14:09 +00:00
a5b6dbea78 git-svn-id: svn://svn.icms.temple.edu/lammps-ro/trunk@15333 f3b2605a-c512-4ea7-a41b-209d697bcdaa 2016-07-16 20:00:36 +00:00
d53616b6c8 git-svn-id: svn://svn.icms.temple.edu/lammps-ro/trunk@15331 f3b2605a-c512-4ea7-a41b-209d697bcdaa 2016-07-15 23:18:17 +00:00
4bf96ac144 git-svn-id: svn://svn.icms.temple.edu/lammps-ro/trunk@15330 f3b2605a-c512-4ea7-a41b-209d697bcdaa 2016-07-15 23:17:36 +00:00
ea39843564 git-svn-id: svn://svn.icms.temple.edu/lammps-ro/trunk@15329 f3b2605a-c512-4ea7-a41b-209d697bcdaa 2016-07-15 23:14:13 +00:00
472fe29a89 git-svn-id: svn://svn.icms.temple.edu/lammps-ro/trunk@15328 f3b2605a-c512-4ea7-a41b-209d697bcdaa 2016-07-15 23:12:28 +00:00
e6dc896b49 git-svn-id: svn://svn.icms.temple.edu/lammps-ro/trunk@15327 f3b2605a-c512-4ea7-a41b-209d697bcdaa 2016-07-15 23:02:09 +00:00
80fce623ec git-svn-id: svn://svn.icms.temple.edu/lammps-ro/trunk@15326 f3b2605a-c512-4ea7-a41b-209d697bcdaa 2016-07-15 22:59:07 +00:00
a47e457ab0 git-svn-id: svn://svn.icms.temple.edu/lammps-ro/trunk@15325 f3b2605a-c512-4ea7-a41b-209d697bcdaa 2016-07-15 22:49:00 +00:00
a2782baa64 git-svn-id: svn://svn.icms.temple.edu/lammps-ro/trunk@15324 f3b2605a-c512-4ea7-a41b-209d697bcdaa 2016-07-15 22:47:23 +00:00
2705ddcaab git-svn-id: svn://svn.icms.temple.edu/lammps-ro/trunk@15323 f3b2605a-c512-4ea7-a41b-209d697bcdaa 2016-07-15 22:43:55 +00:00
638b0db535 git-svn-id: svn://svn.icms.temple.edu/lammps-ro/trunk@15322 f3b2605a-c512-4ea7-a41b-209d697bcdaa 2016-07-15 22:40:38 +00:00
8a4dc62d2c git-svn-id: svn://svn.icms.temple.edu/lammps-ro/trunk@15321 f3b2605a-c512-4ea7-a41b-209d697bcdaa 2016-07-15 22:39:08 +00:00
c79a24c7ef git-svn-id: svn://svn.icms.temple.edu/lammps-ro/trunk@15320 f3b2605a-c512-4ea7-a41b-209d697bcdaa 2016-07-15 22:38:58 +00:00
12f3178e86 git-svn-id: svn://svn.icms.temple.edu/lammps-ro/trunk@15319 f3b2605a-c512-4ea7-a41b-209d697bcdaa 2016-07-15 22:38:42 +00:00
eb78022b5a git-svn-id: svn://svn.icms.temple.edu/lammps-ro/trunk@15318 f3b2605a-c512-4ea7-a41b-209d697bcdaa 2016-07-15 22:35:36 +00:00
10bbb28943 git-svn-id: svn://svn.icms.temple.edu/lammps-ro/trunk@15317 f3b2605a-c512-4ea7-a41b-209d697bcdaa 2016-07-15 22:33:52 +00:00
1587bdf2f1 git-svn-id: svn://svn.icms.temple.edu/lammps-ro/trunk@15316 f3b2605a-c512-4ea7-a41b-209d697bcdaa 2016-07-15 22:26:00 +00:00
05398a6863 git-svn-id: svn://svn.icms.temple.edu/lammps-ro/trunk@15315 f3b2605a-c512-4ea7-a41b-209d697bcdaa 2016-07-15 22:25:17 +00:00
d32f1ea4c0 git-svn-id: svn://svn.icms.temple.edu/lammps-ro/trunk@15314 f3b2605a-c512-4ea7-a41b-209d697bcdaa 2016-07-15 16:37:57 +00:00
d326e8a767 git-svn-id: svn://svn.icms.temple.edu/lammps-ro/trunk@15313 f3b2605a-c512-4ea7-a41b-209d697bcdaa 2016-07-14 19:44:08 +00:00
b399e2e496 git-svn-id: svn://svn.icms.temple.edu/lammps-ro/trunk@15312 f3b2605a-c512-4ea7-a41b-209d697bcdaa 2016-07-14 19:43:50 +00:00
7d2809dd09 git-svn-id: svn://svn.icms.temple.edu/lammps-ro/trunk@15311 f3b2605a-c512-4ea7-a41b-209d697bcdaa 2016-07-14 19:43:03 +00:00
4e76c2da53 git-svn-id: svn://svn.icms.temple.edu/lammps-ro/trunk@15310 f3b2605a-c512-4ea7-a41b-209d697bcdaa 2016-07-14 19:42:45 +00:00
2f09c2755a git-svn-id: svn://svn.icms.temple.edu/lammps-ro/trunk@15309 f3b2605a-c512-4ea7-a41b-209d697bcdaa 2016-07-14 19:42:39 +00:00
2020578386 git-svn-id: svn://svn.icms.temple.edu/lammps-ro/trunk@15307 f3b2605a-c512-4ea7-a41b-209d697bcdaa 2016-07-14 00:09:09 +00:00
35bc617478 git-svn-id: svn://svn.icms.temple.edu/lammps-ro/trunk@15306 f3b2605a-c512-4ea7-a41b-209d697bcdaa 2016-07-14 00:08:27 +00:00
4b18665606 git-svn-id: svn://svn.icms.temple.edu/lammps-ro/trunk@15305 f3b2605a-c512-4ea7-a41b-209d697bcdaa 2016-07-13 22:58:08 +00:00
5881c6da6f git-svn-id: svn://svn.icms.temple.edu/lammps-ro/trunk@15304 f3b2605a-c512-4ea7-a41b-209d697bcdaa 2016-07-13 22:55:25 +00:00
c1fb6c967e git-svn-id: svn://svn.icms.temple.edu/lammps-ro/trunk@15303 f3b2605a-c512-4ea7-a41b-209d697bcdaa 2016-07-13 22:38:44 +00:00
6367c3700b git-svn-id: svn://svn.icms.temple.edu/lammps-ro/trunk@15302 f3b2605a-c512-4ea7-a41b-209d697bcdaa 2016-07-13 22:36:46 +00:00
6171c774aa git-svn-id: svn://svn.icms.temple.edu/lammps-ro/trunk@15301 f3b2605a-c512-4ea7-a41b-209d697bcdaa 2016-07-13 21:36:09 +00:00
b3aeb43a93 git-svn-id: svn://svn.icms.temple.edu/lammps-ro/trunk@15300 f3b2605a-c512-4ea7-a41b-209d697bcdaa 2016-07-13 21:35:51 +00:00
020484f254 git-svn-id: svn://svn.icms.temple.edu/lammps-ro/trunk@15299 f3b2605a-c512-4ea7-a41b-209d697bcdaa 2016-07-13 21:34:23 +00:00
92e5aae7bb git-svn-id: svn://svn.icms.temple.edu/lammps-ro/trunk@15298 f3b2605a-c512-4ea7-a41b-209d697bcdaa 2016-07-13 14:43:40 +00:00
e58791e29d git-svn-id: svn://svn.icms.temple.edu/lammps-ro/trunk@15296 f3b2605a-c512-4ea7-a41b-209d697bcdaa 2016-07-12 23:28:10 +00:00
e3e7d0280d git-svn-id: svn://svn.icms.temple.edu/lammps-ro/trunk@15295 f3b2605a-c512-4ea7-a41b-209d697bcdaa 2016-07-12 23:27:30 +00:00
f9047919e8 git-svn-id: svn://svn.icms.temple.edu/lammps-ro/trunk@15294 f3b2605a-c512-4ea7-a41b-209d697bcdaa 2016-07-12 23:27:22 +00:00
070ca26148 git-svn-id: svn://svn.icms.temple.edu/lammps-ro/trunk@15293 f3b2605a-c512-4ea7-a41b-209d697bcdaa 2016-07-12 23:19:38 +00:00
5783c78f44 git-svn-id: svn://svn.icms.temple.edu/lammps-ro/trunk@15292 f3b2605a-c512-4ea7-a41b-209d697bcdaa 2016-07-12 23:16:13 +00:00
eec4d961c0 git-svn-id: svn://svn.icms.temple.edu/lammps-ro/trunk@15291 f3b2605a-c512-4ea7-a41b-209d697bcdaa 2016-07-12 22:59:27 +00:00
6ec2346ca0 git-svn-id: svn://svn.icms.temple.edu/lammps-ro/trunk@15290 f3b2605a-c512-4ea7-a41b-209d697bcdaa 2016-07-12 22:40:29 +00:00
c5aece5b43 git-svn-id: svn://svn.icms.temple.edu/lammps-ro/trunk@15289 f3b2605a-c512-4ea7-a41b-209d697bcdaa 2016-07-12 22:40:15 +00:00
0ff512a6da git-svn-id: svn://svn.icms.temple.edu/lammps-ro/trunk@15288 f3b2605a-c512-4ea7-a41b-209d697bcdaa 2016-07-12 22:36:08 +00:00
568c530bfe git-svn-id: svn://svn.icms.temple.edu/lammps-ro/trunk@15287 f3b2605a-c512-4ea7-a41b-209d697bcdaa 2016-07-12 22:36:00 +00:00
3f802d97c9 git-svn-id: svn://svn.icms.temple.edu/lammps-ro/trunk@15286 f3b2605a-c512-4ea7-a41b-209d697bcdaa 2016-07-12 22:35:56 +00:00
1df9db5091 git-svn-id: svn://svn.icms.temple.edu/lammps-ro/trunk@15285 f3b2605a-c512-4ea7-a41b-209d697bcdaa 2016-07-12 22:35:46 +00:00
bf9e821055 git-svn-id: svn://svn.icms.temple.edu/lammps-ro/trunk@15284 f3b2605a-c512-4ea7-a41b-209d697bcdaa 2016-07-12 17:51:08 +00:00
dfbecf047f git-svn-id: svn://svn.icms.temple.edu/lammps-ro/trunk@15283 f3b2605a-c512-4ea7-a41b-209d697bcdaa 2016-07-12 14:50:23 +00:00
2812b10032 git-svn-id: svn://svn.icms.temple.edu/lammps-ro/trunk@15282 f3b2605a-c512-4ea7-a41b-209d697bcdaa 2016-07-11 20:57:03 +00:00
5ea854de2e git-svn-id: svn://svn.icms.temple.edu/lammps-ro/trunk@15280 f3b2605a-c512-4ea7-a41b-209d697bcdaa 2016-07-09 20:59:58 +00:00
4484686077 git-svn-id: svn://svn.icms.temple.edu/lammps-ro/trunk@15279 f3b2605a-c512-4ea7-a41b-209d697bcdaa 2016-07-09 20:59:53 +00:00
a33753ae6f git-svn-id: svn://svn.icms.temple.edu/lammps-ro/trunk@15278 f3b2605a-c512-4ea7-a41b-209d697bcdaa 2016-07-09 20:57:00 +00:00
80b6719d68 git-svn-id: svn://svn.icms.temple.edu/lammps-ro/trunk@15277 f3b2605a-c512-4ea7-a41b-209d697bcdaa 2016-07-09 20:43:06 +00:00
ee6f9b9621 git-svn-id: svn://svn.icms.temple.edu/lammps-ro/trunk@15276 f3b2605a-c512-4ea7-a41b-209d697bcdaa 2016-07-09 20:41:38 +00:00
28faf10980 git-svn-id: svn://svn.icms.temple.edu/lammps-ro/trunk@15275 f3b2605a-c512-4ea7-a41b-209d697bcdaa 2016-07-09 20:41:21 +00:00
6bc507cd6a git-svn-id: svn://svn.icms.temple.edu/lammps-ro/trunk@15274 f3b2605a-c512-4ea7-a41b-209d697bcdaa 2016-07-09 20:41:07 +00:00
18b0426c99 git-svn-id: svn://svn.icms.temple.edu/lammps-ro/trunk@15273 f3b2605a-c512-4ea7-a41b-209d697bcdaa 2016-07-09 20:40:46 +00:00
cdc9712261 git-svn-id: svn://svn.icms.temple.edu/lammps-ro/trunk@15270 f3b2605a-c512-4ea7-a41b-209d697bcdaa 2016-07-07 14:39:49 +00:00
526f95e050 git-svn-id: svn://svn.icms.temple.edu/lammps-ro/trunk@15269 f3b2605a-c512-4ea7-a41b-209d697bcdaa 2016-07-06 23:27:18 +00:00
65c1e16401 git-svn-id: svn://svn.icms.temple.edu/lammps-ro/trunk@15268 f3b2605a-c512-4ea7-a41b-209d697bcdaa 2016-07-06 23:26:44 +00:00
fd27214f7d git-svn-id: svn://svn.icms.temple.edu/lammps-ro/trunk@15267 f3b2605a-c512-4ea7-a41b-209d697bcdaa 2016-07-06 23:26:27 +00:00
10c19a9530 git-svn-id: svn://svn.icms.temple.edu/lammps-ro/trunk@15266 f3b2605a-c512-4ea7-a41b-209d697bcdaa 2016-07-06 23:25:51 +00:00
9da26677c5 git-svn-id: svn://svn.icms.temple.edu/lammps-ro/trunk@15265 f3b2605a-c512-4ea7-a41b-209d697bcdaa 2016-07-06 15:52:15 +00:00
85cb6786e0 git-svn-id: svn://svn.icms.temple.edu/lammps-ro/trunk@15264 f3b2605a-c512-4ea7-a41b-209d697bcdaa 2016-07-06 14:54:27 +00:00
33ae48afa5 git-svn-id: svn://svn.icms.temple.edu/lammps-ro/trunk@15263 f3b2605a-c512-4ea7-a41b-209d697bcdaa 2016-07-06 14:54:13 +00:00
f31a2c9400 git-svn-id: svn://svn.icms.temple.edu/lammps-ro/trunk@15262 f3b2605a-c512-4ea7-a41b-209d697bcdaa 2016-07-05 14:53:23 +00:00
34f829c5ac git-svn-id: svn://svn.icms.temple.edu/lammps-ro/trunk@15261 f3b2605a-c512-4ea7-a41b-209d697bcdaa 2016-07-05 14:41:13 +00:00
742ad5ac50 git-svn-id: svn://svn.icms.temple.edu/lammps-ro/trunk@15260 f3b2605a-c512-4ea7-a41b-209d697bcdaa 2016-07-05 14:41:02 +00:00
547dd3edc8 git-svn-id: svn://svn.icms.temple.edu/lammps-ro/trunk@15259 f3b2605a-c512-4ea7-a41b-209d697bcdaa 2016-07-02 21:03:48 +00:00
71f3c4556f git-svn-id: svn://svn.icms.temple.edu/lammps-ro/trunk@15256 f3b2605a-c512-4ea7-a41b-209d697bcdaa 2016-07-01 23:37:01 +00:00
b8ad2ecd7a git-svn-id: svn://svn.icms.temple.edu/lammps-ro/trunk@15255 f3b2605a-c512-4ea7-a41b-209d697bcdaa 2016-07-01 23:36:21 +00:00
1a6f8965b3 git-svn-id: svn://svn.icms.temple.edu/lammps-ro/trunk@15254 f3b2605a-c512-4ea7-a41b-209d697bcdaa 2016-07-01 23:32:48 +00:00
a329a724b2 git-svn-id: svn://svn.icms.temple.edu/lammps-ro/trunk@15253 f3b2605a-c512-4ea7-a41b-209d697bcdaa 2016-07-01 23:31:52 +00:00
2a1b47172d git-svn-id: svn://svn.icms.temple.edu/lammps-ro/trunk@15252 f3b2605a-c512-4ea7-a41b-209d697bcdaa 2016-07-01 23:31:41 +00:00
6f3ac03a08 git-svn-id: svn://svn.icms.temple.edu/lammps-ro/trunk@15251 f3b2605a-c512-4ea7-a41b-209d697bcdaa 2016-07-01 23:31:19 +00:00
88812c44fb git-svn-id: svn://svn.icms.temple.edu/lammps-ro/trunk@15250 f3b2605a-c512-4ea7-a41b-209d697bcdaa 2016-07-01 23:31:09 +00:00
a99c922a14 git-svn-id: svn://svn.icms.temple.edu/lammps-ro/trunk@15249 f3b2605a-c512-4ea7-a41b-209d697bcdaa 2016-07-01 23:29:38 +00:00
9656958169 git-svn-id: svn://svn.icms.temple.edu/lammps-ro/trunk@15248 f3b2605a-c512-4ea7-a41b-209d697bcdaa 2016-07-01 23:27:26 +00:00
8366b35459 git-svn-id: svn://svn.icms.temple.edu/lammps-ro/trunk@15247 f3b2605a-c512-4ea7-a41b-209d697bcdaa 2016-07-01 23:25:50 +00:00
cedb420ebc git-svn-id: svn://svn.icms.temple.edu/lammps-ro/trunk@15246 f3b2605a-c512-4ea7-a41b-209d697bcdaa 2016-07-01 23:23:56 +00:00
d11877d5af git-svn-id: svn://svn.icms.temple.edu/lammps-ro/trunk@15245 f3b2605a-c512-4ea7-a41b-209d697bcdaa 2016-07-01 23:22:42 +00:00
99448ee454 git-svn-id: svn://svn.icms.temple.edu/lammps-ro/trunk@15244 f3b2605a-c512-4ea7-a41b-209d697bcdaa 2016-07-01 23:21:59 +00:00
78df7623b3 git-svn-id: svn://svn.icms.temple.edu/lammps-ro/trunk@15243 f3b2605a-c512-4ea7-a41b-209d697bcdaa 2016-07-01 23:21:38 +00:00
b4b30d65da git-svn-id: svn://svn.icms.temple.edu/lammps-ro/trunk@15242 f3b2605a-c512-4ea7-a41b-209d697bcdaa 2016-07-01 22:39:55 +00:00
d291fcb5d9 git-svn-id: svn://svn.icms.temple.edu/lammps-ro/trunk@15241 f3b2605a-c512-4ea7-a41b-209d697bcdaa 2016-07-01 22:39:26 +00:00
29e64df1ba git-svn-id: svn://svn.icms.temple.edu/lammps-ro/trunk@15240 f3b2605a-c512-4ea7-a41b-209d697bcdaa 2016-07-01 21:27:04 +00:00
f0179f6f90 git-svn-id: svn://svn.icms.temple.edu/lammps-ro/trunk@15238 f3b2605a-c512-4ea7-a41b-209d697bcdaa 2016-07-01 16:29:01 +00:00
9fd6803142 git-svn-id: svn://svn.icms.temple.edu/lammps-ro/trunk@15237 f3b2605a-c512-4ea7-a41b-209d697bcdaa 2016-07-01 14:25:15 +00:00
94bdf1237b git-svn-id: svn://svn.icms.temple.edu/lammps-ro/trunk@15236 f3b2605a-c512-4ea7-a41b-209d697bcdaa 2016-07-01 14:21:54 +00:00
cda102364a git-svn-id: svn://svn.icms.temple.edu/lammps-ro/trunk@15235 f3b2605a-c512-4ea7-a41b-209d697bcdaa 2016-06-30 20:43:33 +00:00
f735a669ad Fixed bug in qtype bugfix
git-svn-id: svn://svn.icms.temple.edu/lammps-ro/trunk@15234 f3b2605a-c512-4ea7-a41b-209d697bcdaa
2016-06-30 19:25:44 +00:00
df720a4565 git-svn-id: svn://svn.icms.temple.edu/lammps-ro/trunk@15233 f3b2605a-c512-4ea7-a41b-209d697bcdaa 2016-06-28 22:43:59 +00:00
c242bca4d1 git-svn-id: svn://svn.icms.temple.edu/lammps-ro/trunk@15230 f3b2605a-c512-4ea7-a41b-209d697bcdaa 2016-06-28 13:30:53 +00:00
a01d08aba4 git-svn-id: svn://svn.icms.temple.edu/lammps-ro/trunk@15229 f3b2605a-c512-4ea7-a41b-209d697bcdaa 2016-06-28 13:30:14 +00:00
42071be08c git-svn-id: svn://svn.icms.temple.edu/lammps-ro/trunk@15228 f3b2605a-c512-4ea7-a41b-209d697bcdaa 2016-06-28 13:30:04 +00:00
8c63302c82 git-svn-id: svn://svn.icms.temple.edu/lammps-ro/trunk@15227 f3b2605a-c512-4ea7-a41b-209d697bcdaa 2016-06-28 13:29:30 +00:00
788ba55436 git-svn-id: svn://svn.icms.temple.edu/lammps-ro/trunk@15226 f3b2605a-c512-4ea7-a41b-209d697bcdaa 2016-06-28 13:29:16 +00:00
cec2f2518f git-svn-id: svn://svn.icms.temple.edu/lammps-ro/trunk@15225 f3b2605a-c512-4ea7-a41b-209d697bcdaa 2016-06-28 13:23:54 +00:00
bcac93f7a2 git-svn-id: svn://svn.icms.temple.edu/lammps-ro/trunk@15224 f3b2605a-c512-4ea7-a41b-209d697bcdaa 2016-06-28 13:20:34 +00:00
ab2fe0113e git-svn-id: svn://svn.icms.temple.edu/lammps-ro/trunk@15223 f3b2605a-c512-4ea7-a41b-209d697bcdaa 2016-06-28 13:20:05 +00:00
dae9f7cbea git-svn-id: svn://svn.icms.temple.edu/lammps-ro/trunk@15221 f3b2605a-c512-4ea7-a41b-209d697bcdaa 2016-06-27 22:39:29 +00:00
f3a3b1c838 git-svn-id: svn://svn.icms.temple.edu/lammps-ro/trunk@15220 f3b2605a-c512-4ea7-a41b-209d697bcdaa 2016-06-27 22:38:50 +00:00
1c05f57bcd git-svn-id: svn://svn.icms.temple.edu/lammps-ro/trunk@15218 f3b2605a-c512-4ea7-a41b-209d697bcdaa 2016-06-27 22:32:39 +00:00
b58e008cae git-svn-id: svn://svn.icms.temple.edu/lammps-ro/trunk@15217 f3b2605a-c512-4ea7-a41b-209d697bcdaa 2016-06-27 22:19:57 +00:00
ffc252e784 Adding Kokkos ReaxFF files
git-svn-id: svn://svn.icms.temple.edu/lammps-ro/trunk@15216 f3b2605a-c512-4ea7-a41b-209d697bcdaa
2016-06-23 21:03:58 +00:00
0434eb1689 Adding Kokkos ReaxFF files
git-svn-id: svn://svn.icms.temple.edu/lammps-ro/trunk@15215 f3b2605a-c512-4ea7-a41b-209d697bcdaa
2016-06-23 20:11:01 +00:00
fa7c2cea4d Adding Kokkos mpi-only Makefile
git-svn-id: svn://svn.icms.temple.edu/lammps-ro/trunk@15214 f3b2605a-c512-4ea7-a41b-209d697bcdaa
2016-06-22 16:51:57 +00:00
38c80e129f git-svn-id: svn://svn.icms.temple.edu/lammps-ro/trunk@15210 f3b2605a-c512-4ea7-a41b-209d697bcdaa 2016-06-18 20:45:53 +00:00
4569c3876c git-svn-id: svn://svn.icms.temple.edu/lammps-ro/trunk@15207 f3b2605a-c512-4ea7-a41b-209d697bcdaa 2016-06-17 23:55:12 +00:00
bcf79f62eb git-svn-id: svn://svn.icms.temple.edu/lammps-ro/trunk@15206 f3b2605a-c512-4ea7-a41b-209d697bcdaa 2016-06-17 23:54:32 +00:00
6f6e08652e git-svn-id: svn://svn.icms.temple.edu/lammps-ro/trunk@15205 f3b2605a-c512-4ea7-a41b-209d697bcdaa 2016-06-17 23:51:19 +00:00
143b72e7da git-svn-id: svn://svn.icms.temple.edu/lammps-ro/trunk@15204 f3b2605a-c512-4ea7-a41b-209d697bcdaa 2016-06-17 23:50:52 +00:00
d55f968432 git-svn-id: svn://svn.icms.temple.edu/lammps-ro/trunk@15203 f3b2605a-c512-4ea7-a41b-209d697bcdaa 2016-06-17 23:48:15 +00:00
b161fbb52a git-svn-id: svn://svn.icms.temple.edu/lammps-ro/trunk@15202 f3b2605a-c512-4ea7-a41b-209d697bcdaa 2016-06-17 23:24:05 +00:00
d89ee2a40d git-svn-id: svn://svn.icms.temple.edu/lammps-ro/trunk@15201 f3b2605a-c512-4ea7-a41b-209d697bcdaa 2016-06-17 23:13:02 +00:00
2fcd26f6c4 git-svn-id: svn://svn.icms.temple.edu/lammps-ro/trunk@15200 f3b2605a-c512-4ea7-a41b-209d697bcdaa 2016-06-17 23:07:51 +00:00
e5fb28a6a0 git-svn-id: svn://svn.icms.temple.edu/lammps-ro/trunk@15199 f3b2605a-c512-4ea7-a41b-209d697bcdaa 2016-06-17 23:07:24 +00:00
a90803641c git-svn-id: svn://svn.icms.temple.edu/lammps-ro/trunk@15198 f3b2605a-c512-4ea7-a41b-209d697bcdaa 2016-06-17 23:02:48 +00:00
46a9fe58aa git-svn-id: svn://svn.icms.temple.edu/lammps-ro/trunk@15197 f3b2605a-c512-4ea7-a41b-209d697bcdaa 2016-06-17 23:02:30 +00:00
852b9eec18 git-svn-id: svn://svn.icms.temple.edu/lammps-ro/trunk@15195 f3b2605a-c512-4ea7-a41b-209d697bcdaa 2016-06-17 21:58:41 +00:00
00438d62c0 git-svn-id: svn://svn.icms.temple.edu/lammps-ro/trunk@15194 f3b2605a-c512-4ea7-a41b-209d697bcdaa 2016-06-17 21:58:00 +00:00
530ede191b git-svn-id: svn://svn.icms.temple.edu/lammps-ro/trunk@15192 f3b2605a-c512-4ea7-a41b-209d697bcdaa 2016-06-17 21:56:42 +00:00
8147c8f742 git-svn-id: svn://svn.icms.temple.edu/lammps-ro/trunk@15191 f3b2605a-c512-4ea7-a41b-209d697bcdaa 2016-06-17 21:56:38 +00:00
503cd82065 git-svn-id: svn://svn.icms.temple.edu/lammps-ro/trunk@15190 f3b2605a-c512-4ea7-a41b-209d697bcdaa 2016-06-17 20:56:22 +00:00
b74ea86bcf git-svn-id: svn://svn.icms.temple.edu/lammps-ro/trunk@15189 f3b2605a-c512-4ea7-a41b-209d697bcdaa 2016-06-17 20:54:47 +00:00
e135e3ee79 git-svn-id: svn://svn.icms.temple.edu/lammps-ro/trunk@15188 f3b2605a-c512-4ea7-a41b-209d697bcdaa 2016-06-17 18:29:08 +00:00
951e7c916a git-svn-id: svn://svn.icms.temple.edu/lammps-ro/trunk@15187 f3b2605a-c512-4ea7-a41b-209d697bcdaa 2016-06-17 17:39:55 +00:00
e088eaa53b git-svn-id: svn://svn.icms.temple.edu/lammps-ro/trunk@15186 f3b2605a-c512-4ea7-a41b-209d697bcdaa 2016-06-17 16:45:57 +00:00
6cb38b17d8 git-svn-id: svn://svn.icms.temple.edu/lammps-ro/trunk@15185 f3b2605a-c512-4ea7-a41b-209d697bcdaa 2016-06-16 19:49:06 +00:00
a3df07b9fc git-svn-id: svn://svn.icms.temple.edu/lammps-ro/trunk@15184 f3b2605a-c512-4ea7-a41b-209d697bcdaa 2016-06-16 15:52:38 +00:00
6ba2664921 new orient/bcc for pull request #95
git-svn-id: svn://svn.icms.temple.edu/lammps-ro/trunk@15183 f3b2605a-c512-4ea7-a41b-209d697bcdaa
2016-06-16 15:51:17 +00:00
8330ef7f7e git-svn-id: svn://svn.icms.temple.edu/lammps-ro/trunk@15182 f3b2605a-c512-4ea7-a41b-209d697bcdaa 2016-06-16 15:46:02 +00:00
328224a298 git-svn-id: svn://svn.icms.temple.edu/lammps-ro/trunk@15181 f3b2605a-c512-4ea7-a41b-209d697bcdaa 2016-06-16 15:44:41 +00:00
14994c04b0 git-svn-id: svn://svn.icms.temple.edu/lammps-ro/trunk@15180 f3b2605a-c512-4ea7-a41b-209d697bcdaa 2016-06-15 23:10:12 +00:00
253bd4c335 git-svn-id: svn://svn.icms.temple.edu/lammps-ro/trunk@15179 f3b2605a-c512-4ea7-a41b-209d697bcdaa 2016-06-15 23:09:53 +00:00
94242eb591 git-svn-id: svn://svn.icms.temple.edu/lammps-ro/trunk@15178 f3b2605a-c512-4ea7-a41b-209d697bcdaa 2016-06-14 22:43:51 +00:00
c2c73fd8a3 git-svn-id: svn://svn.icms.temple.edu/lammps-ro/trunk@15176 f3b2605a-c512-4ea7-a41b-209d697bcdaa 2016-06-14 22:39:56 +00:00
58a091c773 git-svn-id: svn://svn.icms.temple.edu/lammps-ro/trunk@15175 f3b2605a-c512-4ea7-a41b-209d697bcdaa 2016-06-14 22:39:15 +00:00
4615a859c3 git-svn-id: svn://svn.icms.temple.edu/lammps-ro/trunk@15174 f3b2605a-c512-4ea7-a41b-209d697bcdaa 2016-06-14 22:36:31 +00:00
0b92c9f075 git-svn-id: svn://svn.icms.temple.edu/lammps-ro/trunk@15173 f3b2605a-c512-4ea7-a41b-209d697bcdaa 2016-06-14 22:35:59 +00:00
80aca27acd git-svn-id: svn://svn.icms.temple.edu/lammps-ro/trunk@15172 f3b2605a-c512-4ea7-a41b-209d697bcdaa 2016-06-14 22:33:59 +00:00
a7633c422c git-svn-id: svn://svn.icms.temple.edu/lammps-ro/trunk@15171 f3b2605a-c512-4ea7-a41b-209d697bcdaa 2016-06-14 22:32:35 +00:00
e8ea08fa8c git-svn-id: svn://svn.icms.temple.edu/lammps-ro/trunk@15170 f3b2605a-c512-4ea7-a41b-209d697bcdaa 2016-06-14 22:14:34 +00:00
c6e7b0c8b9 git-svn-id: svn://svn.icms.temple.edu/lammps-ro/trunk@15169 f3b2605a-c512-4ea7-a41b-209d697bcdaa 2016-06-14 21:25:00 +00:00
ffbce3c10c git-svn-id: svn://svn.icms.temple.edu/lammps-ro/trunk@15168 f3b2605a-c512-4ea7-a41b-209d697bcdaa 2016-06-14 16:40:10 +00:00
422dfdc1f7 git-svn-id: svn://svn.icms.temple.edu/lammps-ro/trunk@15167 f3b2605a-c512-4ea7-a41b-209d697bcdaa 2016-06-14 16:34:58 +00:00
2dbcfdc70c git-svn-id: svn://svn.icms.temple.edu/lammps-ro/trunk@15166 f3b2605a-c512-4ea7-a41b-209d697bcdaa 2016-06-14 16:33:40 +00:00
ecffbbe531 git-svn-id: svn://svn.icms.temple.edu/lammps-ro/trunk@15165 f3b2605a-c512-4ea7-a41b-209d697bcdaa 2016-06-14 16:32:54 +00:00
0e719ed2ef git-svn-id: svn://svn.icms.temple.edu/lammps-ro/trunk@15164 f3b2605a-c512-4ea7-a41b-209d697bcdaa 2016-06-14 14:16:24 +00:00
863a3d3319 git-svn-id: svn://svn.icms.temple.edu/lammps-ro/trunk@15163 f3b2605a-c512-4ea7-a41b-209d697bcdaa 2016-06-14 13:59:39 +00:00
a4b82a95e9 git-svn-id: svn://svn.icms.temple.edu/lammps-ro/trunk@15162 f3b2605a-c512-4ea7-a41b-209d697bcdaa 2016-06-14 13:58:49 +00:00
5631254f2f git-svn-id: svn://svn.icms.temple.edu/lammps-ro/trunk@15161 f3b2605a-c512-4ea7-a41b-209d697bcdaa 2016-06-09 21:55:26 +00:00
84de575cc7 fixed case where no axial components are active
git-svn-id: svn://svn.icms.temple.edu/lammps-ro/trunk@15160 f3b2605a-c512-4ea7-a41b-209d697bcdaa
2016-06-09 18:31:54 +00:00
50ac419d3f git-svn-id: svn://svn.icms.temple.edu/lammps-ro/trunk@15159 f3b2605a-c512-4ea7-a41b-209d697bcdaa 2016-06-08 23:15:55 +00:00
253f93a579 Adding kokkos half list with ghosts
git-svn-id: svn://svn.icms.temple.edu/lammps-ro/trunk@15158 f3b2605a-c512-4ea7-a41b-209d697bcdaa
2016-06-08 17:01:57 +00:00
84f65fe441 git-svn-id: svn://svn.icms.temple.edu/lammps-ro/trunk@15157 f3b2605a-c512-4ea7-a41b-209d697bcdaa 2016-06-08 15:41:04 +00:00
fdab2d6cb0 git-svn-id: svn://svn.icms.temple.edu/lammps-ro/trunk@15156 f3b2605a-c512-4ea7-a41b-209d697bcdaa 2016-06-08 15:41:00 +00:00
6709e70d18 git-svn-id: svn://svn.icms.temple.edu/lammps-ro/trunk@15154 f3b2605a-c512-4ea7-a41b-209d697bcdaa 2016-06-07 18:33:50 +00:00
d84b73a6b0 git-svn-id: svn://svn.icms.temple.edu/lammps-ro/trunk@15153 f3b2605a-c512-4ea7-a41b-209d697bcdaa 2016-06-07 18:33:11 +00:00
3497d6382c git-svn-id: svn://svn.icms.temple.edu/lammps-ro/trunk@15151 f3b2605a-c512-4ea7-a41b-209d697bcdaa 2016-06-07 18:26:15 +00:00
ab7e896a76 git-svn-id: svn://svn.icms.temple.edu/lammps-ro/trunk@15150 f3b2605a-c512-4ea7-a41b-209d697bcdaa 2016-06-07 18:18:53 +00:00
0129ff9696 git-svn-id: svn://svn.icms.temple.edu/lammps-ro/trunk@15149 f3b2605a-c512-4ea7-a41b-209d697bcdaa 2016-06-07 18:17:05 +00:00
b5abf9342b git-svn-id: svn://svn.icms.temple.edu/lammps-ro/trunk@15148 f3b2605a-c512-4ea7-a41b-209d697bcdaa 2016-06-07 18:16:44 +00:00
0288bdba40 git-svn-id: svn://svn.icms.temple.edu/lammps-ro/trunk@15147 f3b2605a-c512-4ea7-a41b-209d697bcdaa 2016-06-07 18:03:42 +00:00
2f225bbc3a git-svn-id: svn://svn.icms.temple.edu/lammps-ro/trunk@15146 f3b2605a-c512-4ea7-a41b-209d697bcdaa 2016-06-07 18:03:05 +00:00
f3d5260813 git-svn-id: svn://svn.icms.temple.edu/lammps-ro/trunk@15145 f3b2605a-c512-4ea7-a41b-209d697bcdaa 2016-06-07 18:00:54 +00:00
a5fef35ebd git-svn-id: svn://svn.icms.temple.edu/lammps-ro/trunk@15144 f3b2605a-c512-4ea7-a41b-209d697bcdaa 2016-06-07 18:00:49 +00:00
b107958e80 git-svn-id: svn://svn.icms.temple.edu/lammps-ro/trunk@15143 f3b2605a-c512-4ea7-a41b-209d697bcdaa 2016-06-07 17:57:47 +00:00
7a51ae1311 git-svn-id: svn://svn.icms.temple.edu/lammps-ro/trunk@15142 f3b2605a-c512-4ea7-a41b-209d697bcdaa 2016-06-07 17:56:45 +00:00
214e8876c0 git-svn-id: svn://svn.icms.temple.edu/lammps-ro/trunk@15141 f3b2605a-c512-4ea7-a41b-209d697bcdaa 2016-06-07 17:56:27 +00:00
8b17319fce git-svn-id: svn://svn.icms.temple.edu/lammps-ro/trunk@15140 f3b2605a-c512-4ea7-a41b-209d697bcdaa 2016-06-07 17:54:53 +00:00
84514d345b git-svn-id: svn://svn.icms.temple.edu/lammps-ro/trunk@15139 f3b2605a-c512-4ea7-a41b-209d697bcdaa 2016-06-07 17:52:40 +00:00
c3f0833b48 git-svn-id: svn://svn.icms.temple.edu/lammps-ro/trunk@15138 f3b2605a-c512-4ea7-a41b-209d697bcdaa 2016-06-07 17:52:00 +00:00
8a7fb8e7e0 git-svn-id: svn://svn.icms.temple.edu/lammps-ro/trunk@15137 f3b2605a-c512-4ea7-a41b-209d697bcdaa 2016-06-07 16:21:00 +00:00
1afe415c07 git-svn-id: svn://svn.icms.temple.edu/lammps-ro/trunk@15135 f3b2605a-c512-4ea7-a41b-209d697bcdaa 2016-06-07 16:07:03 +00:00
c8c29cb084 git-svn-id: svn://svn.icms.temple.edu/lammps-ro/trunk@15134 f3b2605a-c512-4ea7-a41b-209d697bcdaa 2016-06-07 16:06:24 +00:00
27d7c598bf git-svn-id: svn://svn.icms.temple.edu/lammps-ro/trunk@15133 f3b2605a-c512-4ea7-a41b-209d697bcdaa 2016-06-07 16:05:14 +00:00
9c5655db7c git-svn-id: svn://svn.icms.temple.edu/lammps-ro/trunk@15130 f3b2605a-c512-4ea7-a41b-209d697bcdaa 2016-06-07 15:51:43 +00:00
d1428accee git-svn-id: svn://svn.icms.temple.edu/lammps-ro/trunk@15129 f3b2605a-c512-4ea7-a41b-209d697bcdaa 2016-06-07 15:49:45 +00:00
0f5dcedee1 git-svn-id: svn://svn.icms.temple.edu/lammps-ro/trunk@15128 f3b2605a-c512-4ea7-a41b-209d697bcdaa 2016-06-07 15:49:15 +00:00
78c0df7de2 git-svn-id: svn://svn.icms.temple.edu/lammps-ro/trunk@15127 f3b2605a-c512-4ea7-a41b-209d697bcdaa 2016-06-07 15:41:19 +00:00
22fe6d5290 git-svn-id: svn://svn.icms.temple.edu/lammps-ro/trunk@15126 f3b2605a-c512-4ea7-a41b-209d697bcdaa 2016-06-07 15:41:03 +00:00
a8ae4fa419 git-svn-id: svn://svn.icms.temple.edu/lammps-ro/trunk@15125 f3b2605a-c512-4ea7-a41b-209d697bcdaa 2016-06-07 15:39:12 +00:00
7c55da3879 git-svn-id: svn://svn.icms.temple.edu/lammps-ro/trunk@15124 f3b2605a-c512-4ea7-a41b-209d697bcdaa 2016-06-07 15:37:34 +00:00
63c80930f1 git-svn-id: svn://svn.icms.temple.edu/lammps-ro/trunk@15123 f3b2605a-c512-4ea7-a41b-209d697bcdaa 2016-06-07 15:32:15 +00:00
946de18d47 git-svn-id: svn://svn.icms.temple.edu/lammps-ro/trunk@15122 f3b2605a-c512-4ea7-a41b-209d697bcdaa 2016-06-07 15:32:00 +00:00
59497c1076 git-svn-id: svn://svn.icms.temple.edu/lammps-ro/trunk@15121 f3b2605a-c512-4ea7-a41b-209d697bcdaa 2016-06-07 15:31:35 +00:00
7cb2b01ebd git-svn-id: svn://svn.icms.temple.edu/lammps-ro/trunk@15120 f3b2605a-c512-4ea7-a41b-209d697bcdaa 2016-06-07 15:29:39 +00:00
9fbe666677 git-svn-id: svn://svn.icms.temple.edu/lammps-ro/trunk@15119 f3b2605a-c512-4ea7-a41b-209d697bcdaa 2016-06-07 15:20:33 +00:00
2927a9aee1 git-svn-id: svn://svn.icms.temple.edu/lammps-ro/trunk@15118 f3b2605a-c512-4ea7-a41b-209d697bcdaa 2016-06-07 15:20:19 +00:00
97ae23af75 git-svn-id: svn://svn.icms.temple.edu/lammps-ro/trunk@15117 f3b2605a-c512-4ea7-a41b-209d697bcdaa 2016-06-07 14:49:54 +00:00
78c5aafccf git-svn-id: svn://svn.icms.temple.edu/lammps-ro/trunk@15116 f3b2605a-c512-4ea7-a41b-209d697bcdaa 2016-06-07 14:48:10 +00:00
930e2aa033 git-svn-id: svn://svn.icms.temple.edu/lammps-ro/trunk@15115 f3b2605a-c512-4ea7-a41b-209d697bcdaa 2016-06-07 13:44:22 +00:00
34111922f0 git-svn-id: svn://svn.icms.temple.edu/lammps-ro/trunk@15113 f3b2605a-c512-4ea7-a41b-209d697bcdaa 2016-06-06 23:41:23 +00:00
c2cca5dd62 git-svn-id: svn://svn.icms.temple.edu/lammps-ro/trunk@15112 f3b2605a-c512-4ea7-a41b-209d697bcdaa 2016-06-06 23:40:32 +00:00
921133d65d git-svn-id: svn://svn.icms.temple.edu/lammps-ro/trunk@15110 f3b2605a-c512-4ea7-a41b-209d697bcdaa 2016-06-06 23:22:33 +00:00
0f3e4182e8 git-svn-id: svn://svn.icms.temple.edu/lammps-ro/trunk@15109 f3b2605a-c512-4ea7-a41b-209d697bcdaa 2016-06-06 23:21:39 +00:00
1a52795e1d git-svn-id: svn://svn.icms.temple.edu/lammps-ro/trunk@15108 f3b2605a-c512-4ea7-a41b-209d697bcdaa 2016-06-06 23:17:05 +00:00
bb45a5e492 git-svn-id: svn://svn.icms.temple.edu/lammps-ro/trunk@15105 f3b2605a-c512-4ea7-a41b-209d697bcdaa 2016-06-02 14:14:28 +00:00
195c521b8a git-svn-id: svn://svn.icms.temple.edu/lammps-ro/trunk@15104 f3b2605a-c512-4ea7-a41b-209d697bcdaa 2016-06-02 14:14:17 +00:00
10be4c06bd git-svn-id: svn://svn.icms.temple.edu/lammps-ro/trunk@15103 f3b2605a-c512-4ea7-a41b-209d697bcdaa 2016-06-02 14:12:27 +00:00
0c89f35243 git-svn-id: svn://svn.icms.temple.edu/lammps-ro/trunk@15102 f3b2605a-c512-4ea7-a41b-209d697bcdaa 2016-06-02 14:10:24 +00:00
d26d71fde1 git-svn-id: svn://svn.icms.temple.edu/lammps-ro/trunk@15101 f3b2605a-c512-4ea7-a41b-209d697bcdaa 2016-06-02 14:09:19 +00:00
15068b6354 git-svn-id: svn://svn.icms.temple.edu/lammps-ro/trunk@15100 f3b2605a-c512-4ea7-a41b-209d697bcdaa 2016-06-02 14:08:20 +00:00
01642458f0 git-svn-id: svn://svn.icms.temple.edu/lammps-ro/trunk@15099 f3b2605a-c512-4ea7-a41b-209d697bcdaa 2016-06-02 14:07:44 +00:00
2a21b1ab56 git-svn-id: svn://svn.icms.temple.edu/lammps-ro/trunk@15098 f3b2605a-c512-4ea7-a41b-209d697bcdaa 2016-06-02 14:06:48 +00:00
126ae910b8 git-svn-id: svn://svn.icms.temple.edu/lammps-ro/trunk@15097 f3b2605a-c512-4ea7-a41b-209d697bcdaa 2016-06-02 14:05:40 +00:00
a1d64b989e git-svn-id: svn://svn.icms.temple.edu/lammps-ro/trunk@15096 f3b2605a-c512-4ea7-a41b-209d697bcdaa 2016-06-02 14:05:29 +00:00
e653da12a6 git-svn-id: svn://svn.icms.temple.edu/lammps-ro/trunk@15095 f3b2605a-c512-4ea7-a41b-209d697bcdaa 2016-06-02 14:03:00 +00:00
5c19a0d788 git-svn-id: svn://svn.icms.temple.edu/lammps-ro/trunk@15094 f3b2605a-c512-4ea7-a41b-209d697bcdaa 2016-06-01 15:07:48 +00:00
08c8d65fea git-svn-id: svn://svn.icms.temple.edu/lammps-ro/trunk@15092 f3b2605a-c512-4ea7-a41b-209d697bcdaa 2016-05-31 16:53:09 +00:00
9182047905 git-svn-id: svn://svn.icms.temple.edu/lammps-ro/trunk@15091 f3b2605a-c512-4ea7-a41b-209d697bcdaa 2016-05-31 16:52:28 +00:00
0a56b26fe1 git-svn-id: svn://svn.icms.temple.edu/lammps-ro/trunk@15090 f3b2605a-c512-4ea7-a41b-209d697bcdaa 2016-05-31 16:51:19 +00:00
414c1b011f git-svn-id: svn://svn.icms.temple.edu/lammps-ro/trunk@15089 f3b2605a-c512-4ea7-a41b-209d697bcdaa 2016-05-31 16:50:28 +00:00
17e44d1313 git-svn-id: svn://svn.icms.temple.edu/lammps-ro/trunk@15088 f3b2605a-c512-4ea7-a41b-209d697bcdaa 2016-05-31 16:50:06 +00:00
194a512168 git-svn-id: svn://svn.icms.temple.edu/lammps-ro/trunk@15087 f3b2605a-c512-4ea7-a41b-209d697bcdaa 2016-05-31 16:45:20 +00:00
3c11e66b5a git-svn-id: svn://svn.icms.temple.edu/lammps-ro/trunk@15086 f3b2605a-c512-4ea7-a41b-209d697bcdaa 2016-05-31 16:41:25 +00:00
4c107f450c git-svn-id: svn://svn.icms.temple.edu/lammps-ro/trunk@15085 f3b2605a-c512-4ea7-a41b-209d697bcdaa 2016-05-31 16:39:57 +00:00
ac14f5bc36 git-svn-id: svn://svn.icms.temple.edu/lammps-ro/trunk@15084 f3b2605a-c512-4ea7-a41b-209d697bcdaa 2016-05-31 16:38:36 +00:00
876b033ea8 git-svn-id: svn://svn.icms.temple.edu/lammps-ro/trunk@15083 f3b2605a-c512-4ea7-a41b-209d697bcdaa 2016-05-31 16:33:07 +00:00
7e7450948c git-svn-id: svn://svn.icms.temple.edu/lammps-ro/trunk@15082 f3b2605a-c512-4ea7-a41b-209d697bcdaa 2016-05-31 16:26:20 +00:00
fde2d583e5 git-svn-id: svn://svn.icms.temple.edu/lammps-ro/trunk@15081 f3b2605a-c512-4ea7-a41b-209d697bcdaa 2016-05-31 16:21:23 +00:00
04e401ae4b git-svn-id: svn://svn.icms.temple.edu/lammps-ro/trunk@15080 f3b2605a-c512-4ea7-a41b-209d697bcdaa 2016-05-31 16:20:32 +00:00
5b94aa3ae8 git-svn-id: svn://svn.icms.temple.edu/lammps-ro/trunk@15079 f3b2605a-c512-4ea7-a41b-209d697bcdaa 2016-05-31 16:08:29 +00:00
32ba20085d git-svn-id: svn://svn.icms.temple.edu/lammps-ro/trunk@15078 f3b2605a-c512-4ea7-a41b-209d697bcdaa 2016-05-31 16:07:15 +00:00
c91873924b git-svn-id: svn://svn.icms.temple.edu/lammps-ro/trunk@15077 f3b2605a-c512-4ea7-a41b-209d697bcdaa 2016-05-31 16:04:35 +00:00
cbc5b8bdae git-svn-id: svn://svn.icms.temple.edu/lammps-ro/trunk@15076 f3b2605a-c512-4ea7-a41b-209d697bcdaa 2016-05-31 15:30:30 +00:00
27725bea2f git-svn-id: svn://svn.icms.temple.edu/lammps-ro/trunk@15075 f3b2605a-c512-4ea7-a41b-209d697bcdaa 2016-05-31 15:03:16 +00:00
6567493b49 git-svn-id: svn://svn.icms.temple.edu/lammps-ro/trunk@15074 f3b2605a-c512-4ea7-a41b-209d697bcdaa 2016-05-31 15:02:48 +00:00
6f75d4389c git-svn-id: svn://svn.icms.temple.edu/lammps-ro/trunk@15073 f3b2605a-c512-4ea7-a41b-209d697bcdaa 2016-05-31 14:55:17 +00:00
402b7a7231 git-svn-id: svn://svn.icms.temple.edu/lammps-ro/trunk@15072 f3b2605a-c512-4ea7-a41b-209d697bcdaa 2016-05-31 14:53:34 +00:00
9d3d6193f3 git-svn-id: svn://svn.icms.temple.edu/lammps-ro/trunk@15071 f3b2605a-c512-4ea7-a41b-209d697bcdaa 2016-05-31 14:49:49 +00:00
0a01592d0a git-svn-id: svn://svn.icms.temple.edu/lammps-ro/trunk@15070 f3b2605a-c512-4ea7-a41b-209d697bcdaa 2016-05-31 14:49:26 +00:00
8fa54096d3 git-svn-id: svn://svn.icms.temple.edu/lammps-ro/trunk@15069 f3b2605a-c512-4ea7-a41b-209d697bcdaa 2016-05-31 14:48:46 +00:00
d7c5025530 git-svn-id: svn://svn.icms.temple.edu/lammps-ro/trunk@15068 f3b2605a-c512-4ea7-a41b-209d697bcdaa 2016-05-31 13:58:37 +00:00
71c36003b9 Added Axel's improvement
git-svn-id: svn://svn.icms.temple.edu/lammps-ro/trunk@15067 f3b2605a-c512-4ea7-a41b-209d697bcdaa
2016-05-26 17:27:15 +00:00
edbb8bc31d Fixed uninitialized variables
git-svn-id: svn://svn.icms.temple.edu/lammps-ro/trunk@15066 f3b2605a-c512-4ea7-a41b-209d697bcdaa
2016-05-25 20:09:34 +00:00
d60fbe4c22 Blocked fix rigid
git-svn-id: svn://svn.icms.temple.edu/lammps-ro/trunk@15065 f3b2605a-c512-4ea7-a41b-209d697bcdaa
2016-05-24 23:34:10 +00:00
97b8a30b72 Blocked fix rigid
git-svn-id: svn://svn.icms.temple.edu/lammps-ro/trunk@15064 f3b2605a-c512-4ea7-a41b-209d697bcdaa
2016-05-24 23:31:31 +00:00
bbdd574e6d Removed some memory leaks is SSA neighbor list
git-svn-id: svn://svn.icms.temple.edu/lammps-ro/trunk@15063 f3b2605a-c512-4ea7-a41b-209d697bcdaa
2016-05-24 22:54:03 +00:00
f1d424aba7 git-svn-id: svn://svn.icms.temple.edu/lammps-ro/trunk@15061 f3b2605a-c512-4ea7-a41b-209d697bcdaa 2016-05-17 17:02:58 +00:00
eb6d23effc git-svn-id: svn://svn.icms.temple.edu/lammps-ro/trunk@15059 f3b2605a-c512-4ea7-a41b-209d697bcdaa 2016-05-13 15:48:05 +00:00
723f02501f git-svn-id: svn://svn.icms.temple.edu/lammps-ro/trunk@15058 f3b2605a-c512-4ea7-a41b-209d697bcdaa 2016-05-13 15:47:27 +00:00
6334768dd1 git-svn-id: svn://svn.icms.temple.edu/lammps-ro/trunk@15057 f3b2605a-c512-4ea7-a41b-209d697bcdaa 2016-05-13 15:46:48 +00:00
39d713b5a3 git-svn-id: svn://svn.icms.temple.edu/lammps-ro/trunk@15055 f3b2605a-c512-4ea7-a41b-209d697bcdaa 2016-05-13 15:23:49 +00:00
c126ebf50d git-svn-id: svn://svn.icms.temple.edu/lammps-ro/trunk@15054 f3b2605a-c512-4ea7-a41b-209d697bcdaa 2016-05-13 15:23:05 +00:00
74052b0b86 git-svn-id: svn://svn.icms.temple.edu/lammps-ro/trunk@15053 f3b2605a-c512-4ea7-a41b-209d697bcdaa 2016-05-12 16:33:14 +00:00
8976a2ffff git-svn-id: svn://svn.icms.temple.edu/lammps-ro/trunk@15052 f3b2605a-c512-4ea7-a41b-209d697bcdaa 2016-05-12 16:32:15 +00:00
87e06310cb git-svn-id: svn://svn.icms.temple.edu/lammps-ro/trunk@15051 f3b2605a-c512-4ea7-a41b-209d697bcdaa 2016-05-12 16:29:09 +00:00
62b95b6b34 git-svn-id: svn://svn.icms.temple.edu/lammps-ro/trunk@15050 f3b2605a-c512-4ea7-a41b-209d697bcdaa 2016-05-12 16:28:22 +00:00
7ce9b111eb git-svn-id: svn://svn.icms.temple.edu/lammps-ro/trunk@15046 f3b2605a-c512-4ea7-a41b-209d697bcdaa 2016-05-12 14:17:32 +00:00
f4d7009204 git-svn-id: svn://svn.icms.temple.edu/lammps-ro/trunk@15045 f3b2605a-c512-4ea7-a41b-209d697bcdaa 2016-05-12 14:17:27 +00:00
3fc88d9205 git-svn-id: svn://svn.icms.temple.edu/lammps-ro/trunk@15044 f3b2605a-c512-4ea7-a41b-209d697bcdaa 2016-05-12 14:16:38 +00:00
58befb3a2e git-svn-id: svn://svn.icms.temple.edu/lammps-ro/trunk@15043 f3b2605a-c512-4ea7-a41b-209d697bcdaa 2016-05-12 14:14:39 +00:00
b013a40965 git-svn-id: svn://svn.icms.temple.edu/lammps-ro/trunk@15042 f3b2605a-c512-4ea7-a41b-209d697bcdaa 2016-05-12 14:02:27 +00:00
9653976f05 git-svn-id: svn://svn.icms.temple.edu/lammps-ro/trunk@15041 f3b2605a-c512-4ea7-a41b-209d697bcdaa 2016-05-12 13:58:19 +00:00
212f64d2f1 git-svn-id: svn://svn.icms.temple.edu/lammps-ro/trunk@15040 f3b2605a-c512-4ea7-a41b-209d697bcdaa 2016-05-12 13:54:17 +00:00
c3a95c782b git-svn-id: svn://svn.icms.temple.edu/lammps-ro/trunk@15039 f3b2605a-c512-4ea7-a41b-209d697bcdaa 2016-05-12 13:53:40 +00:00
dfb0f528c0 git-svn-id: svn://svn.icms.temple.edu/lammps-ro/trunk@15038 f3b2605a-c512-4ea7-a41b-209d697bcdaa 2016-05-12 13:52:42 +00:00
37ede68299 git-svn-id: svn://svn.icms.temple.edu/lammps-ro/trunk@15036 f3b2605a-c512-4ea7-a41b-209d697bcdaa 2016-05-11 20:44:46 +00:00
f7637bf64b git-svn-id: svn://svn.icms.temple.edu/lammps-ro/trunk@15035 f3b2605a-c512-4ea7-a41b-209d697bcdaa 2016-05-11 20:44:08 +00:00
b0eea89503 git-svn-id: svn://svn.icms.temple.edu/lammps-ro/trunk@15034 f3b2605a-c512-4ea7-a41b-209d697bcdaa 2016-05-11 20:38:08 +00:00
64b58c743f git-svn-id: svn://svn.icms.temple.edu/lammps-ro/trunk@15033 f3b2605a-c512-4ea7-a41b-209d697bcdaa 2016-05-11 20:37:23 +00:00
757c32500a git-svn-id: svn://svn.icms.temple.edu/lammps-ro/trunk@15032 f3b2605a-c512-4ea7-a41b-209d697bcdaa 2016-05-11 20:36:56 +00:00
8caee4c0e1 git-svn-id: svn://svn.icms.temple.edu/lammps-ro/trunk@15031 f3b2605a-c512-4ea7-a41b-209d697bcdaa 2016-05-11 20:36:18 +00:00
3e4b83abaa git-svn-id: svn://svn.icms.temple.edu/lammps-ro/trunk@15030 f3b2605a-c512-4ea7-a41b-209d697bcdaa 2016-05-11 20:34:40 +00:00
51884d562f git-svn-id: svn://svn.icms.temple.edu/lammps-ro/trunk@15029 f3b2605a-c512-4ea7-a41b-209d697bcdaa 2016-05-11 20:34:01 +00:00
4201f775a3 git-svn-id: svn://svn.icms.temple.edu/lammps-ro/trunk@15028 f3b2605a-c512-4ea7-a41b-209d697bcdaa 2016-05-11 20:27:35 +00:00
451593b9f0 git-svn-id: svn://svn.icms.temple.edu/lammps-ro/trunk@15027 f3b2605a-c512-4ea7-a41b-209d697bcdaa 2016-05-11 20:26:11 +00:00
21f3a8318f git-svn-id: svn://svn.icms.temple.edu/lammps-ro/trunk@15026 f3b2605a-c512-4ea7-a41b-209d697bcdaa 2016-05-11 20:25:39 +00:00
6e5353e844 git-svn-id: svn://svn.icms.temple.edu/lammps-ro/trunk@15025 f3b2605a-c512-4ea7-a41b-209d697bcdaa 2016-05-11 20:25:03 +00:00
c0d0b75923 git-svn-id: svn://svn.icms.temple.edu/lammps-ro/trunk@15024 f3b2605a-c512-4ea7-a41b-209d697bcdaa 2016-05-10 23:09:22 +00:00
e2790b1c63 git-svn-id: svn://svn.icms.temple.edu/lammps-ro/trunk@15023 f3b2605a-c512-4ea7-a41b-209d697bcdaa 2016-05-10 23:06:57 +00:00
59e44bf4bf git-svn-id: svn://svn.icms.temple.edu/lammps-ro/trunk@15021 f3b2605a-c512-4ea7-a41b-209d697bcdaa 2016-05-10 22:42:15 +00:00
1463cb2598 git-svn-id: svn://svn.icms.temple.edu/lammps-ro/trunk@15020 f3b2605a-c512-4ea7-a41b-209d697bcdaa 2016-05-10 22:39:37 +00:00
5253f2aae1 git-svn-id: svn://svn.icms.temple.edu/lammps-ro/trunk@15019 f3b2605a-c512-4ea7-a41b-209d697bcdaa 2016-05-10 22:26:37 +00:00
8c83504db4 git-svn-id: svn://svn.icms.temple.edu/lammps-ro/trunk@15017 f3b2605a-c512-4ea7-a41b-209d697bcdaa 2016-05-10 21:42:31 +00:00
a6d45861d2 git-svn-id: svn://svn.icms.temple.edu/lammps-ro/trunk@15016 f3b2605a-c512-4ea7-a41b-209d697bcdaa 2016-05-10 21:41:51 +00:00
b78a35909a git-svn-id: svn://svn.icms.temple.edu/lammps-ro/trunk@14999 f3b2605a-c512-4ea7-a41b-209d697bcdaa 2016-05-10 20:08:59 +00:00
f12fdf7e17 git-svn-id: svn://svn.icms.temple.edu/lammps-ro/trunk@14998 f3b2605a-c512-4ea7-a41b-209d697bcdaa 2016-05-10 20:08:22 +00:00
532b5e9787 git-svn-id: svn://svn.icms.temple.edu/lammps-ro/trunk@14997 f3b2605a-c512-4ea7-a41b-209d697bcdaa 2016-05-10 20:05:38 +00:00
24f1a9795b git-svn-id: svn://svn.icms.temple.edu/lammps-ro/trunk@14996 f3b2605a-c512-4ea7-a41b-209d697bcdaa 2016-05-10 20:05:15 +00:00
3060e9f11b git-svn-id: svn://svn.icms.temple.edu/lammps-ro/trunk@14995 f3b2605a-c512-4ea7-a41b-209d697bcdaa 2016-05-10 20:05:08 +00:00
2b75c78f69 git-svn-id: svn://svn.icms.temple.edu/lammps-ro/trunk@14994 f3b2605a-c512-4ea7-a41b-209d697bcdaa 2016-05-10 20:03:52 +00:00
244889aed4 added escape to special character
git-svn-id: svn://svn.icms.temple.edu/lammps-ro/trunk@14993 f3b2605a-c512-4ea7-a41b-209d697bcdaa
2016-05-10 18:05:50 +00:00
6d3f5ddd85 git-svn-id: svn://svn.icms.temple.edu/lammps-ro/trunk@14991 f3b2605a-c512-4ea7-a41b-209d697bcdaa 2016-05-10 15:10:49 +00:00
e0650f659d git-svn-id: svn://svn.icms.temple.edu/lammps-ro/trunk@14990 f3b2605a-c512-4ea7-a41b-209d697bcdaa 2016-05-10 15:10:11 +00:00
ed31297f86 git-svn-id: svn://svn.icms.temple.edu/lammps-ro/trunk@14989 f3b2605a-c512-4ea7-a41b-209d697bcdaa 2016-05-10 15:04:16 +00:00
432926229e git-svn-id: svn://svn.icms.temple.edu/lammps-ro/trunk@14988 f3b2605a-c512-4ea7-a41b-209d697bcdaa 2016-05-10 15:02:35 +00:00
2c8227b10c git-svn-id: svn://svn.icms.temple.edu/lammps-ro/trunk@14987 f3b2605a-c512-4ea7-a41b-209d697bcdaa 2016-05-10 15:01:22 +00:00
fbcffab45d git-svn-id: svn://svn.icms.temple.edu/lammps-ro/trunk@14986 f3b2605a-c512-4ea7-a41b-209d697bcdaa 2016-05-10 15:00:12 +00:00
e8fe19dc71 git-svn-id: svn://svn.icms.temple.edu/lammps-ro/trunk@14985 f3b2605a-c512-4ea7-a41b-209d697bcdaa 2016-05-10 14:52:43 +00:00
df885d59ff git-svn-id: svn://svn.icms.temple.edu/lammps-ro/trunk@14984 f3b2605a-c512-4ea7-a41b-209d697bcdaa 2016-05-10 14:49:23 +00:00
03cb787025 git-svn-id: svn://svn.icms.temple.edu/lammps-ro/trunk@14983 f3b2605a-c512-4ea7-a41b-209d697bcdaa 2016-05-10 14:39:05 +00:00
2d4910b40d git-svn-id: svn://svn.icms.temple.edu/lammps-ro/trunk@14982 f3b2605a-c512-4ea7-a41b-209d697bcdaa 2016-05-10 14:07:05 +00:00
c2f479c5fc git-svn-id: svn://svn.icms.temple.edu/lammps-ro/trunk@14981 f3b2605a-c512-4ea7-a41b-209d697bcdaa 2016-05-10 14:05:06 +00:00
afb7a2ad39 git-svn-id: svn://svn.icms.temple.edu/lammps-ro/trunk@14980 f3b2605a-c512-4ea7-a41b-209d697bcdaa 2016-05-10 14:00:33 +00:00
43b2e3ecf6 git-svn-id: svn://svn.icms.temple.edu/lammps-ro/trunk@14979 f3b2605a-c512-4ea7-a41b-209d697bcdaa 2016-05-09 23:31:53 +00:00
ba24c077f7 git-svn-id: svn://svn.icms.temple.edu/lammps-ro/trunk@14978 f3b2605a-c512-4ea7-a41b-209d697bcdaa 2016-05-09 22:06:50 +00:00
eca85415be git-svn-id: svn://svn.icms.temple.edu/lammps-ro/trunk@14977 f3b2605a-c512-4ea7-a41b-209d697bcdaa 2016-05-09 22:05:44 +00:00
551feff776 git-svn-id: svn://svn.icms.temple.edu/lammps-ro/trunk@14976 f3b2605a-c512-4ea7-a41b-209d697bcdaa 2016-05-09 22:05:30 +00:00
f35fdb242c git-svn-id: svn://svn.icms.temple.edu/lammps-ro/trunk@14975 f3b2605a-c512-4ea7-a41b-209d697bcdaa 2016-05-09 21:42:26 +00:00
3759b7e2dd git-svn-id: svn://svn.icms.temple.edu/lammps-ro/trunk@14974 f3b2605a-c512-4ea7-a41b-209d697bcdaa 2016-05-09 21:24:22 +00:00
a7cab80347 git-svn-id: svn://svn.icms.temple.edu/lammps-ro/trunk@14973 f3b2605a-c512-4ea7-a41b-209d697bcdaa 2016-05-09 21:18:02 +00:00
69ffae65d0 git-svn-id: svn://svn.icms.temple.edu/lammps-ro/trunk@14972 f3b2605a-c512-4ea7-a41b-209d697bcdaa 2016-05-09 21:12:06 +00:00
410c743f77 git-svn-id: svn://svn.icms.temple.edu/lammps-ro/trunk@14971 f3b2605a-c512-4ea7-a41b-209d697bcdaa 2016-05-09 21:11:20 +00:00
7baaf2dfe2 git-svn-id: svn://svn.icms.temple.edu/lammps-ro/trunk@14970 f3b2605a-c512-4ea7-a41b-209d697bcdaa 2016-05-09 20:20:41 +00:00
04177fcd07 git-svn-id: svn://svn.icms.temple.edu/lammps-ro/trunk@14969 f3b2605a-c512-4ea7-a41b-209d697bcdaa 2016-05-09 20:19:49 +00:00
61c8946720 git-svn-id: svn://svn.icms.temple.edu/lammps-ro/trunk@14968 f3b2605a-c512-4ea7-a41b-209d697bcdaa 2016-05-09 20:18:18 +00:00
7376b3393f git-svn-id: svn://svn.icms.temple.edu/lammps-ro/trunk@14967 f3b2605a-c512-4ea7-a41b-209d697bcdaa 2016-05-09 20:14:31 +00:00
8d18bc0e60 git-svn-id: svn://svn.icms.temple.edu/lammps-ro/trunk@14966 f3b2605a-c512-4ea7-a41b-209d697bcdaa 2016-05-09 20:13:19 +00:00
baa39ad263 git-svn-id: svn://svn.icms.temple.edu/lammps-ro/trunk@14965 f3b2605a-c512-4ea7-a41b-209d697bcdaa 2016-05-09 20:08:38 +00:00
d3d29d35af git-svn-id: svn://svn.icms.temple.edu/lammps-ro/trunk@14964 f3b2605a-c512-4ea7-a41b-209d697bcdaa 2016-05-09 20:07:39 +00:00
a06ff7f880 git-svn-id: svn://svn.icms.temple.edu/lammps-ro/trunk@14963 f3b2605a-c512-4ea7-a41b-209d697bcdaa 2016-05-09 19:57:18 +00:00
b611b8e056 git-svn-id: svn://svn.icms.temple.edu/lammps-ro/trunk@14962 f3b2605a-c512-4ea7-a41b-209d697bcdaa 2016-05-09 19:55:49 +00:00
1d3f94fbb2 git-svn-id: svn://svn.icms.temple.edu/lammps-ro/trunk@14961 f3b2605a-c512-4ea7-a41b-209d697bcdaa 2016-05-09 19:50:19 +00:00
e03b050785 git-svn-id: svn://svn.icms.temple.edu/lammps-ro/trunk@14960 f3b2605a-c512-4ea7-a41b-209d697bcdaa 2016-05-09 19:49:32 +00:00
3761006af3 git-svn-id: svn://svn.icms.temple.edu/lammps-ro/trunk@14959 f3b2605a-c512-4ea7-a41b-209d697bcdaa 2016-05-09 19:44:17 +00:00
27dfeee45d git-svn-id: svn://svn.icms.temple.edu/lammps-ro/trunk@14958 f3b2605a-c512-4ea7-a41b-209d697bcdaa 2016-05-09 19:31:53 +00:00
5b2d4fe6b1 git-svn-id: svn://svn.icms.temple.edu/lammps-ro/trunk@14957 f3b2605a-c512-4ea7-a41b-209d697bcdaa 2016-05-09 19:31:29 +00:00
5bc5c6ba93 git-svn-id: svn://svn.icms.temple.edu/lammps-ro/trunk@14956 f3b2605a-c512-4ea7-a41b-209d697bcdaa 2016-05-09 17:52:34 +00:00
e4081d2f50 git-svn-id: svn://svn.icms.temple.edu/lammps-ro/trunk@14955 f3b2605a-c512-4ea7-a41b-209d697bcdaa 2016-05-09 17:52:26 +00:00
5c3bc9e426 git-svn-id: svn://svn.icms.temple.edu/lammps-ro/trunk@14954 f3b2605a-c512-4ea7-a41b-209d697bcdaa 2016-05-09 17:52:15 +00:00
482d3890e1 git-svn-id: svn://svn.icms.temple.edu/lammps-ro/trunk@14953 f3b2605a-c512-4ea7-a41b-209d697bcdaa 2016-05-09 17:35:59 +00:00
61af3de4b8 git-svn-id: svn://svn.icms.temple.edu/lammps-ro/trunk@14952 f3b2605a-c512-4ea7-a41b-209d697bcdaa 2016-05-09 17:35:17 +00:00
ebf04bdf16 git-svn-id: svn://svn.icms.temple.edu/lammps-ro/trunk@14951 f3b2605a-c512-4ea7-a41b-209d697bcdaa 2016-05-09 17:33:12 +00:00
1cf54d01f4 git-svn-id: svn://svn.icms.temple.edu/lammps-ro/trunk@14950 f3b2605a-c512-4ea7-a41b-209d697bcdaa 2016-05-09 17:22:38 +00:00
d3e96156a7 git-svn-id: svn://svn.icms.temple.edu/lammps-ro/trunk@14949 f3b2605a-c512-4ea7-a41b-209d697bcdaa 2016-05-09 17:03:51 +00:00
055b20301e git-svn-id: svn://svn.icms.temple.edu/lammps-ro/trunk@14948 f3b2605a-c512-4ea7-a41b-209d697bcdaa 2016-05-09 16:56:43 +00:00
bc3c6d7599 git-svn-id: svn://svn.icms.temple.edu/lammps-ro/trunk@14947 f3b2605a-c512-4ea7-a41b-209d697bcdaa 2016-05-09 16:56:09 +00:00
a8ff213e1f git-svn-id: svn://svn.icms.temple.edu/lammps-ro/trunk@14946 f3b2605a-c512-4ea7-a41b-209d697bcdaa 2016-05-09 16:53:54 +00:00
35badaa4c7 git-svn-id: svn://svn.icms.temple.edu/lammps-ro/trunk@14945 f3b2605a-c512-4ea7-a41b-209d697bcdaa 2016-05-09 16:53:20 +00:00
b58cc3e832 git-svn-id: svn://svn.icms.temple.edu/lammps-ro/trunk@14944 f3b2605a-c512-4ea7-a41b-209d697bcdaa 2016-05-09 16:51:41 +00:00
f298c34a4b git-svn-id: svn://svn.icms.temple.edu/lammps-ro/trunk@14943 f3b2605a-c512-4ea7-a41b-209d697bcdaa 2016-05-09 16:50:38 +00:00
9e23f0540e Fixing Kokkos bug
git-svn-id: svn://svn.icms.temple.edu/lammps-ro/trunk@14940 f3b2605a-c512-4ea7-a41b-209d697bcdaa
2016-05-06 19:05:32 +00:00
8ee124a4d8 Fixing Kokkos bug
git-svn-id: svn://svn.icms.temple.edu/lammps-ro/trunk@14939 f3b2605a-c512-4ea7-a41b-209d697bcdaa
2016-05-06 19:00:45 +00:00
bf3187d764 git-svn-id: svn://svn.icms.temple.edu/lammps-ro/trunk@14938 f3b2605a-c512-4ea7-a41b-209d697bcdaa 2016-05-06 14:20:38 +00:00
7829620661 git-svn-id: svn://svn.icms.temple.edu/lammps-ro/trunk@14937 f3b2605a-c512-4ea7-a41b-209d697bcdaa 2016-05-04 14:05:06 +00:00
48d65b7fef Enabling new kokkos view
git-svn-id: svn://svn.icms.temple.edu/lammps-ro/trunk@14935 f3b2605a-c512-4ea7-a41b-209d697bcdaa
2016-05-03 16:24:08 +00:00
c696881836 Enabling new kokkos view
git-svn-id: svn://svn.icms.temple.edu/lammps-ro/trunk@14934 f3b2605a-c512-4ea7-a41b-209d697bcdaa
2016-05-03 16:23:57 +00:00
955c8d92ba git-svn-id: svn://svn.icms.temple.edu/lammps-ro/trunk@14933 f3b2605a-c512-4ea7-a41b-209d697bcdaa 2016-05-03 16:23:31 +00:00
3b03474d79 git-svn-id: svn://svn.icms.temple.edu/lammps-ro/trunk@14932 f3b2605a-c512-4ea7-a41b-209d697bcdaa 2016-05-03 16:22:36 +00:00
58dd900639 git-svn-id: svn://svn.icms.temple.edu/lammps-ro/trunk@14931 f3b2605a-c512-4ea7-a41b-209d697bcdaa 2016-05-03 16:21:05 +00:00
8da30644a7 git-svn-id: svn://svn.icms.temple.edu/lammps-ro/trunk@14930 f3b2605a-c512-4ea7-a41b-209d697bcdaa 2016-05-03 16:20:44 +00:00
7142393b07 git-svn-id: svn://svn.icms.temple.edu/lammps-ro/trunk@14929 f3b2605a-c512-4ea7-a41b-209d697bcdaa 2016-05-03 16:18:30 +00:00
91447edb58 git-svn-id: svn://svn.icms.temple.edu/lammps-ro/trunk@14928 f3b2605a-c512-4ea7-a41b-209d697bcdaa 2016-05-03 16:11:34 +00:00
98877f7e3c git-svn-id: svn://svn.icms.temple.edu/lammps-ro/trunk@14927 f3b2605a-c512-4ea7-a41b-209d697bcdaa 2016-05-03 16:11:25 +00:00
bccca1c712 git-svn-id: svn://svn.icms.temple.edu/lammps-ro/trunk@14926 f3b2605a-c512-4ea7-a41b-209d697bcdaa 2016-05-03 16:01:16 +00:00
ee668fce7f Reverting accidental changes
git-svn-id: svn://svn.icms.temple.edu/lammps-ro/trunk@14924 f3b2605a-c512-4ea7-a41b-209d697bcdaa
2016-05-03 15:39:35 +00:00
1af05e82c2 Collapsing kokkos half and half/thread neighbor lists
git-svn-id: svn://svn.icms.temple.edu/lammps-ro/trunk@14923 f3b2605a-c512-4ea7-a41b-209d697bcdaa
2016-05-03 14:36:31 +00:00
ed75c87cd6 Collapsing kokkos half and half/thread neighbor lists
git-svn-id: svn://svn.icms.temple.edu/lammps-ro/trunk@14922 f3b2605a-c512-4ea7-a41b-209d697bcdaa
2016-05-03 14:36:02 +00:00
33ce33ce36 Reverting back to old kokkos view
git-svn-id: svn://svn.icms.temple.edu/lammps-ro/trunk@14921 f3b2605a-c512-4ea7-a41b-209d697bcdaa
2016-05-02 23:40:38 +00:00
c5c293d16f Adding signal handler to Kokkos package
git-svn-id: svn://svn.icms.temple.edu/lammps-ro/trunk@14920 f3b2605a-c512-4ea7-a41b-209d697bcdaa
2016-05-02 23:01:37 +00:00
fda492ea48 Updating kokkos lib
git-svn-id: svn://svn.icms.temple.edu/lammps-ro/trunk@14919 f3b2605a-c512-4ea7-a41b-209d697bcdaa
2016-05-02 22:10:37 +00:00
0a1b765248 Updating kokkos lib
git-svn-id: svn://svn.icms.temple.edu/lammps-ro/trunk@14918 f3b2605a-c512-4ea7-a41b-209d697bcdaa
2016-05-02 22:06:50 +00:00
c5d0c55bee git-svn-id: svn://svn.icms.temple.edu/lammps-ro/trunk@14916 f3b2605a-c512-4ea7-a41b-209d697bcdaa 2016-04-30 18:14:25 +00:00
e41fe1cb39 git-svn-id: svn://svn.icms.temple.edu/lammps-ro/trunk@14915 f3b2605a-c512-4ea7-a41b-209d697bcdaa 2016-04-30 18:13:29 +00:00
b29c07bd76 git-svn-id: svn://svn.icms.temple.edu/lammps-ro/trunk@14914 f3b2605a-c512-4ea7-a41b-209d697bcdaa 2016-04-30 18:10:54 +00:00
114742bc0d git-svn-id: svn://svn.icms.temple.edu/lammps-ro/trunk@14913 f3b2605a-c512-4ea7-a41b-209d697bcdaa 2016-04-30 18:08:26 +00:00
6ff2012d88 git-svn-id: svn://svn.icms.temple.edu/lammps-ro/trunk@14912 f3b2605a-c512-4ea7-a41b-209d697bcdaa 2016-04-30 18:07:40 +00:00
997099253b git-svn-id: svn://svn.icms.temple.edu/lammps-ro/trunk@14911 f3b2605a-c512-4ea7-a41b-209d697bcdaa 2016-04-30 18:07:04 +00:00
00f38fdaf0 git-svn-id: svn://svn.icms.temple.edu/lammps-ro/trunk@14910 f3b2605a-c512-4ea7-a41b-209d697bcdaa 2016-04-30 18:06:25 +00:00
10ad7e3475 git-svn-id: svn://svn.icms.temple.edu/lammps-ro/trunk@14909 f3b2605a-c512-4ea7-a41b-209d697bcdaa 2016-04-30 18:03:40 +00:00
a49a2ae11d git-svn-id: svn://svn.icms.temple.edu/lammps-ro/trunk@14906 f3b2605a-c512-4ea7-a41b-209d697bcdaa 2016-04-28 15:03:48 +00:00
51f82750bf git-svn-id: svn://svn.icms.temple.edu/lammps-ro/trunk@14905 f3b2605a-c512-4ea7-a41b-209d697bcdaa 2016-04-28 15:02:54 +00:00
f76c6a3f08 git-svn-id: svn://svn.icms.temple.edu/lammps-ro/trunk@14904 f3b2605a-c512-4ea7-a41b-209d697bcdaa 2016-04-28 14:52:08 +00:00
c3007396e2 git-svn-id: svn://svn.icms.temple.edu/lammps-ro/trunk@14903 f3b2605a-c512-4ea7-a41b-209d697bcdaa 2016-04-28 14:51:25 +00:00
19122f203e git-svn-id: svn://svn.icms.temple.edu/lammps-ro/trunk@14902 f3b2605a-c512-4ea7-a41b-209d697bcdaa 2016-04-28 14:49:49 +00:00
6bc6d45116 git-svn-id: svn://svn.icms.temple.edu/lammps-ro/trunk@14901 f3b2605a-c512-4ea7-a41b-209d697bcdaa 2016-04-28 14:49:38 +00:00
17fd5898df git-svn-id: svn://svn.icms.temple.edu/lammps-ro/trunk@14900 f3b2605a-c512-4ea7-a41b-209d697bcdaa 2016-04-28 14:48:56 +00:00
107e28c77a git-svn-id: svn://svn.icms.temple.edu/lammps-ro/trunk@14899 f3b2605a-c512-4ea7-a41b-209d697bcdaa 2016-04-28 14:48:46 +00:00
0d4d950ce0 git-svn-id: svn://svn.icms.temple.edu/lammps-ro/trunk@14898 f3b2605a-c512-4ea7-a41b-209d697bcdaa 2016-04-28 14:47:48 +00:00
71926fe3d5 git-svn-id: svn://svn.icms.temple.edu/lammps-ro/trunk@14897 f3b2605a-c512-4ea7-a41b-209d697bcdaa 2016-04-28 14:46:41 +00:00
78d7b4af20 git-svn-id: svn://svn.icms.temple.edu/lammps-ro/trunk@14896 f3b2605a-c512-4ea7-a41b-209d697bcdaa 2016-04-27 14:56:51 +00:00
5f86c15dd2 git-svn-id: svn://svn.icms.temple.edu/lammps-ro/trunk@14895 f3b2605a-c512-4ea7-a41b-209d697bcdaa 2016-04-27 14:56:23 +00:00
05d35188c7 bugfix from Axel
git-svn-id: svn://svn.icms.temple.edu/lammps-ro/trunk@14894 f3b2605a-c512-4ea7-a41b-209d697bcdaa
2016-04-27 14:03:59 +00:00
579b61641e Fixing restart error
git-svn-id: svn://svn.icms.temple.edu/lammps-ro/trunk@14893 f3b2605a-c512-4ea7-a41b-209d697bcdaa
2016-04-26 23:00:32 +00:00
6b582d19a2 git-svn-id: svn://svn.icms.temple.edu/lammps-ro/trunk@14891 f3b2605a-c512-4ea7-a41b-209d697bcdaa 2016-04-26 22:38:01 +00:00
cab3fef913 git-svn-id: svn://svn.icms.temple.edu/lammps-ro/trunk@14890 f3b2605a-c512-4ea7-a41b-209d697bcdaa 2016-04-26 22:37:08 +00:00
0da53bb3e3 Fixing compile error
git-svn-id: svn://svn.icms.temple.edu/lammps-ro/trunk@14889 f3b2605a-c512-4ea7-a41b-209d697bcdaa
2016-04-26 22:32:03 +00:00
92b45d0163 git-svn-id: svn://svn.icms.temple.edu/lammps-ro/trunk@14888 f3b2605a-c512-4ea7-a41b-209d697bcdaa 2016-04-26 22:10:30 +00:00
d91aced28b git-svn-id: svn://svn.icms.temple.edu/lammps-ro/trunk@14886 f3b2605a-c512-4ea7-a41b-209d697bcdaa 2016-04-26 20:58:58 +00:00
b3732e76f8 git-svn-id: svn://svn.icms.temple.edu/lammps-ro/trunk@14885 f3b2605a-c512-4ea7-a41b-209d697bcdaa 2016-04-26 20:58:03 +00:00
5d6b6a0a0a git-svn-id: svn://svn.icms.temple.edu/lammps-ro/trunk@14883 f3b2605a-c512-4ea7-a41b-209d697bcdaa 2016-04-26 20:55:09 +00:00
d004bfef60 git-svn-id: svn://svn.icms.temple.edu/lammps-ro/trunk@14882 f3b2605a-c512-4ea7-a41b-209d697bcdaa 2016-04-26 19:55:28 +00:00
e7363b4764 git-svn-id: svn://svn.icms.temple.edu/lammps-ro/trunk@14881 f3b2605a-c512-4ea7-a41b-209d697bcdaa 2016-04-26 19:38:50 +00:00
325295bfaa git-svn-id: svn://svn.icms.temple.edu/lammps-ro/trunk@14880 f3b2605a-c512-4ea7-a41b-209d697bcdaa 2016-04-26 18:41:45 +00:00
446d3b305a git-svn-id: svn://svn.icms.temple.edu/lammps-ro/trunk@14879 f3b2605a-c512-4ea7-a41b-209d697bcdaa 2016-04-26 18:34:36 +00:00
1ce42845a9 git-svn-id: svn://svn.icms.temple.edu/lammps-ro/trunk@14878 f3b2605a-c512-4ea7-a41b-209d697bcdaa 2016-04-26 18:34:13 +00:00
e405f015f9 git-svn-id: svn://svn.icms.temple.edu/lammps-ro/trunk@14877 f3b2605a-c512-4ea7-a41b-209d697bcdaa 2016-04-26 18:33:10 +00:00
4314ef01f6 git-svn-id: svn://svn.icms.temple.edu/lammps-ro/trunk@14876 f3b2605a-c512-4ea7-a41b-209d697bcdaa 2016-04-26 18:32:11 +00:00
98ed24437d git-svn-id: svn://svn.icms.temple.edu/lammps-ro/trunk@14875 f3b2605a-c512-4ea7-a41b-209d697bcdaa 2016-04-26 18:28:26 +00:00
a7ce7207f9 git-svn-id: svn://svn.icms.temple.edu/lammps-ro/trunk@14874 f3b2605a-c512-4ea7-a41b-209d697bcdaa 2016-04-26 18:27:44 +00:00
e6ba3ccd33 Fixing Kokkos compile errors
git-svn-id: svn://svn.icms.temple.edu/lammps-ro/trunk@14873 f3b2605a-c512-4ea7-a41b-209d697bcdaa
2016-04-26 16:12:44 +00:00
656f8ed97a git-svn-id: svn://svn.icms.temple.edu/lammps-ro/trunk@14870 f3b2605a-c512-4ea7-a41b-209d697bcdaa 2016-04-22 17:55:51 +00:00
baf8ad52e5 git-svn-id: svn://svn.icms.temple.edu/lammps-ro/trunk@14869 f3b2605a-c512-4ea7-a41b-209d697bcdaa 2016-04-22 17:54:54 +00:00
08271a0200 git-svn-id: svn://svn.icms.temple.edu/lammps-ro/trunk@14868 f3b2605a-c512-4ea7-a41b-209d697bcdaa 2016-04-22 17:54:49 +00:00
184d5dc0f0 git-svn-id: svn://svn.icms.temple.edu/lammps-ro/trunk@14867 f3b2605a-c512-4ea7-a41b-209d697bcdaa 2016-04-22 17:51:49 +00:00
129796adc2 git-svn-id: svn://svn.icms.temple.edu/lammps-ro/trunk@14866 f3b2605a-c512-4ea7-a41b-209d697bcdaa 2016-04-22 17:42:06 +00:00
8d5ca6c7cc git-svn-id: svn://svn.icms.temple.edu/lammps-ro/trunk@14865 f3b2605a-c512-4ea7-a41b-209d697bcdaa 2016-04-22 17:41:11 +00:00
af63793e71 git-svn-id: svn://svn.icms.temple.edu/lammps-ro/trunk@14864 f3b2605a-c512-4ea7-a41b-209d697bcdaa 2016-04-22 17:40:22 +00:00
cc0049dce5 git-svn-id: svn://svn.icms.temple.edu/lammps-ro/trunk@14863 f3b2605a-c512-4ea7-a41b-209d697bcdaa 2016-04-22 16:00:48 +00:00
912ee9621a git-svn-id: svn://svn.icms.temple.edu/lammps-ro/trunk@14862 f3b2605a-c512-4ea7-a41b-209d697bcdaa 2016-04-22 15:56:56 +00:00
f2a4058415 git-svn-id: svn://svn.icms.temple.edu/lammps-ro/trunk@14861 f3b2605a-c512-4ea7-a41b-209d697bcdaa 2016-04-22 15:56:10 +00:00
f70a764d5c git-svn-id: svn://svn.icms.temple.edu/lammps-ro/trunk@14860 f3b2605a-c512-4ea7-a41b-209d697bcdaa 2016-04-22 15:55:54 +00:00
a4af037d1d git-svn-id: svn://svn.icms.temple.edu/lammps-ro/trunk@14859 f3b2605a-c512-4ea7-a41b-209d697bcdaa 2016-04-22 15:45:50 +00:00
6b0a181fac git-svn-id: svn://svn.icms.temple.edu/lammps-ro/trunk@14858 f3b2605a-c512-4ea7-a41b-209d697bcdaa 2016-04-22 15:44:13 +00:00
2ce937a070 git-svn-id: svn://svn.icms.temple.edu/lammps-ro/trunk@14857 f3b2605a-c512-4ea7-a41b-209d697bcdaa 2016-04-20 23:39:50 +00:00
95c1623665 git-svn-id: svn://svn.icms.temple.edu/lammps-ro/trunk@14856 f3b2605a-c512-4ea7-a41b-209d697bcdaa 2016-04-20 23:02:44 +00:00
93bc5c8be6 git-svn-id: svn://svn.icms.temple.edu/lammps-ro/trunk@14855 f3b2605a-c512-4ea7-a41b-209d697bcdaa 2016-04-20 22:46:43 +00:00
99209263b5 git-svn-id: svn://svn.icms.temple.edu/lammps-ro/trunk@14854 f3b2605a-c512-4ea7-a41b-209d697bcdaa 2016-04-20 22:38:35 +00:00
83cd2e5932 git-svn-id: svn://svn.icms.temple.edu/lammps-ro/trunk@14853 f3b2605a-c512-4ea7-a41b-209d697bcdaa 2016-04-20 22:34:03 +00:00
778165ed4b git-svn-id: svn://svn.icms.temple.edu/lammps-ro/trunk@14852 f3b2605a-c512-4ea7-a41b-209d697bcdaa 2016-04-20 22:31:52 +00:00
6bab5880f5 git-svn-id: svn://svn.icms.temple.edu/lammps-ro/trunk@14851 f3b2605a-c512-4ea7-a41b-209d697bcdaa 2016-04-20 22:13:56 +00:00
cc11bb9de8 git-svn-id: svn://svn.icms.temple.edu/lammps-ro/trunk@14850 f3b2605a-c512-4ea7-a41b-209d697bcdaa 2016-04-20 21:02:55 +00:00
268d59866f git-svn-id: svn://svn.icms.temple.edu/lammps-ro/trunk@14849 f3b2605a-c512-4ea7-a41b-209d697bcdaa 2016-04-20 20:48:13 +00:00
b1ab6b98e6 git-svn-id: svn://svn.icms.temple.edu/lammps-ro/trunk@14848 f3b2605a-c512-4ea7-a41b-209d697bcdaa 2016-04-20 19:37:47 +00:00
d5d0a879e8 Added Einstein version of Green-Kubo
git-svn-id: svn://svn.icms.temple.edu/lammps-ro/trunk@14847 f3b2605a-c512-4ea7-a41b-209d697bcdaa
2016-04-18 23:59:56 +00:00
00f3ccf3b0 Added Einstein version of Green-Kubo
git-svn-id: svn://svn.icms.temple.edu/lammps-ro/trunk@14846 f3b2605a-c512-4ea7-a41b-209d697bcdaa
2016-04-18 23:59:16 +00:00
545a273abf git-svn-id: svn://svn.icms.temple.edu/lammps-ro/trunk@14844 f3b2605a-c512-4ea7-a41b-209d697bcdaa 2016-04-18 14:07:11 +00:00
17fd6b2345 git-svn-id: svn://svn.icms.temple.edu/lammps-ro/trunk@14843 f3b2605a-c512-4ea7-a41b-209d697bcdaa 2016-04-18 14:06:15 +00:00
c6f7fb5968 git-svn-id: svn://svn.icms.temple.edu/lammps-ro/trunk@14842 f3b2605a-c512-4ea7-a41b-209d697bcdaa 2016-04-18 14:06:11 +00:00
44ca843073 git-svn-id: svn://svn.icms.temple.edu/lammps-ro/trunk@14841 f3b2605a-c512-4ea7-a41b-209d697bcdaa 2016-04-18 14:00:58 +00:00
b595eb9d0e git-svn-id: svn://svn.icms.temple.edu/lammps-ro/trunk@14840 f3b2605a-c512-4ea7-a41b-209d697bcdaa 2016-04-18 14:00:30 +00:00
30db73adb3 git-svn-id: svn://svn.icms.temple.edu/lammps-ro/trunk@14838 f3b2605a-c512-4ea7-a41b-209d697bcdaa 2016-04-15 16:33:58 +00:00
7a0c8f23d9 git-svn-id: svn://svn.icms.temple.edu/lammps-ro/trunk@14837 f3b2605a-c512-4ea7-a41b-209d697bcdaa 2016-04-15 16:33:03 +00:00
f9468317dd git-svn-id: svn://svn.icms.temple.edu/lammps-ro/trunk@14836 f3b2605a-c512-4ea7-a41b-209d697bcdaa 2016-04-15 16:28:57 +00:00
e18f5a903b git-svn-id: svn://svn.icms.temple.edu/lammps-ro/trunk@14835 f3b2605a-c512-4ea7-a41b-209d697bcdaa 2016-04-15 16:25:31 +00:00
552da3c342 git-svn-id: svn://svn.icms.temple.edu/lammps-ro/trunk@14834 f3b2605a-c512-4ea7-a41b-209d697bcdaa 2016-04-15 16:21:26 +00:00
a4e8eaaf4d git-svn-id: svn://svn.icms.temple.edu/lammps-ro/trunk@14833 f3b2605a-c512-4ea7-a41b-209d697bcdaa 2016-04-15 16:16:52 +00:00
e6ca2d5e08 git-svn-id: svn://svn.icms.temple.edu/lammps-ro/trunk@14832 f3b2605a-c512-4ea7-a41b-209d697bcdaa 2016-04-15 16:09:16 +00:00
eb8fdc4ab6 git-svn-id: svn://svn.icms.temple.edu/lammps-ro/trunk@14831 f3b2605a-c512-4ea7-a41b-209d697bcdaa 2016-04-15 16:08:22 +00:00
f50b03fcab git-svn-id: svn://svn.icms.temple.edu/lammps-ro/trunk@14830 f3b2605a-c512-4ea7-a41b-209d697bcdaa 2016-04-15 16:07:28 +00:00
212a955285 git-svn-id: svn://svn.icms.temple.edu/lammps-ro/trunk@14829 f3b2605a-c512-4ea7-a41b-209d697bcdaa 2016-04-15 16:07:01 +00:00
32509da721 git-svn-id: svn://svn.icms.temple.edu/lammps-ro/trunk@14828 f3b2605a-c512-4ea7-a41b-209d697bcdaa 2016-04-15 16:06:25 +00:00
3e4ce842ff Adding set method in pair for Kokkos
git-svn-id: svn://svn.icms.temple.edu/lammps-ro/trunk@14825 f3b2605a-c512-4ea7-a41b-209d697bcdaa
2016-04-14 17:51:39 +00:00
ef7e119c86 git-svn-id: svn://svn.icms.temple.edu/lammps-ro/trunk@14824 f3b2605a-c512-4ea7-a41b-209d697bcdaa 2016-04-14 14:42:56 +00:00
c46d973c2b git-svn-id: svn://svn.icms.temple.edu/lammps-ro/trunk@14823 f3b2605a-c512-4ea7-a41b-209d697bcdaa 2016-04-13 22:25:46 +00:00
92fa12cc32 git-svn-id: svn://svn.icms.temple.edu/lammps-ro/trunk@14822 f3b2605a-c512-4ea7-a41b-209d697bcdaa 2016-04-13 22:20:59 +00:00
e564fc4de3 Fixing Kokkos default values
git-svn-id: svn://svn.icms.temple.edu/lammps-ro/trunk@14821 f3b2605a-c512-4ea7-a41b-209d697bcdaa
2016-04-13 21:38:58 +00:00
cf26690feb Corrected formula for case lx != ly
git-svn-id: svn://svn.icms.temple.edu/lammps-ro/trunk@14820 f3b2605a-c512-4ea7-a41b-209d697bcdaa
2016-04-12 14:56:22 +00:00
20f13bf13d Added a script to calculate elastic compliance tensor
git-svn-id: svn://svn.icms.temple.edu/lammps-ro/trunk@14819 f3b2605a-c512-4ea7-a41b-209d697bcdaa
2016-04-12 01:28:58 +00:00
e9085c47c1 git-svn-id: svn://svn.icms.temple.edu/lammps-ro/trunk@14817 f3b2605a-c512-4ea7-a41b-209d697bcdaa 2016-04-07 22:26:34 +00:00
dd562698ef git-svn-id: svn://svn.icms.temple.edu/lammps-ro/trunk@14816 f3b2605a-c512-4ea7-a41b-209d697bcdaa 2016-04-07 22:25:40 +00:00
9aa7608bd0 git-svn-id: svn://svn.icms.temple.edu/lammps-ro/trunk@14814 f3b2605a-c512-4ea7-a41b-209d697bcdaa 2016-04-07 22:22:25 +00:00
9c1857fa62 git-svn-id: svn://svn.icms.temple.edu/lammps-ro/trunk@14813 f3b2605a-c512-4ea7-a41b-209d697bcdaa 2016-04-07 22:22:19 +00:00
6a1e85d32a git-svn-id: svn://svn.icms.temple.edu/lammps-ro/trunk@14812 f3b2605a-c512-4ea7-a41b-209d697bcdaa 2016-04-07 21:16:06 +00:00
5dceda8b15 git-svn-id: svn://svn.icms.temple.edu/lammps-ro/trunk@14811 f3b2605a-c512-4ea7-a41b-209d697bcdaa 2016-04-07 21:14:57 +00:00
d8a711e50e git-svn-id: svn://svn.icms.temple.edu/lammps-ro/trunk@14810 f3b2605a-c512-4ea7-a41b-209d697bcdaa 2016-04-07 21:13:52 +00:00
13c5549009 git-svn-id: svn://svn.icms.temple.edu/lammps-ro/trunk@14809 f3b2605a-c512-4ea7-a41b-209d697bcdaa 2016-04-07 21:12:44 +00:00
008896a77d git-svn-id: svn://svn.icms.temple.edu/lammps-ro/trunk@14808 f3b2605a-c512-4ea7-a41b-209d697bcdaa 2016-04-07 21:10:37 +00:00
67a4004f23 git-svn-id: svn://svn.icms.temple.edu/lammps-ro/trunk@14807 f3b2605a-c512-4ea7-a41b-209d697bcdaa 2016-04-07 21:05:19 +00:00
1e180da830 git-svn-id: svn://svn.icms.temple.edu/lammps-ro/trunk@14806 f3b2605a-c512-4ea7-a41b-209d697bcdaa 2016-04-07 21:05:09 +00:00
8e36364f5c git-svn-id: svn://svn.icms.temple.edu/lammps-ro/trunk@14805 f3b2605a-c512-4ea7-a41b-209d697bcdaa 2016-04-07 21:04:44 +00:00
65f40aa34d Kokkos enhancements
git-svn-id: svn://svn.icms.temple.edu/lammps-ro/trunk@14804 f3b2605a-c512-4ea7-a41b-209d697bcdaa
2016-04-06 23:08:54 +00:00
c855d51211 Corrected wurtzite example
git-svn-id: svn://svn.icms.temple.edu/lammps-ro/trunk@14803 f3b2605a-c512-4ea7-a41b-209d697bcdaa
2016-04-05 15:11:14 +00:00
83796feb88 Fixing bug in fix efield
git-svn-id: svn://svn.icms.temple.edu/lammps-ro/trunk@14802 f3b2605a-c512-4ea7-a41b-209d697bcdaa
2016-04-04 22:24:03 +00:00
5ebf01ec79 fixed bug in triclinic case
git-svn-id: svn://svn.icms.temple.edu/lammps-ro/trunk@14801 f3b2605a-c512-4ea7-a41b-209d697bcdaa
2016-04-04 21:07:40 +00:00
25992ae9f7 fixed bug in triclinic case
git-svn-id: svn://svn.icms.temple.edu/lammps-ro/trunk@14800 f3b2605a-c512-4ea7-a41b-209d697bcdaa
2016-04-04 21:01:02 +00:00
aaa32cfd8b git-svn-id: svn://svn.icms.temple.edu/lammps-ro/trunk@14799 f3b2605a-c512-4ea7-a41b-209d697bcdaa 2016-03-29 16:16:50 +00:00
cedacbc26e git-svn-id: svn://svn.icms.temple.edu/lammps-ro/trunk@14795 f3b2605a-c512-4ea7-a41b-209d697bcdaa 2016-03-28 15:23:57 +00:00
0c3875c602 git-svn-id: svn://svn.icms.temple.edu/lammps-ro/trunk@14794 f3b2605a-c512-4ea7-a41b-209d697bcdaa 2016-03-28 15:23:01 +00:00
c1afa2c1c3 git-svn-id: svn://svn.icms.temple.edu/lammps-ro/trunk@14793 f3b2605a-c512-4ea7-a41b-209d697bcdaa 2016-03-28 15:09:50 +00:00
f28b6cf0e5 git-svn-id: svn://svn.icms.temple.edu/lammps-ro/trunk@14792 f3b2605a-c512-4ea7-a41b-209d697bcdaa 2016-03-28 15:09:29 +00:00
aa833f17a1 git-svn-id: svn://svn.icms.temple.edu/lammps-ro/trunk@14791 f3b2605a-c512-4ea7-a41b-209d697bcdaa 2016-03-28 15:07:06 +00:00
b5591e4518 git-svn-id: svn://svn.icms.temple.edu/lammps-ro/trunk@14790 f3b2605a-c512-4ea7-a41b-209d697bcdaa 2016-03-28 15:05:50 +00:00
7cea607190 git-svn-id: svn://svn.icms.temple.edu/lammps-ro/trunk@14789 f3b2605a-c512-4ea7-a41b-209d697bcdaa 2016-03-28 15:04:09 +00:00
1103448232 git-svn-id: svn://svn.icms.temple.edu/lammps-ro/trunk@14788 f3b2605a-c512-4ea7-a41b-209d697bcdaa 2016-03-28 14:23:28 +00:00
ff4498fcdb Added note about ev
git-svn-id: svn://svn.icms.temple.edu/lammps-ro/trunk@14785 f3b2605a-c512-4ea7-a41b-209d697bcdaa
2016-03-24 22:40:27 +00:00
154eb1f886 Updated Adri's contact info
git-svn-id: svn://svn.icms.temple.edu/lammps-ro/trunk@14784 f3b2605a-c512-4ea7-a41b-209d697bcdaa
2016-03-23 16:09:54 +00:00
0528cde331 git-svn-id: svn://svn.icms.temple.edu/lammps-ro/trunk@14783 f3b2605a-c512-4ea7-a41b-209d697bcdaa 2016-03-22 14:29:41 +00:00
bd51271d4a git-svn-id: svn://svn.icms.temple.edu/lammps-ro/trunk@14781 f3b2605a-c512-4ea7-a41b-209d697bcdaa 2016-03-22 13:48:20 +00:00
c57e61a51f git-svn-id: svn://svn.icms.temple.edu/lammps-ro/trunk@14780 f3b2605a-c512-4ea7-a41b-209d697bcdaa 2016-03-22 13:47:24 +00:00
c1c464d033 git-svn-id: svn://svn.icms.temple.edu/lammps-ro/trunk@14779 f3b2605a-c512-4ea7-a41b-209d697bcdaa 2016-03-22 13:45:31 +00:00
acc29ff266 git-svn-id: svn://svn.icms.temple.edu/lammps-ro/trunk@14778 f3b2605a-c512-4ea7-a41b-209d697bcdaa 2016-03-22 13:44:55 +00:00
4dec943a5e git-svn-id: svn://svn.icms.temple.edu/lammps-ro/trunk@14777 f3b2605a-c512-4ea7-a41b-209d697bcdaa 2016-03-22 13:44:36 +00:00
40a438575d git-svn-id: svn://svn.icms.temple.edu/lammps-ro/trunk@14776 f3b2605a-c512-4ea7-a41b-209d697bcdaa 2016-03-22 13:44:25 +00:00
361ed36bed git-svn-id: svn://svn.icms.temple.edu/lammps-ro/trunk@14775 f3b2605a-c512-4ea7-a41b-209d697bcdaa 2016-03-22 13:43:01 +00:00
b4abf2abd2 git-svn-id: svn://svn.icms.temple.edu/lammps-ro/trunk@14774 f3b2605a-c512-4ea7-a41b-209d697bcdaa 2016-03-22 13:42:54 +00:00
8f3d0f743e git-svn-id: svn://svn.icms.temple.edu/lammps-ro/trunk@14773 f3b2605a-c512-4ea7-a41b-209d697bcdaa 2016-03-22 13:42:13 +00:00
89e3f5b5a2 git-svn-id: svn://svn.icms.temple.edu/lammps-ro/trunk@14772 f3b2605a-c512-4ea7-a41b-209d697bcdaa 2016-03-22 13:28:10 +00:00
5f6a4fb5d8 git-svn-id: svn://svn.icms.temple.edu/lammps-ro/trunk@14771 f3b2605a-c512-4ea7-a41b-209d697bcdaa 2016-03-22 13:27:57 +00:00
13ae93dbdd git-svn-id: svn://svn.icms.temple.edu/lammps-ro/trunk@14769 f3b2605a-c512-4ea7-a41b-209d697bcdaa 2016-03-22 00:20:07 +00:00
eda01bfd1a git-svn-id: svn://svn.icms.temple.edu/lammps-ro/trunk@14768 f3b2605a-c512-4ea7-a41b-209d697bcdaa 2016-03-22 00:19:08 +00:00
86fc7d455c git-svn-id: svn://svn.icms.temple.edu/lammps-ro/trunk@14767 f3b2605a-c512-4ea7-a41b-209d697bcdaa 2016-03-22 00:17:14 +00:00
009f3564e0 git-svn-id: svn://svn.icms.temple.edu/lammps-ro/trunk@14766 f3b2605a-c512-4ea7-a41b-209d697bcdaa 2016-03-22 00:16:15 +00:00
b50f71cd96 git-svn-id: svn://svn.icms.temple.edu/lammps-ro/trunk@14765 f3b2605a-c512-4ea7-a41b-209d697bcdaa 2016-03-22 00:15:54 +00:00
a10c992e38 git-svn-id: svn://svn.icms.temple.edu/lammps-ro/trunk@14764 f3b2605a-c512-4ea7-a41b-209d697bcdaa 2016-03-22 00:05:55 +00:00
a568a8741b git-svn-id: svn://svn.icms.temple.edu/lammps-ro/trunk@14761 f3b2605a-c512-4ea7-a41b-209d697bcdaa 2016-03-21 16:15:41 +00:00
a3ad5b65d3 git-svn-id: svn://svn.icms.temple.edu/lammps-ro/trunk@14760 f3b2605a-c512-4ea7-a41b-209d697bcdaa 2016-03-21 16:14:45 +00:00
847090cd89 git-svn-id: svn://svn.icms.temple.edu/lammps-ro/trunk@14759 f3b2605a-c512-4ea7-a41b-209d697bcdaa 2016-03-21 16:14:31 +00:00
edb3f9559c git-svn-id: svn://svn.icms.temple.edu/lammps-ro/trunk@14758 f3b2605a-c512-4ea7-a41b-209d697bcdaa 2016-03-21 16:10:48 +00:00
cea9e516a8 git-svn-id: svn://svn.icms.temple.edu/lammps-ro/trunk@14757 f3b2605a-c512-4ea7-a41b-209d697bcdaa 2016-03-21 16:06:09 +00:00
36776f425b git-svn-id: svn://svn.icms.temple.edu/lammps-ro/trunk@14756 f3b2605a-c512-4ea7-a41b-209d697bcdaa 2016-03-21 16:01:21 +00:00
784d8cc2f8 git-svn-id: svn://svn.icms.temple.edu/lammps-ro/trunk@14755 f3b2605a-c512-4ea7-a41b-209d697bcdaa 2016-03-21 15:55:33 +00:00
6e3ea06b11 Allowing read_restart to use suffix style with pair style, etc
git-svn-id: svn://svn.icms.temple.edu/lammps-ro/trunk@14754 f3b2605a-c512-4ea7-a41b-209d697bcdaa
2016-03-21 14:51:07 +00:00
1901b7b33e Allowing read_restart to use suffix style with avec
git-svn-id: svn://svn.icms.temple.edu/lammps-ro/trunk@14753 f3b2605a-c512-4ea7-a41b-209d697bcdaa
2016-03-21 14:44:19 +00:00
042f50ac9c Christian's kokkos patch that allows concurrent host and device threading
git-svn-id: svn://svn.icms.temple.edu/lammps-ro/trunk@14752 f3b2605a-c512-4ea7-a41b-209d697bcdaa
2016-03-18 14:50:28 +00:00
b760ca8038 git-svn-id: svn://svn.icms.temple.edu/lammps-ro/trunk@14750 f3b2605a-c512-4ea7-a41b-209d697bcdaa 2016-03-14 19:11:28 +00:00
a59445d4b7 git-svn-id: svn://svn.icms.temple.edu/lammps-ro/trunk@14749 f3b2605a-c512-4ea7-a41b-209d697bcdaa 2016-03-14 19:10:32 +00:00
93e451b594 git-svn-id: svn://svn.icms.temple.edu/lammps-ro/trunk@14748 f3b2605a-c512-4ea7-a41b-209d697bcdaa 2016-03-14 19:06:38 +00:00
ec42f2530c git-svn-id: svn://svn.icms.temple.edu/lammps-ro/trunk@14747 f3b2605a-c512-4ea7-a41b-209d697bcdaa 2016-03-14 18:46:42 +00:00
ad3c57cb63 git-svn-id: svn://svn.icms.temple.edu/lammps-ro/trunk@14746 f3b2605a-c512-4ea7-a41b-209d697bcdaa 2016-03-14 18:45:25 +00:00
61c0608ffa git-svn-id: svn://svn.icms.temple.edu/lammps-ro/trunk@14745 f3b2605a-c512-4ea7-a41b-209d697bcdaa 2016-03-14 18:45:15 +00:00
16b9206d98 git-svn-id: svn://svn.icms.temple.edu/lammps-ro/trunk@14744 f3b2605a-c512-4ea7-a41b-209d697bcdaa 2016-03-14 18:44:45 +00:00
92ef1cb060 git-svn-id: svn://svn.icms.temple.edu/lammps-ro/trunk@14743 f3b2605a-c512-4ea7-a41b-209d697bcdaa 2016-03-14 18:44:13 +00:00
2e3032723c git-svn-id: svn://svn.icms.temple.edu/lammps-ro/trunk@14742 f3b2605a-c512-4ea7-a41b-209d697bcdaa 2016-03-14 18:41:57 +00:00
66b282b36d git-svn-id: svn://svn.icms.temple.edu/lammps-ro/trunk@14741 f3b2605a-c512-4ea7-a41b-209d697bcdaa 2016-03-14 18:41:39 +00:00
e02c66b932 git-svn-id: svn://svn.icms.temple.edu/lammps-ro/trunk@14740 f3b2605a-c512-4ea7-a41b-209d697bcdaa 2016-03-14 18:40:28 +00:00
a946de0b03 git-svn-id: svn://svn.icms.temple.edu/lammps-ro/trunk@14738 f3b2605a-c512-4ea7-a41b-209d697bcdaa 2016-03-10 16:57:36 +00:00
7c9022e3d5 git-svn-id: svn://svn.icms.temple.edu/lammps-ro/trunk@14737 f3b2605a-c512-4ea7-a41b-209d697bcdaa 2016-03-10 16:56:40 +00:00
f5ceb5b292 git-svn-id: svn://svn.icms.temple.edu/lammps-ro/trunk@14736 f3b2605a-c512-4ea7-a41b-209d697bcdaa 2016-03-10 16:34:30 +00:00
109b9bab28 git-svn-id: svn://svn.icms.temple.edu/lammps-ro/trunk@14735 f3b2605a-c512-4ea7-a41b-209d697bcdaa 2016-03-10 16:31:34 +00:00
f620bcc4f9 git-svn-id: svn://svn.icms.temple.edu/lammps-ro/trunk@14734 f3b2605a-c512-4ea7-a41b-209d697bcdaa 2016-03-10 16:30:39 +00:00
047c55383a git-svn-id: svn://svn.icms.temple.edu/lammps-ro/trunk@14732 f3b2605a-c512-4ea7-a41b-209d697bcdaa 2016-03-10 16:13:43 +00:00
b3890e80cb git-svn-id: svn://svn.icms.temple.edu/lammps-ro/trunk@14731 f3b2605a-c512-4ea7-a41b-209d697bcdaa 2016-03-10 16:10:53 +00:00
8412d4a96c git-svn-id: svn://svn.icms.temple.edu/lammps-ro/trunk@14730 f3b2605a-c512-4ea7-a41b-209d697bcdaa 2016-03-10 16:09:52 +00:00
52f20bbbd5 git-svn-id: svn://svn.icms.temple.edu/lammps-ro/trunk@14729 f3b2605a-c512-4ea7-a41b-209d697bcdaa 2016-03-10 16:09:11 +00:00
247bf33d63 git-svn-id: svn://svn.icms.temple.edu/lammps-ro/trunk@14728 f3b2605a-c512-4ea7-a41b-209d697bcdaa 2016-03-10 16:04:18 +00:00
791f18b03d Corrected error in Liouville factorization
git-svn-id: svn://svn.icms.temple.edu/lammps-ro/trunk@14727 f3b2605a-c512-4ea7-a41b-209d697bcdaa
2016-03-08 19:56:06 +00:00
ee1805538c git-svn-id: svn://svn.icms.temple.edu/lammps-ro/trunk@14726 f3b2605a-c512-4ea7-a41b-209d697bcdaa 2016-03-07 17:34:03 +00:00
f4ac444db2 git-svn-id: svn://svn.icms.temple.edu/lammps-ro/trunk@14725 f3b2605a-c512-4ea7-a41b-209d697bcdaa 2016-03-07 17:30:46 +00:00
0e5c36676f git-svn-id: svn://svn.icms.temple.edu/lammps-ro/trunk@14724 f3b2605a-c512-4ea7-a41b-209d697bcdaa 2016-03-07 17:29:34 +00:00
d6a67c2849 Fixing Kokkos bugs
git-svn-id: svn://svn.icms.temple.edu/lammps-ro/trunk@14722 f3b2605a-c512-4ea7-a41b-209d697bcdaa
2016-03-04 17:32:00 +00:00
9b33b04183 git-svn-id: svn://svn.icms.temple.edu/lammps-ro/trunk@14720 f3b2605a-c512-4ea7-a41b-209d697bcdaa 2016-03-02 00:28:45 +00:00
e97ec23ed3 git-svn-id: svn://svn.icms.temple.edu/lammps-ro/trunk@14718 f3b2605a-c512-4ea7-a41b-209d697bcdaa 2016-03-01 22:57:38 +00:00
19d769ac45 git-svn-id: svn://svn.icms.temple.edu/lammps-ro/trunk@14716 f3b2605a-c512-4ea7-a41b-209d697bcdaa 2016-03-01 22:56:43 +00:00
ef273d3095 git-svn-id: svn://svn.icms.temple.edu/lammps-ro/trunk@14715 f3b2605a-c512-4ea7-a41b-209d697bcdaa 2016-03-01 22:56:00 +00:00
9e22165357 git-svn-id: svn://svn.icms.temple.edu/lammps-ro/trunk@14714 f3b2605a-c512-4ea7-a41b-209d697bcdaa 2016-03-01 22:52:13 +00:00
223defd6ee git-svn-id: svn://svn.icms.temple.edu/lammps-ro/trunk@14713 f3b2605a-c512-4ea7-a41b-209d697bcdaa 2016-03-01 22:49:42 +00:00
dd73c4c4c3 git-svn-id: svn://svn.icms.temple.edu/lammps-ro/trunk@14712 f3b2605a-c512-4ea7-a41b-209d697bcdaa 2016-03-01 22:38:04 +00:00
c67331a3bc git-svn-id: svn://svn.icms.temple.edu/lammps-ro/trunk@14711 f3b2605a-c512-4ea7-a41b-209d697bcdaa 2016-03-01 22:30:28 +00:00
1145d82222 git-svn-id: svn://svn.icms.temple.edu/lammps-ro/trunk@14710 f3b2605a-c512-4ea7-a41b-209d697bcdaa 2016-03-01 20:47:30 +00:00
9d64c50cd0 git-svn-id: svn://svn.icms.temple.edu/lammps-ro/trunk@14709 f3b2605a-c512-4ea7-a41b-209d697bcdaa 2016-03-01 20:47:13 +00:00
9dec6db569 git-svn-id: svn://svn.icms.temple.edu/lammps-ro/trunk@14708 f3b2605a-c512-4ea7-a41b-209d697bcdaa 2016-03-01 20:45:46 +00:00
29215e8116 git-svn-id: svn://svn.icms.temple.edu/lammps-ro/trunk@14707 f3b2605a-c512-4ea7-a41b-209d697bcdaa 2016-03-01 20:45:31 +00:00
800c546e19 git-svn-id: svn://svn.icms.temple.edu/lammps-ro/trunk@14706 f3b2605a-c512-4ea7-a41b-209d697bcdaa 2016-03-01 20:45:12 +00:00
45d25c6154 git-svn-id: svn://svn.icms.temple.edu/lammps-ro/trunk@14705 f3b2605a-c512-4ea7-a41b-209d697bcdaa 2016-03-01 20:42:04 +00:00
a3b0234a6a git-svn-id: svn://svn.icms.temple.edu/lammps-ro/trunk@14704 f3b2605a-c512-4ea7-a41b-209d697bcdaa 2016-03-01 20:41:58 +00:00
20beaccf0f git-svn-id: svn://svn.icms.temple.edu/lammps-ro/trunk@14703 f3b2605a-c512-4ea7-a41b-209d697bcdaa 2016-03-01 20:40:25 +00:00
6e2893c768 git-svn-id: svn://svn.icms.temple.edu/lammps-ro/trunk@14702 f3b2605a-c512-4ea7-a41b-209d697bcdaa 2016-03-01 20:40:08 +00:00
756480ba64 git-svn-id: svn://svn.icms.temple.edu/lammps-ro/trunk@14701 f3b2605a-c512-4ea7-a41b-209d697bcdaa 2016-03-01 20:39:41 +00:00
d0646402bc git-svn-id: svn://svn.icms.temple.edu/lammps-ro/trunk@14700 f3b2605a-c512-4ea7-a41b-209d697bcdaa 2016-03-01 20:39:34 +00:00
ee19043605 git-svn-id: svn://svn.icms.temple.edu/lammps-ro/trunk@14699 f3b2605a-c512-4ea7-a41b-209d697bcdaa 2016-03-01 20:38:26 +00:00
0bf65758e8 git-svn-id: svn://svn.icms.temple.edu/lammps-ro/trunk@14698 f3b2605a-c512-4ea7-a41b-209d697bcdaa 2016-03-01 20:35:24 +00:00
7642bc1775 git-svn-id: svn://svn.icms.temple.edu/lammps-ro/trunk@14697 f3b2605a-c512-4ea7-a41b-209d697bcdaa 2016-03-01 18:41:30 +00:00
d1a65e5f6a git-svn-id: svn://svn.icms.temple.edu/lammps-ro/trunk@14696 f3b2605a-c512-4ea7-a41b-209d697bcdaa 2016-03-01 18:22:28 +00:00
23ab6d4c0c git-svn-id: svn://svn.icms.temple.edu/lammps-ro/trunk@14695 f3b2605a-c512-4ea7-a41b-209d697bcdaa 2016-03-01 16:55:50 +00:00
bb0e8a114f git-svn-id: svn://svn.icms.temple.edu/lammps-ro/trunk@14694 f3b2605a-c512-4ea7-a41b-209d697bcdaa 2016-03-01 00:39:43 +00:00
b2e792d6ac fixed problem with exclusion
git-svn-id: svn://svn.icms.temple.edu/lammps-ro/trunk@14693 f3b2605a-c512-4ea7-a41b-209d697bcdaa
2016-02-29 22:10:07 +00:00
855e8ec701 git-svn-id: svn://svn.icms.temple.edu/lammps-ro/trunk@14692 f3b2605a-c512-4ea7-a41b-209d697bcdaa 2016-02-29 15:36:36 +00:00
adbca81654 git-svn-id: svn://svn.icms.temple.edu/lammps-ro/trunk@14691 f3b2605a-c512-4ea7-a41b-209d697bcdaa 2016-02-27 00:43:50 +00:00
94d8894f86 Added exclusion delete function and bounds check for coord2bin
git-svn-id: svn://svn.icms.temple.edu/lammps-ro/trunk@14690 f3b2605a-c512-4ea7-a41b-209d697bcdaa
2016-02-26 23:39:53 +00:00
12d18e27cf git-svn-id: svn://svn.icms.temple.edu/lammps-ro/trunk@14689 f3b2605a-c512-4ea7-a41b-209d697bcdaa 2016-02-26 16:09:00 +00:00
fe00daa91b git-svn-id: svn://svn.icms.temple.edu/lammps-ro/trunk@14688 f3b2605a-c512-4ea7-a41b-209d697bcdaa 2016-02-26 16:04:09 +00:00
4f50ad2a6a git-svn-id: svn://svn.icms.temple.edu/lammps-ro/trunk@14687 f3b2605a-c512-4ea7-a41b-209d697bcdaa 2016-02-26 15:56:01 +00:00
ea62766575 git-svn-id: svn://svn.icms.temple.edu/lammps-ro/trunk@14686 f3b2605a-c512-4ea7-a41b-209d697bcdaa 2016-02-26 15:55:30 +00:00
b6e19b7803 git-svn-id: svn://svn.icms.temple.edu/lammps-ro/trunk@14685 f3b2605a-c512-4ea7-a41b-209d697bcdaa 2016-02-26 15:55:15 +00:00
19ca625699 git-svn-id: svn://svn.icms.temple.edu/lammps-ro/trunk@14684 f3b2605a-c512-4ea7-a41b-209d697bcdaa 2016-02-26 15:53:21 +00:00
2a157729a3 git-svn-id: svn://svn.icms.temple.edu/lammps-ro/trunk@14683 f3b2605a-c512-4ea7-a41b-209d697bcdaa 2016-02-26 15:49:30 +00:00
4351bad69f git-svn-id: svn://svn.icms.temple.edu/lammps-ro/trunk@14681 f3b2605a-c512-4ea7-a41b-209d697bcdaa 2016-02-25 23:58:54 +00:00
37810ba271 git-svn-id: svn://svn.icms.temple.edu/lammps-ro/trunk@14680 f3b2605a-c512-4ea7-a41b-209d697bcdaa 2016-02-25 23:57:59 +00:00
2ce3daca37 git-svn-id: svn://svn.icms.temple.edu/lammps-ro/trunk@14679 f3b2605a-c512-4ea7-a41b-209d697bcdaa 2016-02-25 23:30:00 +00:00
737c083a4f Enabled triclinic cells
git-svn-id: svn://svn.icms.temple.edu/lammps-ro/trunk@14678 f3b2605a-c512-4ea7-a41b-209d697bcdaa
2016-02-25 23:12:38 +00:00
b5e95a0a14 Fixing USER-CUDA compile error
git-svn-id: svn://svn.icms.temple.edu/lammps-ro/trunk@14677 f3b2605a-c512-4ea7-a41b-209d697bcdaa
2016-02-24 18:53:37 +00:00
dbd259948b git-svn-id: svn://svn.icms.temple.edu/lammps-ro/trunk@14675 f3b2605a-c512-4ea7-a41b-209d697bcdaa 2016-02-24 15:27:19 +00:00
98b34b6311 git-svn-id: svn://svn.icms.temple.edu/lammps-ro/trunk@14673 f3b2605a-c512-4ea7-a41b-209d697bcdaa 2016-02-23 23:25:07 +00:00
a4e33d3096 git-svn-id: svn://svn.icms.temple.edu/lammps-ro/trunk@14672 f3b2605a-c512-4ea7-a41b-209d697bcdaa 2016-02-23 23:22:40 +00:00
4ddd88da26 git-svn-id: svn://svn.icms.temple.edu/lammps-ro/trunk@14670 f3b2605a-c512-4ea7-a41b-209d697bcdaa 2016-02-22 23:30:51 +00:00
e0361bb7a5 git-svn-id: svn://svn.icms.temple.edu/lammps-ro/trunk@14669 f3b2605a-c512-4ea7-a41b-209d697bcdaa 2016-02-22 23:30:32 +00:00
43158e38f6 git-svn-id: svn://svn.icms.temple.edu/lammps-ro/trunk@14667 f3b2605a-c512-4ea7-a41b-209d697bcdaa 2016-02-22 23:27:56 +00:00
a7b85690d1 git-svn-id: svn://svn.icms.temple.edu/lammps-ro/trunk@14666 f3b2605a-c512-4ea7-a41b-209d697bcdaa 2016-02-22 23:27:00 +00:00
c200e187b5 git-svn-id: svn://svn.icms.temple.edu/lammps-ro/trunk@14665 f3b2605a-c512-4ea7-a41b-209d697bcdaa 2016-02-22 23:18:57 +00:00
577af878ea git-svn-id: svn://svn.icms.temple.edu/lammps-ro/trunk@14664 f3b2605a-c512-4ea7-a41b-209d697bcdaa 2016-02-22 23:14:55 +00:00
f406a5ed55 git-svn-id: svn://svn.icms.temple.edu/lammps-ro/trunk@14663 f3b2605a-c512-4ea7-a41b-209d697bcdaa 2016-02-22 23:00:59 +00:00
8291f5e69c git-svn-id: svn://svn.icms.temple.edu/lammps-ro/trunk@14662 f3b2605a-c512-4ea7-a41b-209d697bcdaa 2016-02-22 23:00:43 +00:00
dc0dc5a7d9 git-svn-id: svn://svn.icms.temple.edu/lammps-ro/trunk@14661 f3b2605a-c512-4ea7-a41b-209d697bcdaa 2016-02-22 22:13:26 +00:00
d2d5b0cb75 git-svn-id: svn://svn.icms.temple.edu/lammps-ro/trunk@14660 f3b2605a-c512-4ea7-a41b-209d697bcdaa 2016-02-22 21:57:32 +00:00
00fc71e8b8 git-svn-id: svn://svn.icms.temple.edu/lammps-ro/trunk@14659 f3b2605a-c512-4ea7-a41b-209d697bcdaa 2016-02-22 17:10:07 +00:00
f705ef5600 git-svn-id: svn://svn.icms.temple.edu/lammps-ro/trunk@14658 f3b2605a-c512-4ea7-a41b-209d697bcdaa 2016-02-18 22:32:24 +00:00
08b53bb709 git-svn-id: svn://svn.icms.temple.edu/lammps-ro/trunk@14657 f3b2605a-c512-4ea7-a41b-209d697bcdaa 2016-02-18 22:31:59 +00:00
5c78508b40 git-svn-id: svn://svn.icms.temple.edu/lammps-ro/trunk@14656 f3b2605a-c512-4ea7-a41b-209d697bcdaa 2016-02-18 22:31:49 +00:00
e4ea9c0658 git-svn-id: svn://svn.icms.temple.edu/lammps-ro/trunk@14655 f3b2605a-c512-4ea7-a41b-209d697bcdaa 2016-02-17 15:56:48 +00:00
b7e4a33bab git-svn-id: svn://svn.icms.temple.edu/lammps-ro/trunk@14654 f3b2605a-c512-4ea7-a41b-209d697bcdaa 2016-02-17 15:54:00 +00:00
718c87f915 git-svn-id: svn://svn.icms.temple.edu/lammps-ro/trunk@14653 f3b2605a-c512-4ea7-a41b-209d697bcdaa 2016-02-17 15:40:36 +00:00
a43731758c git-svn-id: svn://svn.icms.temple.edu/lammps-ro/trunk@14652 f3b2605a-c512-4ea7-a41b-209d697bcdaa 2016-02-17 00:19:10 +00:00
eee65e309e git-svn-id: svn://svn.icms.temple.edu/lammps-ro/trunk@14651 f3b2605a-c512-4ea7-a41b-209d697bcdaa 2016-02-17 00:18:50 +00:00
e4f7383655 git-svn-id: svn://svn.icms.temple.edu/lammps-ro/trunk@14650 f3b2605a-c512-4ea7-a41b-209d697bcdaa 2016-02-17 00:18:28 +00:00
2ad23266fb Fixing Kokkos bug
git-svn-id: svn://svn.icms.temple.edu/lammps-ro/trunk@14649 f3b2605a-c512-4ea7-a41b-209d697bcdaa
2016-02-16 23:13:28 +00:00
1a1b81cb8f Fixing Kokkos bug
git-svn-id: svn://svn.icms.temple.edu/lammps-ro/trunk@14648 f3b2605a-c512-4ea7-a41b-209d697bcdaa
2016-02-16 21:52:49 +00:00
7b9802562c git-svn-id: svn://svn.icms.temple.edu/lammps-ro/trunk@14647 f3b2605a-c512-4ea7-a41b-209d697bcdaa 2016-02-16 15:06:24 +00:00
9d6dbdb0f4 git-svn-id: svn://svn.icms.temple.edu/lammps-ro/trunk@14646 f3b2605a-c512-4ea7-a41b-209d697bcdaa 2016-02-16 15:05:34 +00:00
f5e8f73677 git-svn-id: svn://svn.icms.temple.edu/lammps-ro/trunk@14645 f3b2605a-c512-4ea7-a41b-209d697bcdaa 2016-02-16 15:05:17 +00:00
abe869aaed git-svn-id: svn://svn.icms.temple.edu/lammps-ro/trunk@14644 f3b2605a-c512-4ea7-a41b-209d697bcdaa 2016-02-16 01:07:46 +00:00
319762c37f git-svn-id: svn://svn.icms.temple.edu/lammps-ro/trunk@14643 f3b2605a-c512-4ea7-a41b-209d697bcdaa 2016-02-16 01:06:35 +00:00
40952680d8 git-svn-id: svn://svn.icms.temple.edu/lammps-ro/trunk@14642 f3b2605a-c512-4ea7-a41b-209d697bcdaa 2016-02-16 01:06:20 +00:00
3311380801 git-svn-id: svn://svn.icms.temple.edu/lammps-ro/trunk@14639 f3b2605a-c512-4ea7-a41b-209d697bcdaa 2016-02-15 23:16:16 +00:00
1abf6d03bc git-svn-id: svn://svn.icms.temple.edu/lammps-ro/trunk@14638 f3b2605a-c512-4ea7-a41b-209d697bcdaa 2016-02-15 23:15:24 +00:00
d242d3b6c4 git-svn-id: svn://svn.icms.temple.edu/lammps-ro/trunk@14636 f3b2605a-c512-4ea7-a41b-209d697bcdaa 2016-02-15 22:59:20 +00:00
ae788f1bdc git-svn-id: svn://svn.icms.temple.edu/lammps-ro/trunk@14635 f3b2605a-c512-4ea7-a41b-209d697bcdaa 2016-02-15 22:47:22 +00:00
9b32dcb314 git-svn-id: svn://svn.icms.temple.edu/lammps-ro/trunk@14634 f3b2605a-c512-4ea7-a41b-209d697bcdaa 2016-02-15 22:46:05 +00:00
53b3a4aec0 git-svn-id: svn://svn.icms.temple.edu/lammps-ro/trunk@14633 f3b2605a-c512-4ea7-a41b-209d697bcdaa 2016-02-15 22:35:06 +00:00
52ca6416a0 git-svn-id: svn://svn.icms.temple.edu/lammps-ro/trunk@14632 f3b2605a-c512-4ea7-a41b-209d697bcdaa 2016-02-15 22:34:20 +00:00
21136ae107 git-svn-id: svn://svn.icms.temple.edu/lammps-ro/trunk@14631 f3b2605a-c512-4ea7-a41b-209d697bcdaa 2016-02-15 22:27:57 +00:00
490b26ed06 git-svn-id: svn://svn.icms.temple.edu/lammps-ro/trunk@14630 f3b2605a-c512-4ea7-a41b-209d697bcdaa 2016-02-15 22:27:37 +00:00
952671afd6 git-svn-id: svn://svn.icms.temple.edu/lammps-ro/trunk@14629 f3b2605a-c512-4ea7-a41b-209d697bcdaa 2016-02-15 22:27:20 +00:00
7512821dd6 git-svn-id: svn://svn.icms.temple.edu/lammps-ro/trunk@14628 f3b2605a-c512-4ea7-a41b-209d697bcdaa 2016-02-15 21:20:17 +00:00
507bde5b5a git-svn-id: svn://svn.icms.temple.edu/lammps-ro/trunk@14627 f3b2605a-c512-4ea7-a41b-209d697bcdaa 2016-02-15 21:19:55 +00:00
ed9e162025 git-svn-id: svn://svn.icms.temple.edu/lammps-ro/trunk@14626 f3b2605a-c512-4ea7-a41b-209d697bcdaa 2016-02-15 21:18:29 +00:00
74d9b4656a git-svn-id: svn://svn.icms.temple.edu/lammps-ro/trunk@14624 f3b2605a-c512-4ea7-a41b-209d697bcdaa 2016-02-15 20:16:10 +00:00
0be549d08f git-svn-id: svn://svn.icms.temple.edu/lammps-ro/trunk@14622 f3b2605a-c512-4ea7-a41b-209d697bcdaa 2016-02-15 18:38:15 +00:00
384d5103dd git-svn-id: svn://svn.icms.temple.edu/lammps-ro/trunk@14621 f3b2605a-c512-4ea7-a41b-209d697bcdaa 2016-02-15 18:37:24 +00:00
c43da51b44 git-svn-id: svn://svn.icms.temple.edu/lammps-ro/trunk@14620 f3b2605a-c512-4ea7-a41b-209d697bcdaa 2016-02-15 18:37:06 +00:00
531cd5fe2c git-svn-id: svn://svn.icms.temple.edu/lammps-ro/trunk@14619 f3b2605a-c512-4ea7-a41b-209d697bcdaa 2016-02-15 18:36:42 +00:00
e77d1d2ad2 git-svn-id: svn://svn.icms.temple.edu/lammps-ro/trunk@14618 f3b2605a-c512-4ea7-a41b-209d697bcdaa 2016-02-15 18:35:14 +00:00
2d43f007d1 git-svn-id: svn://svn.icms.temple.edu/lammps-ro/trunk@14617 f3b2605a-c512-4ea7-a41b-209d697bcdaa 2016-02-15 18:32:58 +00:00
9ac8547feb git-svn-id: svn://svn.icms.temple.edu/lammps-ro/trunk@14616 f3b2605a-c512-4ea7-a41b-209d697bcdaa 2016-02-15 18:32:42 +00:00
e7eced3711 git-svn-id: svn://svn.icms.temple.edu/lammps-ro/trunk@14615 f3b2605a-c512-4ea7-a41b-209d697bcdaa 2016-02-15 18:16:07 +00:00
9fc1ec8d30 git-svn-id: svn://svn.icms.temple.edu/lammps-ro/trunk@14614 f3b2605a-c512-4ea7-a41b-209d697bcdaa 2016-02-15 18:15:02 +00:00
2ed1861320 git-svn-id: svn://svn.icms.temple.edu/lammps-ro/trunk@14613 f3b2605a-c512-4ea7-a41b-209d697bcdaa 2016-02-15 18:14:03 +00:00
be035371f9 git-svn-id: svn://svn.icms.temple.edu/lammps-ro/trunk@14612 f3b2605a-c512-4ea7-a41b-209d697bcdaa 2016-02-15 18:12:57 +00:00
701b4a00a1 git-svn-id: svn://svn.icms.temple.edu/lammps-ro/trunk@14611 f3b2605a-c512-4ea7-a41b-209d697bcdaa 2016-02-15 18:11:41 +00:00
bee7ffaee5 git-svn-id: svn://svn.icms.temple.edu/lammps-ro/trunk@14610 f3b2605a-c512-4ea7-a41b-209d697bcdaa 2016-02-15 18:06:11 +00:00
b1922ae7ca git-svn-id: svn://svn.icms.temple.edu/lammps-ro/trunk@14609 f3b2605a-c512-4ea7-a41b-209d697bcdaa 2016-02-15 18:04:59 +00:00
77fe3db02c git-svn-id: svn://svn.icms.temple.edu/lammps-ro/trunk@14608 f3b2605a-c512-4ea7-a41b-209d697bcdaa 2016-02-15 17:21:52 +00:00
85a60a9415 git-svn-id: svn://svn.icms.temple.edu/lammps-ro/trunk@14607 f3b2605a-c512-4ea7-a41b-209d697bcdaa 2016-02-15 17:21:29 +00:00
de853a373b git-svn-id: svn://svn.icms.temple.edu/lammps-ro/trunk@14605 f3b2605a-c512-4ea7-a41b-209d697bcdaa 2016-02-15 16:20:47 +00:00
9d46c0b938 git-svn-id: svn://svn.icms.temple.edu/lammps-ro/trunk@14604 f3b2605a-c512-4ea7-a41b-209d697bcdaa 2016-02-15 16:19:54 +00:00
468ef144d4 git-svn-id: svn://svn.icms.temple.edu/lammps-ro/trunk@14602 f3b2605a-c512-4ea7-a41b-209d697bcdaa 2016-02-15 16:10:43 +00:00
5cd677f369 git-svn-id: svn://svn.icms.temple.edu/lammps-ro/trunk@14601 f3b2605a-c512-4ea7-a41b-209d697bcdaa 2016-02-15 16:09:46 +00:00
6a5d853c5d git-svn-id: svn://svn.icms.temple.edu/lammps-ro/trunk@14600 f3b2605a-c512-4ea7-a41b-209d697bcdaa 2016-02-15 15:50:13 +00:00
53f0833088 git-svn-id: svn://svn.icms.temple.edu/lammps-ro/trunk@14599 f3b2605a-c512-4ea7-a41b-209d697bcdaa 2016-02-15 15:49:59 +00:00
3db8850f09 git-svn-id: svn://svn.icms.temple.edu/lammps-ro/trunk@14598 f3b2605a-c512-4ea7-a41b-209d697bcdaa 2016-02-15 15:47:38 +00:00
e98647efb2 git-svn-id: svn://svn.icms.temple.edu/lammps-ro/trunk@14597 f3b2605a-c512-4ea7-a41b-209d697bcdaa 2016-02-15 15:36:35 +00:00
d615b660c1 git-svn-id: svn://svn.icms.temple.edu/lammps-ro/trunk@14596 f3b2605a-c512-4ea7-a41b-209d697bcdaa 2016-02-15 15:33:36 +00:00
e813620f24 git-svn-id: svn://svn.icms.temple.edu/lammps-ro/trunk@14595 f3b2605a-c512-4ea7-a41b-209d697bcdaa 2016-02-15 15:32:58 +00:00
4e42c28f66 git-svn-id: svn://svn.icms.temple.edu/lammps-ro/trunk@14594 f3b2605a-c512-4ea7-a41b-209d697bcdaa 2016-02-15 15:32:49 +00:00
cd14e4314d git-svn-id: svn://svn.icms.temple.edu/lammps-ro/trunk@14593 f3b2605a-c512-4ea7-a41b-209d697bcdaa 2016-02-15 15:32:27 +00:00
a70623ee43 git-svn-id: svn://svn.icms.temple.edu/lammps-ro/trunk@14592 f3b2605a-c512-4ea7-a41b-209d697bcdaa 2016-02-15 15:30:15 +00:00
729ba2cc21 git-svn-id: svn://svn.icms.temple.edu/lammps-ro/trunk@14590 f3b2605a-c512-4ea7-a41b-209d697bcdaa 2016-02-15 15:28:30 +00:00
70aa37e4c8 git-svn-id: svn://svn.icms.temple.edu/lammps-ro/trunk@14589 f3b2605a-c512-4ea7-a41b-209d697bcdaa 2016-02-15 15:26:57 +00:00
ce579ea42e Adding Kokkos error check
git-svn-id: svn://svn.icms.temple.edu/lammps-ro/trunk@14588 f3b2605a-c512-4ea7-a41b-209d697bcdaa
2016-02-11 16:42:26 +00:00
707dee68c9 Adding Kokkos error check
git-svn-id: svn://svn.icms.temple.edu/lammps-ro/trunk@14587 f3b2605a-c512-4ea7-a41b-209d697bcdaa
2016-02-11 16:40:45 +00:00
7a5587d65e Adding Kokkos error check
git-svn-id: svn://svn.icms.temple.edu/lammps-ro/trunk@14586 f3b2605a-c512-4ea7-a41b-209d697bcdaa
2016-02-11 15:43:25 +00:00
e2117af23f Fixing Kokkos bug
git-svn-id: svn://svn.icms.temple.edu/lammps-ro/trunk@14585 f3b2605a-c512-4ea7-a41b-209d697bcdaa
2016-02-10 23:34:03 +00:00
b6be86264f Fixing Kokkos bug
git-svn-id: svn://svn.icms.temple.edu/lammps-ro/trunk@14584 f3b2605a-c512-4ea7-a41b-209d697bcdaa
2016-02-10 22:50:51 +00:00
ea12e28c6d git-svn-id: svn://svn.icms.temple.edu/lammps-ro/trunk@14583 f3b2605a-c512-4ea7-a41b-209d697bcdaa 2016-02-10 15:02:22 +00:00
9e20448834 git-svn-id: svn://svn.icms.temple.edu/lammps-ro/trunk@14582 f3b2605a-c512-4ea7-a41b-209d697bcdaa 2016-02-08 16:18:20 +00:00
da41fcb74b Fixing Kokkos bugs
git-svn-id: svn://svn.icms.temple.edu/lammps-ro/trunk@14580 f3b2605a-c512-4ea7-a41b-209d697bcdaa
2016-02-05 17:40:10 +00:00
384aef4381 Fixing Kokkos bugs
git-svn-id: svn://svn.icms.temple.edu/lammps-ro/trunk@14579 f3b2605a-c512-4ea7-a41b-209d697bcdaa
2016-02-05 17:28:51 +00:00
753429e654 git-svn-id: svn://svn.icms.temple.edu/lammps-ro/trunk@14555 f3b2605a-c512-4ea7-a41b-209d697bcdaa 2016-02-04 21:09:19 +00:00
02408d39bf git-svn-id: svn://svn.icms.temple.edu/lammps-ro/trunk@14554 f3b2605a-c512-4ea7-a41b-209d697bcdaa 2016-02-04 21:09:03 +00:00
b2d417f5d7 git-svn-id: svn://svn.icms.temple.edu/lammps-ro/trunk@14552 f3b2605a-c512-4ea7-a41b-209d697bcdaa 2016-02-04 20:39:00 +00:00
6b593f0c91 git-svn-id: svn://svn.icms.temple.edu/lammps-ro/trunk@14550 f3b2605a-c512-4ea7-a41b-209d697bcdaa 2016-02-04 17:17:35 +00:00
ddb83e1b72 git-svn-id: svn://svn.icms.temple.edu/lammps-ro/trunk@14548 f3b2605a-c512-4ea7-a41b-209d697bcdaa 2016-02-03 21:58:23 +00:00
21ee4aa974 git-svn-id: svn://svn.icms.temple.edu/lammps-ro/trunk@14547 f3b2605a-c512-4ea7-a41b-209d697bcdaa 2016-02-03 21:57:30 +00:00
c1aebdd391 git-svn-id: svn://svn.icms.temple.edu/lammps-ro/trunk@14544 f3b2605a-c512-4ea7-a41b-209d697bcdaa 2016-02-03 21:39:32 +00:00
b25125e529 git-svn-id: svn://svn.icms.temple.edu/lammps-ro/trunk@14543 f3b2605a-c512-4ea7-a41b-209d697bcdaa 2016-02-03 21:37:16 +00:00
de6ff01ba7 git-svn-id: svn://svn.icms.temple.edu/lammps-ro/trunk@14542 f3b2605a-c512-4ea7-a41b-209d697bcdaa 2016-02-03 21:33:41 +00:00
94da81bae4 git-svn-id: svn://svn.icms.temple.edu/lammps-ro/trunk@14541 f3b2605a-c512-4ea7-a41b-209d697bcdaa 2016-02-03 21:27:32 +00:00
bb11aa3147 git-svn-id: svn://svn.icms.temple.edu/lammps-ro/trunk@14540 f3b2605a-c512-4ea7-a41b-209d697bcdaa 2016-02-03 21:27:19 +00:00
22961a9267 git-svn-id: svn://svn.icms.temple.edu/lammps-ro/trunk@14539 f3b2605a-c512-4ea7-a41b-209d697bcdaa 2016-02-03 21:23:06 +00:00
0963784ff1 git-svn-id: svn://svn.icms.temple.edu/lammps-ro/trunk@14538 f3b2605a-c512-4ea7-a41b-209d697bcdaa 2016-02-03 21:22:35 +00:00
0662467254 git-svn-id: svn://svn.icms.temple.edu/lammps-ro/trunk@14537 f3b2605a-c512-4ea7-a41b-209d697bcdaa 2016-02-03 21:20:11 +00:00
1a82fbf0ac git-svn-id: svn://svn.icms.temple.edu/lammps-ro/trunk@14536 f3b2605a-c512-4ea7-a41b-209d697bcdaa 2016-02-03 21:19:53 +00:00
216f83d10d git-svn-id: svn://svn.icms.temple.edu/lammps-ro/trunk@14535 f3b2605a-c512-4ea7-a41b-209d697bcdaa 2016-02-03 21:18:58 +00:00
8e9f616a2e git-svn-id: svn://svn.icms.temple.edu/lammps-ro/trunk@14534 f3b2605a-c512-4ea7-a41b-209d697bcdaa 2016-02-03 20:55:24 +00:00
3e9a18b5bb git-svn-id: svn://svn.icms.temple.edu/lammps-ro/trunk@14533 f3b2605a-c512-4ea7-a41b-209d697bcdaa 2016-02-03 20:53:35 +00:00
9cf43f2a0b git-svn-id: svn://svn.icms.temple.edu/lammps-ro/trunk@14532 f3b2605a-c512-4ea7-a41b-209d697bcdaa 2016-02-02 16:05:07 +00:00
0766f95adc git-svn-id: svn://svn.icms.temple.edu/lammps-ro/trunk@14531 f3b2605a-c512-4ea7-a41b-209d697bcdaa 2016-02-02 16:01:46 +00:00
2da2a3b03a git-svn-id: svn://svn.icms.temple.edu/lammps-ro/trunk@14530 f3b2605a-c512-4ea7-a41b-209d697bcdaa 2016-02-02 16:00:13 +00:00
60ec762876 git-svn-id: svn://svn.icms.temple.edu/lammps-ro/trunk@14529 f3b2605a-c512-4ea7-a41b-209d697bcdaa 2016-01-29 20:37:11 +00:00
029faeb7f0 git-svn-id: svn://svn.icms.temple.edu/lammps-ro/trunk@14528 f3b2605a-c512-4ea7-a41b-209d697bcdaa 2016-01-29 18:27:37 +00:00
3d7d897fe8 Changing Kokkos default
git-svn-id: svn://svn.icms.temple.edu/lammps-ro/trunk@14527 f3b2605a-c512-4ea7-a41b-209d697bcdaa
2016-01-29 17:39:04 +00:00
a8d0bf986f Changing Kokkos default
git-svn-id: svn://svn.icms.temple.edu/lammps-ro/trunk@14526 f3b2605a-c512-4ea7-a41b-209d697bcdaa
2016-01-29 17:38:58 +00:00
daaadcdd86 Adding Kokkos output and error checks
git-svn-id: svn://svn.icms.temple.edu/lammps-ro/trunk@14525 f3b2605a-c512-4ea7-a41b-209d697bcdaa
2016-01-29 17:24:04 +00:00
7ae6db5a98 git-svn-id: svn://svn.icms.temple.edu/lammps-ro/trunk@14523 f3b2605a-c512-4ea7-a41b-209d697bcdaa 2016-01-29 17:14:30 +00:00
7b091eb651 git-svn-id: svn://svn.icms.temple.edu/lammps-ro/trunk@14522 f3b2605a-c512-4ea7-a41b-209d697bcdaa 2016-01-29 17:13:39 +00:00
0f877f6fde git-svn-id: svn://svn.icms.temple.edu/lammps-ro/trunk@14521 f3b2605a-c512-4ea7-a41b-209d697bcdaa 2016-01-29 17:09:55 +00:00
ccf5f31eb9 git-svn-id: svn://svn.icms.temple.edu/lammps-ro/trunk@14520 f3b2605a-c512-4ea7-a41b-209d697bcdaa 2016-01-29 16:55:57 +00:00
dd2d454fc4 git-svn-id: svn://svn.icms.temple.edu/lammps-ro/trunk@14519 f3b2605a-c512-4ea7-a41b-209d697bcdaa 2016-01-29 16:44:50 +00:00
ce471ffcdb git-svn-id: svn://svn.icms.temple.edu/lammps-ro/trunk@14518 f3b2605a-c512-4ea7-a41b-209d697bcdaa 2016-01-29 16:31:49 +00:00
cf77d882eb Adding error checks for Kokkos-incompatible features
git-svn-id: svn://svn.icms.temple.edu/lammps-ro/trunk@14517 f3b2605a-c512-4ea7-a41b-209d697bcdaa
2016-01-28 23:32:16 +00:00
def0d4ed80 Fixing logic error with suffix styles
git-svn-id: svn://svn.icms.temple.edu/lammps-ro/trunk@14516 f3b2605a-c512-4ea7-a41b-209d697bcdaa
2016-01-28 23:28:12 +00:00
f86a9f4283 git-svn-id: svn://svn.icms.temple.edu/lammps-ro/trunk@14515 f3b2605a-c512-4ea7-a41b-209d697bcdaa 2016-01-28 21:24:58 +00:00
bc15585acd git-svn-id: svn://svn.icms.temple.edu/lammps-ro/trunk@14514 f3b2605a-c512-4ea7-a41b-209d697bcdaa 2016-01-28 21:24:27 +00:00
237743e6fd git-svn-id: svn://svn.icms.temple.edu/lammps-ro/trunk@14513 f3b2605a-c512-4ea7-a41b-209d697bcdaa 2016-01-28 21:22:58 +00:00
997ce626ac git-svn-id: svn://svn.icms.temple.edu/lammps-ro/trunk@14512 f3b2605a-c512-4ea7-a41b-209d697bcdaa 2016-01-28 21:20:27 +00:00
7e67afad7b git-svn-id: svn://svn.icms.temple.edu/lammps-ro/trunk@14511 f3b2605a-c512-4ea7-a41b-209d697bcdaa 2016-01-28 21:19:43 +00:00
d42ebe4e5e git-svn-id: svn://svn.icms.temple.edu/lammps-ro/trunk@14510 f3b2605a-c512-4ea7-a41b-209d697bcdaa 2016-01-28 21:17:32 +00:00
db773d752e git-svn-id: svn://svn.icms.temple.edu/lammps-ro/trunk@14509 f3b2605a-c512-4ea7-a41b-209d697bcdaa 2016-01-28 21:14:10 +00:00
b355fa25f4 Squelching annoying kokkos output
git-svn-id: svn://svn.icms.temple.edu/lammps-ro/trunk@14508 f3b2605a-c512-4ea7-a41b-209d697bcdaa
2016-01-28 21:07:07 +00:00
ed17939b75 git-svn-id: svn://svn.icms.temple.edu/lammps-ro/trunk@14507 f3b2605a-c512-4ea7-a41b-209d697bcdaa 2016-01-27 23:15:03 +00:00
fa7543b714 git-svn-id: svn://svn.icms.temple.edu/lammps-ro/trunk@14506 f3b2605a-c512-4ea7-a41b-209d697bcdaa 2016-01-27 20:33:55 +00:00
dca90d44b7 Fixing Kokkos bug
git-svn-id: svn://svn.icms.temple.edu/lammps-ro/trunk@14505 f3b2605a-c512-4ea7-a41b-209d697bcdaa
2016-01-27 18:40:46 +00:00
0541d52c17 Fixing Kokkos bug
git-svn-id: svn://svn.icms.temple.edu/lammps-ro/trunk@14504 f3b2605a-c512-4ea7-a41b-209d697bcdaa
2016-01-26 23:49:57 +00:00
8c4d592613 Fixing Kokkos bug
git-svn-id: svn://svn.icms.temple.edu/lammps-ro/trunk@14503 f3b2605a-c512-4ea7-a41b-209d697bcdaa
2016-01-26 22:07:21 +00:00
b5c90fac9c git-svn-id: svn://svn.icms.temple.edu/lammps-ro/trunk@14501 f3b2605a-c512-4ea7-a41b-209d697bcdaa 2016-01-26 15:13:45 +00:00
f7731ba58e git-svn-id: svn://svn.icms.temple.edu/lammps-ro/trunk@14500 f3b2605a-c512-4ea7-a41b-209d697bcdaa 2016-01-26 15:12:53 +00:00
db9fe7ac9f git-svn-id: svn://svn.icms.temple.edu/lammps-ro/trunk@14499 f3b2605a-c512-4ea7-a41b-209d697bcdaa 2016-01-26 15:11:58 +00:00
ef429798c0 git-svn-id: svn://svn.icms.temple.edu/lammps-ro/trunk@14498 f3b2605a-c512-4ea7-a41b-209d697bcdaa 2016-01-26 15:07:02 +00:00
e18635c518 git-svn-id: svn://svn.icms.temple.edu/lammps-ro/trunk@14496 f3b2605a-c512-4ea7-a41b-209d697bcdaa 2016-01-26 00:42:04 +00:00
c25e431864 git-svn-id: svn://svn.icms.temple.edu/lammps-ro/trunk@14495 f3b2605a-c512-4ea7-a41b-209d697bcdaa 2016-01-26 00:41:12 +00:00
49051197a9 git-svn-id: svn://svn.icms.temple.edu/lammps-ro/trunk@14494 f3b2605a-c512-4ea7-a41b-209d697bcdaa 2016-01-25 23:04:56 +00:00
2d78e1a215 git-svn-id: svn://svn.icms.temple.edu/lammps-ro/trunk@14493 f3b2605a-c512-4ea7-a41b-209d697bcdaa 2016-01-25 23:04:42 +00:00
0fbc48723e git-svn-id: svn://svn.icms.temple.edu/lammps-ro/trunk@14492 f3b2605a-c512-4ea7-a41b-209d697bcdaa 2016-01-25 22:50:16 +00:00
12858669c0 git-svn-id: svn://svn.icms.temple.edu/lammps-ro/trunk@14491 f3b2605a-c512-4ea7-a41b-209d697bcdaa 2016-01-25 22:45:10 +00:00
bf9517b562 Fixing Kokkos bugs
git-svn-id: svn://svn.icms.temple.edu/lammps-ro/trunk@14490 f3b2605a-c512-4ea7-a41b-209d697bcdaa
2016-01-25 21:36:10 +00:00
92ae19e280 Fixing Kokkos bugs
git-svn-id: svn://svn.icms.temple.edu/lammps-ro/trunk@14489 f3b2605a-c512-4ea7-a41b-209d697bcdaa
2016-01-25 20:58:13 +00:00
9eb8702f2f Fixing Kokkos bugs
git-svn-id: svn://svn.icms.temple.edu/lammps-ro/trunk@14488 f3b2605a-c512-4ea7-a41b-209d697bcdaa
2016-01-25 16:13:12 +00:00
775b163eae git-svn-id: svn://svn.icms.temple.edu/lammps-ro/trunk@14487 f3b2605a-c512-4ea7-a41b-209d697bcdaa 2016-01-25 15:46:03 +00:00
aca24423f4 git-svn-id: svn://svn.icms.temple.edu/lammps-ro/trunk@14485 f3b2605a-c512-4ea7-a41b-209d697bcdaa 2016-01-23 01:00:50 +00:00
f593ffb67b git-svn-id: svn://svn.icms.temple.edu/lammps-ro/trunk@14484 f3b2605a-c512-4ea7-a41b-209d697bcdaa 2016-01-23 00:59:59 +00:00
67d5513df1 git-svn-id: svn://svn.icms.temple.edu/lammps-ro/trunk@14483 f3b2605a-c512-4ea7-a41b-209d697bcdaa 2016-01-23 00:59:04 +00:00
b31f864e83 git-svn-id: svn://svn.icms.temple.edu/lammps-ro/trunk@14482 f3b2605a-c512-4ea7-a41b-209d697bcdaa 2016-01-23 00:55:48 +00:00
8b7e9fea4c git-svn-id: svn://svn.icms.temple.edu/lammps-ro/trunk@14481 f3b2605a-c512-4ea7-a41b-209d697bcdaa 2016-01-23 00:49:56 +00:00
3621171480 git-svn-id: svn://svn.icms.temple.edu/lammps-ro/trunk@14480 f3b2605a-c512-4ea7-a41b-209d697bcdaa 2016-01-23 00:49:16 +00:00
5d99bf664e git-svn-id: svn://svn.icms.temple.edu/lammps-ro/trunk@14479 f3b2605a-c512-4ea7-a41b-209d697bcdaa 2016-01-23 00:49:10 +00:00
7f049bb31b git-svn-id: svn://svn.icms.temple.edu/lammps-ro/trunk@14478 f3b2605a-c512-4ea7-a41b-209d697bcdaa 2016-01-23 00:48:23 +00:00
627561054c git-svn-id: svn://svn.icms.temple.edu/lammps-ro/trunk@14477 f3b2605a-c512-4ea7-a41b-209d697bcdaa 2016-01-23 00:48:06 +00:00
d4a5571aab git-svn-id: svn://svn.icms.temple.edu/lammps-ro/trunk@14476 f3b2605a-c512-4ea7-a41b-209d697bcdaa 2016-01-23 00:47:55 +00:00
28d6cef697 git-svn-id: svn://svn.icms.temple.edu/lammps-ro/trunk@14475 f3b2605a-c512-4ea7-a41b-209d697bcdaa 2016-01-23 00:47:38 +00:00
79466c37d8 git-svn-id: svn://svn.icms.temple.edu/lammps-ro/trunk@14474 f3b2605a-c512-4ea7-a41b-209d697bcdaa 2016-01-23 00:47:25 +00:00
ad1d7ffdc7 Fixed another bug in edge_histo
git-svn-id: svn://svn.icms.temple.edu/lammps-ro/trunk@14473 f3b2605a-c512-4ea7-a41b-209d697bcdaa
2016-01-23 00:09:48 +00:00
34785af4a6 Fixing Kokkos bugs
git-svn-id: svn://svn.icms.temple.edu/lammps-ro/trunk@14472 f3b2605a-c512-4ea7-a41b-209d697bcdaa
2016-01-22 23:35:38 +00:00
c8fb84783e Fixing Kokkos bug
git-svn-id: svn://svn.icms.temple.edu/lammps-ro/trunk@14471 f3b2605a-c512-4ea7-a41b-209d697bcdaa
2016-01-22 22:36:08 +00:00
e3e687f7b3 Added peratom keyword
git-svn-id: svn://svn.icms.temple.edu/lammps-ro/trunk@14470 f3b2605a-c512-4ea7-a41b-209d697bcdaa
2016-01-22 22:30:40 +00:00
56ef0d40fc Added description of peratom keyword
git-svn-id: svn://svn.icms.temple.edu/lammps-ro/trunk@14469 f3b2605a-c512-4ea7-a41b-209d697bcdaa
2016-01-22 03:14:16 +00:00
4da36d1c01 Added data output example
git-svn-id: svn://svn.icms.temple.edu/lammps-ro/trunk@14468 f3b2605a-c512-4ea7-a41b-209d697bcdaa
2016-01-22 02:49:52 +00:00
eb105410ad Fixing Kokkos bug
git-svn-id: svn://svn.icms.temple.edu/lammps-ro/trunk@14467 f3b2605a-c512-4ea7-a41b-209d697bcdaa
2016-01-21 20:15:34 +00:00
eebf110e02 Further tweaked the flangevin issue
git-svn-id: svn://svn.icms.temple.edu/lammps-ro/trunk@14466 f3b2605a-c512-4ea7-a41b-209d697bcdaa
2016-01-21 18:16:07 +00:00
d13b2ecfc9 Added faces as local compute
git-svn-id: svn://svn.icms.temple.edu/lammps-ro/trunk@14465 f3b2605a-c512-4ea7-a41b-209d697bcdaa
2016-01-21 02:34:41 +00:00
66a048bbb9 Added faces as local compute
git-svn-id: svn://svn.icms.temple.edu/lammps-ro/trunk@14464 f3b2605a-c512-4ea7-a41b-209d697bcdaa
2016-01-21 01:59:48 +00:00
005f9d5ac5 Added faces as local compute
git-svn-id: svn://svn.icms.temple.edu/lammps-ro/trunk@14463 f3b2605a-c512-4ea7-a41b-209d697bcdaa
2016-01-21 01:58:43 +00:00
fe10d8c87b Removed flangevin==NULL check
git-svn-id: svn://svn.icms.temple.edu/lammps-ro/trunk@14462 f3b2605a-c512-4ea7-a41b-209d697bcdaa
2016-01-20 17:29:40 +00:00
fa01d915e3 Changed behavior for non-periodic systems
git-svn-id: svn://svn.icms.temple.edu/lammps-ro/trunk@14461 f3b2605a-c512-4ea7-a41b-209d697bcdaa
2016-01-20 00:03:33 +00:00
81d55a7904 Changed behavior for non-periodic systems
git-svn-id: svn://svn.icms.temple.edu/lammps-ro/trunk@14460 f3b2605a-c512-4ea7-a41b-209d697bcdaa
2016-01-20 00:01:34 +00:00
a3c0f3e43e git-svn-id: svn://svn.icms.temple.edu/lammps-ro/trunk@14459 f3b2605a-c512-4ea7-a41b-209d697bcdaa 2016-01-19 23:12:58 +00:00
7c810d6198 Added 2d example
git-svn-id: svn://svn.icms.temple.edu/lammps-ro/trunk@14458 f3b2605a-c512-4ea7-a41b-209d697bcdaa
2016-01-19 22:26:48 +00:00
0d1c7af98b Updated description of 2d and free surface cases
git-svn-id: svn://svn.icms.temple.edu/lammps-ro/trunk@14457 f3b2605a-c512-4ea7-a41b-209d697bcdaa
2016-01-19 22:18:55 +00:00
f16130ab67 git-svn-id: svn://svn.icms.temple.edu/lammps-ro/trunk@14456 f3b2605a-c512-4ea7-a41b-209d697bcdaa 2016-01-19 20:52:08 +00:00
f6f88c2f28 git-svn-id: svn://svn.icms.temple.edu/lammps-ro/trunk@14455 f3b2605a-c512-4ea7-a41b-209d697bcdaa 2016-01-15 19:40:18 +00:00
0d79f78f73 git-svn-id: svn://svn.icms.temple.edu/lammps-ro/trunk@14453 f3b2605a-c512-4ea7-a41b-209d697bcdaa 2016-01-15 16:06:39 +00:00
b66f1e351e git-svn-id: svn://svn.icms.temple.edu/lammps-ro/trunk@14452 f3b2605a-c512-4ea7-a41b-209d697bcdaa 2016-01-15 16:05:47 +00:00
fb0230aed2 git-svn-id: svn://svn.icms.temple.edu/lammps-ro/trunk@14451 f3b2605a-c512-4ea7-a41b-209d697bcdaa 2016-01-15 16:05:33 +00:00
1174eacfbb git-svn-id: svn://svn.icms.temple.edu/lammps-ro/trunk@14450 f3b2605a-c512-4ea7-a41b-209d697bcdaa 2016-01-15 15:58:24 +00:00
984a1afa72 git-svn-id: svn://svn.icms.temple.edu/lammps-ro/trunk@14449 f3b2605a-c512-4ea7-a41b-209d697bcdaa 2016-01-15 15:53:23 +00:00
bfa988207e git-svn-id: svn://svn.icms.temple.edu/lammps-ro/trunk@14448 f3b2605a-c512-4ea7-a41b-209d697bcdaa 2016-01-15 15:39:44 +00:00
ef6beaad54 git-svn-id: svn://svn.icms.temple.edu/lammps-ro/trunk@14447 f3b2605a-c512-4ea7-a41b-209d697bcdaa 2016-01-15 15:32:31 +00:00
c077947285 git-svn-id: svn://svn.icms.temple.edu/lammps-ro/trunk@14446 f3b2605a-c512-4ea7-a41b-209d697bcdaa 2016-01-15 15:30:46 +00:00
ba87515112 git-svn-id: svn://svn.icms.temple.edu/lammps-ro/trunk@14444 f3b2605a-c512-4ea7-a41b-209d697bcdaa 2016-01-15 00:35:21 +00:00
51a2504a01 git-svn-id: svn://svn.icms.temple.edu/lammps-ro/trunk@14443 f3b2605a-c512-4ea7-a41b-209d697bcdaa 2016-01-15 00:34:30 +00:00
3e36b2b16b git-svn-id: svn://svn.icms.temple.edu/lammps-ro/trunk@14442 f3b2605a-c512-4ea7-a41b-209d697bcdaa 2016-01-15 00:28:30 +00:00
5bba67f290 git-svn-id: svn://svn.icms.temple.edu/lammps-ro/trunk@14441 f3b2605a-c512-4ea7-a41b-209d697bcdaa 2016-01-14 22:35:15 +00:00
09fb68df71 git-svn-id: svn://svn.icms.temple.edu/lammps-ro/trunk@14440 f3b2605a-c512-4ea7-a41b-209d697bcdaa 2016-01-14 22:33:46 +00:00
cd1f4ae7f0 git-svn-id: svn://svn.icms.temple.edu/lammps-ro/trunk@14438 f3b2605a-c512-4ea7-a41b-209d697bcdaa 2016-01-12 22:29:26 +00:00
7bd6ae2d6a git-svn-id: svn://svn.icms.temple.edu/lammps-ro/trunk@14437 f3b2605a-c512-4ea7-a41b-209d697bcdaa 2016-01-12 22:28:35 +00:00
843814f7a3 git-svn-id: svn://svn.icms.temple.edu/lammps-ro/trunk@14436 f3b2605a-c512-4ea7-a41b-209d697bcdaa 2016-01-12 22:25:33 +00:00
9dcf6ee717 git-svn-id: svn://svn.icms.temple.edu/lammps-ro/trunk@14435 f3b2605a-c512-4ea7-a41b-209d697bcdaa 2016-01-12 22:25:10 +00:00
a34b9f17da git-svn-id: svn://svn.icms.temple.edu/lammps-ro/trunk@14433 f3b2605a-c512-4ea7-a41b-209d697bcdaa 2016-01-12 16:48:49 +00:00
dc385fcbd1 git-svn-id: svn://svn.icms.temple.edu/lammps-ro/trunk@14432 f3b2605a-c512-4ea7-a41b-209d697bcdaa 2016-01-12 16:47:56 +00:00
395397d30b git-svn-id: svn://svn.icms.temple.edu/lammps-ro/trunk@14431 f3b2605a-c512-4ea7-a41b-209d697bcdaa 2016-01-12 16:34:59 +00:00
2520edf91e git-svn-id: svn://svn.icms.temple.edu/lammps-ro/trunk@14430 f3b2605a-c512-4ea7-a41b-209d697bcdaa 2016-01-12 16:27:53 +00:00
b9ab7a0cd5 git-svn-id: svn://svn.icms.temple.edu/lammps-ro/trunk@14429 f3b2605a-c512-4ea7-a41b-209d697bcdaa 2016-01-12 16:27:19 +00:00
fbd03b390f git-svn-id: svn://svn.icms.temple.edu/lammps-ro/trunk@14428 f3b2605a-c512-4ea7-a41b-209d697bcdaa 2016-01-12 16:24:33 +00:00
05240f9467 git-svn-id: svn://svn.icms.temple.edu/lammps-ro/trunk@14427 f3b2605a-c512-4ea7-a41b-209d697bcdaa 2016-01-12 00:43:44 +00:00
0da05dcdf4 git-svn-id: svn://svn.icms.temple.edu/lammps-ro/trunk@14426 f3b2605a-c512-4ea7-a41b-209d697bcdaa 2016-01-12 00:29:49 +00:00
20cc9e51cd git-svn-id: svn://svn.icms.temple.edu/lammps-ro/trunk@14425 f3b2605a-c512-4ea7-a41b-209d697bcdaa 2016-01-12 00:13:11 +00:00
4e199dd5bd git-svn-id: svn://svn.icms.temple.edu/lammps-ro/trunk@14424 f3b2605a-c512-4ea7-a41b-209d697bcdaa 2016-01-12 00:12:15 +00:00
12c8aaf29d git-svn-id: svn://svn.icms.temple.edu/lammps-ro/trunk@14423 f3b2605a-c512-4ea7-a41b-209d697bcdaa 2016-01-11 23:38:16 +00:00
681ebfaf8f git-svn-id: svn://svn.icms.temple.edu/lammps-ro/trunk@14422 f3b2605a-c512-4ea7-a41b-209d697bcdaa 2016-01-11 23:34:59 +00:00
e72cef0c3a git-svn-id: svn://svn.icms.temple.edu/lammps-ro/trunk@14421 f3b2605a-c512-4ea7-a41b-209d697bcdaa 2016-01-11 22:00:29 +00:00
e24ed78dd1 git-svn-id: svn://svn.icms.temple.edu/lammps-ro/trunk@14420 f3b2605a-c512-4ea7-a41b-209d697bcdaa 2016-01-11 22:00:15 +00:00
d3f534f5cc git-svn-id: svn://svn.icms.temple.edu/lammps-ro/trunk@14419 f3b2605a-c512-4ea7-a41b-209d697bcdaa 2016-01-11 22:00:09 +00:00
f5946c73f7 git-svn-id: svn://svn.icms.temple.edu/lammps-ro/trunk@14418 f3b2605a-c512-4ea7-a41b-209d697bcdaa 2016-01-11 22:00:04 +00:00
7ae65d4dcf git-svn-id: svn://svn.icms.temple.edu/lammps-ro/trunk@14417 f3b2605a-c512-4ea7-a41b-209d697bcdaa 2016-01-11 22:00:02 +00:00
b5086e3d69 git-svn-id: svn://svn.icms.temple.edu/lammps-ro/trunk@14416 f3b2605a-c512-4ea7-a41b-209d697bcdaa 2016-01-11 21:59:49 +00:00
173d4861a2 git-svn-id: svn://svn.icms.temple.edu/lammps-ro/trunk@14415 f3b2605a-c512-4ea7-a41b-209d697bcdaa 2016-01-11 21:59:38 +00:00
9e4140c954 git-svn-id: svn://svn.icms.temple.edu/lammps-ro/trunk@14413 f3b2605a-c512-4ea7-a41b-209d697bcdaa 2016-01-11 18:57:20 +00:00
eb20bcf183 git-svn-id: svn://svn.icms.temple.edu/lammps-ro/trunk@14412 f3b2605a-c512-4ea7-a41b-209d697bcdaa 2016-01-11 18:56:25 +00:00
eb063200bf git-svn-id: svn://svn.icms.temple.edu/lammps-ro/trunk@14411 f3b2605a-c512-4ea7-a41b-209d697bcdaa 2016-01-11 17:59:15 +00:00
e86cfe431a git-svn-id: svn://svn.icms.temple.edu/lammps-ro/trunk@14410 f3b2605a-c512-4ea7-a41b-209d697bcdaa 2016-01-11 17:58:37 +00:00
51745e2c95 git-svn-id: svn://svn.icms.temple.edu/lammps-ro/trunk@14409 f3b2605a-c512-4ea7-a41b-209d697bcdaa 2016-01-11 16:01:39 +00:00
06e04df4e3 git-svn-id: svn://svn.icms.temple.edu/lammps-ro/trunk@14406 f3b2605a-c512-4ea7-a41b-209d697bcdaa 2015-12-24 00:03:25 +00:00
41a026ee14 git-svn-id: svn://svn.icms.temple.edu/lammps-ro/trunk@14405 f3b2605a-c512-4ea7-a41b-209d697bcdaa 2015-12-24 00:02:10 +00:00
ce72ce4e13 git-svn-id: svn://svn.icms.temple.edu/lammps-ro/trunk@14404 f3b2605a-c512-4ea7-a41b-209d697bcdaa 2015-12-24 00:00:27 +00:00
ca0d0dcb63 git-svn-id: svn://svn.icms.temple.edu/lammps-ro/trunk@14402 f3b2605a-c512-4ea7-a41b-209d697bcdaa 2015-12-23 17:23:17 +00:00
f0cd77adef git-svn-id: svn://svn.icms.temple.edu/lammps-ro/trunk@14401 f3b2605a-c512-4ea7-a41b-209d697bcdaa 2015-12-23 17:21:50 +00:00
10c246dddb git-svn-id: svn://svn.icms.temple.edu/lammps-ro/trunk@14400 f3b2605a-c512-4ea7-a41b-209d697bcdaa 2015-12-23 17:21:36 +00:00
57dc7d32d4 git-svn-id: svn://svn.icms.temple.edu/lammps-ro/trunk@14399 f3b2605a-c512-4ea7-a41b-209d697bcdaa 2015-12-23 17:07:47 +00:00
29b4d60141 git-svn-id: svn://svn.icms.temple.edu/lammps-ro/trunk@14398 f3b2605a-c512-4ea7-a41b-209d697bcdaa 2015-12-23 17:04:56 +00:00
202bc7b713 Added comment about momentum under NPT
git-svn-id: svn://svn.icms.temple.edu/lammps-ro/trunk@14397 f3b2605a-c512-4ea7-a41b-209d697bcdaa
2015-12-23 01:40:29 +00:00
ca9b0adad7 Adding changes from Mike Brown to USER-INTEL for gcc compatibility
git-svn-id: svn://svn.icms.temple.edu/lammps-ro/trunk@14396 f3b2605a-c512-4ea7-a41b-209d697bcdaa
2015-12-22 20:16:56 +00:00
0fd57a6fe4 git-svn-id: svn://svn.icms.temple.edu/lammps-ro/trunk@14395 f3b2605a-c512-4ea7-a41b-209d697bcdaa 2015-12-21 16:44:00 +00:00
6f10e999c9 git-svn-id: svn://svn.icms.temple.edu/lammps-ro/trunk@14394 f3b2605a-c512-4ea7-a41b-209d697bcdaa 2015-12-21 15:20:41 +00:00
0656a3e0b5 git-svn-id: svn://svn.icms.temple.edu/lammps-ro/trunk@14392 f3b2605a-c512-4ea7-a41b-209d697bcdaa 2015-12-19 18:43:03 +00:00
e2ec56c171 git-svn-id: svn://svn.icms.temple.edu/lammps-ro/trunk@14391 f3b2605a-c512-4ea7-a41b-209d697bcdaa 2015-12-19 18:42:24 +00:00
2926854b8d git-svn-id: svn://svn.icms.temple.edu/lammps-ro/trunk@14390 f3b2605a-c512-4ea7-a41b-209d697bcdaa 2015-12-19 18:32:29 +00:00
df8b58a8c6 git-svn-id: svn://svn.icms.temple.edu/lammps-ro/trunk@14389 f3b2605a-c512-4ea7-a41b-209d697bcdaa 2015-12-19 18:32:22 +00:00
528053b65f git-svn-id: svn://svn.icms.temple.edu/lammps-ro/trunk@14388 f3b2605a-c512-4ea7-a41b-209d697bcdaa 2015-12-19 18:31:40 +00:00
15063188b8 git-svn-id: svn://svn.icms.temple.edu/lammps-ro/trunk@14387 f3b2605a-c512-4ea7-a41b-209d697bcdaa 2015-12-19 18:31:14 +00:00
72ebb1dd9b git-svn-id: svn://svn.icms.temple.edu/lammps-ro/trunk@14386 f3b2605a-c512-4ea7-a41b-209d697bcdaa 2015-12-19 18:30:53 +00:00
fb2cb218c8 git-svn-id: svn://svn.icms.temple.edu/lammps-ro/trunk@14383 f3b2605a-c512-4ea7-a41b-209d697bcdaa 2015-12-17 23:18:09 +00:00
71b325e52f git-svn-id: svn://svn.icms.temple.edu/lammps-ro/trunk@14382 f3b2605a-c512-4ea7-a41b-209d697bcdaa 2015-12-17 23:09:35 +00:00
7e9270e813 git-svn-id: svn://svn.icms.temple.edu/lammps-ro/trunk@14381 f3b2605a-c512-4ea7-a41b-209d697bcdaa 2015-12-17 23:05:20 +00:00
e37c9faf05 git-svn-id: svn://svn.icms.temple.edu/lammps-ro/trunk@14380 f3b2605a-c512-4ea7-a41b-209d697bcdaa 2015-12-17 23:05:03 +00:00
f7e17e97e8 git-svn-id: svn://svn.icms.temple.edu/lammps-ro/trunk@14379 f3b2605a-c512-4ea7-a41b-209d697bcdaa 2015-12-17 15:46:12 +00:00
df7e89f9f7 git-svn-id: svn://svn.icms.temple.edu/lammps-ro/trunk@14378 f3b2605a-c512-4ea7-a41b-209d697bcdaa 2015-12-16 16:47:05 +00:00
8dff51d5c9 git-svn-id: svn://svn.icms.temple.edu/lammps-ro/trunk@14377 f3b2605a-c512-4ea7-a41b-209d697bcdaa 2015-12-15 23:48:11 +00:00
9e3b61d926 git-svn-id: svn://svn.icms.temple.edu/lammps-ro/trunk@14376 f3b2605a-c512-4ea7-a41b-209d697bcdaa 2015-12-15 23:32:24 +00:00
f3af23a41b git-svn-id: svn://svn.icms.temple.edu/lammps-ro/trunk@14375 f3b2605a-c512-4ea7-a41b-209d697bcdaa 2015-12-15 23:31:16 +00:00
ead53bc9e7 git-svn-id: svn://svn.icms.temple.edu/lammps-ro/trunk@14373 f3b2605a-c512-4ea7-a41b-209d697bcdaa 2015-12-15 22:33:45 +00:00
0939eb1ee9 git-svn-id: svn://svn.icms.temple.edu/lammps-ro/trunk@14372 f3b2605a-c512-4ea7-a41b-209d697bcdaa 2015-12-15 22:30:45 +00:00
b5a1ba9bfa git-svn-id: svn://svn.icms.temple.edu/lammps-ro/trunk@14371 f3b2605a-c512-4ea7-a41b-209d697bcdaa 2015-12-15 22:29:37 +00:00
06a217aa08 git-svn-id: svn://svn.icms.temple.edu/lammps-ro/trunk@14370 f3b2605a-c512-4ea7-a41b-209d697bcdaa 2015-12-15 22:26:37 +00:00
31f22919ab git-svn-id: svn://svn.icms.temple.edu/lammps-ro/trunk@14369 f3b2605a-c512-4ea7-a41b-209d697bcdaa 2015-12-15 16:03:53 +00:00
9f38620e13 git-svn-id: svn://svn.icms.temple.edu/lammps-ro/trunk@14368 f3b2605a-c512-4ea7-a41b-209d697bcdaa 2015-12-15 16:01:54 +00:00
074d5962c8 git-svn-id: svn://svn.icms.temple.edu/lammps-ro/trunk@14367 f3b2605a-c512-4ea7-a41b-209d697bcdaa 2015-12-15 15:59:11 +00:00
5f802f86b5 git-svn-id: svn://svn.icms.temple.edu/lammps-ro/trunk@14366 f3b2605a-c512-4ea7-a41b-209d697bcdaa 2015-12-15 15:59:01 +00:00
75de12f26a git-svn-id: svn://svn.icms.temple.edu/lammps-ro/trunk@14365 f3b2605a-c512-4ea7-a41b-209d697bcdaa 2015-12-15 15:49:29 +00:00
06370a8d7a git-svn-id: svn://svn.icms.temple.edu/lammps-ro/trunk@14363 f3b2605a-c512-4ea7-a41b-209d697bcdaa 2015-12-11 23:33:38 +00:00
b7b610f3b9 git-svn-id: svn://svn.icms.temple.edu/lammps-ro/trunk@14362 f3b2605a-c512-4ea7-a41b-209d697bcdaa 2015-12-11 23:33:06 +00:00
4e835935c6 git-svn-id: svn://svn.icms.temple.edu/lammps-ro/trunk@14361 f3b2605a-c512-4ea7-a41b-209d697bcdaa 2015-12-11 23:32:49 +00:00
aecaf0bf52 git-svn-id: svn://svn.icms.temple.edu/lammps-ro/trunk@14360 f3b2605a-c512-4ea7-a41b-209d697bcdaa 2015-12-11 23:29:44 +00:00
921bcfb12d Allowed non-full_energy with triclinic
git-svn-id: svn://svn.icms.temple.edu/lammps-ro/trunk@14359 f3b2605a-c512-4ea7-a41b-209d697bcdaa
2015-12-11 23:15:38 +00:00
762f48e528 git-svn-id: svn://svn.icms.temple.edu/lammps-ro/trunk@14358 f3b2605a-c512-4ea7-a41b-209d697bcdaa 2015-12-11 22:20:11 +00:00
3b6a3f4e55 git-svn-id: svn://svn.icms.temple.edu/lammps-ro/trunk@14357 f3b2605a-c512-4ea7-a41b-209d697bcdaa 2015-12-10 22:38:44 +00:00
4c1ad17581 git-svn-id: svn://svn.icms.temple.edu/lammps-ro/trunk@14354 f3b2605a-c512-4ea7-a41b-209d697bcdaa 2015-12-10 17:44:45 +00:00
d94f71a6a8 git-svn-id: svn://svn.icms.temple.edu/lammps-ro/trunk@14352 f3b2605a-c512-4ea7-a41b-209d697bcdaa 2015-12-10 17:42:14 +00:00
2cf6305dc1 git-svn-id: svn://svn.icms.temple.edu/lammps-ro/trunk@14351 f3b2605a-c512-4ea7-a41b-209d697bcdaa 2015-12-10 17:40:52 +00:00
03a9006979 git-svn-id: svn://svn.icms.temple.edu/lammps-ro/trunk@14350 f3b2605a-c512-4ea7-a41b-209d697bcdaa 2015-12-10 17:38:13 +00:00
28bf73fbf7 git-svn-id: svn://svn.icms.temple.edu/lammps-ro/trunk@14349 f3b2605a-c512-4ea7-a41b-209d697bcdaa 2015-12-10 17:23:56 +00:00
8550cacf18 git-svn-id: svn://svn.icms.temple.edu/lammps-ro/trunk@14348 f3b2605a-c512-4ea7-a41b-209d697bcdaa 2015-12-10 17:07:56 +00:00
86a4507b00 git-svn-id: svn://svn.icms.temple.edu/lammps-ro/trunk@14347 f3b2605a-c512-4ea7-a41b-209d697bcdaa 2015-12-10 16:55:37 +00:00
807e00de93 git-svn-id: svn://svn.icms.temple.edu/lammps-ro/trunk@14346 f3b2605a-c512-4ea7-a41b-209d697bcdaa 2015-12-10 16:54:13 +00:00
59e5d36671 Fixing region issue
git-svn-id: svn://svn.icms.temple.edu/lammps-ro/trunk@14343 f3b2605a-c512-4ea7-a41b-209d697bcdaa
2015-12-10 16:04:31 +00:00
760071bb4c git-svn-id: svn://svn.icms.temple.edu/lammps-ro/trunk@14342 f3b2605a-c512-4ea7-a41b-209d697bcdaa 2015-12-09 23:57:59 +00:00
0dcabb7080 git-svn-id: svn://svn.icms.temple.edu/lammps-ro/trunk@14340 f3b2605a-c512-4ea7-a41b-209d697bcdaa 2015-12-09 23:29:22 +00:00
fc1a176d04 git-svn-id: svn://svn.icms.temple.edu/lammps-ro/trunk@14339 f3b2605a-c512-4ea7-a41b-209d697bcdaa 2015-12-09 22:46:13 +00:00
93919c35bf git-svn-id: svn://svn.icms.temple.edu/lammps-ro/trunk@14338 f3b2605a-c512-4ea7-a41b-209d697bcdaa 2015-12-09 22:38:18 +00:00
676cdda5aa git-svn-id: svn://svn.icms.temple.edu/lammps-ro/trunk@14337 f3b2605a-c512-4ea7-a41b-209d697bcdaa 2015-12-09 22:32:13 +00:00
ab01075956 git-svn-id: svn://svn.icms.temple.edu/lammps-ro/trunk@14336 f3b2605a-c512-4ea7-a41b-209d697bcdaa 2015-12-09 22:31:43 +00:00
704f170053 git-svn-id: svn://svn.icms.temple.edu/lammps-ro/trunk@14335 f3b2605a-c512-4ea7-a41b-209d697bcdaa 2015-12-09 22:31:34 +00:00
82d5d73bba git-svn-id: svn://svn.icms.temple.edu/lammps-ro/trunk@14334 f3b2605a-c512-4ea7-a41b-209d697bcdaa 2015-12-09 22:31:24 +00:00
8349024265 git-svn-id: svn://svn.icms.temple.edu/lammps-ro/trunk@14333 f3b2605a-c512-4ea7-a41b-209d697bcdaa 2015-12-09 21:48:11 +00:00
b1dc501880 git-svn-id: svn://svn.icms.temple.edu/lammps-ro/trunk@14332 f3b2605a-c512-4ea7-a41b-209d697bcdaa 2015-12-09 21:47:11 +00:00
493f9c469b git-svn-id: svn://svn.icms.temple.edu/lammps-ro/trunk@14331 f3b2605a-c512-4ea7-a41b-209d697bcdaa 2015-12-09 21:44:39 +00:00
87a47d8b99 git-svn-id: svn://svn.icms.temple.edu/lammps-ro/trunk@14330 f3b2605a-c512-4ea7-a41b-209d697bcdaa 2015-12-09 21:33:03 +00:00
c00aabe736 git-svn-id: svn://svn.icms.temple.edu/lammps-ro/trunk@14329 f3b2605a-c512-4ea7-a41b-209d697bcdaa 2015-12-09 21:24:31 +00:00
900a3aaf9d git-svn-id: svn://svn.icms.temple.edu/lammps-ro/trunk@14328 f3b2605a-c512-4ea7-a41b-209d697bcdaa 2015-12-09 21:10:06 +00:00
6eed163c7b git-svn-id: svn://svn.icms.temple.edu/lammps-ro/trunk@14327 f3b2605a-c512-4ea7-a41b-209d697bcdaa 2015-12-09 21:09:58 +00:00
e000e4a04f git-svn-id: svn://svn.icms.temple.edu/lammps-ro/trunk@14326 f3b2605a-c512-4ea7-a41b-209d697bcdaa 2015-12-09 21:09:50 +00:00
582ba55e4a git-svn-id: svn://svn.icms.temple.edu/lammps-ro/trunk@14325 f3b2605a-c512-4ea7-a41b-209d697bcdaa 2015-12-09 20:28:38 +00:00
7920c13b04 git-svn-id: svn://svn.icms.temple.edu/lammps-ro/trunk@14323 f3b2605a-c512-4ea7-a41b-209d697bcdaa 2015-12-09 18:17:05 +00:00
11fe1e862d git-svn-id: svn://svn.icms.temple.edu/lammps-ro/trunk@14322 f3b2605a-c512-4ea7-a41b-209d697bcdaa 2015-12-09 18:17:02 +00:00
549ea8787b git-svn-id: svn://svn.icms.temple.edu/lammps-ro/trunk@14320 f3b2605a-c512-4ea7-a41b-209d697bcdaa 2015-12-09 17:45:54 +00:00
90100b4eb0 git-svn-id: svn://svn.icms.temple.edu/lammps-ro/trunk@14319 f3b2605a-c512-4ea7-a41b-209d697bcdaa 2015-12-09 17:45:52 +00:00
c2db79391f git-svn-id: svn://svn.icms.temple.edu/lammps-ro/trunk@14318 f3b2605a-c512-4ea7-a41b-209d697bcdaa 2015-12-09 17:45:02 +00:00
21260340e4 git-svn-id: svn://svn.icms.temple.edu/lammps-ro/trunk@14317 f3b2605a-c512-4ea7-a41b-209d697bcdaa 2015-12-09 17:30:15 +00:00
fe6f196439 git-svn-id: svn://svn.icms.temple.edu/lammps-ro/trunk@14316 f3b2605a-c512-4ea7-a41b-209d697bcdaa 2015-12-09 17:29:59 +00:00
d43c4a1bf5 git-svn-id: svn://svn.icms.temple.edu/lammps-ro/trunk@14315 f3b2605a-c512-4ea7-a41b-209d697bcdaa 2015-12-09 17:22:00 +00:00
6f443bf17f git-svn-id: svn://svn.icms.temple.edu/lammps-ro/trunk@14314 f3b2605a-c512-4ea7-a41b-209d697bcdaa 2015-12-09 17:18:22 +00:00
3d659aba57 git-svn-id: svn://svn.icms.temple.edu/lammps-ro/trunk@14313 f3b2605a-c512-4ea7-a41b-209d697bcdaa 2015-12-09 17:16:42 +00:00
bac5a15bfc git-svn-id: svn://svn.icms.temple.edu/lammps-ro/trunk@14312 f3b2605a-c512-4ea7-a41b-209d697bcdaa 2015-12-09 17:16:11 +00:00
f7355d5761 git-svn-id: svn://svn.icms.temple.edu/lammps-ro/trunk@14311 f3b2605a-c512-4ea7-a41b-209d697bcdaa 2015-12-09 17:11:45 +00:00
15c3a77823 git-svn-id: svn://svn.icms.temple.edu/lammps-ro/trunk@14310 f3b2605a-c512-4ea7-a41b-209d697bcdaa 2015-12-09 17:09:56 +00:00
c78f2449b2 git-svn-id: svn://svn.icms.temple.edu/lammps-ro/trunk@14309 f3b2605a-c512-4ea7-a41b-209d697bcdaa 2015-12-09 17:05:37 +00:00
a473a42a9c git-svn-id: svn://svn.icms.temple.edu/lammps-ro/trunk@14308 f3b2605a-c512-4ea7-a41b-209d697bcdaa 2015-12-09 16:59:47 +00:00
1860c2d764 git-svn-id: svn://svn.icms.temple.edu/lammps-ro/trunk@14307 f3b2605a-c512-4ea7-a41b-209d697bcdaa 2015-12-09 15:41:43 +00:00
d24d85da64 git-svn-id: svn://svn.icms.temple.edu/lammps-ro/trunk@14305 f3b2605a-c512-4ea7-a41b-209d697bcdaa 2015-12-09 00:27:17 +00:00
6647 changed files with 916483 additions and 488496 deletions

34
.gitignore vendored Normal file
View File

@ -0,0 +1,34 @@
*~
*.o
*.so
*.cu_o
*.ptx
*_ptx.h
*.a
*.d
*.x
*.exe
*.dll
*.pyc
__pycache__
Obj_*
log.lammps
log.cite
*.bz2
*.gz
*.tar
.*.swp
*.orig
*.rej
.vagrant
\#*#
.#*
.DS_Store
.DS_Store?
._*
.Spotlight-V100
.Trashes
ehthumbs.db
Thumbs.db

View File

@ -0,0 +1,78 @@
LAMMPS (15 Feb 2016)
# FENE beadspring benchmark
units lj
atom_style bond
special_bonds fene
read_data data.chain
orthogonal box = (-16.796 -16.796 -16.796) to (16.796 16.796 16.796)
1 by 1 by 1 MPI processor grid
reading atoms ...
32000 atoms
reading velocities ...
32000 velocities
scanning bonds ...
1 = max bonds/atom
reading bonds ...
31680 bonds
2 = max # of 1-2 neighbors
2 = max # of special neighbors
neighbor 0.4 bin
neigh_modify every 1 delay 1
bond_style fene
bond_coeff 1 30.0 1.5 1.0 1.0
pair_style lj/cut 1.12
pair_modify shift yes
pair_coeff 1 1 1.0 1.0 1.12
fix 1 all nve
fix 2 all langevin 1.0 1.0 10.0 904297
thermo 100
timestep 0.012
run 100
Neighbor list info ...
1 neighbor list requests
update every 1 steps, delay 1 steps, check yes
max neighbors/atom: 2000, page size: 100000
master list distance cutoff = 1.52
ghost atom cutoff = 1.52
binsize = 0.76 -> bins = 45 45 45
Memory usage per processor = 11.5189 Mbytes
Step Temp E_pair E_mol TotEng Press
0 0.97029772 0.44484087 20.494523 22.394765 4.6721833
100 0.9729966 0.4361122 20.507698 22.40326 4.6548819
Loop time of 0.978585 on 1 procs for 100 steps with 32000 atoms
Performance: 105948.895 tau/day, 102.188 timesteps/s
100.0% CPU use with 1 MPI tasks x no OpenMP threads
MPI task timing breakdown:
Section | min time | avg time | max time |%varavg| %total
---------------------------------------------------------------
Pair | 0.19562 | 0.19562 | 0.19562 | 0.0 | 19.99
Bond | 0.087475 | 0.087475 | 0.087475 | 0.0 | 8.94
Neigh | 0.44861 | 0.44861 | 0.44861 | 0.0 | 45.84
Comm | 0.032932 | 0.032932 | 0.032932 | 0.0 | 3.37
Output | 0.00010395 | 0.00010395 | 0.00010395 | 0.0 | 0.01
Modify | 0.19413 | 0.19413 | 0.19413 | 0.0 | 19.84
Other | | 0.01972 | | | 2.02
Nlocal: 32000 ave 32000 max 32000 min
Histogram: 1 0 0 0 0 0 0 0 0 0
Nghost: 9493 ave 9493 max 9493 min
Histogram: 1 0 0 0 0 0 0 0 0 0
Neighs: 155873 ave 155873 max 155873 min
Histogram: 1 0 0 0 0 0 0 0 0 0
Total # of neighbors = 155873
Ave neighs/atom = 4.87103
Ave special neighs/atom = 1.98
Neighbor list builds = 25
Dangerous builds = 0
Total wall time: 0:00:01

View File

@ -0,0 +1,78 @@
LAMMPS (15 Feb 2016)
# FENE beadspring benchmark
units lj
atom_style bond
special_bonds fene
read_data data.chain
orthogonal box = (-16.796 -16.796 -16.796) to (16.796 16.796 16.796)
1 by 2 by 2 MPI processor grid
reading atoms ...
32000 atoms
reading velocities ...
32000 velocities
scanning bonds ...
1 = max bonds/atom
reading bonds ...
31680 bonds
2 = max # of 1-2 neighbors
2 = max # of special neighbors
neighbor 0.4 bin
neigh_modify every 1 delay 1
bond_style fene
bond_coeff 1 30.0 1.5 1.0 1.0
pair_style lj/cut 1.12
pair_modify shift yes
pair_coeff 1 1 1.0 1.0 1.12
fix 1 all nve
fix 2 all langevin 1.0 1.0 10.0 904297
thermo 100
timestep 0.012
run 100
Neighbor list info ...
1 neighbor list requests
update every 1 steps, delay 1 steps, check yes
max neighbors/atom: 2000, page size: 100000
master list distance cutoff = 1.52
ghost atom cutoff = 1.52
binsize = 0.76 -> bins = 45 45 45
Memory usage per processor = 3.91518 Mbytes
Step Temp E_pair E_mol TotEng Press
0 0.97029772 0.44484087 20.494523 22.394765 4.6721833
100 0.97145835 0.43803883 20.502691 22.397872 4.626988
Loop time of 0.271187 on 4 procs for 100 steps with 32000 atoms
Performance: 382319.453 tau/day, 368.749 timesteps/s
99.6% CPU use with 4 MPI tasks x no OpenMP threads
MPI task timing breakdown:
Section | min time | avg time | max time |%varavg| %total
---------------------------------------------------------------
Pair | 0.048621 | 0.050076 | 0.051229 | 0.4 | 18.47
Bond | 0.022254 | 0.022942 | 0.023567 | 0.3 | 8.46
Neigh | 0.11873 | 0.11881 | 0.11887 | 0.0 | 43.81
Comm | 0.019066 | 0.021357 | 0.024297 | 1.3 | 7.88
Output | 5.0068e-05 | 5.5015e-05 | 6.1035e-05 | 0.1 | 0.02
Modify | 0.048737 | 0.050198 | 0.051231 | 0.4 | 18.51
Other | | 0.007751 | | | 2.86
Nlocal: 8000 ave 8030 max 7974 min
Histogram: 1 0 0 1 0 1 0 0 0 1
Nghost: 4177 ave 4191 max 4160 min
Histogram: 1 0 0 0 1 0 0 1 0 1
Neighs: 38995.8 ave 39169 max 38852 min
Histogram: 1 0 0 1 1 0 0 0 0 1
Total # of neighbors = 155983
Ave neighs/atom = 4.87447
Ave special neighs/atom = 1.98
Neighbor list builds = 25
Dangerous builds = 0
Total wall time: 0:00:00

View File

@ -0,0 +1,94 @@
LAMMPS (15 Feb 2016)
# FENE beadspring benchmark
variable x index 1
variable y index 1
variable z index 1
units lj
atom_style bond
atom_modify map hash
special_bonds fene
read_data data.chain
orthogonal box = (-16.796 -16.796 -16.796) to (16.796 16.796 16.796)
1 by 2 by 2 MPI processor grid
reading atoms ...
32000 atoms
reading velocities ...
32000 velocities
scanning bonds ...
1 = max bonds/atom
reading bonds ...
31680 bonds
2 = max # of 1-2 neighbors
2 = max # of special neighbors
replicate $x $y $z
replicate 2 $y $z
replicate 2 2 $z
replicate 2 2 1
orthogonal box = (-16.796 -16.796 -16.796) to (50.388 50.388 16.796)
2 by 2 by 1 MPI processor grid
128000 atoms
126720 bonds
2 = max # of 1-2 neighbors
2 = max # of special neighbors
neighbor 0.4 bin
neigh_modify every 1 delay 1
bond_style fene
bond_coeff 1 30.0 1.5 1.0 1.0
pair_style lj/cut 1.12
pair_modify shift yes
pair_coeff 1 1 1.0 1.0 1.12
fix 1 all nve
fix 2 all langevin 1.0 1.0 10.0 904297
thermo 100
timestep 0.012
run 100
Neighbor list info ...
1 neighbor list requests
update every 1 steps, delay 1 steps, check yes
max neighbors/atom: 2000, page size: 100000
master list distance cutoff = 1.52
ghost atom cutoff = 1.52
binsize = 0.76 -> bins = 89 89 45
Memory usage per processor = 12.8735 Mbytes
Step Temp E_pair E_mol TotEng Press
0 0.97027498 0.44484087 20.494523 22.394765 4.6721833
100 0.97682955 0.44239968 20.500229 22.407862 4.6527025
Loop time of 1.20889 on 4 procs for 100 steps with 128000 atoms
Performance: 85764.410 tau/day, 82.720 timesteps/s
99.8% CPU use with 4 MPI tasks x no OpenMP threads
MPI task timing breakdown:
Section | min time | avg time | max time |%varavg| %total
---------------------------------------------------------------
Pair | 0.21738 | 0.23306 | 0.23926 | 1.9 | 19.28
Bond | 0.094536 | 0.10196 | 0.10534 | 1.4 | 8.43
Neigh | 0.52311 | 0.52392 | 0.52519 | 0.1 | 43.34
Comm | 0.090161 | 0.10022 | 0.12557 | 4.7 | 8.29
Output | 0.00012207 | 0.00017327 | 0.00019598 | 0.2 | 0.01
Modify | 0.19662 | 0.20262 | 0.20672 | 0.8 | 16.76
Other | | 0.04694 | | | 3.88
Nlocal: 32000 ave 32015 max 31983 min
Histogram: 1 0 1 0 0 0 0 0 1 1
Nghost: 9492 ave 9522 max 9432 min
Histogram: 1 0 0 0 0 0 1 0 0 2
Neighs: 155837 ave 156079 max 155506 min
Histogram: 1 0 0 0 0 1 0 0 1 1
Total # of neighbors = 623349
Ave neighs/atom = 4.86991
Ave special neighs/atom = 1.98
Neighbor list builds = 25
Dangerous builds = 0
Total wall time: 0:00:01

View File

@ -0,0 +1,80 @@
LAMMPS (15 Feb 2016)
# LAMMPS benchmark of granular flow
# chute flow of 32000 atoms with frozen base at 26 degrees
units lj
atom_style sphere
boundary p p fs
newton off
comm_modify vel yes
read_data data.chute
orthogonal box = (0 0 0) to (40 20 37.2886)
1 by 1 by 1 MPI processor grid
reading atoms ...
32000 atoms
reading velocities ...
32000 velocities
pair_style gran/hooke/history 200000.0 NULL 50.0 NULL 0.5 0
pair_coeff * *
neighbor 0.1 bin
neigh_modify every 1 delay 0
timestep 0.0001
group bottom type 2
912 atoms in group bottom
group active subtract all bottom
31088 atoms in group active
neigh_modify exclude group bottom bottom
fix 1 all gravity 1.0 chute 26.0
fix 2 bottom freeze
fix 3 active nve/sphere
compute 1 all erotate/sphere
thermo_style custom step atoms ke c_1 vol
thermo_modify norm no
thermo 100
run 100
Neighbor list info ...
2 neighbor list requests
update every 1 steps, delay 0 steps, check yes
max neighbors/atom: 2000, page size: 100000
master list distance cutoff = 1.1
ghost atom cutoff = 1.1
binsize = 0.55 -> bins = 73 37 68
Memory usage per processor = 15.567 Mbytes
Step Atoms KinEng 1 Volume
0 32000 784139.13 1601.1263 29833.783
100 32000 784292.08 1571.0968 29834.707
Loop time of 0.550482 on 1 procs for 100 steps with 32000 atoms
Performance: 1569.534 tau/day, 181.659 timesteps/s
100.1% CPU use with 1 MPI tasks x no OpenMP threads
MPI task timing breakdown:
Section | min time | avg time | max time |%varavg| %total
---------------------------------------------------------------
Pair | 0.33849 | 0.33849 | 0.33849 | 0.0 | 61.49
Neigh | 0.040353 | 0.040353 | 0.040353 | 0.0 | 7.33
Comm | 0.018023 | 0.018023 | 0.018023 | 0.0 | 3.27
Output | 0.00020385 | 0.00020385 | 0.00020385 | 0.0 | 0.04
Modify | 0.13155 | 0.13155 | 0.13155 | 0.0 | 23.90
Other | | 0.02186 | | | 3.97
Nlocal: 32000 ave 32000 max 32000 min
Histogram: 1 0 0 0 0 0 0 0 0 0
Nghost: 5463 ave 5463 max 5463 min
Histogram: 1 0 0 0 0 0 0 0 0 0
Neighs: 115133 ave 115133 max 115133 min
Histogram: 1 0 0 0 0 0 0 0 0 0
Total # of neighbors = 115133
Ave neighs/atom = 3.59791
Neighbor list builds = 2
Dangerous builds = 0
Total wall time: 0:00:00

View File

@ -0,0 +1,80 @@
LAMMPS (15 Feb 2016)
# LAMMPS benchmark of granular flow
# chute flow of 32000 atoms with frozen base at 26 degrees
units lj
atom_style sphere
boundary p p fs
newton off
comm_modify vel yes
read_data data.chute
orthogonal box = (0 0 0) to (40 20 37.2886)
2 by 1 by 2 MPI processor grid
reading atoms ...
32000 atoms
reading velocities ...
32000 velocities
pair_style gran/hooke/history 200000.0 NULL 50.0 NULL 0.5 0
pair_coeff * *
neighbor 0.1 bin
neigh_modify every 1 delay 0
timestep 0.0001
group bottom type 2
912 atoms in group bottom
group active subtract all bottom
31088 atoms in group active
neigh_modify exclude group bottom bottom
fix 1 all gravity 1.0 chute 26.0
fix 2 bottom freeze
fix 3 active nve/sphere
compute 1 all erotate/sphere
thermo_style custom step atoms ke c_1 vol
thermo_modify norm no
thermo 100
run 100
Neighbor list info ...
2 neighbor list requests
update every 1 steps, delay 0 steps, check yes
max neighbors/atom: 2000, page size: 100000
master list distance cutoff = 1.1
ghost atom cutoff = 1.1
binsize = 0.55 -> bins = 73 37 68
Memory usage per processor = 6.81783 Mbytes
Step Atoms KinEng 1 Volume
0 32000 784139.13 1601.1263 29833.783
100 32000 784292.08 1571.0968 29834.707
Loop time of 0.13141 on 4 procs for 100 steps with 32000 atoms
Performance: 6574.833 tau/day, 760.976 timesteps/s
99.3% CPU use with 4 MPI tasks x no OpenMP threads
MPI task timing breakdown:
Section | min time | avg time | max time |%varavg| %total
---------------------------------------------------------------
Pair | 0.062505 | 0.067 | 0.07152 | 1.5 | 50.99
Neigh | 0.010041 | 0.0101 | 0.010178 | 0.1 | 7.69
Comm | 0.012347 | 0.012895 | 0.013444 | 0.5 | 9.81
Output | 6.3896e-05 | 0.00010294 | 0.00014091 | 0.3 | 0.08
Modify | 0.031802 | 0.032348 | 0.032897 | 0.3 | 24.62
Other | | 0.008965 | | | 6.82
Nlocal: 8000 ave 8008 max 7992 min
Histogram: 2 0 0 0 0 0 0 0 0 2
Nghost: 2439 ave 2450 max 2428 min
Histogram: 2 0 0 0 0 0 0 0 0 2
Neighs: 29500.5 ave 30488 max 28513 min
Histogram: 2 0 0 0 0 0 0 0 0 2
Total # of neighbors = 118002
Ave neighs/atom = 3.68756
Neighbor list builds = 2
Dangerous builds = 0
Total wall time: 0:00:00

View File

@ -0,0 +1,90 @@
LAMMPS (15 Feb 2016)
# LAMMPS benchmark of granular flow
# chute flow of 32000 atoms with frozen base at 26 degrees
variable x index 1
variable y index 1
units lj
atom_style sphere
boundary p p fs
newton off
comm_modify vel yes
read_data data.chute
orthogonal box = (0 0 0) to (40 20 37.2886)
2 by 1 by 2 MPI processor grid
reading atoms ...
32000 atoms
reading velocities ...
32000 velocities
replicate $x $y 1
replicate 2 $y 1
replicate 2 2 1
orthogonal box = (0 0 0) to (80 40 37.2922)
2 by 2 by 1 MPI processor grid
128000 atoms
pair_style gran/hooke/history 200000.0 NULL 50.0 NULL 0.5 0
pair_coeff * *
neighbor 0.1 bin
neigh_modify every 1 delay 0
timestep 0.0001
group bottom type 2
3648 atoms in group bottom
group active subtract all bottom
124352 atoms in group active
neigh_modify exclude group bottom bottom
fix 1 all gravity 1.0 chute 26.0
fix 2 bottom freeze
fix 3 active nve/sphere
compute 1 all erotate/sphere
thermo_style custom step atoms ke c_1 vol
thermo_modify norm no
thermo 100
run 100
Neighbor list info ...
2 neighbor list requests
update every 1 steps, delay 0 steps, check yes
max neighbors/atom: 2000, page size: 100000
master list distance cutoff = 1.1
ghost atom cutoff = 1.1
binsize = 0.55 -> bins = 146 73 68
Memory usage per processor = 15.7007 Mbytes
Step Atoms KinEng 1 Volume
0 128000 3136556.5 6404.5051 119335.13
100 128000 3137168.3 6284.3873 119338.83
Loop time of 0.906913 on 4 procs for 100 steps with 128000 atoms
Performance: 952.683 tau/day, 110.264 timesteps/s
99.7% CPU use with 4 MPI tasks x no OpenMP threads
MPI task timing breakdown:
Section | min time | avg time | max time |%varavg| %total
---------------------------------------------------------------
Pair | 0.51454 | 0.53094 | 0.55381 | 2.0 | 58.54
Neigh | 0.042597 | 0.043726 | 0.045801 | 0.6 | 4.82
Comm | 0.063027 | 0.064657 | 0.067367 | 0.7 | 7.13
Output | 0.00024891 | 0.00059718 | 0.00086498 | 1.0 | 0.07
Modify | 0.16508 | 0.17656 | 0.1925 | 2.6 | 19.47
Other | | 0.09043 | | | 9.97
Nlocal: 32000 ave 32000 max 32000 min
Histogram: 4 0 0 0 0 0 0 0 0 0
Nghost: 5463 ave 5463 max 5463 min
Histogram: 4 0 0 0 0 0 0 0 0 0
Neighs: 115133 ave 115133 max 115133 min
Histogram: 4 0 0 0 0 0 0 0 0 0
Total # of neighbors = 460532
Ave neighs/atom = 3.59791
Neighbor list builds = 2
Dangerous builds = 0
Total wall time: 0:00:01

View File

@ -0,0 +1,83 @@
LAMMPS (15 Feb 2016)
# bulk Cu lattice
variable x index 1
variable y index 1
variable z index 1
variable xx equal 20*$x
variable xx equal 20*1
variable yy equal 20*$y
variable yy equal 20*1
variable zz equal 20*$z
variable zz equal 20*1
units metal
atom_style atomic
lattice fcc 3.615
Lattice spacing in x,y,z = 3.615 3.615 3.615
region box block 0 ${xx} 0 ${yy} 0 ${zz}
region box block 0 20 0 ${yy} 0 ${zz}
region box block 0 20 0 20 0 ${zz}
region box block 0 20 0 20 0 20
create_box 1 box
Created orthogonal box = (0 0 0) to (72.3 72.3 72.3)
1 by 1 by 1 MPI processor grid
create_atoms 1 box
Created 32000 atoms
pair_style eam
pair_coeff 1 1 Cu_u3.eam
Reading potential file Cu_u3.eam with DATE: 2007-06-11
velocity all create 1600.0 376847 loop geom
neighbor 1.0 bin
neigh_modify every 1 delay 5 check yes
fix 1 all nve
timestep 0.005
thermo 50
run 100
Neighbor list info ...
1 neighbor list requests
update every 1 steps, delay 5 steps, check yes
max neighbors/atom: 2000, page size: 100000
master list distance cutoff = 5.95
ghost atom cutoff = 5.95
binsize = 2.975 -> bins = 25 25 25
Memory usage per processor = 10.2238 Mbytes
Step Temp E_pair E_mol TotEng Press
0 1600 -113280 0 -106662.09 18703.573
50 781.69049 -109873.35 0 -106640.13 52273.088
100 801.832 -109957.3 0 -106640.77 51322.821
Loop time of 5.90097 on 1 procs for 100 steps with 32000 atoms
Performance: 7.321 ns/day, 3.278 hours/ns, 16.946 timesteps/s
99.9% CPU use with 1 MPI tasks x no OpenMP threads
MPI task timing breakdown:
Section | min time | avg time | max time |%varavg| %total
---------------------------------------------------------------
Pair | 5.2121 | 5.2121 | 5.2121 | 0.0 | 88.33
Neigh | 0.58212 | 0.58212 | 0.58212 | 0.0 | 9.86
Comm | 0.030392 | 0.030392 | 0.030392 | 0.0 | 0.52
Output | 0.00023389 | 0.00023389 | 0.00023389 | 0.0 | 0.00
Modify | 0.060871 | 0.060871 | 0.060871 | 0.0 | 1.03
Other | | 0.01527 | | | 0.26
Nlocal: 32000 ave 32000 max 32000 min
Histogram: 1 0 0 0 0 0 0 0 0 0
Nghost: 19909 ave 19909 max 19909 min
Histogram: 1 0 0 0 0 0 0 0 0 0
Neighs: 1.20778e+06 ave 1.20778e+06 max 1.20778e+06 min
Histogram: 1 0 0 0 0 0 0 0 0 0
Total # of neighbors = 1207784
Ave neighs/atom = 37.7433
Neighbor list builds = 13
Dangerous builds = 0
Total wall time: 0:00:06

View File

@ -0,0 +1,83 @@
LAMMPS (15 Feb 2016)
# bulk Cu lattice
variable x index 1
variable y index 1
variable z index 1
variable xx equal 20*$x
variable xx equal 20*1
variable yy equal 20*$y
variable yy equal 20*1
variable zz equal 20*$z
variable zz equal 20*1
units metal
atom_style atomic
lattice fcc 3.615
Lattice spacing in x,y,z = 3.615 3.615 3.615
region box block 0 ${xx} 0 ${yy} 0 ${zz}
region box block 0 20 0 ${yy} 0 ${zz}
region box block 0 20 0 20 0 ${zz}
region box block 0 20 0 20 0 20
create_box 1 box
Created orthogonal box = (0 0 0) to (72.3 72.3 72.3)
1 by 2 by 2 MPI processor grid
create_atoms 1 box
Created 32000 atoms
pair_style eam
pair_coeff 1 1 Cu_u3.eam
Reading potential file Cu_u3.eam with DATE: 2007-06-11
velocity all create 1600.0 376847 loop geom
neighbor 1.0 bin
neigh_modify every 1 delay 5 check yes
fix 1 all nve
timestep 0.005
thermo 50
run 100
Neighbor list info ...
1 neighbor list requests
update every 1 steps, delay 5 steps, check yes
max neighbors/atom: 2000, page size: 100000
master list distance cutoff = 5.95
ghost atom cutoff = 5.95
binsize = 2.975 -> bins = 25 25 25
Memory usage per processor = 5.09629 Mbytes
Step Temp E_pair E_mol TotEng Press
0 1600 -113280 0 -106662.09 18703.573
50 781.69049 -109873.35 0 -106640.13 52273.088
100 801.832 -109957.3 0 -106640.77 51322.821
Loop time of 1.58019 on 4 procs for 100 steps with 32000 atoms
Performance: 27.338 ns/day, 0.878 hours/ns, 63.284 timesteps/s
99.8% CPU use with 4 MPI tasks x no OpenMP threads
MPI task timing breakdown:
Section | min time | avg time | max time |%varavg| %total
---------------------------------------------------------------
Pair | 1.3617 | 1.366 | 1.3723 | 0.4 | 86.45
Neigh | 0.15123 | 0.15232 | 0.15374 | 0.2 | 9.64
Comm | 0.033429 | 0.041275 | 0.047066 | 2.7 | 2.61
Output | 0.00011301 | 0.0001573 | 0.000211 | 0.3 | 0.01
Modify | 0.014694 | 0.015085 | 0.015421 | 0.2 | 0.95
Other | | 0.005342 | | | 0.34
Nlocal: 8000 ave 8008 max 7993 min
Histogram: 2 0 0 0 0 0 0 0 1 1
Nghost: 9130.25 ave 9138 max 9122 min
Histogram: 2 0 0 0 0 0 0 0 0 2
Neighs: 301946 ave 302392 max 301360 min
Histogram: 1 0 0 0 1 0 0 0 1 1
Total # of neighbors = 1207784
Ave neighs/atom = 37.7433
Neighbor list builds = 13
Dangerous builds = 0
Total wall time: 0:00:01

View File

@ -0,0 +1,83 @@
LAMMPS (15 Feb 2016)
# bulk Cu lattice
variable x index 1
variable y index 1
variable z index 1
variable xx equal 20*$x
variable xx equal 20*2
variable yy equal 20*$y
variable yy equal 20*2
variable zz equal 20*$z
variable zz equal 20*1
units metal
atom_style atomic
lattice fcc 3.615
Lattice spacing in x,y,z = 3.615 3.615 3.615
region box block 0 ${xx} 0 ${yy} 0 ${zz}
region box block 0 40 0 ${yy} 0 ${zz}
region box block 0 40 0 40 0 ${zz}
region box block 0 40 0 40 0 20
create_box 1 box
Created orthogonal box = (0 0 0) to (144.6 144.6 72.3)
2 by 2 by 1 MPI processor grid
create_atoms 1 box
Created 128000 atoms
pair_style eam
pair_coeff 1 1 Cu_u3.eam
Reading potential file Cu_u3.eam with DATE: 2007-06-11
velocity all create 1600.0 376847 loop geom
neighbor 1.0 bin
neigh_modify every 1 delay 5 check yes
fix 1 all nve
timestep 0.005
thermo 50
run 100
Neighbor list info ...
1 neighbor list requests
update every 1 steps, delay 5 steps, check yes
max neighbors/atom: 2000, page size: 100000
master list distance cutoff = 5.95
ghost atom cutoff = 5.95
binsize = 2.975 -> bins = 49 49 25
Memory usage per processor = 10.1402 Mbytes
Step Temp E_pair E_mol TotEng Press
0 1600 -453120 0 -426647.73 18704.012
50 779.50001 -439457.02 0 -426560.06 52355.276
100 797.97828 -439764.76 0 -426562.07 51474.74
Loop time of 6.46849 on 4 procs for 100 steps with 128000 atoms
Performance: 6.679 ns/day, 3.594 hours/ns, 15.460 timesteps/s
99.9% CPU use with 4 MPI tasks x no OpenMP threads
MPI task timing breakdown:
Section | min time | avg time | max time |%varavg| %total
---------------------------------------------------------------
Pair | 5.581 | 5.5997 | 5.6265 | 0.8 | 86.57
Neigh | 0.65287 | 0.658 | 0.66374 | 0.5 | 10.17
Comm | 0.075706 | 0.11015 | 0.13655 | 7.2 | 1.70
Output | 0.00026488 | 0.00028312 | 0.00029302 | 0.1 | 0.00
Modify | 0.069607 | 0.072407 | 0.074555 | 0.7 | 1.12
Other | | 0.02794 | | | 0.43
Nlocal: 32000 ave 32092 max 31914 min
Histogram: 1 0 0 1 0 1 0 0 0 1
Nghost: 19910 ave 19997 max 19818 min
Histogram: 1 0 0 0 1 0 1 0 0 1
Neighs: 1.20728e+06 ave 1.21142e+06 max 1.2036e+06 min
Histogram: 1 0 0 1 1 0 0 0 0 1
Total # of neighbors = 4829126
Ave neighs/atom = 37.7275
Neighbor list builds = 14
Dangerous builds = 0
Total wall time: 0:00:06

View File

@ -0,0 +1,79 @@
LAMMPS (15 Feb 2016)
# 3d Lennard-Jones melt
variable x index 1
variable y index 1
variable z index 1
variable xx equal 20*$x
variable xx equal 20*1
variable yy equal 20*$y
variable yy equal 20*1
variable zz equal 20*$z
variable zz equal 20*1
units lj
atom_style atomic
lattice fcc 0.8442
Lattice spacing in x,y,z = 1.6796 1.6796 1.6796
region box block 0 ${xx} 0 ${yy} 0 ${zz}
region box block 0 20 0 ${yy} 0 ${zz}
region box block 0 20 0 20 0 ${zz}
region box block 0 20 0 20 0 20
create_box 1 box
Created orthogonal box = (0 0 0) to (33.5919 33.5919 33.5919)
1 by 1 by 1 MPI processor grid
create_atoms 1 box
Created 32000 atoms
mass 1 1.0
velocity all create 1.44 87287 loop geom
pair_style lj/cut 2.5
pair_coeff 1 1 1.0 1.0 2.5
neighbor 0.3 bin
neigh_modify delay 0 every 20 check no
fix 1 all nve
run 100
Neighbor list info ...
1 neighbor list requests
update every 20 steps, delay 0 steps, check no
max neighbors/atom: 2000, page size: 100000
master list distance cutoff = 2.8
ghost atom cutoff = 2.8
binsize = 1.4 -> bins = 24 24 24
Memory usage per processor = 8.21387 Mbytes
Step Temp E_pair E_mol TotEng Press
0 1.44 -6.7733681 0 -4.6134356 -5.0197073
100 0.7574531 -5.7585055 0 -4.6223613 0.20726105
Loop time of 2.26309 on 1 procs for 100 steps with 32000 atoms
Performance: 19088.920 tau/day, 44.187 timesteps/s
99.9% CPU use with 1 MPI tasks x no OpenMP threads
MPI task timing breakdown:
Section | min time | avg time | max time |%varavg| %total
---------------------------------------------------------------
Pair | 1.9341 | 1.9341 | 1.9341 | 0.0 | 85.46
Neigh | 0.2442 | 0.2442 | 0.2442 | 0.0 | 10.79
Comm | 0.024158 | 0.024158 | 0.024158 | 0.0 | 1.07
Output | 0.00011611 | 0.00011611 | 0.00011611 | 0.0 | 0.01
Modify | 0.053222 | 0.053222 | 0.053222 | 0.0 | 2.35
Other | | 0.007258 | | | 0.32
Nlocal: 32000 ave 32000 max 32000 min
Histogram: 1 0 0 0 0 0 0 0 0 0
Nghost: 19657 ave 19657 max 19657 min
Histogram: 1 0 0 0 0 0 0 0 0 0
Neighs: 1.20283e+06 ave 1.20283e+06 max 1.20283e+06 min
Histogram: 1 0 0 0 0 0 0 0 0 0
Total # of neighbors = 1202833
Ave neighs/atom = 37.5885
Neighbor list builds = 5
Dangerous builds not checked
Total wall time: 0:00:02

View File

@ -0,0 +1,79 @@
LAMMPS (15 Feb 2016)
# 3d Lennard-Jones melt
variable x index 1
variable y index 1
variable z index 1
variable xx equal 20*$x
variable xx equal 20*1
variable yy equal 20*$y
variable yy equal 20*1
variable zz equal 20*$z
variable zz equal 20*1
units lj
atom_style atomic
lattice fcc 0.8442
Lattice spacing in x,y,z = 1.6796 1.6796 1.6796
region box block 0 ${xx} 0 ${yy} 0 ${zz}
region box block 0 20 0 ${yy} 0 ${zz}
region box block 0 20 0 20 0 ${zz}
region box block 0 20 0 20 0 20
create_box 1 box
Created orthogonal box = (0 0 0) to (33.5919 33.5919 33.5919)
1 by 2 by 2 MPI processor grid
create_atoms 1 box
Created 32000 atoms
mass 1 1.0
velocity all create 1.44 87287 loop geom
pair_style lj/cut 2.5
pair_coeff 1 1 1.0 1.0 2.5
neighbor 0.3 bin
neigh_modify delay 0 every 20 check no
fix 1 all nve
run 100
Neighbor list info ...
1 neighbor list requests
update every 20 steps, delay 0 steps, check no
max neighbors/atom: 2000, page size: 100000
master list distance cutoff = 2.8
ghost atom cutoff = 2.8
binsize = 1.4 -> bins = 24 24 24
Memory usage per processor = 4.09506 Mbytes
Step Temp E_pair E_mol TotEng Press
0 1.44 -6.7733681 0 -4.6134356 -5.0197073
100 0.7574531 -5.7585055 0 -4.6223613 0.20726105
Loop time of 0.640733 on 4 procs for 100 steps with 32000 atoms
Performance: 67422.779 tau/day, 156.071 timesteps/s
99.7% CPU use with 4 MPI tasks x no OpenMP threads
MPI task timing breakdown:
Section | min time | avg time | max time |%varavg| %total
---------------------------------------------------------------
Pair | 0.49487 | 0.51733 | 0.5322 | 1.9 | 80.74
Neigh | 0.061131 | 0.063685 | 0.065433 | 0.6 | 9.94
Comm | 0.02457 | 0.042349 | 0.069598 | 8.1 | 6.61
Output | 5.9843e-05 | 6.3181e-05 | 6.6996e-05 | 0.0 | 0.01
Modify | 0.012961 | 0.013863 | 0.014491 | 0.5 | 2.16
Other | | 0.003448 | | | 0.54
Nlocal: 8000 ave 8037 max 7964 min
Histogram: 2 0 0 0 0 0 0 0 1 1
Nghost: 9007.5 ave 9050 max 8968 min
Histogram: 1 1 0 0 0 0 0 1 0 1
Neighs: 300708 ave 305113 max 297203 min
Histogram: 1 0 0 1 1 0 0 0 0 1
Total # of neighbors = 1202833
Ave neighs/atom = 37.5885
Neighbor list builds = 5
Dangerous builds not checked
Total wall time: 0:00:00

View File

@ -0,0 +1,79 @@
LAMMPS (15 Feb 2016)
# 3d Lennard-Jones melt
variable x index 1
variable y index 1
variable z index 1
variable xx equal 20*$x
variable xx equal 20*2
variable yy equal 20*$y
variable yy equal 20*2
variable zz equal 20*$z
variable zz equal 20*1
units lj
atom_style atomic
lattice fcc 0.8442
Lattice spacing in x,y,z = 1.6796 1.6796 1.6796
region box block 0 ${xx} 0 ${yy} 0 ${zz}
region box block 0 40 0 ${yy} 0 ${zz}
region box block 0 40 0 40 0 ${zz}
region box block 0 40 0 40 0 20
create_box 1 box
Created orthogonal box = (0 0 0) to (67.1838 67.1838 33.5919)
2 by 2 by 1 MPI processor grid
create_atoms 1 box
Created 128000 atoms
mass 1 1.0
velocity all create 1.44 87287 loop geom
pair_style lj/cut 2.5
pair_coeff 1 1 1.0 1.0 2.5
neighbor 0.3 bin
neigh_modify delay 0 every 20 check no
fix 1 all nve
run 100
Neighbor list info ...
1 neighbor list requests
update every 20 steps, delay 0 steps, check no
max neighbors/atom: 2000, page size: 100000
master list distance cutoff = 2.8
ghost atom cutoff = 2.8
binsize = 1.4 -> bins = 48 48 24
Memory usage per processor = 8.13678 Mbytes
Step Temp E_pair E_mol TotEng Press
0 1.44 -6.7733681 0 -4.6133849 -5.0196788
100 0.75841891 -5.759957 0 -4.6223375 0.20008866
Loop time of 2.57914 on 4 procs for 100 steps with 128000 atoms
Performance: 16749.768 tau/day, 38.773 timesteps/s
99.8% CPU use with 4 MPI tasks x no OpenMP threads
MPI task timing breakdown:
Section | min time | avg time | max time |%varavg| %total
---------------------------------------------------------------
Pair | 2.042 | 2.1092 | 2.1668 | 3.1 | 81.78
Neigh | 0.23982 | 0.24551 | 0.25233 | 1.0 | 9.52
Comm | 0.067088 | 0.13887 | 0.22681 | 15.7 | 5.38
Output | 0.00013185 | 0.00021666 | 0.00027108 | 0.4 | 0.01
Modify | 0.060348 | 0.071269 | 0.077063 | 2.5 | 2.76
Other | | 0.01403 | | | 0.54
Nlocal: 32000 ave 32060 max 31939 min
Histogram: 1 0 1 0 0 0 0 1 0 1
Nghost: 19630.8 ave 19681 max 19562 min
Histogram: 1 0 0 0 1 0 0 0 1 1
Neighs: 1.20195e+06 ave 1.20354e+06 max 1.19931e+06 min
Histogram: 1 0 0 0 0 0 0 2 0 1
Total # of neighbors = 4807797
Ave neighs/atom = 37.5609
Neighbor list builds = 5
Dangerous builds not checked
Total wall time: 0:00:02

View File

@ -0,0 +1,121 @@
LAMMPS (15 Feb 2016)
# Rhodopsin model
units real
neigh_modify delay 5 every 1
atom_style full
bond_style harmonic
angle_style charmm
dihedral_style charmm
improper_style harmonic
pair_style lj/charmm/coul/long 8.0 10.0
pair_modify mix arithmetic
kspace_style pppm 1e-4
read_data data.rhodo
orthogonal box = (-27.5 -38.5 -36.3646) to (27.5 38.5 36.3615)
1 by 1 by 1 MPI processor grid
reading atoms ...
32000 atoms
reading velocities ...
32000 velocities
scanning bonds ...
4 = max bonds/atom
scanning angles ...
8 = max angles/atom
scanning dihedrals ...
18 = max dihedrals/atom
scanning impropers ...
2 = max impropers/atom
reading bonds ...
27723 bonds
reading angles ...
40467 angles
reading dihedrals ...
56829 dihedrals
reading impropers ...
1034 impropers
4 = max # of 1-2 neighbors
12 = max # of 1-3 neighbors
24 = max # of 1-4 neighbors
26 = max # of special neighbors
fix 1 all shake 0.0001 5 0 m 1.0 a 232
1617 = # of size 2 clusters
3633 = # of size 3 clusters
747 = # of size 4 clusters
4233 = # of frozen angles
fix 2 all npt temp 300.0 300.0 100.0 z 0.0 0.0 1000.0 mtk no pchain 0 tchain 1
special_bonds charmm
thermo 50
thermo_style multi
timestep 2.0
run 100
PPPM initialization ...
G vector (1/distance) = 0.248835
grid = 25 32 32
stencil order = 5
estimated absolute RMS force accuracy = 0.0355478
estimated relative force accuracy = 0.000107051
using double precision FFTs
3d grid and FFT values/proc = 41070 25600
Neighbor list info ...
1 neighbor list requests
update every 1 steps, delay 5 steps, check yes
max neighbors/atom: 2000, page size: 100000
master list distance cutoff = 12
ghost atom cutoff = 12
binsize = 6 -> bins = 10 13 13
Memory usage per processor = 91.7487 Mbytes
---------------- Step 0 ----- CPU = 0.0000 (sec) ----------------
TotEng = -25356.2064 KinEng = 21444.8313 Temp = 299.0397
PotEng = -46801.0377 E_bond = 2537.9940 E_angle = 10921.3742
E_dihed = 5211.7865 E_impro = 213.5116 E_vdwl = -2307.8634
E_coul = 207025.8927 E_long = -270403.7333 Press = -142.6035
Volume = 307995.0335
---------------- Step 50 ----- CPU = 17.6362 (sec) ----------------
TotEng = -25330.0828 KinEng = 21501.0029 Temp = 299.8230
PotEng = -46831.0857 E_bond = 2471.7004 E_angle = 10836.4975
E_dihed = 5239.6299 E_impro = 227.1218 E_vdwl = -1993.2754
E_coul = 206797.6331 E_long = -270410.3930 Press = 237.6701
Volume = 308031.5639
---------------- Step 100 ----- CPU = 35.9089 (sec) ----------------
TotEng = -25290.7593 KinEng = 21592.0117 Temp = 301.0920
PotEng = -46882.7709 E_bond = 2567.9807 E_angle = 10781.9408
E_dihed = 5198.7432 E_impro = 216.7834 E_vdwl = -1902.4783
E_coul = 206659.2326 E_long = -270404.9733 Press = 6.9960
Volume = 308133.9888
Loop time of 35.9089 on 1 procs for 100 steps with 32000 atoms
Performance: 0.481 ns/day, 49.874 hours/ns, 2.785 timesteps/s
99.9% CPU use with 1 MPI tasks x no OpenMP threads
MPI task timing breakdown:
Section | min time | avg time | max time |%varavg| %total
---------------------------------------------------------------
Pair | 25.731 | 25.731 | 25.731 | 0.0 | 71.66
Bond | 1.2771 | 1.2771 | 1.2771 | 0.0 | 3.56
Kspace | 3.2094 | 3.2094 | 3.2094 | 0.0 | 8.94
Neigh | 4.4538 | 4.4538 | 4.4538 | 0.0 | 12.40
Comm | 0.068507 | 0.068507 | 0.068507 | 0.0 | 0.19
Output | 0.00025916 | 0.00025916 | 0.00025916 | 0.0 | 0.00
Modify | 1.1417 | 1.1417 | 1.1417 | 0.0 | 3.18
Other | | 0.027 | | | 0.08
Nlocal: 32000 ave 32000 max 32000 min
Histogram: 1 0 0 0 0 0 0 0 0 0
Nghost: 47958 ave 47958 max 47958 min
Histogram: 1 0 0 0 0 0 0 0 0 0
Neighs: 1.20281e+07 ave 1.20281e+07 max 1.20281e+07 min
Histogram: 1 0 0 0 0 0 0 0 0 0
Total # of neighbors = 12028107
Ave neighs/atom = 375.878
Ave special neighs/atom = 7.43187
Neighbor list builds = 11
Dangerous builds = 0
Total wall time: 0:00:37

View File

@ -0,0 +1,121 @@
LAMMPS (15 Feb 2016)
# Rhodopsin model
units real
neigh_modify delay 5 every 1
atom_style full
bond_style harmonic
angle_style charmm
dihedral_style charmm
improper_style harmonic
pair_style lj/charmm/coul/long 8.0 10.0
pair_modify mix arithmetic
kspace_style pppm 1e-4
read_data data.rhodo
orthogonal box = (-27.5 -38.5 -36.3646) to (27.5 38.5 36.3615)
1 by 2 by 2 MPI processor grid
reading atoms ...
32000 atoms
reading velocities ...
32000 velocities
scanning bonds ...
4 = max bonds/atom
scanning angles ...
8 = max angles/atom
scanning dihedrals ...
18 = max dihedrals/atom
scanning impropers ...
2 = max impropers/atom
reading bonds ...
27723 bonds
reading angles ...
40467 angles
reading dihedrals ...
56829 dihedrals
reading impropers ...
1034 impropers
4 = max # of 1-2 neighbors
12 = max # of 1-3 neighbors
24 = max # of 1-4 neighbors
26 = max # of special neighbors
fix 1 all shake 0.0001 5 0 m 1.0 a 232
1617 = # of size 2 clusters
3633 = # of size 3 clusters
747 = # of size 4 clusters
4233 = # of frozen angles
fix 2 all npt temp 300.0 300.0 100.0 z 0.0 0.0 1000.0 mtk no pchain 0 tchain 1
special_bonds charmm
thermo 50
thermo_style multi
timestep 2.0
run 100
PPPM initialization ...
G vector (1/distance) = 0.248835
grid = 25 32 32
stencil order = 5
estimated absolute RMS force accuracy = 0.0355478
estimated relative force accuracy = 0.000107051
using double precision FFTs
3d grid and FFT values/proc = 13230 6400
Neighbor list info ...
1 neighbor list requests
update every 1 steps, delay 5 steps, check yes
max neighbors/atom: 2000, page size: 100000
master list distance cutoff = 12
ghost atom cutoff = 12
binsize = 6 -> bins = 10 13 13
Memory usage per processor = 36.629 Mbytes
---------------- Step 0 ----- CPU = 0.0000 (sec) ----------------
TotEng = -25356.2064 KinEng = 21444.8313 Temp = 299.0397
PotEng = -46801.0377 E_bond = 2537.9940 E_angle = 10921.3742
E_dihed = 5211.7865 E_impro = 213.5116 E_vdwl = -2307.8634
E_coul = 207025.8927 E_long = -270403.7333 Press = -142.6035
Volume = 307995.0335
---------------- Step 50 ----- CPU = 4.7461 (sec) ----------------
TotEng = -25330.0828 KinEng = 21501.0029 Temp = 299.8230
PotEng = -46831.0857 E_bond = 2471.7004 E_angle = 10836.4975
E_dihed = 5239.6299 E_impro = 227.1218 E_vdwl = -1993.2754
E_coul = 206797.6331 E_long = -270410.3930 Press = 237.6701
Volume = 308031.5639
---------------- Step 100 ----- CPU = 9.6332 (sec) ----------------
TotEng = -25290.7591 KinEng = 21592.0117 Temp = 301.0920
PotEng = -46882.7708 E_bond = 2567.9807 E_angle = 10781.9408
E_dihed = 5198.7432 E_impro = 216.7834 E_vdwl = -1902.4783
E_coul = 206659.2327 E_long = -270404.9733 Press = 6.9960
Volume = 308133.9888
Loop time of 9.63322 on 4 procs for 100 steps with 32000 atoms
Performance: 1.794 ns/day, 13.379 hours/ns, 10.381 timesteps/s
99.9% CPU use with 4 MPI tasks x no OpenMP threads
MPI task timing breakdown:
Section | min time | avg time | max time |%varavg| %total
---------------------------------------------------------------
Pair | 6.4364 | 6.5993 | 6.7208 | 4.7 | 68.51
Bond | 0.30755 | 0.32435 | 0.35704 | 3.4 | 3.37
Kspace | 0.92248 | 1.0782 | 1.2597 | 13.0 | 11.19
Neigh | 1.1669 | 1.1672 | 1.1675 | 0.0 | 12.12
Comm | 0.094674 | 0.098065 | 0.10543 | 1.4 | 1.02
Output | 0.00015521 | 0.00016224 | 0.00018215 | 0.1 | 0.00
Modify | 0.32982 | 0.34654 | 0.35365 | 1.6 | 3.60
Other | | 0.01943 | | | 0.20
Nlocal: 8000 ave 8143 max 7933 min
Histogram: 1 2 0 0 0 0 0 0 0 1
Nghost: 22733.5 ave 22769 max 22693 min
Histogram: 1 0 0 0 0 2 0 0 0 1
Neighs: 3.00703e+06 ave 3.0975e+06 max 2.96493e+06 min
Histogram: 1 2 0 0 0 0 0 0 0 1
Total # of neighbors = 12028107
Ave neighs/atom = 375.878
Ave special neighs/atom = 7.43187
Neighbor list builds = 11
Dangerous builds = 0
Total wall time: 0:00:10

View File

@ -0,0 +1,142 @@
LAMMPS (15 Feb 2016)
# Rhodopsin model
variable x index 1
variable y index 1
variable z index 1
units real
neigh_modify delay 5 every 1
atom_style full
atom_modify map hash
bond_style harmonic
angle_style charmm
dihedral_style charmm
improper_style harmonic
pair_style lj/charmm/coul/long 8.0 10.0
pair_modify mix arithmetic
kspace_style pppm 1e-4
read_data data.rhodo
orthogonal box = (-27.5 -38.5 -36.3646) to (27.5 38.5 36.3615)
1 by 2 by 2 MPI processor grid
reading atoms ...
32000 atoms
reading velocities ...
32000 velocities
scanning bonds ...
4 = max bonds/atom
scanning angles ...
8 = max angles/atom
scanning dihedrals ...
18 = max dihedrals/atom
scanning impropers ...
2 = max impropers/atom
reading bonds ...
27723 bonds
reading angles ...
40467 angles
reading dihedrals ...
56829 dihedrals
reading impropers ...
1034 impropers
4 = max # of 1-2 neighbors
12 = max # of 1-3 neighbors
24 = max # of 1-4 neighbors
26 = max # of special neighbors
replicate $x $y $z
replicate 2 $y $z
replicate 2 2 $z
replicate 2 2 1
orthogonal box = (-27.5 -38.5 -36.3646) to (82.5 115.5 36.3615)
2 by 2 by 1 MPI processor grid
128000 atoms
110892 bonds
161868 angles
227316 dihedrals
4136 impropers
4 = max # of 1-2 neighbors
12 = max # of 1-3 neighbors
24 = max # of 1-4 neighbors
26 = max # of special neighbors
fix 1 all shake 0.0001 5 0 m 1.0 a 232
6468 = # of size 2 clusters
14532 = # of size 3 clusters
2988 = # of size 4 clusters
16932 = # of frozen angles
fix 2 all npt temp 300.0 300.0 100.0 z 0.0 0.0 1000.0 mtk no pchain 0 tchain 1
special_bonds charmm
thermo 50
thermo_style multi
timestep 2.0
run 100
PPPM initialization ...
G vector (1/distance) = 0.248593
grid = 48 60 36
stencil order = 5
estimated absolute RMS force accuracy = 0.0359793
estimated relative force accuracy = 0.00010835
using double precision FFTs
3d grid and FFT values/proc = 41615 25920
Neighbor list info ...
1 neighbor list requests
update every 1 steps, delay 5 steps, check yes
max neighbors/atom: 2000, page size: 100000
master list distance cutoff = 12
ghost atom cutoff = 12
binsize = 6 -> bins = 19 26 13
Memory usage per processor = 95.5339 Mbytes
---------------- Step 0 ----- CPU = 0.0000 (sec) ----------------
TotEng = -101425.4887 KinEng = 85779.3251 Temp = 299.0304
PotEng = -187204.8138 E_bond = 10151.9760 E_angle = 43685.4968
E_dihed = 20847.1460 E_impro = 854.0463 E_vdwl = -9231.4537
E_coul = 827053.5824 E_long = -1080565.6077 Press = -142.3092
Volume = 1231980.1340
---------------- Step 50 ----- CPU = 18.7806 (sec) ----------------
TotEng = -101320.2677 KinEng = 86003.4837 Temp = 299.8118
PotEng = -187323.7514 E_bond = 9887.1072 E_angle = 43346.7922
E_dihed = 20958.7032 E_impro = 908.4715 E_vdwl = -7973.4457
E_coul = 826141.3831 E_long = -1080592.7629 Press = 238.0161
Volume = 1232126.1855
---------------- Step 100 ----- CPU = 38.3684 (sec) ----------------
TotEng = -101158.1849 KinEng = 86355.6149 Temp = 301.0393
PotEng = -187513.7998 E_bond = 10272.0693 E_angle = 43128.6454
E_dihed = 20793.9759 E_impro = 867.0826 E_vdwl = -7586.7186
E_coul = 825583.7122 E_long = -1080572.5667 Press = 15.2151
Volume = 1232535.8423
Loop time of 38.3684 on 4 procs for 100 steps with 128000 atoms
Performance: 0.450 ns/day, 53.289 hours/ns, 2.606 timesteps/s
99.9% CPU use with 4 MPI tasks x no OpenMP threads
MPI task timing breakdown:
Section | min time | avg time | max time |%varavg| %total
---------------------------------------------------------------
Pair | 26.205 | 26.538 | 26.911 | 5.0 | 69.17
Bond | 1.298 | 1.3125 | 1.3277 | 1.0 | 3.42
Kspace | 3.7099 | 4.0992 | 4.4422 | 13.3 | 10.68
Neigh | 4.6137 | 4.6144 | 4.615 | 0.0 | 12.03
Comm | 0.21398 | 0.21992 | 0.22886 | 1.2 | 0.57
Output | 0.00030518 | 0.00031543 | 0.00033307 | 0.1 | 0.00
Modify | 1.5066 | 1.5232 | 1.5388 | 1.0 | 3.97
Other | | 0.06051 | | | 0.16
Nlocal: 32000 ave 32000 max 32000 min
Histogram: 4 0 0 0 0 0 0 0 0 0
Nghost: 47957 ave 47957 max 47957 min
Histogram: 4 0 0 0 0 0 0 0 0 0
Neighs: 1.20281e+07 ave 1.20572e+07 max 1.1999e+07 min
Histogram: 2 0 0 0 0 0 0 0 0 2
Total # of neighbors = 48112472
Ave neighs/atom = 375.879
Ave special neighs/atom = 7.43187
Neighbor list builds = 11
Dangerous builds = 0
Total wall time: 0:00:39

View File

@ -1,67 +0,0 @@
LAMMPS (30 Apr 2015)
# FENE beadspring benchmark
units lj
atom_style bond
special_bonds fene
read_data data.chain
orthogonal box = (-16.796 -16.796 -16.796) to (16.796 16.796 16.796)
1 by 1 by 1 MPI processor grid
reading atoms ...
32000 atoms
reading velocities ...
32000 velocities
scanning bonds ...
1 = max bonds/atom
reading bonds ...
31680 bonds
2 = max # of 1-2 neighbors
2 = max # of special neighbors
neighbor 0.4 bin
neigh_modify every 1 delay 1
bond_style fene
bond_coeff 1 30.0 1.5 1.0 1.0
pair_style lj/cut 1.12
pair_modify shift yes
pair_coeff 1 1 1.0 1.0 1.12
fix 1 all nve
fix 2 all langevin 1.0 1.0 10.0 904297
thermo 100
timestep 0.012
run 100
Neighbor list info ...
1 neighbor list requests
update every 1 steps, delay 1 steps, check yes
master list distance cutoff = 1.52
Memory usage per processor = 11.5189 Mbytes
Step Temp E_pair E_mol TotEng Press
0 0.97029772 0.44484087 20.494523 22.394765 4.6721833
100 0.9729966 0.4361122 20.507698 22.40326 4.6548819
Loop time of 0.978717 on 1 procs for 100 steps with 32000 atoms
Pair time (%) = 0.195673 (19.9928)
Bond time (%) = 0.0878832 (8.97943)
Neigh time (%) = 0.448004 (45.7746)
Comm time (%) = 0.0329976 (3.37152)
Outpt time (%) = 0.000105143 (0.0107429)
Other time (%) = 0.214054 (21.8709)
Nlocal: 32000 ave 32000 max 32000 min
Histogram: 1 0 0 0 0 0 0 0 0 0
Nghost: 9493 ave 9493 max 9493 min
Histogram: 1 0 0 0 0 0 0 0 0 0
Neighs: 155873 ave 155873 max 155873 min
Histogram: 1 0 0 0 0 0 0 0 0 0
Total # of neighbors = 155873
Ave neighs/atom = 4.87103
Ave special neighs/atom = 1.98
Neighbor list builds = 25
Dangerous builds = 0

View File

@ -1,67 +0,0 @@
LAMMPS (30 Apr 2015)
# FENE beadspring benchmark
units lj
atom_style bond
special_bonds fene
read_data data.chain
orthogonal box = (-16.796 -16.796 -16.796) to (16.796 16.796 16.796)
1 by 2 by 2 MPI processor grid
reading atoms ...
32000 atoms
reading velocities ...
32000 velocities
scanning bonds ...
1 = max bonds/atom
reading bonds ...
31680 bonds
2 = max # of 1-2 neighbors
2 = max # of special neighbors
neighbor 0.4 bin
neigh_modify every 1 delay 1
bond_style fene
bond_coeff 1 30.0 1.5 1.0 1.0
pair_style lj/cut 1.12
pair_modify shift yes
pair_coeff 1 1 1.0 1.0 1.12
fix 1 all nve
fix 2 all langevin 1.0 1.0 10.0 904297
thermo 100
timestep 0.012
run 100
Neighbor list info ...
1 neighbor list requests
update every 1 steps, delay 1 steps, check yes
master list distance cutoff = 1.52
Memory usage per processor = 3.91518 Mbytes
Step Temp E_pair E_mol TotEng Press
0 0.97029772 0.44484087 20.494523 22.394765 4.6721833
100 0.97145835 0.43803883 20.502691 22.397872 4.626988
Loop time of 0.274371 on 4 procs for 100 steps with 32000 atoms
Pair time (%) = 0.0504887 (18.4016)
Bond time (%) = 0.0229129 (8.35106)
Neigh time (%) = 0.119957 (43.7206)
Comm time (%) = 0.020835 (7.59373)
Outpt time (%) = 5.74589e-05 (0.0209421)
Other time (%) = 0.0601202 (21.912)
Nlocal: 8000 ave 8030 max 7974 min
Histogram: 1 0 0 1 0 1 0 0 0 1
Nghost: 4177 ave 4191 max 4160 min
Histogram: 1 0 0 0 1 0 0 1 0 1
Neighs: 38995.8 ave 39169 max 38852 min
Histogram: 1 0 0 1 1 0 0 0 0 1
Total # of neighbors = 155983
Ave neighs/atom = 4.87447
Ave special neighs/atom = 1.98
Neighbor list builds = 25
Dangerous builds = 0

View File

@ -1,83 +0,0 @@
LAMMPS (30 Apr 2015)
# FENE beadspring benchmark
variable x index 1
variable y index 1
variable z index 1
units lj
atom_style bond
atom_modify map hash
special_bonds fene
read_data data.chain
orthogonal box = (-16.796 -16.796 -16.796) to (16.796 16.796 16.796)
1 by 2 by 2 MPI processor grid
reading atoms ...
32000 atoms
reading velocities ...
32000 velocities
scanning bonds ...
1 = max bonds/atom
reading bonds ...
31680 bonds
2 = max # of 1-2 neighbors
2 = max # of special neighbors
replicate $x $y $z
replicate 2 $y $z
replicate 2 2 $z
replicate 2 2 1
orthogonal box = (-16.796 -16.796 -16.796) to (50.388 50.388 16.796)
2 by 2 by 1 MPI processor grid
128000 atoms
126720 bonds
2 = max # of 1-2 neighbors
2 = max # of special neighbors
neighbor 0.4 bin
neigh_modify every 1 delay 1
bond_style fene
bond_coeff 1 30.0 1.5 1.0 1.0
pair_style lj/cut 1.12
pair_modify shift yes
pair_coeff 1 1 1.0 1.0 1.12
fix 1 all nve
fix 2 all langevin 1.0 1.0 10.0 904297
thermo 100
timestep 0.012
run 100
Neighbor list info ...
1 neighbor list requests
update every 1 steps, delay 1 steps, check yes
master list distance cutoff = 1.52
Memory usage per processor = 12.8735 Mbytes
Step Temp E_pair E_mol TotEng Press
0 0.97027498 0.44484087 20.494523 22.394765 4.6721833
100 0.97682955 0.44239968 20.500229 22.407862 4.6527025
Loop time of 1.19919 on 4 procs for 100 steps with 128000 atoms
Pair time (%) = 0.227794 (18.9957)
Bond time (%) = 0.0981662 (8.18606)
Neigh time (%) = 0.527868 (44.0188)
Comm time (%) = 0.0980042 (8.17255)
Outpt time (%) = 0.000200272 (0.0167006)
Other time (%) = 0.247155 (20.6102)
Nlocal: 32000 ave 32015 max 31983 min
Histogram: 1 0 1 0 0 0 0 0 1 1
Nghost: 9492 ave 9522 max 9432 min
Histogram: 1 0 0 0 0 0 1 0 0 2
Neighs: 155837 ave 156079 max 155506 min
Histogram: 1 0 0 0 0 1 0 0 1 1
Total # of neighbors = 623349
Ave neighs/atom = 4.86991
Ave special neighs/atom = 1.98
Neighbor list builds = 25
Dangerous builds = 0

View File

@ -1,69 +0,0 @@
LAMMPS (30 Apr 2015)
# LAMMPS benchmark of granular flow
# chute flow of 32000 atoms with frozen base at 26 degrees
units lj
atom_style sphere
boundary p p fs
newton off
comm_modify vel yes
read_data data.chute
orthogonal box = (0 0 0) to (40 20 37.2886)
1 by 1 by 1 MPI processor grid
reading atoms ...
32000 atoms
reading velocities ...
32000 velocities
pair_style gran/hooke/history 200000.0 NULL 50.0 NULL 0.5 0
pair_coeff * *
neighbor 0.1 bin
neigh_modify every 1 delay 0
timestep 0.0001
group bottom type 2
912 atoms in group bottom
group active subtract all bottom
31088 atoms in group active
neigh_modify exclude group bottom bottom
fix 1 all gravity 1.0 chute 26.0
fix 2 bottom freeze
fix 3 active nve/sphere
compute 1 all erotate/sphere
thermo_style custom step atoms ke c_1 vol
thermo_modify norm no
thermo 100
run 100
Neighbor list info ...
2 neighbor list requests
update every 1 steps, delay 0 steps, check yes
master list distance cutoff = 1.1
Memory usage per processor = 15.567 Mbytes
Step Atoms KinEng 1 Volume
0 32000 784139.13 1601.1263 29833.783
100 32000 784292.08 1571.0968 29834.707
Loop time of 0.539647 on 1 procs for 100 steps with 32000 atoms
Pair time (%) = 0.328789 (60.9267)
Neigh time (%) = 0.0401711 (7.44397)
Comm time (%) = 0.0179052 (3.31795)
Outpt time (%) = 0.00019908 (0.0368907)
Other time (%) = 0.152582 (28.2745)
Nlocal: 32000 ave 32000 max 32000 min
Histogram: 1 0 0 0 0 0 0 0 0 0
Nghost: 5463 ave 5463 max 5463 min
Histogram: 1 0 0 0 0 0 0 0 0 0
Neighs: 115133 ave 115133 max 115133 min
Histogram: 1 0 0 0 0 0 0 0 0 0
Total # of neighbors = 115133
Ave neighs/atom = 3.59791
Neighbor list builds = 2
Dangerous builds = 0

View File

@ -1,69 +0,0 @@
LAMMPS (30 Apr 2015)
# LAMMPS benchmark of granular flow
# chute flow of 32000 atoms with frozen base at 26 degrees
units lj
atom_style sphere
boundary p p fs
newton off
comm_modify vel yes
read_data data.chute
orthogonal box = (0 0 0) to (40 20 37.2886)
2 by 1 by 2 MPI processor grid
reading atoms ...
32000 atoms
reading velocities ...
32000 velocities
pair_style gran/hooke/history 200000.0 NULL 50.0 NULL 0.5 0
pair_coeff * *
neighbor 0.1 bin
neigh_modify every 1 delay 0
timestep 0.0001
group bottom type 2
912 atoms in group bottom
group active subtract all bottom
31088 atoms in group active
neigh_modify exclude group bottom bottom
fix 1 all gravity 1.0 chute 26.0
fix 2 bottom freeze
fix 3 active nve/sphere
compute 1 all erotate/sphere
thermo_style custom step atoms ke c_1 vol
thermo_modify norm no
thermo 100
run 100
Neighbor list info ...
2 neighbor list requests
update every 1 steps, delay 0 steps, check yes
master list distance cutoff = 1.1
Memory usage per processor = 6.81783 Mbytes
Step Atoms KinEng 1 Volume
0 32000 784139.13 1601.1263 29833.783
100 32000 784292.08 1571.0968 29834.707
Loop time of 0.146584 on 4 procs for 100 steps with 32000 atoms
Pair time (%) = 0.0737562 (50.3167)
Neigh time (%) = 0.0105147 (7.17314)
Comm time (%) = 0.0147474 (10.0607)
Outpt time (%) = 0.000131965 (0.0900267)
Other time (%) = 0.0474337 (32.3594)
Nlocal: 8000 ave 8008 max 7992 min
Histogram: 2 0 0 0 0 0 0 0 0 2
Nghost: 2439 ave 2450 max 2428 min
Histogram: 2 0 0 0 0 0 0 0 0 2
Neighs: 29500.5 ave 30488 max 28513 min
Histogram: 2 0 0 0 0 0 0 0 0 2
Total # of neighbors = 118002
Ave neighs/atom = 3.68756
Neighbor list builds = 2
Dangerous builds = 0

View File

@ -1,79 +0,0 @@
LAMMPS (30 Apr 2015)
# LAMMPS benchmark of granular flow
# chute flow of 32000 atoms with frozen base at 26 degrees
variable x index 1
variable y index 1
units lj
atom_style sphere
boundary p p fs
newton off
comm_modify vel yes
read_data data.chute
orthogonal box = (0 0 0) to (40 20 37.2886)
2 by 1 by 2 MPI processor grid
reading atoms ...
32000 atoms
reading velocities ...
32000 velocities
replicate $x $y 1
replicate 2 $y 1
replicate 2 2 1
orthogonal box = (0 0 0) to (80 40 37.2922)
2 by 2 by 1 MPI processor grid
128000 atoms
pair_style gran/hooke/history 200000.0 NULL 50.0 NULL 0.5 0
pair_coeff * *
neighbor 0.1 bin
neigh_modify every 1 delay 0
timestep 0.0001
group bottom type 2
3648 atoms in group bottom
group active subtract all bottom
124352 atoms in group active
neigh_modify exclude group bottom bottom
fix 1 all gravity 1.0 chute 26.0
fix 2 bottom freeze
fix 3 active nve/sphere
compute 1 all erotate/sphere
thermo_style custom step atoms ke c_1 vol
thermo_modify norm no
thermo 100
run 100
Neighbor list info ...
2 neighbor list requests
update every 1 steps, delay 0 steps, check yes
master list distance cutoff = 1.1
Memory usage per processor = 15.7007 Mbytes
Step Atoms KinEng 1 Volume
0 128000 3136556.5 6404.5051 119335.13
100 128000 3137168.3 6284.3873 119338.83
Loop time of 0.899154 on 4 procs for 100 steps with 128000 atoms
Pair time (%) = 0.523338 (58.2033)
Neigh time (%) = 0.0433982 (4.82656)
Comm time (%) = 0.0642623 (7.14697)
Outpt time (%) = 0.000541449 (0.0602175)
Other time (%) = 0.267615 (29.7629)
Nlocal: 32000 ave 32000 max 32000 min
Histogram: 4 0 0 0 0 0 0 0 0 0
Nghost: 5463 ave 5463 max 5463 min
Histogram: 4 0 0 0 0 0 0 0 0 0
Neighs: 115133 ave 115133 max 115133 min
Histogram: 4 0 0 0 0 0 0 0 0 0
Total # of neighbors = 460532
Ave neighs/atom = 3.59791
Neighbor list builds = 2
Dangerous builds = 0

View File

@ -1,71 +0,0 @@
LAMMPS (30 Apr 2015)
# bulk Cu lattice
variable x index 1
variable y index 1
variable z index 1
variable xx equal 20*$x
variable xx equal 20*1
variable yy equal 20*$y
variable yy equal 20*1
variable zz equal 20*$z
variable zz equal 20*1
units metal
atom_style atomic
lattice fcc 3.615
Lattice spacing in x,y,z = 3.615 3.615 3.615
region box block 0 ${xx} 0 ${yy} 0 ${zz}
region box block 0 20 0 ${yy} 0 ${zz}
region box block 0 20 0 20 0 ${zz}
region box block 0 20 0 20 0 20
create_box 1 box
Created orthogonal box = (0 0 0) to (72.3 72.3 72.3)
1 by 1 by 1 MPI processor grid
create_atoms 1 box
Created 32000 atoms
pair_style eam
pair_coeff 1 1 Cu_u3.eam
velocity all create 1600.0 376847 loop geom
neighbor 1.0 bin
neigh_modify every 1 delay 5 check yes
fix 1 all nve
timestep 0.005
thermo 50
run 100
Neighbor list info ...
1 neighbor list requests
update every 1 steps, delay 5 steps, check yes
master list distance cutoff = 5.95
Memory usage per processor = 10.2238 Mbytes
Step Temp E_pair E_mol TotEng Press
0 1600 -113280 0 -106662.09 18703.573
50 781.69049 -109873.35 0 -106640.13 52273.088
100 801.832 -109957.3 0 -106640.77 51322.821
Loop time of 5.89995 on 1 procs for 100 steps with 32000 atoms
Pair time (%) = 5.21525 (88.3948)
Neigh time (%) = 0.579447 (9.82122)
Comm time (%) = 0.0302751 (0.513142)
Outpt time (%) = 0.000234127 (0.00396829)
Other time (%) = 0.0747423 (1.26683)
Nlocal: 32000 ave 32000 max 32000 min
Histogram: 1 0 0 0 0 0 0 0 0 0
Nghost: 19909 ave 19909 max 19909 min
Histogram: 1 0 0 0 0 0 0 0 0 0
Neighs: 1.20778e+06 ave 1.20778e+06 max 1.20778e+06 min
Histogram: 1 0 0 0 0 0 0 0 0 0
Total # of neighbors = 1207784
Ave neighs/atom = 37.7433
Neighbor list builds = 13
Dangerous builds = 0

View File

@ -1,71 +0,0 @@
LAMMPS (30 Apr 2015)
# bulk Cu lattice
variable x index 1
variable y index 1
variable z index 1
variable xx equal 20*$x
variable xx equal 20*1
variable yy equal 20*$y
variable yy equal 20*1
variable zz equal 20*$z
variable zz equal 20*1
units metal
atom_style atomic
lattice fcc 3.615
Lattice spacing in x,y,z = 3.615 3.615 3.615
region box block 0 ${xx} 0 ${yy} 0 ${zz}
region box block 0 20 0 ${yy} 0 ${zz}
region box block 0 20 0 20 0 ${zz}
region box block 0 20 0 20 0 20
create_box 1 box
Created orthogonal box = (0 0 0) to (72.3 72.3 72.3)
1 by 2 by 2 MPI processor grid
create_atoms 1 box
Created 32000 atoms
pair_style eam
pair_coeff 1 1 Cu_u3.eam
velocity all create 1600.0 376847 loop geom
neighbor 1.0 bin
neigh_modify every 1 delay 5 check yes
fix 1 all nve
timestep 0.005
thermo 50
run 100
Neighbor list info ...
1 neighbor list requests
update every 1 steps, delay 5 steps, check yes
master list distance cutoff = 5.95
Memory usage per processor = 5.09629 Mbytes
Step Temp E_pair E_mol TotEng Press
0 1600 -113280 0 -106662.09 18703.573
50 781.69049 -109873.35 0 -106640.13 52273.088
100 801.832 -109957.3 0 -106640.77 51322.821
Loop time of 1.57597 on 4 procs for 100 steps with 32000 atoms
Pair time (%) = 1.36786 (86.7953)
Neigh time (%) = 0.152391 (9.6697)
Comm time (%) = 0.0353726 (2.2445)
Outpt time (%) = 0.000111699 (0.00708766)
Other time (%) = 0.0202255 (1.28337)
Nlocal: 8000 ave 8008 max 7993 min
Histogram: 2 0 0 0 0 0 0 0 1 1
Nghost: 9130.25 ave 9138 max 9122 min
Histogram: 2 0 0 0 0 0 0 0 0 2
Neighs: 301946 ave 302392 max 301360 min
Histogram: 1 0 0 0 1 0 0 0 1 1
Total # of neighbors = 1207784
Ave neighs/atom = 37.7433
Neighbor list builds = 13
Dangerous builds = 0

View File

@ -1,71 +0,0 @@
LAMMPS (30 Apr 2015)
# bulk Cu lattice
variable x index 1
variable y index 1
variable z index 1
variable xx equal 20*$x
variable xx equal 20*2
variable yy equal 20*$y
variable yy equal 20*2
variable zz equal 20*$z
variable zz equal 20*1
units metal
atom_style atomic
lattice fcc 3.615
Lattice spacing in x,y,z = 3.615 3.615 3.615
region box block 0 ${xx} 0 ${yy} 0 ${zz}
region box block 0 40 0 ${yy} 0 ${zz}
region box block 0 40 0 40 0 ${zz}
region box block 0 40 0 40 0 20
create_box 1 box
Created orthogonal box = (0 0 0) to (144.6 144.6 72.3)
2 by 2 by 1 MPI processor grid
create_atoms 1 box
Created 128000 atoms
pair_style eam
pair_coeff 1 1 Cu_u3.eam
velocity all create 1600.0 376847 loop geom
neighbor 1.0 bin
neigh_modify every 1 delay 5 check yes
fix 1 all nve
timestep 0.005
thermo 50
run 100
Neighbor list info ...
1 neighbor list requests
update every 1 steps, delay 5 steps, check yes
master list distance cutoff = 5.95
Memory usage per processor = 10.1402 Mbytes
Step Temp E_pair E_mol TotEng Press
0 1600 -453120 0 -426647.73 18704.012
50 779.50001 -439457.02 0 -426560.06 52355.276
100 797.97828 -439764.76 0 -426562.07 51474.74
Loop time of 6.4972 on 4 procs for 100 steps with 128000 atoms
Pair time (%) = 5.61297 (86.3906)
Neigh time (%) = 0.655333 (10.0864)
Comm time (%) = 0.130434 (2.00755)
Outpt time (%) = 0.000279069 (0.00429522)
Other time (%) = 0.0981811 (1.51113)
Nlocal: 32000 ave 32092 max 31914 min
Histogram: 1 0 0 1 0 1 0 0 0 1
Nghost: 19910 ave 19997 max 19818 min
Histogram: 1 0 0 0 1 0 1 0 0 1
Neighs: 1.20728e+06 ave 1.21142e+06 max 1.2036e+06 min
Histogram: 1 0 0 1 1 0 0 0 0 1
Total # of neighbors = 4829126
Ave neighs/atom = 37.7275
Neighbor list builds = 14
Dangerous builds = 0

View File

@ -1,68 +0,0 @@
LAMMPS (30 Apr 2015)
# 3d Lennard-Jones melt
variable x index 1
variable y index 1
variable z index 1
variable xx equal 20*$x
variable xx equal 20*1
variable yy equal 20*$y
variable yy equal 20*1
variable zz equal 20*$z
variable zz equal 20*1
units lj
atom_style atomic
lattice fcc 0.8442
Lattice spacing in x,y,z = 1.6796 1.6796 1.6796
region box block 0 ${xx} 0 ${yy} 0 ${zz}
region box block 0 20 0 ${yy} 0 ${zz}
region box block 0 20 0 20 0 ${zz}
region box block 0 20 0 20 0 20
create_box 1 box
Created orthogonal box = (0 0 0) to (33.5919 33.5919 33.5919)
1 by 1 by 1 MPI processor grid
create_atoms 1 box
Created 32000 atoms
mass 1 1.0
velocity all create 1.44 87287 loop geom
pair_style lj/cut 2.5
pair_coeff 1 1 1.0 1.0 2.5
neighbor 0.3 bin
neigh_modify delay 0 every 20 check no
fix 1 all nve
run 100
Neighbor list info ...
1 neighbor list requests
update every 20 steps, delay 0 steps, check no
master list distance cutoff = 2.8
Memory usage per processor = 8.21387 Mbytes
Step Temp E_pair E_mol TotEng Press
0 1.44 -6.7733681 0 -4.6134356 -5.0197073
100 0.7574531 -5.7585055 0 -4.6223613 0.20726105
Loop time of 2.25588 on 1 procs for 100 steps with 32000 atoms
Pair time (%) = 1.93512 (85.7815)
Neigh time (%) = 0.236483 (10.483)
Comm time (%) = 0.0239627 (1.06224)
Outpt time (%) = 0.000118017 (0.00523155)
Other time (%) = 0.0601869 (2.66801)
Nlocal: 32000 ave 32000 max 32000 min
Histogram: 1 0 0 0 0 0 0 0 0 0
Nghost: 19657 ave 19657 max 19657 min
Histogram: 1 0 0 0 0 0 0 0 0 0
Neighs: 1.20283e+06 ave 1.20283e+06 max 1.20283e+06 min
Histogram: 1 0 0 0 0 0 0 0 0 0
Total # of neighbors = 1202833
Ave neighs/atom = 37.5885
Neighbor list builds = 5
Dangerous builds = 0

View File

@ -1,68 +0,0 @@
LAMMPS (30 Apr 2015)
# 3d Lennard-Jones melt
variable x index 1
variable y index 1
variable z index 1
variable xx equal 20*$x
variable xx equal 20*1
variable yy equal 20*$y
variable yy equal 20*1
variable zz equal 20*$z
variable zz equal 20*1
units lj
atom_style atomic
lattice fcc 0.8442
Lattice spacing in x,y,z = 1.6796 1.6796 1.6796
region box block 0 ${xx} 0 ${yy} 0 ${zz}
region box block 0 20 0 ${yy} 0 ${zz}
region box block 0 20 0 20 0 ${zz}
region box block 0 20 0 20 0 20
create_box 1 box
Created orthogonal box = (0 0 0) to (33.5919 33.5919 33.5919)
1 by 2 by 2 MPI processor grid
create_atoms 1 box
Created 32000 atoms
mass 1 1.0
velocity all create 1.44 87287 loop geom
pair_style lj/cut 2.5
pair_coeff 1 1 1.0 1.0 2.5
neighbor 0.3 bin
neigh_modify delay 0 every 20 check no
fix 1 all nve
run 100
Neighbor list info ...
1 neighbor list requests
update every 20 steps, delay 0 steps, check no
master list distance cutoff = 2.8
Memory usage per processor = 4.09506 Mbytes
Step Temp E_pair E_mol TotEng Press
0 1.44 -6.7733681 0 -4.6134356 -5.0197073
100 0.7574531 -5.7585055 0 -4.6223613 0.20726105
Loop time of 0.623887 on 4 procs for 100 steps with 32000 atoms
Pair time (%) = 0.50691 (81.2504)
Neigh time (%) = 0.0619052 (9.92251)
Comm time (%) = 0.0389298 (6.23989)
Outpt time (%) = 5.85914e-05 (0.00939135)
Other time (%) = 0.0160829 (2.57785)
Nlocal: 8000 ave 8037 max 7964 min
Histogram: 2 0 0 0 0 0 0 0 1 1
Nghost: 9007.5 ave 9050 max 8968 min
Histogram: 1 1 0 0 0 0 0 1 0 1
Neighs: 300708 ave 305113 max 297203 min
Histogram: 1 0 0 1 1 0 0 0 0 1
Total # of neighbors = 1202833
Ave neighs/atom = 37.5885
Neighbor list builds = 5
Dangerous builds = 0

View File

@ -1,68 +0,0 @@
LAMMPS (30 Apr 2015)
# 3d Lennard-Jones melt
variable x index 1
variable y index 1
variable z index 1
variable xx equal 20*$x
variable xx equal 20*2
variable yy equal 20*$y
variable yy equal 20*2
variable zz equal 20*$z
variable zz equal 20*1
units lj
atom_style atomic
lattice fcc 0.8442
Lattice spacing in x,y,z = 1.6796 1.6796 1.6796
region box block 0 ${xx} 0 ${yy} 0 ${zz}
region box block 0 40 0 ${yy} 0 ${zz}
region box block 0 40 0 40 0 ${zz}
region box block 0 40 0 40 0 20
create_box 1 box
Created orthogonal box = (0 0 0) to (67.1838 67.1838 33.5919)
2 by 2 by 1 MPI processor grid
create_atoms 1 box
Created 128000 atoms
mass 1 1.0
velocity all create 1.44 87287 loop geom
pair_style lj/cut 2.5
pair_coeff 1 1 1.0 1.0 2.5
neighbor 0.3 bin
neigh_modify delay 0 every 20 check no
fix 1 all nve
run 100
Neighbor list info ...
1 neighbor list requests
update every 20 steps, delay 0 steps, check no
master list distance cutoff = 2.8
Memory usage per processor = 8.13678 Mbytes
Step Temp E_pair E_mol TotEng Press
0 1.44 -6.7733681 0 -4.6133849 -5.0196788
100 0.75841891 -5.759957 0 -4.6223375 0.20008866
Loop time of 2.53011 on 4 procs for 100 steps with 128000 atoms
Pair time (%) = 2.09024 (82.6146)
Neigh time (%) = 0.24414 (9.64939)
Comm time (%) = 0.111739 (4.41638)
Outpt time (%) = 0.000135601 (0.00535947)
Other time (%) = 0.0838551 (3.31428)
Nlocal: 32000 ave 32060 max 31939 min
Histogram: 1 0 1 0 0 0 0 1 0 1
Nghost: 19630.8 ave 19681 max 19562 min
Histogram: 1 0 0 0 1 0 0 0 1 1
Neighs: 1.20195e+06 ave 1.20354e+06 max 1.19931e+06 min
Histogram: 1 0 0 0 0 0 0 2 0 1
Total # of neighbors = 4807797
Ave neighs/atom = 37.5609
Neighbor list builds = 5
Dangerous builds = 0

View File

@ -1,110 +0,0 @@
LAMMPS (30 Apr 2015)
# Rhodopsin model
units real
neigh_modify delay 5 every 1
atom_style full
bond_style harmonic
angle_style charmm
dihedral_style charmm
improper_style harmonic
pair_style lj/charmm/coul/long 8.0 10.0
pair_modify mix arithmetic
kspace_style pppm 1e-4
read_data data.rhodo
orthogonal box = (-27.5 -38.5 -36.3646) to (27.5 38.5 36.3615)
1 by 1 by 1 MPI processor grid
reading atoms ...
32000 atoms
reading velocities ...
32000 velocities
scanning bonds ...
4 = max bonds/atom
scanning angles ...
8 = max angles/atom
scanning dihedrals ...
18 = max dihedrals/atom
scanning impropers ...
2 = max impropers/atom
reading bonds ...
27723 bonds
reading angles ...
40467 angles
reading dihedrals ...
56829 dihedrals
reading impropers ...
1034 impropers
4 = max # of 1-2 neighbors
12 = max # of 1-3 neighbors
24 = max # of 1-4 neighbors
26 = max # of special neighbors
fix 1 all shake 0.0001 5 0 m 1.0 a 232
1617 = # of size 2 clusters
3633 = # of size 3 clusters
747 = # of size 4 clusters
4233 = # of frozen angles
fix 2 all npt temp 300.0 300.0 100.0 z 0.0 0.0 1000.0 mtk no pchain 0 tchain 1
special_bonds charmm
thermo 50
thermo_style multi
timestep 2.0
run 100
PPPM initialization ...
G vector (1/distance) = 0.248835
grid = 25 32 32
stencil order = 5
estimated absolute RMS force accuracy = 0.0355478
estimated relative force accuracy = 0.000107051
using double precision FFTs
3d grid and FFT values/proc = 41070 25600
Neighbor list info ...
1 neighbor list requests
update every 1 steps, delay 5 steps, check yes
master list distance cutoff = 12
Memory usage per processor = 91.7487 Mbytes
---------------- Step 0 ----- CPU = 0.0000 (sec) ----------------
TotEng = -25356.2064 KinEng = 21444.8313 Temp = 299.0397
PotEng = -46801.0377 E_bond = 2537.9940 E_angle = 10921.3742
E_dihed = 5211.7865 E_impro = 213.5116 E_vdwl = -2307.8634
E_coul = 207025.8927 E_long = -270403.7333 Press = -142.6035
Volume = 307995.0335
---------------- Step 50 ----- CPU = 17.3751 (sec) ----------------
TotEng = -25330.0828 KinEng = 21501.0029 Temp = 299.8230
PotEng = -46831.0857 E_bond = 2471.7004 E_angle = 10836.4975
E_dihed = 5239.6299 E_impro = 227.1218 E_vdwl = -1993.2754
E_coul = 206797.6331 E_long = -270410.3930 Press = 237.6701
Volume = 308031.5639
---------------- Step 100 ----- CPU = 35.3771 (sec) ----------------
TotEng = -25290.7593 KinEng = 21592.0117 Temp = 301.0920
PotEng = -46882.7709 E_bond = 2567.9807 E_angle = 10781.9408
E_dihed = 5198.7432 E_impro = 216.7834 E_vdwl = -1902.4783
E_coul = 206659.2326 E_long = -270404.9733 Press = 6.9960
Volume = 308133.9888
Loop time of 35.3771 on 1 procs for 100 steps with 32000 atoms
Pair time (%) = 25.4765 (72.0139)
Bond time (%) = 1.27905 (3.61547)
Kspce time (%) = 3.22381 (9.11269)
Neigh time (%) = 4.26655 (12.0602)
Comm time (%) = 0.0692198 (0.195663)
Outpt time (%) = 0.000253916 (0.00071774)
Other time (%) = 1.06179 (3.00134)
Nlocal: 32000 ave 32000 max 32000 min
Histogram: 1 0 0 0 0 0 0 0 0 0
Nghost: 47958 ave 47958 max 47958 min
Histogram: 1 0 0 0 0 0 0 0 0 0
Neighs: 1.20281e+07 ave 1.20281e+07 max 1.20281e+07 min
Histogram: 1 0 0 0 0 0 0 0 0 0
Total # of neighbors = 12028107
Ave neighs/atom = 375.878
Ave special neighs/atom = 7.43187
Neighbor list builds = 11
Dangerous builds = 0

View File

@ -1,110 +0,0 @@
LAMMPS (30 Apr 2015)
# Rhodopsin model
units real
neigh_modify delay 5 every 1
atom_style full
bond_style harmonic
angle_style charmm
dihedral_style charmm
improper_style harmonic
pair_style lj/charmm/coul/long 8.0 10.0
pair_modify mix arithmetic
kspace_style pppm 1e-4
read_data data.rhodo
orthogonal box = (-27.5 -38.5 -36.3646) to (27.5 38.5 36.3615)
1 by 2 by 2 MPI processor grid
reading atoms ...
32000 atoms
reading velocities ...
32000 velocities
scanning bonds ...
4 = max bonds/atom
scanning angles ...
8 = max angles/atom
scanning dihedrals ...
18 = max dihedrals/atom
scanning impropers ...
2 = max impropers/atom
reading bonds ...
27723 bonds
reading angles ...
40467 angles
reading dihedrals ...
56829 dihedrals
reading impropers ...
1034 impropers
4 = max # of 1-2 neighbors
12 = max # of 1-3 neighbors
24 = max # of 1-4 neighbors
26 = max # of special neighbors
fix 1 all shake 0.0001 5 0 m 1.0 a 232
1617 = # of size 2 clusters
3633 = # of size 3 clusters
747 = # of size 4 clusters
4233 = # of frozen angles
fix 2 all npt temp 300.0 300.0 100.0 z 0.0 0.0 1000.0 mtk no pchain 0 tchain 1
special_bonds charmm
thermo 50
thermo_style multi
timestep 2.0
run 100
PPPM initialization ...
G vector (1/distance) = 0.248835
grid = 25 32 32
stencil order = 5
estimated absolute RMS force accuracy = 0.0355478
estimated relative force accuracy = 0.000107051
using double precision FFTs
3d grid and FFT values/proc = 13230 6400
Neighbor list info ...
1 neighbor list requests
update every 1 steps, delay 5 steps, check yes
master list distance cutoff = 12
Memory usage per processor = 36.629 Mbytes
---------------- Step 0 ----- CPU = 0.0000 (sec) ----------------
TotEng = -25356.2064 KinEng = 21444.8313 Temp = 299.0397
PotEng = -46801.0377 E_bond = 2537.9940 E_angle = 10921.3742
E_dihed = 5211.7865 E_impro = 213.5116 E_vdwl = -2307.8634
E_coul = 207025.8927 E_long = -270403.7333 Press = -142.6035
Volume = 307995.0335
---------------- Step 50 ----- CPU = 4.6438 (sec) ----------------
TotEng = -25330.0828 KinEng = 21501.0029 Temp = 299.8230
PotEng = -46831.0857 E_bond = 2471.7004 E_angle = 10836.4975
E_dihed = 5239.6299 E_impro = 227.1218 E_vdwl = -1993.2754
E_coul = 206797.6331 E_long = -270410.3930 Press = 237.6701
Volume = 308031.5639
---------------- Step 100 ----- CPU = 9.4301 (sec) ----------------
TotEng = -25290.7591 KinEng = 21592.0117 Temp = 301.0920
PotEng = -46882.7708 E_bond = 2567.9807 E_angle = 10781.9408
E_dihed = 5198.7432 E_impro = 216.7834 E_vdwl = -1902.4783
E_coul = 206659.2327 E_long = -270404.9733 Press = 6.9960
Volume = 308133.9888
Loop time of 9.43015 on 4 procs for 100 steps with 32000 atoms
Pair time (%) = 6.53815 (69.3324)
Bond time (%) = 0.323679 (3.43239)
Kspce time (%) = 1.02664 (10.8868)
Neigh time (%) = 1.11839 (11.8597)
Comm time (%) = 0.0812459 (0.861554)
Outpt time (%) = 0.000150442 (0.00159533)
Other time (%) = 0.341896 (3.62557)
Nlocal: 8000 ave 8143 max 7933 min
Histogram: 1 2 0 0 0 0 0 0 0 1
Nghost: 22733.5 ave 22769 max 22693 min
Histogram: 1 0 0 0 0 2 0 0 0 1
Neighs: 3.00703e+06 ave 3.0975e+06 max 2.96493e+06 min
Histogram: 1 2 0 0 0 0 0 0 0 1
Total # of neighbors = 12028107
Ave neighs/atom = 375.878
Ave special neighs/atom = 7.43187
Neighbor list builds = 11
Dangerous builds = 0

View File

@ -1,131 +0,0 @@
LAMMPS (30 Apr 2015)
# Rhodopsin model
variable x index 1
variable y index 1
variable z index 1
units real
neigh_modify delay 5 every 1
atom_style full
atom_modify map hash
bond_style harmonic
angle_style charmm
dihedral_style charmm
improper_style harmonic
pair_style lj/charmm/coul/long 8.0 10.0
pair_modify mix arithmetic
kspace_style pppm 1e-4
read_data data.rhodo
orthogonal box = (-27.5 -38.5 -36.3646) to (27.5 38.5 36.3615)
1 by 2 by 2 MPI processor grid
reading atoms ...
32000 atoms
reading velocities ...
32000 velocities
scanning bonds ...
4 = max bonds/atom
scanning angles ...
8 = max angles/atom
scanning dihedrals ...
18 = max dihedrals/atom
scanning impropers ...
2 = max impropers/atom
reading bonds ...
27723 bonds
reading angles ...
40467 angles
reading dihedrals ...
56829 dihedrals
reading impropers ...
1034 impropers
4 = max # of 1-2 neighbors
12 = max # of 1-3 neighbors
24 = max # of 1-4 neighbors
26 = max # of special neighbors
replicate $x $y $z
replicate 2 $y $z
replicate 2 2 $z
replicate 2 2 1
orthogonal box = (-27.5 -38.5 -36.3646) to (82.5 115.5 36.3615)
2 by 2 by 1 MPI processor grid
128000 atoms
110892 bonds
161868 angles
227316 dihedrals
4136 impropers
4 = max # of 1-2 neighbors
12 = max # of 1-3 neighbors
24 = max # of 1-4 neighbors
26 = max # of special neighbors
fix 1 all shake 0.0001 5 0 m 1.0 a 232
6468 = # of size 2 clusters
14532 = # of size 3 clusters
2988 = # of size 4 clusters
16932 = # of frozen angles
fix 2 all npt temp 300.0 300.0 100.0 z 0.0 0.0 1000.0 mtk no pchain 0 tchain 1
special_bonds charmm
thermo 50
thermo_style multi
timestep 2.0
run 100
PPPM initialization ...
G vector (1/distance) = 0.248593
grid = 48 60 36
stencil order = 5
estimated absolute RMS force accuracy = 0.0359793
estimated relative force accuracy = 0.00010835
using double precision FFTs
3d grid and FFT values/proc = 41615 25920
Neighbor list info ...
1 neighbor list requests
update every 1 steps, delay 5 steps, check yes
master list distance cutoff = 12
Memory usage per processor = 95.5339 Mbytes
---------------- Step 0 ----- CPU = 0.0000 (sec) ----------------
TotEng = -101425.4887 KinEng = 85779.3251 Temp = 299.0304
PotEng = -187204.8138 E_bond = 10151.9760 E_angle = 43685.4968
E_dihed = 20847.1460 E_impro = 854.0463 E_vdwl = -9231.4537
E_coul = 827053.5824 E_long = -1080565.6077 Press = -142.3092
Volume = 1231980.1340
---------------- Step 50 ----- CPU = 18.5923 (sec) ----------------
TotEng = -101320.2677 KinEng = 86003.4837 Temp = 299.8118
PotEng = -187323.7514 E_bond = 9887.1072 E_angle = 43346.7922
E_dihed = 20958.7032 E_impro = 908.4715 E_vdwl = -7973.4457
E_coul = 826141.3831 E_long = -1080592.7629 Press = 238.0161
Volume = 1232126.1855
---------------- Step 100 ----- CPU = 38.1551 (sec) ----------------
TotEng = -101158.1849 KinEng = 86355.6149 Temp = 301.0393
PotEng = -187513.7998 E_bond = 10272.0693 E_angle = 43128.6454
E_dihed = 20793.9759 E_impro = 867.0826 E_vdwl = -7586.7186
E_coul = 825583.7122 E_long = -1080572.5667 Press = 15.2151
Volume = 1232535.8423
Loop time of 38.1551 on 4 procs for 100 steps with 128000 atoms
Pair time (%) = 26.4472 (69.3149)
Bond time (%) = 1.31402 (3.44388)
Kspce time (%) = 4.23553 (11.1008)
Neigh time (%) = 4.45503 (11.6761)
Comm time (%) = 0.208946 (0.547622)
Outpt time (%) = 0.000290096 (0.000760307)
Other time (%) = 1.49411 (3.91587)
Nlocal: 32000 ave 32000 max 32000 min
Histogram: 4 0 0 0 0 0 0 0 0 0
Nghost: 47957 ave 47957 max 47957 min
Histogram: 4 0 0 0 0 0 0 0 0 0
Neighs: 1.20281e+07 ave 1.20572e+07 max 1.1999e+07 min
Histogram: 2 0 0 0 0 0 0 0 0 2
Total # of neighbors = 48112472
Ave neighs/atom = 375.879
Ave special neighs/atom = 7.43187
Neighbor list builds = 11
Dangerous builds = 0

1
doc/.gitignore vendored Normal file
View File

@ -0,0 +1 @@

View File

@ -1,14 +0,0 @@
\documentclass[12pt,article]{article}
\usepackage{indentfirst}
\usepackage{amsmath}
\begin{document}
\begin{eqnarray*}
r_{c}^{fcc} & = & \frac{1}{2} \left(\frac{\sqrt{2}}{2} + 1\right) \mathrm{a} \simeq 0.8536 \:\mathrm{a} \\
r_{c}^{bcc} & = & \frac{1}{2}(\sqrt{2} + 1) \mathrm{a} \simeq 1.207 \:\mathrm{a} \\
r_{c}^{hcp} & = & \frac{1}{2}\left(1+\sqrt{\frac{4+2x^{2}}{3}}\right) \mathrm{a}
\end{eqnarray*}
\end{document}

View File

@ -1,12 +0,0 @@
\documentclass[12pt,article]{article}
\usepackage{indentfirst}
\usepackage{amsmath}
\begin{document}
$$
Rc + Rs > 2*{\rm cutoff}
$$
\end{document}

Binary file not shown.

Before

Width:  |  Height:  |  Size: 19 KiB

View File

@ -1,9 +0,0 @@
\documentstyle[12pt]{article}
\begin{document}
$$
E_{Pauli(ECP_s)}=p_1\exp\left(-\frac{p_2r^2}{p_3+s^2} \right)
$$
\end{document}

View File

@ -1,8 +0,0 @@
\documentstyle[12pt]{article}
\begin{document}
$$
E_{Pauli(ECP_p)}=p_1\left( \frac{2}{p_2/s+s/p_2} \right)\left( r-p_3s\right)^2\exp \left[ -\frac{p_4\left( r-p_3s \right)^2}{p_5+s^2} \right]
$$

Binary file not shown.

Before

Width:  |  Height:  |  Size: 30 KiB

View File

@ -1,35 +0,0 @@
\documentclass[24pt]{article}
\pagestyle{empty}
\Huge
\begin{document}
\mathchardef\mhyphen="2D
% The imaginary unit
\providecommand*{\iu}%
{\ensuremath{{\rm i}}}
\begin{eqnarray*}
\exp \left(\iu{} L \Delta t \right) &=&<EFBFBD>
\exp \left(\iu{} L_{\rm T\mhyphen baro} \frac{\Delta t}{2} \right)
\exp \left(\iu{} L_{\rm T\mhyphen part} \frac{\Delta t}{2} \right)
\exp \left(\iu{} L_{\epsilon , 2} \frac{\Delta t}{2} \right)
\exp \left(\iu{} L_{2}^{(2)} \frac{\Delta t}{2} \right) \\
&&\times \left[
\exp \left(\iu{} L_{2}^{(1)} \frac{\Delta t}{2n} \right)
\exp \left(\iu{} L_{\epsilon , 1} \frac{\Delta t}{n} \right)
\exp \left(\iu{} L_1 \frac{\Delta t}{n} \right)
\exp \left(\iu{} L_{2}^{(1)} \frac{\Delta t}{2n} \right)
\right]^n \\
&&\times
\exp \left(\iu{} L_{2}^{(2)} \frac{\Delta t}{2} \right)
\exp \left(\iu{} L_{\epsilon , 2} \frac{\Delta t}{2} \right)
\exp \left(\iu{} L_{\rm T\mhyphen part} \frac{\Delta t}{2} \right)
\exp \left(\iu{} L_{\rm T\mhyphen baro} \frac{\Delta t}{2} \right) \\
&&+ \mathcal{O} \left(\Delta t^3 \right)
\end{eqnarray*}
\end{document}

Binary file not shown.

Before

Width:  |  Height:  |  Size: 1.7 KiB

View File

@ -1,11 +0,0 @@
\documentclass[12pt]{article}
\usepackage{amsmath}
\begin{document}
$$
F_{\text{total}} = \lambda F_{\text{int}}
$$
\end{document}

Binary file not shown.

Before

Width:  |  Height:  |  Size: 2.5 KiB

View File

@ -1,9 +0,0 @@
\documentclass[12pt]{article}
\begin{document}
$$
\lambda(\tau) = \lambda_i + \tau \left( \lambda_f - \lambda_i \right)
$$
\end{document}

Binary file not shown.

Before

Width:  |  Height:  |  Size: 3.2 KiB

View File

@ -1,9 +0,0 @@
\documentclass[12pt]{article}
\begin{document}
$$
\lambda(\tau) = \frac{\lambda_i}{1 + \tau \left( \frac{\lambda_i}{\lambda_f} - 1 \right)}
$$
\end{document}

Binary file not shown.

Before

Width:  |  Height:  |  Size: 4.7 KiB

View File

@ -1,9 +0,0 @@
\documentclass[12pt]{article}
\begin{document}
$$
\lambda(\tau) = \frac{\lambda_i}{ 1 + \log_2(1+\tau) \left( \frac{\lambda_i}{\lambda_f} - 1 \right)}
$$
\end{document}

Binary file not shown.

Before

Width:  |  Height:  |  Size: 3.5 KiB

View File

@ -1,11 +0,0 @@
\documentclass[12pt]{article}
\usepackage{amsmath}
\begin{document}
$$
F_{\text{total}} = \left( 1-\lambda \right) F_{\text{solid}} + \lambda F_{\text{harm}}
$$
\end{document}

Binary file not shown.

Before

Width:  |  Height:  |  Size: 1.0 KiB

Binary file not shown.

Before

Width:  |  Height:  |  Size: 5.6 KiB

View File

@ -1,12 +0,0 @@
\documentstyle[12pt]{article}
\begin{document}
$$
E = 2 \pi \epsilon \left[ \frac{2}{5} \left(\frac{\sigma}{r}\right)^{10} -
\left(\frac{\sigma}{r}\right)^4 -
\frac{\sqrt(2)\sigma^3}{3\left(r+\left(0.61/\sqrt(2)\right)\sigma\right)^3}\right]
\qquad r < r_c
$$
\end{document}

View File

@ -1,11 +0,0 @@
\documentclass[12pt]{article}
\begin{document}
$$
\kappa = \frac{V}{k_B T^2} \int_0^\infty \langle J_x(0) J_x(t) \rangle \, dt
= \frac{V}{3 k_B T^2} \int_0^\infty \langle \mathbf{J}(0) \cdot \mathbf{J}(t) \rangle \, dt
$$
\end{document}

View File

@ -1,13 +0,0 @@
\documentstyle[12pt]{article}
\begin{document}
$$
E=\frac{1}{2}K\left( \frac{1+cos\omega_0}{sin\omega_0}\right) ^2 \left( cos\omega - cos\omega_0\right) \qquad \omega_0 \neq 0^o
$$
$$
E=K\left( 1-cos\omega\right) \qquad \omega_0 = 0^o
$$
\end{document}

View File

@ -1,7 +0,0 @@
\documentclass[12pt]{article}
\begin{document} \large
\begin{eqnarray*}
E_T & = & \sum_i [ E_i^{self} (q_i) + \sum_{j>i} [E_{ij}^{short} (r_{ij}, q_i, q_j) + E_{ij}^{Coul} (r_{ij}, q_i, q_j)] + \\
&& E^{polar} (q_i, r_{ij}) + E^{vdW} (r_{ij}) + E^{barr} (q_i) + E^{corr} (r_{ij}, \theta_{jik})] \\
\end{eqnarray*}
\end{document}

View File

@ -1,23 +0,0 @@
\documentclass[10pt]{article}
\begin{document}
\begin{table}[h]
\begin{tabular}{|c|c|c|c|c|c|c|c|c|}
\hline
& $O$ & $Cu$ & $N$ & $C$ & $H$ & $Ti$ & $Zn$ & $Zr$ \\ \hline
$O$ & F & F & F & F & F & F & F & F\\ \hline
$Cu$ & F & F & P & F & F & P & F & P \\ \hline
$N$ & F & P & F & M & F & P & P & P \\ \hline
$C$ & F & F & M & F & F & M & M & M \\ \hline
$H$ & F & F & F & F & F & M & M & F \\ \hline
$Ti$ & F & P & P & M & M & F & P & P \\ \hline
$Zn$ & F & F & P & M & M & P & F & P \\ \hline
$Zr$ & F & P & P & M & F & P & P & F \\ \hline
\multicolumn{9}{l}{F: Fully optimized} \\
\multicolumn{9}{l}{M: Only optimized for dimer molecule} \\
\multicolumn{9}{l}{P: in Progress but have it from mixing rule} \\
\end{tabular}
\end{table}
\end{document}

87
doc/Makefile Normal file
View File

@ -0,0 +1,87 @@
# Makefile for LAMMPS documentation
SHA1 = $(shell echo $USER-$PWD | python utils/sha1sum.py)
BUILDDIR = /tmp/lammps-docs-$(SHA1)
RSTDIR = $(BUILDDIR)/rst
VENV = $(BUILDDIR)/docenv
TXT2RST = $(VENV)/bin/txt2rst
PYTHON = $(shell which python3)
ifeq ($(shell which python3 >/dev/null 2>&1; echo $$?), 1)
$(error Python3 was not found! Please check README.md for further instructions)
endif
ifeq ($(shell which virtualenv >/dev/null 2>&1; echo $$?), 1)
$(error virtualenv was not found! Please check README.md for further instructions)
endif
SOURCES=$(wildcard src/*.txt)
OBJECTS=$(SOURCES:src/%.txt=$(RSTDIR)/%.rst)
.PHONY: help clean-all clean html pdf venv
help:
@echo "Please use \`make <target>' where <target> is one of"
@echo " html to make HTML version of documentation using Sphinx"
@echo " pdf to make Manual.pdf"
@echo " txt2html to build txt2html tool"
@echo " clean to remove all generated RST files"
@echo " clean-all to reset the entire build environment"
clean-all:
rm -rf $(BUILDDIR)/* utils/txt2html/txt2html.exe
clean:
rm -rf $(RSTDIR)
txt2html: utils/txt2html/txt2html.exe
html: $(OBJECTS)
@(\
. $(VENV)/bin/activate ;\
cp -r src/* $(RSTDIR)/ ;\
sphinx-build -j 8 -b html -c utils/sphinx-config -d $(BUILDDIR)/doctrees $(RSTDIR) html ;\
deactivate ;\
)
-rm html/searchindex.js
-rm -rf html/_sources
@echo "Build finished. The HTML pages are in doc/html."
pdf: utils/txt2html/txt2html.exe
@(\
cd src; \
../utils/txt2html/txt2html.exe -b *.txt; \
htmldoc --batch ../lammps.book; \
for s in `echo *.txt | sed -e 's,\.txt,\.html,g'` ; \
do grep -q $$s ../lammps.book || \
echo doc file $$s missing in lammps.book; done; \
rm *.html; \
)
utils/txt2html/txt2html.exe: utils/txt2html/txt2html.cpp
g++ -O -Wall -o $@ $<
$(RSTDIR)/%.rst : src/%.txt $(TXT2RST)
@(\
mkdir -p $(RSTDIR) ; \
. $(VENV)/bin/activate ;\
txt2rst $< > $@ ;\
deactivate ;\
)
$(VENV):
@( \
virtualenv -p $(PYTHON) $(VENV); \
. $(VENV)/bin/activate; \
pip install Sphinx; \
pip install sphinxcontrib-images; \
deactivate;\
)
$(TXT2RST): $(VENV)
@( \
. $(VENV)/bin/activate; \
(cd utils/converters;\
python setup.py develop);\
deactivate;\
)

View File

@ -1,456 +0,0 @@
<!DOCTYPE html>
<!--[if IE 8]><html class="no-js lt-ie9" lang="en" > <![endif]-->
<!--[if gt IE 8]><!--> <html class="no-js" lang="en" > <!--<![endif]-->
<head>
<meta charset="utf-8">
<meta name="viewport" content="width=device-width, initial-scale=1.0">
<title>LAMMPS Documentation &mdash; LAMMPS 15 May 2015 version documentation</title>
<link rel="stylesheet" href="_static/css/theme.css" type="text/css" />
<link rel="stylesheet" href="_static/sphinxcontrib-images/LightBox2/lightbox2/css/lightbox.css" type="text/css" />
<link rel="top" title="LAMMPS 15 May 2015 version documentation" href="index.html"/>
<link rel="next" title="1. Introduction" href="Section_intro.html"/>
<script src="_static/js/modernizr.min.js"></script>
</head>
<body class="wy-body-for-nav" role="document">
<div class="wy-grid-for-nav">
<nav data-toggle="wy-nav-shift" class="wy-nav-side">
<div class="wy-side-nav-search">
<a href="#" class="icon icon-home"> LAMMPS
</a>
<div role="search">
<form id="rtd-search-form" class="wy-form" action="search.html" method="get">
<input type="text" name="q" placeholder="Search docs" />
<input type="hidden" name="check_keywords" value="yes" />
<input type="hidden" name="area" value="default" />
</form>
</div>
</div>
<div class="wy-menu wy-menu-vertical" data-spy="affix" role="navigation" aria-label="main navigation">
<ul>
<li class="toctree-l1"><a class="reference internal" href="Section_intro.html">1. Introduction</a></li>
<li class="toctree-l1"><a class="reference internal" href="Section_start.html">2. Getting Started</a></li>
<li class="toctree-l1"><a class="reference internal" href="Section_commands.html">3. Commands</a></li>
<li class="toctree-l1"><a class="reference internal" href="Section_packages.html">4. Packages</a></li>
<li class="toctree-l1"><a class="reference internal" href="Section_accelerate.html">5. Accelerating LAMMPS performance</a></li>
<li class="toctree-l1"><a class="reference internal" href="Section_howto.html">6. How-to discussions</a></li>
<li class="toctree-l1"><a class="reference internal" href="Section_example.html">7. Example problems</a></li>
<li class="toctree-l1"><a class="reference internal" href="Section_perf.html">8. Performance &amp; scalability</a></li>
<li class="toctree-l1"><a class="reference internal" href="Section_tools.html">9. Additional tools</a></li>
<li class="toctree-l1"><a class="reference internal" href="Section_modify.html">10. Modifying &amp; extending LAMMPS</a></li>
<li class="toctree-l1"><a class="reference internal" href="Section_python.html">11. Python interface to LAMMPS</a></li>
<li class="toctree-l1"><a class="reference internal" href="Section_errors.html">12. Errors</a></li>
<li class="toctree-l1"><a class="reference internal" href="Section_history.html">13. Future and history</a></li>
</ul>
</div>
&nbsp;
</nav>
<section data-toggle="wy-nav-shift" class="wy-nav-content-wrap">
<nav class="wy-nav-top" role="navigation" aria-label="top navigation">
<i data-toggle="wy-nav-top" class="fa fa-bars"></i>
<a href="#">LAMMPS</a>
</nav>
<div class="wy-nav-content">
<div class="rst-content">
<div role="navigation" aria-label="breadcrumbs navigation">
<ul class="wy-breadcrumbs">
<li><a href="#">Docs</a> &raquo;</li>
<li>LAMMPS Documentation</li>
<li class="wy-breadcrumbs-aside">
<a href="http://lammps.sandia.gov">Website</a>
<a href="Section_commands.html#comm">Commands</a>
</li>
</ul>
<hr/>
<div class="rst-footer-buttons" style="margin-bottom: 1em" role="navigation" aria-label="footer navigation">
<a href="Section_intro.html" class="btn btn-neutral float-right" title="1. Introduction" accesskey="n">Next <span class="fa fa-arrow-circle-right"></span></a>
</div>
</div>
<div role="main" class="document" itemscope="itemscope" itemtype="http://schema.org/Article">
<div itemprop="articleBody">
<H1></H1><div class="section" id="lammps-documentation">
<h1>LAMMPS Documentation<a class="headerlink" href="#lammps-documentation" title="Permalink to this headline"></a></h1>
<div class="section" id="dec-2015-version">
<h2>7 Dec 2015 version<a class="headerlink" href="#dec-2015-version" title="Permalink to this headline"></a></h2>
</div>
<div class="section" id="version-info">
<h2>Version info:<a class="headerlink" href="#version-info" title="Permalink to this headline"></a></h2>
<p>The LAMMPS &#8220;version&#8221; is the date when it was released, such as 1 May
2010. LAMMPS is updated continuously. Whenever we fix a bug or add a
feature, we release it immediately, and post a notice on <a class="reference external" href="http://lammps.sandia.gov/bug.html">this page of the WWW site</a>. Each dated copy of LAMMPS contains all the
features and bug-fixes up to and including that version date. The
version date is printed to the screen and logfile every time you run
LAMMPS. It is also in the file src/version.h and in the LAMMPS
directory name created when you unpack a tarball, and at the top of
the first page of the manual (this page).</p>
<ul class="simple">
<li>If you browse the HTML doc pages on the LAMMPS WWW site, they always
describe the most current version of LAMMPS.</li>
<li>If you browse the HTML doc pages included in your tarball, they
describe the version you have.</li>
<li>The <a class="reference external" href="Manual.pdf">PDF file</a> on the WWW site or in the tarball is updated
about once per month. This is because it is large, and we don&#8217;t want
it to be part of every patch.</li>
<li>There is also a <a class="reference external" href="Developer.pdf">Developer.pdf</a> file in the doc
directory, which describes the internal structure and algorithms of
LAMMPS.</li>
</ul>
<p>LAMMPS stands for Large-scale Atomic/Molecular Massively Parallel
Simulator.</p>
<p>LAMMPS is a classical molecular dynamics simulation code designed to
run efficiently on parallel computers. It was developed at Sandia
National Laboratories, a US Department of Energy facility, with
funding from the DOE. It is an open-source code, distributed freely
under the terms of the GNU Public License (GPL).</p>
<p>The primary developers of LAMMPS are <a class="reference external" href="http://www.sandia.gov/~sjplimp">Steve Plimpton</a>, Aidan
Thompson, and Paul Crozier who can be contacted at
sjplimp,athomps,pscrozi at sandia.gov. The <a class="reference external" href="http://lammps.sandia.gov">LAMMPS WWW Site</a> at
<a class="reference external" href="http://lammps.sandia.gov">http://lammps.sandia.gov</a> has more information about the code and its
uses.</p>
<hr class="docutils" />
<p>The LAMMPS documentation is organized into the following sections. If
you find errors or omissions in this manual or have suggestions for
useful information to add, please send an email to the developers so
we can improve the LAMMPS documentation.</p>
<p>Once you are familiar with LAMMPS, you may want to bookmark <a class="reference internal" href="Section_commands.html#comm"><span>this page</span></a> at Section_commands.html#comm since
it gives quick access to documentation for all LAMMPS commands.</p>
<p><a class="reference external" href="Manual.pdf">PDF file</a> of the entire manual, generated by
<a class="reference external" href="http://freecode.com/projects/htmldoc">htmldoc</a></p>
<div class="toctree-wrapper compound">
<ul>
<li class="toctree-l1"><a class="reference internal" href="Section_intro.html">1. Introduction</a><ul>
<li class="toctree-l2"><a class="reference internal" href="Section_intro.html#what-is-lammps">1.1. What is LAMMPS</a></li>
<li class="toctree-l2"><a class="reference internal" href="Section_intro.html#lammps-features">1.2. LAMMPS features</a></li>
<li class="toctree-l2"><a class="reference internal" href="Section_intro.html#lammps-non-features">1.3. LAMMPS non-features</a></li>
<li class="toctree-l2"><a class="reference internal" href="Section_intro.html#open-source-distribution">1.4. Open source distribution</a></li>
<li class="toctree-l2"><a class="reference internal" href="Section_intro.html#acknowledgments-and-citations">1.5. Acknowledgments and citations</a></li>
</ul>
</li>
<li class="toctree-l1"><a class="reference internal" href="Section_start.html">2. Getting Started</a><ul>
<li class="toctree-l2"><a class="reference internal" href="Section_start.html#what-s-in-the-lammps-distribution">2.1. What&#8217;s in the LAMMPS distribution</a></li>
<li class="toctree-l2"><a class="reference internal" href="Section_start.html#making-lammps">2.2. Making LAMMPS</a></li>
<li class="toctree-l2"><a class="reference internal" href="Section_start.html#making-lammps-with-optional-packages">2.3. Making LAMMPS with optional packages</a></li>
<li class="toctree-l2"><a class="reference internal" href="Section_start.html#building-lammps-via-the-make-py-tool">2.4. Building LAMMPS via the Make.py tool</a></li>
<li class="toctree-l2"><a class="reference internal" href="Section_start.html#building-lammps-as-a-library">2.5. Building LAMMPS as a library</a></li>
<li class="toctree-l2"><a class="reference internal" href="Section_start.html#running-lammps">2.6. Running LAMMPS</a></li>
<li class="toctree-l2"><a class="reference internal" href="Section_start.html#command-line-options">2.7. Command-line options</a></li>
<li class="toctree-l2"><a class="reference internal" href="Section_start.html#lammps-screen-output">2.8. LAMMPS screen output</a></li>
<li class="toctree-l2"><a class="reference internal" href="Section_start.html#tips-for-users-of-previous-lammps-versions">2.9. Tips for users of previous LAMMPS versions</a></li>
</ul>
</li>
<li class="toctree-l1"><a class="reference internal" href="Section_commands.html">3. Commands</a><ul>
<li class="toctree-l2"><a class="reference internal" href="Section_commands.html#lammps-input-script">3.1. LAMMPS input script</a></li>
<li class="toctree-l2"><a class="reference internal" href="Section_commands.html#parsing-rules">3.2. Parsing rules</a></li>
<li class="toctree-l2"><a class="reference internal" href="Section_commands.html#input-script-structure">3.3. Input script structure</a></li>
<li class="toctree-l2"><a class="reference internal" href="Section_commands.html#commands-listed-by-category">3.4. Commands listed by category</a></li>
<li class="toctree-l2"><a class="reference internal" href="Section_commands.html#individual-commands">3.5. Individual commands</a></li>
<li class="toctree-l2"><a class="reference internal" href="Section_commands.html#fix-styles">3.6. Fix styles</a></li>
<li class="toctree-l2"><a class="reference internal" href="Section_commands.html#compute-styles">3.7. Compute styles</a></li>
<li class="toctree-l2"><a class="reference internal" href="Section_commands.html#pair-style-potentials">3.8. Pair_style potentials</a></li>
<li class="toctree-l2"><a class="reference internal" href="Section_commands.html#bond-style-potentials">3.9. Bond_style potentials</a></li>
<li class="toctree-l2"><a class="reference internal" href="Section_commands.html#angle-style-potentials">3.10. Angle_style potentials</a></li>
<li class="toctree-l2"><a class="reference internal" href="Section_commands.html#dihedral-style-potentials">3.11. Dihedral_style potentials</a></li>
<li class="toctree-l2"><a class="reference internal" href="Section_commands.html#improper-style-potentials">3.12. Improper_style potentials</a></li>
<li class="toctree-l2"><a class="reference internal" href="Section_commands.html#kspace-solvers">3.13. Kspace solvers</a></li>
</ul>
</li>
<li class="toctree-l1"><a class="reference internal" href="Section_packages.html">4. Packages</a><ul>
<li class="toctree-l2"><a class="reference internal" href="Section_packages.html#standard-packages">4.1. Standard packages</a></li>
<li class="toctree-l2"><a class="reference internal" href="Section_packages.html#build-instructions-for-compress-package">4.2. Build instructions for COMPRESS package</a></li>
<li class="toctree-l2"><a class="reference internal" href="Section_packages.html#build-instructions-for-gpu-package">4.3. Build instructions for GPU package</a></li>
<li class="toctree-l2"><a class="reference internal" href="Section_packages.html#build-instructions-for-kim-package">4.4. Build instructions for KIM package</a></li>
<li class="toctree-l2"><a class="reference internal" href="Section_packages.html#build-instructions-for-kokkos-package">4.5. Build instructions for KOKKOS package</a></li>
<li class="toctree-l2"><a class="reference internal" href="Section_packages.html#build-instructions-for-kspace-package">4.6. Build instructions for KSPACE package</a></li>
<li class="toctree-l2"><a class="reference internal" href="Section_packages.html#build-instructions-for-meam-package">4.7. Build instructions for MEAM package</a></li>
<li class="toctree-l2"><a class="reference internal" href="Section_packages.html#build-instructions-for-poems-package">4.8. Build instructions for POEMS package</a></li>
<li class="toctree-l2"><a class="reference internal" href="Section_packages.html#build-instructions-for-python-package">4.9. Build instructions for PYTHON package</a></li>
<li class="toctree-l2"><a class="reference internal" href="Section_packages.html#build-instructions-for-reax-package">4.10. Build instructions for REAX package</a></li>
<li class="toctree-l2"><a class="reference internal" href="Section_packages.html#build-instructions-for-voronoi-package">4.11. Build instructions for VORONOI package</a></li>
<li class="toctree-l2"><a class="reference internal" href="Section_packages.html#build-instructions-for-xtc-package">4.12. Build instructions for XTC package</a></li>
<li class="toctree-l2"><a class="reference internal" href="Section_packages.html#user-packages">4.13. User packages</a></li>
<li class="toctree-l2"><a class="reference internal" href="Section_packages.html#user-atc-package">4.14. USER-ATC package</a></li>
<li class="toctree-l2"><a class="reference internal" href="Section_packages.html#user-awpmd-package">4.15. USER-AWPMD package</a></li>
<li class="toctree-l2"><a class="reference internal" href="Section_packages.html#user-cg-cmm-package">4.16. USER-CG-CMM package</a></li>
<li class="toctree-l2"><a class="reference internal" href="Section_packages.html#user-colvars-package">4.17. USER-COLVARS package</a></li>
<li class="toctree-l2"><a class="reference internal" href="Section_packages.html#user-cuda-package">4.18. USER-CUDA package</a></li>
<li class="toctree-l2"><a class="reference internal" href="Section_packages.html#user-diffraction-package">4.19. USER-DIFFRACTION package</a></li>
<li class="toctree-l2"><a class="reference internal" href="Section_packages.html#user-drude-package">4.20. USER-DRUDE package</a></li>
<li class="toctree-l2"><a class="reference internal" href="Section_packages.html#user-eff-package">4.21. USER-EFF package</a></li>
<li class="toctree-l2"><a class="reference internal" href="Section_packages.html#user-fep-package">4.22. USER-FEP package</a></li>
<li class="toctree-l2"><a class="reference internal" href="Section_packages.html#user-h5md-package">4.23. USER-H5MD package</a></li>
<li class="toctree-l2"><a class="reference internal" href="Section_packages.html#user-intel-package">4.24. USER-INTEL package</a></li>
<li class="toctree-l2"><a class="reference internal" href="Section_packages.html#user-lb-package">4.25. USER-LB package</a></li>
<li class="toctree-l2"><a class="reference internal" href="Section_packages.html#user-mgpt-package">4.26. USER-MGPT package</a></li>
<li class="toctree-l2"><a class="reference internal" href="Section_packages.html#user-misc-package">4.27. USER-MISC package</a></li>
<li class="toctree-l2"><a class="reference internal" href="Section_packages.html#user-molfile-package">4.28. USER-MOLFILE package</a></li>
<li class="toctree-l2"><a class="reference internal" href="Section_packages.html#user-omp-package">4.29. USER-OMP package</a></li>
<li class="toctree-l2"><a class="reference internal" href="Section_packages.html#user-phonon-package">4.30. USER-PHONON package</a></li>
<li class="toctree-l2"><a class="reference internal" href="Section_packages.html#user-qmmm-package">4.31. USER-QMMM package</a></li>
<li class="toctree-l2"><a class="reference internal" href="Section_packages.html#user-qtb-package">4.32. USER-QTB package</a></li>
<li class="toctree-l2"><a class="reference internal" href="Section_packages.html#user-reaxc-package">4.33. USER-REAXC package</a></li>
<li class="toctree-l2"><a class="reference internal" href="Section_packages.html#user-smd-package">4.34. USER-SMD package</a></li>
<li class="toctree-l2"><a class="reference internal" href="Section_packages.html#user-smtbq-package">4.35. USER-SMTBQ package</a></li>
<li class="toctree-l2"><a class="reference internal" href="Section_packages.html#user-sph-package">4.36. USER-SPH package</a></li>
</ul>
</li>
<li class="toctree-l1"><a class="reference internal" href="Section_accelerate.html">5. Accelerating LAMMPS performance</a><ul>
<li class="toctree-l2"><a class="reference internal" href="Section_accelerate.html#measuring-performance">5.1. Measuring performance</a></li>
<li class="toctree-l2"><a class="reference internal" href="Section_accelerate.html#general-strategies">5.2. General strategies</a></li>
<li class="toctree-l2"><a class="reference internal" href="Section_accelerate.html#packages-with-optimized-styles">5.3. Packages with optimized styles</a></li>
<li class="toctree-l2"><a class="reference internal" href="Section_accelerate.html#comparison-of-various-accelerator-packages">5.4. Comparison of various accelerator packages</a></li>
</ul>
</li>
<li class="toctree-l1"><a class="reference internal" href="Section_howto.html">6. How-to discussions</a><ul>
<li class="toctree-l2"><a class="reference internal" href="Section_howto.html#restarting-a-simulation">6.1. Restarting a simulation</a></li>
<li class="toctree-l2"><a class="reference internal" href="Section_howto.html#d-simulations">6.2. 2d simulations</a></li>
<li class="toctree-l2"><a class="reference internal" href="Section_howto.html#charmm-amber-and-dreiding-force-fields">6.3. CHARMM, AMBER, and DREIDING force fields</a></li>
<li class="toctree-l2"><a class="reference internal" href="Section_howto.html#running-multiple-simulations-from-one-input-script">6.4. Running multiple simulations from one input script</a></li>
<li class="toctree-l2"><a class="reference internal" href="Section_howto.html#multi-replica-simulations">6.5. Multi-replica simulations</a></li>
<li class="toctree-l2"><a class="reference internal" href="Section_howto.html#granular-models">6.6. Granular models</a></li>
<li class="toctree-l2"><a class="reference internal" href="Section_howto.html#tip3p-water-model">6.7. TIP3P water model</a></li>
<li class="toctree-l2"><a class="reference internal" href="Section_howto.html#tip4p-water-model">6.8. TIP4P water model</a></li>
<li class="toctree-l2"><a class="reference internal" href="Section_howto.html#spc-water-model">6.9. SPC water model</a></li>
<li class="toctree-l2"><a class="reference internal" href="Section_howto.html#coupling-lammps-to-other-codes">6.10. Coupling LAMMPS to other codes</a></li>
<li class="toctree-l2"><a class="reference internal" href="Section_howto.html#visualizing-lammps-snapshots">6.11. Visualizing LAMMPS snapshots</a></li>
<li class="toctree-l2"><a class="reference internal" href="Section_howto.html#triclinic-non-orthogonal-simulation-boxes">6.12. Triclinic (non-orthogonal) simulation boxes</a></li>
<li class="toctree-l2"><a class="reference internal" href="Section_howto.html#nemd-simulations">6.13. NEMD simulations</a></li>
<li class="toctree-l2"><a class="reference internal" href="Section_howto.html#finite-size-spherical-and-aspherical-particles">6.14. Finite-size spherical and aspherical particles</a></li>
<li class="toctree-l2"><a class="reference internal" href="Section_howto.html#output-from-lammps-thermo-dumps-computes-fixes-variables">6.15. Output from LAMMPS (thermo, dumps, computes, fixes, variables)</a></li>
<li class="toctree-l2"><a class="reference internal" href="Section_howto.html#thermostatting-barostatting-and-computing-temperature">6.16. Thermostatting, barostatting, and computing temperature</a></li>
<li class="toctree-l2"><a class="reference internal" href="Section_howto.html#walls">6.17. Walls</a></li>
<li class="toctree-l2"><a class="reference internal" href="Section_howto.html#elastic-constants">6.18. Elastic constants</a></li>
<li class="toctree-l2"><a class="reference internal" href="Section_howto.html#library-interface-to-lammps">6.19. Library interface to LAMMPS</a></li>
<li class="toctree-l2"><a class="reference internal" href="Section_howto.html#calculating-thermal-conductivity">6.20. Calculating thermal conductivity</a></li>
<li class="toctree-l2"><a class="reference internal" href="Section_howto.html#calculating-viscosity">6.21. Calculating viscosity</a></li>
<li class="toctree-l2"><a class="reference internal" href="Section_howto.html#calculating-a-diffusion-coefficient">6.22. Calculating a diffusion coefficient</a></li>
<li class="toctree-l2"><a class="reference internal" href="Section_howto.html#using-chunks-to-calculate-system-properties">6.23. Using chunks to calculate system properties</a></li>
<li class="toctree-l2"><a class="reference internal" href="Section_howto.html#setting-parameters-for-the-kspace-style-pppm-disp-command">6.24. Setting parameters for the <code class="docutils literal"><span class="pre">kspace_style</span> <span class="pre">pppm/disp</span></code> command</a></li>
<li class="toctree-l2"><a class="reference internal" href="Section_howto.html#polarizable-models">6.25. Polarizable models</a></li>
<li class="toctree-l2"><a class="reference internal" href="Section_howto.html#adiabatic-core-shell-model">6.26. Adiabatic core/shell model</a></li>
<li class="toctree-l2"><a class="reference internal" href="Section_howto.html#drude-induced-dipoles">6.27. Drude induced dipoles</a></li>
</ul>
</li>
<li class="toctree-l1"><a class="reference internal" href="Section_example.html">7. Example problems</a></li>
<li class="toctree-l1"><a class="reference internal" href="Section_perf.html">8. Performance &amp; scalability</a></li>
<li class="toctree-l1"><a class="reference internal" href="Section_tools.html">9. Additional tools</a><ul>
<li class="toctree-l2"><a class="reference internal" href="Section_tools.html#amber2lmp-tool">9.1. amber2lmp tool</a></li>
<li class="toctree-l2"><a class="reference internal" href="Section_tools.html#binary2txt-tool">9.2. binary2txt tool</a></li>
<li class="toctree-l2"><a class="reference internal" href="Section_tools.html#ch2lmp-tool">9.3. ch2lmp tool</a></li>
<li class="toctree-l2"><a class="reference internal" href="Section_tools.html#chain-tool">9.4. chain tool</a></li>
<li class="toctree-l2"><a class="reference internal" href="Section_tools.html#colvars-tools">9.5. colvars tools</a></li>
<li class="toctree-l2"><a class="reference internal" href="Section_tools.html#createatoms-tool">9.6. createatoms tool</a></li>
<li class="toctree-l2"><a class="reference internal" href="Section_tools.html#data2xmovie-tool">9.7. data2xmovie tool</a></li>
<li class="toctree-l2"><a class="reference internal" href="Section_tools.html#eam-database-tool">9.8. eam database tool</a></li>
<li class="toctree-l2"><a class="reference internal" href="Section_tools.html#eam-generate-tool">9.9. eam generate tool</a></li>
<li class="toctree-l2"><a class="reference internal" href="Section_tools.html#eff-tool">9.10. eff tool</a></li>
<li class="toctree-l2"><a class="reference internal" href="Section_tools.html#emacs-tool">9.11. emacs tool</a></li>
<li class="toctree-l2"><a class="reference internal" href="Section_tools.html#fep-tool">9.12. fep tool</a></li>
<li class="toctree-l2"><a class="reference internal" href="Section_tools.html#i-pi-tool">9.13. i-pi tool</a></li>
<li class="toctree-l2"><a class="reference internal" href="Section_tools.html#ipp-tool">9.14. ipp tool</a></li>
<li class="toctree-l2"><a class="reference internal" href="Section_tools.html#kate-tool">9.15. kate tool</a></li>
<li class="toctree-l2"><a class="reference internal" href="Section_tools.html#lmp2arc-tool">9.16. lmp2arc tool</a></li>
<li class="toctree-l2"><a class="reference internal" href="Section_tools.html#lmp2cfg-tool">9.17. lmp2cfg tool</a></li>
<li class="toctree-l2"><a class="reference internal" href="Section_tools.html#lmp2vmd-tool">9.18. lmp2vmd tool</a></li>
<li class="toctree-l2"><a class="reference internal" href="Section_tools.html#matlab-tool">9.19. matlab tool</a></li>
<li class="toctree-l2"><a class="reference internal" href="Section_tools.html#micelle2d-tool">9.20. micelle2d tool</a></li>
<li class="toctree-l2"><a class="reference internal" href="Section_tools.html#moltemplate-tool">9.21. moltemplate tool</a></li>
<li class="toctree-l2"><a class="reference internal" href="Section_tools.html#msi2lmp-tool">9.22. msi2lmp tool</a></li>
<li class="toctree-l2"><a class="reference internal" href="Section_tools.html#phonon-tool">9.23. phonon tool</a></li>
<li class="toctree-l2"><a class="reference internal" href="Section_tools.html#polymer-bonding-tool">9.24. polymer bonding tool</a></li>
<li class="toctree-l2"><a class="reference internal" href="Section_tools.html#pymol-asphere-tool">9.25. pymol_asphere tool</a></li>
<li class="toctree-l2"><a class="reference internal" href="Section_tools.html#python-tool">9.26. python tool</a></li>
<li class="toctree-l2"><a class="reference internal" href="Section_tools.html#reax-tool">9.27. reax tool</a></li>
<li class="toctree-l2"><a class="reference internal" href="Section_tools.html#restart2data-tool">9.28. restart2data tool</a></li>
<li class="toctree-l2"><a class="reference internal" href="Section_tools.html#vim-tool">9.29. vim tool</a></li>
<li class="toctree-l2"><a class="reference internal" href="Section_tools.html#xmgrace-tool">9.30. xmgrace tool</a></li>
<li class="toctree-l2"><a class="reference internal" href="Section_tools.html#xmovie-tool">9.31. xmovie tool</a></li>
</ul>
</li>
<li class="toctree-l1"><a class="reference internal" href="Section_modify.html">10. Modifying &amp; extending LAMMPS</a><ul>
<li class="toctree-l2"><a class="reference internal" href="Section_modify.html#atom-styles">10.1. Atom styles</a></li>
<li class="toctree-l2"><a class="reference internal" href="Section_modify.html#bond-angle-dihedral-improper-potentials">10.2. Bond, angle, dihedral, improper potentials</a></li>
<li class="toctree-l2"><a class="reference internal" href="Section_modify.html#compute-styles">10.3. Compute styles</a></li>
<li class="toctree-l2"><a class="reference internal" href="Section_modify.html#dump-styles">10.4. Dump styles</a></li>
<li class="toctree-l2"><a class="reference internal" href="Section_modify.html#dump-custom-output-options">10.5. Dump custom output options</a></li>
<li class="toctree-l2"><a class="reference internal" href="Section_modify.html#fix-styles">10.6. Fix styles</a></li>
<li class="toctree-l2"><a class="reference internal" href="Section_modify.html#input-script-commands">10.7. Input script commands</a></li>
<li class="toctree-l2"><a class="reference internal" href="Section_modify.html#kspace-computations">10.8. Kspace computations</a></li>
<li class="toctree-l2"><a class="reference internal" href="Section_modify.html#minimization-styles">10.9. Minimization styles</a></li>
<li class="toctree-l2"><a class="reference internal" href="Section_modify.html#pairwise-potentials">10.10. Pairwise potentials</a></li>
<li class="toctree-l2"><a class="reference internal" href="Section_modify.html#region-styles">10.11. Region styles</a></li>
<li class="toctree-l2"><a class="reference internal" href="Section_modify.html#body-styles">10.12. Body styles</a></li>
<li class="toctree-l2"><a class="reference internal" href="Section_modify.html#thermodynamic-output-options">10.13. Thermodynamic output options</a></li>
<li class="toctree-l2"><a class="reference internal" href="Section_modify.html#variable-options">10.14. Variable options</a></li>
<li class="toctree-l2"><a class="reference internal" href="Section_modify.html#submitting-new-features-for-inclusion-in-lammps">10.15. Submitting new features for inclusion in LAMMPS</a></li>
</ul>
</li>
<li class="toctree-l1"><a class="reference internal" href="Section_python.html">11. Python interface to LAMMPS</a><ul>
<li class="toctree-l2"><a class="reference internal" href="Section_python.html#overview-of-running-lammps-from-python">11.1. Overview of running LAMMPS from Python</a></li>
<li class="toctree-l2"><a class="reference internal" href="Section_python.html#overview-of-using-python-from-a-lammps-script">11.2. Overview of using Python from a LAMMPS script</a></li>
<li class="toctree-l2"><a class="reference internal" href="Section_python.html#building-lammps-as-a-shared-library">11.3. Building LAMMPS as a shared library</a></li>
<li class="toctree-l2"><a class="reference internal" href="Section_python.html#installing-the-python-wrapper-into-python">11.4. Installing the Python wrapper into Python</a></li>
<li class="toctree-l2"><a class="reference internal" href="Section_python.html#extending-python-with-mpi-to-run-in-parallel">11.5. Extending Python with MPI to run in parallel</a></li>
<li class="toctree-l2"><a class="reference internal" href="Section_python.html#testing-the-python-lammps-interface">11.6. Testing the Python-LAMMPS interface</a></li>
<li class="toctree-l2"><a class="reference internal" href="Section_python.html#using-lammps-from-python">11.7. Using LAMMPS from Python</a></li>
<li class="toctree-l2"><a class="reference internal" href="Section_python.html#example-python-scripts-that-use-lammps">11.8. Example Python scripts that use LAMMPS</a></li>
</ul>
</li>
<li class="toctree-l1"><a class="reference internal" href="Section_errors.html">12. Errors</a><ul>
<li class="toctree-l2"><a class="reference internal" href="Section_errors.html#common-problems">12.1. Common problems</a></li>
<li class="toctree-l2"><a class="reference internal" href="Section_errors.html#reporting-bugs">12.2. Reporting bugs</a></li>
<li class="toctree-l2"><a class="reference internal" href="Section_errors.html#error-warning-messages">12.3. Error &amp; warning messages</a></li>
<li class="toctree-l2"><a class="reference internal" href="Section_errors.html#error">12.4. Errors:</a></li>
<li class="toctree-l2"><a class="reference internal" href="Section_errors.html#warnings">12.5. Warnings:</a></li>
</ul>
</li>
<li class="toctree-l1"><a class="reference internal" href="Section_history.html">13. Future and history</a><ul>
<li class="toctree-l2"><a class="reference internal" href="Section_history.html#coming-attractions">13.1. Coming attractions</a></li>
<li class="toctree-l2"><a class="reference internal" href="Section_history.html#past-versions">13.2. Past versions</a></li>
</ul>
</li>
</ul>
</div>
</div>
</div>
<div class="section" id="indices-and-tables">
<h1>Indices and tables<a class="headerlink" href="#indices-and-tables" title="Permalink to this headline"></a></h1>
<ul class="simple">
<li><a class="reference internal" href="genindex.html"><span>Index</span></a></li>
<li><a class="reference internal" href="search.html"><span>Search Page</span></a></li>
</ul>
</BODY></div>
</div>
</div>
<footer>
<div class="rst-footer-buttons" role="navigation" aria-label="footer navigation">
<a href="Section_intro.html" class="btn btn-neutral float-right" title="1. Introduction" accesskey="n">Next <span class="fa fa-arrow-circle-right"></span></a>
</div>
<hr/>
<div role="contentinfo">
<p>
&copy; Copyright .
</p>
</div>
Built with <a href="http://sphinx-doc.org/">Sphinx</a> using a <a href="https://github.com/snide/sphinx_rtd_theme">theme</a> provided by <a href="https://readthedocs.org">Read the Docs</a>.
</footer>
</div>
</div>
</section>
</div>
<script type="text/javascript">
var DOCUMENTATION_OPTIONS = {
URL_ROOT:'./',
VERSION:'15 May 2015 version',
COLLAPSE_INDEX:false,
FILE_SUFFIX:'.html',
HAS_SOURCE: true
};
</script>
<script type="text/javascript" src="_static/jquery.js"></script>
<script type="text/javascript" src="_static/underscore.js"></script>
<script type="text/javascript" src="_static/doctools.js"></script>
<script type="text/javascript" src="https://cdn.mathjax.org/mathjax/latest/MathJax.js?config=TeX-AMS-MML_HTMLorMML"></script>
<script type="text/javascript" src="_static/sphinxcontrib-images/LightBox2/lightbox2/js/jquery-1.11.0.min.js"></script>
<script type="text/javascript" src="_static/sphinxcontrib-images/LightBox2/lightbox2/js/lightbox.min.js"></script>
<script type="text/javascript" src="_static/sphinxcontrib-images/LightBox2/lightbox2-customize/jquery-noconflict.js"></script>
<script type="text/javascript" src="_static/js/theme.js"></script>
<script type="text/javascript">
jQuery(function () {
SphinxRtdTheme.StickyNav.enable();
});
</script>
</body>
</html>

Binary file not shown.

File diff suppressed because it is too large Load Diff

48
doc/README.md Normal file
View File

@ -0,0 +1,48 @@
# Generation of LAMMPS Documentation
The generation of all the documentation is managed by the Makefile inside the
`doc/` folder.
## Usage:
```bash
make html # generate HTML using Sphinx
make pdf # generate PDF using htmldoc
make clean # remove generated RST files
make clean-all # remove entire build folder and any cached data
```
## Installing prerequisites
To run the documention build toolchain, Python 3 and virtualenv have
to be installed. Here are instructions for common setups:
### Ubuntu
```bash
sudo apt-get install python-virtualenv
```
### Fedora (up to version 21), Red Hat Enterprise Linux or CentOS (up to version 7.x)
```bash
sudo yum install python3-virtualenv
```
### Fedora (since version 22)
```bash
sudo dnf install python3-virtualenv
```
### MacOS X
## Python 3
Download the latest Python 3 MacOS X package from https://www.python.org and install it.
This will install both Python 3 and pip3.
## virtualenv
Once Python 3 is installed, open a Terminal and type `pip3 install virtualenv`. This will
install virtualenv from the Python Package Index.

View File

@ -1,89 +0,0 @@
#!/usr/bin/env python
"""
function:
parse the block of thermo data in a lammps logfile and perform auto- and
cross correlation of the specified column data. The total sum of the
correlation is also computed which can be converted to an integral by
multiplying by the timestep.
output:
standard output contains column data for the auto- & cross correlations
plus the total sum of each. Note, only the upper triangle of the
correlation matrix is computed.
usage:
correlate.py [-c col] <-c col2> <-s max_correlation_time> [logfile]
"""
import sys
import re
import array
# parse command line
maxCorrelationTime = 0
cols = array.array("I")
nCols = 0
args = sys.argv[1:]
index = 0
while index < len(args):
arg = args[index]
index += 1
if (arg == "-c"):
cols.append(int(args[index])-1)
nCols += 1
index += 1
elif (arg == "-s"):
maxCorrelationTime = int(args[index])
index += 1
else :
filename = arg
if (nCols < 1): raise RuntimeError, 'no data columns requested'
data = [array.array("d")]
for s in range(1,nCols) : data.append( array.array("d") )
# read data block from log file
start = False
input = open(filename)
nSamples = 0
pattern = re.compile('\d')
line = input.readline()
while line :
columns = line.split()
if (columns and pattern.match(columns[0])) :
for i in range(nCols):
data[i].append( float(columns[cols[i]]) )
nSamples += 1
start = True
else :
if (start) : break
line = input.readline()
print "# read :",nSamples," samples of ", nCols," data"
if( maxCorrelationTime < 1): maxCorrelationTime = int(nSamples/2);
# correlate and integrate
correlationPairs = []
for i in range(0,nCols):
for j in range(i,nCols): # note only upper triangle of the correlation matrix
correlationPairs.append([i,j])
header = "# "
for k in range(len(correlationPairs)):
i = str(correlationPairs[k][0]+1)
j = str(correlationPairs[k][1]+1)
header += " C"+i+j+" sum_C"+i+j
print header
nCorrelationPairs = len(correlationPairs)
sum = [0.0] * nCorrelationPairs
for s in range(maxCorrelationTime) :
correlation = [0.0] * nCorrelationPairs
nt = nSamples-s
for t in range(0,nt) :
for p in range(nCorrelationPairs):
i = correlationPairs[p][0]
j = correlationPairs[p][1]
correlation[p] += data[i][t]*data[j][s+t]
output = ""
for p in range(0,nCorrelationPairs):
correlation[p] /= nt
sum[p] += correlation[p]
output += str(correlation[p]) + " " + str(sum[p]) + " "
print output

View File

@ -1,642 +0,0 @@
<!DOCTYPE html>
<!--[if IE 8]><html class="no-js lt-ie9" lang="en" > <![endif]-->
<!--[if gt IE 8]><!--> <html class="no-js" lang="en" > <!--<![endif]-->
<head>
<meta charset="utf-8">
<meta name="viewport" content="width=device-width, initial-scale=1.0">
<title>5. Accelerating LAMMPS performance &mdash; LAMMPS 15 May 2015 version documentation</title>
<link rel="stylesheet" href="_static/css/theme.css" type="text/css" />
<link rel="stylesheet" href="_static/sphinxcontrib-images/LightBox2/lightbox2/css/lightbox.css" type="text/css" />
<link rel="top" title="LAMMPS 15 May 2015 version documentation" href="index.html"/>
<link rel="next" title="6. How-to discussions" href="Section_howto.html"/>
<link rel="prev" title="4. Packages" href="Section_packages.html"/>
<script src="_static/js/modernizr.min.js"></script>
</head>
<body class="wy-body-for-nav" role="document">
<div class="wy-grid-for-nav">
<nav data-toggle="wy-nav-shift" class="wy-nav-side">
<div class="wy-side-nav-search">
<a href="Manual.html" class="icon icon-home"> LAMMPS
</a>
<div role="search">
<form id="rtd-search-form" class="wy-form" action="search.html" method="get">
<input type="text" name="q" placeholder="Search docs" />
<input type="hidden" name="check_keywords" value="yes" />
<input type="hidden" name="area" value="default" />
</form>
</div>
</div>
<div class="wy-menu wy-menu-vertical" data-spy="affix" role="navigation" aria-label="main navigation">
<ul class="current">
<li class="toctree-l1"><a class="reference internal" href="Section_intro.html">1. Introduction</a></li>
<li class="toctree-l1"><a class="reference internal" href="Section_start.html">2. Getting Started</a></li>
<li class="toctree-l1"><a class="reference internal" href="Section_commands.html">3. Commands</a></li>
<li class="toctree-l1"><a class="reference internal" href="Section_packages.html">4. Packages</a></li>
<li class="toctree-l1 current"><a class="current reference internal" href="">5. Accelerating LAMMPS performance</a><ul>
<li class="toctree-l2"><a class="reference internal" href="#measuring-performance">5.1. Measuring performance</a></li>
<li class="toctree-l2"><a class="reference internal" href="#general-strategies">5.2. General strategies</a></li>
<li class="toctree-l2"><a class="reference internal" href="#packages-with-optimized-styles">5.3. Packages with optimized styles</a></li>
<li class="toctree-l2"><a class="reference internal" href="#comparison-of-various-accelerator-packages">5.4. Comparison of various accelerator packages</a><ul>
<li class="toctree-l3"><a class="reference internal" href="#examples">5.4.1. Examples</a></li>
</ul>
</li>
</ul>
</li>
<li class="toctree-l1"><a class="reference internal" href="Section_howto.html">6. How-to discussions</a></li>
<li class="toctree-l1"><a class="reference internal" href="Section_example.html">7. Example problems</a></li>
<li class="toctree-l1"><a class="reference internal" href="Section_perf.html">8. Performance &amp; scalability</a></li>
<li class="toctree-l1"><a class="reference internal" href="Section_tools.html">9. Additional tools</a></li>
<li class="toctree-l1"><a class="reference internal" href="Section_modify.html">10. Modifying &amp; extending LAMMPS</a></li>
<li class="toctree-l1"><a class="reference internal" href="Section_python.html">11. Python interface to LAMMPS</a></li>
<li class="toctree-l1"><a class="reference internal" href="Section_errors.html">12. Errors</a></li>
<li class="toctree-l1"><a class="reference internal" href="Section_history.html">13. Future and history</a></li>
</ul>
</div>
&nbsp;
</nav>
<section data-toggle="wy-nav-shift" class="wy-nav-content-wrap">
<nav class="wy-nav-top" role="navigation" aria-label="top navigation">
<i data-toggle="wy-nav-top" class="fa fa-bars"></i>
<a href="Manual.html">LAMMPS</a>
</nav>
<div class="wy-nav-content">
<div class="rst-content">
<div role="navigation" aria-label="breadcrumbs navigation">
<ul class="wy-breadcrumbs">
<li><a href="Manual.html">Docs</a> &raquo;</li>
<li>5. Accelerating LAMMPS performance</li>
<li class="wy-breadcrumbs-aside">
<a href="http://lammps.sandia.gov">Website</a>
<a href="Section_commands.html#comm">Commands</a>
</li>
</ul>
<hr/>
<div class="rst-footer-buttons" style="margin-bottom: 1em" role="navigation" aria-label="footer navigation">
<a href="Section_howto.html" class="btn btn-neutral float-right" title="6. How-to discussions" accesskey="n">Next <span class="fa fa-arrow-circle-right"></span></a>
<a href="Section_packages.html" class="btn btn-neutral" title="4. Packages" accesskey="p"><span class="fa fa-arrow-circle-left"></span> Previous</a>
</div>
</div>
<div role="main" class="document" itemscope="itemscope" itemtype="http://schema.org/Article">
<div itemprop="articleBody">
<div class="section" id="accelerating-lammps-performance">
<h1>5. Accelerating LAMMPS performance<a class="headerlink" href="#accelerating-lammps-performance" title="Permalink to this headline"></a></h1>
<p>This section describes various methods for improving LAMMPS
performance for different classes of problems running on different
kinds of machines.</p>
<p>There are two thrusts to the discussion that follows. The
first is using code options that implement alternate algorithms
that can speed-up a simulation. The second is to use one
of the several accelerator packages provided with LAMMPS that
contain code optimized for certain kinds of hardware, including
multi-core CPUs, GPUs, and Intel Xeon Phi coprocessors.</p>
<ul class="simple">
<li>5.1 <a class="reference internal" href="#acc-1"><span>Measuring performance</span></a></li>
<li>5.2 <a class="reference internal" href="#acc-2"><span>Algorithms and code options to boost performace</span></a></li>
<li>5.3 <a class="reference internal" href="#acc-3"><span>Accelerator packages with optimized styles</span></a></li>
<li>5.3.1 <a class="reference internal" href="accelerate_cuda.html"><em>USER-CUDA package</em></a></li>
<li>5.3.2 <a class="reference internal" href="accelerate_gpu.html"><em>GPU package</em></a></li>
<li>5.3.3 <a class="reference internal" href="accelerate_intel.html"><em>USER-INTEL package</em></a></li>
<li>5.3.4 <a class="reference internal" href="accelerate_kokkos.html"><em>KOKKOS package</em></a></li>
<li>5.3.5 <a class="reference internal" href="accelerate_omp.html"><em>USER-OMP package</em></a></li>
<li>5.3.6 <a class="reference internal" href="accelerate_opt.html"><em>OPT package</em></a></li>
<li>5.4 <a class="reference internal" href="#acc-4"><span>Comparison of various accelerator packages</span></a></li>
</ul>
<p>The <a class="reference external" href="http://lammps.sandia.gov/bench.html">Benchmark page</a> of the LAMMPS
web site gives performance results for the various accelerator
packages discussed in Section 5.2, for several of the standard LAMMPS
benchmark problems, as a function of problem size and number of
compute nodes, on different hardware platforms.</p>
<div class="section" id="measuring-performance">
<span id="acc-1"></span><h2>5.1. Measuring performance<a class="headerlink" href="#measuring-performance" title="Permalink to this headline"></a></h2>
<p>Before trying to make your simulation run faster, you should
understand how it currently performs and where the bottlenecks are.</p>
<p>The best way to do this is run the your system (actual number of
atoms) for a modest number of timesteps (say 100 steps) on several
different processor counts, including a single processor if possible.
Do this for an equilibrium version of your system, so that the
100-step timings are representative of a much longer run. There is
typically no need to run for 1000s of timesteps to get accurate
timings; you can simply extrapolate from short runs.</p>
<p>For the set of runs, look at the timing data printed to the screen and
log file at the end of each LAMMPS run. <a class="reference internal" href="Section_start.html#start-8"><span>This section</span></a> of the manual has an overview.</p>
<p>Running on one (or a few processors) should give a good estimate of
the serial performance and what portions of the timestep are taking
the most time. Running the same problem on a few different processor
counts should give an estimate of parallel scalability. I.e. if the
simulation runs 16x faster on 16 processors, its 100% parallel
efficient; if it runs 8x faster on 16 processors, it&#8217;s 50% efficient.</p>
<p>The most important data to look at in the timing info is the timing
breakdown and relative percentages. For example, trying different
options for speeding up the long-range solvers will have little impact
if they only consume 10% of the run time. If the pairwise time is
dominating, you may want to look at GPU or OMP versions of the pair
style, as discussed below. Comparing how the percentages change as
you increase the processor count gives you a sense of how different
operations within the timestep are scaling. Note that if you are
running with a Kspace solver, there is additional output on the
breakdown of the Kspace time. For PPPM, this includes the fraction
spent on FFTs, which can be communication intensive.</p>
<p>Another important detail in the timing info are the histograms of
atoms counts and neighbor counts. If these vary widely across
processors, you have a load-imbalance issue. This often results in
inaccurate relative timing data, because processors have to wait when
communication occurs for other processors to catch up. Thus the
reported times for &#8220;Communication&#8221; or &#8220;Other&#8221; may be higher than they
really are, due to load-imbalance. If this is an issue, you can
uncomment the MPI_Barrier() lines in src/timer.cpp, and recompile
LAMMPS, to obtain synchronized timings.</p>
<hr class="docutils" />
</div>
<div class="section" id="general-strategies">
<span id="acc-2"></span><h2>5.2. General strategies<a class="headerlink" href="#general-strategies" title="Permalink to this headline"></a></h2>
<div class="admonition note">
<p class="first admonition-title">Note</p>
<p class="last">this section 5.2 is still a work in progress</p>
</div>
<p>Here is a list of general ideas for improving simulation performance.
Most of them are only applicable to certain models and certain
bottlenecks in the current performance, so let the timing data you
generate be your guide. It is hard, if not impossible, to predict how
much difference these options will make, since it is a function of
problem size, number of processors used, and your machine. There is
no substitute for identifying performance bottlenecks, and trying out
various options.</p>
<ul class="simple">
<li>rRESPA</li>
<li>2-FFT PPPM</li>
<li>Staggered PPPM</li>
<li>single vs double PPPM</li>
<li>partial charge PPPM</li>
<li>verlet/split run style</li>
<li>processor command for proc layout and numa layout</li>
<li>load-balancing: balance and fix balance</li>
</ul>
<p>2-FFT PPPM, also called <em>analytic differentiation</em> or <em>ad</em> PPPM, uses
2 FFTs instead of the 4 FFTs used by the default <em>ik differentiation</em>
PPPM. However, 2-FFT PPPM also requires a slightly larger mesh size to
achieve the same accuracy as 4-FFT PPPM. For problems where the FFT
cost is the performance bottleneck (typically large problems running
on many processors), 2-FFT PPPM may be faster than 4-FFT PPPM.</p>
<p>Staggered PPPM performs calculations using two different meshes, one
shifted slightly with respect to the other. This can reduce force
aliasing errors and increase the accuracy of the method, but also
doubles the amount of work required. For high relative accuracy, using
staggered PPPM allows one to half the mesh size in each dimension as
compared to regular PPPM, which can give around a 4x speedup in the
kspace time. However, for low relative accuracy, using staggered PPPM
gives little benefit and can be up to 2x slower in the kspace
time. For example, the rhodopsin benchmark was run on a single
processor, and results for kspace time vs. relative accuracy for the
different methods are shown in the figure below. For this system,
staggered PPPM (using ik differentiation) becomes useful when using a
relative accuracy of slightly greater than 1e-5 and above.</p>
<img alt="_images/rhodo_staggered.jpg" class="align-center" src="_images/rhodo_staggered.jpg" />
<div class="admonition warning">
<p class="first admonition-title">Warning</p>
<p class="last">Using staggered PPPM may not give the same increase in
accuracy of energy and pressure as it does in forces, so some caution
must be used if energy and/or pressure are quantities of interest,
such as when using a barostat.</p>
</div>
<hr class="docutils" />
</div>
<div class="section" id="packages-with-optimized-styles">
<span id="acc-3"></span><h2>5.3. Packages with optimized styles<a class="headerlink" href="#packages-with-optimized-styles" title="Permalink to this headline"></a></h2>
<p>Accelerated versions of various <a class="reference internal" href="pair_style.html"><em>pair_style</em></a>,
<a class="reference internal" href="fix.html"><em>fixes</em></a>, <a class="reference internal" href="compute.html"><em>computes</em></a>, and other commands have
been added to LAMMPS, which will typically run faster than the
standard non-accelerated versions. Some require appropriate hardware
to be present on your system, e.g. GPUs or Intel Xeon Phi
coprocessors.</p>
<p>All of these commands are in packages provided with LAMMPS. An
overview of packages is give in <a class="reference internal" href="Section_packages.html"><em>Section packages</em></a>.</p>
<p>These are the accelerator packages
currently in LAMMPS, either as standard or user packages:</p>
<table border="1" class="docutils">
<colgroup>
<col width="44%" />
<col width="56%" />
</colgroup>
<tbody valign="top">
<tr class="row-odd"><td><a class="reference internal" href="accelerate_cuda.html"><em>USER-CUDA</em></a></td>
<td>for NVIDIA GPUs</td>
</tr>
<tr class="row-even"><td><a class="reference internal" href="accelerate_gpu.html"><em>GPU</em></a></td>
<td>for NVIDIA GPUs as well as OpenCL support</td>
</tr>
<tr class="row-odd"><td><a class="reference internal" href="accelerate_intel.html"><em>USER-INTEL</em></a></td>
<td>for Intel CPUs and Intel Xeon Phi</td>
</tr>
<tr class="row-even"><td><a class="reference internal" href="accelerate_kokkos.html"><em>KOKKOS</em></a></td>
<td>for GPUs, Intel Xeon Phi, and OpenMP threading</td>
</tr>
<tr class="row-odd"><td><a class="reference internal" href="accelerate_omp.html"><em>USER-OMP</em></a></td>
<td>for OpenMP threading</td>
</tr>
<tr class="row-even"><td><a class="reference internal" href="accelerate_opt.html"><em>OPT</em></a></td>
<td>generic CPU optimizations</td>
</tr>
</tbody>
</table>
<p>Inverting this list, LAMMPS currently has acceleration support for
three kinds of hardware, via the listed packages:</p>
<table border="1" class="docutils">
<colgroup>
<col width="10%" />
<col width="90%" />
</colgroup>
<tbody valign="top">
<tr class="row-odd"><td>Many-core CPUs</td>
<td><a class="reference internal" href="accelerate_intel.html"><em>USER-INTEL</em></a>, <a class="reference internal" href="accelerate_kokkos.html"><em>KOKKOS</em></a>, <a class="reference internal" href="accelerate_omp.html"><em>USER-OMP</em></a>, <a class="reference internal" href="accelerate_opt.html"><em>OPT</em></a> packages</td>
</tr>
<tr class="row-even"><td>NVIDIA GPUs</td>
<td><a class="reference internal" href="accelerate_cuda.html"><em>USER-CUDA</em></a>, <a class="reference internal" href="accelerate_gpu.html"><em>GPU</em></a>, <a class="reference internal" href="accelerate_kokkos.html"><em>KOKKOS</em></a> packages</td>
</tr>
<tr class="row-odd"><td>Intel Phi</td>
<td><a class="reference internal" href="accelerate_intel.html"><em>USER-INTEL</em></a>, <a class="reference internal" href="accelerate_kokkos.html"><em>KOKKOS</em></a> packages</td>
</tr>
</tbody>
</table>
<p>Which package is fastest for your hardware may depend on the size
problem you are running and what commands (accelerated and
non-accelerated) are invoked by your input script. While these doc
pages include performance guidelines, there is no substitute for
trying out the different packages appropriate to your hardware.</p>
<p>Any accelerated style has the same name as the corresponding standard
style, except that a suffix is appended. Otherwise, the syntax for
the command that uses the style is identical, their functionality is
the same, and the numerical results it produces should also be the
same, except for precision and round-off effects.</p>
<p>For example, all of these styles are accelerated variants of the
Lennard-Jones <a class="reference internal" href="pair_lj.html"><em>pair_style lj/cut</em></a>:</p>
<ul class="simple">
<li><a class="reference internal" href="pair_lj.html"><em>pair_style lj/cut/cuda</em></a></li>
<li><a class="reference internal" href="pair_lj.html"><em>pair_style lj/cut/gpu</em></a></li>
<li><a class="reference internal" href="pair_lj.html"><em>pair_style lj/cut/intel</em></a></li>
<li><a class="reference internal" href="pair_lj.html"><em>pair_style lj/cut/kk</em></a></li>
<li><a class="reference internal" href="pair_lj.html"><em>pair_style lj/cut/omp</em></a></li>
<li><a class="reference internal" href="pair_lj.html"><em>pair_style lj/cut/opt</em></a></li>
</ul>
<p>To see what accelerate styles are currently available, see
<a class="reference internal" href="Section_commands.html#cmd-5"><span>Section_commands 5</span></a> of the manual. The
doc pages for individual commands (e.g. <a class="reference internal" href="pair_lj.html"><em>pair lj/cut</em></a> or
<a class="reference internal" href="fix_nve.html"><em>fix nve</em></a>) also list any accelerated variants available
for that style.</p>
<p>To use an accelerator package in LAMMPS, and one or more of the styles
it provides, follow these general steps. Details vary from package to
package and are explained in the individual accelerator doc pages,
listed above:</p>
<table border="1" class="docutils">
<colgroup>
<col width="26%" />
<col width="74%" />
</colgroup>
<tbody valign="top">
<tr class="row-odd"><td>build the accelerator library</td>
<td>only for USER-CUDA and GPU packages</td>
</tr>
<tr class="row-even"><td>install the accelerator package</td>
<td>make yes-opt, make yes-user-intel, etc</td>
</tr>
</tbody>
</table>
<div class="line-block">
<div class="line">install the accelerator package | make yes-opt, make yes-user-intel, etc |</div>
</div>
<blockquote>
<div>only for USER-INTEL, KOKKOS, USER-OMP, OPT packages |</div></blockquote>
<table border="1" class="docutils">
<colgroup>
<col width="26%" />
<col width="74%" />
</colgroup>
<tbody valign="top">
<tr class="row-odd"><td>re-build LAMMPS</td>
<td>make machine</td>
</tr>
</tbody>
</table>
<div class="line-block">
<div class="line">re-build LAMMPS | make machine |</div>
</div>
<blockquote>
<div>mpirun -np 32 lmp_machine -in in.script |</div></blockquote>
<blockquote>
<div>only for USER-CUDA and KOKKOS packages |</div></blockquote>
<blockquote>
<div><a class="reference internal" href="package.html"><em>package</em></a> command, &lt;br&gt;
only if defaults need to be changed |</div></blockquote>
<blockquote>
<div><a class="reference internal" href="suffix.html"><em>suffix</em></a> command |</div></blockquote>
<table border="1" class="docutils">
<colgroup>
</colgroup>
<tbody valign="top">
</tbody>
</table>
<p>Note that the first 4 steps can be done as a single command, using the
src/Make.py tool. This tool is discussed in <a class="reference internal" href="Section_start.html#start-4"><span>Section 2.4</span></a> of the manual, and its use is
illustrated in the individual accelerator sections. Typically these
steps only need to be done once, to create an executable that uses one
or more accelerator packages.</p>
<p>The last 4 steps can all be done from the command-line when LAMMPS is
launched, without changing your input script, as illustrated in the
individual accelerator sections. Or you can add
<a class="reference internal" href="package.html"><em>package</em></a> and <a class="reference internal" href="suffix.html"><em>suffix</em></a> commands to your input
script.</p>
<div class="admonition warning">
<p class="first admonition-title">Warning</p>
<p class="last">With a few exceptions, you can build a single LAMMPS
executable with all its accelerator packages installed. Note however
that the USER-INTEL and KOKKOS packages require you to choose one of
their hardware options when building for a specific platform.
I.e. CPU or Phi option for the USER-INTEL package. Or the OpenMP,
Cuda, or Phi option for the KOKKOS package.</p>
</div>
<p>These are the exceptions. You cannot build a single executable with:</p>
<ul class="simple">
<li>both the USER-INTEL Phi and KOKKOS Phi options</li>
<li>the USER-INTEL Phi or Kokkos Phi option, and either the USER-CUDA or GPU packages</li>
</ul>
<p>See the examples/accelerate/README and make.list files for sample
Make.py commands that build LAMMPS with any or all of the accelerator
packages. As an example, here is a command that builds with all the
GPU related packages installed (USER-CUDA, GPU, KOKKOS with Cuda),
including settings to build the needed auxiliary USER-CUDA and GPU
libraries for Kepler GPUs:</p>
<pre class="literal-block">
Make.py -j 16 -p omp gpu cuda kokkos -cc nvcc wrap=mpi -cuda mode=double arch=35 -gpu mode=double arch=35 -kokkos cuda arch=35 lib-all file mpi
</pre>
<p>The examples/accelerate directory also has input scripts that can be
used with all of the accelerator packages. See its README file for
details.</p>
<p>Likewise, the bench directory has FERMI and KEPLER and PHI
sub-directories with Make.py commands and input scripts for using all
the accelerator packages on various machines. See the README files in
those dirs.</p>
<p>As mentioned above, the <a class="reference external" href="http://lammps.sandia.gov/bench.html">Benchmark page</a> of the LAMMPS web site gives
performance results for the various accelerator packages for several
of the standard LAMMPS benchmark problems, as a function of problem
size and number of compute nodes, on different hardware platforms.</p>
<p>Here is a brief summary of what the various packages provide. Details
are in the individual accelerator sections.</p>
<ul class="simple">
<li>Styles with a &#8220;cuda&#8221; or &#8220;gpu&#8221; suffix are part of the USER-CUDA or GPU
packages, and can be run on NVIDIA GPUs. The speed-up on a GPU
depends on a variety of factors, discussed in the accelerator
sections.</li>
<li>Styles with an &#8220;intel&#8221; suffix are part of the USER-INTEL
package. These styles support vectorized single and mixed precision
calculations, in addition to full double precision. In extreme cases,
this can provide speedups over 3.5x on CPUs. The package also
supports acceleration in &#8220;offload&#8221; mode to Intel(R) Xeon Phi(TM)
coprocessors. This can result in additional speedup over 2x depending
on the hardware configuration.</li>
<li>Styles with a &#8220;kk&#8221; suffix are part of the KOKKOS package, and can be
run using OpenMP on multicore CPUs, on an NVIDIA GPU, or on an Intel
Xeon Phi in &#8220;native&#8221; mode. The speed-up depends on a variety of
factors, as discussed on the KOKKOS accelerator page.</li>
<li>Styles with an &#8220;omp&#8221; suffix are part of the USER-OMP package and allow
a pair-style to be run in multi-threaded mode using OpenMP. This can
be useful on nodes with high-core counts when using less MPI processes
than cores is advantageous, e.g. when running with PPPM so that FFTs
are run on fewer MPI processors or when the many MPI tasks would
overload the available bandwidth for communication.</li>
<li>Styles with an &#8220;opt&#8221; suffix are part of the OPT package and typically
speed-up the pairwise calculations of your simulation by 5-25% on a
CPU.</li>
</ul>
<p>The individual accelerator package doc pages explain:</p>
<ul class="simple">
<li>what hardware and software the accelerated package requires</li>
<li>how to build LAMMPS with the accelerated package</li>
<li>how to run with the accelerated package either via command-line switches or modifying the input script</li>
<li>speed-ups to expect</li>
<li>guidelines for best performance</li>
<li>restrictions</li>
</ul>
<hr class="docutils" />
</div>
<div class="section" id="comparison-of-various-accelerator-packages">
<span id="acc-4"></span><h2>5.4. Comparison of various accelerator packages<a class="headerlink" href="#comparison-of-various-accelerator-packages" title="Permalink to this headline"></a></h2>
<div class="admonition note">
<p class="first admonition-title">Note</p>
<p class="last">this section still needs to be re-worked with additional KOKKOS
and USER-INTEL information.</p>
</div>
<p>The next section compares and contrasts the various accelerator
options, since there are multiple ways to perform OpenMP threading,
run on GPUs, and run on Intel Xeon Phi coprocessors.</p>
<p>All 3 of these packages accelerate a LAMMPS calculation using NVIDIA
hardware, but they do it in different ways.</p>
<p>As a consequence, for a particular simulation on specific hardware,
one package may be faster than the other. We give guidelines below,
but the best way to determine which package is faster for your input
script is to try both of them on your machine. See the benchmarking
section below for examples where this has been done.</p>
<p><strong>Guidelines for using each package optimally:</strong></p>
<ul class="simple">
<li>The GPU package allows you to assign multiple CPUs (cores) to a single
GPU (a common configuration for &#8220;hybrid&#8221; nodes that contain multicore
CPU(s) and GPU(s)) and works effectively in this mode. The USER-CUDA
package does not allow this; you can only use one CPU per GPU.</li>
<li>The GPU package moves per-atom data (coordinates, forces)
back-and-forth between the CPU and GPU every timestep. The USER-CUDA
package only does this on timesteps when a CPU calculation is required
(e.g. to invoke a fix or compute that is non-GPU-ized). Hence, if you
can formulate your input script to only use GPU-ized fixes and
computes, and avoid doing I/O too often (thermo output, dump file
snapshots, restart files), then the data transfer cost of the
USER-CUDA package can be very low, causing it to run faster than the
GPU package.</li>
<li>The GPU package is often faster than the USER-CUDA package, if the
number of atoms per GPU is smaller. The crossover point, in terms of
atoms/GPU at which the USER-CUDA package becomes faster depends
strongly on the pair style. For example, for a simple Lennard Jones
system the crossover (in single precision) is often about 50K-100K
atoms per GPU. When performing double precision calculations the
crossover point can be significantly smaller.</li>
<li>Both packages compute bonded interactions (bonds, angles, etc) on the
CPU. This means a model with bonds will force the USER-CUDA package
to transfer per-atom data back-and-forth between the CPU and GPU every
timestep. If the GPU package is running with several MPI processes
assigned to one GPU, the cost of computing the bonded interactions is
spread across more CPUs and hence the GPU package can run faster.</li>
<li>When using the GPU package with multiple CPUs assigned to one GPU, its
performance depends to some extent on high bandwidth between the CPUs
and the GPU. Hence its performance is affected if full 16 PCIe lanes
are not available for each GPU. In HPC environments this can be the
case if S2050/70 servers are used, where two devices generally share
one PCIe 2.0 16x slot. Also many multi-GPU mainboards do not provide
full 16 lanes to each of the PCIe 2.0 16x slots.</li>
</ul>
<p><strong>Differences between the two packages:</strong></p>
<ul class="simple">
<li>The GPU package accelerates only pair force, neighbor list, and PPPM
calculations. The USER-CUDA package currently supports a wider range
of pair styles and can also accelerate many fix styles and some
compute styles, as well as neighbor list and PPPM calculations.</li>
<li>The USER-CUDA package does not support acceleration for minimization.</li>
<li>The USER-CUDA package does not support hybrid pair styles.</li>
<li>The USER-CUDA package can order atoms in the neighbor list differently
from run to run resulting in a different order for force accumulation.</li>
<li>The USER-CUDA package has a limit on the number of atom types that can be
used in a simulation.</li>
<li>The GPU package requires neighbor lists to be built on the CPU when using
exclusion lists or a triclinic simulation box.</li>
<li>The GPU package uses more GPU memory than the USER-CUDA package. This
is generally not a problem since typical runs are computation-limited
rather than memory-limited.</li>
</ul>
<div class="section" id="examples">
<h3>5.4.1. Examples<a class="headerlink" href="#examples" title="Permalink to this headline"></a></h3>
<p>The LAMMPS distribution has two directories with sample input scripts
for the GPU and USER-CUDA packages.</p>
<ul class="simple">
<li>lammps/examples/gpu = GPU package files</li>
<li>lammps/examples/USER/cuda = USER-CUDA package files</li>
</ul>
<p>These contain input scripts for identical systems, so they can be used
to benchmark the performance of both packages on your system.</p>
</div>
</div>
</div>
</div>
</div>
<footer>
<div class="rst-footer-buttons" role="navigation" aria-label="footer navigation">
<a href="Section_howto.html" class="btn btn-neutral float-right" title="6. How-to discussions" accesskey="n">Next <span class="fa fa-arrow-circle-right"></span></a>
<a href="Section_packages.html" class="btn btn-neutral" title="4. Packages" accesskey="p"><span class="fa fa-arrow-circle-left"></span> Previous</a>
</div>
<hr/>
<div role="contentinfo">
<p>
&copy; Copyright .
</p>
</div>
Built with <a href="http://sphinx-doc.org/">Sphinx</a> using a <a href="https://github.com/snide/sphinx_rtd_theme">theme</a> provided by <a href="https://readthedocs.org">Read the Docs</a>.
</footer>
</div>
</div>
</section>
</div>
<script type="text/javascript">
var DOCUMENTATION_OPTIONS = {
URL_ROOT:'./',
VERSION:'15 May 2015 version',
COLLAPSE_INDEX:false,
FILE_SUFFIX:'.html',
HAS_SOURCE: true
};
</script>
<script type="text/javascript" src="_static/jquery.js"></script>
<script type="text/javascript" src="_static/underscore.js"></script>
<script type="text/javascript" src="_static/doctools.js"></script>
<script type="text/javascript" src="https://cdn.mathjax.org/mathjax/latest/MathJax.js?config=TeX-AMS-MML_HTMLorMML"></script>
<script type="text/javascript" src="_static/sphinxcontrib-images/LightBox2/lightbox2/js/jquery-1.11.0.min.js"></script>
<script type="text/javascript" src="_static/sphinxcontrib-images/LightBox2/lightbox2/js/lightbox.min.js"></script>
<script type="text/javascript" src="_static/sphinxcontrib-images/LightBox2/lightbox2-customize/jquery-noconflict.js"></script>
<script type="text/javascript" src="_static/js/theme.js"></script>
<script type="text/javascript">
jQuery(function () {
SphinxRtdTheme.StickyNav.enable();
});
</script>
</body>
</html>

View File

@ -1,415 +0,0 @@
"Previous Section"_Section_packages.html - "LAMMPS WWW Site"_lws -
"LAMMPS Documentation"_ld - "LAMMPS Commands"_lc - "Next
Section"_Section_howto.html :c
:link(lws,http://lammps.sandia.gov)
:link(ld,Manual.html)
:link(lc,Section_commands.html#comm)
:line
5. Accelerating LAMMPS performance :h3
This section describes various methods for improving LAMMPS
performance for different classes of problems running on different
kinds of machines.
There are two thrusts to the discussion that follows. The
first is using code options that implement alternate algorithms
that can speed-up a simulation. The second is to use one
of the several accelerator packages provided with LAMMPS that
contain code optimized for certain kinds of hardware, including
multi-core CPUs, GPUs, and Intel Xeon Phi coprocessors.
5.1 "Measuring performance"_#acc_1 :ulb,l
5.2 "Algorithms and code options to boost performace"_#acc_2 :l
5.3 "Accelerator packages with optimized styles"_#acc_3 :l
5.3.1 "USER-CUDA package"_accelerate_cuda.html :ulb,l
5.3.2 "GPU package"_accelerate_gpu.html :l
5.3.3 "USER-INTEL package"_accelerate_intel.html :l
5.3.4 "KOKKOS package"_accelerate_kokkos.html :l
5.3.5 "USER-OMP package"_accelerate_omp.html :l
5.3.6 "OPT package"_accelerate_opt.html :l,ule
5.4 "Comparison of various accelerator packages"_#acc_4 :l,ule
The "Benchmark page"_http://lammps.sandia.gov/bench.html of the LAMMPS
web site gives performance results for the various accelerator
packages discussed in Section 5.2, for several of the standard LAMMPS
benchmark problems, as a function of problem size and number of
compute nodes, on different hardware platforms.
:line
:line
5.1 Measuring performance :h4,link(acc_1)
Before trying to make your simulation run faster, you should
understand how it currently performs and where the bottlenecks are.
The best way to do this is run the your system (actual number of
atoms) for a modest number of timesteps (say 100 steps) on several
different processor counts, including a single processor if possible.
Do this for an equilibrium version of your system, so that the
100-step timings are representative of a much longer run. There is
typically no need to run for 1000s of timesteps to get accurate
timings; you can simply extrapolate from short runs.
For the set of runs, look at the timing data printed to the screen and
log file at the end of each LAMMPS run. "This
section"_Section_start.html#start_8 of the manual has an overview.
Running on one (or a few processors) should give a good estimate of
the serial performance and what portions of the timestep are taking
the most time. Running the same problem on a few different processor
counts should give an estimate of parallel scalability. I.e. if the
simulation runs 16x faster on 16 processors, its 100% parallel
efficient; if it runs 8x faster on 16 processors, it's 50% efficient.
The most important data to look at in the timing info is the timing
breakdown and relative percentages. For example, trying different
options for speeding up the long-range solvers will have little impact
if they only consume 10% of the run time. If the pairwise time is
dominating, you may want to look at GPU or OMP versions of the pair
style, as discussed below. Comparing how the percentages change as
you increase the processor count gives you a sense of how different
operations within the timestep are scaling. Note that if you are
running with a Kspace solver, there is additional output on the
breakdown of the Kspace time. For PPPM, this includes the fraction
spent on FFTs, which can be communication intensive.
Another important detail in the timing info are the histograms of
atoms counts and neighbor counts. If these vary widely across
processors, you have a load-imbalance issue. This often results in
inaccurate relative timing data, because processors have to wait when
communication occurs for other processors to catch up. Thus the
reported times for "Communication" or "Other" may be higher than they
really are, due to load-imbalance. If this is an issue, you can
uncomment the MPI_Barrier() lines in src/timer.cpp, and recompile
LAMMPS, to obtain synchronized timings.
:line
5.2 General strategies :h4,link(acc_2)
NOTE: this section 5.2 is still a work in progress
Here is a list of general ideas for improving simulation performance.
Most of them are only applicable to certain models and certain
bottlenecks in the current performance, so let the timing data you
generate be your guide. It is hard, if not impossible, to predict how
much difference these options will make, since it is a function of
problem size, number of processors used, and your machine. There is
no substitute for identifying performance bottlenecks, and trying out
various options.
rRESPA
2-FFT PPPM
Staggered PPPM
single vs double PPPM
partial charge PPPM
verlet/split run style
processor command for proc layout and numa layout
load-balancing: balance and fix balance :ul
2-FFT PPPM, also called {analytic differentiation} or {ad} PPPM, uses
2 FFTs instead of the 4 FFTs used by the default {ik differentiation}
PPPM. However, 2-FFT PPPM also requires a slightly larger mesh size to
achieve the same accuracy as 4-FFT PPPM. For problems where the FFT
cost is the performance bottleneck (typically large problems running
on many processors), 2-FFT PPPM may be faster than 4-FFT PPPM.
Staggered PPPM performs calculations using two different meshes, one
shifted slightly with respect to the other. This can reduce force
aliasing errors and increase the accuracy of the method, but also
doubles the amount of work required. For high relative accuracy, using
staggered PPPM allows one to half the mesh size in each dimension as
compared to regular PPPM, which can give around a 4x speedup in the
kspace time. However, for low relative accuracy, using staggered PPPM
gives little benefit and can be up to 2x slower in the kspace
time. For example, the rhodopsin benchmark was run on a single
processor, and results for kspace time vs. relative accuracy for the
different methods are shown in the figure below. For this system,
staggered PPPM (using ik differentiation) becomes useful when using a
relative accuracy of slightly greater than 1e-5 and above.
:c,image(JPG/rhodo_staggered.jpg)
IMPORTANT NOTE: Using staggered PPPM may not give the same increase in
accuracy of energy and pressure as it does in forces, so some caution
must be used if energy and/or pressure are quantities of interest,
such as when using a barostat.
:line
5.3 Packages with optimized styles :h4,link(acc_3)
Accelerated versions of various "pair_style"_pair_style.html,
"fixes"_fix.html, "computes"_compute.html, and other commands have
been added to LAMMPS, which will typically run faster than the
standard non-accelerated versions. Some require appropriate hardware
to be present on your system, e.g. GPUs or Intel Xeon Phi
coprocessors.
All of these commands are in packages provided with LAMMPS. An
overview of packages is give in "Section
packages"_Section_packages.html.
These are the accelerator packages
currently in LAMMPS, either as standard or user packages:
"USER-CUDA"_accelerate_cuda.html : for NVIDIA GPUs
"GPU"_accelerate_gpu.html : for NVIDIA GPUs as well as OpenCL support
"USER-INTEL"_accelerate_intel.html : for Intel CPUs and Intel Xeon Phi
"KOKKOS"_accelerate_kokkos.html : for GPUs, Intel Xeon Phi, and OpenMP threading
"USER-OMP"_accelerate_omp.html : for OpenMP threading
"OPT"_accelerate_opt.html : generic CPU optimizations :tb(s=:)
Inverting this list, LAMMPS currently has acceleration support for
three kinds of hardware, via the listed packages:
Many-core CPUs : "USER-INTEL"_accelerate_intel.html, "KOKKOS"_accelerate_kokkos.html, "USER-OMP"_accelerate_omp.html, "OPT"_accelerate_opt.html packages
NVIDIA GPUs : "USER-CUDA"_accelerate_cuda.html, "GPU"_accelerate_gpu.html, "KOKKOS"_accelerate_kokkos.html packages
Intel Phi : "USER-INTEL"_accelerate_intel.html, "KOKKOS"_accelerate_kokkos.html packages :tb(s=:)
Which package is fastest for your hardware may depend on the size
problem you are running and what commands (accelerated and
non-accelerated) are invoked by your input script. While these doc
pages include performance guidelines, there is no substitute for
trying out the different packages appropriate to your hardware.
Any accelerated style has the same name as the corresponding standard
style, except that a suffix is appended. Otherwise, the syntax for
the command that uses the style is identical, their functionality is
the same, and the numerical results it produces should also be the
same, except for precision and round-off effects.
For example, all of these styles are accelerated variants of the
Lennard-Jones "pair_style lj/cut"_pair_lj.html:
"pair_style lj/cut/cuda"_pair_lj.html
"pair_style lj/cut/gpu"_pair_lj.html
"pair_style lj/cut/intel"_pair_lj.html
"pair_style lj/cut/kk"_pair_lj.html
"pair_style lj/cut/omp"_pair_lj.html
"pair_style lj/cut/opt"_pair_lj.html :ul
To see what accelerate styles are currently available, see
"Section_commands 5"_Section_commands.html#cmd_5 of the manual. The
doc pages for individual commands (e.g. "pair lj/cut"_pair_lj.html or
"fix nve"_fix_nve.html) also list any accelerated variants available
for that style.
To use an accelerator package in LAMMPS, and one or more of the styles
it provides, follow these general steps. Details vary from package to
package and are explained in the individual accelerator doc pages,
listed above:
build the accelerator library |
only for USER-CUDA and GPU packages |
install the accelerator package |
make yes-opt, make yes-user-intel, etc |
add compile/link flags to Makefile.machine |
in src/MAKE, <br>
only for USER-INTEL, KOKKOS, USER-OMP, OPT packages |
re-build LAMMPS |
make machine |
run a LAMMPS simulation |
lmp_machine < in.script <br>
mpirun -np 32 lmp_machine -in in.script |
enable the accelerator package |
via "-c on" and "-k on" "command-line switches"_Section_start.html#start_7, <br>
only for USER-CUDA and KOKKOS packages |
set any needed options for the package |
via "-pk" "command-line switch"_Section_start.html#start_7 or
"package"_package.html command, <br>
only if defaults need to be changed |
use accelerated styles in your input script |
via "-sf" "command-line switch"_Section_start.html#start_7 or
"suffix"_suffix.html command :tb(c=2,s=|)
Note that the first 4 steps can be done as a single command, using the
src/Make.py tool. This tool is discussed in "Section
2.4"_Section_start.html#start_4 of the manual, and its use is
illustrated in the individual accelerator sections. Typically these
steps only need to be done once, to create an executable that uses one
or more accelerator packages.
The last 4 steps can all be done from the command-line when LAMMPS is
launched, without changing your input script, as illustrated in the
individual accelerator sections. Or you can add
"package"_package.html and "suffix"_suffix.html commands to your input
script.
IMPORTANT NOTE: With a few exceptions, you can build a single LAMMPS
executable with all its accelerator packages installed. Note however
that the USER-INTEL and KOKKOS packages require you to choose one of
their hardware options when building for a specific platform.
I.e. CPU or Phi option for the USER-INTEL package. Or the OpenMP,
Cuda, or Phi option for the KOKKOS package.
These are the exceptions. You cannot build a single executable with:
both the USER-INTEL Phi and KOKKOS Phi options
the USER-INTEL Phi or Kokkos Phi option, and either the USER-CUDA or GPU packages :ul
See the examples/accelerate/README and make.list files for sample
Make.py commands that build LAMMPS with any or all of the accelerator
packages. As an example, here is a command that builds with all the
GPU related packages installed (USER-CUDA, GPU, KOKKOS with Cuda),
including settings to build the needed auxiliary USER-CUDA and GPU
libraries for Kepler GPUs:
Make.py -j 16 -p omp gpu cuda kokkos -cc nvcc wrap=mpi \
-cuda mode=double arch=35 -gpu mode=double arch=35 \\
-kokkos cuda arch=35 lib-all file mpi :pre
The examples/accelerate directory also has input scripts that can be
used with all of the accelerator packages. See its README file for
details.
Likewise, the bench directory has FERMI and KEPLER and PHI
sub-directories with Make.py commands and input scripts for using all
the accelerator packages on various machines. See the README files in
those dirs.
As mentioned above, the "Benchmark
page"_http://lammps.sandia.gov/bench.html of the LAMMPS web site gives
performance results for the various accelerator packages for several
of the standard LAMMPS benchmark problems, as a function of problem
size and number of compute nodes, on different hardware platforms.
Here is a brief summary of what the various packages provide. Details
are in the individual accelerator sections.
Styles with a "cuda" or "gpu" suffix are part of the USER-CUDA or GPU
packages, and can be run on NVIDIA GPUs. The speed-up on a GPU
depends on a variety of factors, discussed in the accelerator
sections. :ulb,l
Styles with an "intel" suffix are part of the USER-INTEL
package. These styles support vectorized single and mixed precision
calculations, in addition to full double precision. In extreme cases,
this can provide speedups over 3.5x on CPUs. The package also
supports acceleration in "offload" mode to Intel(R) Xeon Phi(TM)
coprocessors. This can result in additional speedup over 2x depending
on the hardware configuration. :l
Styles with a "kk" suffix are part of the KOKKOS package, and can be
run using OpenMP on multicore CPUs, on an NVIDIA GPU, or on an Intel
Xeon Phi in "native" mode. The speed-up depends on a variety of
factors, as discussed on the KOKKOS accelerator page. :l
Styles with an "omp" suffix are part of the USER-OMP package and allow
a pair-style to be run in multi-threaded mode using OpenMP. This can
be useful on nodes with high-core counts when using less MPI processes
than cores is advantageous, e.g. when running with PPPM so that FFTs
are run on fewer MPI processors or when the many MPI tasks would
overload the available bandwidth for communication. :l
Styles with an "opt" suffix are part of the OPT package and typically
speed-up the pairwise calculations of your simulation by 5-25% on a
CPU. :l,ule
The individual accelerator package doc pages explain:
what hardware and software the accelerated package requires
how to build LAMMPS with the accelerated package
how to run with the accelerated package either via command-line switches or modifying the input script
speed-ups to expect
guidelines for best performance
restrictions :ul
:line
5.4 Comparison of various accelerator packages :h4,link(acc_4)
NOTE: this section still needs to be re-worked with additional KOKKOS
and USER-INTEL information.
The next section compares and contrasts the various accelerator
options, since there are multiple ways to perform OpenMP threading,
run on GPUs, and run on Intel Xeon Phi coprocessors.
All 3 of these packages accelerate a LAMMPS calculation using NVIDIA
hardware, but they do it in different ways.
As a consequence, for a particular simulation on specific hardware,
one package may be faster than the other. We give guidelines below,
but the best way to determine which package is faster for your input
script is to try both of them on your machine. See the benchmarking
section below for examples where this has been done.
[Guidelines for using each package optimally:]
The GPU package allows you to assign multiple CPUs (cores) to a single
GPU (a common configuration for "hybrid" nodes that contain multicore
CPU(s) and GPU(s)) and works effectively in this mode. The USER-CUDA
package does not allow this; you can only use one CPU per GPU. :ulb,l
The GPU package moves per-atom data (coordinates, forces)
back-and-forth between the CPU and GPU every timestep. The USER-CUDA
package only does this on timesteps when a CPU calculation is required
(e.g. to invoke a fix or compute that is non-GPU-ized). Hence, if you
can formulate your input script to only use GPU-ized fixes and
computes, and avoid doing I/O too often (thermo output, dump file
snapshots, restart files), then the data transfer cost of the
USER-CUDA package can be very low, causing it to run faster than the
GPU package. :l
The GPU package is often faster than the USER-CUDA package, if the
number of atoms per GPU is smaller. The crossover point, in terms of
atoms/GPU at which the USER-CUDA package becomes faster depends
strongly on the pair style. For example, for a simple Lennard Jones
system the crossover (in single precision) is often about 50K-100K
atoms per GPU. When performing double precision calculations the
crossover point can be significantly smaller. :l
Both packages compute bonded interactions (bonds, angles, etc) on the
CPU. This means a model with bonds will force the USER-CUDA package
to transfer per-atom data back-and-forth between the CPU and GPU every
timestep. If the GPU package is running with several MPI processes
assigned to one GPU, the cost of computing the bonded interactions is
spread across more CPUs and hence the GPU package can run faster. :l
When using the GPU package with multiple CPUs assigned to one GPU, its
performance depends to some extent on high bandwidth between the CPUs
and the GPU. Hence its performance is affected if full 16 PCIe lanes
are not available for each GPU. In HPC environments this can be the
case if S2050/70 servers are used, where two devices generally share
one PCIe 2.0 16x slot. Also many multi-GPU mainboards do not provide
full 16 lanes to each of the PCIe 2.0 16x slots. :l,ule
[Differences between the two packages:]
The GPU package accelerates only pair force, neighbor list, and PPPM
calculations. The USER-CUDA package currently supports a wider range
of pair styles and can also accelerate many fix styles and some
compute styles, as well as neighbor list and PPPM calculations. :ulb,l
The USER-CUDA package does not support acceleration for minimization. :l
The USER-CUDA package does not support hybrid pair styles. :l
The USER-CUDA package can order atoms in the neighbor list differently
from run to run resulting in a different order for force accumulation. :l
The USER-CUDA package has a limit on the number of atom types that can be
used in a simulation. :l
The GPU package requires neighbor lists to be built on the CPU when using
exclusion lists or a triclinic simulation box. :l
The GPU package uses more GPU memory than the USER-CUDA package. This
is generally not a problem since typical runs are computation-limited
rather than memory-limited. :l,ule
[Examples:]
The LAMMPS distribution has two directories with sample input scripts
for the GPU and USER-CUDA packages.
lammps/examples/gpu = GPU package files
lammps/examples/USER/cuda = USER-CUDA package files :ul
These contain input scripts for identical systems, so they can be used
to benchmark the performance of both packages on your system.

File diff suppressed because it is too large Load Diff

File diff suppressed because it is too large Load Diff

View File

@ -1,401 +0,0 @@
<!DOCTYPE html>
<!--[if IE 8]><html class="no-js lt-ie9" lang="en" > <![endif]-->
<!--[if gt IE 8]><!--> <html class="no-js" lang="en" > <!--<![endif]-->
<head>
<meta charset="utf-8">
<meta name="viewport" content="width=device-width, initial-scale=1.0">
<title>7. Example problems &mdash; LAMMPS 15 May 2015 version documentation</title>
<link rel="stylesheet" href="_static/css/theme.css" type="text/css" />
<link rel="stylesheet" href="_static/sphinxcontrib-images/LightBox2/lightbox2/css/lightbox.css" type="text/css" />
<link rel="top" title="LAMMPS 15 May 2015 version documentation" href="index.html"/>
<link rel="next" title="8. Performance &amp; scalability" href="Section_perf.html"/>
<link rel="prev" title="6. How-to discussions" href="Section_howto.html"/>
<script src="_static/js/modernizr.min.js"></script>
</head>
<body class="wy-body-for-nav" role="document">
<div class="wy-grid-for-nav">
<nav data-toggle="wy-nav-shift" class="wy-nav-side">
<div class="wy-side-nav-search">
<a href="Manual.html" class="icon icon-home"> LAMMPS
</a>
<div role="search">
<form id="rtd-search-form" class="wy-form" action="search.html" method="get">
<input type="text" name="q" placeholder="Search docs" />
<input type="hidden" name="check_keywords" value="yes" />
<input type="hidden" name="area" value="default" />
</form>
</div>
</div>
<div class="wy-menu wy-menu-vertical" data-spy="affix" role="navigation" aria-label="main navigation">
<ul class="current">
<li class="toctree-l1"><a class="reference internal" href="Section_intro.html">1. Introduction</a></li>
<li class="toctree-l1"><a class="reference internal" href="Section_start.html">2. Getting Started</a></li>
<li class="toctree-l1"><a class="reference internal" href="Section_commands.html">3. Commands</a></li>
<li class="toctree-l1"><a class="reference internal" href="Section_packages.html">4. Packages</a></li>
<li class="toctree-l1"><a class="reference internal" href="Section_accelerate.html">5. Accelerating LAMMPS performance</a></li>
<li class="toctree-l1"><a class="reference internal" href="Section_howto.html">6. How-to discussions</a></li>
<li class="toctree-l1 current"><a class="current reference internal" href="">7. Example problems</a></li>
<li class="toctree-l1"><a class="reference internal" href="Section_perf.html">8. Performance &amp; scalability</a></li>
<li class="toctree-l1"><a class="reference internal" href="Section_tools.html">9. Additional tools</a></li>
<li class="toctree-l1"><a class="reference internal" href="Section_modify.html">10. Modifying &amp; extending LAMMPS</a></li>
<li class="toctree-l1"><a class="reference internal" href="Section_python.html">11. Python interface to LAMMPS</a></li>
<li class="toctree-l1"><a class="reference internal" href="Section_errors.html">12. Errors</a></li>
<li class="toctree-l1"><a class="reference internal" href="Section_history.html">13. Future and history</a></li>
</ul>
</div>
&nbsp;
</nav>
<section data-toggle="wy-nav-shift" class="wy-nav-content-wrap">
<nav class="wy-nav-top" role="navigation" aria-label="top navigation">
<i data-toggle="wy-nav-top" class="fa fa-bars"></i>
<a href="Manual.html">LAMMPS</a>
</nav>
<div class="wy-nav-content">
<div class="rst-content">
<div role="navigation" aria-label="breadcrumbs navigation">
<ul class="wy-breadcrumbs">
<li><a href="Manual.html">Docs</a> &raquo;</li>
<li>7. Example problems</li>
<li class="wy-breadcrumbs-aside">
<a href="http://lammps.sandia.gov">Website</a>
<a href="Section_commands.html#comm">Commands</a>
</li>
</ul>
<hr/>
<div class="rst-footer-buttons" style="margin-bottom: 1em" role="navigation" aria-label="footer navigation">
<a href="Section_perf.html" class="btn btn-neutral float-right" title="8. Performance &amp; scalability" accesskey="n">Next <span class="fa fa-arrow-circle-right"></span></a>
<a href="Section_howto.html" class="btn btn-neutral" title="6. How-to discussions" accesskey="p"><span class="fa fa-arrow-circle-left"></span> Previous</a>
</div>
</div>
<div role="main" class="document" itemscope="itemscope" itemtype="http://schema.org/Article">
<div itemprop="articleBody">
<div class="section" id="example-problems">
<h1>7. Example problems<a class="headerlink" href="#example-problems" title="Permalink to this headline"></a></h1>
<p>The LAMMPS distribution includes an examples sub-directory with
several sample problems. Each problem is in a sub-directory of its
own. Most are 2d models so that they run quickly, requiring at most a
couple of minutes to run on a desktop machine. Each problem has an
input script (in.*) and produces a log file (log.*) and dump file
(dump.*) when it runs. Some use a data file (data.*) of initial
coordinates as additional input. A few sample log file outputs on
different machines and different numbers of processors are included in
the directories to compare your answers to. E.g. a log file like
log.crack.foo.P means it ran on P processors of machine &#8220;foo&#8221;.</p>
<p>For examples that use input data files, many of them were produced by
<a class="reference external" href="http://pizza.sandia.gov">Pizza.py</a> or setup tools described in the
<a class="reference internal" href="Section_tools.html"><em>Additional Tools</em></a> section of the LAMMPS
documentation and provided with the LAMMPS distribution.</p>
<p>If you uncomment the <a class="reference internal" href="dump.html"><em>dump</em></a> command in the input script, a
text dump file will be produced, which can be animated by various
<a class="reference external" href="http://lammps.sandia.gov/viz.html">visualization programs</a>. It can
also be animated using the xmovie tool described in the <a class="reference internal" href="Section_tools.html"><em>Additional Tools</em></a> section of the LAMMPS documentation.</p>
<p>If you uncomment the <a class="reference internal" href="dump.html"><em>dump image</em></a> command in the input
script, and assuming you have built LAMMPS with a JPG library, JPG
snapshot images will be produced when the simulation runs. They can
be quickly post-processed into a movie using commands described on the
<a class="reference internal" href="dump_image.html"><em>dump image</em></a> doc page.</p>
<p>Animations of many of these examples can be viewed on the Movies
section of the <a class="reference external" href="http://lammps.sandia.gov">LAMMPS WWW Site</a>.</p>
<p>These are the sample problems in the examples sub-directories:</p>
<table border="1" class="docutils">
<colgroup>
<col width="15%" />
<col width="85%" />
</colgroup>
<tbody valign="top">
<tr class="row-odd"><td>balance</td>
<td>dynamic load balancing, 2d system</td>
</tr>
<tr class="row-even"><td>body</td>
<td>body particles, 2d system</td>
</tr>
<tr class="row-odd"><td>colloid</td>
<td>big colloid particles in a small particle solvent, 2d system</td>
</tr>
<tr class="row-even"><td>comb</td>
<td>models using the COMB potential</td>
</tr>
<tr class="row-odd"><td>crack</td>
<td>crack propagation in a 2d solid</td>
</tr>
<tr class="row-even"><td>cuda</td>
<td>use of the USER-CUDA package for GPU acceleration</td>
</tr>
<tr class="row-odd"><td>dipole</td>
<td>point dipolar particles, 2d system</td>
</tr>
<tr class="row-even"><td>dreiding</td>
<td>methanol via Dreiding FF</td>
</tr>
<tr class="row-odd"><td>eim</td>
<td>NaCl using the EIM potential</td>
</tr>
<tr class="row-even"><td>ellipse</td>
<td>ellipsoidal particles in spherical solvent, 2d system</td>
</tr>
<tr class="row-odd"><td>flow</td>
<td>Couette and Poiseuille flow in a 2d channel</td>
</tr>
<tr class="row-even"><td>friction</td>
<td>frictional contact of spherical asperities between 2d surfaces</td>
</tr>
<tr class="row-odd"><td>gpu</td>
<td>use of the GPU package for GPU acceleration</td>
</tr>
<tr class="row-even"><td>hugoniostat</td>
<td>Hugoniostat shock dynamics</td>
</tr>
<tr class="row-odd"><td>indent</td>
<td>spherical indenter into a 2d solid</td>
</tr>
<tr class="row-even"><td>intel</td>
<td>use of the USER-INTEL package for CPU or Intel(R) Xeon Phi(TM) coprocessor</td>
</tr>
<tr class="row-odd"><td>kim</td>
<td>use of potentials in Knowledge Base for Interatomic Models (KIM)</td>
</tr>
<tr class="row-even"><td>line</td>
<td>line segment particles in 2d rigid bodies</td>
</tr>
<tr class="row-odd"><td>meam</td>
<td>MEAM test for SiC and shear (same as shear examples)</td>
</tr>
<tr class="row-even"><td>melt</td>
<td>rapid melt of 3d LJ system</td>
</tr>
<tr class="row-odd"><td>micelle</td>
<td>self-assembly of small lipid-like molecules into 2d bilayers</td>
</tr>
<tr class="row-even"><td>min</td>
<td>energy minimization of 2d LJ melt</td>
</tr>
<tr class="row-odd"><td>msst</td>
<td>MSST shock dynamics</td>
</tr>
<tr class="row-even"><td>nb3b</td>
<td>use of nonbonded 3-body harmonic pair style</td>
</tr>
<tr class="row-odd"><td>neb</td>
<td>nudged elastic band (NEB) calculation for barrier finding</td>
</tr>
<tr class="row-even"><td>nemd</td>
<td>non-equilibrium MD of 2d sheared system</td>
</tr>
<tr class="row-odd"><td>obstacle</td>
<td>flow around two voids in a 2d channel</td>
</tr>
<tr class="row-even"><td>peptide</td>
<td>dynamics of a small solvated peptide chain (5-mer)</td>
</tr>
<tr class="row-odd"><td>peri</td>
<td>Peridynamic model of cylinder impacted by indenter</td>
</tr>
<tr class="row-even"><td>pour</td>
<td>pouring of granular particles into a 3d box, then chute flow</td>
</tr>
<tr class="row-odd"><td>prd</td>
<td>parallel replica dynamics of vacancy diffusion in bulk Si</td>
</tr>
<tr class="row-even"><td>qeq</td>
<td>use of the QEQ pacakge for charge equilibration</td>
</tr>
<tr class="row-odd"><td>reax</td>
<td>RDX and TATB models using the ReaxFF</td>
</tr>
<tr class="row-even"><td>rigid</td>
<td>rigid bodies modeled as independent or coupled</td>
</tr>
<tr class="row-odd"><td>shear</td>
<td>sideways shear applied to 2d solid, with and without a void</td>
</tr>
<tr class="row-even"><td>snap</td>
<td>NVE dynamics for BCC tantalum crystal using SNAP potential</td>
</tr>
<tr class="row-odd"><td>srd</td>
<td>stochastic rotation dynamics (SRD) particles as solvent</td>
</tr>
<tr class="row-even"><td>tad</td>
<td>temperature-accelerated dynamics of vacancy diffusion in bulk Si</td>
</tr>
<tr class="row-odd"><td>tri</td>
<td>triangular particles in rigid bodies</td>
</tr>
</tbody>
</table>
<p>vashishta: models using the Vashishta potential</p>
<p>Here is how you might run and visualize one of the sample problems:</p>
<div class="highlight-python"><div class="highlight"><pre>cd indent
cp ../../src/lmp_linux . # copy LAMMPS executable to this dir
lmp_linux -in in.indent # run the problem
</pre></div>
</div>
<p>Running the simulation produces the files <em>dump.indent</em> and
<em>log.lammps</em>. You can visualize the dump file as follows:</p>
<div class="highlight-python"><div class="highlight"><pre>../../tools/xmovie/xmovie -scale dump.indent
</pre></div>
</div>
<p>If you uncomment the <a class="reference internal" href="dump_image.html"><em>dump image</em></a> line(s) in the input
script a series of JPG images will be produced by the run. These can
be viewed individually or turned into a movie or animated by tools
like ImageMagick or QuickTime or various Windows-based tools. See the
<a class="reference internal" href="dump_image.html"><em>dump image</em></a> doc page for more details. E.g. this
Imagemagick command would create a GIF file suitable for viewing in a
browser.</p>
<div class="highlight-python"><div class="highlight"><pre>% convert -loop 1 *.jpg foo.gif
</pre></div>
</div>
<hr class="docutils" />
<p>There is also a COUPLE directory with examples of how to use LAMMPS as
a library, either by itself or in tandem with another code or library.
See the COUPLE/README file to get started.</p>
<p>There is also an ELASTIC directory with an example script for
computing elastic constants at zero temperature, using an Si example. See
the ELASTIC/in.elastic file for more info.</p>
<p>There is also an ELASTIC_T directory with an example script for
computing elastic constants at finite temperature, using an Si example. See
the ELASTIC_T/in.elastic file for more info.</p>
<p>There is also a USER directory which contains subdirectories of
user-provided examples for user packages. See the README files in
those directories for more info. See the
<a class="reference internal" href="Section_start.html"><em>Section_start.html</em></a> file for more info about user
packages.</p>
</div>
</div>
</div>
<footer>
<div class="rst-footer-buttons" role="navigation" aria-label="footer navigation">
<a href="Section_perf.html" class="btn btn-neutral float-right" title="8. Performance &amp; scalability" accesskey="n">Next <span class="fa fa-arrow-circle-right"></span></a>
<a href="Section_howto.html" class="btn btn-neutral" title="6. How-to discussions" accesskey="p"><span class="fa fa-arrow-circle-left"></span> Previous</a>
</div>
<hr/>
<div role="contentinfo">
<p>
&copy; Copyright .
</p>
</div>
Built with <a href="http://sphinx-doc.org/">Sphinx</a> using a <a href="https://github.com/snide/sphinx_rtd_theme">theme</a> provided by <a href="https://readthedocs.org">Read the Docs</a>.
</footer>
</div>
</div>
</section>
</div>
<script type="text/javascript">
var DOCUMENTATION_OPTIONS = {
URL_ROOT:'./',
VERSION:'15 May 2015 version',
COLLAPSE_INDEX:false,
FILE_SUFFIX:'.html',
HAS_SOURCE: true
};
</script>
<script type="text/javascript" src="_static/jquery.js"></script>
<script type="text/javascript" src="_static/underscore.js"></script>
<script type="text/javascript" src="_static/doctools.js"></script>
<script type="text/javascript" src="https://cdn.mathjax.org/mathjax/latest/MathJax.js?config=TeX-AMS-MML_HTMLorMML"></script>
<script type="text/javascript" src="_static/sphinxcontrib-images/LightBox2/lightbox2/js/jquery-1.11.0.min.js"></script>
<script type="text/javascript" src="_static/sphinxcontrib-images/LightBox2/lightbox2/js/lightbox.min.js"></script>
<script type="text/javascript" src="_static/sphinxcontrib-images/LightBox2/lightbox2-customize/jquery-noconflict.js"></script>
<script type="text/javascript" src="_static/js/theme.js"></script>
<script type="text/javascript">
jQuery(function () {
SphinxRtdTheme.StickyNav.enable();
});
</script>
</body>
</html>

View File

@ -1,124 +0,0 @@
"Previous Section"_Section_howto.html - "LAMMPS WWW Site"_lws - "LAMMPS Documentation"_ld - "LAMMPS Commands"_lc - "Next Section"_Section_perf.html :c
:link(lws,http://lammps.sandia.gov)
:link(ld,Manual.html)
:link(lc,Section_commands.html#comm)
:line
7. Example problems :h3
The LAMMPS distribution includes an examples sub-directory with
several sample problems. Each problem is in a sub-directory of its
own. Most are 2d models so that they run quickly, requiring at most a
couple of minutes to run on a desktop machine. Each problem has an
input script (in.*) and produces a log file (log.*) and dump file
(dump.*) when it runs. Some use a data file (data.*) of initial
coordinates as additional input. A few sample log file outputs on
different machines and different numbers of processors are included in
the directories to compare your answers to. E.g. a log file like
log.crack.foo.P means it ran on P processors of machine "foo".
For examples that use input data files, many of them were produced by
"Pizza.py"_http://pizza.sandia.gov or setup tools described in the
"Additional Tools"_Section_tools.html section of the LAMMPS
documentation and provided with the LAMMPS distribution.
If you uncomment the "dump"_dump.html command in the input script, a
text dump file will be produced, which can be animated by various
"visualization programs"_http://lammps.sandia.gov/viz.html. It can
also be animated using the xmovie tool described in the "Additional
Tools"_Section_tools.html section of the LAMMPS documentation.
If you uncomment the "dump image"_dump.html command in the input
script, and assuming you have built LAMMPS with a JPG library, JPG
snapshot images will be produced when the simulation runs. They can
be quickly post-processed into a movie using commands described on the
"dump image"_dump_image.html doc page.
Animations of many of these examples can be viewed on the Movies
section of the "LAMMPS WWW Site"_lws.
These are the sample problems in the examples sub-directories:
balance: dynamic load balancing, 2d system
body: body particles, 2d system
colloid: big colloid particles in a small particle solvent, 2d system
comb: models using the COMB potential
crack: crack propagation in a 2d solid
cuda: use of the USER-CUDA package for GPU acceleration
dipole: point dipolar particles, 2d system
dreiding: methanol via Dreiding FF
eim: NaCl using the EIM potential
ellipse: ellipsoidal particles in spherical solvent, 2d system
flow: Couette and Poiseuille flow in a 2d channel
friction: frictional contact of spherical asperities between 2d surfaces
gpu: use of the GPU package for GPU acceleration
hugoniostat: Hugoniostat shock dynamics
indent: spherical indenter into a 2d solid
intel: use of the USER-INTEL package for CPU or Intel(R) Xeon Phi(TM) coprocessor
kim: use of potentials in Knowledge Base for Interatomic Models (KIM)
line: line segment particles in 2d rigid bodies
meam: MEAM test for SiC and shear (same as shear examples)
melt: rapid melt of 3d LJ system
micelle: self-assembly of small lipid-like molecules into 2d bilayers
min: energy minimization of 2d LJ melt
msst: MSST shock dynamics
nb3b: use of nonbonded 3-body harmonic pair style
neb: nudged elastic band (NEB) calculation for barrier finding
nemd: non-equilibrium MD of 2d sheared system
obstacle: flow around two voids in a 2d channel
peptide: dynamics of a small solvated peptide chain (5-mer)
peri: Peridynamic model of cylinder impacted by indenter
pour: pouring of granular particles into a 3d box, then chute flow
prd: parallel replica dynamics of vacancy diffusion in bulk Si
qeq: use of the QEQ pacakge for charge equilibration
reax: RDX and TATB models using the ReaxFF
rigid: rigid bodies modeled as independent or coupled
shear: sideways shear applied to 2d solid, with and without a void
snap: NVE dynamics for BCC tantalum crystal using SNAP potential
srd: stochastic rotation dynamics (SRD) particles as solvent
tad: temperature-accelerated dynamics of vacancy diffusion in bulk Si
tri: triangular particles in rigid bodies :tb(s=:)
vashishta: models using the Vashishta potential
Here is how you might run and visualize one of the sample problems:
cd indent
cp ../../src/lmp_linux . # copy LAMMPS executable to this dir
lmp_linux -in in.indent # run the problem :pre
Running the simulation produces the files {dump.indent} and
{log.lammps}. You can visualize the dump file as follows:
../../tools/xmovie/xmovie -scale dump.indent :pre
If you uncomment the "dump image"_dump_image.html line(s) in the input
script a series of JPG images will be produced by the run. These can
be viewed individually or turned into a movie or animated by tools
like ImageMagick or QuickTime or various Windows-based tools. See the
"dump image"_dump_image.html doc page for more details. E.g. this
Imagemagick command would create a GIF file suitable for viewing in a
browser.
% convert -loop 1 *.jpg foo.gif :pre
:line
There is also a COUPLE directory with examples of how to use LAMMPS as
a library, either by itself or in tandem with another code or library.
See the COUPLE/README file to get started.
There is also an ELASTIC directory with an example script for
computing elastic constants at zero temperature, using an Si example. See
the ELASTIC/in.elastic file for more info.
There is also an ELASTIC_T directory with an example script for
computing elastic constants at finite temperature, using an Si example. See
the ELASTIC_T/in.elastic file for more info.
There is also a USER directory which contains subdirectories of
user-provided examples for user packages. See the README files in
those directories for more info. See the
"Section_start.html"_Section_start.html file for more info about user
packages.

View File

@ -1,313 +0,0 @@
<!DOCTYPE html>
<!--[if IE 8]><html class="no-js lt-ie9" lang="en" > <![endif]-->
<!--[if gt IE 8]><!--> <html class="no-js" lang="en" > <!--<![endif]-->
<head>
<meta charset="utf-8">
<meta name="viewport" content="width=device-width, initial-scale=1.0">
<title>13. Future and history &mdash; LAMMPS 15 May 2015 version documentation</title>
<link rel="stylesheet" href="_static/css/theme.css" type="text/css" />
<link rel="stylesheet" href="_static/sphinxcontrib-images/LightBox2/lightbox2/css/lightbox.css" type="text/css" />
<link rel="top" title="LAMMPS 15 May 2015 version documentation" href="index.html"/>
<link rel="prev" title="12. Errors" href="Section_errors.html"/>
<script src="_static/js/modernizr.min.js"></script>
</head>
<body class="wy-body-for-nav" role="document">
<div class="wy-grid-for-nav">
<nav data-toggle="wy-nav-shift" class="wy-nav-side">
<div class="wy-side-nav-search">
<a href="Manual.html" class="icon icon-home"> LAMMPS
</a>
<div role="search">
<form id="rtd-search-form" class="wy-form" action="search.html" method="get">
<input type="text" name="q" placeholder="Search docs" />
<input type="hidden" name="check_keywords" value="yes" />
<input type="hidden" name="area" value="default" />
</form>
</div>
</div>
<div class="wy-menu wy-menu-vertical" data-spy="affix" role="navigation" aria-label="main navigation">
<ul class="current">
<li class="toctree-l1"><a class="reference internal" href="Section_intro.html">1. Introduction</a></li>
<li class="toctree-l1"><a class="reference internal" href="Section_start.html">2. Getting Started</a></li>
<li class="toctree-l1"><a class="reference internal" href="Section_commands.html">3. Commands</a></li>
<li class="toctree-l1"><a class="reference internal" href="Section_packages.html">4. Packages</a></li>
<li class="toctree-l1"><a class="reference internal" href="Section_accelerate.html">5. Accelerating LAMMPS performance</a></li>
<li class="toctree-l1"><a class="reference internal" href="Section_howto.html">6. How-to discussions</a></li>
<li class="toctree-l1"><a class="reference internal" href="Section_example.html">7. Example problems</a></li>
<li class="toctree-l1"><a class="reference internal" href="Section_perf.html">8. Performance &amp; scalability</a></li>
<li class="toctree-l1"><a class="reference internal" href="Section_tools.html">9. Additional tools</a></li>
<li class="toctree-l1"><a class="reference internal" href="Section_modify.html">10. Modifying &amp; extending LAMMPS</a></li>
<li class="toctree-l1"><a class="reference internal" href="Section_python.html">11. Python interface to LAMMPS</a></li>
<li class="toctree-l1"><a class="reference internal" href="Section_errors.html">12. Errors</a></li>
<li class="toctree-l1 current"><a class="current reference internal" href="">13. Future and history</a><ul>
<li class="toctree-l2"><a class="reference internal" href="#coming-attractions">13.1. Coming attractions</a></li>
<li class="toctree-l2"><a class="reference internal" href="#past-versions">13.2. Past versions</a></li>
</ul>
</li>
</ul>
</div>
&nbsp;
</nav>
<section data-toggle="wy-nav-shift" class="wy-nav-content-wrap">
<nav class="wy-nav-top" role="navigation" aria-label="top navigation">
<i data-toggle="wy-nav-top" class="fa fa-bars"></i>
<a href="Manual.html">LAMMPS</a>
</nav>
<div class="wy-nav-content">
<div class="rst-content">
<div role="navigation" aria-label="breadcrumbs navigation">
<ul class="wy-breadcrumbs">
<li><a href="Manual.html">Docs</a> &raquo;</li>
<li>13. Future and history</li>
<li class="wy-breadcrumbs-aside">
<a href="http://lammps.sandia.gov">Website</a>
<a href="Section_commands.html#comm">Commands</a>
</li>
</ul>
<hr/>
<div class="rst-footer-buttons" style="margin-bottom: 1em" role="navigation" aria-label="footer navigation">
<a href="Section_errors.html" class="btn btn-neutral" title="12. Errors" accesskey="p"><span class="fa fa-arrow-circle-left"></span> Previous</a>
</div>
</div>
<div role="main" class="document" itemscope="itemscope" itemtype="http://schema.org/Article">
<div itemprop="articleBody">
<div class="section" id="future-and-history">
<h1>13. Future and history<a class="headerlink" href="#future-and-history" title="Permalink to this headline"></a></h1>
<p>This section lists features we plan to add to LAMMPS, features of
previous versions of LAMMPS, and features of other parallel molecular
dynamics codes our group has distributed.</p>
<div class="line-block">
<div class="line">13.1 <a class="reference internal" href="#hist-1"><span>Coming attractions</span></a></div>
<div class="line">13.2 <a class="reference internal" href="#hist-2"><span>Past versions</span></a></div>
<div class="line"><br /></div>
</div>
<div class="section" id="coming-attractions">
<span id="hist-1"></span><h2>13.1. Coming attractions<a class="headerlink" href="#coming-attractions" title="Permalink to this headline"></a></h2>
<p>The <a class="reference external" href="http://lammps.sandia.gov/future.html">Wish list link</a> on the
LAMMPS WWW page gives a list of features we are hoping to add to
LAMMPS in the future, including contact names of individuals you can
email if you are interested in contributing to the developement or
would be a future user of that feature.</p>
<p>You can also send <a class="reference external" href="http://lammps.sandia.gov/authors.html">email to the developers</a> if you want to add
your wish to the list.</p>
<hr class="docutils" />
</div>
<div class="section" id="past-versions">
<span id="hist-2"></span><h2>13.2. Past versions<a class="headerlink" href="#past-versions" title="Permalink to this headline"></a></h2>
<p>LAMMPS development began in the mid 1990s under a cooperative research
&amp; development agreement (CRADA) between two DOE labs (Sandia and LLNL)
and 3 companies (Cray, Bristol Myers Squibb, and Dupont). The goal was
to develop a large-scale parallel classical MD code; the coding effort
was led by Steve Plimpton at Sandia.</p>
<p>After the CRADA ended, a final F77 version, LAMMPS 99, was
released. As development of LAMMPS continued at Sandia, its memory
management was converted to F90; a final F90 version was released as
LAMMPS 2001.</p>
<p>The current LAMMPS is a rewrite in C++ and was first publicly released
as an open source code in 2004. It includes many new features beyond
those in LAMMPS 99 or 2001. It also includes features from older
parallel MD codes written at Sandia, namely ParaDyn, Warp, and
GranFlow (see below).</p>
<p>In late 2006 we began merging new capabilities into LAMMPS that were
developed by Aidan Thompson at Sandia for his MD code GRASP, which has
a parallel framework similar to LAMMPS. Most notably, these have
included many-body potentials - Stillinger-Weber, Tersoff, ReaxFF -
and the associated charge-equilibration routines needed for ReaxFF.</p>
<p>The <a class="reference external" href="http://lammps.sandia.gov/history.html">History link</a> on the
LAMMPS WWW page gives a timeline of features added to the
C++ open-source version of LAMMPS over the last several years.</p>
<p>These older codes are available for download from the <a class="reference external" href="http://lammps.sandia.gov">LAMMPS WWW site</a>, except for Warp &amp; GranFlow which were primarily used
internally. A brief listing of their features is given here.</p>
<p>LAMMPS 2001</p>
<ul class="simple">
<li>F90 + MPI</li>
<li>dynamic memory</li>
<li>spatial-decomposition parallelism</li>
<li>NVE, NVT, NPT, NPH, rRESPA integrators</li>
<li>LJ and Coulombic pairwise force fields</li>
<li>all-atom, united-atom, bead-spring polymer force fields</li>
<li>CHARMM-compatible force fields</li>
<li>class 2 force fields</li>
<li>3d/2d Ewald &amp; PPPM</li>
<li>various force and temperature constraints</li>
<li>SHAKE</li>
<li>Hessian-free truncated-Newton minimizer</li>
<li>user-defined diagnostics</li>
</ul>
<p>LAMMPS 99</p>
<ul class="simple">
<li>F77 + MPI</li>
<li>static memory allocation</li>
<li>spatial-decomposition parallelism</li>
<li>most of the LAMMPS 2001 features with a few exceptions</li>
<li>no 2d Ewald &amp; PPPM</li>
<li>molecular force fields are missing a few CHARMM terms</li>
<li>no SHAKE</li>
</ul>
<p>Warp</p>
<ul class="simple">
<li>F90 + MPI</li>
<li>spatial-decomposition parallelism</li>
<li>embedded atom method (EAM) metal potentials + LJ</li>
<li>lattice and grain-boundary atom creation</li>
<li>NVE, NVT integrators</li>
<li>boundary conditions for applying shear stresses</li>
<li>temperature controls for actively sheared systems</li>
<li>per-atom energy and centro-symmetry computation and output</li>
</ul>
<p>ParaDyn</p>
<ul class="simple">
<li>F77 + MPI</li>
<li>atom- and force-decomposition parallelism</li>
<li>embedded atom method (EAM) metal potentials</li>
<li>lattice atom creation</li>
<li>NVE, NVT, NPT integrators</li>
<li>all serial DYNAMO features for controls and constraints</li>
</ul>
<p>GranFlow</p>
<ul class="simple">
<li>F90 + MPI</li>
<li>spatial-decomposition parallelism</li>
<li>frictional granular potentials</li>
<li>NVE integrator</li>
<li>boundary conditions for granular flow and packing and walls</li>
<li>particle insertion</li>
</ul>
</div>
</div>
</div>
</div>
<footer>
<div class="rst-footer-buttons" role="navigation" aria-label="footer navigation">
<a href="Section_errors.html" class="btn btn-neutral" title="12. Errors" accesskey="p"><span class="fa fa-arrow-circle-left"></span> Previous</a>
</div>
<hr/>
<div role="contentinfo">
<p>
&copy; Copyright .
</p>
</div>
Built with <a href="http://sphinx-doc.org/">Sphinx</a> using a <a href="https://github.com/snide/sphinx_rtd_theme">theme</a> provided by <a href="https://readthedocs.org">Read the Docs</a>.
</footer>
</div>
</div>
</section>
</div>
<script type="text/javascript">
var DOCUMENTATION_OPTIONS = {
URL_ROOT:'./',
VERSION:'15 May 2015 version',
COLLAPSE_INDEX:false,
FILE_SUFFIX:'.html',
HAS_SOURCE: true
};
</script>
<script type="text/javascript" src="_static/jquery.js"></script>
<script type="text/javascript" src="_static/underscore.js"></script>
<script type="text/javascript" src="_static/doctools.js"></script>
<script type="text/javascript" src="https://cdn.mathjax.org/mathjax/latest/MathJax.js?config=TeX-AMS-MML_HTMLorMML"></script>
<script type="text/javascript" src="_static/sphinxcontrib-images/LightBox2/lightbox2/js/jquery-1.11.0.min.js"></script>
<script type="text/javascript" src="_static/sphinxcontrib-images/LightBox2/lightbox2/js/lightbox.min.js"></script>
<script type="text/javascript" src="_static/sphinxcontrib-images/LightBox2/lightbox2-customize/jquery-noconflict.js"></script>
<script type="text/javascript" src="_static/js/theme.js"></script>
<script type="text/javascript">
jQuery(function () {
SphinxRtdTheme.StickyNav.enable();
});
</script>
</body>
</html>

File diff suppressed because it is too large Load Diff

View File

@ -1,695 +0,0 @@
<!DOCTYPE html>
<!--[if IE 8]><html class="no-js lt-ie9" lang="en" > <![endif]-->
<!--[if gt IE 8]><!--> <html class="no-js" lang="en" > <!--<![endif]-->
<head>
<meta charset="utf-8">
<meta name="viewport" content="width=device-width, initial-scale=1.0">
<title>1. Introduction &mdash; LAMMPS 15 May 2015 version documentation</title>
<link rel="stylesheet" href="_static/css/theme.css" type="text/css" />
<link rel="stylesheet" href="_static/sphinxcontrib-images/LightBox2/lightbox2/css/lightbox.css" type="text/css" />
<link rel="top" title="LAMMPS 15 May 2015 version documentation" href="index.html"/>
<link rel="next" title="2. Getting Started" href="Section_start.html"/>
<link rel="prev" title="LAMMPS Documentation" href="Manual.html"/>
<script src="_static/js/modernizr.min.js"></script>
</head>
<body class="wy-body-for-nav" role="document">
<div class="wy-grid-for-nav">
<nav data-toggle="wy-nav-shift" class="wy-nav-side">
<div class="wy-side-nav-search">
<a href="Manual.html" class="icon icon-home"> LAMMPS
</a>
<div role="search">
<form id="rtd-search-form" class="wy-form" action="search.html" method="get">
<input type="text" name="q" placeholder="Search docs" />
<input type="hidden" name="check_keywords" value="yes" />
<input type="hidden" name="area" value="default" />
</form>
</div>
</div>
<div class="wy-menu wy-menu-vertical" data-spy="affix" role="navigation" aria-label="main navigation">
<ul class="current">
<li class="toctree-l1 current"><a class="current reference internal" href="">1. Introduction</a><ul>
<li class="toctree-l2"><a class="reference internal" href="#what-is-lammps">1.1. What is LAMMPS</a></li>
<li class="toctree-l2"><a class="reference internal" href="#lammps-features">1.2. LAMMPS features</a><ul>
<li class="toctree-l3"><a class="reference internal" href="#general-features">1.2.1. General features</a></li>
<li class="toctree-l3"><a class="reference internal" href="#particle-and-model-types">1.2.2. Particle and model types</a></li>
<li class="toctree-l3"><a class="reference internal" href="#force-fields">1.2.3. Force fields</a></li>
<li class="toctree-l3"><a class="reference internal" href="#atom-creation">1.2.4. Atom creation</a></li>
<li class="toctree-l3"><a class="reference internal" href="#ensembles-constraints-and-boundary-conditions">1.2.5. Ensembles, constraints, and boundary conditions</a></li>
<li class="toctree-l3"><a class="reference internal" href="#integrators">1.2.6. Integrators</a></li>
<li class="toctree-l3"><a class="reference internal" href="#diagnostics">1.2.7. Diagnostics</a></li>
<li class="toctree-l3"><a class="reference internal" href="#output">1.2.8. Output</a></li>
<li class="toctree-l3"><a class="reference internal" href="#multi-replica-models">1.2.9. Multi-replica models</a></li>
<li class="toctree-l3"><a class="reference internal" href="#pre-and-post-processing">1.2.10. Pre- and post-processing</a></li>
<li class="toctree-l3"><a class="reference internal" href="#specialized-features">1.2.11. Specialized features</a></li>
</ul>
</li>
<li class="toctree-l2"><a class="reference internal" href="#lammps-non-features">1.3. LAMMPS non-features</a></li>
<li class="toctree-l2"><a class="reference internal" href="#open-source-distribution">1.4. Open source distribution</a></li>
<li class="toctree-l2"><a class="reference internal" href="#acknowledgments-and-citations">1.5. Acknowledgments and citations</a></li>
</ul>
</li>
<li class="toctree-l1"><a class="reference internal" href="Section_start.html">2. Getting Started</a></li>
<li class="toctree-l1"><a class="reference internal" href="Section_commands.html">3. Commands</a></li>
<li class="toctree-l1"><a class="reference internal" href="Section_packages.html">4. Packages</a></li>
<li class="toctree-l1"><a class="reference internal" href="Section_accelerate.html">5. Accelerating LAMMPS performance</a></li>
<li class="toctree-l1"><a class="reference internal" href="Section_howto.html">6. How-to discussions</a></li>
<li class="toctree-l1"><a class="reference internal" href="Section_example.html">7. Example problems</a></li>
<li class="toctree-l1"><a class="reference internal" href="Section_perf.html">8. Performance &amp; scalability</a></li>
<li class="toctree-l1"><a class="reference internal" href="Section_tools.html">9. Additional tools</a></li>
<li class="toctree-l1"><a class="reference internal" href="Section_modify.html">10. Modifying &amp; extending LAMMPS</a></li>
<li class="toctree-l1"><a class="reference internal" href="Section_python.html">11. Python interface to LAMMPS</a></li>
<li class="toctree-l1"><a class="reference internal" href="Section_errors.html">12. Errors</a></li>
<li class="toctree-l1"><a class="reference internal" href="Section_history.html">13. Future and history</a></li>
</ul>
</div>
&nbsp;
</nav>
<section data-toggle="wy-nav-shift" class="wy-nav-content-wrap">
<nav class="wy-nav-top" role="navigation" aria-label="top navigation">
<i data-toggle="wy-nav-top" class="fa fa-bars"></i>
<a href="Manual.html">LAMMPS</a>
</nav>
<div class="wy-nav-content">
<div class="rst-content">
<div role="navigation" aria-label="breadcrumbs navigation">
<ul class="wy-breadcrumbs">
<li><a href="Manual.html">Docs</a> &raquo;</li>
<li>1. Introduction</li>
<li class="wy-breadcrumbs-aside">
<a href="http://lammps.sandia.gov">Website</a>
<a href="Section_commands.html#comm">Commands</a>
</li>
</ul>
<hr/>
<div class="rst-footer-buttons" style="margin-bottom: 1em" role="navigation" aria-label="footer navigation">
<a href="Section_start.html" class="btn btn-neutral float-right" title="2. Getting Started" accesskey="n">Next <span class="fa fa-arrow-circle-right"></span></a>
<a href="Manual.html" class="btn btn-neutral" title="LAMMPS Documentation" accesskey="p"><span class="fa fa-arrow-circle-left"></span> Previous</a>
</div>
</div>
<div role="main" class="document" itemscope="itemscope" itemtype="http://schema.org/Article">
<div itemprop="articleBody">
<div class="section" id="introduction">
<h1>1. Introduction<a class="headerlink" href="#introduction" title="Permalink to this headline"></a></h1>
<p>This section provides an overview of what LAMMPS can and can&#8217;t do,
describes what it means for LAMMPS to be an open-source code, and
acknowledges the funding and people who have contributed to LAMMPS
over the years.</p>
<div class="line-block">
<div class="line">1.1 <a class="reference internal" href="#intro-1"><span>What is LAMMPS</span></a></div>
<div class="line">1.2 <a class="reference internal" href="#intro-2"><span>LAMMPS features</span></a></div>
<div class="line">1.3 <a class="reference internal" href="#intro-3"><span>LAMMPS non-features</span></a></div>
<div class="line">1.4 <a class="reference internal" href="#intro-4"><span>Open source distribution</span></a></div>
<div class="line">1.5 <a class="reference internal" href="#intro-5"><span>Acknowledgments and citations</span></a></div>
<div class="line"><br /></div>
</div>
<div class="section" id="what-is-lammps">
<span id="intro-1"></span><h2>1.1. What is LAMMPS<a class="headerlink" href="#what-is-lammps" title="Permalink to this headline"></a></h2>
<p>LAMMPS is a classical molecular dynamics code that models an ensemble
of particles in a liquid, solid, or gaseous state. It can model
atomic, polymeric, biological, metallic, granular, and coarse-grained
systems using a variety of force fields and boundary conditions.</p>
<p>For examples of LAMMPS simulations, see the Publications page of the
<a class="reference external" href="http://lammps.sandia.gov">LAMMPS WWW Site</a>.</p>
<p>LAMMPS runs efficiently on single-processor desktop or laptop
machines, but is designed for parallel computers. It will run on any
parallel machine that compiles C++ and supports the <a class="reference external" href="http://www-unix.mcs.anl.gov/mpi">MPI</a>
message-passing library. This includes distributed- or shared-memory
parallel machines and Beowulf-style clusters.</p>
<p>LAMMPS can model systems with only a few particles up to millions or
billions. See <a class="reference internal" href="Section_perf.html"><em>Section_perf</em></a> for information on
LAMMPS performance and scalability, or the Benchmarks section of the
<a class="reference external" href="http://lammps.sandia.gov">LAMMPS WWW Site</a>.</p>
<p>LAMMPS is a freely-available open-source code, distributed under the
terms of the <a class="reference external" href="http://www.gnu.org/copyleft/gpl.html">GNU Public License</a>, which means you can use or
modify the code however you wish. See <a class="reference internal" href="#intro-4"><span>this section</span></a> for a
brief discussion of the open-source philosophy.</p>
<p>LAMMPS is designed to be easy to modify or extend with new
capabilities, such as new force fields, atom types, boundary
conditions, or diagnostics. See <a class="reference internal" href="Section_modify.html"><em>Section_modify</em></a>
for more details.</p>
<p>The current version of LAMMPS is written in C++. Earlier versions
were written in F77 and F90. See
<a class="reference internal" href="Section_history.html"><em>Section_history</em></a> for more information on
different versions. All versions can be downloaded from the <a class="reference external" href="http://lammps.sandia.gov">LAMMPS WWW Site</a>.</p>
<p>LAMMPS was originally developed under a US Department of Energy CRADA
(Cooperative Research and Development Agreement) between two DOE labs
and 3 companies. It is distributed by <a class="reference external" href="http://www.sandia.gov">Sandia National Labs</a>.
See <a class="reference internal" href="#intro-5"><span>this section</span></a> for more information on LAMMPS funding and
individuals who have contributed to LAMMPS.</p>
<p>In the most general sense, LAMMPS integrates Newton&#8217;s equations of
motion for collections of atoms, molecules, or macroscopic particles
that interact via short- or long-range forces with a variety of
initial and/or boundary conditions. For computational efficiency
LAMMPS uses neighbor lists to keep track of nearby particles. The
lists are optimized for systems with particles that are repulsive at
short distances, so that the local density of particles never becomes
too large. On parallel machines, LAMMPS uses spatial-decomposition
techniques to partition the simulation domain into small 3d
sub-domains, one of which is assigned to each processor. Processors
communicate and store &#8220;ghost&#8221; atom information for atoms that border
their sub-domain. LAMMPS is most efficient (in a parallel sense) for
systems whose particles fill a 3d rectangular box with roughly uniform
density. Papers with technical details of the algorithms used in
LAMMPS are listed in <a class="reference internal" href="#intro-5"><span>this section</span></a>.</p>
<hr class="docutils" />
</div>
<div class="section" id="lammps-features">
<span id="intro-2"></span><h2>1.2. LAMMPS features<a class="headerlink" href="#lammps-features" title="Permalink to this headline"></a></h2>
<p>This section highlights LAMMPS features, with pointers to specific
commands which give more details. If LAMMPS doesn&#8217;t have your
favorite interatomic potential, boundary condition, or atom type, see
<a class="reference internal" href="Section_modify.html"><em>Section_modify</em></a>, which describes how you can add
it to LAMMPS.</p>
<div class="section" id="general-features">
<h3>1.2.1. General features<a class="headerlink" href="#general-features" title="Permalink to this headline"></a></h3>
<ul class="simple">
<li>runs on a single processor or in parallel</li>
<li>distributed-memory message-passing parallelism (MPI)</li>
<li>spatial-decomposition of simulation domain for parallelism</li>
<li>open-source distribution</li>
<li>highly portable C++</li>
<li>optional libraries used: MPI and single-processor FFT</li>
<li>GPU (CUDA and OpenCL), Intel(R) Xeon Phi(TM) coprocessors, and OpenMP support for many code features</li>
<li>easy to extend with new features and functionality</li>
<li>runs from an input script</li>
<li>syntax for defining and using variables and formulas</li>
<li>syntax for looping over runs and breaking out of loops</li>
<li>run one or multiple simulations simultaneously (in parallel) from one script</li>
<li>build as library, invoke LAMMPS thru library interface or provided Python wrapper</li>
<li>couple with other codes: LAMMPS calls other code, other code calls LAMMPS, umbrella code calls both</li>
</ul>
</div>
<div class="section" id="particle-and-model-types">
<h3>1.2.2. Particle and model types<a class="headerlink" href="#particle-and-model-types" title="Permalink to this headline"></a></h3>
<p>(<a class="reference internal" href="atom_style.html"><em>atom style</em></a> command)</p>
<ul class="simple">
<li>atoms</li>
<li>coarse-grained particles (e.g. bead-spring polymers)</li>
<li>united-atom polymers or organic molecules</li>
<li>all-atom polymers, organic molecules, proteins, DNA</li>
<li>metals</li>
<li>granular materials</li>
<li>coarse-grained mesoscale models</li>
<li>finite-size spherical and ellipsoidal particles</li>
<li>finite-size line segment (2d) and triangle (3d) particles</li>
<li>point dipole particles</li>
<li>rigid collections of particles</li>
<li>hybrid combinations of these</li>
</ul>
</div>
<div class="section" id="force-fields">
<h3>1.2.3. Force fields<a class="headerlink" href="#force-fields" title="Permalink to this headline"></a></h3>
<p>(<a class="reference internal" href="pair_style.html"><em>pair style</em></a>, <a class="reference internal" href="bond_style.html"><em>bond style</em></a>,
<a class="reference internal" href="angle_style.html"><em>angle style</em></a>, <a class="reference internal" href="dihedral_style.html"><em>dihedral style</em></a>,
<a class="reference internal" href="improper_style.html"><em>improper style</em></a>, <a class="reference internal" href="kspace_style.html"><em>kspace style</em></a>
commands)</p>
<ul class="simple">
<li>pairwise potentials: Lennard-Jones, Buckingham, Morse, Born-Mayer-Huggins, Yukawa, soft, class 2 (COMPASS), hydrogen bond, tabulated</li>
<li>charged pairwise potentials: Coulombic, point-dipole</li>
<li>manybody potentials: EAM, Finnis/Sinclair EAM, modified EAM (MEAM), embedded ion method (EIM), EDIP, ADP, Stillinger-Weber, Tersoff, REBO, AIREBO, ReaxFF, COMB, SNAP, Streitz-Mintmire, 3-body polymorphic</li>
<li>long-range interactions for charge, point-dipoles, and LJ dispersion: Ewald, Wolf, PPPM (similar to particle-mesh Ewald)</li>
<li>polarization models: <a class="reference internal" href="fix_qeq.html"><em>QEq</em></a>, <a class="reference internal" href="Section_howto.html#howto-26"><span>core/shell model</span></a>, <a class="reference internal" href="Section_howto.html#howto-27"><span>Drude dipole model</span></a></li>
<li>charge equilibration (QEq via dynamic, point, shielded, Slater methods)</li>
<li>coarse-grained potentials: DPD, GayBerne, REsquared, colloidal, DLVO</li>
<li>mesoscopic potentials: granular, Peridynamics, SPH</li>
<li>electron force field (eFF, AWPMD)</li>
<li>bond potentials: harmonic, FENE, Morse, nonlinear, class 2, quartic (breakable)</li>
<li>angle potentials: harmonic, CHARMM, cosine, cosine/squared, cosine/periodic, class 2 (COMPASS)</li>
<li>dihedral potentials: harmonic, CHARMM, multi-harmonic, helix, class 2 (COMPASS), OPLS</li>
<li>improper potentials: harmonic, cvff, umbrella, class 2 (COMPASS)</li>
<li>polymer potentials: all-atom, united-atom, bead-spring, breakable</li>
<li>water potentials: TIP3P, TIP4P, SPC</li>
<li>implicit solvent potentials: hydrodynamic lubrication, Debye</li>
<li>force-field compatibility with common CHARMM, AMBER, DREIDING, OPLS, GROMACS, COMPASS options</li>
<li>access to <a class="reference external" href="http://openkim.org">KIM archive</a> of potentials via <a class="reference internal" href="pair_kim.html"><em>pair kim</em></a></li>
<li>hybrid potentials: multiple pair, bond, angle, dihedral, improper potentials can be used in one simulation</li>
<li>overlaid potentials: superposition of multiple pair potentials</li>
</ul>
</div>
<div class="section" id="atom-creation">
<h3>1.2.4. Atom creation<a class="headerlink" href="#atom-creation" title="Permalink to this headline"></a></h3>
<p>(<a class="reference internal" href="read_data.html"><em>read_data</em></a>, <a class="reference internal" href="lattice.html"><em>lattice</em></a>,
<a class="reference internal" href="create_atoms.html"><em>create_atoms</em></a>, <a class="reference internal" href="delete_atoms.html"><em>delete_atoms</em></a>,
<a class="reference internal" href="displace_atoms.html"><em>displace_atoms</em></a>, <a class="reference internal" href="replicate.html"><em>replicate</em></a> commands)</p>
<ul class="simple">
<li>read in atom coords from files</li>
<li>create atoms on one or more lattices (e.g. grain boundaries)</li>
<li>delete geometric or logical groups of atoms (e.g. voids)</li>
<li>replicate existing atoms multiple times</li>
<li>displace atoms</li>
</ul>
</div>
<div class="section" id="ensembles-constraints-and-boundary-conditions">
<h3>1.2.5. Ensembles, constraints, and boundary conditions<a class="headerlink" href="#ensembles-constraints-and-boundary-conditions" title="Permalink to this headline"></a></h3>
<p>(<a class="reference internal" href="fix.html"><em>fix</em></a> command)</p>
<ul class="simple">
<li>2d or 3d systems</li>
<li>orthogonal or non-orthogonal (triclinic symmetry) simulation domains</li>
<li>constant NVE, NVT, NPT, NPH, Parinello/Rahman integrators</li>
<li>thermostatting options for groups and geometric regions of atoms</li>
<li>pressure control via Nose/Hoover or Berendsen barostatting in 1 to 3 dimensions</li>
<li>simulation box deformation (tensile and shear)</li>
<li>harmonic (umbrella) constraint forces</li>
<li>rigid body constraints</li>
<li>SHAKE bond and angle constraints</li>
<li>Monte Carlo bond breaking, formation, swapping</li>
<li>atom/molecule insertion and deletion</li>
<li>walls of various kinds</li>
<li>non-equilibrium molecular dynamics (NEMD)</li>
<li>variety of additional boundary conditions and constraints</li>
</ul>
</div>
<div class="section" id="integrators">
<h3>1.2.6. Integrators<a class="headerlink" href="#integrators" title="Permalink to this headline"></a></h3>
<p>(<a class="reference internal" href="run.html"><em>run</em></a>, <a class="reference internal" href="run_style.html"><em>run_style</em></a>, <a class="reference internal" href="minimize.html"><em>minimize</em></a> commands)</p>
<ul class="simple">
<li>velocity-Verlet integrator</li>
<li>Brownian dynamics</li>
<li>rigid body integration</li>
<li>energy minimization via conjugate gradient or steepest descent relaxation</li>
<li>rRESPA hierarchical timestepping</li>
<li>rerun command for post-processing of dump files</li>
</ul>
</div>
<div class="section" id="diagnostics">
<h3>1.2.7. Diagnostics<a class="headerlink" href="#diagnostics" title="Permalink to this headline"></a></h3>
<ul class="simple">
<li>see the various flavors of the <a class="reference internal" href="fix.html"><em>fix</em></a> and <a class="reference internal" href="compute.html"><em>compute</em></a> commands</li>
</ul>
</div>
<div class="section" id="output">
<h3>1.2.8. Output<a class="headerlink" href="#output" title="Permalink to this headline"></a></h3>
<p>(<a class="reference internal" href="dump.html"><em>dump</em></a>, <a class="reference internal" href="restart.html"><em>restart</em></a> commands)</p>
<ul class="simple">
<li>log file of thermodynamic info</li>
<li>text dump files of atom coords, velocities, other per-atom quantities</li>
<li>binary restart files</li>
<li>parallel I/O of dump and restart files</li>
<li>per-atom quantities (energy, stress, centro-symmetry parameter, CNA, etc)</li>
<li>user-defined system-wide (log file) or per-atom (dump file) calculations</li>
<li>spatial and time averaging of per-atom quantities</li>
<li>time averaging of system-wide quantities</li>
<li>atom snapshots in native, XYZ, XTC, DCD, CFG formats</li>
</ul>
</div>
<div class="section" id="multi-replica-models">
<h3>1.2.9. Multi-replica models<a class="headerlink" href="#multi-replica-models" title="Permalink to this headline"></a></h3>
<p><a class="reference internal" href="neb.html"><em>nudged elastic band</em></a>
<a class="reference internal" href="prd.html"><em>parallel replica dynamics</em></a>
<a class="reference internal" href="tad.html"><em>temperature accelerated dynamics</em></a>
<a class="reference internal" href="temper.html"><em>parallel tempering</em></a></p>
</div>
<div class="section" id="pre-and-post-processing">
<h3>1.2.10. Pre- and post-processing<a class="headerlink" href="#pre-and-post-processing" title="Permalink to this headline"></a></h3>
<ul class="simple">
<li>Various pre- and post-processing serial tools are packaged
with LAMMPS; see these <a class="reference internal" href="Section_tools.html"><em>doc pages</em></a>.</li>
<li>Our group has also written and released a separate toolkit called
<a class="reference external" href="http://www.sandia.gov/~sjplimp/pizza.html">Pizza.py</a> which provides tools for doing setup, analysis,
plotting, and visualization for LAMMPS simulations. Pizza.py is
written in <a class="reference external" href="http://www.python.org">Python</a> and is available for download from <a class="reference external" href="http://www.sandia.gov/~sjplimp/pizza.html">the Pizza.py WWW site</a>.</li>
</ul>
</div>
<div class="section" id="specialized-features">
<h3>1.2.11. Specialized features<a class="headerlink" href="#specialized-features" title="Permalink to this headline"></a></h3>
<p>These are LAMMPS capabilities which you may not think of as typical
molecular dynamics options:</p>
<ul class="simple">
<li><a class="reference internal" href="balance.html"><em>static</em></a> and <a class="reference internal" href="fix_balance.html"><em>dynamic load-balancing</em></a></li>
<li><a class="reference internal" href="body.html"><em>generalized aspherical particles</em></a></li>
<li><a class="reference internal" href="fix_srd.html"><em>stochastic rotation dynamics (SRD)</em></a></li>
<li><a class="reference internal" href="fix_imd.html"><em>real-time visualization and interactive MD</em></a></li>
<li>calculate <a class="reference internal" href="compute_xrd.html"><em>virtual diffraction patterns</em></a></li>
<li><a class="reference internal" href="fix_atc.html"><em>atom-to-continuum coupling</em></a> with finite elements</li>
<li>coupled rigid body integration via the <a class="reference internal" href="fix_poems.html"><em>POEMS</em></a> library</li>
<li><a class="reference internal" href="fix_qmmm.html"><em>QM/MM coupling</em></a></li>
<li><a class="reference internal" href="fix_ipi.html"><em>path-integral molecular dynamics (PIMD)</em></a> and <a class="reference internal" href="fix_pimd.html"><em>this as well</em></a></li>
<li>Monte Carlo via <a class="reference internal" href="fix_gcmc.html"><em>GCMC</em></a> and <a class="reference internal" href="fix_tfmc.html"><em>tfMC</em></a> and <code class="xref doc docutils literal"><span class="pre">atom</span> <span class="pre">swapping</span></code></li>
<li><a class="reference internal" href="pair_dsmc.html"><em>Direct Simulation Monte Carlo</em></a> for low-density fluids</li>
<li><a class="reference internal" href="pair_peri.html"><em>Peridynamics mesoscale modeling</em></a></li>
<li><a class="reference internal" href="fix_lb_fluid.html"><em>Lattice Boltzmann fluid</em></a></li>
<li><a class="reference internal" href="fix_tmd.html"><em>targeted</em></a> and <a class="reference internal" href="fix_smd.html"><em>steered</em></a> molecular dynamics</li>
<li><a class="reference internal" href="fix_ttm.html"><em>two-temperature electron model</em></a></li>
</ul>
<hr class="docutils" />
</div>
</div>
<div class="section" id="lammps-non-features">
<span id="intro-3"></span><h2>1.3. LAMMPS non-features<a class="headerlink" href="#lammps-non-features" title="Permalink to this headline"></a></h2>
<p>LAMMPS is designed to efficiently compute Newton&#8217;s equations of motion
for a system of interacting particles. Many of the tools needed to
pre- and post-process the data for such simulations are not included
in the LAMMPS kernel for several reasons:</p>
<ul class="simple">
<li>the desire to keep LAMMPS simple</li>
<li>they are not parallel operations</li>
<li>other codes already do them</li>
<li>limited development resources</li>
</ul>
<p>Specifically, LAMMPS itself does not:</p>
<ul class="simple">
<li>run thru a GUI</li>
<li>build molecular systems</li>
<li>assign force-field coefficients automagically</li>
<li>perform sophisticated analyses of your MD simulation</li>
<li>visualize your MD simulation</li>
<li>plot your output data</li>
</ul>
<p>A few tools for pre- and post-processing tasks are provided as part of
the LAMMPS package; they are described in <a class="reference internal" href="Section_tools.html"><em>this section</em></a>. However, many people use other codes or
write their own tools for these tasks.</p>
<p>As noted above, our group has also written and released a separate
toolkit called <a class="reference external" href="http://www.sandia.gov/~sjplimp/pizza.html">Pizza.py</a> which addresses some of the listed
bullets. It provides tools for doing setup, analysis, plotting, and
visualization for LAMMPS simulations. Pizza.py is written in
<a class="reference external" href="http://www.python.org">Python</a> and is available for download from <a class="reference external" href="http://www.sandia.gov/~sjplimp/pizza.html">the Pizza.py WWW site</a>.</p>
<p>LAMMPS requires as input a list of initial atom coordinates and types,
molecular topology information, and force-field coefficients assigned
to all atoms and bonds. LAMMPS will not build molecular systems and
assign force-field parameters for you.</p>
<p>For atomic systems LAMMPS provides a <a class="reference internal" href="create_atoms.html"><em>create_atoms</em></a>
command which places atoms on solid-state lattices (fcc, bcc,
user-defined, etc). Assigning small numbers of force field
coefficients can be done via the <a class="reference internal" href="pair_coeff.html"><em>pair coeff</em></a>, <a class="reference internal" href="bond_coeff.html"><em>bond coeff</em></a>, <a class="reference internal" href="angle_coeff.html"><em>angle coeff</em></a>, etc commands.
For molecular systems or more complicated simulation geometries, users
typically use another code as a builder and convert its output to
LAMMPS input format, or write their own code to generate atom
coordinate and molecular topology for LAMMPS to read in.</p>
<p>For complicated molecular systems (e.g. a protein), a multitude of
topology information and hundreds of force-field coefficients must
typically be specified. We suggest you use a program like
<a class="reference external" href="http://www.scripps.edu/brooks">CHARMM</a> or <a class="reference external" href="http://amber.scripps.edu">AMBER</a> or other molecular builders to setup
such problems and dump its information to a file. You can then
reformat the file as LAMMPS input. Some of the tools in <a class="reference internal" href="Section_tools.html"><em>this section</em></a> can assist in this process.</p>
<p>Similarly, LAMMPS creates output files in a simple format. Most users
post-process these files with their own analysis tools or re-format
them for input into other programs, including visualization packages.
If you are convinced you need to compute something on-the-fly as
LAMMPS runs, see <a class="reference internal" href="Section_modify.html"><em>Section_modify</em></a> for a discussion
of how you can use the <a class="reference internal" href="dump.html"><em>dump</em></a> and <a class="reference internal" href="compute.html"><em>compute</em></a> and
<a class="reference internal" href="fix.html"><em>fix</em></a> commands to print out data of your choosing. Keep in
mind that complicated computations can slow down the molecular
dynamics timestepping, particularly if the computations are not
parallel, so it is often better to leave such analysis to
post-processing codes.</p>
<p>A very simple (yet fast) visualizer is provided with the LAMMPS
package - see the <a class="reference internal" href="Section_tools.html#xmovie"><span>xmovie</span></a> tool in <a class="reference internal" href="Section_tools.html"><em>this section</em></a>. It creates xyz projection views of
atomic coordinates and animates them. We find it very useful for
debugging purposes. For high-quality visualization we recommend the
following packages:</p>
<ul class="simple">
<li><a class="reference external" href="http://www.ks.uiuc.edu/Research/vmd">VMD</a></li>
<li><a class="reference external" href="http://mt.seas.upenn.edu/Archive/Graphics/A">AtomEye</a></li>
<li><a class="reference external" href="http://pymol.sourceforge.net">PyMol</a></li>
<li><a class="reference external" href="http://www.bmsc.washington.edu/raster3d/raster3d.html">Raster3d</a></li>
<li><a class="reference external" href="http://www.openrasmol.org">RasMol</a></li>
</ul>
<p>Other features that LAMMPS does not yet (and may never) support are
discussed in <a class="reference internal" href="Section_history.html"><em>Section_history</em></a>.</p>
<p>Finally, these are freely-available molecular dynamics codes, most of
them parallel, which may be well-suited to the problems you want to
model. They can also be used in conjunction with LAMMPS to perform
complementary modeling tasks.</p>
<ul class="simple">
<li><a class="reference external" href="http://www.scripps.edu/brooks">CHARMM</a></li>
<li><a class="reference external" href="http://amber.scripps.edu">AMBER</a></li>
<li><a class="reference external" href="http://www.ks.uiuc.edu/Research/namd/">NAMD</a></li>
<li><a class="reference external" href="http://www.emsl.pnl.gov/docs/nwchem/nwchem.html">NWCHEM</a></li>
<li><a class="reference external" href="http://www.cse.clrc.ac.uk/msi/software/DL_POLY">DL_POLY</a></li>
<li><a class="reference external" href="http://dasher.wustl.edu/tinker">Tinker</a></li>
</ul>
<p>CHARMM, AMBER, NAMD, NWCHEM, and Tinker are designed primarily for
modeling biological molecules. CHARMM and AMBER use
atom-decomposition (replicated-data) strategies for parallelism; NAMD
and NWCHEM use spatial-decomposition approaches, similar to LAMMPS.
Tinker is a serial code. DL_POLY includes potentials for a variety of
biological and non-biological materials; both a replicated-data and
spatial-decomposition version exist.</p>
<hr class="docutils" />
</div>
<div class="section" id="open-source-distribution">
<span id="intro-4"></span><h2>1.4. Open source distribution<a class="headerlink" href="#open-source-distribution" title="Permalink to this headline"></a></h2>
<p>LAMMPS comes with no warranty of any kind. As each source file states
in its header, it is a copyrighted code that is distributed free-of-
charge, under the terms of the <a class="reference external" href="http://www.gnu.org/copyleft/gpl.html">GNU Public License</a> (GPL). This
is often referred to as open-source distribution - see
<a class="reference external" href="http://www.gnu.org">www.gnu.org</a> or <a class="reference external" href="http://www.opensource.org">www.opensource.org</a> for more
details. The legal text of the GPL is in the LICENSE file that is
included in the LAMMPS distribution.</p>
<p>Here is a summary of what the GPL means for LAMMPS users:</p>
<p>(1) Anyone is free to use, modify, or extend LAMMPS in any way they
choose, including for commercial purposes.</p>
<p>(2) If you distribute a modified version of LAMMPS, it must remain
open-source, meaning you distribute it under the terms of the GPL.
You should clearly annotate such a code as a derivative version of
LAMMPS.</p>
<p>(3) If you release any code that includes LAMMPS source code, then it
must also be open-sourced, meaning you distribute it under the terms
of the GPL.</p>
<p>(4) If you give LAMMPS files to someone else, the GPL LICENSE file and
source file headers (including the copyright and GPL notices) should
remain part of the code.</p>
<p>In the spirit of an open-source code, these are various ways you can
contribute to making LAMMPS better. You can send email to the
<a class="reference external" href="http://lammps.sandia.gov/authors.html">developers</a> on any of these
items.</p>
<ul class="simple">
<li>Point prospective users to the <a class="reference external" href="http://lammps.sandia.gov">LAMMPS WWW Site</a>. Mention it in
talks or link to it from your WWW site.</li>
<li>If you find an error or omission in this manual or on the <a class="reference external" href="http://lammps.sandia.gov">LAMMPS WWW Site</a>, or have a suggestion for something to clarify or include,
send an email to the
<a class="reference external" href="http://lammps.sandia.gov/authors.html">developers</a>.</li>
<li>If you find a bug, <a class="reference internal" href="Section_errors.html#err-2"><span>Section_errors 2</span></a>
describes how to report it.</li>
<li>If you publish a paper using LAMMPS results, send the citation (and
any cool pictures or movies if you like) to add to the Publications,
Pictures, and Movies pages of the <a class="reference external" href="http://lammps.sandia.gov">LAMMPS WWW Site</a>, with links
and attributions back to you.</li>
<li>Create a new Makefile.machine that can be added to the src/MAKE
directory.</li>
<li>The tools sub-directory of the LAMMPS distribution has various
stand-alone codes for pre- and post-processing of LAMMPS data. More
details are given in <a class="reference internal" href="Section_tools.html"><em>Section_tools</em></a>. If you write
a new tool that users will find useful, it can be added to the LAMMPS
distribution.</li>
<li>LAMMPS is designed to be easy to extend with new code for features
like potentials, boundary conditions, diagnostic computations, etc.
<a class="reference internal" href="Section_modify.html"><em>This section</em></a> gives details. If you add a
feature of general interest, it can be added to the LAMMPS
distribution.</li>
<li>The Benchmark page of the <a class="reference external" href="http://lammps.sandia.gov">LAMMPS WWW Site</a> lists LAMMPS
performance on various platforms. The files needed to run the
benchmarks are part of the LAMMPS distribution. If your machine is
sufficiently different from those listed, your timing data can be
added to the page.</li>
<li>You can send feedback for the User Comments page of the <a class="reference external" href="http://lammps.sandia.gov">LAMMPS WWW Site</a>. It might be added to the page. No promises.</li>
<li>Cash. Small denominations, unmarked bills preferred. Paper sack OK.
Leave on desk. VISA also accepted. Chocolate chip cookies
encouraged.</li>
</ul>
<hr class="docutils" />
</div>
<div class="section" id="acknowledgments-and-citations">
<span id="intro-5"></span><h2>1.5. Acknowledgments and citations<a class="headerlink" href="#acknowledgments-and-citations" title="Permalink to this headline"></a></h2>
<p>LAMMPS development has been funded by the <a class="reference external" href="http://www.doe.gov">US Department of Energy</a> (DOE), through its CRADA, LDRD, ASCI, and Genomes-to-Life
programs and its <a class="reference external" href="http://www.sc.doe.gov/ascr/home.html">OASCR</a> and <a class="reference external" href="http://www.er.doe.gov/production/ober/ober_top.html">OBER</a> offices.</p>
<p>Specifically, work on the latest version was funded in part by the US
Department of Energy&#8217;s Genomics:GTL program
(<a class="reference external" href="http://www.doegenomestolife.org">www.doegenomestolife.org</a>) under the <a class="reference external" href="http://www.genomes2life.org">project</a>, &#8220;Carbon
Sequestration in Synechococcus Sp.: From Molecular Machines to
Hierarchical Modeling&#8221;.</p>
<p>The following paper describe the basic parallel algorithms used in
LAMMPS. If you use LAMMPS results in your published work, please cite
this paper and include a pointer to the <a class="reference external" href="http://lammps.sandia.gov">LAMMPS WWW Site</a>
(<a class="reference external" href="http://lammps.sandia.gov">http://lammps.sandia.gov</a>):</p>
<p>S. J. Plimpton, <strong>Fast Parallel Algorithms for Short-Range Molecular
Dynamics</strong>, J Comp Phys, 117, 1-19 (1995).</p>
<p>Other papers describing specific algorithms used in LAMMPS are listed
under the <a class="reference external" href="http://lammps.sandia.gov/cite.html">Citing LAMMPS link</a> of
the LAMMPS WWW page.</p>
<p>The <a class="reference external" href="http://lammps.sandia.gov/papers.html">Publications link</a> on the
LAMMPS WWW page lists papers that have cited LAMMPS. If your paper is
not listed there for some reason, feel free to send us the info. If
the simulations in your paper produced cool pictures or animations,
we&#8217;ll be pleased to add them to the
<a class="reference external" href="http://lammps.sandia.gov/pictures.html">Pictures</a> or
<a class="reference external" href="http://lammps.sandia.gov/movies.html">Movies</a> pages of the LAMMPS WWW
site.</p>
<p>The core group of LAMMPS developers is at Sandia National Labs:</p>
<ul class="simple">
<li>Steve Plimpton, sjplimp at sandia.gov</li>
<li>Aidan Thompson, athomps at sandia.gov</li>
<li>Paul Crozier, pscrozi at sandia.gov</li>
</ul>
<p>The following folks are responsible for significant contributions to
the code, or other aspects of the LAMMPS development effort. Many of
the packages they have written are somewhat unique to LAMMPS and the
code would not be as general-purpose as it is without their expertise
and efforts.</p>
<ul class="simple">
<li>Axel Kohlmeyer (Temple U), akohlmey at gmail.com, SVN and Git repositories, indefatigable mail list responder, USER-CG-CMM and USER-OMP packages</li>
<li>Roy Pollock (LLNL), Ewald and PPPM solvers</li>
<li>Mike Brown (ORNL), brownw at ornl.gov, GPU package</li>
<li>Greg Wagner (Sandia), gjwagne at sandia.gov, MEAM package for MEAM potential</li>
<li>Mike Parks (Sandia), mlparks at sandia.gov, PERI package for Peridynamics</li>
<li>Rudra Mukherjee (JPL), Rudranarayan.M.Mukherjee at jpl.nasa.gov, POEMS package for articulated rigid body motion</li>
<li>Reese Jones (Sandia) and collaborators, rjones at sandia.gov, USER-ATC package for atom/continuum coupling</li>
<li>Ilya Valuev (JIHT), valuev at physik.hu-berlin.de, USER-AWPMD package for wave-packet MD</li>
<li>Christian Trott (U Tech Ilmenau), christian.trott at tu-ilmenau.de, USER-CUDA package</li>
<li>Andres Jaramillo-Botero (Caltech), ajaramil at wag.caltech.edu, USER-EFF package for electron force field</li>
<li>Christoph Kloss (JKU), Christoph.Kloss at jku.at, USER-LIGGGHTS package for granular models and granular/fluid coupling</li>
<li>Metin Aktulga (LBL), hmaktulga at lbl.gov, USER-REAXC package for C version of ReaxFF</li>
<li>Georg Gunzenmuller (EMI), georg.ganzenmueller at emi.fhg.de, USER-SPH package</li>
</ul>
<p>As discussed in <a class="reference internal" href="Section_history.html"><em>Section_history</em></a>, LAMMPS
originated as a cooperative project between DOE labs and industrial
partners. Folks involved in the design and testing of the original
version of LAMMPS were the following:</p>
<ul class="simple">
<li>John Carpenter (Mayo Clinic, formerly at Cray Research)</li>
<li>Terry Stouch (Lexicon Pharmaceuticals, formerly at Bristol Myers Squibb)</li>
<li>Steve Lustig (Dupont)</li>
<li>Jim Belak (LLNL)</li>
</ul>
</div>
</div>
</div>
</div>
<footer>
<div class="rst-footer-buttons" role="navigation" aria-label="footer navigation">
<a href="Section_start.html" class="btn btn-neutral float-right" title="2. Getting Started" accesskey="n">Next <span class="fa fa-arrow-circle-right"></span></a>
<a href="Manual.html" class="btn btn-neutral" title="LAMMPS Documentation" accesskey="p"><span class="fa fa-arrow-circle-left"></span> Previous</a>
</div>
<hr/>
<div role="contentinfo">
<p>
&copy; Copyright .
</p>
</div>
Built with <a href="http://sphinx-doc.org/">Sphinx</a> using a <a href="https://github.com/snide/sphinx_rtd_theme">theme</a> provided by <a href="https://readthedocs.org">Read the Docs</a>.
</footer>
</div>
</div>
</section>
</div>
<script type="text/javascript">
var DOCUMENTATION_OPTIONS = {
URL_ROOT:'./',
VERSION:'15 May 2015 version',
COLLAPSE_INDEX:false,
FILE_SUFFIX:'.html',
HAS_SOURCE: true
};
</script>
<script type="text/javascript" src="_static/jquery.js"></script>
<script type="text/javascript" src="_static/underscore.js"></script>
<script type="text/javascript" src="_static/doctools.js"></script>
<script type="text/javascript" src="https://cdn.mathjax.org/mathjax/latest/MathJax.js?config=TeX-AMS-MML_HTMLorMML"></script>
<script type="text/javascript" src="_static/sphinxcontrib-images/LightBox2/lightbox2/js/jquery-1.11.0.min.js"></script>
<script type="text/javascript" src="_static/sphinxcontrib-images/LightBox2/lightbox2/js/lightbox.min.js"></script>
<script type="text/javascript" src="_static/sphinxcontrib-images/LightBox2/lightbox2-customize/jquery-noconflict.js"></script>
<script type="text/javascript" src="_static/js/theme.js"></script>
<script type="text/javascript">
jQuery(function () {
SphinxRtdTheme.StickyNav.enable();
});
</script>
</body>
</html>

File diff suppressed because it is too large Load Diff

File diff suppressed because it is too large Load Diff

View File

@ -1,858 +0,0 @@
"Previous Section"_Section_commands.html - "LAMMPS WWW Site"_lws -
"LAMMPS Documentation"_ld - "LAMMPS Commands"_lc - "Next
Section"_Section_accelerate.html :c
:link(lws,http://lammps.sandia.gov)
:link(ld,Manual.html)
:link(lc,Section_commands.html#comm)
:line
4. Packages :h3
This section gives a quick overview of the add-on packages that extend
LAMMPS functionality.
4.1 "Standard packages"_#pkg_1
4.2 "User packages"_#pkg_2 :all(b)
LAMMPS includes many optional packages, which are groups of files that
enable a specific set of features. For example, force fields for
molecular systems or granular systems are in packages. You can see
the list of all packages by typing "make package" from within the src
directory of the LAMMPS distribution.
See "Section_start 3"_Section_start.html#start_3 of the manual for
details on how to include/exclude specific packages as part of the
LAMMPS build process, and for more details about the differences
between standard packages and user packages.
Unless otherwise noted below, every package is independent of all the
others. I.e. any package can be included or excluded in a LAMMPS
build, independent of all other packages. However, note that some
packages include commands derived from commands in other packages. If
the other package is not installed, the derived command from the new
package will also not be installed when you include the new one.
E.g. the pair lj/cut/coul/long/omp command from the USER-OMP package
will not be installed as part of the USER-OMP package if the KSPACE
package is not also installed, since it contains the pair
lj/cut/coul/long command. If you later install the KSPACE pacakge and
the USER-OMP package is already installed, both the pair
lj/cut/coul/long and lj/cut/coul/long/omp commands will be installed.
The two tables below list currently available packages in LAMMPS, with
a one-line descriptions of each. The sections below give a few more
details, including instructions for building LAMMPS with the package,
either via the make command or the Make.py tool described in "Section
2.4"_Section_start.html#start_4.
:line
:line
4.1 Standard packages :h4,link(pkg_1)
The current list of standard packages is as follows.
Package, Description, Author(s), Doc page, Example, Library
ASPHERE, aspherical particles, -, "Section_howto 6.14"_Section_howto.html#howto_14, ellipse, -
BODY, body-style particles, -, "body"_body.html, body, -
CLASS2, class 2 force fields, -, "pair_style lj/class2"_pair_class2.html, -, -
COLLOID, colloidal particles, -, "atom_style colloid"_atom_style.html, colloid, -
COMPRESS, I/O compression, Axel Kohlmeyer (Temple U), "dump */gz"_dump.html, -, -
CORESHELL, adiabatic core/shell model, Hendrik Heenen (Technical U of Munich), "Section_howto 6.25"_Section_howto.html#howto_25, coreshell, -
DIPOLE, point dipole particles, -, "pair_style dipole/cut"_pair_dipole.html, dipole, -
FLD, Fast Lubrication Dynamics, Kumar & Bybee & Higdon (1), "pair_style lubricateU"_pair_lubricateU.html, -, -
GPU, GPU-enabled styles, Mike Brown (ORNL), "Section accelerate"_accelerate_gpu.html, gpu, lib/gpu
GRANULAR, granular systems, -, "Section_howto 6.6"_Section_howto.html#howto_6, pour, -
KIM, openKIM potentials, Smirichinski & Elliot & Tadmor (3), "pair_style kim"_pair_kim.html, kim, KIM
KOKKOS, Kokkos-enabled styles, Trott & Edwards (4), "Section_accelerate"_accelerate_kokkos.html, kokkos, lib/kokkos
KSPACE, long-range Coulombic solvers, -, "kspace_style"_kspace_style.html, peptide, -
MANYBODY, many-body potentials, -, "pair_style tersoff"_pair_tersoff.html, shear, -
MEAM, modified EAM potential, Greg Wagner (Sandia), "pair_style meam"_pair_meam.html, meam, lib/meam
MC, Monte Carlo options, -, "fix gcmc"_fix_gcmc.html, -, -
MOLECULE, molecular system force fields, -, "Section_howto 6.3"_Section_howto.html#howto_3, peptide, -
OPT, optimized pair styles, Fischer & Richie & Natoli (2), "Section accelerate"_accelerate_opt.html, -, -
PERI, Peridynamics models, Mike Parks (Sandia), "pair_style peri"_pair_peri.html, peri, -
POEMS, coupled rigid body motion, Rudra Mukherjee (JPL), "fix poems"_fix_poems.html, rigid, lib/poems
PYTHON, embed Python code in an input script, -, "python"_python.html, python, lib/python
REAX, ReaxFF potential, Aidan Thompson (Sandia), "pair_style reax"_pair_reax.html, reax, lib/reax
REPLICA, multi-replica methods, -, "Section_howto 6.5"_Section_howto.html#howto_5, tad, -
RIGID, rigid bodies, -, "fix rigid"_fix_rigid.html, rigid, -
SHOCK, shock loading methods, -, "fix msst"_fix_msst.html, -, -
SNAP, quantum-fit potential, Aidan Thompson (Sandia), "pair snap"_pair_snap.html, snap, -
SRD, stochastic rotation dynamics, -, "fix srd"_fix_srd.html, srd, -
VORONOI, Voronoi tesselations, Daniel Schwen (LANL), "compute voronoi/atom"_compute_voronoi_atom.html, -, Voro++
XTC, dumps in XTC format, -, "dump"_dump.html, -, -
:tb(ea=c)
The "Authors" column lists a name(s) if a specific person is
responible for creating and maintaining the package.
More details on
multiple authors are give below.
(1) The FLD package was created by Amit Kumar and Michael Bybee from
Jonathan Higdon's group at UIUC.
(2) The OPT package was created by James Fischer (High Performance
Technologies), David Richie, and Vincent Natoli (Stone Ridge
Technolgy).
(3) The KIM package was created by Valeriu Smirichinski, Ryan Elliott,
and Ellad Tadmor (U Minn).
(4) The KOKKOS package was created primarily by Christian Trott
(Sandia). It uses the Kokkos library which was developed by Carter
Edwards, Christian, and collaborators at Sandia.
The "Doc page" column links to either a portion of the
"Section_howto"_Section_howto.html of the manual, or an input script
command implemented as part of the package.
The "Example" column is a sub-directory in the examples directory of
the distribution which has an input script that uses the package.
E.g. "peptide" refers to the examples/peptide directory.
The "Library" column lists an external library which must be built
first and which LAMMPS links to when it is built. If it is listed as
lib/package, then the code for the library is under the lib directory
of the LAMMPS distribution. See the lib/package/README file for info
on how to build the library. If it is not listed as lib/package, then
it is a third-party library not included in the LAMMPS distribution.
See the src/package/README or src/package/Makefile.lammps file for
info on where to download the library. "Section
start"_Section_start.html#start_3_3 of the manual also gives details
on how to build LAMMPS with both kinds of auxiliary libraries.
Except where explained below, all of these packages can be installed,
and LAMMPS re-built, by issuing these commands from the src dir.
make yes-package
make machine
or
Make.py -p package -a machine :pre
To un-install the package and re-build LAMMPS without it:
make no-package
make machine
or
Make.py -p ^package -a machine :pre
"Package" is the name of the package in lower-case letters,
e.g. asphere or rigid, and "machine" is the build target, e.g. mpi or
serial.
:line
:line
Build instructions for COMPRESS package :h4
:line
Build instructions for GPU package :h4
:line
Build instructions for KIM package :h4
:line
Build instructions for KOKKOS package :h4
:line
Build instructions for KSPACE package :h4
:line
Build instructions for MEAM package :h4
:line
Build instructions for POEMS package :h4
:line
Build instructions for PYTHON package :h4
:line
Build instructions for REAX package :h4
:line
Build instructions for VORONOI package :h4
:line
Build instructions for XTC package :h4
:line
:line
4.2 User packages :h4,link(pkg_2)
The current list of user-contributed packages is as follows:
Package, Description, Author(s), Doc page, Example, Pic/movie, Library
USER-ATC, atom-to-continuum coupling, Jones & Templeton & Zimmerman (1), "fix atc"_fix_atc.html, USER/atc, "atc"_atc, lib/atc
USER-AWPMD, wave-packet MD, Ilya Valuev (JIHT), "pair_style awpmd/cut"_pair_awpmd.html, USER/awpmd, -, lib/awpmd
USER-CG-CMM, coarse-graining model, Axel Kohlmeyer (Temple U), "pair_style lj/sdk"_pair_sdk.html, USER/cg-cmm, "cg"_cg, -
USER-COLVARS, collective variables, Fiorin & Henin & Kohlmeyer (2), "fix colvars"_fix_colvars.html, USER/colvars, "colvars"_colvars, lib/colvars
USER-CUDA, NVIDIA GPU styles, Christian Trott (U Tech Ilmenau), "Section accelerate"_accelerate_cuda.html, USER/cuda, -, lib/cuda
USER-DIFFRACTION, virutal x-ray and electron diffraction, Shawn Coleman (ARL),"compute xrd"_compute_xrd.html, USER/diffraction, -, -
USER-DRUDE, Drude oscillators, Dequidt & Devemy & Padua (3), "tutorial"_tutorial_drude.html, USER/drude, -, -
USER-EFF, electron force field, Andres Jaramillo-Botero (Caltech), "pair_style eff/cut"_pair_eff.html, USER/eff, "eff"_eff, -
USER-FEP, free energy perturbation, Agilio Padua (U Blaise Pascal Clermont-Ferrand), "compute fep"_compute_fep.html, USER/fep, -, -
USER-H5MD, dump output via HDF5, Pierre de Buyl (KU Leuven), "dump h5md"_dump_h5md.html, -, -, lib/h5md
USER-INTEL, Vectorized CPU and Intel(R) coprocessor styles, W. Michael Brown (Intel), "Section accelerate"_accelerate_intel.html, examples/intel, -, -
USER-LB, Lattice Boltzmann fluid, Colin Denniston (U Western Ontario), "fix lb/fluid"_fix_lb_fluid.html, USER/lb, -, -
USER-MGPT, fast MGPT multi-ion potentials, Tomas Oppelstrup & John Moriarty (LLNL), "pair_style mgpt"_pair_mgpt.html, USER/mgpt, -, -
USER-MISC, single-file contributions, USER-MISC/README, USER-MISC/README, -, -, -
USER-MOLFILE, "VMD"_VMD molfile plug-ins, Axel Kohlmeyer (Temple U), "dump molfile"_dump_molfile.html, -, -, VMD-MOLFILE
USER-OMP, OpenMP threaded styles, Axel Kohlmeyer (Temple U), "Section accelerate"_accelerate_omp.html, -, -, -
USER-PHONON, phonon dynamical matrix, Ling-Ti Kong (Shanghai Jiao Tong U), "fix phonon"_fix_phonon.html, USER/phonon, -, -
USER-QMMM, QM/MM coupling, Axel Kohlmeyer (Temple U), "fix qmmm"_fix_qmmm.html, USER/qmmm, -, lib/qmmm
USER-QTB, quantum nuclear effects, Yuan Shen (Stanford), "fix qtb"_fix_qtb.html "fix_qbmsst"_fix_qbmsst.html, qtb, -, -
USER-QUIP, QUIP/libatoms interface, Albert Bartok-Partay (U Cambridge), "pair_style quip"_pair_quip.html, USER/quip, -, lib/quip
USER-REAXC, C version of ReaxFF, Metin Aktulga (LBNL), "pair_style reaxc"_pair_reax_c.html, reax, -, -
USER-SMD, smoothed Mach dynamics, Georg Ganzenmuller (EMI), "userguide.pdf"_PDF/SMD_LAMMPS_userguide.pdf, USER/smd, -, -
USER-SMTBQ, Second Moment Tight Binding - QEq potential, Salles & Maras & Politano & Tetot (4), "pair_style smtbq"_pair_smtbq.html, USER/smtbq, -, -
USER-SPH, smoothed particle hydrodynamics, Georg Ganzenmuller (EMI), "userguide.pdf"_PDF/SPH_LAMMPS_userguide.pdf, USER/sph, "sph"_sph, -
USER-TALLY, Pairwise tallied computes, Axel Kohlmeyer (Temple U), "compute <...>/tally"_compute_tally.html, USER/tally, -, -
:tb(ea=c)
:link(atc,http://lammps.sandia.gov/pictures.html#atc)
:link(cg,http://lammps.sandia.gov/pictures.html#cg)
:link(eff,http://lammps.sandia.gov/movies.html#eff)
:link(sph,http://lammps.sandia.gov/movies.html#sph)
:link(VMD,http://www.ks.uiuc.edu/Research/vmd)
The "Authors" column lists a name(s) if a specific person is
responible for creating and maintaining the package.
(1) The ATC package was created by Reese Jones, Jeremy Templeton, and
Jon Zimmerman (Sandia).
(2) The COLVARS package was created by Axel Kohlmeyer (Temple U) using
the colvars module library written by Giacomo Fiorin (Temple U) and
Jerome Henin (LISM, Marseille, France).
(3) The DRUDE package was created by Alain Dequidt (U Blaise Pascal
Clermont-Ferrand) and co-authors Julien Devemy (CNRS) and Agilio Padua
(U Blaise Pascal).
(4) The SMTBQ package was created by Nicolas Salles, Emile Maras,
Olivier Politano, and Robert Tetot (LAAS-CNRS, France).
If the Library is not listed as lib/package, then it is a third-party
library not included in the LAMMPS distribution. See the
src/package/Makefile.lammps file for info on where to download the
library from.
The "Doc page" column links to either a portion of the
"Section_howto"_Section_howto.html of the manual, or an input script
command implemented as part of the package, or to additional
documentation provided within the package.
The "Example" column is a sub-directory in the examples directory of
the distribution which has an input script that uses the package.
E.g. "peptide" refers to the examples/peptide directory. USER/cuda
refers to the examples/USER/cuda directory.
The "Library" column lists an external library which must be built
first and which LAMMPS links to when it is built. If it is listed as
lib/package, then the code for the library is under the lib directory
of the LAMMPS distribution. See the lib/package/README file for info
on how to build the library. If it is not listed as lib/package, then
it is a third-party library not included in the LAMMPS distribution.
See the src/package/Makefile.lammps file for info on where to download
the library. "Section start"_Section_start.html#start_3_3 of the
manual also gives details on how to build LAMMPS with both kinds of
auxiliary libraries.
Except where explained below, all of these packages can be installed,
and LAMMPS re-built, by issuing these commands from the src dir.
make yes-user-package
make machine
or
Make.py -p package -a machine :pre
To un-install the package and re-build LAMMPS without it:
make no-user-package
make machine
or
Make.py -p ^package -a machine :pre
"Package" is the name of the package (in this case without the user
prefix) in lower-case letters, e.g. drude or phonon, and "machine" is
the build target, e.g. mpi or serial.
:line
:line
USER-ATC package :h4
This package implements a "fix atc" command which can be used in a
LAMMPS input script. This fix can be employed to either do concurrent
coupling of MD with FE-based physics surrogates or on-the-fly
post-processing of atomic information to continuum fields.
See the doc page for the fix atc command to get started. At the
bottom of the doc page are many links to additional documentation
contained in the doc/USER/atc directory.
There are example scripts for using this package in examples/USER/atc.
This package uses an external library in lib/atc which must be
compiled before making LAMMPS. See the lib/atc/README file and the
LAMMPS manual for information on building LAMMPS with external
libraries.
The primary people who created this package are Reese Jones (rjones at
sandia.gov), Jeremy Templeton (jatempl at sandia.gov) and Jon
Zimmerman (jzimmer at sandia.gov) at Sandia. Contact them directly if
you have questions.
:line
USER-AWPMD package :h4
This package contains a LAMMPS implementation of the Antisymmetrized
Wave Packet Molecular Dynamics (AWPMD) method.
See the doc page for the pair_style awpmd/cut command to get started.
There are example scripts for using this package in examples/USER/awpmd.
This package uses an external library in lib/awpmd which must be
compiled before making LAMMPS. See the lib/awpmd/README file and the
LAMMPS manual for information on building LAMMPS with external
libraries.
The person who created this package is Ilya Valuev at the JIHT in
Russia (valuev at physik.hu-berlin.de). Contact him directly if you
have questions.
:line
USER-CG-CMM package :h4
This package implements 3 commands which can be used in a LAMMPS input
script:
pair_style lj/sdk
pair_style lj/sdk/coul/long
angle_style sdk :ul
These styles allow coarse grained MD simulations with the
parametrization of Shinoda, DeVane, Klein, Mol Sim, 33, 27 (2007)
(SDK), with extensions to simulate ionic liquids, electrolytes, lipids
and charged amino acids.
See the doc pages for these commands for details.
There are example scripts for using this package in
examples/USER/cg-cmm.
This is the second generation implementation reducing the the clutter
of the previous version. For many systems with electrostatics, it will
be faster to use pair_style hybrid/overlay with lj/sdk and coul/long
instead of the combined lj/sdk/coul/long style. since the number of
charged atom types is usually small. For any other coulomb
interactions this is now required. To exploit this property, the use
of the kspace_style pppm/cg is recommended over regular pppm. For all
new styles, input file backward compatibility is provided. The old
implementation is still available through appending the /old
suffix. These will be discontinued and removed after the new
implementation has been fully validated.
The current version of this package should be considered beta
quality. The CG potentials work correctly for "normal" situations, but
have not been testing with all kinds of potential parameters and
simulation systems.
The person who created this package is Axel Kohlmeyer at Temple U
(akohlmey at gmail.com). Contact him directly if you have questions.
:line
USER-COLVARS package :h4
This package implements the "fix colvars" command which can be
used in a LAMMPS input script.
This fix allows to use "collective variables" to implement
Adaptive Biasing Force, Metadynamics, Steered MD, Umbrella
Sampling and Restraints. This code consists of two parts:
A portable collective variable module library written and maintained
by Giacomo Fiorin (ICMS, Temple University, Philadelphia, PA, USA) and
Jerome Henin (LISM, CNRS, Marseille, France). This code is located in
the directory lib/colvars and needs to be compiled first. The colvars
fix and an interface layer, exchanges information between LAMMPS and
the collective variable module. :ul
See the doc page of "fix colvars"_fix_colvars.html for more details.
There are example scripts for using this package in
examples/USER/colvars
This is a very new interface that does not yet support all
features in the module and will see future optimizations
and improvements. The colvars module library is also available
in NAMD has been thoroughly used and tested there. Bugs and
problems are likely due to the interface layers code.
Thus the current version of this package should be considered
beta quality.
The person who created this package is Axel Kohlmeyer at Temple U
(akohlmey at gmail.com). Contact him directly if you have questions.
:line
USER-CUDA package :h4
This package provides acceleration of various LAMMPS pair styles, fix
styles, compute styles, and long-range Coulombics via PPPM for NVIDIA
GPUs.
See this section of the manual to get started:
"Section_accelerate"_Section_accelerate.html#acc_7
There are example scripts for using this package in
examples/USER/cuda.
This package uses an external library in lib/cuda which must be
compiled before making LAMMPS. See the lib/cuda/README file and the
LAMMPS manual for information on building LAMMPS with external
libraries.
The person who created this package is Christian Trott at the
University of Technology Ilmenau, Germany (christian.trott at
tu-ilmenau.de). Contact him directly if you have questions.
:line
USER-DIFFRACTION package :h4
This package contains the commands neeed to calculate x-ray and
electron diffraction intensities based on kinematic diffraction
theory.
See these doc pages and their related commands to get started:
"compute xrd"_compute_xrd.html
"compute saed"_compute_saed.html
"fix saed/vtk"_fix_saed_vtk.html :ul
The person who created this package is Shawn P. Coleman
(shawn.p.coleman8.ctr at mail.mil) while at the University of
Arkansas. Contact him directly if you have questions.
:line
USER-DRUDE package :h4
This package implements methods for simulating polarizable systems
in LAMMPS using thermalized Drude oscillators.
See these doc pages and their related commands to get started:
"Drude tutorial"_tutorial_drude.html
"fix drude"_fix_drude.html
"compute temp/drude"_compute_temp_drude.html
"fix langevin/drude"_fix_langevin_drude.html
"fix drude/transform/..."_fix_drude_transform.html
"pair thole"_pair_thole.html :ul
There are auxiliary tools for using this package in tools/drude.
The person who created this package is Alain Dequidt at Universite
Blaise Pascal Clermont-Ferrand (alain.dequidt at univ-bpclermont.fr)
Contact him directly if you have questions. Co-authors: Julien Devemy,
Agilio Padua.
:line
USER-EFF package :h4
This package contains a LAMMPS implementation of the electron Force
Field (eFF) currently under development at Caltech, as described in
A. Jaramillo-Botero, J. Su, Q. An, and W.A. Goddard III, JCC,
2010. The eFF potential was first introduced by Su and Goddard, in
2007.
eFF can be viewed as an approximation to QM wave packet dynamics and
Fermionic molecular dynamics, combining the ability of electronic
structure methods to describe atomic structure, bonding, and chemistry
in materials, and of plasma methods to describe nonequilibrium
dynamics of large systems with a large number of highly excited
electrons. We classify it as a mixed QM-classical approach rather than
a conventional force field method, which introduces QM-based terms (a
spin-dependent repulsion term to account for the Pauli exclusion
principle and the electron wavefunction kinetic energy associated with
the Heisenberg principle) that reduce, along with classical
electrostatic terms between nuclei and electrons, to the sum of a set
of effective pairwise potentials. This makes eFF uniquely suited to
simulate materials over a wide range of temperatures and pressures
where electronically excited and ionized states of matter can occur
and coexist.
The necessary customizations to the LAMMPS core are in place to
enable the correct handling of explicit electron properties during
minimization and dynamics.
See the doc page for the pair_style eff/cut command to get started.
There are example scripts for using this package in
examples/USER/eff.
There are auxiliary tools for using this package in tools/eff.
The person who created this package is Andres Jaramillo-Botero at
CalTech (ajaramil at wag.caltech.edu). Contact him directly if you
have questions.
:line
USER-FEP package :h4
This package provides methods for performing free energy perturbation
simulations with soft-core pair potentials in LAMMPS.
See these doc pages and their related commands to get started:
"fix adapt/fep"_fix_adapt_fep.html
"compute fep"_compute_fep.html
"soft pair styles"_pair_lj_soft.html :ul
The person who created this package is Agilio Padua at Universite
Blaise Pascal Clermont-Ferrand (agilio.padua at univ-bpclermont.fr)
Contact him directly if you have questions.
:line
USER-H5MD package :h4
This package contains a "dump h5md"_dump_h5md.html command for
performing a dump of atom properties in HDF5 format. "HDF5
files"_HDF5 are binary, portable and self-describing and can be
examined and used by a variety of auxiliary tools. The output HDF5
files are structured in a format called H5MD, which was designed to
store molecular data, and can be used and produced by various MD and
MD-related codes. The "dump h5md"_doc/dump_h5md.html command gives a
citation to a paper describing the format.
:link(HDF5,http://www.hdfgroup.org/HDF5/)
The person who created this package and the underlying H5MD format is
Pierre de Buyl at KU Leuven (see http://pdebuyl.be). Contact him
directly if you have questions.
:line
USER-INTEL package :h4
This package provides options for performing neighbor list and
non-bonded force calculations in single, mixed, or double precision
and also a capability for accelerating calculations with an
Intel(R) Xeon Phi(TM) coprocessor.
See this section of the manual to get started:
"Section_accelerate"_Section_accelerate.html#acc_9
The person who created this package is W. Michael Brown at Intel
(michael.w.brown at intel.com). Contact him directly if you have questions.
:line
USER-LB package :h4
This package contains a LAMMPS implementation of a background
Lattice-Boltzmann fluid, which can be used to model MD particles
influenced by hydrodynamic forces.
See this doc page and its related commands to get started:
"fix lb/fluid"_fix_lb_fluid.html
The people who created this package are Frances Mackay (fmackay at
uwo.ca) and Colin (cdennist at uwo.ca) Denniston, University of
Western Ontario. Contact them directly if you have questions.
:line
USER-MGPT package :h4
This package contains a fast implementation for LAMMPS of
quantum-based MGPT multi-ion potentials. The MGPT or model GPT method
derives from first-principles DFT-based generalized pseudopotential
theory (GPT) through a series of systematic approximations valid for
mid-period transition metals with nearly half-filled d bands. The
MGPT method was originally developed by John Moriarty at Lawrence
Livermore National Lab (LLNL).
In the general matrix representation of MGPT, which can also be
applied to f-band actinide metals, the multi-ion potentials are
evaluated on the fly during a simulation through d- or f-state matrix
multiplication, and the forces that move the ions are determined
analytically. The {mgpt} pair style in this package calculates forces
and energies using an optimized matrix-MGPT algorithm due to Tomas
Oppelstrup at LLNL.
See this doc page to get started:
"pair_style mgpt"_pair_mgpt.html
The persons who created the USER-MGPT package are Tomas Oppelstrup
(oppelstrup2@llnl.gov) and John Moriarty (moriarty2@llnl.gov)
Contact them directly if you have any questions.
:line
USER-MISC package :h4
The files in this package are a potpourri of (mostly) unrelated
features contributed to LAMMPS by users. Each feature is a single
pair of files (*.cpp and *.h).
More information about each feature can be found by reading its doc
page in the LAMMPS doc directory. The doc page which lists all LAMMPS
input script commands is as follows:
"Section_commands"_Section_commands.html#cmd_5
User-contributed features are listed at the bottom of the fix,
compute, pair, etc sections.
The list of features and author of each is given in the
src/USER-MISC/README file.
You should contact the author directly if you have specific questions
about the feature or its coding.
:line
USER-MOLFILE package :h4
This package contains a dump molfile command which uses molfile
plugins that are bundled with the
"VMD"_http://www.ks.uiuc.edu/Research/vmd molecular visualization and
analysis program, to enable LAMMPS to dump its information in formats
compatible with various molecular simulation tools.
The package only provides the interface code, not the plugins. These
can be obtained from a VMD installation which has to match the
platform that you are using to compile LAMMPS for. By adding plugins
to VMD, support for new file formats can be added to LAMMPS (or VMD or
other programs that use them) without having to recompile the
application itself.
See this doc page to get started:
"dump molfile"_dump_molfile.html#acc_5
The person who created this package is Axel Kohlmeyer at Temple U
(akohlmey at gmail.com). Contact him directly if you have questions.
:line
USER-OMP package :h4
This package provides OpenMP multi-threading support and
other optimizations of various LAMMPS pair styles, dihedral
styles, and fix styles.
See this section of the manual to get started:
"Section_accelerate"_Section_accelerate.html#acc_5
The person who created this package is Axel Kohlmeyer at Temple U
(akohlmey at gmail.com). Contact him directly if you have questions.
:line
USER-PHONON package :h4
This package contains a fix phonon command that calculates dynamical
matrices, which can then be used to compute phonon dispersion
relations, directly from molecular dynamics simulations.
See this doc page to get started:
"fix phonon"_fix_phonon.html
The person who created this package is Ling-Ti Kong (konglt at
sjtu.edu.cn) at Shanghai Jiao Tong University. Contact him directly
if you have questions.
:line
USER-QMMM package :h4
This package provides a fix qmmm command which allows LAMMPS to be
used in a QM/MM simulation, currently only in combination with pw.x
code from the "Quantum ESPRESSO"_espresso package.
:link(espresso,http://www.quantum-espresso.org)
The current implementation only supports an ONIOM style mechanical
coupling to the Quantum ESPRESSO plane wave DFT package.
Electrostatic coupling is in preparation and the interface has been
written in a manner that coupling to other QM codes should be possible
without changes to LAMMPS itself.
See this doc page to get started:
"fix qmmm"_fix_qmmm.html
as well as the lib/qmmm/README file.
The person who created this package is Axel Kohlmeyer at Temple U
(akohlmey at gmail.com). Contact him directly if you have questions.
:line
USER-QTB package :h4
This package provides a self-consistent quantum treatment of the
vibrational modes in a classical molecular dynamics simulation. By
coupling the MD simulation to a colored thermostat, it introduces zero
point energy into the system, alter the energy power spectrum and the
heat capacity towards their quantum nature. This package could be of
interest if one wants to model systems at temperatures lower than
their classical limits or when temperatures ramp up across the
classical limits in the simulation.
See these two doc pages to get started:
"fix qtb"_fix_qtb.html provides quantum nulcear correction through a
colored thermostat and can be used with other time integration schemes
like "fix nve"_fix_nve.html or "fix nph"_fix_nh.html.
"fix qbmsst"_fix_qbmsst.html enables quantum nuclear correction of a
multi-scale shock technique simulation by coupling the quantum thermal
bath with the shocked system.
The person who created this package is Yuan Shen (sy0302 at
stanford.edu) at Stanford University. Contact him directly if you
have questions.
:line
USER-REAXC package :h4
This package contains a implementation for LAMMPS of the ReaxFF force
field. ReaxFF uses distance-dependent bond-order functions to
represent the contributions of chemical bonding to the potential
energy. It was originally developed by Adri van Duin and the Goddard
group at CalTech.
The USER-REAXC version of ReaxFF (pair_style reax/c), implemented in
C, should give identical or very similar results to pair_style reax,
which is a ReaxFF implementation on top of a Fortran library, a
version of which library was originally authored by Adri van Duin.
The reax/c version should be somewhat faster and more scalable,
particularly with respect to the charge equilibration calculation. It
should also be easier to build and use since there are no complicating
issues with Fortran memory allocation or linking to a Fortran library.
For technical details about this implemention of ReaxFF, see
this paper:
Parallel and Scalable Reactive Molecular Dynamics: Numerical Methods
and Algorithmic Techniques, H. M. Aktulga, J. C. Fogarty,
S. A. Pandit, A. Y. Grama, Parallel Computing, in press (2011).
See the doc page for the pair_style reax/c command for details
of how to use it in LAMMPS.
The person who created this package is Hasan Metin Aktulga (hmaktulga
at lbl.gov), while at Purdue University. Contact him directly, or
Aidan Thompson at Sandia (athomps at sandia.gov), if you have
questions.
:line
USER-SMD package :h4
This package implements smoothed Mach dynamics (SMD) in
LAMMPS. Currently, the package has the following features:
* Does liquids via traditional Smooth Particle Hydrodynamics (SPH)
* Also solves solids mechanics problems via a state of the art
stabilized meshless method with hourglass control.
* Can specify hydrostatic interactions independently from material
strength models, i.e. pressure and deviatoric stresses are separated.
* Many material models available (Johnson-Cook, plasticity with
hardening, Mie-Grueneisen, Polynomial EOS). Easy to add new
material models.
* Rigid boundary conditions (walls) can be loaded as surface geometries
from *.STL files.
See the file doc/PDF/SMD_LAMMPS_userguide.pdf to get started.
There are example scripts for using this package in examples/USER/smd.
The person who created this package is Georg Ganzenmuller at the
Fraunhofer-Institute for High-Speed Dynamics, Ernst Mach Institute in
Germany (georg.ganzenmueller at emi.fhg.de). Contact him directly if
you have questions.
:line
USER-SMTBQ package :h4
This package implements the Second Moment Tight Binding - QEq (SMTB-Q)
potential for the description of ionocovalent bonds in oxides.
There are example scripts for using this package in
examples/USER/smtbq.
See this doc page to get started:
"pair_style smtbq"_pair_smtbq.html
The persons who created the USER-SMTBQ package are Nicolas Salles,
Emile Maras, Olivier Politano, Robert Tetot, who can be contacted at
these email addreses: lammps@u-bourgogne.fr, nsalles@laas.fr. Contact
them directly if you have any questions.
:line
USER-SPH package :h4
This package implements smoothed particle hydrodynamics (SPH) in
LAMMPS. Currently, the package has the following features:
* Tait, ideal gas, Lennard-Jones equation of states, full support for
complete (i.e. internal-energy dependent) equations of state
* Plain or Monaghans XSPH integration of the equations of motion
* Density continuity or density summation to propagate the density field
* Commands to set internal energy and density of particles from the
input script
* Output commands to access internal energy and density for dumping and
thermo output
See the file doc/PDF/SPH_LAMMPS_userguide.pdf to get started.
There are example scripts for using this package in examples/USER/sph.
The person who created this package is Georg Ganzenmuller at the
Fraunhofer-Institute for High-Speed Dynamics, Ernst Mach Institute in
Germany (georg.ganzenmueller at emi.fhg.de). Contact him directly if
you have questions.

View File

@ -1,304 +0,0 @@
<!DOCTYPE html>
<!--[if IE 8]><html class="no-js lt-ie9" lang="en" > <![endif]-->
<!--[if gt IE 8]><!--> <html class="no-js" lang="en" > <!--<![endif]-->
<head>
<meta charset="utf-8">
<meta name="viewport" content="width=device-width, initial-scale=1.0">
<title>8. Performance &amp; scalability &mdash; LAMMPS 15 May 2015 version documentation</title>
<link rel="stylesheet" href="_static/css/theme.css" type="text/css" />
<link rel="stylesheet" href="_static/sphinxcontrib-images/LightBox2/lightbox2/css/lightbox.css" type="text/css" />
<link rel="top" title="LAMMPS 15 May 2015 version documentation" href="index.html"/>
<link rel="next" title="9. Additional tools" href="Section_tools.html"/>
<link rel="prev" title="7. Example problems" href="Section_example.html"/>
<script src="_static/js/modernizr.min.js"></script>
</head>
<body class="wy-body-for-nav" role="document">
<div class="wy-grid-for-nav">
<nav data-toggle="wy-nav-shift" class="wy-nav-side">
<div class="wy-side-nav-search">
<a href="Manual.html" class="icon icon-home"> LAMMPS
</a>
<div role="search">
<form id="rtd-search-form" class="wy-form" action="search.html" method="get">
<input type="text" name="q" placeholder="Search docs" />
<input type="hidden" name="check_keywords" value="yes" />
<input type="hidden" name="area" value="default" />
</form>
</div>
</div>
<div class="wy-menu wy-menu-vertical" data-spy="affix" role="navigation" aria-label="main navigation">
<ul class="current">
<li class="toctree-l1"><a class="reference internal" href="Section_intro.html">1. Introduction</a></li>
<li class="toctree-l1"><a class="reference internal" href="Section_start.html">2. Getting Started</a></li>
<li class="toctree-l1"><a class="reference internal" href="Section_commands.html">3. Commands</a></li>
<li class="toctree-l1"><a class="reference internal" href="Section_packages.html">4. Packages</a></li>
<li class="toctree-l1"><a class="reference internal" href="Section_accelerate.html">5. Accelerating LAMMPS performance</a></li>
<li class="toctree-l1"><a class="reference internal" href="Section_howto.html">6. How-to discussions</a></li>
<li class="toctree-l1"><a class="reference internal" href="Section_example.html">7. Example problems</a></li>
<li class="toctree-l1 current"><a class="current reference internal" href="">8. Performance &amp; scalability</a></li>
<li class="toctree-l1"><a class="reference internal" href="Section_tools.html">9. Additional tools</a></li>
<li class="toctree-l1"><a class="reference internal" href="Section_modify.html">10. Modifying &amp; extending LAMMPS</a></li>
<li class="toctree-l1"><a class="reference internal" href="Section_python.html">11. Python interface to LAMMPS</a></li>
<li class="toctree-l1"><a class="reference internal" href="Section_errors.html">12. Errors</a></li>
<li class="toctree-l1"><a class="reference internal" href="Section_history.html">13. Future and history</a></li>
</ul>
</div>
&nbsp;
</nav>
<section data-toggle="wy-nav-shift" class="wy-nav-content-wrap">
<nav class="wy-nav-top" role="navigation" aria-label="top navigation">
<i data-toggle="wy-nav-top" class="fa fa-bars"></i>
<a href="Manual.html">LAMMPS</a>
</nav>
<div class="wy-nav-content">
<div class="rst-content">
<div role="navigation" aria-label="breadcrumbs navigation">
<ul class="wy-breadcrumbs">
<li><a href="Manual.html">Docs</a> &raquo;</li>
<li>8. Performance &amp; scalability</li>
<li class="wy-breadcrumbs-aside">
<a href="http://lammps.sandia.gov">Website</a>
<a href="Section_commands.html#comm">Commands</a>
</li>
</ul>
<hr/>
<div class="rst-footer-buttons" style="margin-bottom: 1em" role="navigation" aria-label="footer navigation">
<a href="Section_tools.html" class="btn btn-neutral float-right" title="9. Additional tools" accesskey="n">Next <span class="fa fa-arrow-circle-right"></span></a>
<a href="Section_example.html" class="btn btn-neutral" title="7. Example problems" accesskey="p"><span class="fa fa-arrow-circle-left"></span> Previous</a>
</div>
</div>
<div role="main" class="document" itemscope="itemscope" itemtype="http://schema.org/Article">
<div itemprop="articleBody">
<div class="section" id="performance-scalability">
<h1>8. Performance &amp; scalability<a class="headerlink" href="#performance-scalability" title="Permalink to this headline"></a></h1>
<p>LAMMPS performance on several prototypical benchmarks and machines is
discussed on the Benchmarks page of the <a class="reference external" href="http://lammps.sandia.gov">LAMMPS WWW Site</a> where
CPU timings and parallel efficiencies are listed. Here, the
benchmarks are described briefly and some useful rules of thumb about
their performance are highlighted.</p>
<p>These are the 5 benchmark problems:</p>
<ol class="arabic simple">
<li>LJ = atomic fluid, Lennard-Jones potential with 2.5 sigma cutoff (55</li>
</ol>
<blockquote>
<div>neighbors per atom), NVE integration</div></blockquote>
<ol class="arabic simple">
<li>Chain = bead-spring polymer melt of 100-mer chains, FENE bonds and LJ
pairwise interactions with a 2^(1/6) sigma cutoff (5 neighbors per
atom), NVE integration</li>
<li>EAM = metallic solid, Cu EAM potential with 4.95 Angstrom cutoff (45
neighbors per atom), NVE integration</li>
<li>Chute = granular chute flow, frictional history potential with 1.1
sigma cutoff (7 neighbors per atom), NVE integration</li>
<li>Rhodo = rhodopsin protein in solvated lipid bilayer, CHARMM force
field with a 10 Angstrom LJ cutoff (440 neighbors per atom),
particle-particle particle-mesh (PPPM) for long-range Coulombics, NPT
integration</li>
</ol>
<p>The input files for running the benchmarks are included in the LAMMPS
distribution, as are sample output files. Each of the 5 problems has
32,000 atoms and runs for 100 timesteps. Each can be run as a serial
benchmarks (on one processor) or in parallel. In parallel, each
benchmark can be run as a fixed-size or scaled-size problem. For
fixed-size benchmarking, the same 32K atom problem is run on various
numbers of processors. For scaled-size benchmarking, the model size
is increased with the number of processors. E.g. on 8 processors, a
256K-atom problem is run; on 1024 processors, a 32-million atom
problem is run, etc.</p>
<p>A useful metric from the benchmarks is the CPU cost per atom per
timestep. Since LAMMPS performance scales roughly linearly with
problem size and timesteps, the run time of any problem using the same
model (atom style, force field, cutoff, etc) can then be estimated.
For example, on a 1.7 GHz Pentium desktop machine (Intel icc compiler
under Red Hat Linux), the CPU run-time in seconds/atom/timestep for
the 5 problems is</p>
<table border="1" class="docutils">
<colgroup>
<col width="25%" />
<col width="14%" />
<col width="14%" />
<col width="14%" />
<col width="14%" />
<col width="17%" />
</colgroup>
<tbody valign="top">
<tr class="row-odd"><td>Problem:</td>
<td>LJ</td>
<td>Chain</td>
<td>EAM</td>
<td>Chute</td>
<td>Rhodopsin</td>
</tr>
<tr class="row-even"><td>CPU/atom/step:</td>
<td>4.55E-6</td>
<td>2.18E-6</td>
<td>9.38E-6</td>
<td>2.18E-6</td>
<td>1.11E-4</td>
</tr>
<tr class="row-odd"><td>Ratio to LJ:</td>
<td>1.0</td>
<td>0.48</td>
<td>2.06</td>
<td>0.48</td>
<td>24.5</td>
</tr>
</tbody>
</table>
<p>The ratios mean that if the atomic LJ system has a normalized cost of
1.0, the bead-spring chains and granular systems run 2x faster, while
the EAM metal and solvated protein models run 2x and 25x slower
respectively. The bulk of these cost differences is due to the
expense of computing a particular pairwise force field for a given
number of neighbors per atom.</p>
<p>Performance on a parallel machine can also be predicted from the
one-processor timings if the parallel efficiency can be estimated.
The communication bandwidth and latency of a particular parallel
machine affects the efficiency. On most machines LAMMPS will give
fixed-size parallel efficiencies on these benchmarks above 50% so long
as the atoms/processor count is a few 100 or greater - i.e. on 64 to
128 processors. Likewise, scaled-size parallel efficiencies will
typically be 80% or greater up to very large processor counts. The
benchmark data on the <a class="reference external" href="http://lammps.sandia.gov">LAMMPS WWW Site</a> gives specific examples on
some different machines, including a run of 3/4 of a billion LJ atoms
on 1500 processors that ran at 85% parallel efficiency.</p>
</div>
</div>
</div>
<footer>
<div class="rst-footer-buttons" role="navigation" aria-label="footer navigation">
<a href="Section_tools.html" class="btn btn-neutral float-right" title="9. Additional tools" accesskey="n">Next <span class="fa fa-arrow-circle-right"></span></a>
<a href="Section_example.html" class="btn btn-neutral" title="7. Example problems" accesskey="p"><span class="fa fa-arrow-circle-left"></span> Previous</a>
</div>
<hr/>
<div role="contentinfo">
<p>
&copy; Copyright .
</p>
</div>
Built with <a href="http://sphinx-doc.org/">Sphinx</a> using a <a href="https://github.com/snide/sphinx_rtd_theme">theme</a> provided by <a href="https://readthedocs.org">Read the Docs</a>.
</footer>
</div>
</div>
</section>
</div>
<script type="text/javascript">
var DOCUMENTATION_OPTIONS = {
URL_ROOT:'./',
VERSION:'15 May 2015 version',
COLLAPSE_INDEX:false,
FILE_SUFFIX:'.html',
HAS_SOURCE: true
};
</script>
<script type="text/javascript" src="_static/jquery.js"></script>
<script type="text/javascript" src="_static/underscore.js"></script>
<script type="text/javascript" src="_static/doctools.js"></script>
<script type="text/javascript" src="https://cdn.mathjax.org/mathjax/latest/MathJax.js?config=TeX-AMS-MML_HTMLorMML"></script>
<script type="text/javascript" src="_static/sphinxcontrib-images/LightBox2/lightbox2/js/jquery-1.11.0.min.js"></script>
<script type="text/javascript" src="_static/sphinxcontrib-images/LightBox2/lightbox2/js/lightbox.min.js"></script>
<script type="text/javascript" src="_static/sphinxcontrib-images/LightBox2/lightbox2-customize/jquery-noconflict.js"></script>
<script type="text/javascript" src="_static/js/theme.js"></script>
<script type="text/javascript">
jQuery(function () {
SphinxRtdTheme.StickyNav.enable();
});
</script>
</body>
</html>

View File

@ -1,77 +0,0 @@
"Previous Section"_Section_example.html - "LAMMPS WWW Site"_lws - "LAMMPS Documentation"_ld - "LAMMPS Commands"_lc - "Next Section"_Section_tools.html :c
:link(lws,http://lammps.sandia.gov)
:link(ld,Manual.html)
:link(lc,Section_commands.html#comm)
:line
8. Performance & scalability :h3
LAMMPS performance on several prototypical benchmarks and machines is
discussed on the Benchmarks page of the "LAMMPS WWW Site"_lws where
CPU timings and parallel efficiencies are listed. Here, the
benchmarks are described briefly and some useful rules of thumb about
their performance are highlighted.
These are the 5 benchmark problems:
LJ = atomic fluid, Lennard-Jones potential with 2.5 sigma cutoff (55
neighbors per atom), NVE integration :olb,l
Chain = bead-spring polymer melt of 100-mer chains, FENE bonds and LJ
pairwise interactions with a 2^(1/6) sigma cutoff (5 neighbors per
atom), NVE integration :l
EAM = metallic solid, Cu EAM potential with 4.95 Angstrom cutoff (45
neighbors per atom), NVE integration :l
Chute = granular chute flow, frictional history potential with 1.1
sigma cutoff (7 neighbors per atom), NVE integration :l
Rhodo = rhodopsin protein in solvated lipid bilayer, CHARMM force
field with a 10 Angstrom LJ cutoff (440 neighbors per atom),
particle-particle particle-mesh (PPPM) for long-range Coulombics, NPT
integration :ole,l
The input files for running the benchmarks are included in the LAMMPS
distribution, as are sample output files. Each of the 5 problems has
32,000 atoms and runs for 100 timesteps. Each can be run as a serial
benchmarks (on one processor) or in parallel. In parallel, each
benchmark can be run as a fixed-size or scaled-size problem. For
fixed-size benchmarking, the same 32K atom problem is run on various
numbers of processors. For scaled-size benchmarking, the model size
is increased with the number of processors. E.g. on 8 processors, a
256K-atom problem is run; on 1024 processors, a 32-million atom
problem is run, etc.
A useful metric from the benchmarks is the CPU cost per atom per
timestep. Since LAMMPS performance scales roughly linearly with
problem size and timesteps, the run time of any problem using the same
model (atom style, force field, cutoff, etc) can then be estimated.
For example, on a 1.7 GHz Pentium desktop machine (Intel icc compiler
under Red Hat Linux), the CPU run-time in seconds/atom/timestep for
the 5 problems is
Problem:, LJ, Chain, EAM, Chute, Rhodopsin
CPU/atom/step:, 4.55E-6, 2.18E-6, 9.38E-6, 2.18E-6, 1.11E-4
Ratio to LJ:, 1.0, 0.48, 2.06, 0.48, 24.5 :tb(ea=c,ca1=r)
The ratios mean that if the atomic LJ system has a normalized cost of
1.0, the bead-spring chains and granular systems run 2x faster, while
the EAM metal and solvated protein models run 2x and 25x slower
respectively. The bulk of these cost differences is due to the
expense of computing a particular pairwise force field for a given
number of neighbors per atom.
Performance on a parallel machine can also be predicted from the
one-processor timings if the parallel efficiency can be estimated.
The communication bandwidth and latency of a particular parallel
machine affects the efficiency. On most machines LAMMPS will give
fixed-size parallel efficiencies on these benchmarks above 50% so long
as the atoms/processor count is a few 100 or greater - i.e. on 64 to
128 processors. Likewise, scaled-size parallel efficiencies will
typically be 80% or greater up to very large processor counts. The
benchmark data on the "LAMMPS WWW Site"_lws gives specific examples on
some different machines, including a run of 3/4 of a billion LJ atoms
on 1500 processors that ran at 85% parallel efficiency.

View File

@ -1,978 +0,0 @@
<!DOCTYPE html>
<!--[if IE 8]><html class="no-js lt-ie9" lang="en" > <![endif]-->
<!--[if gt IE 8]><!--> <html class="no-js" lang="en" > <!--<![endif]-->
<head>
<meta charset="utf-8">
<meta name="viewport" content="width=device-width, initial-scale=1.0">
<title>11. Python interface to LAMMPS &mdash; LAMMPS 15 May 2015 version documentation</title>
<link rel="stylesheet" href="_static/css/theme.css" type="text/css" />
<link rel="stylesheet" href="_static/sphinxcontrib-images/LightBox2/lightbox2/css/lightbox.css" type="text/css" />
<link rel="top" title="LAMMPS 15 May 2015 version documentation" href="index.html"/>
<link rel="next" title="12. Errors" href="Section_errors.html"/>
<link rel="prev" title="10. Modifying &amp; extending LAMMPS" href="Section_modify.html"/>
<script src="_static/js/modernizr.min.js"></script>
</head>
<body class="wy-body-for-nav" role="document">
<div class="wy-grid-for-nav">
<nav data-toggle="wy-nav-shift" class="wy-nav-side">
<div class="wy-side-nav-search">
<a href="Manual.html" class="icon icon-home"> LAMMPS
</a>
<div role="search">
<form id="rtd-search-form" class="wy-form" action="search.html" method="get">
<input type="text" name="q" placeholder="Search docs" />
<input type="hidden" name="check_keywords" value="yes" />
<input type="hidden" name="area" value="default" />
</form>
</div>
</div>
<div class="wy-menu wy-menu-vertical" data-spy="affix" role="navigation" aria-label="main navigation">
<ul class="current">
<li class="toctree-l1"><a class="reference internal" href="Section_intro.html">1. Introduction</a></li>
<li class="toctree-l1"><a class="reference internal" href="Section_start.html">2. Getting Started</a></li>
<li class="toctree-l1"><a class="reference internal" href="Section_commands.html">3. Commands</a></li>
<li class="toctree-l1"><a class="reference internal" href="Section_packages.html">4. Packages</a></li>
<li class="toctree-l1"><a class="reference internal" href="Section_accelerate.html">5. Accelerating LAMMPS performance</a></li>
<li class="toctree-l1"><a class="reference internal" href="Section_howto.html">6. How-to discussions</a></li>
<li class="toctree-l1"><a class="reference internal" href="Section_example.html">7. Example problems</a></li>
<li class="toctree-l1"><a class="reference internal" href="Section_perf.html">8. Performance &amp; scalability</a></li>
<li class="toctree-l1"><a class="reference internal" href="Section_tools.html">9. Additional tools</a></li>
<li class="toctree-l1"><a class="reference internal" href="Section_modify.html">10. Modifying &amp; extending LAMMPS</a></li>
<li class="toctree-l1 current"><a class="current reference internal" href="">11. Python interface to LAMMPS</a><ul>
<li class="toctree-l2"><a class="reference internal" href="#overview-of-running-lammps-from-python">11.1. Overview of running LAMMPS from Python</a></li>
<li class="toctree-l2"><a class="reference internal" href="#overview-of-using-python-from-a-lammps-script">11.2. Overview of using Python from a LAMMPS script</a></li>
<li class="toctree-l2"><a class="reference internal" href="#building-lammps-as-a-shared-library">11.3. Building LAMMPS as a shared library</a></li>
<li class="toctree-l2"><a class="reference internal" href="#installing-the-python-wrapper-into-python">11.4. Installing the Python wrapper into Python</a></li>
<li class="toctree-l2"><a class="reference internal" href="#extending-python-with-mpi-to-run-in-parallel">11.5. Extending Python with MPI to run in parallel</a></li>
<li class="toctree-l2"><a class="reference internal" href="#testing-the-python-lammps-interface">11.6. Testing the Python-LAMMPS interface</a><ul>
<li class="toctree-l3"><a class="reference internal" href="#test-lammps-and-python-in-serial">11.6.1. <strong>Test LAMMPS and Python in serial:</strong></a></li>
<li class="toctree-l3"><a class="reference internal" href="#test-lammps-and-python-in-parallel">11.6.2. <strong>Test LAMMPS and Python in parallel:</strong></a></li>
<li class="toctree-l3"><a class="reference internal" href="#running-python-scripts">11.6.3. <strong>Running Python scripts:</strong></a></li>
</ul>
</li>
<li class="toctree-l2"><a class="reference internal" href="#using-lammps-from-python">11.7. Using LAMMPS from Python</a></li>
<li class="toctree-l2"><a class="reference internal" href="#example-python-scripts-that-use-lammps">11.8. Example Python scripts that use LAMMPS</a></li>
</ul>
</li>
<li class="toctree-l1"><a class="reference internal" href="Section_errors.html">12. Errors</a></li>
<li class="toctree-l1"><a class="reference internal" href="Section_history.html">13. Future and history</a></li>
</ul>
</div>
&nbsp;
</nav>
<section data-toggle="wy-nav-shift" class="wy-nav-content-wrap">
<nav class="wy-nav-top" role="navigation" aria-label="top navigation">
<i data-toggle="wy-nav-top" class="fa fa-bars"></i>
<a href="Manual.html">LAMMPS</a>
</nav>
<div class="wy-nav-content">
<div class="rst-content">
<div role="navigation" aria-label="breadcrumbs navigation">
<ul class="wy-breadcrumbs">
<li><a href="Manual.html">Docs</a> &raquo;</li>
<li>11. Python interface to LAMMPS</li>
<li class="wy-breadcrumbs-aside">
<a href="http://lammps.sandia.gov">Website</a>
<a href="Section_commands.html#comm">Commands</a>
</li>
</ul>
<hr/>
<div class="rst-footer-buttons" style="margin-bottom: 1em" role="navigation" aria-label="footer navigation">
<a href="Section_errors.html" class="btn btn-neutral float-right" title="12. Errors" accesskey="n">Next <span class="fa fa-arrow-circle-right"></span></a>
<a href="Section_modify.html" class="btn btn-neutral" title="10. Modifying &amp; extending LAMMPS" accesskey="p"><span class="fa fa-arrow-circle-left"></span> Previous</a>
</div>
</div>
<div role="main" class="document" itemscope="itemscope" itemtype="http://schema.org/Article">
<div itemprop="articleBody">
<div class="section" id="python-interface-to-lammps">
<h1>11. Python interface to LAMMPS<a class="headerlink" href="#python-interface-to-lammps" title="Permalink to this headline"></a></h1>
<p>LAMMPS can work together with Python in two ways. First, Python can
wrap LAMMPS through the <a class="reference internal" href="Section_howto.html#howto-19"><span>LAMMPS library interface</span></a>, so that a Python script can
create one or more instances of LAMMPS and launch one or more
simulations. In Python lingo, this is &#8220;extending&#8221; Python with LAMMPS.</p>
<p>Second, LAMMPS can use the Python interpreter, so that a LAMMPS input
script can invoke Python code, and pass information back-and-forth
between the input script and Python functions you write. The Python
code can also callback to LAMMPS to query or change its attributes.
In Python lingo, this is &#8220;embedding&#8221; Python in LAMMPS.</p>
<p>This section describes how to do both.</p>
<ul class="simple">
<li>11.1 <a class="reference internal" href="#py-1"><span>Overview of running LAMMPS from Python</span></a></li>
<li>11.2 <a class="reference internal" href="#py-2"><span>Overview of using Python from a LAMMPS script</span></a></li>
<li>11.3 <a class="reference internal" href="#py-3"><span>Building LAMMPS as a shared library</span></a></li>
<li>11.4 <a class="reference internal" href="#py-4"><span>Installing the Python wrapper into Python</span></a></li>
<li>11.5 <a class="reference internal" href="#py-5"><span>Extending Python with MPI to run in parallel</span></a></li>
<li>11.6 <a class="reference internal" href="#py-6"><span>Testing the Python-LAMMPS interface</span></a></li>
<li>11.7 <a class="reference internal" href="#py-7"><span>Using LAMMPS from Python</span></a></li>
<li>11.8 <a class="reference internal" href="#py-8"><span>Example Python scripts that use LAMMPS</span></a></li>
</ul>
<p>If you are not familiar with it, <a class="reference external" href="http://www.python.org">Python</a> is a
powerful scripting and programming language which can essentially do
anything that faster, lower-level languages like C or C++ can do, but
typically with much fewer lines of code. When used in embedded mode,
Python can perform operations that the simplistic LAMMPS input script
syntax cannot. Python can be also be used as a &#8220;glue&#8221; language to
drive a program through its library interface, or to hook multiple
pieces of software together, such as a simulation package plus a
visualization package, or to run a coupled multiscale or multiphysics
model.</p>
<p>See <a class="reference internal" href="Section_howto.html#howto-10"><span>Section_howto 10</span></a> of the manual and
the couple directory of the distribution for more ideas about coupling
LAMMPS to other codes. See <a class="reference internal" href="Section_howto.html#howto-19"><span>Section_howto 19</span></a> for a description of the LAMMPS
library interface provided in src/library.cpp and src/library.h, and
how to extend it for your needs. As described below, that interface
is what is exposed to Python either when calling LAMMPS from Python or
when calling Python from a LAMMPS input script and then calling back
to LAMMPS from Python code. The library interface is designed to be
easy to add functions to. Thus the Python interface to LAMMPS is also
easy to extend as well.</p>
<p>If you create interesting Python scripts that run LAMMPS or
interesting Python functions that can be called from a LAMMPS input
script, that you think would be useful to other users, please <a class="reference external" href="http://lammps.sandia.gov/authors.html">email them to the developers</a>. We can
include them in the LAMMPS distribution.</p>
<div class="section" id="overview-of-running-lammps-from-python">
<span id="py-1"></span><h2>11.1. Overview of running LAMMPS from Python<a class="headerlink" href="#overview-of-running-lammps-from-python" title="Permalink to this headline"></a></h2>
<p>The LAMMPS distribution includes a python directory with all you need
to run LAMMPS from Python. The python/lammps.py file wraps the LAMMPS
library interface, with one wrapper function per LAMMPS library
function. This file makes it is possible to do the following either
from a Python script, or interactively from a Python prompt: create
one or more instances of LAMMPS, invoke LAMMPS commands or give it an
input script, run LAMMPS incrementally, extract LAMMPS results, an
modify internal LAMMPS variables. From a Python script you can do
this in serial or parallel. Running Python interactively in parallel
does not generally work, unless you have a version of Python that
extends standard Python to enable multiple instances of Python to read
what you type.</p>
<p>To do all of this, you must first build LAMMPS as a shared library,
then insure that your Python can find the python/lammps.py file and
the shared library. These steps are explained in subsequent sections
11.3 and 11.4. Sections 11.5 and 11.6 discuss using MPI from a
parallel Python program and how to test that you are ready to use
LAMMPS from Python. Section 11.7 lists all the functions in the
current LAMMPS library interface and how to call them from Python.</p>
<p>Section 11.8 gives some examples of coupling LAMMPS to other tools via
Python. For example, LAMMPS can easily be coupled to a GUI or other
visualization tools that display graphs or animations in real time as
LAMMPS runs. Examples of such scripts are inlcluded in the python
directory.</p>
<p>Two advantages of using Python to run LAMMPS are how concise the
language is, and that it can be run interactively, enabling rapid
development and debugging of programs. If you use it to mostly invoke
costly operations within LAMMPS, such as running a simulation for a
reasonable number of timesteps, then the overhead cost of invoking
LAMMPS thru Python will be negligible.</p>
<p>The Python wrapper for LAMMPS uses the amazing and magical (to me)
&#8220;ctypes&#8221; package in Python, which auto-generates the interface code
needed between Python and a set of C interface routines for a library.
Ctypes is part of standard Python for versions 2.5 and later. You can
check which version of Python you have installed, by simply typing
&#8220;python&#8221; at a shell prompt.</p>
<hr class="docutils" />
</div>
<div class="section" id="overview-of-using-python-from-a-lammps-script">
<span id="py-2"></span><h2>11.2. Overview of using Python from a LAMMPS script<a class="headerlink" href="#overview-of-using-python-from-a-lammps-script" title="Permalink to this headline"></a></h2>
<div class="admonition warning">
<p class="first admonition-title">Warning</p>
<p class="last">It is not currently possible to use the
<a class="reference internal" href="python.html"><em>python</em></a> command described in this section with Python 3,
only with Python 2. The C API changed from Python 2 to 3 and the
LAMMPS code is not compatible with both.</p>
</div>
<p>LAMMPS has a <a class="reference internal" href="python.html"><em>python</em></a> command which can be used in an
input script to define and execute a Python function that you write
the code for. The Python function can also be assigned to a LAMMPS
python-style variable via the <a class="reference internal" href="variable.html"><em>variable</em></a> command. Each
time the variable is evaluated, either in the LAMMPS input script
itself, or by another LAMMPS command that uses the variable, this will
trigger the Python function to be invoked.</p>
<p>The Python code for the function can be included directly in the input
script or in an auxiliary file. The function can have arguments which
are mapped to LAMMPS variables (also defined in the input script) and
it can return a value to a LAMMPS variable. This is thus a mechanism
for your input script to pass information to a piece of Python code,
ask Python to execute the code, and return information to your input
script.</p>
<p>Note that a Python function can be arbitrarily complex. It can import
other Python modules, instantiate Python classes, call other Python
functions, etc. The Python code that you provide can contain more
code than the single function. It can contain other functions or
Python classes, as well as global variables or other mechanisms for
storing state between calls from LAMMPS to the function.</p>
<p>The Python function you provide can consist of &#8220;pure&#8221; Python code that
only performs operations provided by standard Python. However, the
Python function can also &#8220;call back&#8221; to LAMMPS through its
Python-wrapped library interface, in the manner described in the
previous section 11.1. This means it can issue LAMMPS input script
commands or query and set internal LAMMPS state. As an example, this
can be useful in an input script to create a more complex loop with
branching logic, than can be created using the simple looping and
brancing logic enabled by the <a class="reference internal" href="next.html"><em>next</em></a> and <a class="reference internal" href="if.html"><em>if</em></a>
commands.</p>
<p>See the <a class="reference internal" href="python.html"><em>python</em></a> doc page and the <a class="reference internal" href="variable.html"><em>variable</em></a>
doc page for its python-style variables for more info, including
examples of Python code you can write for both pure Python operations
and callbacks to LAMMPS.</p>
<p>To run pure Python code from LAMMPS, you only need to build LAMMPS
with the PYTHON package installed:</p>
<p>make yes-python
make machine</p>
<p>Note that this will link LAMMPS with the Python library on your
system, which typically requires several auxiliary system libraries to
also be linked. The list of these libraries and the paths to find
them are specified in the lib/python/Makefile.lammps file. You need
to insure that file contains the correct information for your version
of Python and your machine to successfully build LAMMPS. See the
lib/python/README file for more info.</p>
<p>If you want to write Python code with callbacks to LAMMPS, then you
must also follow the steps overviewed in the preceeding section (11.1)
for running LAMMPS from Python. I.e. you must build LAMMPS as a
shared library and insure that Python can find the python/lammps.py
file and the shared library.</p>
<hr class="docutils" />
</div>
<div class="section" id="building-lammps-as-a-shared-library">
<span id="py-3"></span><h2>11.3. Building LAMMPS as a shared library<a class="headerlink" href="#building-lammps-as-a-shared-library" title="Permalink to this headline"></a></h2>
<p>Instructions on how to build LAMMPS as a shared library are given in
<a class="reference internal" href="Section_start.html#start-5"><span>Section_start 5</span></a>. A shared library is one
that is dynamically loadable, which is what Python requires to wrap
LAMMPS. On Linux this is a library file that ends in &#8221;.so&#8221;, not &#8221;.a&#8221;.</p>
<p>&gt;From the src directory, type</p>
<div class="highlight-python"><div class="highlight"><pre>make foo mode=shlib
</pre></div>
</div>
<p>where foo is the machine target name, such as linux or g++ or serial.
This should create the file liblammps_foo.so in the src directory, as
well as a soft link liblammps.so, which is what the Python wrapper will
load by default. Note that if you are building multiple machine
versions of the shared library, the soft link is always set to the
most recently built version.</p>
<p>If this fails, see <a class="reference internal" href="Section_start.html#start-5"><span>Section_start 5</span></a> for
more details, especially if your LAMMPS build uses auxiliary libraries
like MPI or FFTW which may not be built as shared libraries on your
system.</p>
<hr class="docutils" />
</div>
<div class="section" id="installing-the-python-wrapper-into-python">
<span id="py-4"></span><h2>11.4. Installing the Python wrapper into Python<a class="headerlink" href="#installing-the-python-wrapper-into-python" title="Permalink to this headline"></a></h2>
<p>For Python to invoke LAMMPS, there are 2 files it needs to know about:</p>
<ul class="simple">
<li>python/lammps.py</li>
<li>src/liblammps.so</li>
</ul>
<p>Lammps.py is the Python wrapper on the LAMMPS library interface.
Liblammps.so is the shared LAMMPS library that Python loads, as
described above.</p>
<p>You can insure Python can find these files in one of two ways:</p>
<ul class="simple">
<li>set two environment variables</li>
<li>run the python/install.py script</li>
</ul>
<p>If you set the paths to these files as environment variables, you only
have to do it once. For the csh or tcsh shells, add something like
this to your ~/.cshrc file, one line for each of the two files:</p>
<div class="highlight-python"><div class="highlight"><pre>setenv PYTHONPATH ${PYTHONPATH}:/home/sjplimp/lammps/python
setenv LD_LIBRARY_PATH ${LD_LIBRARY_PATH}:/home/sjplimp/lammps/src
</pre></div>
</div>
<p>If you use the python/install.py script, you need to invoke it every
time you rebuild LAMMPS (as a shared library) or make changes to the
python/lammps.py file.</p>
<p>You can invoke install.py from the python directory as</p>
<div class="highlight-python"><div class="highlight"><pre>% python install.py [libdir] [pydir]
</pre></div>
</div>
<p>The optional libdir is where to copy the LAMMPS shared library to; the
default is /usr/local/lib. The optional pydir is where to copy the
lammps.py file to; the default is the site-packages directory of the
version of Python that is running the install script.</p>
<p>Note that libdir must be a location that is in your default
LD_LIBRARY_PATH, like /usr/local/lib or /usr/lib. And pydir must be a
location that Python looks in by default for imported modules, like
its site-packages dir. If you want to copy these files to
non-standard locations, such as within your own user space, you will
need to set your PYTHONPATH and LD_LIBRARY_PATH environment variables
accordingly, as above.</p>
<p>If the install.py script does not allow you to copy files into system
directories, prefix the python command with &#8220;sudo&#8221;. If you do this,
make sure that the Python that root runs is the same as the Python you
run. E.g. you may need to do something like</p>
<div class="highlight-python"><div class="highlight"><pre>% sudo /usr/local/bin/python install.py [libdir] [pydir]
</pre></div>
</div>
<p>You can also invoke install.py from the make command in the src
directory as</p>
<div class="highlight-python"><div class="highlight"><pre>% make install-python
</pre></div>
</div>
<p>In this mode you cannot append optional arguments. Again, you may
need to prefix this with &#8220;sudo&#8221;. In this mode you cannot control
which Python is invoked by root.</p>
<p>Note that if you want Python to be able to load different versions of
the LAMMPS shared library (see <a class="reference internal" href="#py-5"><span>this section</span></a> below), you will
need to manually copy files like liblammps_g++.so into the appropriate
system directory. This is not needed if you set the LD_LIBRARY_PATH
environment variable as described above.</p>
<hr class="docutils" />
</div>
<div class="section" id="extending-python-with-mpi-to-run-in-parallel">
<span id="py-5"></span><h2>11.5. Extending Python with MPI to run in parallel<a class="headerlink" href="#extending-python-with-mpi-to-run-in-parallel" title="Permalink to this headline"></a></h2>
<p>If you wish to run LAMMPS in parallel from Python, you need to extend
your Python with an interface to MPI. This also allows you to
make MPI calls directly from Python in your script, if you desire.</p>
<p>There are several Python packages available that purport to wrap MPI
as a library and allow MPI functions to be called from Python.</p>
<p>These include</p>
<ul class="simple">
<li><a class="reference external" href="http://pympi.sourceforge.net/">pyMPI</a></li>
<li><a class="reference external" href="http://code.google.com/p/maroonmpi/">maroonmpi</a></li>
<li><a class="reference external" href="http://code.google.com/p/mpi4py/">mpi4py</a></li>
<li><a class="reference external" href="http://nbcr.sdsc.edu/forum/viewtopic.php?t=89&amp;sid=c997fefc3933bd66204875b436940f16">myMPI</a></li>
<li><a class="reference external" href="http://code.google.com/p/pypar">Pypar</a></li>
</ul>
<p>All of these except pyMPI work by wrapping the MPI library and
exposing (some portion of) its interface to your Python script. This
means Python cannot be used interactively in parallel, since they do
not address the issue of interactive input to multiple instances of
Python running on different processors. The one exception is pyMPI,
which alters the Python interpreter to address this issue, and (I
believe) creates a new alternate executable (in place of &#8220;python&#8221;
itself) as a result.</p>
<p>In principle any of these Python/MPI packages should work to invoke
LAMMPS in parallel and to make MPI calls themselves from a Python
script which is itself running in parallel. However, when I
downloaded and looked at a few of them, their documentation was
incomplete and I had trouble with their installation. It&#8217;s not clear
if some of the packages are still being actively developed and
supported.</p>
<p>The one I recommend, since I have successfully used it with LAMMPS, is
Pypar. Pypar requires the ubiquitous <a class="reference external" href="http://numpy.scipy.org">Numpy package</a> be installed in your Python. After
launching python, type</p>
<div class="highlight-python"><div class="highlight"><pre><span class="kn">import</span> <span class="nn">numpy</span>
</pre></div>
</div>
<p>to see if it is installed. If not, here is how to install it (version
1.3.0b1 as of April 2009). Unpack the numpy tarball and from its
top-level directory, type</p>
<div class="highlight-python"><div class="highlight"><pre>python setup.py build
sudo python setup.py install
</pre></div>
</div>
<p>The &#8220;sudo&#8221; is only needed if required to copy Numpy files into your
Python distribution&#8217;s site-packages directory.</p>
<p>To install Pypar (version pypar-2.1.4_94 as of Aug 2012), unpack it
and from its &#8220;source&#8221; directory, type</p>
<div class="highlight-python"><div class="highlight"><pre>python setup.py build
sudo python setup.py install
</pre></div>
</div>
<p>Again, the &#8220;sudo&#8221; is only needed if required to copy Pypar files into
your Python distribution&#8217;s site-packages directory.</p>
<p>If you have successully installed Pypar, you should be able to run
Python and type</p>
<div class="highlight-python"><div class="highlight"><pre><span class="kn">import</span> <span class="nn">pypar</span>
</pre></div>
</div>
<p>without error. You should also be able to run python in parallel
on a simple test script</p>
<div class="highlight-python"><div class="highlight"><pre>% mpirun -np 4 python test.py
</pre></div>
</div>
<p>where test.py contains the lines</p>
<div class="highlight-python"><div class="highlight"><pre>import pypar
print &quot;Proc %d out of %d procs&quot; % (pypar.rank(),pypar.size())
</pre></div>
</div>
<p>and see one line of output for each processor you run on.</p>
<div class="admonition warning">
<p class="first admonition-title">Warning</p>
<p class="last">To use Pypar and LAMMPS in parallel from Python, you
must insure both are using the same version of MPI. If you only have
one MPI installed on your system, this is not an issue, but it can be
if you have multiple MPIs. Your LAMMPS build is explicit about which
MPI it is using, since you specify the details in your lo-level
src/MAKE/Makefile.foo file. Pypar uses the &#8220;mpicc&#8221; command to find
information about the MPI it uses to build against. And it tries to
load &#8220;libmpi.so&#8221; from the LD_LIBRARY_PATH. This may or may not find
the MPI library that LAMMPS is using. If you have problems running
both Pypar and LAMMPS together, this is an issue you may need to
address, e.g. by moving other MPI installations so that Pypar finds
the right one.</p>
</div>
<hr class="docutils" />
</div>
<div class="section" id="testing-the-python-lammps-interface">
<span id="py-6"></span><h2>11.6. Testing the Python-LAMMPS interface<a class="headerlink" href="#testing-the-python-lammps-interface" title="Permalink to this headline"></a></h2>
<p>To test if LAMMPS is callable from Python, launch Python interactively
and type:</p>
<div class="highlight-python"><div class="highlight"><pre><span class="gp">&gt;&gt;&gt; </span><span class="kn">from</span> <span class="nn">lammps</span> <span class="kn">import</span> <span class="n">lammps</span>
<span class="gp">&gt;&gt;&gt; </span><span class="n">lmp</span> <span class="o">=</span> <span class="n">lammps</span><span class="p">()</span>
</pre></div>
</div>
<p>If you get no errors, you&#8217;re ready to use LAMMPS from Python. If the
2nd command fails, the most common error to see is</p>
<div class="highlight-python"><div class="highlight"><pre>OSError: Could not load LAMMPS dynamic library
</pre></div>
</div>
<p>which means Python was unable to load the LAMMPS shared library. This
typically occurs if the system can&#8217;t find the LAMMPS shared library or
one of the auxiliary shared libraries it depends on, or if something
about the library is incompatible with your Python. The error message
should give you an indication of what went wrong.</p>
<p>You can also test the load directly in Python as follows, without
first importing from the lammps.py file:</p>
<div class="highlight-python"><div class="highlight"><pre><span class="gp">&gt;&gt;&gt; </span><span class="kn">from</span> <span class="nn">ctypes</span> <span class="kn">import</span> <span class="n">CDLL</span>
<span class="gp">&gt;&gt;&gt; </span><span class="n">CDLL</span><span class="p">(</span><span class="s">&quot;liblammps.so&quot;</span><span class="p">)</span>
</pre></div>
</div>
<p>If an error occurs, carefully go thru the steps in <a class="reference internal" href="Section_start.html#start-5"><span>Section_start 5</span></a> and above about building a shared
library and about insuring Python can find the necessary two files
it needs.</p>
<div class="section" id="test-lammps-and-python-in-serial">
<h3>11.6.1. <strong>Test LAMMPS and Python in serial:</strong><a class="headerlink" href="#test-lammps-and-python-in-serial" title="Permalink to this headline"></a></h3>
<p>To run a LAMMPS test in serial, type these lines into Python
interactively from the bench directory:</p>
<div class="highlight-python"><div class="highlight"><pre><span class="gp">&gt;&gt;&gt; </span><span class="kn">from</span> <span class="nn">lammps</span> <span class="kn">import</span> <span class="n">lammps</span>
<span class="gp">&gt;&gt;&gt; </span><span class="n">lmp</span> <span class="o">=</span> <span class="n">lammps</span><span class="p">()</span>
<span class="gp">&gt;&gt;&gt; </span><span class="n">lmp</span><span class="o">.</span><span class="n">file</span><span class="p">(</span><span class="s">&quot;in.lj&quot;</span><span class="p">)</span>
</pre></div>
</div>
<p>Or put the same lines in the file test.py and run it as</p>
<div class="highlight-python"><div class="highlight"><pre>% python test.py
</pre></div>
</div>
<p>Either way, you should see the results of running the in.lj benchmark
on a single processor appear on the screen, the same as if you had
typed something like:</p>
<div class="highlight-python"><div class="highlight"><pre>lmp_g++ -in in.lj
</pre></div>
</div>
</div>
<div class="section" id="test-lammps-and-python-in-parallel">
<h3>11.6.2. <strong>Test LAMMPS and Python in parallel:</strong><a class="headerlink" href="#test-lammps-and-python-in-parallel" title="Permalink to this headline"></a></h3>
<p>To run LAMMPS in parallel, assuming you have installed the
<a class="reference external" href="http://datamining.anu.edu.au/~ole/pypar">Pypar</a> package as discussed
above, create a test.py file containing these lines:</p>
<div class="highlight-python"><div class="highlight"><pre>import pypar
from lammps import lammps
lmp = lammps()
lmp.file(&quot;in.lj&quot;)
print &quot;Proc %d out of %d procs has&quot; % (pypar.rank(),pypar.size()),lmp
pypar.finalize()
</pre></div>
</div>
<p>You can then run it in parallel as:</p>
<div class="highlight-python"><div class="highlight"><pre>% mpirun -np 4 python test.py
</pre></div>
</div>
<p>and you should see the same output as if you had typed</p>
<div class="highlight-python"><div class="highlight"><pre>% mpirun -np 4 lmp_g++ -in in.lj
</pre></div>
</div>
<p>Note that if you leave out the 3 lines from test.py that specify Pypar
commands you will instantiate and run LAMMPS independently on each of
the P processors specified in the mpirun command. In this case you
should get 4 sets of output, each showing that a LAMMPS run was made
on a single processor, instead of one set of output showing that
LAMMPS ran on 4 processors. If the 1-processor outputs occur, it
means that Pypar is not working correctly.</p>
<p>Also note that once you import the PyPar module, Pypar initializes MPI
for you, and you can use MPI calls directly in your Python script, as
described in the Pypar documentation. The last line of your Python
script should be pypar.finalize(), to insure MPI is shut down
correctly.</p>
</div>
<div class="section" id="running-python-scripts">
<h3>11.6.3. <strong>Running Python scripts:</strong><a class="headerlink" href="#running-python-scripts" title="Permalink to this headline"></a></h3>
<p>Note that any Python script (not just for LAMMPS) can be invoked in
one of several ways:</p>
<div class="highlight-python"><div class="highlight"><pre>% python foo.script
% python -i foo.script
% foo.script
</pre></div>
</div>
<p>The last command requires that the first line of the script be
something like this:</p>
<div class="highlight-python"><div class="highlight"><pre><span class="c">#!/usr/local/bin/python</span>
<span class="c">#!/usr/local/bin/python -i</span>
</pre></div>
</div>
<p>where the path points to where you have Python installed, and that you
have made the script file executable:</p>
<div class="highlight-python"><div class="highlight"><pre>% chmod +x foo.script
</pre></div>
</div>
<p>Without the &#8220;-i&#8221; flag, Python will exit when the script finishes.
With the &#8220;-i&#8221; flag, you will be left in the Python interpreter when
the script finishes, so you can type subsequent commands. As
mentioned above, you can only run Python interactively when running
Python on a single processor, not in parallel.</p>
</div>
</div>
<div class="section" id="using-lammps-from-python">
<span id="py-7"></span><h2>11.7. Using LAMMPS from Python<a class="headerlink" href="#using-lammps-from-python" title="Permalink to this headline"></a></h2>
<p>As described above, the Python interface to LAMMPS consists of a
Python &#8220;lammps&#8221; module, the source code for which is in
python/lammps.py, which creates a &#8220;lammps&#8221; object, with a set of
methods that can be invoked on that object. The sample Python code
below assumes you have first imported the &#8220;lammps&#8221; module in your
Python script, as follows:</p>
<div class="highlight-python"><div class="highlight"><pre><span class="kn">from</span> <span class="nn">lammps</span> <span class="kn">import</span> <span class="n">lammps</span>
</pre></div>
</div>
<p>These are the methods defined by the lammps module. If you look at
the files src/library.cpp and src/library.h you will see that they
correspond one-to-one with calls you can make to the LAMMPS library
from a C++ or C or Fortran program.</p>
<div class="highlight-python"><div class="highlight"><pre><span class="n">lmp</span> <span class="o">=</span> <span class="n">lammps</span><span class="p">()</span> <span class="c"># create a LAMMPS object using the default liblammps.so library</span>
<span class="n">lmp</span> <span class="o">=</span> <span class="n">lammps</span><span class="p">(</span><span class="n">ptr</span><span class="o">=</span><span class="n">lmpptr</span><span class="p">)</span> <span class="c"># ditto, but use lmpptr as previously created LAMMPS object</span>
<span class="n">lmp</span> <span class="o">=</span> <span class="n">lammps</span><span class="p">(</span><span class="s">&quot;g++&quot;</span><span class="p">)</span> <span class="c"># create a LAMMPS object using the liblammps_g++.so library</span>
<span class="n">lmp</span> <span class="o">=</span> <span class="n">lammps</span><span class="p">(</span><span class="s">&quot;&quot;</span><span class="p">,</span><span class="nb">list</span><span class="p">)</span> <span class="c"># ditto, with command-line args, e.g. list = [&quot;-echo&quot;,&quot;screen&quot;]</span>
<span class="n">lmp</span> <span class="o">=</span> <span class="n">lammps</span><span class="p">(</span><span class="s">&quot;g++&quot;</span><span class="p">,</span><span class="nb">list</span><span class="p">)</span>
</pre></div>
</div>
<div class="highlight-python"><div class="highlight"><pre><span class="n">lmp</span><span class="o">.</span><span class="n">close</span><span class="p">()</span> <span class="c"># destroy a LAMMPS object</span>
</pre></div>
</div>
<p>version = lmp.version() # return the numerical version id, e.g. LAMMPS 2 Sep 2015 -&gt; 20150902</p>
<div class="highlight-python"><div class="highlight"><pre><span class="n">lmp</span><span class="o">.</span><span class="n">file</span><span class="p">(</span><span class="nb">file</span><span class="p">)</span> <span class="c"># run an entire input script, file = &quot;in.lj&quot;</span>
<span class="n">lmp</span><span class="o">.</span><span class="n">command</span><span class="p">(</span><span class="n">cmd</span><span class="p">)</span> <span class="c"># invoke a single LAMMPS command, cmd = &quot;run 100&quot;</span>
</pre></div>
</div>
<div class="highlight-python"><div class="highlight"><pre><span class="n">xlo</span> <span class="o">=</span> <span class="n">lmp</span><span class="o">.</span><span class="n">extract_global</span><span class="p">(</span><span class="n">name</span><span class="p">,</span><span class="nb">type</span><span class="p">)</span> <span class="c"># extract a global quantity</span>
<span class="c"># name = &quot;boxxlo&quot;, &quot;nlocal&quot;, etc</span>
<span class="c"># type = 0 = int</span>
<span class="c"># 1 = double</span>
</pre></div>
</div>
<div class="highlight-python"><div class="highlight"><pre><span class="n">coords</span> <span class="o">=</span> <span class="n">lmp</span><span class="o">.</span><span class="n">extract_atom</span><span class="p">(</span><span class="n">name</span><span class="p">,</span><span class="nb">type</span><span class="p">)</span> <span class="c"># extract a per-atom quantity</span>
<span class="c"># name = &quot;x&quot;, &quot;type&quot;, etc</span>
<span class="c"># type = 0 = vector of ints</span>
<span class="c"># 1 = array of ints</span>
<span class="c"># 2 = vector of doubles</span>
<span class="c"># 3 = array of doubles</span>
</pre></div>
</div>
<div class="highlight-python"><div class="highlight"><pre><span class="n">eng</span> <span class="o">=</span> <span class="n">lmp</span><span class="o">.</span><span class="n">extract_compute</span><span class="p">(</span><span class="nb">id</span><span class="p">,</span><span class="n">style</span><span class="p">,</span><span class="nb">type</span><span class="p">)</span> <span class="c"># extract value(s) from a compute</span>
<span class="n">v3</span> <span class="o">=</span> <span class="n">lmp</span><span class="o">.</span><span class="n">extract_fix</span><span class="p">(</span><span class="nb">id</span><span class="p">,</span><span class="n">style</span><span class="p">,</span><span class="nb">type</span><span class="p">,</span><span class="n">i</span><span class="p">,</span><span class="n">j</span><span class="p">)</span> <span class="c"># extract value(s) from a fix</span>
<span class="c"># id = ID of compute or fix</span>
<span class="c"># style = 0 = global data</span>
<span class="c"># 1 = per-atom data</span>
<span class="c"># 2 = local data</span>
<span class="c"># type = 0 = scalar</span>
<span class="c"># 1 = vector</span>
<span class="c"># 2 = array</span>
<span class="c"># i,j = indices of value in global vector or array</span>
</pre></div>
</div>
<div class="highlight-python"><div class="highlight"><pre><span class="n">var</span> <span class="o">=</span> <span class="n">lmp</span><span class="o">.</span><span class="n">extract_variable</span><span class="p">(</span><span class="n">name</span><span class="p">,</span><span class="n">group</span><span class="p">,</span><span class="n">flag</span><span class="p">)</span> <span class="c"># extract value(s) from a variable</span>
<span class="c"># name = name of variable</span>
<span class="c"># group = group ID (ignored for equal-style variables)</span>
<span class="c"># flag = 0 = equal-style variable</span>
<span class="c"># 1 = atom-style variable</span>
</pre></div>
</div>
<div class="highlight-python"><div class="highlight"><pre><span class="n">flag</span> <span class="o">=</span> <span class="n">lmp</span><span class="o">.</span><span class="n">set_variable</span><span class="p">(</span><span class="n">name</span><span class="p">,</span><span class="n">value</span><span class="p">)</span> <span class="c"># set existing named string-style variable to value, flag = 0 if successful</span>
<span class="n">natoms</span> <span class="o">=</span> <span class="n">lmp</span><span class="o">.</span><span class="n">get_natoms</span><span class="p">()</span> <span class="c"># total # of atoms as int</span>
<span class="n">data</span> <span class="o">=</span> <span class="n">lmp</span><span class="o">.</span><span class="n">gather_atoms</span><span class="p">(</span><span class="n">name</span><span class="p">,</span><span class="nb">type</span><span class="p">,</span><span class="n">count</span><span class="p">)</span> <span class="c"># return atom attribute of all atoms gathered into data, ordered by atom ID</span>
<span class="c"># name = &quot;x&quot;, &quot;charge&quot;, &quot;type&quot;, etc</span>
<span class="c"># count = # of per-atom values, 1 or 3, etc</span>
<span class="n">lmp</span><span class="o">.</span><span class="n">scatter_atoms</span><span class="p">(</span><span class="n">name</span><span class="p">,</span><span class="nb">type</span><span class="p">,</span><span class="n">count</span><span class="p">,</span><span class="n">data</span><span class="p">)</span> <span class="c"># scatter atom attribute of all atoms from data, ordered by atom ID</span>
<span class="c"># name = &quot;x&quot;, &quot;charge&quot;, &quot;type&quot;, etc</span>
<span class="c"># count = # of per-atom values, 1 or 3, etc</span>
</pre></div>
</div>
<hr class="docutils" />
<div class="admonition warning">
<p class="first admonition-title">Warning</p>
<p class="last">Currently, the creation of a LAMMPS object from within
lammps.py does not take an MPI communicator as an argument. There
should be a way to do this, so that the LAMMPS instance runs on a
subset of processors if desired, but I don&#8217;t know how to do it from
Pypar. So for now, it runs with MPI_COMM_WORLD, which is all the
processors. If someone figures out how to do this with one or more of
the Python wrappers for MPI, like Pypar, please let us know and we
will amend these doc pages.</p>
</div>
<p>The lines</p>
<div class="highlight-python"><div class="highlight"><pre><span class="kn">from</span> <span class="nn">lammps</span> <span class="kn">import</span> <span class="n">lammps</span>
<span class="n">lmp</span> <span class="o">=</span> <span class="n">lammps</span><span class="p">()</span>
</pre></div>
</div>
<p>create an instance of LAMMPS, wrapped in a Python class by the lammps
Python module, and return an instance of the Python class as lmp. It
is used to make all subequent calls to the LAMMPS library.</p>
<p>Additional arguments can be used to tell Python the name of the shared
library to load or to pass arguments to the LAMMPS instance, the same
as if LAMMPS were launched from a command-line prompt.</p>
<p>If the ptr argument is set like this:</p>
<div class="highlight-python"><div class="highlight"><pre><span class="n">lmp</span> <span class="o">=</span> <span class="n">lammps</span><span class="p">(</span><span class="n">ptr</span><span class="o">=</span><span class="n">lmpptr</span><span class="p">)</span>
</pre></div>
</div>
<p>then lmpptr must be an argument passed to Python via the LAMMPS
<a class="reference internal" href="python.html"><em>python</em></a> command, when it is used to define a Python
function that is invoked by the LAMMPS input script. This mode of
using Python with LAMMPS is described above in 11.2. The variable
lmpptr refers to the instance of LAMMPS that called the embedded
Python interpreter. Using it as an argument to lammps() allows the
returned Python class instance &#8220;lmp&#8221; to make calls to that instance of
LAMMPS. See the <a class="reference internal" href="python.html"><em>python</em></a> command doc page for examples
using this syntax.</p>
<p>Note that you can create multiple LAMMPS objects in your Python
script, and coordinate and run multiple simulations, e.g.</p>
<div class="highlight-python"><div class="highlight"><pre><span class="kn">from</span> <span class="nn">lammps</span> <span class="kn">import</span> <span class="n">lammps</span>
<span class="n">lmp1</span> <span class="o">=</span> <span class="n">lammps</span><span class="p">()</span>
<span class="n">lmp2</span> <span class="o">=</span> <span class="n">lammps</span><span class="p">()</span>
<span class="n">lmp1</span><span class="o">.</span><span class="n">file</span><span class="p">(</span><span class="s">&quot;in.file1&quot;</span><span class="p">)</span>
<span class="n">lmp2</span><span class="o">.</span><span class="n">file</span><span class="p">(</span><span class="s">&quot;in.file2&quot;</span><span class="p">)</span>
</pre></div>
</div>
<p>The file() and command() methods allow an input script or single
commands to be invoked.</p>
<p>The extract_global(), extract_atom(), extract_compute(),
extract_fix(), and extract_variable() methods return values or
pointers to data structures internal to LAMMPS.</p>
<p>For extract_global() see the src/library.cpp file for the list of
valid names. New names could easily be added. A double or integer is
returned. You need to specify the appropriate data type via the type
argument.</p>
<p>For extract_atom(), a pointer to internal LAMMPS atom-based data is
returned, which you can use via normal Python subscripting. See the
extract() method in the src/atom.cpp file for a list of valid names.
Again, new names could easily be added. A pointer to a vector of
doubles or integers, or a pointer to an array of doubles (double <a href="#id2"><span class="problematic" id="id3">**</span></a>)
or integers (int <a href="#id4"><span class="problematic" id="id5">**</span></a>) is returned. You need to specify the appropriate
data type via the type argument.</p>
<p>For extract_compute() and extract_fix(), the global, per-atom, or
local data calulated by the compute or fix can be accessed. What is
returned depends on whether the compute or fix calculates a scalar or
vector or array. For a scalar, a single double value is returned. If
the compute or fix calculates a vector or array, a pointer to the
internal LAMMPS data is returned, which you can use via normal Python
subscripting. The one exception is that for a fix that calculates a
global vector or array, a single double value from the vector or array
is returned, indexed by I (vector) or I and J (array). I,J are
zero-based indices. The I,J arguments can be left out if not needed.
See <a class="reference internal" href="Section_howto.html#howto-15"><span>Section_howto 15</span></a> of the manual for a
discussion of global, per-atom, and local data, and of scalar, vector,
and array data types. See the doc pages for individual
<a class="reference internal" href="compute.html"><em>computes</em></a> and <a class="reference internal" href="fix.html"><em>fixes</em></a> for a description of what
they calculate and store.</p>
<p>For extract_variable(), an <a class="reference internal" href="variable.html"><em>equal-style or atom-style variable</em></a> is evaluated and its result returned.</p>
<p>For equal-style variables a single double value is returned and the
group argument is ignored. For atom-style variables, a vector of
doubles is returned, one value per atom, which you can use via normal
Python subscripting. The values will be zero for atoms not in the
specified group.</p>
<p>The get_natoms() method returns the total number of atoms in the
simulation, as an int.</p>
<p>The gather_atoms() method returns a ctypes vector of ints or doubles
as specified by type, of length count*natoms, for the property of all
the atoms in the simulation specified by name, ordered by count and
then by atom ID. The vector can be used via normal Python
subscripting. If atom IDs are not consecutively ordered within
LAMMPS, a None is returned as indication of an error.</p>
<p>Note that the data structure gather_atoms(&#8220;x&#8221;) returns is different
from the data structure returned by extract_atom(&#8220;x&#8221;) in four ways.
(1) Gather_atoms() returns a vector which you index as x[i];
extract_atom() returns an array which you index as x[i][j]. (2)
Gather_atoms() orders the atoms by atom ID while extract_atom() does
not. (3) Gathert_atoms() returns a list of all atoms in the
simulation; extract_atoms() returns just the atoms local to each
processor. (4) Finally, the gather_atoms() data structure is a copy
of the atom coords stored internally in LAMMPS, whereas extract_atom()
returns an array that effectively points directly to the internal
data. This means you can change values inside LAMMPS from Python by
assigning a new values to the extract_atom() array. To do this with
the gather_atoms() vector, you need to change values in the vector,
then invoke the scatter_atoms() method.</p>
<p>The scatter_atoms() method takes a vector of ints or doubles as
specified by type, of length count*natoms, for the property of all the
atoms in the simulation specified by name, ordered by bount and then
by atom ID. It uses the vector of data to overwrite the corresponding
properties for each atom inside LAMMPS. This requires LAMMPS to have
its &#8220;map&#8221; option enabled; see the <a class="reference internal" href="atom_modify.html"><em>atom_modify</em></a>
command for details. If it is not, or if atom IDs are not
consecutively ordered, no coordinates are reset.</p>
<p>The array of coordinates passed to scatter_atoms() must be a ctypes
vector of ints or doubles, allocated and initialized something like
this:</p>
<div class="highlight-python"><div class="highlight"><pre>from ctypes import *
natoms = lmp.get_natoms()
n3 = 3*natoms
x = (n3*c_double)()
x[0] = x coord of atom with ID 1
x[1] = y coord of atom with ID 1
x[2] = z coord of atom with ID 1
x[3] = x coord of atom with ID 2
...
x[n3-1] = z coord of atom with ID natoms
lmp.scatter_coords(&quot;x&quot;,1,3,x)
</pre></div>
</div>
<p>Alternatively, you can just change values in the vector returned by
gather_atoms(&#8220;x&#8221;,1,3), since it is a ctypes vector of doubles.</p>
<hr class="docutils" />
<p>As noted above, these Python class methods correspond one-to-one with
the functions in the LAMMPS library interface in src/library.cpp and
library.h. This means you can extend the Python wrapper via the
following steps:</p>
<ul class="simple">
<li>Add a new interface function to src/library.cpp and
src/library.h.</li>
<li>Rebuild LAMMPS as a shared library.</li>
<li>Add a wrapper method to python/lammps.py for this interface
function.</li>
<li>You should now be able to invoke the new interface function from a
Python script. Isn&#8217;t ctypes amazing?</li>
</ul>
</div>
<div class="section" id="example-python-scripts-that-use-lammps">
<span id="py-8"></span><h2>11.8. Example Python scripts that use LAMMPS<a class="headerlink" href="#example-python-scripts-that-use-lammps" title="Permalink to this headline"></a></h2>
<p>These are the Python scripts included as demos in the python/examples
directory of the LAMMPS distribution, to illustrate the kinds of
things that are possible when Python wraps LAMMPS. If you create your
own scripts, send them to us and we can include them in the LAMMPS
distribution.</p>
<table border="1" class="docutils">
<colgroup>
<col width="27%" />
<col width="73%" />
</colgroup>
<tbody valign="top">
<tr class="row-odd"><td>trivial.py</td>
<td>read/run a LAMMPS input script thru Python</td>
</tr>
<tr class="row-even"><td>demo.py</td>
<td>invoke various LAMMPS library interface routines</td>
</tr>
<tr class="row-odd"><td>simple.py</td>
<td>mimic operation of couple/simple/simple.cpp in Python</td>
</tr>
<tr class="row-even"><td>gui.py</td>
<td>GUI go/stop/temperature-slider to control LAMMPS</td>
</tr>
<tr class="row-odd"><td>plot.py</td>
<td>real-time temeperature plot with GnuPlot via Pizza.py</td>
</tr>
<tr class="row-even"><td>viz_tool.py</td>
<td>real-time viz via some viz package</td>
</tr>
<tr class="row-odd"><td>vizplotgui_tool.py</td>
<td>combination of viz_tool.py and plot.py and gui.py</td>
</tr>
</tbody>
</table>
<hr class="docutils" />
<p>For the viz_tool.py and vizplotgui_tool.py commands, replace &#8220;tool&#8221;
with &#8220;gl&#8221; or &#8220;atomeye&#8221; or &#8220;pymol&#8221; or &#8220;vmd&#8221;, depending on what
visualization package you have installed.</p>
<p>Note that for GL, you need to be able to run the Pizza.py GL tool,
which is included in the pizza sub-directory. See the <a class="reference external" href="http://www.sandia.gov/~sjplimp/pizza.html">Pizza.py doc pages</a> for more info:</p>
<p>Note that for AtomEye, you need version 3, and there is a line in the
scripts that specifies the path and name of the executable. See the
AtomEye WWW pages <a class="reference external" href="http://mt.seas.upenn.edu/Archive/Graphics/A">here</a> or <a class="reference external" href="http://mt.seas.upenn.edu/Archive/Graphics/A3/A3.html">here</a> for more details:</p>
<div class="highlight-python"><div class="highlight"><pre>http://mt.seas.upenn.edu/Archive/Graphics/A
http://mt.seas.upenn.edu/Archive/Graphics/A3/A3.html
</pre></div>
</div>
<p>The latter link is to AtomEye 3 which has the scriping
capability needed by these Python scripts.</p>
<p>Note that for PyMol, you need to have built and installed the
open-source version of PyMol in your Python, so that you can import it
from a Python script. See the PyMol WWW pages <a class="reference external" href="http://www.pymol.org">here</a> or
<a class="reference external" href="http://sourceforge.net/scm/?type=svn&amp;group_id=4546">here</a> for more details:</p>
<div class="highlight-python"><div class="highlight"><pre>http://www.pymol.org
http://sourceforge.net/scm/?type=svn&amp;group_id=4546
</pre></div>
</div>
<p>The latter link is to the open-source version.</p>
<p>Note that for VMD, you need a fairly current version (1.8.7 works for
me) and there are some lines in the pizza/vmd.py script for 4 PIZZA
variables that have to match the VMD installation on your system.</p>
<hr class="docutils" />
<p>See the python/README file for instructions on how to run them and the
source code for individual scripts for comments about what they do.</p>
<p>Here are screenshots of the vizplotgui_tool.py script in action for
different visualization package options. Click to see larger images:</p>
<a data-lightbox="group-default"
href="_images/screenshot_gl.jpg"
class=""
title=""
data-title=""
><img src="_images/screenshot_gl.jpg"
class=""
width="25%"
height="auto"
alt=""/>
</a><a data-lightbox="group-default"
href="_images/screenshot_atomeye.jpg"
class=""
title=""
data-title=""
><img src="_images/screenshot_atomeye.jpg"
class=""
width="25%"
height="auto"
alt=""/>
</a><a data-lightbox="group-default"
href="_images/screenshot_pymol.jpg"
class=""
title=""
data-title=""
><img src="_images/screenshot_pymol.jpg"
class=""
width="25%"
height="auto"
alt=""/>
</a><a data-lightbox="group-default"
href="_images/screenshot_vmd.jpg"
class=""
title=""
data-title=""
><img src="_images/screenshot_vmd.jpg"
class=""
width="25%"
height="auto"
alt=""/>
</a></div>
</div>
</div>
</div>
<footer>
<div class="rst-footer-buttons" role="navigation" aria-label="footer navigation">
<a href="Section_errors.html" class="btn btn-neutral float-right" title="12. Errors" accesskey="n">Next <span class="fa fa-arrow-circle-right"></span></a>
<a href="Section_modify.html" class="btn btn-neutral" title="10. Modifying &amp; extending LAMMPS" accesskey="p"><span class="fa fa-arrow-circle-left"></span> Previous</a>
</div>
<hr/>
<div role="contentinfo">
<p>
&copy; Copyright .
</p>
</div>
Built with <a href="http://sphinx-doc.org/">Sphinx</a> using a <a href="https://github.com/snide/sphinx_rtd_theme">theme</a> provided by <a href="https://readthedocs.org">Read the Docs</a>.
</footer>
</div>
</div>
</section>
</div>
<script type="text/javascript">
var DOCUMENTATION_OPTIONS = {
URL_ROOT:'./',
VERSION:'15 May 2015 version',
COLLAPSE_INDEX:false,
FILE_SUFFIX:'.html',
HAS_SOURCE: true
};
</script>
<script type="text/javascript" src="_static/jquery.js"></script>
<script type="text/javascript" src="_static/underscore.js"></script>
<script type="text/javascript" src="_static/doctools.js"></script>
<script type="text/javascript" src="https://cdn.mathjax.org/mathjax/latest/MathJax.js?config=TeX-AMS-MML_HTMLorMML"></script>
<script type="text/javascript" src="_static/sphinxcontrib-images/LightBox2/lightbox2/js/jquery-1.11.0.min.js"></script>
<script type="text/javascript" src="_static/sphinxcontrib-images/LightBox2/lightbox2/js/lightbox.min.js"></script>
<script type="text/javascript" src="_static/sphinxcontrib-images/LightBox2/lightbox2-customize/jquery-noconflict.js"></script>
<script type="text/javascript" src="_static/js/theme.js"></script>
<script type="text/javascript">
jQuery(function () {
SphinxRtdTheme.StickyNav.enable();
});
</script>
</body>
</html>

File diff suppressed because it is too large Load Diff

View File

@ -1,707 +0,0 @@
<!DOCTYPE html>
<!--[if IE 8]><html class="no-js lt-ie9" lang="en" > <![endif]-->
<!--[if gt IE 8]><!--> <html class="no-js" lang="en" > <!--<![endif]-->
<head>
<meta charset="utf-8">
<meta name="viewport" content="width=device-width, initial-scale=1.0">
<title>9. Additional tools &mdash; LAMMPS 15 May 2015 version documentation</title>
<link rel="stylesheet" href="_static/css/theme.css" type="text/css" />
<link rel="stylesheet" href="_static/sphinxcontrib-images/LightBox2/lightbox2/css/lightbox.css" type="text/css" />
<link rel="top" title="LAMMPS 15 May 2015 version documentation" href="index.html"/>
<link rel="next" title="10. Modifying &amp; extending LAMMPS" href="Section_modify.html"/>
<link rel="prev" title="8. Performance &amp; scalability" href="Section_perf.html"/>
<script src="_static/js/modernizr.min.js"></script>
</head>
<body class="wy-body-for-nav" role="document">
<div class="wy-grid-for-nav">
<nav data-toggle="wy-nav-shift" class="wy-nav-side">
<div class="wy-side-nav-search">
<a href="Manual.html" class="icon icon-home"> LAMMPS
</a>
<div role="search">
<form id="rtd-search-form" class="wy-form" action="search.html" method="get">
<input type="text" name="q" placeholder="Search docs" />
<input type="hidden" name="check_keywords" value="yes" />
<input type="hidden" name="area" value="default" />
</form>
</div>
</div>
<div class="wy-menu wy-menu-vertical" data-spy="affix" role="navigation" aria-label="main navigation">
<ul class="current">
<li class="toctree-l1"><a class="reference internal" href="Section_intro.html">1. Introduction</a></li>
<li class="toctree-l1"><a class="reference internal" href="Section_start.html">2. Getting Started</a></li>
<li class="toctree-l1"><a class="reference internal" href="Section_commands.html">3. Commands</a></li>
<li class="toctree-l1"><a class="reference internal" href="Section_packages.html">4. Packages</a></li>
<li class="toctree-l1"><a class="reference internal" href="Section_accelerate.html">5. Accelerating LAMMPS performance</a></li>
<li class="toctree-l1"><a class="reference internal" href="Section_howto.html">6. How-to discussions</a></li>
<li class="toctree-l1"><a class="reference internal" href="Section_example.html">7. Example problems</a></li>
<li class="toctree-l1"><a class="reference internal" href="Section_perf.html">8. Performance &amp; scalability</a></li>
<li class="toctree-l1 current"><a class="current reference internal" href="">9. Additional tools</a><ul>
<li class="toctree-l2"><a class="reference internal" href="#amber2lmp-tool">9.1. amber2lmp tool</a></li>
<li class="toctree-l2"><a class="reference internal" href="#binary2txt-tool">9.2. binary2txt tool</a></li>
<li class="toctree-l2"><a class="reference internal" href="#ch2lmp-tool">9.3. ch2lmp tool</a></li>
<li class="toctree-l2"><a class="reference internal" href="#chain-tool">9.4. chain tool</a></li>
<li class="toctree-l2"><a class="reference internal" href="#colvars-tools">9.5. colvars tools</a></li>
<li class="toctree-l2"><a class="reference internal" href="#createatoms-tool">9.6. createatoms tool</a></li>
<li class="toctree-l2"><a class="reference internal" href="#data2xmovie-tool">9.7. data2xmovie tool</a></li>
<li class="toctree-l2"><a class="reference internal" href="#eam-database-tool">9.8. eam database tool</a></li>
<li class="toctree-l2"><a class="reference internal" href="#eam-generate-tool">9.9. eam generate tool</a></li>
<li class="toctree-l2"><a class="reference internal" href="#eff-tool">9.10. eff tool</a></li>
<li class="toctree-l2"><a class="reference internal" href="#emacs-tool">9.11. emacs tool</a></li>
<li class="toctree-l2"><a class="reference internal" href="#fep-tool">9.12. fep tool</a></li>
<li class="toctree-l2"><a class="reference internal" href="#i-pi-tool">9.13. i-pi tool</a></li>
<li class="toctree-l2"><a class="reference internal" href="#ipp-tool">9.14. ipp tool</a></li>
<li class="toctree-l2"><a class="reference internal" href="#kate-tool">9.15. kate tool</a></li>
<li class="toctree-l2"><a class="reference internal" href="#lmp2arc-tool">9.16. lmp2arc tool</a></li>
<li class="toctree-l2"><a class="reference internal" href="#lmp2cfg-tool">9.17. lmp2cfg tool</a></li>
<li class="toctree-l2"><a class="reference internal" href="#lmp2vmd-tool">9.18. lmp2vmd tool</a></li>
<li class="toctree-l2"><a class="reference internal" href="#matlab-tool">9.19. matlab tool</a></li>
<li class="toctree-l2"><a class="reference internal" href="#micelle2d-tool">9.20. micelle2d tool</a></li>
<li class="toctree-l2"><a class="reference internal" href="#moltemplate-tool">9.21. moltemplate tool</a></li>
<li class="toctree-l2"><a class="reference internal" href="#msi2lmp-tool">9.22. msi2lmp tool</a></li>
<li class="toctree-l2"><a class="reference internal" href="#phonon-tool">9.23. phonon tool</a></li>
<li class="toctree-l2"><a class="reference internal" href="#polymer-bonding-tool">9.24. polymer bonding tool</a></li>
<li class="toctree-l2"><a class="reference internal" href="#pymol-asphere-tool">9.25. pymol_asphere tool</a></li>
<li class="toctree-l2"><a class="reference internal" href="#python-tool">9.26. python tool</a></li>
<li class="toctree-l2"><a class="reference internal" href="#reax-tool">9.27. reax tool</a></li>
<li class="toctree-l2"><a class="reference internal" href="#restart2data-tool">9.28. restart2data tool</a></li>
<li class="toctree-l2"><a class="reference internal" href="#vim-tool">9.29. vim tool</a></li>
<li class="toctree-l2"><a class="reference internal" href="#xmgrace-tool">9.30. xmgrace tool</a></li>
<li class="toctree-l2"><a class="reference internal" href="#xmovie-tool">9.31. xmovie tool</a></li>
</ul>
</li>
<li class="toctree-l1"><a class="reference internal" href="Section_modify.html">10. Modifying &amp; extending LAMMPS</a></li>
<li class="toctree-l1"><a class="reference internal" href="Section_python.html">11. Python interface to LAMMPS</a></li>
<li class="toctree-l1"><a class="reference internal" href="Section_errors.html">12. Errors</a></li>
<li class="toctree-l1"><a class="reference internal" href="Section_history.html">13. Future and history</a></li>
</ul>
</div>
&nbsp;
</nav>
<section data-toggle="wy-nav-shift" class="wy-nav-content-wrap">
<nav class="wy-nav-top" role="navigation" aria-label="top navigation">
<i data-toggle="wy-nav-top" class="fa fa-bars"></i>
<a href="Manual.html">LAMMPS</a>
</nav>
<div class="wy-nav-content">
<div class="rst-content">
<div role="navigation" aria-label="breadcrumbs navigation">
<ul class="wy-breadcrumbs">
<li><a href="Manual.html">Docs</a> &raquo;</li>
<li>9. Additional tools</li>
<li class="wy-breadcrumbs-aside">
<a href="http://lammps.sandia.gov">Website</a>
<a href="Section_commands.html#comm">Commands</a>
</li>
</ul>
<hr/>
<div class="rst-footer-buttons" style="margin-bottom: 1em" role="navigation" aria-label="footer navigation">
<a href="Section_modify.html" class="btn btn-neutral float-right" title="10. Modifying &amp; extending LAMMPS" accesskey="n">Next <span class="fa fa-arrow-circle-right"></span></a>
<a href="Section_perf.html" class="btn btn-neutral" title="8. Performance &amp; scalability" accesskey="p"><span class="fa fa-arrow-circle-left"></span> Previous</a>
</div>
</div>
<div role="main" class="document" itemscope="itemscope" itemtype="http://schema.org/Article">
<div itemprop="articleBody">
<div class="section" id="additional-tools">
<h1>9. Additional tools<a class="headerlink" href="#additional-tools" title="Permalink to this headline"></a></h1>
<p>LAMMPS is designed to be a computational kernel for performing
molecular dynamics computations. Additional pre- and post-processing
steps are often necessary to setup and analyze a simulation. A few
additional tools are provided with the LAMMPS distribution and are
described in this section.</p>
<p>Our group has also written and released a separate toolkit called
<a class="reference external" href="http://www.sandia.gov/~sjplimp/pizza.html">Pizza.py</a> which provides tools for doing setup, analysis,
plotting, and visualization for LAMMPS simulations. Pizza.py is
written in <a class="reference external" href="http://www.python.org">Python</a> and is available for download from <a class="reference external" href="http://www.sandia.gov/~sjplimp/pizza.html">the Pizza.py WWW site</a>.</p>
<p>Note that many users write their own setup or analysis tools or use
other existing codes and convert their output to a LAMMPS input format
or vice versa. The tools listed here are included in the LAMMPS
distribution as examples of auxiliary tools. Some of them are not
actively supported by Sandia, as they were contributed by LAMMPS
users. If you have problems using them, we can direct you to the
authors.</p>
<p>The source code for each of these codes is in the tools sub-directory
of the LAMMPS distribution. There is a Makefile (which you may need
to edit for your platform) which will build several of the tools which
reside in that directory. Some of them are larger packages in their
own sub-directories with their own Makefiles.</p>
<ul class="simple">
<li><a class="reference internal" href="#amber"><span>amber2lmp</span></a></li>
<li><a class="reference internal" href="#binary"><span>binary2txt</span></a></li>
<li><a class="reference internal" href="#charmm"><span>ch2lmp</span></a></li>
<li><a class="reference internal" href="#chain"><span>chain</span></a></li>
<li><a class="reference internal" href="#colvars"><span>colvars</span></a></li>
<li><a class="reference internal" href="#create"><span>createatoms</span></a></li>
<li><a class="reference internal" href="#data"><span>data2xmovie</span></a></li>
<li><a class="reference internal" href="#eamdb"><span>eam database</span></a></li>
<li><a class="reference internal" href="#eamgn"><span>eam generate</span></a></li>
<li><a class="reference internal" href="#eff"><span>eff</span></a></li>
<li><a class="reference internal" href="#emacs"><span>emacs</span></a></li>
<li><a class="reference internal" href="#fep"><span>fep</span></a></li>
<li><a class="reference internal" href="fix_ipi.html#ipi"><span>i-pi</span></a></li>
<li><a class="reference internal" href="#ipp"><span>ipp</span></a></li>
<li><a class="reference internal" href="#kate"><span>kate</span></a></li>
<li><a class="reference internal" href="#arc"><span>lmp2arc</span></a></li>
<li><a class="reference internal" href="#cfg"><span>lmp2cfg</span></a></li>
<li><a class="reference internal" href="#vmd"><span>lmp2vmd</span></a></li>
<li><span class="xref std std-ref">matlab</span></li>
<li><a class="reference internal" href="#micelle"><span>micelle2d</span></a></li>
<li><a class="reference internal" href="#moltemplate"><span>moltemplate</span></a></li>
<li><a class="reference internal" href="#msi"><span>msi2lmp</span></a></li>
<li><a class="reference internal" href="#phonon"><span>phonon</span></a></li>
<li><a class="reference internal" href="#polybond"><span>polymer bonding</span></a></li>
<li><span class="xref std std-ref">pymol_asphere</span></li>
<li><a class="reference internal" href="#pythontools"><span>python</span></a></li>
<li><a class="reference internal" href="#reax"><span>reax</span></a></li>
<li><a class="reference internal" href="#restart"><span>restart2data</span></a></li>
<li><a class="reference internal" href="#vim"><span>vim</span></a></li>
<li><a class="reference internal" href="#xmgrace"><span>xmgrace</span></a></li>
<li><a class="reference internal" href="#xmovie"><span>xmovie</span></a></li>
</ul>
<hr class="docutils" />
<div class="section" id="amber2lmp-tool">
<span id="amber"></span><h2>9.1. amber2lmp tool<a class="headerlink" href="#amber2lmp-tool" title="Permalink to this headline"></a></h2>
<p>The amber2lmp sub-directory contains two Python scripts for converting
files back-and-forth between the AMBER MD code and LAMMPS. See the
README file in amber2lmp for more information.</p>
<p>These tools were written by Keir Novik while he was at Queen Mary
University of London. Keir is no longer there and cannot support
these tools which are out-of-date with respect to the current LAMMPS
version (and maybe with respect to AMBER as well). Since we don&#8217;t use
these tools at Sandia, you&#8217;ll need to experiment with them and make
necessary modifications yourself.</p>
<hr class="docutils" />
</div>
<div class="section" id="binary2txt-tool">
<span id="binary"></span><h2>9.2. binary2txt tool<a class="headerlink" href="#binary2txt-tool" title="Permalink to this headline"></a></h2>
<p>The file binary2txt.cpp converts one or more binary LAMMPS dump file
into ASCII text files. The syntax for running the tool is</p>
<div class="highlight-python"><div class="highlight"><pre>binary2txt file1 file2 ...
</pre></div>
</div>
<p>which creates file1.txt, file2.txt, etc. This tool must be compiled
on a platform that can read the binary file created by a LAMMPS run,
since binary files are not compatible across all platforms.</p>
<hr class="docutils" />
</div>
<div class="section" id="ch2lmp-tool">
<span id="charmm"></span><h2>9.3. ch2lmp tool<a class="headerlink" href="#ch2lmp-tool" title="Permalink to this headline"></a></h2>
<p>The ch2lmp sub-directory contains tools for converting files
back-and-forth between the CHARMM MD code and LAMMPS.</p>
<p>They are intended to make it easy to use CHARMM as a builder and as a
post-processor for LAMMPS. Using charmm2lammps.pl, you can convert an
ensemble built in CHARMM into its LAMMPS equivalent. Using
lammps2pdb.pl you can convert LAMMPS atom dumps into pdb files.</p>
<p>See the README file in the ch2lmp sub-directory for more information.</p>
<p>These tools were created by Pieter in&#8217;t Veld (pjintve at sandia.gov)
and Paul Crozier (pscrozi at sandia.gov) at Sandia.</p>
<hr class="docutils" />
</div>
<div class="section" id="chain-tool">
<span id="chain"></span><h2>9.4. chain tool<a class="headerlink" href="#chain-tool" title="Permalink to this headline"></a></h2>
<p>The file chain.f creates a LAMMPS data file containing bead-spring
polymer chains and/or monomer solvent atoms. It uses a text file
containing chain definition parameters as an input. The created
chains and solvent atoms can strongly overlap, so LAMMPS needs to run
the system initially with a &#8220;soft&#8221; pair potential to un-overlap it.
The syntax for running the tool is</p>
<div class="highlight-python"><div class="highlight"><pre>chain &lt; def.chain &gt; data.file
</pre></div>
</div>
<p>See the def.chain or def.chain.ab files in the tools directory for
examples of definition files. This tool was used to create the
system for the <a class="reference internal" href="Section_perf.html"><em>chain benchmark</em></a>.</p>
<hr class="docutils" />
</div>
<div class="section" id="colvars-tools">
<span id="colvars"></span><h2>9.5. colvars tools<a class="headerlink" href="#colvars-tools" title="Permalink to this headline"></a></h2>
<p>The colvars directory contains a collection of tools for postprocessing
data produced by the colvars collective variable library.
To compile the tools, edit the makefile for your system and run &#8220;make&#8221;.</p>
<p>Please report problems and issues the colvars library and its tools
at: <a class="reference external" href="https://github.com/colvars/colvars/issues">https://github.com/colvars/colvars/issues</a></p>
<p>abf_integrate:</p>
<p>MC-based integration of multidimensional free energy gradient
Version 20110511</p>
<div class="highlight-python"><div class="highlight"><pre>Syntax: ./abf_integrate &lt; filename &gt; [-n &lt; nsteps &gt;] [-t &lt; temp &gt;] [-m [0|1] (metadynamics)] [-h &lt; hill_height &gt;] [-f &lt; variable_hill_factor &gt;]
</pre></div>
</div>
<p>The LAMMPS interface to the colvars collective variable library, as
well as these tools, were created by Axel Kohlmeyer (akohlmey at
gmail.com) at ICTP, Italy.</p>
<hr class="docutils" />
</div>
<div class="section" id="createatoms-tool">
<span id="create"></span><h2>9.6. createatoms tool<a class="headerlink" href="#createatoms-tool" title="Permalink to this headline"></a></h2>
<p>The tools/createatoms directory contains a Fortran program called
createAtoms.f which can generate a variety of interesting crystal
structures and geometries and output the resulting list of atom
coordinates in LAMMPS or other formats.</p>
<p>See the included Manual.pdf for details.</p>
<p>The tool is authored by Xiaowang Zhou (Sandia), xzhou at sandia.gov.</p>
<hr class="docutils" />
</div>
<div class="section" id="data2xmovie-tool">
<span id="data"></span><h2>9.7. data2xmovie tool<a class="headerlink" href="#data2xmovie-tool" title="Permalink to this headline"></a></h2>
<p>The file data2xmovie.c converts a LAMMPS data file into a snapshot
suitable for visualizing with the <a class="reference internal" href="#xmovie"><span>xmovie</span></a> tool, as if it had
been output with a dump command from LAMMPS itself. The syntax for
running the tool is</p>
<div class="highlight-python"><div class="highlight"><pre><span class="n">data2xmovie</span> <span class="p">[</span><span class="n">options</span><span class="p">]</span> <span class="o">&lt;</span> <span class="n">infile</span> <span class="o">&gt;</span> <span class="n">outfile</span>
</pre></div>
</div>
<p>See the top of the data2xmovie.c file for a discussion of the options.</p>
<hr class="docutils" />
</div>
<div class="section" id="eam-database-tool">
<span id="eamdb"></span><h2>9.8. eam database tool<a class="headerlink" href="#eam-database-tool" title="Permalink to this headline"></a></h2>
<p>The tools/eam_database directory contains a Fortran program that will
generate EAM alloy setfl potential files for any combination of 16
elements: Cu, Ag, Au, Ni, Pd, Pt, Al, Pb, Fe, Mo, Ta, W, Mg, Co, Ti,
Zr. The files can then be used with the <a class="reference internal" href="pair_eam.html"><em>pair_style eam/alloy</em></a> command.</p>
<p>The tool is authored by Xiaowang Zhou (Sandia), xzhou at sandia.gov,
and is based on his paper:</p>
<p>X. W. Zhou, R. A. Johnson, and H. N. G. Wadley, Phys. Rev. B, 69,
144113 (2004).</p>
<hr class="docutils" />
</div>
<div class="section" id="eam-generate-tool">
<span id="eamgn"></span><h2>9.9. eam generate tool<a class="headerlink" href="#eam-generate-tool" title="Permalink to this headline"></a></h2>
<p>The tools/eam_generate directory contains several one-file C programs
that convert an analytic formula into a tabulated <a class="reference internal" href="pair_eam.html"><em>embedded atom method (EAM)</em></a> setfl potential file. The potentials they
produce are in the potentials directory, and can be used with the
<a class="reference internal" href="pair_eam.html"><em>pair_style eam/alloy</em></a> command.</p>
<p>The source files and potentials were provided by Gerolf Ziegenhain
(gerolf at ziegenhain.com).</p>
<hr class="docutils" />
</div>
<div class="section" id="eff-tool">
<span id="eff"></span><h2>9.10. eff tool<a class="headerlink" href="#eff-tool" title="Permalink to this headline"></a></h2>
<p>The tools/eff directory contains various scripts for generating
structures and post-processing output for simulations using the
electron force field (eFF).</p>
<p>These tools were provided by Andres Jaramillo-Botero at CalTech
(ajaramil at wag.caltech.edu).</p>
<hr class="docutils" />
</div>
<div class="section" id="emacs-tool">
<span id="emacs"></span><h2>9.11. emacs tool<a class="headerlink" href="#emacs-tool" title="Permalink to this headline"></a></h2>
<p>The tools/emacs directory contains a Lips add-on file for Emacs that
enables a lammps-mode for editing of input scripts when using Emacs,
with various highlighting options setup.</p>
<p>These tools were provided by Aidan Thompson at Sandia
(athomps at sandia.gov).</p>
<hr class="docutils" />
</div>
<div class="section" id="fep-tool">
<span id="fep"></span><h2>9.12. fep tool<a class="headerlink" href="#fep-tool" title="Permalink to this headline"></a></h2>
<p>The tools/fep directory contains Python scripts useful for
post-processing results from performing free-energy perturbation
simulations using the USER-FEP package.</p>
<p>The scripts were contributed by Agilio Padua (Universite Blaise
Pascal Clermont-Ferrand), agilio.padua at univ-bpclermont.fr.</p>
<p>See README file in the tools/fep directory.</p>
<hr class="docutils" />
</div>
<div class="section" id="i-pi-tool">
<span id="ipi"></span><h2>9.13. i-pi tool<a class="headerlink" href="#i-pi-tool" title="Permalink to this headline"></a></h2>
<p>The tools/i-pi directory contains a version of the i-PI package, with
all the LAMMPS-unrelated files removed. It is provided so that it can
be used with the <a class="reference internal" href="fix_ipi.html"><em>fix ipi</em></a> command to perform
path-integral molecular dynamics (PIMD).</p>
<p>The i-PI package was created and is maintained by Michele Ceriotti,
michele.ceriotti at gmail.com, to interface to a variety of molecular
dynamics codes.</p>
<p>See the tools/i-pi/manual.pdf file for an overview of i-PI, and the
<a class="reference internal" href="fix_ipi.html"><em>fix ipi</em></a> doc page for further details on running PIMD
calculations with LAMMPS.</p>
<hr class="docutils" />
</div>
<div class="section" id="ipp-tool">
<span id="ipp"></span><h2>9.14. ipp tool<a class="headerlink" href="#ipp-tool" title="Permalink to this headline"></a></h2>
<p>The tools/ipp directory contains a Perl script ipp which can be used
to facilitate the creation of a complicated file (say, a lammps input
script or tools/createatoms input file) using a template file.</p>
<p>ipp was created and is maintained by Reese Jones (Sandia), rjones at
sandia.gov.</p>
<p>See two examples in the tools/ipp directory. One of them is for the
tools/createatoms tool&#8217;s input file.</p>
<hr class="docutils" />
</div>
<div class="section" id="kate-tool">
<span id="kate"></span><h2>9.15. kate tool<a class="headerlink" href="#kate-tool" title="Permalink to this headline"></a></h2>
<p>The file in the tools/kate directory is an add-on to the Kate editor
in the KDE suite that allow syntax highlighting of LAMMPS input
scripts. See the README.txt file for details.</p>
<p>The file was provided by Alessandro Luigi Sellerio
(alessandro.sellerio at ieni.cnr.it).</p>
<hr class="docutils" />
</div>
<div class="section" id="lmp2arc-tool">
<span id="arc"></span><h2>9.16. lmp2arc tool<a class="headerlink" href="#lmp2arc-tool" title="Permalink to this headline"></a></h2>
<p>The lmp2arc sub-directory contains a tool for converting LAMMPS output
files to the format for Accelrys&#8217; Insight MD code (formerly
MSI/Biosym and its Discover MD code). See the README file for more
information.</p>
<p>This tool was written by John Carpenter (Cray), Michael Peachey
(Cray), and Steve Lustig (Dupont). John is now at the Mayo Clinic
(jec at mayo.edu), but still fields questions about the tool.</p>
<p>This tool was updated for the current LAMMPS C++ version by Jeff
Greathouse at Sandia (jagreat at sandia.gov).</p>
<hr class="docutils" />
</div>
<div class="section" id="lmp2cfg-tool">
<span id="cfg"></span><h2>9.17. lmp2cfg tool<a class="headerlink" href="#lmp2cfg-tool" title="Permalink to this headline"></a></h2>
<p>The lmp2cfg sub-directory contains a tool for converting LAMMPS output
files into a series of <a href="#id1"><span class="problematic" id="id2">*</span></a>.cfg files which can be read into the
<a class="reference external" href="http://mt.seas.upenn.edu/Archive/Graphics/A">AtomEye</a> visualizer. See
the README file for more information.</p>
<p>This tool was written by Ara Kooser at Sandia (askoose at sandia.gov).</p>
<hr class="docutils" />
</div>
<div class="section" id="lmp2vmd-tool">
<span id="vmd"></span><h2>9.18. lmp2vmd tool<a class="headerlink" href="#lmp2vmd-tool" title="Permalink to this headline"></a></h2>
<p>The lmp2vmd sub-directory contains a README.txt file that describes
details of scripts and plugin support within the <a class="reference external" href="http://www.ks.uiuc.edu/Research/vmd">VMD package</a> for visualizing LAMMPS
dump files.</p>
<p>The VMD plugins and other supporting scripts were written by Axel
Kohlmeyer (akohlmey at cmm.chem.upenn.edu) at U Penn.</p>
<hr class="docutils" />
</div>
<div class="section" id="matlab-tool">
<span id="matlab"></span><h2>9.19. matlab tool<a class="headerlink" href="#matlab-tool" title="Permalink to this headline"></a></h2>
<p>The matlab sub-directory contains several <span class="xref std std-ref">MATLAB</span> scripts for
post-processing LAMMPS output. The scripts include readers for log
and dump files, a reader for EAM potential files, and a converter that
reads LAMMPS dump files and produces CFG files that can be visualized
with the <a class="reference external" href="http://mt.seas.upenn.edu/Archive/Graphics/A">AtomEye</a>
visualizer.</p>
<p>See the README.pdf file for more information.</p>
<p>These scripts were written by Arun Subramaniyan at Purdue Univ
(asubrama at purdue.edu).</p>
<hr class="docutils" />
</div>
<div class="section" id="micelle2d-tool">
<span id="micelle"></span><h2>9.20. micelle2d tool<a class="headerlink" href="#micelle2d-tool" title="Permalink to this headline"></a></h2>
<p>The file micelle2d.f creates a LAMMPS data file containing short lipid
chains in a monomer solution. It uses a text file containing lipid
definition parameters as an input. The created molecules and solvent
atoms can strongly overlap, so LAMMPS needs to run the system
initially with a &#8220;soft&#8221; pair potential to un-overlap it. The syntax
for running the tool is</p>
<div class="highlight-python"><div class="highlight"><pre>micelle2d &lt; def.micelle2d &gt; data.file
</pre></div>
</div>
<p>See the def.micelle2d file in the tools directory for an example of a
definition file. This tool was used to create the system for the
<a class="reference internal" href="Section_example.html"><em>micelle example</em></a>.</p>
<hr class="docutils" />
</div>
<div class="section" id="moltemplate-tool">
<span id="moltemplate"></span><h2>9.21. moltemplate tool<a class="headerlink" href="#moltemplate-tool" title="Permalink to this headline"></a></h2>
<p>The moltemplate sub-directory contains a Python-based tool for
building molecular systems based on a text-file description, and
creating LAMMPS data files that encode their molecular topology as
lists of bonds, angles, dihedrals, etc. See the README.TXT file for
more information.</p>
<p>This tool was written by Andrew Jewett (jewett.aij at gmail.com), who
supports it. It has its own WWW page at
<a class="reference external" href="http://moltemplate.org">http://moltemplate.org</a>.</p>
<hr class="docutils" />
</div>
<div class="section" id="msi2lmp-tool">
<span id="msi"></span><h2>9.22. msi2lmp tool<a class="headerlink" href="#msi2lmp-tool" title="Permalink to this headline"></a></h2>
<p>The msi2lmp sub-directory contains a tool for creating LAMMPS input
data files from Accelrys&#8217; Insight MD code (formerly MSI/Biosym and
its Discover MD code). See the README file for more information.</p>
<p>This tool was written by John Carpenter (Cray), Michael Peachey
(Cray), and Steve Lustig (Dupont). John is now at the Mayo Clinic
(jec at mayo.edu), but still fields questions about the tool.</p>
<p>This tool may be out-of-date with respect to the current LAMMPS and
Insight versions. Since we don&#8217;t use it at Sandia, you&#8217;ll need to
experiment with it yourself.</p>
<hr class="docutils" />
</div>
<div class="section" id="phonon-tool">
<span id="phonon"></span><h2>9.23. phonon tool<a class="headerlink" href="#phonon-tool" title="Permalink to this headline"></a></h2>
<p>The phonon sub-directory contains a post-processing tool useful for
analyzing the output of the <a class="reference internal" href="fix_phonon.html"><em>fix phonon</em></a> command in
the USER-PHONON package.</p>
<p>See the README file for instruction on building the tool and what
library it needs. And see the examples/USER/phonon directory
for example problems that can be post-processed with this tool.</p>
<p>This tool was written by Ling-Ti Kong at Shanghai Jiao Tong
University.</p>
<hr class="docutils" />
</div>
<div class="section" id="polymer-bonding-tool">
<span id="polybond"></span><h2>9.24. polymer bonding tool<a class="headerlink" href="#polymer-bonding-tool" title="Permalink to this headline"></a></h2>
<p>The polybond sub-directory contains a Python-based tool useful for
performing &#8220;programmable polymer bonding&#8221;. The Python file
lmpsdata.py provides a &#8220;Lmpsdata&#8221; class with various methods which can
be invoked by a user-written Python script to create data files with
complex bonding topologies.</p>
<p>See the Manual.pdf for details and example scripts.</p>
<p>This tool was written by Zachary Kraus at Georgia Tech.</p>
<hr class="docutils" />
</div>
<div class="section" id="pymol-asphere-tool">
<span id="pymol"></span><h2>9.25. pymol_asphere tool<a class="headerlink" href="#pymol-asphere-tool" title="Permalink to this headline"></a></h2>
<p>The pymol_asphere sub-directory contains a tool for converting a
LAMMPS dump file that contains orientation info for ellipsoidal
particles into an input file for the <span class="xref std std-ref">PyMol visualization package</span>.</p>
<p>Specifically, the tool triangulates the ellipsoids so they can be
viewed as true ellipsoidal particles within PyMol. See the README and
examples directory within pymol_asphere for more information.</p>
<p>This tool was written by Mike Brown at Sandia.</p>
<hr class="docutils" />
</div>
<div class="section" id="python-tool">
<span id="pythontools"></span><h2>9.26. python tool<a class="headerlink" href="#python-tool" title="Permalink to this headline"></a></h2>
<p>The python sub-directory contains several Python scripts
that perform common LAMMPS post-processing tasks, such as:</p>
<ul class="simple">
<li>extract thermodynamic info from a log file as columns of numbers</li>
<li>plot two columns of thermodynamic info from a log file using GnuPlot</li>
<li>sort the snapshots in a dump file by atom ID</li>
<li>convert multiple <a class="reference internal" href="neb.html"><em>NEB</em></a> dump files into one dump file for viz</li>
<li>convert dump files into XYZ, CFG, or PDB format for viz by other packages</li>
</ul>
<p>These are simple scripts built on <a class="reference external" href="http://www.sandia.gov/~sjplimp/pizza.html">Pizza.py</a> modules. See the
README for more info on Pizza.py and how to use these scripts.</p>
<hr class="docutils" />
</div>
<div class="section" id="reax-tool">
<span id="reax"></span><h2>9.27. reax tool<a class="headerlink" href="#reax-tool" title="Permalink to this headline"></a></h2>
<p>The reax sub-directory contains stand-alond codes that can
post-process the output of the <a class="reference internal" href="fix_reax_bonds.html"><em>fix reax/bonds</em></a>
command from a LAMMPS simulation using <a class="reference internal" href="pair_reax.html"><em>ReaxFF</em></a>. See
the README.txt file for more info.</p>
<p>These tools were written by Aidan Thompson at Sandia.</p>
<hr class="docutils" />
</div>
<div class="section" id="restart2data-tool">
<span id="restart"></span><h2>9.28. restart2data tool<a class="headerlink" href="#restart2data-tool" title="Permalink to this headline"></a></h2>
<div class="admonition warning">
<p class="first admonition-title">Warning</p>
<p class="last">This tool is now obsolete and is not included in the
current LAMMPS distribution. This is becaues there is now a
<a class="reference internal" href="write_data.html"><em>write_data</em></a> command, which can create a data file
from within an input script. Running LAMMPS with the &#8220;-r&#8221;
<a class="reference internal" href="Section_start.html#start-7"><span>command-line switch</span></a> as follows:</p>
</div>
<p>lmp_g++ -r restartfile datafile</p>
<p>is the same as running a 2-line input script:</p>
<p>read_restart restartfile
write_data datafile</p>
<p>which will produce the same data file that the restart2data tool used
to create. The following information is included in case you have an
older version of LAMMPS which still includes the restart2data tool.</p>
<p>The file restart2data.cpp converts a binary LAMMPS restart file into
an ASCII data file. The syntax for running the tool is</p>
<div class="highlight-python"><div class="highlight"><pre>restart2data restart-file data-file (input-file)
</pre></div>
</div>
<p>Input-file is optional and if specified will contain LAMMPS input
commands for the masses and force field parameters, instead of putting
those in the data-file. Only a few force field styles currently
support this option.</p>
<p>This tool must be compiled on a platform that can read the binary file
created by a LAMMPS run, since binary files are not compatible across
all platforms.</p>
<p>Note that a text data file has less precision than a binary restart
file. Hence, continuing a run from a converted data file will
typically not conform as closely to a previous run as will restarting
from a binary restart file.</p>
<p>If a &#8220;%&#8221; appears in the specified restart-file, the tool expects a set
of multiple files to exist. See the <a class="reference internal" href="restart.html"><em>restart</em></a> and
<a class="reference internal" href="write_restart.html"><em>write_restart</em></a> commands for info on how such sets
of files are written by LAMMPS, and how the files are named.</p>
<hr class="docutils" />
</div>
<div class="section" id="vim-tool">
<span id="vim"></span><h2>9.29. vim tool<a class="headerlink" href="#vim-tool" title="Permalink to this headline"></a></h2>
<p>The files in the tools/vim directory are add-ons to the VIM editor
that allow easier editing of LAMMPS input scripts. See the README.txt
file for details.</p>
<p>These files were provided by Gerolf Ziegenhain (gerolf at
ziegenhain.com)</p>
<hr class="docutils" />
</div>
<div class="section" id="xmgrace-tool">
<span id="xmgrace"></span><h2>9.30. xmgrace tool<a class="headerlink" href="#xmgrace-tool" title="Permalink to this headline"></a></h2>
<p>The files in the tools/xmgrace directory can be used to plot the
thermodynamic data in LAMMPS log files via the xmgrace plotting
package. There are several tools in the directory that can be used in
post-processing mode. The lammpsplot.cpp file can be compiled and
used to create plots from the current state of a running LAMMPS
simulation.</p>
<p>See the README file for details.</p>
<p>These files were provided by Vikas Varshney (vv0210 at gmail.com)</p>
<hr class="docutils" />
</div>
<div class="section" id="xmovie-tool">
<span id="xmovie"></span><h2>9.31. xmovie tool<a class="headerlink" href="#xmovie-tool" title="Permalink to this headline"></a></h2>
<p>The xmovie tool is an X-based visualization package that can read
LAMMPS dump files and animate them. It is in its own sub-directory
with the tools directory. You may need to modify its Makefile so that
it can find the appropriate X libraries to link against.</p>
<p>The syntax for running xmovie is</p>
<div class="highlight-python"><div class="highlight"><pre>xmovie [options] dump.file1 dump.file2 ...
</pre></div>
</div>
<p>If you just type &#8220;xmovie&#8221; you will see a list of options. Note that
by default, LAMMPS dump files are in scaled coordinates, so you
typically need to use the -scale option with xmovie. When xmovie runs
it opens a visualization window and a control window. The control
options are straightforward to use.</p>
<p>Xmovie was mostly written by Mike Uttormark (U Wisconsin) while he
spent a summer at Sandia. It displays 2d projections of a 3d domain.
While simple in design, it is an amazingly fast program that can
render large numbers of atoms very quickly. It&#8217;s a useful tool for
debugging LAMMPS input and output and making sure your simulation is
doing what you think it should. The animations on the Examples page
of the <a class="reference external" href="http://lammps.sandia.gov">LAMMPS WWW site</a> were created with xmovie.</p>
<p>I&#8217;ve lost contact with Mike, so I hope he&#8217;s comfortable with us
distributing his great tool!</p>
</div>
</div>
</div>
</div>
<footer>
<div class="rst-footer-buttons" role="navigation" aria-label="footer navigation">
<a href="Section_modify.html" class="btn btn-neutral float-right" title="10. Modifying &amp; extending LAMMPS" accesskey="n">Next <span class="fa fa-arrow-circle-right"></span></a>
<a href="Section_perf.html" class="btn btn-neutral" title="8. Performance &amp; scalability" accesskey="p"><span class="fa fa-arrow-circle-left"></span> Previous</a>
</div>
<hr/>
<div role="contentinfo">
<p>
&copy; Copyright .
</p>
</div>
Built with <a href="http://sphinx-doc.org/">Sphinx</a> using a <a href="https://github.com/snide/sphinx_rtd_theme">theme</a> provided by <a href="https://readthedocs.org">Read the Docs</a>.
</footer>
</div>
</div>
</section>
</div>
<script type="text/javascript">
var DOCUMENTATION_OPTIONS = {
URL_ROOT:'./',
VERSION:'15 May 2015 version',
COLLAPSE_INDEX:false,
FILE_SUFFIX:'.html',
HAS_SOURCE: true
};
</script>
<script type="text/javascript" src="_static/jquery.js"></script>
<script type="text/javascript" src="_static/underscore.js"></script>
<script type="text/javascript" src="_static/doctools.js"></script>
<script type="text/javascript" src="https://cdn.mathjax.org/mathjax/latest/MathJax.js?config=TeX-AMS-MML_HTMLorMML"></script>
<script type="text/javascript" src="_static/sphinxcontrib-images/LightBox2/lightbox2/js/jquery-1.11.0.min.js"></script>
<script type="text/javascript" src="_static/sphinxcontrib-images/LightBox2/lightbox2/js/lightbox.min.js"></script>
<script type="text/javascript" src="_static/sphinxcontrib-images/LightBox2/lightbox2-customize/jquery-noconflict.js"></script>
<script type="text/javascript" src="_static/js/theme.js"></script>
<script type="text/javascript">
jQuery(function () {
SphinxRtdTheme.StickyNav.enable();
});
</script>
</body>
</html>

Binary file not shown.

Before

Width:  |  Height:  |  Size: 19 KiB

Binary file not shown.

Before

Width:  |  Height:  |  Size: 30 KiB

Binary file not shown.

Before

Width:  |  Height:  |  Size: 1.7 KiB

Binary file not shown.

Before

Width:  |  Height:  |  Size: 2.5 KiB

Binary file not shown.

Before

Width:  |  Height:  |  Size: 3.2 KiB

Binary file not shown.

Before

Width:  |  Height:  |  Size: 4.7 KiB

Binary file not shown.

Before

Width:  |  Height:  |  Size: 3.5 KiB

Binary file not shown.

Before

Width:  |  Height:  |  Size: 1.0 KiB

Binary file not shown.

Before

Width:  |  Height:  |  Size: 5.6 KiB

File diff suppressed because it is too large Load Diff

View File

@ -1,372 +0,0 @@
<!DOCTYPE html>
<!--[if IE 8]><html class="no-js lt-ie9" lang="en" > <![endif]-->
<!--[if gt IE 8]><!--> <html class="no-js" lang="en" > <!--<![endif]-->
<head>
<meta charset="utf-8">
<meta name="viewport" content="width=device-width, initial-scale=1.0">
<title>5.USER-CUDA package &mdash; LAMMPS 15 May 2015 version documentation</title>
<link rel="stylesheet" href="_static/css/theme.css" type="text/css" />
<link rel="stylesheet" href="_static/sphinxcontrib-images/LightBox2/lightbox2/css/lightbox.css" type="text/css" />
<link rel="top" title="LAMMPS 15 May 2015 version documentation" href="index.html"/>
<script src="_static/js/modernizr.min.js"></script>
</head>
<body class="wy-body-for-nav" role="document">
<div class="wy-grid-for-nav">
<nav data-toggle="wy-nav-shift" class="wy-nav-side">
<div class="wy-side-nav-search">
<a href="Manual.html" class="icon icon-home"> LAMMPS
</a>
<div role="search">
<form id="rtd-search-form" class="wy-form" action="search.html" method="get">
<input type="text" name="q" placeholder="Search docs" />
<input type="hidden" name="check_keywords" value="yes" />
<input type="hidden" name="area" value="default" />
</form>
</div>
</div>
<div class="wy-menu wy-menu-vertical" data-spy="affix" role="navigation" aria-label="main navigation">
<ul>
<li class="toctree-l1"><a class="reference internal" href="Section_intro.html">1. Introduction</a></li>
<li class="toctree-l1"><a class="reference internal" href="Section_start.html">2. Getting Started</a></li>
<li class="toctree-l1"><a class="reference internal" href="Section_commands.html">3. Commands</a></li>
<li class="toctree-l1"><a class="reference internal" href="Section_packages.html">4. Packages</a></li>
<li class="toctree-l1"><a class="reference internal" href="Section_accelerate.html">5. Accelerating LAMMPS performance</a></li>
<li class="toctree-l1"><a class="reference internal" href="Section_howto.html">6. How-to discussions</a></li>
<li class="toctree-l1"><a class="reference internal" href="Section_example.html">7. Example problems</a></li>
<li class="toctree-l1"><a class="reference internal" href="Section_perf.html">8. Performance &amp; scalability</a></li>
<li class="toctree-l1"><a class="reference internal" href="Section_tools.html">9. Additional tools</a></li>
<li class="toctree-l1"><a class="reference internal" href="Section_modify.html">10. Modifying &amp; extending LAMMPS</a></li>
<li class="toctree-l1"><a class="reference internal" href="Section_python.html">11. Python interface to LAMMPS</a></li>
<li class="toctree-l1"><a class="reference internal" href="Section_errors.html">12. Errors</a></li>
<li class="toctree-l1"><a class="reference internal" href="Section_history.html">13. Future and history</a></li>
</ul>
</div>
&nbsp;
</nav>
<section data-toggle="wy-nav-shift" class="wy-nav-content-wrap">
<nav class="wy-nav-top" role="navigation" aria-label="top navigation">
<i data-toggle="wy-nav-top" class="fa fa-bars"></i>
<a href="Manual.html">LAMMPS</a>
</nav>
<div class="wy-nav-content">
<div class="rst-content">
<div role="navigation" aria-label="breadcrumbs navigation">
<ul class="wy-breadcrumbs">
<li><a href="Manual.html">Docs</a> &raquo;</li>
<li>5.USER-CUDA package</li>
<li class="wy-breadcrumbs-aside">
<a href="http://lammps.sandia.gov">Website</a>
<a href="Section_commands.html#comm">Commands</a>
</li>
</ul>
<hr/>
</div>
<div role="main" class="document" itemscope="itemscope" itemtype="http://schema.org/Article">
<div itemprop="articleBody">
<p><a class="reference internal" href="Section_accelerate.html"><em>Return to Section accelerate overview</em></a></p>
<div class="section" id="user-cuda-package">
<h1>5.USER-CUDA package<a class="headerlink" href="#user-cuda-package" title="Permalink to this headline"></a></h1>
<p>The USER-CUDA package was developed by Christian Trott (Sandia) while
at U Technology Ilmenau in Germany. It provides NVIDIA GPU versions
of many pair styles, many fixes, a few computes, and for long-range
Coulombics via the PPPM command. It has the following general
features:</p>
<ul class="simple">
<li>The package is designed to allow an entire LAMMPS calculation, for
many timesteps, to run entirely on the GPU (except for inter-processor
MPI communication), so that atom-based data (e.g. coordinates, forces)
do not have to move back-and-forth between the CPU and GPU.</li>
<li>The speed-up advantage of this approach is typically better when the
number of atoms per GPU is large</li>
<li>Data will stay on the GPU until a timestep where a non-USER-CUDA fix
or compute is invoked. Whenever a non-GPU operation occurs (fix,
compute, output), data automatically moves back to the CPU as needed.
This may incur a performance penalty, but should otherwise work
transparently.</li>
<li>Neighbor lists are constructed on the GPU.</li>
<li>The package only supports use of a single MPI task, running on a
single CPU (core), assigned to each GPU.</li>
</ul>
<p>Here is a quick overview of how to use the USER-CUDA package:</p>
<ul class="simple">
<li>build the library in lib/cuda for your GPU hardware with desired precision</li>
<li>include the USER-CUDA package and build LAMMPS</li>
<li>use the mpirun command to specify 1 MPI task per GPU (on each node)</li>
<li>enable the USER-CUDA package via the &#8220;-c on&#8221; command-line switch</li>
<li>specify the # of GPUs per node</li>
<li>use USER-CUDA styles in your input script</li>
</ul>
<p>The latter two steps can be done using the &#8220;-pk cuda&#8221; and &#8220;-sf cuda&#8221;
<a class="reference internal" href="Section_start.html#start-7"><span>command-line switches</span></a> respectively. Or
the effect of the &#8220;-pk&#8221; or &#8220;-sf&#8221; switches can be duplicated by adding
the <a class="reference internal" href="package.html"><em>package cuda</em></a> or <a class="reference internal" href="suffix.html"><em>suffix cuda</em></a> commands
respectively to your input script.</p>
<p><strong>Required hardware/software:</strong></p>
<p>To use this package, you need to have one or more NVIDIA GPUs and
install the NVIDIA Cuda software on your system:</p>
<p>Your NVIDIA GPU needs to support Compute Capability 1.3. This list may
help you to find out the Compute Capability of your card:</p>
<p><a class="reference external" href="http://en.wikipedia.org/wiki/Comparison_of_Nvidia_graphics_processing_units">http://en.wikipedia.org/wiki/Comparison_of_Nvidia_graphics_processing_units</a></p>
<p>Install the Nvidia Cuda Toolkit (version 3.2 or higher) and the
corresponding GPU drivers. The Nvidia Cuda SDK is not required, but
we recommend it also be installed. You can then make sure its sample
projects can be compiled without problems.</p>
<p><strong>Building LAMMPS with the USER-CUDA package:</strong></p>
<p>This requires two steps (a,b): build the USER-CUDA library, then build
LAMMPS with the USER-CUDA package.</p>
<p>You can do both these steps in one line, using the src/Make.py script,
described in <a class="reference internal" href="Section_start.html#start-4"><span>Section 2.4</span></a> of the manual.
Type &#8220;Make.py -h&#8221; for help. If run from the src directory, this
command will create src/lmp_cuda using src/MAKE/Makefile.mpi as the
starting Makefile.machine:</p>
<div class="highlight-python"><div class="highlight"><pre>Make.py -p cuda -cuda mode=single arch=20 -o cuda -a lib-cuda file mpi
</pre></div>
</div>
<p>Or you can follow these two (a,b) steps:</p>
<ol class="loweralpha simple">
<li>Build the USER-CUDA library</li>
</ol>
<p>The USER-CUDA library is in lammps/lib/cuda. If your <em>CUDA</em> toolkit
is not installed in the default system directoy <em>/usr/local/cuda</em> edit
the file <em>lib/cuda/Makefile.common</em> accordingly.</p>
<p>To build the library with the settings in lib/cuda/Makefile.default,
simply type:</p>
<div class="highlight-python"><div class="highlight"><pre><span class="n">make</span>
</pre></div>
</div>
<p>To set options when the library is built, type &#8220;make OPTIONS&#8221;, where
<em>OPTIONS</em> are one or more of the following. The settings will be
written to the <em>lib/cuda/Makefile.defaults</em> before the build.</p>
<pre class="literal-block">
<em>precision=N</em> to set the precision level
N = 1 for single precision (default)
N = 2 for double precision
N = 3 for positions in double precision
N = 4 for positions and velocities in double precision
<em>arch=M</em> to set GPU compute capability
M = 35 for Kepler GPUs
M = 20 for CC2.0 (GF100/110, e.g. C2050,GTX580,GTX470) (default)
M = 21 for CC2.1 (GF104/114, e.g. GTX560, GTX460, GTX450)
M = 13 for CC1.3 (GF200, e.g. C1060, GTX285)
<em>prec_timer=0/1</em> to use hi-precision timers
0 = do not use them (default)
1 = use them
this is usually only useful for Mac machines
<em>dbg=0/1</em> to activate debug mode
0 = no debug mode (default)
1 = yes debug mode
this is only useful for developers
<em>cufft=1</em> for use of the CUDA FFT library
0 = no CUFFT support (default)
in the future other CUDA-enabled FFT libraries might be supported
</pre>
<p>If the build is successful, it will produce the files liblammpscuda.a and
Makefile.lammps.</p>
<p>Note that if you change any of the options (like precision), you need
to re-build the entire library. Do a &#8220;make clean&#8221; first, followed by
&#8220;make&#8221;.</p>
<ol class="loweralpha simple" start="2">
<li>Build LAMMPS with the USER-CUDA package</li>
</ol>
<div class="highlight-python"><div class="highlight"><pre>cd lammps/src
make yes-user-cuda
make machine
</pre></div>
</div>
<p>No additional compile/link flags are needed in Makefile.machine.</p>
<p>Note that if you change the USER-CUDA library precision (discussed
above) and rebuild the USER-CUDA library, then you also need to
re-install the USER-CUDA package and re-build LAMMPS, so that all
affected files are re-compiled and linked to the new USER-CUDA
library.</p>
<p><strong>Run with the USER-CUDA package from the command line:</strong></p>
<p>The mpirun or mpiexec command sets the total number of MPI tasks used
by LAMMPS (one or multiple per compute node) and the number of MPI
tasks used per node. E.g. the mpirun command in MPICH does this via
its -np and -ppn switches. Ditto for OpenMPI via -np and -npernode.</p>
<p>When using the USER-CUDA package, you must use exactly one MPI task
per physical GPU.</p>
<p>You must use the &#8220;-c on&#8221; <a class="reference internal" href="Section_start.html#start-7"><span>command-line switch</span></a> to enable the USER-CUDA package.
The &#8220;-c on&#8221; switch also issues a default <a class="reference internal" href="package.html"><em>package cuda 1</em></a>
command which sets various USER-CUDA options to default values, as
discussed on the <a class="reference internal" href="package.html"><em>package</em></a> command doc page.</p>
<p>Use the &#8220;-sf cuda&#8221; <a class="reference internal" href="Section_start.html#start-7"><span>command-line switch</span></a>,
which will automatically append &#8220;cuda&#8221; to styles that support it. Use
the &#8220;-pk cuda Ng&#8221; <a class="reference internal" href="Section_start.html#start-7"><span>command-line switch</span></a> to
set Ng = # of GPUs per node to a different value than the default set
by the &#8220;-c on&#8221; switch (1 GPU) or change other <a class="reference internal" href="package.html"><em>package cuda</em></a> options.</p>
<div class="highlight-python"><div class="highlight"><pre>lmp_machine -c on -sf cuda -pk cuda 1 -in in.script # 1 MPI task uses 1 GPU
mpirun -np 2 lmp_machine -c on -sf cuda -pk cuda 2 -in in.script # 2 MPI tasks use 2 GPUs on a single 16-core (or whatever) node
mpirun -np 24 -ppn 2 lmp_machine -c on -sf cuda -pk cuda 2 -in in.script # ditto on 12 16-core nodes
</pre></div>
</div>
<p>The syntax for the &#8220;-pk&#8221; switch is the same as same as the &#8220;package
cuda&#8221; command. See the <a class="reference internal" href="package.html"><em>package</em></a> command doc page for
details, including the default values used for all its options if it
is not specified.</p>
<p>Note that the default for the <a class="reference internal" href="package.html"><em>package cuda</em></a> command is
to set the Newton flag to &#8220;off&#8221; for both pairwise and bonded
interactions. This typically gives fastest performance. If the
<a class="reference internal" href="newton.html"><em>newton</em></a> command is used in the input script, it can
override these defaults.</p>
<p><strong>Or run with the USER-CUDA package by editing an input script:</strong></p>
<p>The discussion above for the mpirun/mpiexec command and the requirement
of one MPI task per GPU is the same.</p>
<p>You must still use the &#8220;-c on&#8221; <a class="reference internal" href="Section_start.html#start-7"><span>command-line switch</span></a> to enable the USER-CUDA package.</p>
<p>Use the <a class="reference internal" href="suffix.html"><em>suffix cuda</em></a> command, or you can explicitly add a
&#8220;cuda&#8221; suffix to individual styles in your input script, e.g.</p>
<div class="highlight-python"><div class="highlight"><pre>pair_style lj/cut/cuda 2.5
</pre></div>
</div>
<p>You only need to use the <a class="reference internal" href="package.html"><em>package cuda</em></a> command if you
wish to change any of its option defaults, including the number of
GPUs/node (default = 1), as set by the &#8220;-c on&#8221; <a class="reference internal" href="Section_start.html#start-7"><span>command-line switch</span></a>.</p>
<p><strong>Speed-ups to expect:</strong></p>
<p>The performance of a GPU versus a multi-core CPU is a function of your
hardware, which pair style is used, the number of atoms/GPU, and the
precision used on the GPU (double, single, mixed).</p>
<p>See the <a class="reference external" href="http://lammps.sandia.gov/bench.html">Benchmark page</a> of the
LAMMPS web site for performance of the USER-CUDA package on different
hardware.</p>
<p><strong>Guidelines for best performance:</strong></p>
<ul class="simple">
<li>The USER-CUDA package offers more speed-up relative to CPU performance
when the number of atoms per GPU is large, e.g. on the order of tens
or hundreds of 1000s.</li>
<li>As noted above, this package will continue to run a simulation
entirely on the GPU(s) (except for inter-processor MPI communication),
for multiple timesteps, until a CPU calculation is required, either by
a fix or compute that is non-GPU-ized, or until output is performed
(thermo or dump snapshot or restart file). The less often this
occurs, the faster your simulation will run.</li>
</ul>
<div class="section" id="restrictions">
<h2>Restrictions<a class="headerlink" href="#restrictions" title="Permalink to this headline"></a></h2>
<p>None.</p>
</div>
</div>
</div>
</div>
<footer>
<hr/>
<div role="contentinfo">
<p>
&copy; Copyright .
</p>
</div>
Built with <a href="http://sphinx-doc.org/">Sphinx</a> using a <a href="https://github.com/snide/sphinx_rtd_theme">theme</a> provided by <a href="https://readthedocs.org">Read the Docs</a>.
</footer>
</div>
</div>
</section>
</div>
<script type="text/javascript">
var DOCUMENTATION_OPTIONS = {
URL_ROOT:'./',
VERSION:'15 May 2015 version',
COLLAPSE_INDEX:false,
FILE_SUFFIX:'.html',
HAS_SOURCE: true
};
</script>
<script type="text/javascript" src="_static/jquery.js"></script>
<script type="text/javascript" src="_static/underscore.js"></script>
<script type="text/javascript" src="_static/doctools.js"></script>
<script type="text/javascript" src="https://cdn.mathjax.org/mathjax/latest/MathJax.js?config=TeX-AMS-MML_HTMLorMML"></script>
<script type="text/javascript" src="_static/sphinxcontrib-images/LightBox2/lightbox2/js/jquery-1.11.0.min.js"></script>
<script type="text/javascript" src="_static/sphinxcontrib-images/LightBox2/lightbox2/js/lightbox.min.js"></script>
<script type="text/javascript" src="_static/sphinxcontrib-images/LightBox2/lightbox2-customize/jquery-noconflict.js"></script>
<script type="text/javascript" src="_static/js/theme.js"></script>
<script type="text/javascript">
jQuery(function () {
SphinxRtdTheme.StickyNav.enable();
});
</script>
</body>
</html>

View File

@ -1,223 +0,0 @@
"Previous Section"_Section_packages.html - "LAMMPS WWW Site"_lws -
"LAMMPS Documentation"_ld - "LAMMPS Commands"_lc :c
:link(lws,http://lammps.sandia.gov)
:link(ld,Manual.html)
:link(lc,Section_commands.html#comm)
:line
"Return to Section accelerate overview"_Section_accelerate.html
5.3.1 USER-CUDA package :h4
The USER-CUDA package was developed by Christian Trott (Sandia) while
at U Technology Ilmenau in Germany. It provides NVIDIA GPU versions
of many pair styles, many fixes, a few computes, and for long-range
Coulombics via the PPPM command. It has the following general
features:
The package is designed to allow an entire LAMMPS calculation, for
many timesteps, to run entirely on the GPU (except for inter-processor
MPI communication), so that atom-based data (e.g. coordinates, forces)
do not have to move back-and-forth between the CPU and GPU. :ulb,l
The speed-up advantage of this approach is typically better when the
number of atoms per GPU is large :l
Data will stay on the GPU until a timestep where a non-USER-CUDA fix
or compute is invoked. Whenever a non-GPU operation occurs (fix,
compute, output), data automatically moves back to the CPU as needed.
This may incur a performance penalty, but should otherwise work
transparently. :l
Neighbor lists are constructed on the GPU. :l
The package only supports use of a single MPI task, running on a
single CPU (core), assigned to each GPU. :l,ule
Here is a quick overview of how to use the USER-CUDA package:
build the library in lib/cuda for your GPU hardware with desired precision
include the USER-CUDA package and build LAMMPS
use the mpirun command to specify 1 MPI task per GPU (on each node)
enable the USER-CUDA package via the "-c on" command-line switch
specify the # of GPUs per node
use USER-CUDA styles in your input script :ul
The latter two steps can be done using the "-pk cuda" and "-sf cuda"
"command-line switches"_Section_start.html#start_7 respectively. Or
the effect of the "-pk" or "-sf" switches can be duplicated by adding
the "package cuda"_package.html or "suffix cuda"_suffix.html commands
respectively to your input script.
[Required hardware/software:]
To use this package, you need to have one or more NVIDIA GPUs and
install the NVIDIA Cuda software on your system:
Your NVIDIA GPU needs to support Compute Capability 1.3. This list may
help you to find out the Compute Capability of your card:
http://en.wikipedia.org/wiki/Comparison_of_Nvidia_graphics_processing_units
Install the Nvidia Cuda Toolkit (version 3.2 or higher) and the
corresponding GPU drivers. The Nvidia Cuda SDK is not required, but
we recommend it also be installed. You can then make sure its sample
projects can be compiled without problems.
[Building LAMMPS with the USER-CUDA package:]
This requires two steps (a,b): build the USER-CUDA library, then build
LAMMPS with the USER-CUDA package.
You can do both these steps in one line, using the src/Make.py script,
described in "Section 2.4"_Section_start.html#start_4 of the manual.
Type "Make.py -h" for help. If run from the src directory, this
command will create src/lmp_cuda using src/MAKE/Makefile.mpi as the
starting Makefile.machine:
Make.py -p cuda -cuda mode=single arch=20 -o cuda -a lib-cuda file mpi :pre
Or you can follow these two (a,b) steps:
(a) Build the USER-CUDA library
The USER-CUDA library is in lammps/lib/cuda. If your {CUDA} toolkit
is not installed in the default system directoy {/usr/local/cuda} edit
the file {lib/cuda/Makefile.common} accordingly.
To build the library with the settings in lib/cuda/Makefile.default,
simply type:
make :pre
To set options when the library is built, type "make OPTIONS", where
{OPTIONS} are one or more of the following. The settings will be
written to the {lib/cuda/Makefile.defaults} before the build.
{precision=N} to set the precision level
N = 1 for single precision (default)
N = 2 for double precision
N = 3 for positions in double precision
N = 4 for positions and velocities in double precision
{arch=M} to set GPU compute capability
M = 35 for Kepler GPUs
M = 20 for CC2.0 (GF100/110, e.g. C2050,GTX580,GTX470) (default)
M = 21 for CC2.1 (GF104/114, e.g. GTX560, GTX460, GTX450)
M = 13 for CC1.3 (GF200, e.g. C1060, GTX285)
{prec_timer=0/1} to use hi-precision timers
0 = do not use them (default)
1 = use them
this is usually only useful for Mac machines
{dbg=0/1} to activate debug mode
0 = no debug mode (default)
1 = yes debug mode
this is only useful for developers
{cufft=1} for use of the CUDA FFT library
0 = no CUFFT support (default)
in the future other CUDA-enabled FFT libraries might be supported :pre
If the build is successful, it will produce the files liblammpscuda.a and
Makefile.lammps.
Note that if you change any of the options (like precision), you need
to re-build the entire library. Do a "make clean" first, followed by
"make".
(b) Build LAMMPS with the USER-CUDA package
cd lammps/src
make yes-user-cuda
make machine :pre
No additional compile/link flags are needed in Makefile.machine.
Note that if you change the USER-CUDA library precision (discussed
above) and rebuild the USER-CUDA library, then you also need to
re-install the USER-CUDA package and re-build LAMMPS, so that all
affected files are re-compiled and linked to the new USER-CUDA
library.
[Run with the USER-CUDA package from the command line:]
The mpirun or mpiexec command sets the total number of MPI tasks used
by LAMMPS (one or multiple per compute node) and the number of MPI
tasks used per node. E.g. the mpirun command in MPICH does this via
its -np and -ppn switches. Ditto for OpenMPI via -np and -npernode.
When using the USER-CUDA package, you must use exactly one MPI task
per physical GPU.
You must use the "-c on" "command-line
switch"_Section_start.html#start_7 to enable the USER-CUDA package.
The "-c on" switch also issues a default "package cuda 1"_package.html
command which sets various USER-CUDA options to default values, as
discussed on the "package"_package.html command doc page.
Use the "-sf cuda" "command-line switch"_Section_start.html#start_7,
which will automatically append "cuda" to styles that support it. Use
the "-pk cuda Ng" "command-line switch"_Section_start.html#start_7 to
set Ng = # of GPUs per node to a different value than the default set
by the "-c on" switch (1 GPU) or change other "package
cuda"_package.html options.
lmp_machine -c on -sf cuda -pk cuda 1 -in in.script # 1 MPI task uses 1 GPU
mpirun -np 2 lmp_machine -c on -sf cuda -pk cuda 2 -in in.script # 2 MPI tasks use 2 GPUs on a single 16-core (or whatever) node
mpirun -np 24 -ppn 2 lmp_machine -c on -sf cuda -pk cuda 2 -in in.script # ditto on 12 16-core nodes :pre
The syntax for the "-pk" switch is the same as same as the "package
cuda" command. See the "package"_package.html command doc page for
details, including the default values used for all its options if it
is not specified.
Note that the default for the "package cuda"_package.html command is
to set the Newton flag to "off" for both pairwise and bonded
interactions. This typically gives fastest performance. If the
"newton"_newton.html command is used in the input script, it can
override these defaults.
[Or run with the USER-CUDA package by editing an input script:]
The discussion above for the mpirun/mpiexec command and the requirement
of one MPI task per GPU is the same.
You must still use the "-c on" "command-line
switch"_Section_start.html#start_7 to enable the USER-CUDA package.
Use the "suffix cuda"_suffix.html command, or you can explicitly add a
"cuda" suffix to individual styles in your input script, e.g.
pair_style lj/cut/cuda 2.5 :pre
You only need to use the "package cuda"_package.html command if you
wish to change any of its option defaults, including the number of
GPUs/node (default = 1), as set by the "-c on" "command-line
switch"_Section_start.html#start_7.
[Speed-ups to expect:]
The performance of a GPU versus a multi-core CPU is a function of your
hardware, which pair style is used, the number of atoms/GPU, and the
precision used on the GPU (double, single, mixed).
See the "Benchmark page"_http://lammps.sandia.gov/bench.html of the
LAMMPS web site for performance of the USER-CUDA package on different
hardware.
[Guidelines for best performance:]
The USER-CUDA package offers more speed-up relative to CPU performance
when the number of atoms per GPU is large, e.g. on the order of tens
or hundreds of 1000s. :ulb,l
As noted above, this package will continue to run a simulation
entirely on the GPU(s) (except for inter-processor MPI communication),
for multiple timesteps, until a CPU calculation is required, either by
a fix or compute that is non-GPU-ized, or until output is performed
(thermo or dump snapshot or restart file). The less often this
occurs, the faster your simulation will run. :l,ule
[Restrictions:]
None.

View File

@ -1,401 +0,0 @@
<!DOCTYPE html>
<!--[if IE 8]><html class="no-js lt-ie9" lang="en" > <![endif]-->
<!--[if gt IE 8]><!--> <html class="no-js" lang="en" > <!--<![endif]-->
<head>
<meta charset="utf-8">
<meta name="viewport" content="width=device-width, initial-scale=1.0">
<title>5.GPU package &mdash; LAMMPS 15 May 2015 version documentation</title>
<link rel="stylesheet" href="_static/css/theme.css" type="text/css" />
<link rel="stylesheet" href="_static/sphinxcontrib-images/LightBox2/lightbox2/css/lightbox.css" type="text/css" />
<link rel="top" title="LAMMPS 15 May 2015 version documentation" href="index.html"/>
<script src="_static/js/modernizr.min.js"></script>
</head>
<body class="wy-body-for-nav" role="document">
<div class="wy-grid-for-nav">
<nav data-toggle="wy-nav-shift" class="wy-nav-side">
<div class="wy-side-nav-search">
<a href="Manual.html" class="icon icon-home"> LAMMPS
</a>
<div role="search">
<form id="rtd-search-form" class="wy-form" action="search.html" method="get">
<input type="text" name="q" placeholder="Search docs" />
<input type="hidden" name="check_keywords" value="yes" />
<input type="hidden" name="area" value="default" />
</form>
</div>
</div>
<div class="wy-menu wy-menu-vertical" data-spy="affix" role="navigation" aria-label="main navigation">
<ul>
<li class="toctree-l1"><a class="reference internal" href="Section_intro.html">1. Introduction</a></li>
<li class="toctree-l1"><a class="reference internal" href="Section_start.html">2. Getting Started</a></li>
<li class="toctree-l1"><a class="reference internal" href="Section_commands.html">3. Commands</a></li>
<li class="toctree-l1"><a class="reference internal" href="Section_packages.html">4. Packages</a></li>
<li class="toctree-l1"><a class="reference internal" href="Section_accelerate.html">5. Accelerating LAMMPS performance</a></li>
<li class="toctree-l1"><a class="reference internal" href="Section_howto.html">6. How-to discussions</a></li>
<li class="toctree-l1"><a class="reference internal" href="Section_example.html">7. Example problems</a></li>
<li class="toctree-l1"><a class="reference internal" href="Section_perf.html">8. Performance &amp; scalability</a></li>
<li class="toctree-l1"><a class="reference internal" href="Section_tools.html">9. Additional tools</a></li>
<li class="toctree-l1"><a class="reference internal" href="Section_modify.html">10. Modifying &amp; extending LAMMPS</a></li>
<li class="toctree-l1"><a class="reference internal" href="Section_python.html">11. Python interface to LAMMPS</a></li>
<li class="toctree-l1"><a class="reference internal" href="Section_errors.html">12. Errors</a></li>
<li class="toctree-l1"><a class="reference internal" href="Section_history.html">13. Future and history</a></li>
</ul>
</div>
&nbsp;
</nav>
<section data-toggle="wy-nav-shift" class="wy-nav-content-wrap">
<nav class="wy-nav-top" role="navigation" aria-label="top navigation">
<i data-toggle="wy-nav-top" class="fa fa-bars"></i>
<a href="Manual.html">LAMMPS</a>
</nav>
<div class="wy-nav-content">
<div class="rst-content">
<div role="navigation" aria-label="breadcrumbs navigation">
<ul class="wy-breadcrumbs">
<li><a href="Manual.html">Docs</a> &raquo;</li>
<li>5.GPU package</li>
<li class="wy-breadcrumbs-aside">
<a href="http://lammps.sandia.gov">Website</a>
<a href="Section_commands.html#comm">Commands</a>
</li>
</ul>
<hr/>
</div>
<div role="main" class="document" itemscope="itemscope" itemtype="http://schema.org/Article">
<div itemprop="articleBody">
<p><a class="reference internal" href="Section_accelerate.html"><em>Return to Section accelerate overview</em></a></p>
<div class="section" id="gpu-package">
<h1>5.GPU package<a class="headerlink" href="#gpu-package" title="Permalink to this headline"></a></h1>
<p>The GPU package was developed by Mike Brown at ORNL and his
collaborators, particularly Trung Nguyen (ORNL). It provides GPU
versions of many pair styles, including the 3-body Stillinger-Weber
pair style, and for <a class="reference internal" href="kspace_style.html"><em>kspace_style pppm</em></a> for
long-range Coulombics. It has the following general features:</p>
<ul class="simple">
<li>It is designed to exploit common GPU hardware configurations where one
or more GPUs are coupled to many cores of one or more multi-core CPUs,
e.g. within a node of a parallel machine.</li>
<li>Atom-based data (e.g. coordinates, forces) moves back-and-forth
between the CPU(s) and GPU every timestep.</li>
<li>Neighbor lists can be built on the CPU or on the GPU</li>
<li>The charge assignement and force interpolation portions of PPPM can be
run on the GPU. The FFT portion, which requires MPI communication
between processors, runs on the CPU.</li>
<li>Asynchronous force computations can be performed simultaneously on the
CPU(s) and GPU.</li>
<li>It allows for GPU computations to be performed in single or double
precision, or in mixed-mode precision, where pairwise forces are
computed in single precision, but accumulated into double-precision
force vectors.</li>
<li>LAMMPS-specific code is in the GPU package. It makes calls to a
generic GPU library in the lib/gpu directory. This library provides
NVIDIA support as well as more general OpenCL support, so that the
same functionality can eventually be supported on a variety of GPU
hardware.</li>
</ul>
<p>Here is a quick overview of how to use the GPU package:</p>
<ul class="simple">
<li>build the library in lib/gpu for your GPU hardware wity desired precision</li>
<li>include the GPU package and build LAMMPS</li>
<li>use the mpirun command to set the number of MPI tasks/node which determines the number of MPI tasks/GPU</li>
<li>specify the # of GPUs per node</li>
<li>use GPU styles in your input script</li>
</ul>
<p>The latter two steps can be done using the &#8220;-pk gpu&#8221; and &#8220;-sf gpu&#8221;
<a class="reference internal" href="Section_start.html#start-7"><span>command-line switches</span></a> respectively. Or
the effect of the &#8220;-pk&#8221; or &#8220;-sf&#8221; switches can be duplicated by adding
the <a class="reference internal" href="package.html"><em>package gpu</em></a> or <a class="reference internal" href="suffix.html"><em>suffix gpu</em></a> commands
respectively to your input script.</p>
<p><strong>Required hardware/software:</strong></p>
<p>To use this package, you currently need to have an NVIDIA GPU and
install the NVIDIA Cuda software on your system:</p>
<ul class="simple">
<li>Check if you have an NVIDIA GPU: cat /proc/driver/nvidia/gpus/0/information</li>
<li>Go to <a class="reference external" href="http://www.nvidia.com/object/cuda_get.html">http://www.nvidia.com/object/cuda_get.html</a></li>
<li>Install a driver and toolkit appropriate for your system (SDK is not necessary)</li>
<li>Run lammps/lib/gpu/nvc_get_devices (after building the GPU library, see below) to list supported devices and properties</li>
</ul>
<p><strong>Building LAMMPS with the GPU package:</strong></p>
<p>This requires two steps (a,b): build the GPU library, then build
LAMMPS with the GPU package.</p>
<p>You can do both these steps in one line, using the src/Make.py script,
described in <a class="reference internal" href="Section_start.html#start-4"><span>Section 2.4</span></a> of the manual.
Type &#8220;Make.py -h&#8221; for help. If run from the src directory, this
command will create src/lmp_gpu using src/MAKE/Makefile.mpi as the
starting Makefile.machine:</p>
<div class="highlight-python"><div class="highlight"><pre>Make.py -p gpu -gpu mode=single arch=31 -o gpu -a lib-gpu file mpi
</pre></div>
</div>
<p>Or you can follow these two (a,b) steps:</p>
<ol class="loweralpha simple">
<li>Build the GPU library</li>
</ol>
<p>The GPU library is in lammps/lib/gpu. Select a Makefile.machine (in
lib/gpu) appropriate for your system. You should pay special
attention to 3 settings in this makefile.</p>
<ul class="simple">
<li>CUDA_HOME = needs to be where NVIDIA Cuda software is installed on your system</li>
<li>CUDA_ARCH = needs to be appropriate to your GPUs</li>
<li>CUDA_PREC = precision (double, mixed, single) you desire</li>
</ul>
<p>See lib/gpu/Makefile.linux.double for examples of the ARCH settings
for different GPU choices, e.g. Fermi vs Kepler. It also lists the
possible precision settings:</p>
<div class="highlight-python"><div class="highlight"><pre><span class="n">CUDA_PREC</span> <span class="o">=</span> <span class="o">-</span><span class="n">D_SINGLE_SINGLE</span> <span class="c"># single precision for all calculations</span>
<span class="n">CUDA_PREC</span> <span class="o">=</span> <span class="o">-</span><span class="n">D_DOUBLE_DOUBLE</span> <span class="c"># double precision for all calculations</span>
<span class="n">CUDA_PREC</span> <span class="o">=</span> <span class="o">-</span><span class="n">D_SINGLE_DOUBLE</span> <span class="c"># accumulation of forces, etc, in double</span>
</pre></div>
</div>
<p>The last setting is the mixed mode referred to above. Note that your
GPU must support double precision to use either the 2nd or 3rd of
these settings.</p>
<p>To build the library, type:</p>
<div class="highlight-python"><div class="highlight"><pre>make -f Makefile.machine
</pre></div>
</div>
<p>If successful, it will produce the files libgpu.a and Makefile.lammps.</p>
<p>The latter file has 3 settings that need to be appropriate for the
paths and settings for the CUDA system software on your machine.
Makefile.lammps is a copy of the file specified by the EXTRAMAKE
setting in Makefile.machine. You can change EXTRAMAKE or create your
own Makefile.lammps.machine if needed.</p>
<p>Note that to change the precision of the GPU library, you need to
re-build the entire library. Do a &#8220;clean&#8221; first, e.g. &#8220;make -f
Makefile.linux clean&#8221;, followed by the make command above.</p>
<ol class="loweralpha simple" start="2">
<li>Build LAMMPS with the GPU package</li>
</ol>
<div class="highlight-python"><div class="highlight"><pre>cd lammps/src
make yes-gpu
make machine
</pre></div>
</div>
<p>No additional compile/link flags are needed in Makefile.machine.</p>
<p>Note that if you change the GPU library precision (discussed above)
and rebuild the GPU library, then you also need to re-install the GPU
package and re-build LAMMPS, so that all affected files are
re-compiled and linked to the new GPU library.</p>
<p><strong>Run with the GPU package from the command line:</strong></p>
<p>The mpirun or mpiexec command sets the total number of MPI tasks used
by LAMMPS (one or multiple per compute node) and the number of MPI
tasks used per node. E.g. the mpirun command in MPICH does this via
its -np and -ppn switches. Ditto for OpenMPI via -np and -npernode.</p>
<p>When using the GPU package, you cannot assign more than one GPU to a
single MPI task. However multiple MPI tasks can share the same GPU,
and in many cases it will be more efficient to run this way. Likewise
it may be more efficient to use less MPI tasks/node than the available
# of CPU cores. Assignment of multiple MPI tasks to a GPU will happen
automatically if you create more MPI tasks/node than there are
GPUs/mode. E.g. with 8 MPI tasks/node and 2 GPUs, each GPU will be
shared by 4 MPI tasks.</p>
<p>Use the &#8220;-sf gpu&#8221; <a class="reference internal" href="Section_start.html#start-7"><span>command-line switch</span></a>,
which will automatically append &#8220;gpu&#8221; to styles that support it. Use
the &#8220;-pk gpu Ng&#8221; <a class="reference internal" href="Section_start.html#start-7"><span>command-line switch</span></a> to
set Ng = # of GPUs/node to use.</p>
<div class="highlight-python"><div class="highlight"><pre>lmp_machine -sf gpu -pk gpu 1 -in in.script # 1 MPI task uses 1 GPU
mpirun -np 12 lmp_machine -sf gpu -pk gpu 2 -in in.script # 12 MPI tasks share 2 GPUs on a single 16-core (or whatever) node
mpirun -np 48 -ppn 12 lmp_machine -sf gpu -pk gpu 2 -in in.script # ditto on 4 16-core nodes
</pre></div>
</div>
<p>Note that if the &#8220;-sf gpu&#8221; switch is used, it also issues a default
<a class="reference internal" href="package.html"><em>package gpu 1</em></a> command, which sets the number of
GPUs/node to 1.</p>
<p>Using the &#8220;-pk&#8221; switch explicitly allows for setting of the number of
GPUs/node to use and additional options. Its syntax is the same as
same as the &#8220;package gpu&#8221; command. See the <a class="reference internal" href="package.html"><em>package</em></a>
command doc page for details, including the default values used for
all its options if it is not specified.</p>
<p>Note that the default for the <a class="reference internal" href="package.html"><em>package gpu</em></a> command is to
set the Newton flag to &#8220;off&#8221; pairwise interactions. It does not
affect the setting for bonded interactions (LAMMPS default is &#8220;on&#8221;).
The &#8220;off&#8221; setting for pairwise interaction is currently required for
GPU package pair styles.</p>
<p><strong>Or run with the GPU package by editing an input script:</strong></p>
<p>The discussion above for the mpirun/mpiexec command, MPI tasks/node,
and use of multiple MPI tasks/GPU is the same.</p>
<p>Use the <a class="reference internal" href="suffix.html"><em>suffix gpu</em></a> command, or you can explicitly add an
&#8220;gpu&#8221; suffix to individual styles in your input script, e.g.</p>
<div class="highlight-python"><div class="highlight"><pre>pair_style lj/cut/gpu 2.5
</pre></div>
</div>
<p>You must also use the <a class="reference internal" href="package.html"><em>package gpu</em></a> command to enable the
GPU package, unless the &#8220;-sf gpu&#8221; or &#8220;-pk gpu&#8221; <a class="reference internal" href="Section_start.html#start-7"><span>command-line switches</span></a> were used. It specifies the
number of GPUs/node to use, as well as other options.</p>
<p><strong>Speed-ups to expect:</strong></p>
<p>The performance of a GPU versus a multi-core CPU is a function of your
hardware, which pair style is used, the number of atoms/GPU, and the
precision used on the GPU (double, single, mixed).</p>
<p>See the <a class="reference external" href="http://lammps.sandia.gov/bench.html">Benchmark page</a> of the
LAMMPS web site for performance of the GPU package on various
hardware, including the Titan HPC platform at ORNL.</p>
<p>You should also experiment with how many MPI tasks per GPU to use to
give the best performance for your problem and machine. This is also
a function of the problem size and the pair style being using.
Likewise, you should experiment with the precision setting for the GPU
library to see if single or mixed precision will give accurate
results, since they will typically be faster.</p>
<p><strong>Guidelines for best performance:</strong></p>
<ul class="simple">
<li>Using multiple MPI tasks per GPU will often give the best performance,
as allowed my most multi-core CPU/GPU configurations.</li>
<li>If the number of particles per MPI task is small (e.g. 100s of
particles), it can be more efficient to run with fewer MPI tasks per
GPU, even if you do not use all the cores on the compute node.</li>
<li>The <a class="reference internal" href="package.html"><em>package gpu</em></a> command has several options for tuning
performance. Neighbor lists can be built on the GPU or CPU. Force
calculations can be dynamically balanced across the CPU cores and
GPUs. GPU-specific settings can be made which can be optimized
for different hardware. See the <a class="reference internal" href="package.html"><em>packakge</em></a> command
doc page for details.</li>
<li>As described by the <a class="reference internal" href="package.html"><em>package gpu</em></a> command, GPU
accelerated pair styles can perform computations asynchronously with
CPU computations. The &#8220;Pair&#8221; time reported by LAMMPS will be the
maximum of the time required to complete the CPU pair style
computations and the time required to complete the GPU pair style
computations. Any time spent for GPU-enabled pair styles for
computations that run simultaneously with <a class="reference internal" href="bond_style.html"><em>bond</em></a>,
<a class="reference internal" href="angle_style.html"><em>angle</em></a>, <a class="reference internal" href="dihedral_style.html"><em>dihedral</em></a>,
<a class="reference internal" href="improper_style.html"><em>improper</em></a>, and <a class="reference internal" href="kspace_style.html"><em>long-range</em></a>
calculations will not be included in the &#8220;Pair&#8221; time.</li>
<li>When the <em>mode</em> setting for the package gpu command is force/neigh,
the time for neighbor list calculations on the GPU will be added into
the &#8220;Pair&#8221; time, not the &#8220;Neigh&#8221; time. An additional breakdown of the
times required for various tasks on the GPU (data copy, neighbor
calculations, force computations, etc) are output only with the LAMMPS
screen output (not in the log file) at the end of each run. These
timings represent total time spent on the GPU for each routine,
regardless of asynchronous CPU calculations.</li>
<li>The output section &#8220;GPU Time Info (average)&#8221; reports &#8220;Max Mem / Proc&#8221;.
This is the maximum memory used at one time on the GPU for data
storage by a single MPI process.</li>
</ul>
<div class="section" id="restrictions">
<h2>Restrictions<a class="headerlink" href="#restrictions" title="Permalink to this headline"></a></h2>
<p>None.</p>
</div>
</div>
</div>
</div>
<footer>
<hr/>
<div role="contentinfo">
<p>
&copy; Copyright .
</p>
</div>
Built with <a href="http://sphinx-doc.org/">Sphinx</a> using a <a href="https://github.com/snide/sphinx_rtd_theme">theme</a> provided by <a href="https://readthedocs.org">Read the Docs</a>.
</footer>
</div>
</div>
</section>
</div>
<script type="text/javascript">
var DOCUMENTATION_OPTIONS = {
URL_ROOT:'./',
VERSION:'15 May 2015 version',
COLLAPSE_INDEX:false,
FILE_SUFFIX:'.html',
HAS_SOURCE: true
};
</script>
<script type="text/javascript" src="_static/jquery.js"></script>
<script type="text/javascript" src="_static/underscore.js"></script>
<script type="text/javascript" src="_static/doctools.js"></script>
<script type="text/javascript" src="https://cdn.mathjax.org/mathjax/latest/MathJax.js?config=TeX-AMS-MML_HTMLorMML"></script>
<script type="text/javascript" src="_static/sphinxcontrib-images/LightBox2/lightbox2/js/jquery-1.11.0.min.js"></script>
<script type="text/javascript" src="_static/sphinxcontrib-images/LightBox2/lightbox2/js/lightbox.min.js"></script>
<script type="text/javascript" src="_static/sphinxcontrib-images/LightBox2/lightbox2-customize/jquery-noconflict.js"></script>
<script type="text/javascript" src="_static/js/theme.js"></script>
<script type="text/javascript">
jQuery(function () {
SphinxRtdTheme.StickyNav.enable();
});
</script>
</body>
</html>

View File

@ -1,461 +0,0 @@
<!DOCTYPE html>
<!--[if IE 8]><html class="no-js lt-ie9" lang="en" > <![endif]-->
<!--[if gt IE 8]><!--> <html class="no-js" lang="en" > <!--<![endif]-->
<head>
<meta charset="utf-8">
<meta name="viewport" content="width=device-width, initial-scale=1.0">
<title>5.USER-INTEL package &mdash; LAMMPS 15 May 2015 version documentation</title>
<link rel="stylesheet" href="_static/css/theme.css" type="text/css" />
<link rel="stylesheet" href="_static/sphinxcontrib-images/LightBox2/lightbox2/css/lightbox.css" type="text/css" />
<link rel="top" title="LAMMPS 15 May 2015 version documentation" href="index.html"/>
<script src="_static/js/modernizr.min.js"></script>
</head>
<body class="wy-body-for-nav" role="document">
<div class="wy-grid-for-nav">
<nav data-toggle="wy-nav-shift" class="wy-nav-side">
<div class="wy-side-nav-search">
<a href="Manual.html" class="icon icon-home"> LAMMPS
</a>
<div role="search">
<form id="rtd-search-form" class="wy-form" action="search.html" method="get">
<input type="text" name="q" placeholder="Search docs" />
<input type="hidden" name="check_keywords" value="yes" />
<input type="hidden" name="area" value="default" />
</form>
</div>
</div>
<div class="wy-menu wy-menu-vertical" data-spy="affix" role="navigation" aria-label="main navigation">
<ul>
<li class="toctree-l1"><a class="reference internal" href="Section_intro.html">1. Introduction</a></li>
<li class="toctree-l1"><a class="reference internal" href="Section_start.html">2. Getting Started</a></li>
<li class="toctree-l1"><a class="reference internal" href="Section_commands.html">3. Commands</a></li>
<li class="toctree-l1"><a class="reference internal" href="Section_packages.html">4. Packages</a></li>
<li class="toctree-l1"><a class="reference internal" href="Section_accelerate.html">5. Accelerating LAMMPS performance</a></li>
<li class="toctree-l1"><a class="reference internal" href="Section_howto.html">6. How-to discussions</a></li>
<li class="toctree-l1"><a class="reference internal" href="Section_example.html">7. Example problems</a></li>
<li class="toctree-l1"><a class="reference internal" href="Section_perf.html">8. Performance &amp; scalability</a></li>
<li class="toctree-l1"><a class="reference internal" href="Section_tools.html">9. Additional tools</a></li>
<li class="toctree-l1"><a class="reference internal" href="Section_modify.html">10. Modifying &amp; extending LAMMPS</a></li>
<li class="toctree-l1"><a class="reference internal" href="Section_python.html">11. Python interface to LAMMPS</a></li>
<li class="toctree-l1"><a class="reference internal" href="Section_errors.html">12. Errors</a></li>
<li class="toctree-l1"><a class="reference internal" href="Section_history.html">13. Future and history</a></li>
</ul>
</div>
&nbsp;
</nav>
<section data-toggle="wy-nav-shift" class="wy-nav-content-wrap">
<nav class="wy-nav-top" role="navigation" aria-label="top navigation">
<i data-toggle="wy-nav-top" class="fa fa-bars"></i>
<a href="Manual.html">LAMMPS</a>
</nav>
<div class="wy-nav-content">
<div class="rst-content">
<div role="navigation" aria-label="breadcrumbs navigation">
<ul class="wy-breadcrumbs">
<li><a href="Manual.html">Docs</a> &raquo;</li>
<li>5.USER-INTEL package</li>
<li class="wy-breadcrumbs-aside">
<a href="http://lammps.sandia.gov">Website</a>
<a href="Section_commands.html#comm">Commands</a>
</li>
</ul>
<hr/>
</div>
<div role="main" class="document" itemscope="itemscope" itemtype="http://schema.org/Article">
<div itemprop="articleBody">
<p><a class="reference internal" href="Section_accelerate.html"><em>Return to Section accelerate overview</em></a></p>
<div class="section" id="user-intel-package">
<h1>5.USER-INTEL package<a class="headerlink" href="#user-intel-package" title="Permalink to this headline"></a></h1>
<p>The USER-INTEL package was developed by Mike Brown at Intel
Corporation. It provides two methods for accelerating simulations,
depending on the hardware you have. The first is acceleration on
Intel(R) CPUs by running in single, mixed, or double precision with
vectorization. The second is acceleration on Intel(R) Xeon Phi(TM)
coprocessors via offloading neighbor list and non-bonded force
calculations to the Phi. The same C++ code is used in both cases.
When offloading to a coprocessor from a CPU, the same routine is run
twice, once on the CPU and once with an offload flag.</p>
<p>Note that the USER-INTEL package supports use of the Phi in &#8220;offload&#8221;
mode, not &#8220;native&#8221; mode like the <a class="reference internal" href="accelerate_kokkos.html"><em>KOKKOS package</em></a>.</p>
<p>Also note that the USER-INTEL package can be used in tandem with the
<a class="reference internal" href="accelerate_omp.html"><em>USER-OMP package</em></a>. This is useful when
offloading pair style computations to the Phi, so that other styles
not supported by the USER-INTEL package, e.g. bond, angle, dihedral,
improper, and long-range electrostatics, can run simultaneously in
threaded mode on the CPU cores. Since less MPI tasks than CPU cores
will typically be invoked when running with coprocessors, this enables
the extra CPU cores to be used for useful computation.</p>
<p>As illustrated below, if LAMMPS is built with both the USER-INTEL and
USER-OMP packages, this dual mode of operation is made easier to use,
via the &#8220;-suffix hybrid intel omp&#8221; <a class="reference internal" href="Section_start.html#start-7"><span>command-line switch</span></a> or the <a class="reference internal" href="suffix.html"><em>suffix hybrid intel omp</em></a> command. Both set a second-choice suffix to &#8220;omp&#8221; so
that styles from the USER-INTEL package will be used if available,
with styles from the USER-OMP package as a second choice.</p>
<p>Here is a quick overview of how to use the USER-INTEL package for CPU
acceleration, assuming one or more 16-core nodes. More details
follow.</p>
<div class="highlight-python"><div class="highlight"><pre>use an Intel compiler
use these CCFLAGS settings in Makefile.machine: -fopenmp, -DLAMMPS_MEMALIGN=64, -restrict, -xHost, -fno-alias, -ansi-alias, -override-limits
use these LINKFLAGS settings in Makefile.machine: -fopenmp, -xHost
make yes-user-intel yes-user-omp # including user-omp is optional
make mpi # build with the USER-INTEL package, if settings (including compiler) added to Makefile.mpi
make intel_cpu # or Makefile.intel_cpu already has settings, uses Intel MPI wrapper
Make.py -v -p intel omp -intel cpu -a file mpich_icc # or one-line build via Make.py for MPICH
Make.py -v -p intel omp -intel cpu -a file ompi_icc # or for OpenMPI
Make.py -v -p intel omp -intel cpu -a file intel_cpu # or for Intel MPI wrapper
</pre></div>
</div>
<div class="highlight-python"><div class="highlight"><pre>lmp_machine -sf intel -pk intel 0 omp 16 -in in.script # 1 node, 1 MPI task/node, 16 threads/task, no USER-OMP
mpirun -np 32 lmp_machine -sf intel -in in.script # 2 nodess, 16 MPI tasks/node, no threads, no USER-OMP
lmp_machine -sf hybrid intel omp -pk intel 0 omp 16 -pk omp 16 -in in.script # 1 node, 1 MPI task/node, 16 threads/task, with USER-OMP
mpirun -np 32 -ppn 4 lmp_machine -sf hybrid intel omp -pk omp 4 -pk omp 4 -in in.script # 8 nodes, 4 MPI tasks/node, 4 threads/task, with USER-OMP
</pre></div>
</div>
<p>Here is a quick overview of how to use the USER-INTEL package for the
same CPUs as above (16 cores/node), with an additional Xeon Phi(TM)
coprocessor per node. More details follow.</p>
<div class="highlight-python"><div class="highlight"><pre>Same as above for building, with these additions/changes:
add the flag -DLMP_INTEL_OFFLOAD to CCFLAGS in Makefile.machine
add the flag -offload to LINKFLAGS in Makefile.machine
for Make.py change &quot;-intel cpu&quot; to &quot;-intel phi&quot;, and &quot;file intel_cpu&quot; to &quot;file intel_phi&quot;
</pre></div>
</div>
<div class="highlight-python"><div class="highlight"><pre>mpirun -np 32 lmp_machine -sf intel -pk intel 1 -in in.script # 2 nodes, 16 MPI tasks/node, 240 total threads on coprocessor, no USER-OMP
mpirun -np 16 -ppn 8 lmp_machine -sf intel -pk intel 1 omp 2 -in in.script # 2 nodes, 8 MPI tasks/node, 2 threads/task, 240 total threads on coprocessor, no USER-OMP
mpirun -np 32 -ppn 8 lmp_machine -sf hybrid intel omp -pk intel 1 omp 2 -pk omp 2 -in in.script # 4 nodes, 8 MPI tasks/node, 2 threads/task, 240 total threads on coprocessor, with USER-OMP
</pre></div>
</div>
<p><strong>Required hardware/software:</strong></p>
<p>Your compiler must support the OpenMP interface. Use of an Intel(R)
C++ compiler is recommended, but not required. However, g++ will not
recognize some of the settings listed above, so they cannot be used.
Optimizations for vectorization have only been tested with the
Intel(R) compiler. Use of other compilers may not result in
vectorization, or give poor performance.</p>
<p>The recommended version of the Intel(R) compiler is 14.0.1.106.
Versions 15.0.1.133 and later are also supported. If using Intel(R)
MPI, versions 15.0.2.044 and later are recommended.</p>
<p>To use the offload option, you must have one or more Intel(R) Xeon
Phi(TM) coprocessors and use an Intel(R) C++ compiler.</p>
<p><strong>Building LAMMPS with the USER-INTEL package:</strong></p>
<p>The lines above illustrate how to include/build with the USER-INTEL
package, for either CPU or Phi support, in two steps, using the &#8220;make&#8221;
command. Or how to do it with one command via the src/Make.py script,
described in <a class="reference internal" href="Section_start.html#start-4"><span>Section 2.4</span></a> of the manual.
Type &#8220;Make.py -h&#8221; for help. Because the mechanism for specifing what
compiler to use (Intel in this case) is different for different MPI
wrappers, 3 versions of the Make.py command are shown.</p>
<p>Note that if you build with support for a Phi coprocessor, the same
binary can be used on nodes with or without coprocessors installed.
However, if you do not have coprocessors on your system, building
without offload support will produce a smaller binary.</p>
<p>If you also build with the USER-OMP package, you can use styles from
both packages, as described below.</p>
<p>Note that the CCFLAGS and LINKFLAGS settings in Makefile.machine must
include &#8220;-fopenmp&#8221;. Likewise, if you use an Intel compiler, the
CCFLAGS setting must include &#8220;-restrict&#8221;. For Phi support, the
&#8220;-DLMP_INTEL_OFFLOAD&#8221; (CCFLAGS) and &#8220;-offload&#8221; (LINKFLAGS) settings
are required. The other settings listed above are optional, but will
typically improve performance. The Make.py command will add all of
these automatically.</p>
<p>If you are compiling on the same architecture that will be used for
the runs, adding the flag <em>-xHost</em> to CCFLAGS enables vectorization
with the Intel(R) compiler. Otherwise, you must provide the correct
compute node architecture to the -x option (e.g. -xAVX).</p>
<p>Example machines makefiles Makefile.intel_cpu and Makefile.intel_phi
are included in the src/MAKE/OPTIONS directory with settings that
perform well with the Intel(R) compiler. The latter has support for
offload to Phi coprocessors; the former does not.</p>
<p><strong>Run with the USER-INTEL package from the command line:</strong></p>
<p>The mpirun or mpiexec command sets the total number of MPI tasks used
by LAMMPS (one or multiple per compute node) and the number of MPI
tasks used per node. E.g. the mpirun command in MPICH does this via
its -np and -ppn switches. Ditto for OpenMPI via -np and -npernode.</p>
<p>If you compute (any portion of) pairwise interactions using USER-INTEL
pair styles on the CPU, or use USER-OMP styles on the CPU, you need to
choose how many OpenMP threads per MPI task to use. If both packages
are used, it must be done for both packages, and the same thread count
value should be used for both. Note that the product of MPI tasks *
threads/task should not exceed the physical number of cores (on a
node), otherwise performance will suffer.</p>
<p>When using the USER-INTEL package for the Phi, you also need to
specify the number of coprocessor/node and optionally the number of
coprocessor threads per MPI task to use. Note that coprocessor
threads (which run on the coprocessor) are totally independent from
OpenMP threads (which run on the CPU). The default values for the
settings that affect coprocessor threads are typically fine, as
discussed below.</p>
<p>As in the lines above, use the &#8220;-sf intel&#8221; or &#8220;-sf hybrid intel omp&#8221;
<a class="reference internal" href="Section_start.html#start-7"><span>command-line switch</span></a>, which will
automatically append &#8220;intel&#8221; to styles that support it. In the second
case, &#8220;omp&#8221; will be appended if an &#8220;intel&#8221; style does not exist.</p>
<p>Note that if either switch is used, it also invokes a default command:
<a class="reference internal" href="package.html"><em>package intel 1</em></a>. If the &#8220;-sf hybrid intel omp&#8221; switch
is used, the default USER-OMP command <a class="reference internal" href="package.html"><em>package omp 0</em></a> is
also invoked (if LAMMPS was built with USER-OMP). Both set the number
of OpenMP threads per MPI task via the OMP_NUM_THREADS environment
variable. The first command sets the number of Xeon Phi(TM)
coprocessors/node to 1 (ignored if USER-INTEL is built for CPU-only),
and the precision mode to &#8220;mixed&#8221; (default value).</p>
<p>You can also use the &#8220;-pk intel Nphi&#8221; <a class="reference internal" href="Section_start.html#start-7"><span>command-line switch</span></a> to explicitly set Nphi = # of Xeon
Phi(TM) coprocessors/node, as well as additional options. Nphi should
be &gt;= 1 if LAMMPS was built with coprocessor support, otherswise Nphi
= 0 for a CPU-only build. All the available coprocessor threads on
each Phi will be divided among MPI tasks, unless the <em>tptask</em> option
of the &#8220;-pk intel&#8221; <a class="reference internal" href="Section_start.html#start-7"><span>command-line switch</span></a> is
used to limit the coprocessor threads per MPI task. See the <a class="reference internal" href="package.html"><em>package intel</em></a> command for details, including the default values
used for all its options if not specified, and how to set the number
of OpenMP threads via the OMP_NUM_THREADS environment variable if
desired.</p>
<p>If LAMMPS was built with the USER-OMP package, you can also use the
&#8220;-pk omp Nt&#8221; <a class="reference internal" href="Section_start.html#start-7"><span>command-line switch</span></a> to
explicitly set Nt = # of OpenMP threads per MPI task to use, as well
as additional options. Nt should be the same threads per MPI task as
set for the USER-INTEL package, e.g. via the &#8220;-pk intel Nphi omp Nt&#8221;
command. Again, see the <a class="reference internal" href="package.html"><em>package omp</em></a> command for
details, including the default values used for all its options if not
specified, and how to set the number of OpenMP threads via the
OMP_NUM_THREADS environment variable if desired.</p>
<p><strong>Or run with the USER-INTEL package by editing an input script:</strong></p>
<p>The discussion above for the mpirun/mpiexec command, MPI tasks/node,
OpenMP threads per MPI task, and coprocessor threads per MPI task is
the same.</p>
<p>Use the <a class="reference internal" href="suffix.html"><em>suffix intel</em></a> or <a class="reference internal" href="suffix.html"><em>suffix hybrid intel omp</em></a> commands, or you can explicitly add an &#8220;intel&#8221; or
&#8220;omp&#8221; suffix to individual styles in your input script, e.g.</p>
<div class="highlight-python"><div class="highlight"><pre>pair_style lj/cut/intel 2.5
</pre></div>
</div>
<p>You must also use the <a class="reference internal" href="package.html"><em>package intel</em></a> command, unless the
&#8220;-sf intel&#8221; or &#8220;-pk intel&#8221; <a class="reference internal" href="Section_start.html#start-7"><span>command-line switches</span></a> were used. It specifies how many
coprocessors/node to use, as well as other OpenMP threading and
coprocessor options. The <a class="reference internal" href="package.html"><em>package</em></a> doc page explains how
to set the number of OpenMP threads via an environment variable if
desired.</p>
<p>If LAMMPS was also built with the USER-OMP package, you must also use
the <a class="reference internal" href="package.html"><em>package omp</em></a> command to enable that package, unless
the &#8220;-sf hybrid intel omp&#8221; or &#8220;-pk omp&#8221; <a class="reference internal" href="Section_start.html#start-7"><span>command-line switches</span></a> were used. It specifies how many
OpenMP threads per MPI task to use (should be same as the setting for
the USER-INTEL package), as well as other options. Its doc page
explains how to set the number of OpenMP threads via an environment
variable if desired.</p>
<p><strong>Speed-ups to expect:</strong></p>
<p>If LAMMPS was not built with coprocessor support (CPU only) when
including the USER-INTEL package, then acclerated styles will run on
the CPU using vectorization optimizations and the specified precision.
This may give a substantial speed-up for a pair style, particularly if
mixed or single precision is used.</p>
<p>If LAMMPS was built with coproccesor support, the pair styles will run
on one or more Intel(R) Xeon Phi(TM) coprocessors (per node). The
performance of a Xeon Phi versus a multi-core CPU is a function of
your hardware, which pair style is used, the number of
atoms/coprocessor, and the precision used on the coprocessor (double,
single, mixed).</p>
<p>See the <a class="reference external" href="http://lammps.sandia.gov/bench.html">Benchmark page</a> of the
LAMMPS web site for performance of the USER-INTEL package on different
hardware.</p>
<div class="admonition warning">
<p class="first admonition-title">Warning</p>
<p class="last">Setting core affinity is often used to pin MPI tasks
and OpenMP threads to a core or group of cores so that memory access
can be uniform. Unless disabled at build time, affinity for MPI tasks
and OpenMP threads on the host (CPU) will be set by default on the
host when using offload to a coprocessor. In this case, it is
unnecessary to use other methods to control affinity (e.g. taskset,
numactl, I_MPI_PIN_DOMAIN, etc.). This can be disabled in an input
script with the <em>no_affinity</em> option to the <a class="reference internal" href="package.html"><em>package intel</em></a> command or by disabling the option at build time
(by adding -DINTEL_OFFLOAD_NOAFFINITY to the CCFLAGS line of your
Makefile). Disabling this option is not recommended, especially when
running on a machine with hyperthreading disabled.</p>
</div>
<p><strong>Guidelines for best performance on an Intel(R) Xeon Phi(TM)
coprocessor:</strong></p>
<ul class="simple">
<li>The default for the <a class="reference internal" href="package.html"><em>package intel</em></a> command is to have
all the MPI tasks on a given compute node use a single Xeon Phi(TM)
coprocessor. In general, running with a large number of MPI tasks on
each node will perform best with offload. Each MPI task will
automatically get affinity to a subset of the hardware threads
available on the coprocessor. For example, if your card has 61 cores,
with 60 cores available for offload and 4 hardware threads per core
(240 total threads), running with 24 MPI tasks per node will cause
each MPI task to use a subset of 10 threads on the coprocessor. Fine
tuning of the number of threads to use per MPI task or the number of
threads to use per core can be accomplished with keyword settings of
the <a class="reference internal" href="package.html"><em>package intel</em></a> command.</li>
<li>If desired, only a fraction of the pair style computation can be
offloaded to the coprocessors. This is accomplished by using the
<em>balance</em> keyword in the <a class="reference internal" href="package.html"><em>package intel</em></a> command. A
balance of 0 runs all calculations on the CPU. A balance of 1 runs
all calculations on the coprocessor. A balance of 0.5 runs half of
the calculations on the coprocessor. Setting the balance to -1 (the
default) will enable dynamic load balancing that continously adjusts
the fraction of offloaded work throughout the simulation. This option
typically produces results within 5 to 10 percent of the optimal fixed
balance.</li>
<li>When using offload with CPU hyperthreading disabled, it may help
performance to use fewer MPI tasks and OpenMP threads than available
cores. This is due to the fact that additional threads are generated
internally to handle the asynchronous offload tasks.</li>
<li>If running short benchmark runs with dynamic load balancing, adding a
short warm-up run (10-20 steps) will allow the load-balancer to find a
near-optimal setting that will carry over to additional runs.</li>
<li>If pair computations are being offloaded to an Intel(R) Xeon Phi(TM)
coprocessor, a diagnostic line is printed to the screen (not to the
log file), during the setup phase of a run, indicating that offload
mode is being used and indicating the number of coprocessor threads
per MPI task. Additionally, an offload timing summary is printed at
the end of each run. When offloading, the frequency for <a class="reference internal" href="atom_modify.html"><em>atom sorting</em></a> is changed to 1 so that the per-atom data is
effectively sorted at every rebuild of the neighbor lists.</li>
<li>For simulations with long-range electrostatics or bond, angle,
dihedral, improper calculations, computation and data transfer to the
coprocessor will run concurrently with computations and MPI
communications for these calculations on the host CPU. The USER-INTEL
package has two modes for deciding which atoms will be handled by the
coprocessor. This choice is controlled with the <em>ghost</em> keyword of
the <a class="reference internal" href="package.html"><em>package intel</em></a> command. When set to 0, ghost atoms
(atoms at the borders between MPI tasks) are not offloaded to the
card. This allows for overlap of MPI communication of forces with
computation on the coprocessor when the <a class="reference internal" href="newton.html"><em>newton</em></a> setting
is &#8220;on&#8221;. The default is dependent on the style being used, however,
better performance may be achieved by setting this option
explictly.</li>
</ul>
<div class="section" id="restrictions">
<h2>Restrictions<a class="headerlink" href="#restrictions" title="Permalink to this headline"></a></h2>
<p>When offloading to a coprocessor, <a class="reference internal" href="pair_hybrid.html"><em>hybrid</em></a> styles
that require skip lists for neighbor builds cannot be offloaded.
Using <a class="reference internal" href="pair_hybrid.html"><em>hybrid/overlay</em></a> is allowed. Only one intel
accelerated style may be used with hybrid styles.
<a class="reference internal" href="special_bonds.html"><em>Special_bonds</em></a> exclusion lists are not currently
supported with offload, however, the same effect can often be
accomplished by setting cutoffs for excluded atom types to 0. None of
the pair styles in the USER-INTEL package currently support the
&#8220;inner&#8221;, &#8220;middle&#8221;, &#8220;outer&#8221; options for rRESPA integration via the
<a class="reference internal" href="run_style.html"><em>run_style respa</em></a> command; only the &#8220;pair&#8221; option is
supported.</p>
</div>
</div>
</div>
</div>
<footer>
<hr/>
<div role="contentinfo">
<p>
&copy; Copyright .
</p>
</div>
Built with <a href="http://sphinx-doc.org/">Sphinx</a> using a <a href="https://github.com/snide/sphinx_rtd_theme">theme</a> provided by <a href="https://readthedocs.org">Read the Docs</a>.
</footer>
</div>
</div>
</section>
</div>
<script type="text/javascript">
var DOCUMENTATION_OPTIONS = {
URL_ROOT:'./',
VERSION:'15 May 2015 version',
COLLAPSE_INDEX:false,
FILE_SUFFIX:'.html',
HAS_SOURCE: true
};
</script>
<script type="text/javascript" src="_static/jquery.js"></script>
<script type="text/javascript" src="_static/underscore.js"></script>
<script type="text/javascript" src="_static/doctools.js"></script>
<script type="text/javascript" src="https://cdn.mathjax.org/mathjax/latest/MathJax.js?config=TeX-AMS-MML_HTMLorMML"></script>
<script type="text/javascript" src="_static/sphinxcontrib-images/LightBox2/lightbox2/js/jquery-1.11.0.min.js"></script>
<script type="text/javascript" src="_static/sphinxcontrib-images/LightBox2/lightbox2/js/lightbox.min.js"></script>
<script type="text/javascript" src="_static/sphinxcontrib-images/LightBox2/lightbox2-customize/jquery-noconflict.js"></script>
<script type="text/javascript" src="_static/js/theme.js"></script>
<script type="text/javascript">
jQuery(function () {
SphinxRtdTheme.StickyNav.enable();
});
</script>
</body>
</html>

View File

@ -1,321 +0,0 @@
"Previous Section"_Section_packages.html - "LAMMPS WWW Site"_lws -
"LAMMPS Documentation"_ld - "LAMMPS Commands"_lc :c
:link(lws,http://lammps.sandia.gov)
:link(ld,Manual.html)
:link(lc,Section_commands.html#comm)
:line
"Return to Section accelerate overview"_Section_accelerate.html
5.3.3 USER-INTEL package :h4
The USER-INTEL package was developed by Mike Brown at Intel
Corporation. It provides two methods for accelerating simulations,
depending on the hardware you have. The first is acceleration on
Intel(R) CPUs by running in single, mixed, or double precision with
vectorization. The second is acceleration on Intel(R) Xeon Phi(TM)
coprocessors via offloading neighbor list and non-bonded force
calculations to the Phi. The same C++ code is used in both cases.
When offloading to a coprocessor from a CPU, the same routine is run
twice, once on the CPU and once with an offload flag.
Note that the USER-INTEL package supports use of the Phi in "offload"
mode, not "native" mode like the "KOKKOS
package"_accelerate_kokkos.html.
Also note that the USER-INTEL package can be used in tandem with the
"USER-OMP package"_accelerate_omp.html. This is useful when
offloading pair style computations to the Phi, so that other styles
not supported by the USER-INTEL package, e.g. bond, angle, dihedral,
improper, and long-range electrostatics, can run simultaneously in
threaded mode on the CPU cores. Since less MPI tasks than CPU cores
will typically be invoked when running with coprocessors, this enables
the extra CPU cores to be used for useful computation.
As illustrated below, if LAMMPS is built with both the USER-INTEL and
USER-OMP packages, this dual mode of operation is made easier to use,
via the "-suffix hybrid intel omp" "command-line
switch"_Section_start.html#start_7 or the "suffix hybrid intel
omp"_suffix.html command. Both set a second-choice suffix to "omp" so
that styles from the USER-INTEL package will be used if available,
with styles from the USER-OMP package as a second choice.
Here is a quick overview of how to use the USER-INTEL package for CPU
acceleration, assuming one or more 16-core nodes. More details
follow.
use an Intel compiler
use these CCFLAGS settings in Makefile.machine: -fopenmp, -DLAMMPS_MEMALIGN=64, -restrict, -xHost, -fno-alias, -ansi-alias, -override-limits
use these LINKFLAGS settings in Makefile.machine: -fopenmp, -xHost
make yes-user-intel yes-user-omp # including user-omp is optional
make mpi # build with the USER-INTEL package, if settings (including compiler) added to Makefile.mpi
make intel_cpu # or Makefile.intel_cpu already has settings, uses Intel MPI wrapper
Make.py -v -p intel omp -intel cpu -a file mpich_icc # or one-line build via Make.py for MPICH
Make.py -v -p intel omp -intel cpu -a file ompi_icc # or for OpenMPI
Make.py -v -p intel omp -intel cpu -a file intel_cpu # or for Intel MPI wrapper :pre
lmp_machine -sf intel -pk intel 0 omp 16 -in in.script # 1 node, 1 MPI task/node, 16 threads/task, no USER-OMP
mpirun -np 32 lmp_machine -sf intel -in in.script # 2 nodess, 16 MPI tasks/node, no threads, no USER-OMP
lmp_machine -sf hybrid intel omp -pk intel 0 omp 16 -pk omp 16 -in in.script # 1 node, 1 MPI task/node, 16 threads/task, with USER-OMP
mpirun -np 32 -ppn 4 lmp_machine -sf hybrid intel omp -pk omp 4 -pk omp 4 -in in.script # 8 nodes, 4 MPI tasks/node, 4 threads/task, with USER-OMP :pre
Here is a quick overview of how to use the USER-INTEL package for the
same CPUs as above (16 cores/node), with an additional Xeon Phi(TM)
coprocessor per node. More details follow.
Same as above for building, with these additions/changes:
add the flag -DLMP_INTEL_OFFLOAD to CCFLAGS in Makefile.machine
add the flag -offload to LINKFLAGS in Makefile.machine
for Make.py change "-intel cpu" to "-intel phi", and "file intel_cpu" to "file intel_phi" :pre
mpirun -np 32 lmp_machine -sf intel -pk intel 1 -in in.script # 2 nodes, 16 MPI tasks/node, 240 total threads on coprocessor, no USER-OMP
mpirun -np 16 -ppn 8 lmp_machine -sf intel -pk intel 1 omp 2 -in in.script # 2 nodes, 8 MPI tasks/node, 2 threads/task, 240 total threads on coprocessor, no USER-OMP
mpirun -np 32 -ppn 8 lmp_machine -sf hybrid intel omp -pk intel 1 omp 2 -pk omp 2 -in in.script # 4 nodes, 8 MPI tasks/node, 2 threads/task, 240 total threads on coprocessor, with USER-OMP :pre
[Required hardware/software:]
Your compiler must support the OpenMP interface. Use of an Intel(R)
C++ compiler is recommended, but not required. However, g++ will not
recognize some of the settings listed above, so they cannot be used.
Optimizations for vectorization have only been tested with the
Intel(R) compiler. Use of other compilers may not result in
vectorization, or give poor performance.
The recommended version of the Intel(R) compiler is 14.0.1.106.
Versions 15.0.1.133 and later are also supported. If using Intel(R)
MPI, versions 15.0.2.044 and later are recommended.
To use the offload option, you must have one or more Intel(R) Xeon
Phi(TM) coprocessors and use an Intel(R) C++ compiler.
[Building LAMMPS with the USER-INTEL package:]
The lines above illustrate how to include/build with the USER-INTEL
package, for either CPU or Phi support, in two steps, using the "make"
command. Or how to do it with one command via the src/Make.py script,
described in "Section 2.4"_Section_start.html#start_4 of the manual.
Type "Make.py -h" for help. Because the mechanism for specifing what
compiler to use (Intel in this case) is different for different MPI
wrappers, 3 versions of the Make.py command are shown.
Note that if you build with support for a Phi coprocessor, the same
binary can be used on nodes with or without coprocessors installed.
However, if you do not have coprocessors on your system, building
without offload support will produce a smaller binary.
If you also build with the USER-OMP package, you can use styles from
both packages, as described below.
Note that the CCFLAGS and LINKFLAGS settings in Makefile.machine must
include "-fopenmp". Likewise, if you use an Intel compiler, the
CCFLAGS setting must include "-restrict". For Phi support, the
"-DLMP_INTEL_OFFLOAD" (CCFLAGS) and "-offload" (LINKFLAGS) settings
are required. The other settings listed above are optional, but will
typically improve performance. The Make.py command will add all of
these automatically.
If you are compiling on the same architecture that will be used for
the runs, adding the flag {-xHost} to CCFLAGS enables vectorization
with the Intel(R) compiler. Otherwise, you must provide the correct
compute node architecture to the -x option (e.g. -xAVX).
Example machines makefiles Makefile.intel_cpu and Makefile.intel_phi
are included in the src/MAKE/OPTIONS directory with settings that
perform well with the Intel(R) compiler. The latter has support for
offload to Phi coprocessors; the former does not.
[Run with the USER-INTEL package from the command line:]
The mpirun or mpiexec command sets the total number of MPI tasks used
by LAMMPS (one or multiple per compute node) and the number of MPI
tasks used per node. E.g. the mpirun command in MPICH does this via
its -np and -ppn switches. Ditto for OpenMPI via -np and -npernode.
If you compute (any portion of) pairwise interactions using USER-INTEL
pair styles on the CPU, or use USER-OMP styles on the CPU, you need to
choose how many OpenMP threads per MPI task to use. If both packages
are used, it must be done for both packages, and the same thread count
value should be used for both. Note that the product of MPI tasks *
threads/task should not exceed the physical number of cores (on a
node), otherwise performance will suffer.
When using the USER-INTEL package for the Phi, you also need to
specify the number of coprocessor/node and optionally the number of
coprocessor threads per MPI task to use. Note that coprocessor
threads (which run on the coprocessor) are totally independent from
OpenMP threads (which run on the CPU). The default values for the
settings that affect coprocessor threads are typically fine, as
discussed below.
As in the lines above, use the "-sf intel" or "-sf hybrid intel omp"
"command-line switch"_Section_start.html#start_7, which will
automatically append "intel" to styles that support it. In the second
case, "omp" will be appended if an "intel" style does not exist.
Note that if either switch is used, it also invokes a default command:
"package intel 1"_package.html. If the "-sf hybrid intel omp" switch
is used, the default USER-OMP command "package omp 0"_package.html is
also invoked (if LAMMPS was built with USER-OMP). Both set the number
of OpenMP threads per MPI task via the OMP_NUM_THREADS environment
variable. The first command sets the number of Xeon Phi(TM)
coprocessors/node to 1 (ignored if USER-INTEL is built for CPU-only),
and the precision mode to "mixed" (default value).
You can also use the "-pk intel Nphi" "command-line
switch"_Section_start.html#start_7 to explicitly set Nphi = # of Xeon
Phi(TM) coprocessors/node, as well as additional options. Nphi should
be >= 1 if LAMMPS was built with coprocessor support, otherswise Nphi
= 0 for a CPU-only build. All the available coprocessor threads on
each Phi will be divided among MPI tasks, unless the {tptask} option
of the "-pk intel" "command-line switch"_Section_start.html#start_7 is
used to limit the coprocessor threads per MPI task. See the "package
intel"_package.html command for details, including the default values
used for all its options if not specified, and how to set the number
of OpenMP threads via the OMP_NUM_THREADS environment variable if
desired.
If LAMMPS was built with the USER-OMP package, you can also use the
"-pk omp Nt" "command-line switch"_Section_start.html#start_7 to
explicitly set Nt = # of OpenMP threads per MPI task to use, as well
as additional options. Nt should be the same threads per MPI task as
set for the USER-INTEL package, e.g. via the "-pk intel Nphi omp Nt"
command. Again, see the "package omp"_package.html command for
details, including the default values used for all its options if not
specified, and how to set the number of OpenMP threads via the
OMP_NUM_THREADS environment variable if desired.
[Or run with the USER-INTEL package by editing an input script:]
The discussion above for the mpirun/mpiexec command, MPI tasks/node,
OpenMP threads per MPI task, and coprocessor threads per MPI task is
the same.
Use the "suffix intel"_suffix.html or "suffix hybrid intel
omp"_suffix.html commands, or you can explicitly add an "intel" or
"omp" suffix to individual styles in your input script, e.g.
pair_style lj/cut/intel 2.5 :pre
You must also use the "package intel"_package.html command, unless the
"-sf intel" or "-pk intel" "command-line
switches"_Section_start.html#start_7 were used. It specifies how many
coprocessors/node to use, as well as other OpenMP threading and
coprocessor options. The "package"_package.html doc page explains how
to set the number of OpenMP threads via an environment variable if
desired.
If LAMMPS was also built with the USER-OMP package, you must also use
the "package omp"_package.html command to enable that package, unless
the "-sf hybrid intel omp" or "-pk omp" "command-line
switches"_Section_start.html#start_7 were used. It specifies how many
OpenMP threads per MPI task to use (should be same as the setting for
the USER-INTEL package), as well as other options. Its doc page
explains how to set the number of OpenMP threads via an environment
variable if desired.
[Speed-ups to expect:]
If LAMMPS was not built with coprocessor support (CPU only) when
including the USER-INTEL package, then acclerated styles will run on
the CPU using vectorization optimizations and the specified precision.
This may give a substantial speed-up for a pair style, particularly if
mixed or single precision is used.
If LAMMPS was built with coproccesor support, the pair styles will run
on one or more Intel(R) Xeon Phi(TM) coprocessors (per node). The
performance of a Xeon Phi versus a multi-core CPU is a function of
your hardware, which pair style is used, the number of
atoms/coprocessor, and the precision used on the coprocessor (double,
single, mixed).
See the "Benchmark page"_http://lammps.sandia.gov/bench.html of the
LAMMPS web site for performance of the USER-INTEL package on different
hardware.
IMPORTANT NOTE: Setting core affinity is often used to pin MPI tasks
and OpenMP threads to a core or group of cores so that memory access
can be uniform. Unless disabled at build time, affinity for MPI tasks
and OpenMP threads on the host (CPU) will be set by default on the
host when using offload to a coprocessor. In this case, it is
unnecessary to use other methods to control affinity (e.g. taskset,
numactl, I_MPI_PIN_DOMAIN, etc.). This can be disabled in an input
script with the {no_affinity} option to the "package
intel"_package.html command or by disabling the option at build time
(by adding -DINTEL_OFFLOAD_NOAFFINITY to the CCFLAGS line of your
Makefile). Disabling this option is not recommended, especially when
running on a machine with hyperthreading disabled.
[Guidelines for best performance on an Intel(R) Xeon Phi(TM)
coprocessor:]
The default for the "package intel"_package.html command is to have
all the MPI tasks on a given compute node use a single Xeon Phi(TM)
coprocessor. In general, running with a large number of MPI tasks on
each node will perform best with offload. Each MPI task will
automatically get affinity to a subset of the hardware threads
available on the coprocessor. For example, if your card has 61 cores,
with 60 cores available for offload and 4 hardware threads per core
(240 total threads), running with 24 MPI tasks per node will cause
each MPI task to use a subset of 10 threads on the coprocessor. Fine
tuning of the number of threads to use per MPI task or the number of
threads to use per core can be accomplished with keyword settings of
the "package intel"_package.html command. :ulb,l
If desired, only a fraction of the pair style computation can be
offloaded to the coprocessors. This is accomplished by using the
{balance} keyword in the "package intel"_package.html command. A
balance of 0 runs all calculations on the CPU. A balance of 1 runs
all calculations on the coprocessor. A balance of 0.5 runs half of
the calculations on the coprocessor. Setting the balance to -1 (the
default) will enable dynamic load balancing that continously adjusts
the fraction of offloaded work throughout the simulation. This option
typically produces results within 5 to 10 percent of the optimal fixed
balance. :l
When using offload with CPU hyperthreading disabled, it may help
performance to use fewer MPI tasks and OpenMP threads than available
cores. This is due to the fact that additional threads are generated
internally to handle the asynchronous offload tasks. :l
If running short benchmark runs with dynamic load balancing, adding a
short warm-up run (10-20 steps) will allow the load-balancer to find a
near-optimal setting that will carry over to additional runs. :l
If pair computations are being offloaded to an Intel(R) Xeon Phi(TM)
coprocessor, a diagnostic line is printed to the screen (not to the
log file), during the setup phase of a run, indicating that offload
mode is being used and indicating the number of coprocessor threads
per MPI task. Additionally, an offload timing summary is printed at
the end of each run. When offloading, the frequency for "atom
sorting"_atom_modify.html is changed to 1 so that the per-atom data is
effectively sorted at every rebuild of the neighbor lists. :l
For simulations with long-range electrostatics or bond, angle,
dihedral, improper calculations, computation and data transfer to the
coprocessor will run concurrently with computations and MPI
communications for these calculations on the host CPU. The USER-INTEL
package has two modes for deciding which atoms will be handled by the
coprocessor. This choice is controlled with the {ghost} keyword of
the "package intel"_package.html command. When set to 0, ghost atoms
(atoms at the borders between MPI tasks) are not offloaded to the
card. This allows for overlap of MPI communication of forces with
computation on the coprocessor when the "newton"_newton.html setting
is "on". The default is dependent on the style being used, however,
better performance may be achieved by setting this option
explictly. :l,ule
[Restrictions:]
When offloading to a coprocessor, "hybrid"_pair_hybrid.html styles
that require skip lists for neighbor builds cannot be offloaded.
Using "hybrid/overlay"_pair_hybrid.html is allowed. Only one intel
accelerated style may be used with hybrid styles.
"Special_bonds"_special_bonds.html exclusion lists are not currently
supported with offload, however, the same effect can often be
accomplished by setting cutoffs for excluded atom types to 0. None of
the pair styles in the USER-INTEL package currently support the
"inner", "middle", "outer" options for rRESPA integration via the
"run_style respa"_run_style.html command; only the "pair" option is
supported.

View File

@ -1,622 +0,0 @@
<!DOCTYPE html>
<!--[if IE 8]><html class="no-js lt-ie9" lang="en" > <![endif]-->
<!--[if gt IE 8]><!--> <html class="no-js" lang="en" > <!--<![endif]-->
<head>
<meta charset="utf-8">
<meta name="viewport" content="width=device-width, initial-scale=1.0">
<title>5.KOKKOS package &mdash; LAMMPS 15 May 2015 version documentation</title>
<link rel="stylesheet" href="_static/css/theme.css" type="text/css" />
<link rel="stylesheet" href="_static/sphinxcontrib-images/LightBox2/lightbox2/css/lightbox.css" type="text/css" />
<link rel="top" title="LAMMPS 15 May 2015 version documentation" href="index.html"/>
<script src="_static/js/modernizr.min.js"></script>
</head>
<body class="wy-body-for-nav" role="document">
<div class="wy-grid-for-nav">
<nav data-toggle="wy-nav-shift" class="wy-nav-side">
<div class="wy-side-nav-search">
<a href="Manual.html" class="icon icon-home"> LAMMPS
</a>
<div role="search">
<form id="rtd-search-form" class="wy-form" action="search.html" method="get">
<input type="text" name="q" placeholder="Search docs" />
<input type="hidden" name="check_keywords" value="yes" />
<input type="hidden" name="area" value="default" />
</form>
</div>
</div>
<div class="wy-menu wy-menu-vertical" data-spy="affix" role="navigation" aria-label="main navigation">
<ul>
<li class="toctree-l1"><a class="reference internal" href="Section_intro.html">1. Introduction</a></li>
<li class="toctree-l1"><a class="reference internal" href="Section_start.html">2. Getting Started</a></li>
<li class="toctree-l1"><a class="reference internal" href="Section_commands.html">3. Commands</a></li>
<li class="toctree-l1"><a class="reference internal" href="Section_packages.html">4. Packages</a></li>
<li class="toctree-l1"><a class="reference internal" href="Section_accelerate.html">5. Accelerating LAMMPS performance</a></li>
<li class="toctree-l1"><a class="reference internal" href="Section_howto.html">6. How-to discussions</a></li>
<li class="toctree-l1"><a class="reference internal" href="Section_example.html">7. Example problems</a></li>
<li class="toctree-l1"><a class="reference internal" href="Section_perf.html">8. Performance &amp; scalability</a></li>
<li class="toctree-l1"><a class="reference internal" href="Section_tools.html">9. Additional tools</a></li>
<li class="toctree-l1"><a class="reference internal" href="Section_modify.html">10. Modifying &amp; extending LAMMPS</a></li>
<li class="toctree-l1"><a class="reference internal" href="Section_python.html">11. Python interface to LAMMPS</a></li>
<li class="toctree-l1"><a class="reference internal" href="Section_errors.html">12. Errors</a></li>
<li class="toctree-l1"><a class="reference internal" href="Section_history.html">13. Future and history</a></li>
</ul>
</div>
&nbsp;
</nav>
<section data-toggle="wy-nav-shift" class="wy-nav-content-wrap">
<nav class="wy-nav-top" role="navigation" aria-label="top navigation">
<i data-toggle="wy-nav-top" class="fa fa-bars"></i>
<a href="Manual.html">LAMMPS</a>
</nav>
<div class="wy-nav-content">
<div class="rst-content">
<div role="navigation" aria-label="breadcrumbs navigation">
<ul class="wy-breadcrumbs">
<li><a href="Manual.html">Docs</a> &raquo;</li>
<li>5.KOKKOS package</li>
<li class="wy-breadcrumbs-aside">
<a href="http://lammps.sandia.gov">Website</a>
<a href="Section_commands.html#comm">Commands</a>
</li>
</ul>
<hr/>
</div>
<div role="main" class="document" itemscope="itemscope" itemtype="http://schema.org/Article">
<div itemprop="articleBody">
<p><a class="reference internal" href="Section_accelerate.html"><em>Return to Section accelerate overview</em></a></p>
<div class="section" id="kokkos-package">
<h1>5.KOKKOS package<a class="headerlink" href="#kokkos-package" title="Permalink to this headline"></a></h1>
<p>The KOKKOS package was developed primarily by Christian Trott (Sandia)
with contributions of various styles by others, including Sikandar
Mashayak (UIUC), Stan Moore (Sandia), and Ray Shan (Sandia). The
underlying Kokkos library was written primarily by Carter Edwards,
Christian Trott, and Dan Sunderland (all Sandia).</p>
<p>The KOKKOS package contains versions of pair, fix, and atom styles
that use data structures and macros provided by the Kokkos library,
which is included with LAMMPS in lib/kokkos.</p>
<p>The Kokkos library is part of
<a class="reference external" href="http://trilinos.sandia.gov/packages/kokkos">Trilinos</a> and can also be
downloaded from <a class="reference external" href="https://github.com/kokkos/kokkos">Github</a>. Kokkos is a
templated C++ library that provides two key abstractions for an
application like LAMMPS. First, it allows a single implementation of
an application kernel (e.g. a pair style) to run efficiently on
different kinds of hardware, such as a GPU, Intel Phi, or many-core
CPU.</p>
<p>The Kokkos library also provides data abstractions to adjust (at
compile time) the memory layout of basic data structures like 2d and
3d arrays and allow the transparent utilization of special hardware
load and store operations. Such data structures are used in LAMMPS to
store atom coordinates or forces or neighbor lists. The layout is
chosen to optimize performance on different platforms. Again this
functionality is hidden from the developer, and does not affect how
the kernel is coded.</p>
<p>These abstractions are set at build time, when LAMMPS is compiled with
the KOKKOS package installed. All Kokkos operations occur within the
context of an individual MPI task running on a single node of the
machine. The total number of MPI tasks used by LAMMPS (one or
multiple per compute node) is set in the usual manner via the mpirun
or mpiexec commands, and is independent of Kokkos.</p>
<p>Kokkos currently provides support for 3 modes of execution (per MPI
task). These are OpenMP (for many-core CPUs), Cuda (for NVIDIA GPUs),
and OpenMP (for Intel Phi). Note that the KOKKOS package supports
running on the Phi in native mode, not offload mode like the
USER-INTEL package supports. You choose the mode at build time to
produce an executable compatible with specific hardware.</p>
<p>Here is a quick overview of how to use the KOKKOS package
for CPU acceleration, assuming one or more 16-core nodes.
More details follow.</p>
<p>use a C++11 compatible compiler
make yes-kokkos
make mpi KOKKOS_DEVICES=OpenMP # build with the KOKKOS package
make kokkos_omp # or Makefile.kokkos_omp already has variable set
Make.py -v -p kokkos -kokkos omp -o mpi -a file mpi # or one-line build via Make.py</p>
<div class="highlight-python"><div class="highlight"><pre>mpirun -np 16 lmp_mpi -k on -sf kk -in in.lj # 1 node, 16 MPI tasks/node, no threads
mpirun -np 2 -ppn 1 lmp_mpi -k on t 16 -sf kk -in in.lj # 2 nodes, 1 MPI task/node, 16 threads/task
mpirun -np 2 lmp_mpi -k on t 8 -sf kk -in in.lj # 1 node, 2 MPI tasks/node, 8 threads/task
mpirun -np 32 -ppn 4 lmp_mpi -k on t 4 -sf kk -in in.lj # 8 nodes, 4 MPI tasks/node, 4 threads/task
</pre></div>
</div>
<ul class="simple">
<li>specify variables and settings in your Makefile.machine that enable OpenMP, GPU, or Phi support</li>
<li>include the KOKKOS package and build LAMMPS</li>
<li>enable the KOKKOS package and its hardware options via the &#8220;-k on&#8221; command-line switch use KOKKOS styles in your input script</li>
</ul>
<p>Here is a quick overview of how to use the KOKKOS package for GPUs,
assuming one or more nodes, each with 16 cores and a GPU. More
details follow.</p>
<p>discuss use of NVCC, which Makefiles to examine</p>
<p>use a C++11 compatible compiler
KOKKOS_DEVICES = Cuda, OpenMP
KOKKOS_ARCH = Kepler35
make yes-kokkos
make machine
Make.py -p kokkos -kokkos cuda arch=31 -o kokkos_cuda -a file kokkos_cuda</p>
<div class="highlight-python"><div class="highlight"><pre>mpirun -np 1 lmp_cuda -k on t 6 -sf kk -in in.lj # one MPI task, 6 threads on CPU
mpirun -np 4 -ppn 1 lmp_cuda -k on t 6 -sf kk -in in.lj # ditto on 4 nodes
</pre></div>
</div>
<div class="highlight-python"><div class="highlight"><pre>mpirun -np 2 lmp_cuda -k on t 8 g 2 -sf kk -in in.lj # two MPI tasks, 8 threads per CPU
mpirun -np 32 -ppn 2 lmp_cuda -k on t 8 g 2 -sf kk -in in.lj # ditto on 16 nodes
</pre></div>
</div>
<p>Here is a quick overview of how to use the KOKKOS package
for the Intel Phi:</p>
<div class="highlight-python"><div class="highlight"><pre>use a C++11 compatible compiler
KOKKOS_DEVICES = OpenMP
KOKKOS_ARCH = KNC
make yes-kokkos
make machine
Make.py -p kokkos -kokkos phi -o kokkos_phi -a file mpi
</pre></div>
</div>
<p>host=MIC, Intel Phi with 61 cores (240 threads/phi via 4x hardware threading):
mpirun -np 1 lmp_g++ -k on t 240 -sf kk -in in.lj # 1 MPI task on 1 Phi, 1*240 = 240
mpirun -np 30 lmp_g++ -k on t 8 -sf kk -in in.lj # 30 MPI tasks on 1 Phi, 30*8 = 240
mpirun -np 12 lmp_g++ -k on t 20 -sf kk -in in.lj # 12 MPI tasks on 1 Phi, 12*20 = 240
mpirun -np 96 -ppn 12 lmp_g++ -k on t 20 -sf kk -in in.lj # ditto on 8 Phis</p>
<p><strong>Required hardware/software:</strong></p>
<p>Kokkos support within LAMMPS must be built with a C++11 compatible
compiler. If using gcc, version 4.8.1 or later is required.</p>
<p>To build with Kokkos support for CPUs, your compiler must support the
OpenMP interface. You should have one or more multi-core CPUs so that
multiple threads can be launched by each MPI task running on a CPU.</p>
<p>To build with Kokkos support for NVIDIA GPUs, NVIDIA Cuda software
version 6.5 or later must be installed on your system. See the
discussion for the <a class="reference internal" href="accelerate_cuda.html"><em>USER-CUDA</em></a> and
<a class="reference internal" href="accelerate_gpu.html"><em>GPU</em></a> packages for details of how to check and do
this.</p>
<div class="admonition warning">
<p class="first admonition-title">Warning</p>
<p class="last">For good performance of the KOKKOS package on GPUs,
you must have Kepler generation GPUs (or later). The Kokkos library
exploits texture cache options not supported by Telsa generation GPUs
(or older).</p>
</div>
<p>To build with Kokkos support for Intel Xeon Phi coprocessors, your
sysmte must be configured to use them in &#8220;native&#8221; mode, not &#8220;offload&#8221;
mode like the USER-INTEL package supports.</p>
<p><strong>Building LAMMPS with the KOKKOS package:</strong></p>
<p>You must choose at build time whether to build for CPUs (OpenMP),
GPUs, or Phi.</p>
<p>You can do any of these in one line, using the src/Make.py script,
described in <a class="reference internal" href="Section_start.html#start-4"><span>Section 2.4</span></a> of the manual.
Type &#8220;Make.py -h&#8221; for help. If run from the src directory, these
commands will create src/lmp_kokkos_omp, lmp_kokkos_cuda, and
lmp_kokkos_phi. Note that the OMP and PHI options use
src/MAKE/Makefile.mpi as the starting Makefile.machine. The CUDA
option uses src/MAKE/OPTIONS/Makefile.kokkos_cuda.</p>
<p>The latter two steps can be done using the &#8220;-k on&#8221;, &#8220;-pk kokkos&#8221; and
&#8220;-sf kk&#8221; <a class="reference internal" href="Section_start.html#start-7"><span>command-line switches</span></a>
respectively. Or the effect of the &#8220;-pk&#8221; or &#8220;-sf&#8221; switches can be
duplicated by adding the <a class="reference internal" href="package.html"><em>package kokkos</em></a> or <a class="reference internal" href="suffix.html"><em>suffix kk</em></a> commands respectively to your input script.</p>
<p>Or you can follow these steps:</p>
<p>CPU-only (run all-MPI or with OpenMP threading):</p>
<div class="highlight-python"><div class="highlight"><pre>cd lammps/src
make yes-kokkos
make g++ KOKKOS_DEVICES=OpenMP
</pre></div>
</div>
<p>Intel Xeon Phi:</p>
<div class="highlight-python"><div class="highlight"><pre>cd lammps/src
make yes-kokkos
make g++ KOKKOS_DEVICES=OpenMP KOKKOS_ARCH=KNC
</pre></div>
</div>
<p>CPUs and GPUs:</p>
<div class="highlight-python"><div class="highlight"><pre>cd lammps/src
make yes-kokkos
make cuda KOKKOS_DEVICES=Cuda
</pre></div>
</div>
<p>These examples set the KOKKOS-specific OMP, MIC, CUDA variables on the
make command line which requires a GNU-compatible make command. Try
&#8220;gmake&#8221; if your system&#8217;s standard make complains.</p>
<div class="admonition warning">
<p class="first admonition-title">Warning</p>
<p class="last">If you build using make line variables and re-build
LAMMPS twice with different KOKKOS options and the <em>same</em> target,
e.g. g++ in the first two examples above, then you <em>must</em> perform a
&#8220;make clean-all&#8221; or &#8220;make clean-machine&#8221; before each build. This is
to force all the KOKKOS-dependent files to be re-compiled with the new
options.</p>
</div>
<p>You can also hardwire these make variables in the specified machine
makefile, e.g. src/MAKE/Makefile.g++ in the first two examples above,
with a line like:</p>
<div class="highlight-python"><div class="highlight"><pre><span class="n">KOKKOS_ARCH</span> <span class="o">=</span> <span class="n">KNC</span>
</pre></div>
</div>
<p>Note that if you build LAMMPS multiple times in this manner, using
different KOKKOS options (defined in different machine makefiles), you
do not have to worry about doing a &#8220;clean&#8221; in between. This is
because the targets will be different.</p>
<div class="admonition warning">
<p class="first admonition-title">Warning</p>
<p class="last">The 3rd example above for a GPU, uses a different
machine makefile, in this case src/MAKE/Makefile.cuda, which is
included in the LAMMPS distribution. To build the KOKKOS package for
a GPU, this makefile must use the NVIDA &#8220;nvcc&#8221; compiler. And it must
have a KOKKOS_ARCH setting that is appropriate for your NVIDIA
hardware and installed software. Typical values for KOKKOS_ARCH are given
below, as well
as other settings that must be included in the machine makefile, if
you create your own.</p>
</div>
<div class="admonition warning">
<p class="first admonition-title">Warning</p>
<p class="last">Currently, there are no precision options with the
KOKKOS package. All compilation and computation is performed in
double precision.</p>
</div>
<p>There are other allowed options when building with the KOKKOS package.
As above, they can be set either as variables on the make command line
or in Makefile.machine. This is the full list of options, including
those discussed above, Each takes a value shown below. The
default value is listed, which is set in the
lib/kokkos/Makefile.kokkos file.</p>
<p>#Default settings specific options
#Options: force_uvm,use_ldg,rdc</p>
<ul class="simple">
<li>KOKKOS_DEVICES, values = <em>OpenMP</em>, <em>Serial</em>, <em>Pthreads</em>, <em>Cuda</em>, default = <em>OpenMP</em></li>
<li>KOKKOS_ARCH, values = <em>KNC</em>, <em>SNB</em>, <em>HSW</em>, <em>Kepler</em>, <em>Kepler30</em>, <em>Kepler32</em>, <em>Kepler35</em>, <em>Kepler37</em>, <em>Maxwell</em>, <em>Maxwell50</em>, <em>Maxwell52</em>, <em>Maxwell53</em>, <em>ARMv8</em>, <em>BGQ</em>, <em>Power7</em>, <em>Power8</em>, default = <em>none</em></li>
<li>KOKKOS_DEBUG, values = <em>yes</em>, <em>no</em>, default = <em>no</em></li>
<li>KOKKOS_USE_TPLS, values = <em>hwloc</em>, <em>librt</em>, default = <em>none</em></li>
<li>KOKKOS_CUDA_OPTIONS, values = <em>force_uvm</em>, <em>use_ldg</em>, <em>rdc</em></li>
</ul>
<p>KOKKOS_DEVICE sets the parallelization method used for Kokkos code
(within LAMMPS). KOKKOS_DEVICES=OpenMP means that OpenMP will be
used. KOKKOS_DEVICES=Pthreads means that pthreads will be used.
KOKKOS_DEVICES=Cuda means an NVIDIA GPU running CUDA will be used.</p>
<p>If KOKKOS_DEVICES=Cuda, then the lo-level Makefile in the src/MAKE
directory must use &#8220;nvcc&#8221; as its compiler, via its CC setting. For
best performance its CCFLAGS setting should use -O3 and have a
KOKKOS_ARCH setting that matches the compute capability of your NVIDIA
hardware and software installation, e.g. KOKKOS_ARCH=Kepler30. Note
the minimal required compute capability is 2.0, but this will give
signicantly reduced performance compared to Kepler generation GPUs
with compute capability 3.x. For the LINK setting, &#8220;nvcc&#8221; should not
be used; instead use g++ or another compiler suitable for linking C++
applications. Often you will want to use your MPI compiler wrapper
for this setting (i.e. mpicxx). Finally, the lo-level Makefile must
also have a &#8220;Compilation rule&#8221; for creating <a href="#id1"><span class="problematic" id="id2">*</span></a>.o files from <a href="#id3"><span class="problematic" id="id4">*</span></a>.cu files.
See src/Makefile.cuda for an example of a lo-level Makefile with all
of these settings.</p>
<p>KOKKOS_USE_TPLS=hwloc binds threads to hardware cores, so they do not
migrate during a simulation. KOKKOS_USE_TPLS=hwloc should always be
used if running with KOKKOS_DEVICES=Pthreads for pthreads. It is not
necessary for KOKKOS_DEVICES=OpenMP for OpenMP, because OpenMP
provides alternative methods via environment variables for binding
threads to hardware cores. More info on binding threads to cores is
given in <span class="xref std std-ref">this section</span>.</p>
<p>KOKKOS_ARCH=KNC enables compiler switches needed when compling for an
Intel Phi processor.</p>
<p>KOKKOS_USE_TPLS=librt enables use of a more accurate timer mechanism
on most Unix platforms. This library is not available on all
platforms.</p>
<p>KOKKOS_DEBUG is only useful when developing a Kokkos-enabled style
within LAMMPS. KOKKOS_DEBUG=yes enables printing of run-time
debugging information that can be useful. It also enables runtime
bounds checking on Kokkos data structures.</p>
<p>KOKKOS_CUDA_OPTIONS are additional options for CUDA.</p>
<p>For more information on Kokkos see the Kokkos programmers&#8217; guide here:
/lib/kokkos/doc/Kokkos_PG.pdf.</p>
<p><strong>Run with the KOKKOS package from the command line:</strong></p>
<p>The mpirun or mpiexec command sets the total number of MPI tasks used
by LAMMPS (one or multiple per compute node) and the number of MPI
tasks used per node. E.g. the mpirun command in MPICH does this via
its -np and -ppn switches. Ditto for OpenMPI via -np and -npernode.</p>
<p>When using KOKKOS built with host=OMP, you need to choose how many
OpenMP threads per MPI task will be used (via the &#8220;-k&#8221; command-line
switch discussed below). Note that the product of MPI tasks * OpenMP
threads/task should not exceed the physical number of cores (on a
node), otherwise performance will suffer.</p>
<p>When using the KOKKOS package built with device=CUDA, you must use
exactly one MPI task per physical GPU.</p>
<p>When using the KOKKOS package built with host=MIC for Intel Xeon Phi
coprocessor support you need to insure there are one or more MPI tasks
per coprocessor, and choose the number of coprocessor threads to use
per MPI task (via the &#8220;-k&#8221; command-line switch discussed below). The
product of MPI tasks * coprocessor threads/task should not exceed the
maximum number of threads the coproprocessor is designed to run,
otherwise performance will suffer. This value is 240 for current
generation Xeon Phi(TM) chips, which is 60 physical cores * 4
threads/core. Note that with the KOKKOS package you do not need to
specify how many Phi coprocessors there are per node; each
coprocessors is simply treated as running some number of MPI tasks.</p>
<p>You must use the &#8220;-k on&#8221; <a class="reference internal" href="Section_start.html#start-7"><span>command-line switch</span></a> to enable the KOKKOS package. It
takes additional arguments for hardware settings appropriate to your
system. Those arguments are <a class="reference internal" href="Section_start.html#start-7"><span>documented here</span></a>. The two most commonly used
options are:</p>
<div class="highlight-python"><div class="highlight"><pre>-k on t Nt g Ng
</pre></div>
</div>
<p>The &#8220;t Nt&#8221; option applies to host=OMP (even if device=CUDA) and
host=MIC. For host=OMP, it specifies how many OpenMP threads per MPI
task to use with a node. For host=MIC, it specifies how many Xeon Phi
threads per MPI task to use within a node. The default is Nt = 1.
Note that for host=OMP this is effectively MPI-only mode which may be
fine. But for host=MIC you will typically end up using far less than
all the 240 available threads, which could give very poor performance.</p>
<p>The &#8220;g Ng&#8221; option applies to device=CUDA. It specifies how many GPUs
per compute node to use. The default is 1, so this only needs to be
specified is you have 2 or more GPUs per compute node.</p>
<p>The &#8220;-k on&#8221; switch also issues a &#8220;package kokkos&#8221; command (with no
additional arguments) which sets various KOKKOS options to default
values, as discussed on the <a class="reference internal" href="package.html"><em>package</em></a> command doc page.</p>
<p>Use the &#8220;-sf kk&#8221; <a class="reference internal" href="Section_start.html#start-7"><span>command-line switch</span></a>,
which will automatically append &#8220;kk&#8221; to styles that support it. Use
the &#8220;-pk kokkos&#8221; <a class="reference internal" href="Section_start.html#start-7"><span>command-line switch</span></a> if
you wish to change any of the default <a class="reference internal" href="package.html"><em>package kokkos</em></a>
optionns set by the &#8220;-k on&#8221; <a class="reference internal" href="Section_start.html#start-7"><span>command-line switch</span></a>.</p>
<p>Note that the default for the <a class="reference internal" href="package.html"><em>package kokkos</em></a> command is
to use &#8220;full&#8221; neighbor lists and set the Newton flag to &#8220;off&#8221; for both
pairwise and bonded interactions. This typically gives fastest
performance. If the <a class="reference internal" href="newton.html"><em>newton</em></a> command is used in the input
script, it can override the Newton flag defaults.</p>
<p>However, when running in MPI-only mode with 1 thread per MPI task, it
will typically be faster to use &#8220;half&#8221; neighbor lists and set the
Newton flag to &#8220;on&#8221;, just as is the case for non-accelerated pair
styles. You can do this with the &#8220;-pk&#8221; <a class="reference internal" href="Section_start.html#start-7"><span>command-line switch</span></a>.</p>
<p><strong>Or run with the KOKKOS package by editing an input script:</strong></p>
<p>The discussion above for the mpirun/mpiexec command and setting
appropriate thread and GPU values for host=OMP or host=MIC or
device=CUDA are the same.</p>
<p>You must still use the &#8220;-k on&#8221; <a class="reference internal" href="Section_start.html#start-7"><span>command-line switch</span></a> to enable the KOKKOS package, and
specify its additional arguments for hardware options appopriate to
your system, as documented above.</p>
<p>Use the <a class="reference internal" href="suffix.html"><em>suffix kk</em></a> command, or you can explicitly add a
&#8220;kk&#8221; suffix to individual styles in your input script, e.g.</p>
<div class="highlight-python"><div class="highlight"><pre>pair_style lj/cut/kk 2.5
</pre></div>
</div>
<p>You only need to use the <a class="reference internal" href="package.html"><em>package kokkos</em></a> command if you
wish to change any of its option defaults, as set by the &#8220;-k on&#8221;
<a class="reference internal" href="Section_start.html#start-7"><span>command-line switch</span></a>.</p>
<p><strong>Speed-ups to expect:</strong></p>
<p>The performance of KOKKOS running in different modes is a function of
your hardware, which KOKKOS-enable styles are used, and the problem
size.</p>
<p>Generally speaking, the following rules of thumb apply:</p>
<ul class="simple">
<li>When running on CPUs only, with a single thread per MPI task,
performance of a KOKKOS style is somewhere between the standard
(un-accelerated) styles (MPI-only mode), and those provided by the
USER-OMP package. However the difference between all 3 is small (less
than 20%).</li>
<li>When running on CPUs only, with multiple threads per MPI task,
performance of a KOKKOS style is a bit slower than the USER-OMP
package.</li>
<li>When running on GPUs, KOKKOS is typically faster than the USER-CUDA
and GPU packages.</li>
<li>When running on Intel Xeon Phi, KOKKOS is not as fast as
the USER-INTEL package, which is optimized for that hardware.</li>
</ul>
<p>See the <a class="reference external" href="http://lammps.sandia.gov/bench.html">Benchmark page</a> of the
LAMMPS web site for performance of the KOKKOS package on different
hardware.</p>
<p><strong>Guidelines for best performance:</strong></p>
<p>Here are guidline for using the KOKKOS package on the different
hardware configurations listed above.</p>
<p>Many of the guidelines use the <a class="reference internal" href="package.html"><em>package kokkos</em></a> command
See its doc page for details and default settings. Experimenting with
its options can provide a speed-up for specific calculations.</p>
<p><strong>Running on a multi-core CPU:</strong></p>
<p>If N is the number of physical cores/node, then the number of MPI
tasks/node * number of threads/task should not exceed N, and should
typically equal N. Note that the default threads/task is 1, as set by
the &#8220;t&#8221; keyword of the &#8220;-k&#8221; <a class="reference internal" href="Section_start.html#start-7"><span>command-line switch</span></a>. If you do not change this, no
additional parallelism (beyond MPI) will be invoked on the host
CPU(s).</p>
<p>You can compare the performance running in different modes:</p>
<ul class="simple">
<li>run with 1 MPI task/node and N threads/task</li>
<li>run with N MPI tasks/node and 1 thread/task</li>
<li>run with settings in between these extremes</li>
</ul>
<p>Examples of mpirun commands in these modes are shown above.</p>
<p>When using KOKKOS to perform multi-threading, it is important for
performance to bind both MPI tasks to physical cores, and threads to
physical cores, so they do not migrate during a simulation.</p>
<p>If you are not certain MPI tasks are being bound (check the defaults
for your MPI installation), binding can be forced with these flags:</p>
<div class="highlight-python"><div class="highlight"><pre>OpenMPI 1.8: mpirun -np 2 -bind-to socket -map-by socket ./lmp_openmpi ...
Mvapich2 2.0: mpiexec -np 2 -bind-to socket -map-by socket ./lmp_mvapich ...
</pre></div>
</div>
<p>For binding threads with the KOKKOS OMP option, use thread affinity
environment variables to force binding. With OpenMP 3.1 (gcc 4.7 or
later, intel 12 or later) setting the environment variable
OMP_PROC_BIND=true should be sufficient. For binding threads with the
KOKKOS pthreads option, compile LAMMPS the KOKKOS HWLOC=yes option, as
discussed in <a class="reference internal" href="Section_start.html#start-3-4"><span>Section 2.3.4</span></a> of the
manual.</p>
<p><strong>Running on GPUs:</strong></p>
<p>Insure the -arch setting in the machine makefile you are using,
e.g. src/MAKE/Makefile.cuda, is correct for your GPU hardware/software
(see <a class="reference internal" href="Section_start.html#start-3-4"><span>this section</span></a> of the manual for
details).</p>
<p>The -np setting of the mpirun command should set the number of MPI
tasks/node to be equal to the # of physical GPUs on the node.</p>
<p>Use the &#8220;-k&#8221; <a class="reference internal" href="Section_start.html#start-7"><span>command-line switch</span></a> to
specify the number of GPUs per node, and the number of threads per MPI
task. As above for multi-core CPUs (and no GPU), if N is the number
of physical cores/node, then the number of MPI tasks/node * number of
threads/task should not exceed N. With one GPU (and one MPI task) it
may be faster to use less than all the available cores, by setting
threads/task to a smaller value. This is because using all the cores
on a dual-socket node will incur extra cost to copy memory from the
2nd socket to the GPU.</p>
<p>Examples of mpirun commands that follow these rules are shown above.</p>
<div class="admonition warning">
<p class="first admonition-title">Warning</p>
<p class="last">When using a GPU, you will achieve the best
performance if your input script does not use any fix or compute
styles which are not yet Kokkos-enabled. This allows data to stay on
the GPU for multiple timesteps, without being copied back to the host
CPU. Invoking a non-Kokkos fix or compute, or performing I/O for
<a class="reference internal" href="thermo_style.html"><em>thermo</em></a> or <a class="reference internal" href="dump.html"><em>dump</em></a> output will cause data
to be copied back to the CPU.</p>
</div>
<p>You cannot yet assign multiple MPI tasks to the same GPU with the
KOKKOS package. We plan to support this in the future, similar to the
GPU package in LAMMPS.</p>
<p>You cannot yet use both the host (multi-threaded) and device (GPU)
together to compute pairwise interactions with the KOKKOS package. We
hope to support this in the future, similar to the GPU package in
LAMMPS.</p>
<p><strong>Running on an Intel Phi:</strong></p>
<p>Kokkos only uses Intel Phi processors in their &#8220;native&#8221; mode, i.e.
not hosted by a CPU.</p>
<p>As illustrated above, build LAMMPS with OMP=yes (the default) and
MIC=yes. The latter insures code is correctly compiled for the Intel
Phi. The OMP setting means OpenMP will be used for parallelization on
the Phi, which is currently the best option within Kokkos. In the
future, other options may be added.</p>
<p>Current-generation Intel Phi chips have either 61 or 57 cores. One
core should be excluded for running the OS, leaving 60 or 56 cores.
Each core is hyperthreaded, so there are effectively N = 240 (4*60) or
N = 224 (4*56) cores to run on.</p>
<p>The -np setting of the mpirun command sets the number of MPI
tasks/node. The &#8220;-k on t Nt&#8221; command-line switch sets the number of
threads/task as Nt. The product of these 2 values should be N, i.e.
240 or 224. Also, the number of threads/task should be a multiple of
4 so that logical threads from more than one MPI task do not run on
the same physical core.</p>
<p>Examples of mpirun commands that follow these rules are shown above.</p>
<div class="section" id="restrictions">
<h2>Restrictions<a class="headerlink" href="#restrictions" title="Permalink to this headline"></a></h2>
<p>As noted above, if using GPUs, the number of MPI tasks per compute
node should equal to the number of GPUs per compute node. In the
future Kokkos will support assigning multiple MPI tasks to a single
GPU.</p>
<p>Currently Kokkos does not support AMD GPUs due to limits in the
available backend programming models. Specifically, Kokkos requires
extensive C++ support from the Kernel language. This is expected to
change in the future.</p>
</div>
</div>
</div>
</div>
<footer>
<hr/>
<div role="contentinfo">
<p>
&copy; Copyright .
</p>
</div>
Built with <a href="http://sphinx-doc.org/">Sphinx</a> using a <a href="https://github.com/snide/sphinx_rtd_theme">theme</a> provided by <a href="https://readthedocs.org">Read the Docs</a>.
</footer>
</div>
</div>
</section>
</div>
<script type="text/javascript">
var DOCUMENTATION_OPTIONS = {
URL_ROOT:'./',
VERSION:'15 May 2015 version',
COLLAPSE_INDEX:false,
FILE_SUFFIX:'.html',
HAS_SOURCE: true
};
</script>
<script type="text/javascript" src="_static/jquery.js"></script>
<script type="text/javascript" src="_static/underscore.js"></script>
<script type="text/javascript" src="_static/doctools.js"></script>
<script type="text/javascript" src="https://cdn.mathjax.org/mathjax/latest/MathJax.js?config=TeX-AMS-MML_HTMLorMML"></script>
<script type="text/javascript" src="_static/sphinxcontrib-images/LightBox2/lightbox2/js/jquery-1.11.0.min.js"></script>
<script type="text/javascript" src="_static/sphinxcontrib-images/LightBox2/lightbox2/js/lightbox.min.js"></script>
<script type="text/javascript" src="_static/sphinxcontrib-images/LightBox2/lightbox2-customize/jquery-noconflict.js"></script>
<script type="text/javascript" src="_static/js/theme.js"></script>
<script type="text/javascript">
jQuery(function () {
SphinxRtdTheme.StickyNav.enable();
});
</script>
</body>
</html>

View File

@ -1,338 +0,0 @@
<!DOCTYPE html>
<!--[if IE 8]><html class="no-js lt-ie9" lang="en" > <![endif]-->
<!--[if gt IE 8]><!--> <html class="no-js" lang="en" > <!--<![endif]-->
<head>
<meta charset="utf-8">
<meta name="viewport" content="width=device-width, initial-scale=1.0">
<title>5.USER-OMP package &mdash; LAMMPS 15 May 2015 version documentation</title>
<link rel="stylesheet" href="_static/css/theme.css" type="text/css" />
<link rel="stylesheet" href="_static/sphinxcontrib-images/LightBox2/lightbox2/css/lightbox.css" type="text/css" />
<link rel="top" title="LAMMPS 15 May 2015 version documentation" href="index.html"/>
<script src="_static/js/modernizr.min.js"></script>
</head>
<body class="wy-body-for-nav" role="document">
<div class="wy-grid-for-nav">
<nav data-toggle="wy-nav-shift" class="wy-nav-side">
<div class="wy-side-nav-search">
<a href="Manual.html" class="icon icon-home"> LAMMPS
</a>
<div role="search">
<form id="rtd-search-form" class="wy-form" action="search.html" method="get">
<input type="text" name="q" placeholder="Search docs" />
<input type="hidden" name="check_keywords" value="yes" />
<input type="hidden" name="area" value="default" />
</form>
</div>
</div>
<div class="wy-menu wy-menu-vertical" data-spy="affix" role="navigation" aria-label="main navigation">
<ul>
<li class="toctree-l1"><a class="reference internal" href="Section_intro.html">1. Introduction</a></li>
<li class="toctree-l1"><a class="reference internal" href="Section_start.html">2. Getting Started</a></li>
<li class="toctree-l1"><a class="reference internal" href="Section_commands.html">3. Commands</a></li>
<li class="toctree-l1"><a class="reference internal" href="Section_packages.html">4. Packages</a></li>
<li class="toctree-l1"><a class="reference internal" href="Section_accelerate.html">5. Accelerating LAMMPS performance</a></li>
<li class="toctree-l1"><a class="reference internal" href="Section_howto.html">6. How-to discussions</a></li>
<li class="toctree-l1"><a class="reference internal" href="Section_example.html">7. Example problems</a></li>
<li class="toctree-l1"><a class="reference internal" href="Section_perf.html">8. Performance &amp; scalability</a></li>
<li class="toctree-l1"><a class="reference internal" href="Section_tools.html">9. Additional tools</a></li>
<li class="toctree-l1"><a class="reference internal" href="Section_modify.html">10. Modifying &amp; extending LAMMPS</a></li>
<li class="toctree-l1"><a class="reference internal" href="Section_python.html">11. Python interface to LAMMPS</a></li>
<li class="toctree-l1"><a class="reference internal" href="Section_errors.html">12. Errors</a></li>
<li class="toctree-l1"><a class="reference internal" href="Section_history.html">13. Future and history</a></li>
</ul>
</div>
&nbsp;
</nav>
<section data-toggle="wy-nav-shift" class="wy-nav-content-wrap">
<nav class="wy-nav-top" role="navigation" aria-label="top navigation">
<i data-toggle="wy-nav-top" class="fa fa-bars"></i>
<a href="Manual.html">LAMMPS</a>
</nav>
<div class="wy-nav-content">
<div class="rst-content">
<div role="navigation" aria-label="breadcrumbs navigation">
<ul class="wy-breadcrumbs">
<li><a href="Manual.html">Docs</a> &raquo;</li>
<li>5.USER-OMP package</li>
<li class="wy-breadcrumbs-aside">
<a href="http://lammps.sandia.gov">Website</a>
<a href="Section_commands.html#comm">Commands</a>
</li>
</ul>
<hr/>
</div>
<div role="main" class="document" itemscope="itemscope" itemtype="http://schema.org/Article">
<div itemprop="articleBody">
<p><a class="reference internal" href="Section_accelerate.html"><em>Return to Section accelerate overview</em></a></p>
<div class="section" id="user-omp-package">
<h1>5.USER-OMP package<a class="headerlink" href="#user-omp-package" title="Permalink to this headline"></a></h1>
<p>The USER-OMP package was developed by Axel Kohlmeyer at Temple
University. It provides multi-threaded versions of most pair styles,
nearly all bonded styles (bond, angle, dihedral, improper), several
Kspace styles, and a few fix styles. The package currently uses the
OpenMP interface for multi-threading.</p>
<p>Here is a quick overview of how to use the USER-OMP package, assuming
one or more 16-core nodes. More details follow.</p>
<div class="highlight-python"><div class="highlight"><pre>use -fopenmp with CCFLAGS and LINKFLAGS in Makefile.machine
make yes-user-omp
make mpi # build with USER-OMP package, if settings added to Makefile.mpi
make omp # or Makefile.omp already has settings
Make.py -v -p omp -o mpi -a file mpi # or one-line build via Make.py
</pre></div>
</div>
<div class="highlight-python"><div class="highlight"><pre>lmp_mpi -sf omp -pk omp 16 &lt; in.script # 1 MPI task, 16 threads
mpirun -np 4 lmp_mpi -sf omp -pk omp 4 -in in.script # 4 MPI tasks, 4 threads/task
mpirun -np 32 -ppn 4 lmp_mpi -sf omp -pk omp 4 -in in.script # 8 nodes, 4 MPI tasks/node, 4 threads/task
</pre></div>
</div>
<p><strong>Required hardware/software:</strong></p>
<p>Your compiler must support the OpenMP interface. You should have one
or more multi-core CPUs so that multiple threads can be launched by
each MPI task running on a CPU.</p>
<p><strong>Building LAMMPS with the USER-OMP package:</strong></p>
<p>The lines above illustrate how to include/build with the USER-OMP
package in two steps, using the &#8220;make&#8221; command. Or how to do it with
one command via the src/Make.py script, described in <a class="reference internal" href="Section_start.html#start-4"><span>Section 2.4</span></a> of the manual. Type &#8220;Make.py -h&#8221; for
help.</p>
<p>Note that the CCFLAGS and LINKFLAGS settings in Makefile.machine must
include &#8220;-fopenmp&#8221;. Likewise, if you use an Intel compiler, the
CCFLAGS setting must include &#8220;-restrict&#8221;. The Make.py command will
add these automatically.</p>
<p><strong>Run with the USER-OMP package from the command line:</strong></p>
<p>The mpirun or mpiexec command sets the total number of MPI tasks used
by LAMMPS (one or multiple per compute node) and the number of MPI
tasks used per node. E.g. the mpirun command in MPICH does this via
its -np and -ppn switches. Ditto for OpenMPI via -np and -npernode.</p>
<p>You need to choose how many OpenMP threads per MPI task will be used
by the USER-OMP package. Note that the product of MPI tasks *
threads/task should not exceed the physical number of cores (on a
node), otherwise performance will suffer.</p>
<p>As in the lines above, use the &#8220;-sf omp&#8221; <a class="reference internal" href="Section_start.html#start-7"><span>command-line switch</span></a>, which will automatically append
&#8220;omp&#8221; to styles that support it. The &#8220;-sf omp&#8221; switch also issues a
default <a class="reference internal" href="package.html"><em>package omp 0</em></a> command, which will set the
number of threads per MPI task via the OMP_NUM_THREADS environment
variable.</p>
<p>You can also use the &#8220;-pk omp Nt&#8221; <a class="reference internal" href="Section_start.html#start-7"><span>command-line switch</span></a>, to explicitly set Nt = # of OpenMP
threads per MPI task to use, as well as additional options. Its
syntax is the same as the <a class="reference internal" href="package.html"><em>package omp</em></a> command whose doc
page gives details, including the default values used if it is not
specified. It also gives more details on how to set the number of
threads via the OMP_NUM_THREADS environment variable.</p>
<p><strong>Or run with the USER-OMP package by editing an input script:</strong></p>
<p>The discussion above for the mpirun/mpiexec command, MPI tasks/node,
and threads/MPI task is the same.</p>
<p>Use the <a class="reference internal" href="suffix.html"><em>suffix omp</em></a> command, or you can explicitly add an
&#8220;omp&#8221; suffix to individual styles in your input script, e.g.</p>
<div class="highlight-python"><div class="highlight"><pre>pair_style lj/cut/omp 2.5
</pre></div>
</div>
<p>You must also use the <a class="reference internal" href="package.html"><em>package omp</em></a> command to enable the
USER-OMP package. When you do this you also specify how many threads
per MPI task to use. The command doc page explains other options and
how to set the number of threads via the OMP_NUM_THREADS environment
variable.</p>
<p><strong>Speed-ups to expect:</strong></p>
<p>Depending on which styles are accelerated, you should look for a
reduction in the &#8220;Pair time&#8221;, &#8220;Bond time&#8221;, &#8220;KSpace time&#8221;, and &#8220;Loop
time&#8221; values printed at the end of a run.</p>
<p>You may see a small performance advantage (5 to 20%) when running a
USER-OMP style (in serial or parallel) with a single thread per MPI
task, versus running standard LAMMPS with its standard un-accelerated
styles (in serial or all-MPI parallelization with 1 task/core). This
is because many of the USER-OMP styles contain similar optimizations
to those used in the OPT package, described in <a class="reference internal" href="accelerate_opt.html"><em>Section accelerate 5.3.6</em></a>.</p>
<p>With multiple threads/task, the optimal choice of number of MPI
tasks/node and OpenMP threads/task can vary a lot and should always be
tested via benchmark runs for a specific simulation running on a
specific machine, paying attention to guidelines discussed in the next
sub-section.</p>
<p>A description of the multi-threading strategy used in the USER-OMP
package and some performance examples are <a class="reference external" href="http://sites.google.com/site/akohlmey/software/lammps-icms/lammps-icms-tms2011-talk.pdf?attredirects=0&amp;d=1">presented here</a></p>
<p><strong>Guidelines for best performance:</strong></p>
<p>For many problems on current generation CPUs, running the USER-OMP
package with a single thread/task is faster than running with multiple
threads/task. This is because the MPI parallelization in LAMMPS is
often more efficient than multi-threading as implemented in the
USER-OMP package. The parallel efficiency (in a threaded sense) also
varies for different USER-OMP styles.</p>
<p>Using multiple threads/task can be more effective under the following
circumstances:</p>
<ul class="simple">
<li>Individual compute nodes have a significant number of CPU cores but
the CPU itself has limited memory bandwidth, e.g. for Intel Xeon 53xx
(Clovertown) and 54xx (Harpertown) quad-core processors. Running one
MPI task per CPU core will result in significant performance
degradation, so that running with 4 or even only 2 MPI tasks per node
is faster. Running in hybrid MPI+OpenMP mode will reduce the
inter-node communication bandwidth contention in the same way, but
offers an additional speedup by utilizing the otherwise idle CPU
cores.</li>
<li>The interconnect used for MPI communication does not provide
sufficient bandwidth for a large number of MPI tasks per node. For
example, this applies to running over gigabit ethernet or on Cray XT4
or XT5 series supercomputers. As in the aforementioned case, this
effect worsens when using an increasing number of nodes.</li>
<li>The system has a spatially inhomogeneous particle density which does
not map well to the <a class="reference internal" href="processors.html"><em>domain decomposition scheme</em></a> or
<a class="reference internal" href="balance.html"><em>load-balancing</em></a> options that LAMMPS provides. This is
because multi-threading achives parallelism over the number of
particles, not via their distribution in space.</li>
<li>A machine is being used in &#8220;capability mode&#8221;, i.e. near the point
where MPI parallelism is maxed out. For example, this can happen when
using the <a class="reference internal" href="kspace_style.html"><em>PPPM solver</em></a> for long-range
electrostatics on large numbers of nodes. The scaling of the KSpace
calculation (see the <a class="reference internal" href="kspace_style.html"><em>kspace_style</em></a> command) becomes
the performance-limiting factor. Using multi-threading allows less
MPI tasks to be invoked and can speed-up the long-range solver, while
increasing overall performance by parallelizing the pairwise and
bonded calculations via OpenMP. Likewise additional speedup can be
sometimes be achived by increasing the length of the Coulombic cutoff
and thus reducing the work done by the long-range solver. Using the
<a class="reference internal" href="run_style.html"><em>run_style verlet/split</em></a> command, which is compatible
with the USER-OMP package, is an alternative way to reduce the number
of MPI tasks assigned to the KSpace calculation.</li>
</ul>
<p>Additional performance tips are as follows:</p>
<ul class="simple">
<li>The best parallel efficiency from <em>omp</em> styles is typically achieved
when there is at least one MPI task per physical CPU chip, i.e. socket
or die.</li>
<li>It is usually most efficient to restrict threading to a single
socket, i.e. use one or more MPI task per socket.</li>
<li>IMPORTANT NOTE: By default, several current MPI implementations use a
processor affinity setting that restricts each MPI task to a single
CPU core. Using multi-threading in this mode will force all threads
to share the one core and thus is likely to be counterproductive.
Instead, binding MPI tasks to a (multi-core) socket, should solve this
issue.</li>
</ul>
<div class="section" id="restrictions">
<h2>Restrictions<a class="headerlink" href="#restrictions" title="Permalink to this headline"></a></h2>
<p>None.</p>
</div>
</div>
</div>
</div>
<footer>
<hr/>
<div role="contentinfo">
<p>
&copy; Copyright .
</p>
</div>
Built with <a href="http://sphinx-doc.org/">Sphinx</a> using a <a href="https://github.com/snide/sphinx_rtd_theme">theme</a> provided by <a href="https://readthedocs.org">Read the Docs</a>.
</footer>
</div>
</div>
</section>
</div>
<script type="text/javascript">
var DOCUMENTATION_OPTIONS = {
URL_ROOT:'./',
VERSION:'15 May 2015 version',
COLLAPSE_INDEX:false,
FILE_SUFFIX:'.html',
HAS_SOURCE: true
};
</script>
<script type="text/javascript" src="_static/jquery.js"></script>
<script type="text/javascript" src="_static/underscore.js"></script>
<script type="text/javascript" src="_static/doctools.js"></script>
<script type="text/javascript" src="https://cdn.mathjax.org/mathjax/latest/MathJax.js?config=TeX-AMS-MML_HTMLorMML"></script>
<script type="text/javascript" src="_static/sphinxcontrib-images/LightBox2/lightbox2/js/jquery-1.11.0.min.js"></script>
<script type="text/javascript" src="_static/sphinxcontrib-images/LightBox2/lightbox2/js/lightbox.min.js"></script>
<script type="text/javascript" src="_static/sphinxcontrib-images/LightBox2/lightbox2-customize/jquery-noconflict.js"></script>
<script type="text/javascript" src="_static/js/theme.js"></script>
<script type="text/javascript">
jQuery(function () {
SphinxRtdTheme.StickyNav.enable();
});
</script>
</body>
</html>

View File

@ -1,239 +0,0 @@
<!DOCTYPE html>
<!--[if IE 8]><html class="no-js lt-ie9" lang="en" > <![endif]-->
<!--[if gt IE 8]><!--> <html class="no-js" lang="en" > <!--<![endif]-->
<head>
<meta charset="utf-8">
<meta name="viewport" content="width=device-width, initial-scale=1.0">
<title>5.OPT package &mdash; LAMMPS 15 May 2015 version documentation</title>
<link rel="stylesheet" href="_static/css/theme.css" type="text/css" />
<link rel="stylesheet" href="_static/sphinxcontrib-images/LightBox2/lightbox2/css/lightbox.css" type="text/css" />
<link rel="top" title="LAMMPS 15 May 2015 version documentation" href="index.html"/>
<script src="_static/js/modernizr.min.js"></script>
</head>
<body class="wy-body-for-nav" role="document">
<div class="wy-grid-for-nav">
<nav data-toggle="wy-nav-shift" class="wy-nav-side">
<div class="wy-side-nav-search">
<a href="Manual.html" class="icon icon-home"> LAMMPS
</a>
<div role="search">
<form id="rtd-search-form" class="wy-form" action="search.html" method="get">
<input type="text" name="q" placeholder="Search docs" />
<input type="hidden" name="check_keywords" value="yes" />
<input type="hidden" name="area" value="default" />
</form>
</div>
</div>
<div class="wy-menu wy-menu-vertical" data-spy="affix" role="navigation" aria-label="main navigation">
<ul>
<li class="toctree-l1"><a class="reference internal" href="Section_intro.html">1. Introduction</a></li>
<li class="toctree-l1"><a class="reference internal" href="Section_start.html">2. Getting Started</a></li>
<li class="toctree-l1"><a class="reference internal" href="Section_commands.html">3. Commands</a></li>
<li class="toctree-l1"><a class="reference internal" href="Section_packages.html">4. Packages</a></li>
<li class="toctree-l1"><a class="reference internal" href="Section_accelerate.html">5. Accelerating LAMMPS performance</a></li>
<li class="toctree-l1"><a class="reference internal" href="Section_howto.html">6. How-to discussions</a></li>
<li class="toctree-l1"><a class="reference internal" href="Section_example.html">7. Example problems</a></li>
<li class="toctree-l1"><a class="reference internal" href="Section_perf.html">8. Performance &amp; scalability</a></li>
<li class="toctree-l1"><a class="reference internal" href="Section_tools.html">9. Additional tools</a></li>
<li class="toctree-l1"><a class="reference internal" href="Section_modify.html">10. Modifying &amp; extending LAMMPS</a></li>
<li class="toctree-l1"><a class="reference internal" href="Section_python.html">11. Python interface to LAMMPS</a></li>
<li class="toctree-l1"><a class="reference internal" href="Section_errors.html">12. Errors</a></li>
<li class="toctree-l1"><a class="reference internal" href="Section_history.html">13. Future and history</a></li>
</ul>
</div>
&nbsp;
</nav>
<section data-toggle="wy-nav-shift" class="wy-nav-content-wrap">
<nav class="wy-nav-top" role="navigation" aria-label="top navigation">
<i data-toggle="wy-nav-top" class="fa fa-bars"></i>
<a href="Manual.html">LAMMPS</a>
</nav>
<div class="wy-nav-content">
<div class="rst-content">
<div role="navigation" aria-label="breadcrumbs navigation">
<ul class="wy-breadcrumbs">
<li><a href="Manual.html">Docs</a> &raquo;</li>
<li>5.OPT package</li>
<li class="wy-breadcrumbs-aside">
<a href="http://lammps.sandia.gov">Website</a>
<a href="Section_commands.html#comm">Commands</a>
</li>
</ul>
<hr/>
</div>
<div role="main" class="document" itemscope="itemscope" itemtype="http://schema.org/Article">
<div itemprop="articleBody">
<p><a class="reference internal" href="Section_accelerate.html"><em>Return to Section accelerate overview</em></a></p>
<div class="section" id="opt-package">
<h1>5.OPT package<a class="headerlink" href="#opt-package" title="Permalink to this headline"></a></h1>
<p>The OPT package was developed by James Fischer (High Performance
Technologies), David Richie, and Vincent Natoli (Stone Ridge
Technologies). It contains a handful of pair styles whose compute()
methods were rewritten in C++ templated form to reduce the overhead
due to if tests and other conditional code.</p>
<p>Here is a quick overview of how to use the OPT package. More details
follow.</p>
<div class="highlight-python"><div class="highlight"><pre>make yes-opt
make mpi # build with the OPT pacakge
Make.py -v -p opt -o mpi -a file mpi # or one-line build via Make.py
</pre></div>
</div>
<div class="highlight-python"><div class="highlight"><pre>lmp_mpi -sf opt -in in.script # run in serial
mpirun -np 4 lmp_mpi -sf opt -in in.script # run in parallel
</pre></div>
</div>
<p><strong>Required hardware/software:</strong></p>
<p>None.</p>
<p><strong>Building LAMMPS with the OPT package:</strong></p>
<p>The lines above illustrate how to build LAMMPS with the OPT package in
two steps, using the &#8220;make&#8221; command. Or how to do it with one command
via the src/Make.py script, described in <a class="reference internal" href="Section_start.html#start-4"><span>Section 2.4</span></a> of the manual. Type &#8220;Make.py -h&#8221; for
help.</p>
<p>Note that if you use an Intel compiler to build with the OPT package,
the CCFLAGS setting in your Makefile.machine must include &#8220;-restrict&#8221;.
The Make.py command will add this automatically.</p>
<p><strong>Run with the OPT package from the command line:</strong></p>
<p>As in the lines above, use the &#8220;-sf opt&#8221; <a class="reference internal" href="Section_start.html#start-7"><span>command-line switch</span></a>, which will automatically append
&#8220;opt&#8221; to styles that support it.</p>
<p><strong>Or run with the OPT package by editing an input script:</strong></p>
<p>Use the <a class="reference internal" href="suffix.html"><em>suffix opt</em></a> command, or you can explicitly add an
&#8220;opt&#8221; suffix to individual styles in your input script, e.g.</p>
<div class="highlight-python"><div class="highlight"><pre>pair_style lj/cut/opt 2.5
</pre></div>
</div>
<p><strong>Speed-ups to expect:</strong></p>
<p>You should see a reduction in the &#8220;Pair time&#8221; value printed at the end
of a run. On most machines for reasonable problem sizes, it will be a
5 to 20% savings.</p>
<p><strong>Guidelines for best performance:</strong></p>
<p>Just try out an OPT pair style to see how it performs.</p>
<div class="section" id="restrictions">
<h2>Restrictions<a class="headerlink" href="#restrictions" title="Permalink to this headline"></a></h2>
<p>None.</p>
</div>
</div>
</div>
</div>
<footer>
<hr/>
<div role="contentinfo">
<p>
&copy; Copyright .
</p>
</div>
Built with <a href="http://sphinx-doc.org/">Sphinx</a> using a <a href="https://github.com/snide/sphinx_rtd_theme">theme</a> provided by <a href="https://readthedocs.org">Read the Docs</a>.
</footer>
</div>
</div>
</section>
</div>
<script type="text/javascript">
var DOCUMENTATION_OPTIONS = {
URL_ROOT:'./',
VERSION:'15 May 2015 version',
COLLAPSE_INDEX:false,
FILE_SUFFIX:'.html',
HAS_SOURCE: true
};
</script>
<script type="text/javascript" src="_static/jquery.js"></script>
<script type="text/javascript" src="_static/underscore.js"></script>
<script type="text/javascript" src="_static/doctools.js"></script>
<script type="text/javascript" src="https://cdn.mathjax.org/mathjax/latest/MathJax.js?config=TeX-AMS-MML_HTMLorMML"></script>
<script type="text/javascript" src="_static/sphinxcontrib-images/LightBox2/lightbox2/js/jquery-1.11.0.min.js"></script>
<script type="text/javascript" src="_static/sphinxcontrib-images/LightBox2/lightbox2/js/lightbox.min.js"></script>
<script type="text/javascript" src="_static/sphinxcontrib-images/LightBox2/lightbox2-customize/jquery-noconflict.js"></script>
<script type="text/javascript" src="_static/js/theme.js"></script>
<script type="text/javascript">
jQuery(function () {
SphinxRtdTheme.StickyNav.enable();
});
</script>
</body>
</html>

View File

@ -1,264 +0,0 @@
<!DOCTYPE html>
<!--[if IE 8]><html class="no-js lt-ie9" lang="en" > <![endif]-->
<!--[if gt IE 8]><!--> <html class="no-js" lang="en" > <!--<![endif]-->
<head>
<meta charset="utf-8">
<meta name="viewport" content="width=device-width, initial-scale=1.0">
<title>angle_style charmm command &mdash; LAMMPS 15 May 2015 version documentation</title>
<link rel="stylesheet" href="_static/css/theme.css" type="text/css" />
<link rel="stylesheet" href="_static/sphinxcontrib-images/LightBox2/lightbox2/css/lightbox.css" type="text/css" />
<link rel="top" title="LAMMPS 15 May 2015 version documentation" href="index.html"/>
<script src="_static/js/modernizr.min.js"></script>
</head>
<body class="wy-body-for-nav" role="document">
<div class="wy-grid-for-nav">
<nav data-toggle="wy-nav-shift" class="wy-nav-side">
<div class="wy-side-nav-search">
<a href="Manual.html" class="icon icon-home"> LAMMPS
</a>
<div role="search">
<form id="rtd-search-form" class="wy-form" action="search.html" method="get">
<input type="text" name="q" placeholder="Search docs" />
<input type="hidden" name="check_keywords" value="yes" />
<input type="hidden" name="area" value="default" />
</form>
</div>
</div>
<div class="wy-menu wy-menu-vertical" data-spy="affix" role="navigation" aria-label="main navigation">
<ul>
<li class="toctree-l1"><a class="reference internal" href="Section_intro.html">1. Introduction</a></li>
<li class="toctree-l1"><a class="reference internal" href="Section_start.html">2. Getting Started</a></li>
<li class="toctree-l1"><a class="reference internal" href="Section_commands.html">3. Commands</a></li>
<li class="toctree-l1"><a class="reference internal" href="Section_packages.html">4. Packages</a></li>
<li class="toctree-l1"><a class="reference internal" href="Section_accelerate.html">5. Accelerating LAMMPS performance</a></li>
<li class="toctree-l1"><a class="reference internal" href="Section_howto.html">6. How-to discussions</a></li>
<li class="toctree-l1"><a class="reference internal" href="Section_example.html">7. Example problems</a></li>
<li class="toctree-l1"><a class="reference internal" href="Section_perf.html">8. Performance &amp; scalability</a></li>
<li class="toctree-l1"><a class="reference internal" href="Section_tools.html">9. Additional tools</a></li>
<li class="toctree-l1"><a class="reference internal" href="Section_modify.html">10. Modifying &amp; extending LAMMPS</a></li>
<li class="toctree-l1"><a class="reference internal" href="Section_python.html">11. Python interface to LAMMPS</a></li>
<li class="toctree-l1"><a class="reference internal" href="Section_errors.html">12. Errors</a></li>
<li class="toctree-l1"><a class="reference internal" href="Section_history.html">13. Future and history</a></li>
</ul>
</div>
&nbsp;
</nav>
<section data-toggle="wy-nav-shift" class="wy-nav-content-wrap">
<nav class="wy-nav-top" role="navigation" aria-label="top navigation">
<i data-toggle="wy-nav-top" class="fa fa-bars"></i>
<a href="Manual.html">LAMMPS</a>
</nav>
<div class="wy-nav-content">
<div class="rst-content">
<div role="navigation" aria-label="breadcrumbs navigation">
<ul class="wy-breadcrumbs">
<li><a href="Manual.html">Docs</a> &raquo;</li>
<li>angle_style charmm command</li>
<li class="wy-breadcrumbs-aside">
<a href="http://lammps.sandia.gov">Website</a>
<a href="Section_commands.html#comm">Commands</a>
</li>
</ul>
<hr/>
</div>
<div role="main" class="document" itemscope="itemscope" itemtype="http://schema.org/Article">
<div itemprop="articleBody">
<div class="section" id="angle-style-charmm-command">
<span id="index-0"></span><h1>angle_style charmm command<a class="headerlink" href="#angle-style-charmm-command" title="Permalink to this headline"></a></h1>
</div>
<div class="section" id="angle-style-charmm-kk-command">
<h1>angle_style charmm/kk command<a class="headerlink" href="#angle-style-charmm-kk-command" title="Permalink to this headline"></a></h1>
</div>
<div class="section" id="angle-style-charmm-omp-command">
<h1>angle_style charmm/omp command<a class="headerlink" href="#angle-style-charmm-omp-command" title="Permalink to this headline"></a></h1>
<div class="section" id="syntax">
<h2>Syntax<a class="headerlink" href="#syntax" title="Permalink to this headline"></a></h2>
<div class="highlight-python"><div class="highlight"><pre>angle_style charmm
</pre></div>
</div>
</div>
<div class="section" id="examples">
<h2>Examples<a class="headerlink" href="#examples" title="Permalink to this headline"></a></h2>
<div class="highlight-python"><div class="highlight"><pre>angle_style charmm
angle_coeff 1 300.0 107.0 50.0 3.0
</pre></div>
</div>
</div>
<div class="section" id="description">
<h2>Description<a class="headerlink" href="#description" title="Permalink to this headline"></a></h2>
<p>The <em>charmm</em> angle style uses the potential</p>
<img alt="_images/angle_charmm.jpg" class="align-center" src="_images/angle_charmm.jpg" />
<p>with an additional Urey_Bradley term based on the distance <em>r</em> between
the 1st and 3rd atoms in the angle. K, theta0, Kub, and Rub are
coefficients defined for each angle type.</p>
<p>See <a class="reference internal" href="special_bonds.html#mackerell"><span>(MacKerell)</span></a> for a description of the CHARMM force
field.</p>
<p>The following coefficients must be defined for each angle type via the
<a class="reference internal" href="angle_coeff.html"><em>angle_coeff</em></a> command as in the example above, or in
the data file or restart files read by the <a class="reference internal" href="read_data.html"><em>read_data</em></a>
or <a class="reference internal" href="read_restart.html"><em>read_restart</em></a> commands:</p>
<ul class="simple">
<li>K (energy/radian^2)</li>
<li>theta0 (degrees)</li>
<li>K_ub (energy/distance^2)</li>
<li>r_ub (distance)</li>
</ul>
<p>Theta0 is specified in degrees, but LAMMPS converts it to radians
internally; hence the units of K are in energy/radian^2.</p>
<hr class="docutils" />
<p>Styles with a <em>cuda</em>, <em>gpu</em>, <em>intel</em>, <em>kk</em>, <em>omp</em>, or <em>opt</em> suffix are
functionally the same as the corresponding style without the suffix.
They have been optimized to run faster, depending on your available
hardware, as discussed in <a class="reference internal" href="Section_accelerate.html"><em>Section_accelerate</em></a>
of the manual. The accelerated styles take the same arguments and
should produce the same results, except for round-off and precision
issues.</p>
<p>These accelerated styles are part of the USER-CUDA, GPU, USER-INTEL,
KOKKOS, USER-OMP and OPT packages, respectively. They are only
enabled if LAMMPS was built with those packages. See the <a class="reference internal" href="Section_start.html#start-3"><span>Making LAMMPS</span></a> section for more info.</p>
<p>You can specify the accelerated styles explicitly in your input script
by including their suffix, or you can use the <a class="reference internal" href="Section_start.html#start-7"><span>-suffix command-line switch</span></a> when you invoke LAMMPS, or you can
use the <a class="reference internal" href="suffix.html"><em>suffix</em></a> command in your input script.</p>
<p>See <a class="reference internal" href="Section_accelerate.html"><em>Section_accelerate</em></a> of the manual for
more instructions on how to use the accelerated styles effectively.</p>
</div>
<hr class="docutils" />
<div class="section" id="restrictions">
<h2>Restrictions<a class="headerlink" href="#restrictions" title="Permalink to this headline"></a></h2>
<p>This angle style can only be used if LAMMPS was built with the
MOLECULE package (which it is by default). See the <a class="reference internal" href="Section_start.html#start-3"><span>Making LAMMPS</span></a> section for more info on packages.</p>
</div>
<div class="section" id="related-commands">
<h2>Related commands<a class="headerlink" href="#related-commands" title="Permalink to this headline"></a></h2>
<p><a class="reference internal" href="angle_coeff.html"><em>angle_coeff</em></a></p>
<p><strong>Default:</strong> none</p>
<hr class="docutils" />
<p id="mackerell"><strong>(MacKerell)</strong> MacKerell, Bashford, Bellott, Dunbrack, Evanseck, Field,
Fischer, Gao, Guo, Ha, et al, J Phys Chem, 102, 3586 (1998).</p>
</div>
</div>
</div>
</div>
<footer>
<hr/>
<div role="contentinfo">
<p>
&copy; Copyright .
</p>
</div>
Built with <a href="http://sphinx-doc.org/">Sphinx</a> using a <a href="https://github.com/snide/sphinx_rtd_theme">theme</a> provided by <a href="https://readthedocs.org">Read the Docs</a>.
</footer>
</div>
</div>
</section>
</div>
<script type="text/javascript">
var DOCUMENTATION_OPTIONS = {
URL_ROOT:'./',
VERSION:'15 May 2015 version',
COLLAPSE_INDEX:false,
FILE_SUFFIX:'.html',
HAS_SOURCE: true
};
</script>
<script type="text/javascript" src="_static/jquery.js"></script>
<script type="text/javascript" src="_static/underscore.js"></script>
<script type="text/javascript" src="_static/doctools.js"></script>
<script type="text/javascript" src="https://cdn.mathjax.org/mathjax/latest/MathJax.js?config=TeX-AMS-MML_HTMLorMML"></script>
<script type="text/javascript" src="_static/sphinxcontrib-images/LightBox2/lightbox2/js/jquery-1.11.0.min.js"></script>
<script type="text/javascript" src="_static/sphinxcontrib-images/LightBox2/lightbox2/js/lightbox.min.js"></script>
<script type="text/javascript" src="_static/sphinxcontrib-images/LightBox2/lightbox2-customize/jquery-noconflict.js"></script>
<script type="text/javascript" src="_static/js/theme.js"></script>
<script type="text/javascript">
jQuery(function () {
SphinxRtdTheme.StickyNav.enable();
});
</script>
</body>
</html>

View File

@ -1,291 +0,0 @@
<!DOCTYPE html>
<!--[if IE 8]><html class="no-js lt-ie9" lang="en" > <![endif]-->
<!--[if gt IE 8]><!--> <html class="no-js" lang="en" > <!--<![endif]-->
<head>
<meta charset="utf-8">
<meta name="viewport" content="width=device-width, initial-scale=1.0">
<title>angle_style class2 command &mdash; LAMMPS 15 May 2015 version documentation</title>
<link rel="stylesheet" href="_static/css/theme.css" type="text/css" />
<link rel="stylesheet" href="_static/sphinxcontrib-images/LightBox2/lightbox2/css/lightbox.css" type="text/css" />
<link rel="top" title="LAMMPS 15 May 2015 version documentation" href="index.html"/>
<script src="_static/js/modernizr.min.js"></script>
</head>
<body class="wy-body-for-nav" role="document">
<div class="wy-grid-for-nav">
<nav data-toggle="wy-nav-shift" class="wy-nav-side">
<div class="wy-side-nav-search">
<a href="Manual.html" class="icon icon-home"> LAMMPS
</a>
<div role="search">
<form id="rtd-search-form" class="wy-form" action="search.html" method="get">
<input type="text" name="q" placeholder="Search docs" />
<input type="hidden" name="check_keywords" value="yes" />
<input type="hidden" name="area" value="default" />
</form>
</div>
</div>
<div class="wy-menu wy-menu-vertical" data-spy="affix" role="navigation" aria-label="main navigation">
<ul>
<li class="toctree-l1"><a class="reference internal" href="Section_intro.html">1. Introduction</a></li>
<li class="toctree-l1"><a class="reference internal" href="Section_start.html">2. Getting Started</a></li>
<li class="toctree-l1"><a class="reference internal" href="Section_commands.html">3. Commands</a></li>
<li class="toctree-l1"><a class="reference internal" href="Section_packages.html">4. Packages</a></li>
<li class="toctree-l1"><a class="reference internal" href="Section_accelerate.html">5. Accelerating LAMMPS performance</a></li>
<li class="toctree-l1"><a class="reference internal" href="Section_howto.html">6. How-to discussions</a></li>
<li class="toctree-l1"><a class="reference internal" href="Section_example.html">7. Example problems</a></li>
<li class="toctree-l1"><a class="reference internal" href="Section_perf.html">8. Performance &amp; scalability</a></li>
<li class="toctree-l1"><a class="reference internal" href="Section_tools.html">9. Additional tools</a></li>
<li class="toctree-l1"><a class="reference internal" href="Section_modify.html">10. Modifying &amp; extending LAMMPS</a></li>
<li class="toctree-l1"><a class="reference internal" href="Section_python.html">11. Python interface to LAMMPS</a></li>
<li class="toctree-l1"><a class="reference internal" href="Section_errors.html">12. Errors</a></li>
<li class="toctree-l1"><a class="reference internal" href="Section_history.html">13. Future and history</a></li>
</ul>
</div>
&nbsp;
</nav>
<section data-toggle="wy-nav-shift" class="wy-nav-content-wrap">
<nav class="wy-nav-top" role="navigation" aria-label="top navigation">
<i data-toggle="wy-nav-top" class="fa fa-bars"></i>
<a href="Manual.html">LAMMPS</a>
</nav>
<div class="wy-nav-content">
<div class="rst-content">
<div role="navigation" aria-label="breadcrumbs navigation">
<ul class="wy-breadcrumbs">
<li><a href="Manual.html">Docs</a> &raquo;</li>
<li>angle_style class2 command</li>
<li class="wy-breadcrumbs-aside">
<a href="http://lammps.sandia.gov">Website</a>
<a href="Section_commands.html#comm">Commands</a>
</li>
</ul>
<hr/>
</div>
<div role="main" class="document" itemscope="itemscope" itemtype="http://schema.org/Article">
<div itemprop="articleBody">
<div class="section" id="angle-style-class2-command">
<span id="index-0"></span><h1>angle_style class2 command<a class="headerlink" href="#angle-style-class2-command" title="Permalink to this headline"></a></h1>
</div>
<div class="section" id="angle-style-class2-omp-command">
<h1>angle_style class2/omp command<a class="headerlink" href="#angle-style-class2-omp-command" title="Permalink to this headline"></a></h1>
<div class="section" id="syntax">
<h2>Syntax<a class="headerlink" href="#syntax" title="Permalink to this headline"></a></h2>
<div class="highlight-python"><div class="highlight"><pre>angle_style class2
</pre></div>
</div>
</div>
<div class="section" id="examples">
<h2>Examples<a class="headerlink" href="#examples" title="Permalink to this headline"></a></h2>
<div class="highlight-python"><div class="highlight"><pre>angle_style class2
angle_coeff * 75.0
angle_coeff 1 bb 10.5872 1.0119 1.5228
angle_coeff * ba 3.6551 24.895 1.0119 1.5228
</pre></div>
</div>
</div>
<div class="section" id="description">
<h2>Description<a class="headerlink" href="#description" title="Permalink to this headline"></a></h2>
<p>The <em>class2</em> angle style uses the potential</p>
<img alt="_images/angle_class2.jpg" class="align-center" src="_images/angle_class2.jpg" />
<p>where Ea is the angle term, Ebb is a bond-bond term, and Eba is a
bond-angle term. Theta0 is the equilibrium angle and r1 and r2 are
the equilibrium bond lengths.</p>
<p>See <a class="reference internal" href="pair_modify.html#sun"><span>(Sun)</span></a> for a description of the COMPASS class2 force field.</p>
<p>Coefficients for the Ea, Ebb, and Eba formulas must be defined for
each angle type via the <a class="reference internal" href="angle_coeff.html"><em>angle_coeff</em></a> command as in
the example above, or in the data file or restart files read by the
<a class="reference internal" href="read_data.html"><em>read_data</em></a> or <a class="reference internal" href="read_restart.html"><em>read_restart</em></a>
commands.</p>
<p>These are the 4 coefficients for the Ea formula:</p>
<ul class="simple">
<li>theta0 (degrees)</li>
<li>K2 (energy/radian^2)</li>
<li>K3 (energy/radian^3)</li>
<li>K4 (energy/radian^4)</li>
</ul>
<p>Theta0 is specified in degrees, but LAMMPS converts it to radians
internally; hence the units of the various K are in per-radian.</p>
<p>For the Ebb formula, each line in a <a class="reference internal" href="angle_coeff.html"><em>angle_coeff</em></a>
command in the input script lists 4 coefficients, the first of which
is &#8220;bb&#8221; to indicate they are BondBond coefficients. In a data file,
these coefficients should be listed under a &#8220;BondBond Coeffs&#8221; heading
and you must leave out the &#8220;bb&#8221;, i.e. only list 3 coefficients after
the angle type.</p>
<ul class="simple">
<li>bb</li>
<li>M (energy/distance^2)</li>
<li>r1 (distance)</li>
<li>r2 (distance)</li>
</ul>
<p>For the Eba formula, each line in a <a class="reference internal" href="angle_coeff.html"><em>angle_coeff</em></a>
command in the input script lists 5 coefficients, the first of which
is &#8220;ba&#8221; to indicate they are BondAngle coefficients. In a data file,
these coefficients should be listed under a &#8220;BondAngle Coeffs&#8221; heading
and you must leave out the &#8220;ba&#8221;, i.e. only list 4 coefficients after
the angle type.</p>
<ul class="simple">
<li>ba</li>
<li>N1 (energy/distance^2)</li>
<li>N2 (energy/distance^2)</li>
<li>r1 (distance)</li>
<li>r2 (distance)</li>
</ul>
<p>The theta0 value in the Eba formula is not specified, since it is the
same value from the Ea formula.</p>
<hr class="docutils" />
<p>Styles with a <em>cuda</em>, <em>gpu</em>, <em>intel</em>, <em>kk</em>, <em>omp</em>, or <em>opt</em> suffix are
functionally the same as the corresponding style without the suffix.
They have been optimized to run faster, depending on your available
hardware, as discussed in <a class="reference internal" href="Section_accelerate.html"><em>Section_accelerate</em></a>
of the manual. The accelerated styles take the same arguments and
should produce the same results, except for round-off and precision
issues.</p>
<p>These accelerated styles are part of the USER-CUDA, GPU, USER-INTEL,
KOKKOS, USER-OMP and OPT packages, respectively. They are only
enabled if LAMMPS was built with those packages. See the <a class="reference internal" href="Section_start.html#start-3"><span>Making LAMMPS</span></a> section for more info.</p>
<p>You can specify the accelerated styles explicitly in your input script
by including their suffix, or you can use the <a class="reference internal" href="Section_start.html#start-7"><span>-suffix command-line switch</span></a> when you invoke LAMMPS, or you can
use the <a class="reference internal" href="suffix.html"><em>suffix</em></a> command in your input script.</p>
<p>See <a class="reference internal" href="Section_accelerate.html"><em>Section_accelerate</em></a> of the manual for
more instructions on how to use the accelerated styles effectively.</p>
</div>
<hr class="docutils" />
<div class="section" id="restrictions">
<h2>Restrictions<a class="headerlink" href="#restrictions" title="Permalink to this headline"></a></h2>
<p>This angle style can only be used if LAMMPS was built with the CLASS2
package. See the <a class="reference internal" href="Section_start.html#start-3"><span>Making LAMMPS</span></a> section
for more info on packages.</p>
</div>
<div class="section" id="related-commands">
<h2>Related commands<a class="headerlink" href="#related-commands" title="Permalink to this headline"></a></h2>
<p><a class="reference internal" href="angle_coeff.html"><em>angle_coeff</em></a></p>
<p><strong>Default:</strong> none</p>
<hr class="docutils" />
<p id="sun"><strong>(Sun)</strong> Sun, J Phys Chem B 102, 7338-7364 (1998).</p>
</div>
</div>
</div>
</div>
<footer>
<hr/>
<div role="contentinfo">
<p>
&copy; Copyright .
</p>
</div>
Built with <a href="http://sphinx-doc.org/">Sphinx</a> using a <a href="https://github.com/snide/sphinx_rtd_theme">theme</a> provided by <a href="https://readthedocs.org">Read the Docs</a>.
</footer>
</div>
</div>
</section>
</div>
<script type="text/javascript">
var DOCUMENTATION_OPTIONS = {
URL_ROOT:'./',
VERSION:'15 May 2015 version',
COLLAPSE_INDEX:false,
FILE_SUFFIX:'.html',
HAS_SOURCE: true
};
</script>
<script type="text/javascript" src="_static/jquery.js"></script>
<script type="text/javascript" src="_static/underscore.js"></script>
<script type="text/javascript" src="_static/doctools.js"></script>
<script type="text/javascript" src="https://cdn.mathjax.org/mathjax/latest/MathJax.js?config=TeX-AMS-MML_HTMLorMML"></script>
<script type="text/javascript" src="_static/sphinxcontrib-images/LightBox2/lightbox2/js/jquery-1.11.0.min.js"></script>
<script type="text/javascript" src="_static/sphinxcontrib-images/LightBox2/lightbox2/js/lightbox.min.js"></script>
<script type="text/javascript" src="_static/sphinxcontrib-images/LightBox2/lightbox2-customize/jquery-noconflict.js"></script>
<script type="text/javascript" src="_static/js/theme.js"></script>
<script type="text/javascript">
jQuery(function () {
SphinxRtdTheme.StickyNav.enable();
});
</script>
</body>
</html>

View File

@ -1,279 +0,0 @@
<!DOCTYPE html>
<!--[if IE 8]><html class="no-js lt-ie9" lang="en" > <![endif]-->
<!--[if gt IE 8]><!--> <html class="no-js" lang="en" > <!--<![endif]-->
<head>
<meta charset="utf-8">
<meta name="viewport" content="width=device-width, initial-scale=1.0">
<title>angle_coeff command &mdash; LAMMPS 15 May 2015 version documentation</title>
<link rel="stylesheet" href="_static/css/theme.css" type="text/css" />
<link rel="stylesheet" href="_static/sphinxcontrib-images/LightBox2/lightbox2/css/lightbox.css" type="text/css" />
<link rel="top" title="LAMMPS 15 May 2015 version documentation" href="index.html"/>
<script src="_static/js/modernizr.min.js"></script>
</head>
<body class="wy-body-for-nav" role="document">
<div class="wy-grid-for-nav">
<nav data-toggle="wy-nav-shift" class="wy-nav-side">
<div class="wy-side-nav-search">
<a href="Manual.html" class="icon icon-home"> LAMMPS
</a>
<div role="search">
<form id="rtd-search-form" class="wy-form" action="search.html" method="get">
<input type="text" name="q" placeholder="Search docs" />
<input type="hidden" name="check_keywords" value="yes" />
<input type="hidden" name="area" value="default" />
</form>
</div>
</div>
<div class="wy-menu wy-menu-vertical" data-spy="affix" role="navigation" aria-label="main navigation">
<ul>
<li class="toctree-l1"><a class="reference internal" href="Section_intro.html">1. Introduction</a></li>
<li class="toctree-l1"><a class="reference internal" href="Section_start.html">2. Getting Started</a></li>
<li class="toctree-l1"><a class="reference internal" href="Section_commands.html">3. Commands</a></li>
<li class="toctree-l1"><a class="reference internal" href="Section_packages.html">4. Packages</a></li>
<li class="toctree-l1"><a class="reference internal" href="Section_accelerate.html">5. Accelerating LAMMPS performance</a></li>
<li class="toctree-l1"><a class="reference internal" href="Section_howto.html">6. How-to discussions</a></li>
<li class="toctree-l1"><a class="reference internal" href="Section_example.html">7. Example problems</a></li>
<li class="toctree-l1"><a class="reference internal" href="Section_perf.html">8. Performance &amp; scalability</a></li>
<li class="toctree-l1"><a class="reference internal" href="Section_tools.html">9. Additional tools</a></li>
<li class="toctree-l1"><a class="reference internal" href="Section_modify.html">10. Modifying &amp; extending LAMMPS</a></li>
<li class="toctree-l1"><a class="reference internal" href="Section_python.html">11. Python interface to LAMMPS</a></li>
<li class="toctree-l1"><a class="reference internal" href="Section_errors.html">12. Errors</a></li>
<li class="toctree-l1"><a class="reference internal" href="Section_history.html">13. Future and history</a></li>
</ul>
</div>
&nbsp;
</nav>
<section data-toggle="wy-nav-shift" class="wy-nav-content-wrap">
<nav class="wy-nav-top" role="navigation" aria-label="top navigation">
<i data-toggle="wy-nav-top" class="fa fa-bars"></i>
<a href="Manual.html">LAMMPS</a>
</nav>
<div class="wy-nav-content">
<div class="rst-content">
<div role="navigation" aria-label="breadcrumbs navigation">
<ul class="wy-breadcrumbs">
<li><a href="Manual.html">Docs</a> &raquo;</li>
<li>angle_coeff command</li>
<li class="wy-breadcrumbs-aside">
<a href="http://lammps.sandia.gov">Website</a>
<a href="Section_commands.html#comm">Commands</a>
</li>
</ul>
<hr/>
</div>
<div role="main" class="document" itemscope="itemscope" itemtype="http://schema.org/Article">
<div itemprop="articleBody">
<div class="section" id="angle-coeff-command">
<span id="index-0"></span><h1>angle_coeff command<a class="headerlink" href="#angle-coeff-command" title="Permalink to this headline"></a></h1>
<div class="section" id="syntax">
<h2>Syntax<a class="headerlink" href="#syntax" title="Permalink to this headline"></a></h2>
<div class="highlight-python"><div class="highlight"><pre>angle_coeff N args
</pre></div>
</div>
<ul class="simple">
<li>N = angle type (see asterisk form below)</li>
<li>args = coefficients for one or more angle types</li>
</ul>
</div>
<div class="section" id="examples">
<h2>Examples<a class="headerlink" href="#examples" title="Permalink to this headline"></a></h2>
<div class="highlight-python"><div class="highlight"><pre>angle_coeff 1 300.0 107.0
angle_coeff * 5.0
angle_coeff 2*10 5.0
</pre></div>
</div>
</div>
<div class="section" id="description">
<h2>Description<a class="headerlink" href="#description" title="Permalink to this headline"></a></h2>
<p>Specify the angle force field coefficients for one or more angle types.
The number and meaning of the coefficients depends on the angle style.
Angle coefficients can also be set in the data file read by the
<a class="reference internal" href="read_data.html"><em>read_data</em></a> command or in a restart file.</p>
<p>N can be specified in one of two ways. An explicit numeric value can
be used, as in the 1st example above. Or a wild-card asterisk can be
used to set the coefficients for multiple angle types. This takes the
form &#8220;*&#8221; or &#8220;<em>n&#8221; or &#8220;n</em>&#8221; or &#8220;m*n&#8221;. If N = the number of angle types,
then an asterisk with no numeric values means all types from 1 to N. A
leading asterisk means all types from 1 to n (inclusive). A trailing
asterisk means all types from n to N (inclusive). A middle asterisk
means all types from m to n (inclusive).</p>
<p>Note that using an angle_coeff command can override a previous setting
for the same angle type. For example, these commands set the coeffs
for all angle types, then overwrite the coeffs for just angle type 2:</p>
<div class="highlight-python"><div class="highlight"><pre>angle_coeff * 200.0 107.0 1.2
angle_coeff 2 50.0 107.0
</pre></div>
</div>
<p>A line in a data file that specifies angle coefficients uses the exact
same format as the arguments of the angle_coeff command in an input
script, except that wild-card asterisks should not be used since
coefficients for all N types must be listed in the file. For example,
under the &#8220;Angle Coeffs&#8221; section of a data file, the line that
corresponds to the 1st example above would be listed as</p>
<div class="highlight-python"><div class="highlight"><pre>1 300.0 107.0
</pre></div>
</div>
<p>The <a class="reference internal" href="angle_class2.html"><em>angle_style class2</em></a> is an exception to this
rule, in that an additional argument is used in the input script to
allow specification of the cross-term coefficients. See its
doc page for details.</p>
<hr class="docutils" />
<p>Here is an alphabetic list of angle styles defined in LAMMPS. Click on
the style to display the formula it computes and coefficients
specified by the associated <a class="reference internal" href=""><em>angle_coeff</em></a> command.</p>
<p>Note that there are also additional angle styles submitted by users
which are included in the LAMMPS distribution. The list of these with
links to the individual styles are given in the angle section of <a class="reference internal" href="Section_commands.html#cmd-5"><span>this page</span></a>.</p>
<ul class="simple">
<li><a class="reference internal" href="angle_none.html"><em>angle_style none</em></a> - turn off angle interactions</li>
<li><a class="reference internal" href="angle_hybrid.html"><em>angle_style hybrid</em></a> - define multiple styles of angle interactions</li>
<li><a class="reference internal" href="angle_charmm.html"><em>angle_style charmm</em></a> - CHARMM angle</li>
<li><a class="reference internal" href="angle_class2.html"><em>angle_style class2</em></a> - COMPASS (class 2) angle</li>
<li><a class="reference internal" href="angle_cosine.html"><em>angle_style cosine</em></a> - cosine angle potential</li>
<li><a class="reference internal" href="angle_cosine_delta.html"><em>angle_style cosine/delta</em></a> - difference of cosines angle potential</li>
<li><a class="reference internal" href="angle_cosine_periodic.html"><em>angle_style cosine/periodic</em></a> - DREIDING angle</li>
<li><a class="reference internal" href="angle_cosine_squared.html"><em>angle_style cosine/squared</em></a> - cosine squared angle potential</li>
<li><a class="reference internal" href="angle_harmonic.html"><em>angle_style harmonic</em></a> - harmonic angle</li>
<li><a class="reference internal" href="angle_table.html"><em>angle_style table</em></a> - tabulated by angle</li>
</ul>
</div>
<hr class="docutils" />
<div class="section" id="restrictions">
<h2>Restrictions<a class="headerlink" href="#restrictions" title="Permalink to this headline"></a></h2>
<p>This command must come after the simulation box is defined by a
<a class="reference internal" href="read_data.html"><em>read_data</em></a>, <a class="reference internal" href="read_restart.html"><em>read_restart</em></a>, or
<a class="reference internal" href="create_box.html"><em>create_box</em></a> command.</p>
<p>An angle style must be defined before any angle coefficients are
set, either in the input script or in a data file.</p>
</div>
<div class="section" id="related-commands">
<h2>Related commands<a class="headerlink" href="#related-commands" title="Permalink to this headline"></a></h2>
<p><a class="reference internal" href="angle_style.html"><em>angle_style</em></a></p>
<p><strong>Default:</strong> none</p>
</div>
</div>
</div>
</div>
<footer>
<hr/>
<div role="contentinfo">
<p>
&copy; Copyright .
</p>
</div>
Built with <a href="http://sphinx-doc.org/">Sphinx</a> using a <a href="https://github.com/snide/sphinx_rtd_theme">theme</a> provided by <a href="https://readthedocs.org">Read the Docs</a>.
</footer>
</div>
</div>
</section>
</div>
<script type="text/javascript">
var DOCUMENTATION_OPTIONS = {
URL_ROOT:'./',
VERSION:'15 May 2015 version',
COLLAPSE_INDEX:false,
FILE_SUFFIX:'.html',
HAS_SOURCE: true
};
</script>
<script type="text/javascript" src="_static/jquery.js"></script>
<script type="text/javascript" src="_static/underscore.js"></script>
<script type="text/javascript" src="_static/doctools.js"></script>
<script type="text/javascript" src="https://cdn.mathjax.org/mathjax/latest/MathJax.js?config=TeX-AMS-MML_HTMLorMML"></script>
<script type="text/javascript" src="_static/sphinxcontrib-images/LightBox2/lightbox2/js/jquery-1.11.0.min.js"></script>
<script type="text/javascript" src="_static/sphinxcontrib-images/LightBox2/lightbox2/js/lightbox.min.js"></script>
<script type="text/javascript" src="_static/sphinxcontrib-images/LightBox2/lightbox2-customize/jquery-noconflict.js"></script>
<script type="text/javascript" src="_static/js/theme.js"></script>
<script type="text/javascript">
jQuery(function () {
SphinxRtdTheme.StickyNav.enable();
});
</script>
</body>
</html>

Some files were not shown because too many files have changed in this diff Show More