turned on verbose to check tolerances and also print that info out to the run.log file
This commit is contained in:
9
.github/workflows/kokkos-regression.yaml
vendored
9
.github/workflows/kokkos-regression.yaml
vendored
@ -75,15 +75,12 @@ jobs:
|
|||||||
--lmp-bin=build/lmp \
|
--lmp-bin=build/lmp \
|
||||||
--config-file=tools/regression-tests/config_kokkos_openmp.yaml \
|
--config-file=tools/regression-tests/config_kokkos_openmp.yaml \
|
||||||
--example-folders="examples/colloid;examples/melt;examples/micelle;examples/threebody" \
|
--example-folders="examples/colloid;examples/melt;examples/micelle;examples/threebody" \
|
||||||
--output-file=output.xml \
|
--output-file=output_kokkos.xml --progress-file=progress_kokkos.yaml --log-file=run_kokkos.log \
|
||||||
--progress-file=progress.yaml \
|
--verbose
|
||||||
--log-file=run.log
|
|
||||||
|
|
||||||
tar -cvf kokkos-regression-test.tar run.log progress.yaml output.xml
|
|
||||||
|
|
||||||
- name: Upload artifacts
|
- name: Upload artifacts
|
||||||
uses: actions/upload-artifact@v4
|
uses: actions/upload-artifact@v4
|
||||||
with:
|
with:
|
||||||
name: kokkos-regression-test-artifact
|
name: kokkos-regression-test-artifact
|
||||||
path: kokkos-regression-test.tar
|
path: *_kokkos.*
|
||||||
|
|
||||||
|
|||||||
@ -659,6 +659,9 @@ def iterate(lmp_binary, input_folder, input_list, config, results, progress_file
|
|||||||
msg = f" {num_abs_failed} abs diff checks failed."
|
msg = f" {num_abs_failed} abs diff checks failed."
|
||||||
print(msg)
|
print(msg)
|
||||||
logger.info(msg)
|
logger.info(msg)
|
||||||
|
for out in failed_abs_output:
|
||||||
|
logger.info(f" - {out}")
|
||||||
|
|
||||||
if verbose == True:
|
if verbose == True:
|
||||||
for out in failed_abs_output:
|
for out in failed_abs_output:
|
||||||
print(f" - {out}")
|
print(f" - {out}")
|
||||||
@ -667,6 +670,9 @@ def iterate(lmp_binary, input_folder, input_list, config, results, progress_file
|
|||||||
msg = f" {num_rel_failed} rel diff checks failed."
|
msg = f" {num_rel_failed} rel diff checks failed."
|
||||||
print(msg)
|
print(msg)
|
||||||
logger.info(msg)
|
logger.info(msg)
|
||||||
|
for out in failed_rel_output:
|
||||||
|
logger.info(f" - {out}")
|
||||||
|
|
||||||
if verbose == True:
|
if verbose == True:
|
||||||
for out in failed_rel_output:
|
for out in failed_rel_output:
|
||||||
print(f" - {out}")
|
print(f" - {out}")
|
||||||
@ -675,11 +681,15 @@ def iterate(lmp_binary, input_folder, input_list, config, results, progress_file
|
|||||||
msg = f" all {num_checks} checks passed."
|
msg = f" all {num_checks} checks passed."
|
||||||
print(msg)
|
print(msg)
|
||||||
logger.info(msg)
|
logger.info(msg)
|
||||||
|
|
||||||
|
result.status = f" 'status': 'passed', 'abs_diff_failed': '{num_abs_failed}', 'rel_diff_failed': '{num_rel_failed}' "
|
||||||
|
|
||||||
num_passed = num_passed + 1
|
num_passed = num_passed + 1
|
||||||
else:
|
else:
|
||||||
|
result.status = f" 'status': 'failed', 'abs_diff_failed': '{num_abs_failed}', 'rel_diff_failed': '{num_rel_failed}' "
|
||||||
num_error = num_error + 1
|
num_error = num_error + 1
|
||||||
|
|
||||||
result.status = f" 'abs_diff_failed': '{num_abs_failed}', 'rel_diff_failed': '{num_rel_failed}' "
|
|
||||||
results.append(result)
|
results.append(result)
|
||||||
|
|
||||||
# check if memleak detects from valgrind run (need to replace "mpirun" -> valgrind --leak-check=yes mpirun")
|
# check if memleak detects from valgrind run (need to replace "mpirun" -> valgrind --leak-check=yes mpirun")
|
||||||
@ -1556,12 +1566,10 @@ if __name__ == "__main__":
|
|||||||
for result in all_results:
|
for result in all_results:
|
||||||
#print(f"{result.name}: {result.status}")
|
#print(f"{result.name}: {result.status}")
|
||||||
case = TestCase(name=result.name, classname=result.name)
|
case = TestCase(name=result.name, classname=result.name)
|
||||||
if result.status == "failed":
|
if "passed" not in result.status:
|
||||||
case.add_failure_info(message="Actual values did not match expected ones.")
|
case.add_failure_info(message=result.status)
|
||||||
if result.status == "skipped":
|
if "skipped" in result.status:
|
||||||
case.add_skipped_info(message="Test was skipped.")
|
case.add_skipped_info(message="Test was skipped.")
|
||||||
if result.status == "error":
|
|
||||||
case.add_skipped_info(message="Test run had errors.")
|
|
||||||
test_cases.append(case)
|
test_cases.append(case)
|
||||||
|
|
||||||
current_timestamp = datetime.datetime.now()
|
current_timestamp = datetime.datetime.now()
|
||||||
|
|||||||
Reference in New Issue
Block a user