Counted the total num checks per test
This commit is contained in:
@ -360,6 +360,7 @@ def iterate(input_list, config, results, removeAnnotatedInput=False):
|
|||||||
num_rel_failed = 0
|
num_rel_failed = 0
|
||||||
failed_abs_output = []
|
failed_abs_output = []
|
||||||
failed_rel_output = []
|
failed_rel_output = []
|
||||||
|
num_checks = 0
|
||||||
for i in range(num_fields):
|
for i in range(num_fields):
|
||||||
quantity = thermo[0]['keywords'][i]
|
quantity = thermo[0]['keywords'][i]
|
||||||
|
|
||||||
@ -378,6 +379,7 @@ def iterate(input_list, config, results, removeAnnotatedInput=False):
|
|||||||
if quantity in config['tolerance']:
|
if quantity in config['tolerance']:
|
||||||
abs_tol = float(config['tolerance'][quantity]['abs'])
|
abs_tol = float(config['tolerance'][quantity]['abs'])
|
||||||
rel_tol = float(config['tolerance'][quantity]['rel'])
|
rel_tol = float(config['tolerance'][quantity]['rel'])
|
||||||
|
num_checks = num_checks + 2
|
||||||
if abs_diff > abs_tol:
|
if abs_diff > abs_tol:
|
||||||
abs_diff_check = "FAILED"
|
abs_diff_check = "FAILED"
|
||||||
reason = f"{quantity}: actual ({abs_diff:0.2e}) > expected ({abs_tol:0.2e})"
|
reason = f"{quantity}: actual ({abs_diff:0.2e}) > expected ({abs_tol:0.2e})"
|
||||||
@ -388,7 +390,6 @@ def iterate(input_list, config, results, removeAnnotatedInput=False):
|
|||||||
reason = f"{quantity}: actual ({rel_diff:0.2e}) > expected ({rel_tol:0.2e})"
|
reason = f"{quantity}: actual ({rel_diff:0.2e}) > expected ({rel_tol:0.2e})"
|
||||||
failed_rel_output.append(f"{reason}")
|
failed_rel_output.append(f"{reason}")
|
||||||
num_rel_failed = num_rel_failed + 1
|
num_rel_failed = num_rel_failed + 1
|
||||||
|
|
||||||
else:
|
else:
|
||||||
abs_diff_check = "N/A"
|
abs_diff_check = "N/A"
|
||||||
rel_diff_check = "N/A"
|
rel_diff_check = "N/A"
|
||||||
@ -407,7 +408,7 @@ def iterate(input_list, config, results, removeAnnotatedInput=False):
|
|||||||
for i in failed_rel_output:
|
for i in failed_rel_output:
|
||||||
print(f"- {i}")
|
print(f"- {i}")
|
||||||
if num_abs_failed == 0 and num_rel_failed == 0:
|
if num_abs_failed == 0 and num_rel_failed == 0:
|
||||||
print("All checks passed.")
|
print(f"All {num_checks} checks passed.")
|
||||||
result.status = "passed"
|
result.status = "passed"
|
||||||
if verbose == True:
|
if verbose == True:
|
||||||
print(" N/A means that tolerances are not defined in the config file.")
|
print(" N/A means that tolerances are not defined in the config file.")
|
||||||
@ -581,6 +582,7 @@ if __name__ == "__main__":
|
|||||||
|
|
||||||
print("Summary:")
|
print("Summary:")
|
||||||
print(f" - {passed_tests} passed / {total_tests} tests")
|
print(f" - {passed_tests} passed / {total_tests} tests")
|
||||||
|
print(f" - Details are given in {output_file}.")
|
||||||
|
|
||||||
# generate a JUnit XML file
|
# generate a JUnit XML file
|
||||||
with open(output_file, 'w') as f:
|
with open(output_file, 'w') as f:
|
||||||
@ -589,11 +591,11 @@ if __name__ == "__main__":
|
|||||||
#print(f"{result.name}: {result.status}")
|
#print(f"{result.name}: {result.status}")
|
||||||
case = TestCase(name=result.name, classname=result.name)
|
case = TestCase(name=result.name, classname=result.name)
|
||||||
if result.status == "failed":
|
if result.status == "failed":
|
||||||
case.add_failure_info('failure')
|
case.add_failure_info(message="Expected value did not match.")
|
||||||
if result.status == "skipped":
|
if result.status == "skipped":
|
||||||
case.add_skipped_info('skipped')
|
case.add_skipped_info(message="Test was skipped.")
|
||||||
if result.status == "error":
|
if result.status == "error":
|
||||||
case.add_skipped_info('error')
|
case.add_skipped_info(message="Test run had errors.")
|
||||||
test_cases.append(case)
|
test_cases.append(case)
|
||||||
|
|
||||||
current_timestamp = datetime.datetime.now()
|
current_timestamp = datetime.datetime.now()
|
||||||
|
|||||||
Reference in New Issue
Block a user