Development Tasks / Testing
Automating software testing is an important software development task because it helps ensure that software performs as intended. It is particular important to ensure that a change in one part of the code does not break another part of the code, and that changes in the computer environment don't break the software.
This documentation contains the following sections:
- Introduction
- Test StateMod by Running Reference Datasets
- Automated Functional Testing Using TSTool
- Test Code Using Unit Tests
Introduction
Automated testing of StateMod software has not been formalized into a software framework in the past.
This documentation describes the current state of testing.
The cdss-app-statemod-fortran-test
repository contains StateMod tests.
Testing generally involves two approaches:
- Create small stand-alone tests that test specific software features, such as a operating rule or model option, to verify specific computational code. A number of these tests are included in the StateMod test repository and can be enhanced similar to TSTool software tests.
- Run full datasets using different StateMod executable versions and compare the results. This helps identify the impacts of a change in one version compared to a baseline version. The baseline version is typically the latest published release that it has been sufficiently tested and verified. This approach is the focus of current automated testing.
The following sections provide more details about testing.
Test StateMod by Running Reference Datasets
The current approach to testing StateMod is to run entire datasets using different StateMod versions
and compare the results to ensure that differences are small or are otherwise explainable.
This approach is described in the
cdss-app-statemod-fortran-test
README
.
Automated Functional Testing Using TSTool
TSTool software can be used to implement automated tests of specific functionality. These smaller tests are self-contained and provide confirmation that specific software features are working. Running many such tests on controlled test data in a test suite provides confirmation that all software features are working as intended.
Running TSTool for Testing
The cdss-app-statemod-fortran-test
described in the previous section includes tests for full datasets.
The scripts/statemod-test.bash
script that is used to test datasets can also be used
to run TSTool for functional tests.
cd scripts
./statemod-test.bash
The interactive menu provides a runtstool
menu item to run TSTool,
which prompts for the StateMod executable to use for testing.
A StateMod executable can be selected from the dataset tests or
executables in the StateMod development folder.
The executable path is set as the StateModExecutable
TSTool property,
which can be used with all command files.
This allows all the tests to be run for any executable.
The following documentation explains how to define and run functional tests. TSTool can be used with existing tests and it is recommended to trying running tests as this documentation is read.
Legacy Tests
The StateMod development effort that has occurred over many years has resulted in many small examples that have been used for testing and as examples for documentation. However, these tests are not automated, and many tests have not been maintained.
The legacy examples are being leveraged to form the basis of new functional tests.
Examples will be migrated to this repository as resources allow.
The legacy Ex1
example is used for illustration below and other tests are similar.
Define a Test from a Legacy Test
The legacy tests can be used to define a new test.
For example, the Ex1
example was converted to ex0001
test as follows.
Note that placeholder folders have been created in the repository tests but the folders have not been populated.
- General:
- Note that legacy example files often have inconsistencies in filenames such as upper and lower case.
The following
ex0001
example generally uses lowercase for filenames. Inconsistencies in legacy examples should be cleaned up as the files are moved into this testing repository. - Be aware of using different filename patterns and impacts on
.gitignore
. It is desirable to avoid accidentally committing dynamic files (especially large files) to the repository, and.gitignore
is used to constrain files that are committed. The repository's main.gitignore
file may need to be updated to ensure that dynamic files are ignored. Be aware of files that are staged to commit to the repository and do not just blindly commit files without reviewing their names. - This process is generally consistent with the TSTool Quality Control documentation.
- Note that legacy example files often have inconsistencies in filenames such as upper and lower case.
The following
- Create a repository folder for the test's files:
- For example:
test/examples/ex0001/
- The folder name starts with
ex
to indicate an example. - Use a folder name with 4 digits padded with zeros to allow sorting.
- If a legacy test uses a different name (for example
TEST
instead ofEX
), a similar approach can be taken. - Do not copy all the legacy files into the new folder. There is typically a need for review and cleanup so only copy the minimum number of files to implement the test, as discussed below.
- For example:
- Create a
README.md
file:- This file will be readable on the repository web page.
- The legacy example typically includes a README file that can be incorporated.
- For example, see:
ex0001/README.md
- Also update the list of tests to track which legacy tests have been implemented in this repository.
- Copy the response file:
- For example, copy
EX1.RSP
toex0001.rsp
. - If necessary, run the
SmNewRsp
utility to convert the response file to new format. It is OK to keep a copy of the old file such asex0001.rsp.old
to ensure that comparisons can occur. TheSmNewRsp
program can be compiled in the development files. The current response file format uses data file descriptors such asControl
whereas the older format required files to be listed in a specific order and is prone to errors. - If necessary, edit the response file and perform additional cleanup,
such as inserting an appropriate comment at the top rather than the default
SmNewRsp
header and removing whitespace at the end of lines that make it difficult to view the file.
- For example, copy
- Create a TSTool command file to run the test:
- Run TSTool from the
scripts/statemod-test.bash
script and pick a StateMod executable. - Can copy and modify a TSTool command file from another test as a starting point.
- Name the file consistent with the example, for example see:
test-ex0001.tstool
- Modify and run the TSTool command file repeatedly until the test is complete. The following steps describe additional steps to complete the test.
- Run TSTool from the
- Implement tests in the TSTool command file for standard StateMod run modes:
- StateMod's main run modes include
-check
,-base
,-sim
, and-report
. - For example, see the
GO*.BAT
batch files that are often included in legacy tests. For example,GOC.BAT
typically runs the-check
mode on a dataset. It is recommended to not copy these batch files into the testing repository because they do not provide the level of functionality needed to do full automated testing, may have errors, and often reference folders and files that are not portable. - Use the TSTool
RunProgram
command as in the abovetest-ex0001.tstool
example to run the StateMod executable that was configured for TSTool. Use full paths to the executable (${StateModExecutable}
) and response file (using${WorkingDir}/ex0001.rsp
) so that there is no confusion about their location. - Implement commands to save results to the
expected-results/
folder, which allows the tests to be run in the future and compare to expected results. It is recommended to create folders for each run mode's expected results, as in theex0001
example in order to keep files organized. For example useexpected-results/base
to save base flow expected results,expected-results/sim
to save simulation expected results, andexpected-results/report
to save report expected results. Don't forget this because it allows other people to run the tests without recreating the expected results. TSTool commands are typically included in a test command file to save/copy expected results. Once verified, the commands are commented out until expected results need to be updated. It may also be appropriate to copy the legacy results to the currentexpected-results
folder. For example, save the version 15.00.01 binary output files from simulation, which should in many cases be correct for basic model functionality. This provides a cross-check. - Based on the run mode, implement appropriate results comparisons:
- For
-check
, use TSTool'sCheckFile
command to check for file patterns indicating a problem. - For
-base
, use TSTool'sCompareFiles
command to compare the output baseflow time series files. - For
-sim
, use TSTool'sCompareTimeSeries
command to compare binary file output time series using an appropriate tolerance. - For
-report
, use TSTool'sCompareFiles
command to compare text files. It may be necessary to implement multiple report comparisons depending on the complexity of the test.
- For
- StateMod's main run modes include
- Run the TSTool command file repeatedly and resolve any issues.
For example, start by copying all the input files referenced in the response file but if any are forgotten,
it should be obvious.
Refer to the startup log file (
statem.log
) and response file for the test (e.g.,ex0001.log
) to help resolve issues.
Define a New Test
The easiest way to define a new test is often to copy a similar test's files and modify as appropriate. See the previous section for an explanation of test details. Care should be taken to not clutter up test folders with unused files, which confuse other software testers.
Creating a Test Suite
It is typical to create a test suite to run all the tests. Otherwise, running each test individually requires a lot of time and is prone to overlooking issues. Running the tests in a suite streamlines testing and should be done as part of the software release checklist.
To create a test suite that reflects current tests,
run TSTool via the statemod-test.bash
script and run the
test/examples-suite/create/create-examples-suite.tstool
command file.
This will create the command file discussed in the next section.
Running a Test Suite
To run the test suite created in the previous step, use TSTool to run the
test/examples-suite/run/run-examples-suite.tstool
command file.
The command status for each command and the output files for the suites contain information about
whether tests passed or failed.
The following is an example of the text file resport:
# File generated by...
# program: TSTool 14.0.0.dev3 (2021-09-09)
# user: sam
# date: Mon Sep 13 00:22:10 America/Denver 2021
# host: RoaringFork
# directory: C:\Users\sam\cdss-dev\StateMod\git-repos\cdss-app-statemod-fortran-test\test\examples-suite\run
# command line: TSTool
# --home C:/CDSS/TSTool-14.0.0.dev3
# # C:/Users/sam/cdss-dev/StateMod/git-repos/cdss-app-statemod-fortran-test/test/examples/0-readme/readme.tstool
# # StateModExecutable==C:/Users/sam/cdss-dev/StateMod/git-repos/cdss-app-statemod-fortran/src/main/fortran/statemod-17.0.3-gfortran-win-64bit.exe
#
# Command file regression test report from StartRegressionTestResultsReport() and RunCommands()
#
# Explanation of columns:
#
# Num: count of the tests
# Enabled: TRUE if test enabled or FALSE if "#@enabled false" in command file
# Run Time: run time in milliseconds - use table output to see this column
# Test Pass/Fail:
# The test status below may be PASS or FAIL (or blank if disabled).
# A test will pass if the command file actual status matches the expected status.
# Disabled tests are not run and do not count as PASS or FAIL.
# Search for *FAIL* to find failed tests.
# Commands Expected Status:
# Default is assumed to be SUCCESS.
# "#@expectedStatus Warning|Failure" comment in command file overrides default.
# Commands Actual Status:
# The most severe status (Success|Warning|Failure) for each command file.
#
# | |Test |Commands |Commands |
# | |Pass/ |Expected |Actual |
# Num|Enabled|Fail |Status |Status |Command File
#----+-------+------+----------+-----------+---------------------------------------------------------------------------------------------
1|TRUE |*FAIL*|SUCCESS |WARNING |C:\Users\sam\cdss-dev\StateMod\git-repos\cdss-app-statemod-fortran-test\test\examples\ex0001\test-ex0001.tstool
2|TRUE |*FAIL*|SUCCESS |WARNING |C:\Users\sam\cdss-dev\StateMod\git-repos\cdss-app-statemod-fortran-test\test\examples\ex0002\test-ex0002.tstool
#----+-------+------+----------+-----------+---------------------------------------------------------------------------------------------
FAIL count = 2, 100.000%
PASS count = 0, 0.000%
Disabled count = 0, 0.000%
#--------------------------------
Total = 2
Example Test Suite Results Text File
The following illustrates the Excel file report:
Example Test Suite Results Excel File (see also the full-size image)
Additional Testing Considerations
Automated tests are prone to breaking for various reasons, including bugs in the software, software changes that impact results and output formats, and other unforeseen issues. Automated tests must be maintained and updated over time. The following considerations may impact tests:
- The legacy example names are not descriptive. In the future it may be desirable to name folders (and perhaps) test files for specific operating rules or other features.
- Expected results are saved from a previous software version and may need to be updated periodically. TSTool command files typically contain commands to save expected results and these commands can be uncommented and commented as necessary to update expected results.
- If it is desired that tests work with old and new software, it may be necessary to run an old repository commit for tests, for example using a repository tag to check out an old version.
- Multiple expected results can be saved,
for example for each major software version, and only the appropriate test run.
This can use TSTool's
@require
feature, although more enhancements may be needed. - Verification of expected results may require additional work,
such a creating an Excel file to duplicate calculations, use TSTool, etc.
Artifacts from such work can be saved in the test's folder
(perhaps the
expected-results
folder) to review as needed.
Test Code Using Unit Tests
Unit tests are compiled code the test an atomic unit of code. Unit tests are a tool for software developers that require access to the develop environment, unlike functional tests, that could be run by users without the full developer environment.
There does not seem to be extensive unit test frameworks for Fortran. See:
Unit tests could be implemented in the future, especially if StateCU is ported to another language that provides a unit test framework.