A tool for Boost. It creates a table of all test results organized by test features
Purpose
Any time one considers using a library as large and complex as the Boost libraries, he must have a way of validating the the library functions in his environment. This should be done when the library is installed and anytime questions are raised regarding its applicabililty and/or its usage. The procedures described here permit a user to run any combination of tests on any or all libraries and generate a set of convenient tables which show which libraries pass which tests under what conditions.
Preliminaries
Generating these tables requires a couple of utility programs: process_jam_log and library_status. These can be built by moving to the directory tools/regression/build and invoking bjam. If all goes well these utility programs will be found in the directory dist/bin. From there they should be moved to a place in the current path. Running Tests for One Library
- Start from your command line environment.
- set the current directory to:/libs//test
- Invoke one of the following: ../../../tools/regression/src/library_test (*nix). ......\tools\regression\src\library_test (windows).
- This will display short help message describing the how to set the command line arguments for the compilers and variants you want to appear in the final table.
- Setting these arguments requires rudimentary knowledge of bjam usage. Hopefully, if you've arrived at this page you've gained the required knowledge during the installation and library build process.
- Rerun the above command with the argument set accordingly.
- When the command terminates, there should be a file named library_status.html
- Display this file with any web browser.
This table was generated by invoking the following command line from within the .../libs/serialization/test directory:
../../../tools/regression/src/library_test --toolset=clang-darwin-03,clang-darwin-11,darwin4.9 address-model=64,32 link=static,shared variant=debug,release
It shows all the combinations for test configurations run. As more tests are run the table gets longer. As more configurations are run, the table gets wider. The cells marked "Missing" correspond to tests that were not run for some reason or another. This is usually because the corresponding Jamfile.v2 excludes this test for the given combination of compiler and build attributes.
Tables are cumulative. That is, if you run one set of tests now and tests with different attributes later, the table will contain all the results to date. The test results are stored in ../bin.v2/libs/test//*. To reinitialize the test results to empty, delete the corresponding files in this directory.
The procedure above assumes that the table are generated within the directory ../libs//test. This is the most common case since this directory contains the Jamfile.v2 as well as the source code that is used by official boost testers. However, this is just a convention. The table can be generated for other directories within the library. One possibility would be to generate the table for all the examples in ../libs//example. Or one might have a special directory of performance tests which take a long time to run and hence are not suitable for running by official boost testers. Just remember that library status table is generated in the directory from which the library_test command is invoked.
Running Tests for All Libraries
For those with *nix or cygwin command line shells, there is shell script that can be run from the boost root directory: tools/regression/src/library_test_all
The command line arguments are the same as for running the test for one library. This script creates all the html files in all the test directories as well as an html page in the status directory named library_status_summary.html. This can be used to browse through all test results for all test in all libraries.
Copyright 2011 Bryce Lelbach.
Copyright 2007-2015 Robert Ramey.
Distributed under the Boost Software License, Version 1.0. (See accompanying file LICENSE_1_0.txt or http://www.boost.org/LICENSE_1_0.txt)