Documentation for 3.20
Automatic Failure Interpretation in TextTest
There are situations when it saves a fair bit of time to
inform TextTest about what your system is likely to log and what
it means. This means that TextTest can find such patterns
automatically and instead of just saying “differences in
stdout” can provide you a much more succinct description
of what went wrong. TextTest can also automatically try to
collect information about what happened from the system: for
example core files that have been dumped on UNIX. This document
aims to give a guide to how TextTest can interpret failures at a
higher level than file differences.
This will just happen. Each test is run with a unique
temporary directory as the current working directory. This means
that TextTest will pick up any core file written there and try
to extract the stack trace from it. If it succeeds, it will
report the test as having “CRASHED” and display the
stack trace in the “Text Info” window.
Since version 3.7, it does this by using an external script
“interpretcore.py” which outputs stack traces given
core files. It has been tested on quite a few flavours of UNIX :
HPUX-11, powerpc, Sparc Solaris and Linux RHEL3, RHEL4 and seems
to be quite portable (much more so than the old code that was
part of TextTest until version 3.6) This script is plugged in by
default via the default value of the “collate_script”entry on UNIX.
It is provided as standard with TextTest, but can clearly
also be used externally to TextTest. It works by using the
standard debugger 'gdb'. If it fails to find the stack trace for
any reason the test will still be reported as “CRASHED”
but the reason for failure will be given in the “Text
Info” window instead.
It can be useful to associate a succinct problem description
with the appearance of certain logged text. For example, your
application may well report “Internal Error!!!” when
something bad happens. In this case you could tell TextTest
about this so it can describe such a test failure as an Internal
Error and save you the trouble of reading the log file.
Another common usage, as the title hints, is for known bugs
in the system under test. This may be needed because the bug
appears only intermittently, or because time has not yet been
found to fix it even though its presence is known. In either
case, telling TextTest about it can prevent lots of people
examining lots of failed tests in order to discover the same
issue again.
The easiest way to do this is by using the dialog in the
static GUI. Right-click the test or testsuite where the bug should
apply (for example, if it applies everywhere, select the root
test suite) and select "Enter Failure Information". We can then fill in the
dialog as shown :
Here we imagine that when our application logs “after
<n> seconds” this implies it has gone into a delay
that isn't appropriate. So we fill in what text it should look for, which file
it should look in, and then in the bottom section provide some description information,
a full text to give and a short description. This creates a special
file knownbugs.<app> in the test or suite's
directory, which has a format that is easy to edit in a normal editor.
If we then run the test and it produces the indicated text,
we then get a nice summary as well as the usual complete file
differences. Note it has used our “brief description”
given above in the Details column of the test tree view, while
the full description appears in the Text Info window at the
bottom right. (Here I've run two copies of the text to show what tests failing with
knownbugs look like in the test tree when not selected)
Ordinarily, you will search for some text that will be produced in a certain file, as
given by the "file to search in" entry. This will search the unfiltered version of the file
(not the diff produced in the Text Info window). This can be inverted via the check box in the top
section, so that the bug is triggered when the text is not present.
In addition to searching in specific files, you can also select "brief text/details" or "full difference
report" for the matching. This will look in (respectively) the text that appears in the Details column
or the text in the bottom right window, i.e. the actual difference report.
Note however that it isn't generally possible to match on the first line of this text,
which is an additional explanatory line added by the GUI and not part of the state of the
test. This line will be different in the case of
running in batch mode, which is why it isn't a good idea for TextTest
to use it for matching.
There is now also a control to trigger reruns when known bugs are found, which is useful for intermittent problems. In that case the test will simply be restarted, up to a maximum of the number of times you give. When running with a grid engine, the tests will not
necessarily be rerun on the same machine so it is also a way to circumvent temporary hardware issues in that case.
When the tests are run, TextTest wll then find all such "knownbugs" files, using its
mechanism for finding
and prioritising files in the hierarchy. All information from all such files will be used,
the files do not overwrite each other as versioned files used to up to version 3.10.
If you have a bug-tracking system with an API of some sort,
you can probably get it to talk to TextTest without
very much effort. Instead of providing textual descriptions you
can then just provide the bug ID from the bug system and
TextTest will extract the (current) information for you. It will
try to determine whether the bug has been fixed (closed in the
bugsystem) and if so it will be reported as “internal
error” rather than as “known bug”, as it is
expected the bug text would not continue to occur if the bug had
been closed.
To set this up, you need to give TextTest the URL to your bug system,
using the config file entry "bug_system_location". The key should be the name of
the bug system, currently one of "bugzilla", "bugzillav2" or "jira". For example, if you use
bugzilla version 3 and later you might add this in your config file:
[bug_system_location]
bugzilla:http://www.mysite.com/bugzilla
[end]
If you use Bugzilla to
track your bugs, there are two plugins already written and bundled
with TextTest. For bugzilla version 3 and onwards there is a plugin called "bugzilla" that calls bugzilla's native
web service API to extract the relevant information. This should work against any bugzilla installation out of the box.
There is also a plugin called "bugzillav2" that can interface with the older versions of bugzilla that don't
have a web service, however for this to work, you also need to install the
“ bugcli”
program. This is essentially an additional open source CGI script that runs on the
bugzilla server. Note that bugcli is not supported by anyone any more: if you find problems you'll
just need to upgrade Bugzilla.
If you use Jira to
track your bugs, help is also at hand. Simply set the location of your Jira installation as below. Note that it has
been found necessary to include the port number (8443 is Jira's default) to make
this work in some circumstances. Jira's webservice also requires a login before it will release information, so you need to also provide the settings "bug_system_username" and "bug_system_password" to TextTest in a similar way. A sample config file might therefore
look like this:
[bug_system_location]
jira:https://jira.blah.com:8443
[bug_system_username]
jira:texttest
[bug_system_password]
jira:the_password
[end]
These kind of settings are often useful to put in a site-specific config file.
If you use some other bug tracker with an API,
it should be fairly easy to copy the "bugzilla.py" or "jira.py" module from the TextTest
source (under default/knownbugs) and change it to implement the 'findBugInfo' method as appropriate for your
bug system. By providing the name of the module in the bug
system field when reporting the bug, it will load this module
and extract the relevant information. Naturally it's appreciated if you submit your changes
back to the project if you do this.
|