Session
1: Tools That Can Help
Tools throughout the lifecycle: Where in the software development
life cycle test activities take place and which are tool supportable.
Tool support for testing: The different types of tool support for
testing activities throughout the software development lifecycle.
Benefits and pitfalls: The benefits and pitfalls that can be expected
with the different types of test tool.
Session 2: How Much Help Can Tools Provide?
What help is required?: The reasons for introducing a test tool
and whether this is the best solution to the problem.
Are tools the best solution?: Two different approaches to comparing
different solutions against a set of problems.
Quantifying the return on investment: A simple approach to estimating
the costs and benefits of testing with and without tool support.
Example business cases: A few simple business cases for different
tool types.
Assessing readiness: How to assess the best time to Introduce change
Session 3: Introducing a Test Tool
Choosing the right tool: The need for a project to select a test
tool and the steps involved.
Role of a pilot project: The test tool implementation process and
the importance of a pilot project.
Developing best practices: A few objectives for the pilot project
and the areas where early adoption of best practices is key to long
term success.
Lessons to learn: A few case histories where things went wrong
and the lessons that can be learnt from them.
Session 4: Automating the Test Process
The test process: A detailed description of the steps taken when
creating and performing tests and their suitability for automation.
Example application: A brief description of a simple application
with an explanation of one simple test that could be performed on
it.
Automating test execution: The problems incurred by using record
/ playback facilities alone.
Automating test verification: Some of the many and varied choices
that can be made.
Automated but not Automatic: The problems involved in making the
same automated test work under different situations and an introduction
to possible solutions.
Session 5: Case History
Background: The reasons behind one organisation's decision to automate
testing and the solutions they implemented.
Results: The effort put into automation and the savings achieved.
Lessons: Things learnt in this case history that need to be addressed
by most other test automation projects.
Session 6: Scripting Techniques
Example application: A brief description of a simple application
with an explanation of one simple test that could be performed on
it.
Introduction to scripting: A few facts about scripting that help
explain its importance in test execution automation.
Scripting techniques: Five approaches to scripting (linear, structured,
shared, data-driven and keyword-driven). The pros and cons of each
technique using examples to illustrate the points made.
Script pre-processing: An explanation of script pre-processing
and the benefits that can be achieved.
Session 7: Automated Comparison
Automated test verification: Some simple guidelines that can help
you avoid making costly mistakes.
Test sensitivity: The importance of achieving the right degree
of verification and how it should be varied across different test
cases.
Implementing post-execution comparison: Alternative approaches
to automating post-execution comparison.
Simple versus complex comparisons: The benefits and pitfalls of
using simple and complex comparisons and a few approaches to achieving
them.
Post execution comparisons: A practical approach to implementing
comparisons
Session 8: Testware Architecture
Introduction: How testware is organised and structured. Three key
issues that have to be addressed in order to avoid serious problems.
An approach to organising testware that addresses each of these
three issues.
Implementation ideas: Explains one possible implementation of testware
architecture.
Session 9: Pre- and Post-Processing
Introduction: The activities that are collectively called pre-
and post-processing and the importance of automating them. Examples
of pre- and post-processing.
Implementation: Two different approaches to implementing pre- and
post-processing.
Session 10: Other Issues
Test selection: The order in which tests should be run and how
to efficiently select them for execution.
Order of running: Considerations to take account of when executing
a number of automated test cases.
Test status: The end result of a test should not be limited to
a simple pass or fail.
Monitoring progress: Metrics for reporting progress and how to
obtain and present them.
Session 11: Test Suite Maintenance
Attributes of test maintenance: Common reasons for high maintenance
costs and how to deal with them
Strategy and tactics: The general approach to keeping on top of
test suite maintenance costs
|