Skip to content

Test_strategy

William French edited this page Aug 7, 2014 · 1 revision

Test Strategy

Main information

Resources:

Goals of Testing

Automation testing:

  • Rapid regression testing to allow systems/applications to continue to change and improve over time without long "testing" phases at the end of each development cycle
  • Finding defects and problems earlier and faster especially when tests can be run on developer machines, and as part of a build on a CI server
  • Ensure the user can interact with the system as expected

Overall, with automated testing we are aiming for increased project delivery speed with built in quality.

Manual Testing:

  • Cover areas where automation testing unavailable

Criteria of success

  • Main functionality is covered by tests
  • Customers don't find a critical bugs

Key Tasks

  • Avoid fake test fails
  • Decrease test execution time
  • Cover "canvas" elements by automated tests

Start of testing criteria

Testing start:

  • New build is deployed
  • Smoke tests (all tasks from status Deployed)
  • Every night or weekend
  • Regression test
  • By request

End of testing criteria

  • All tests are put through
  • A tester does not find the bug for a long time
  • A build is not suitable for testing
  • A new build is deployed

Description of the testing process on the project

Test approach for manual testing:

  • Script testing
  • Exploratory testing
  • Student role
  • Lecturer role
  • Researcher
  • Author

Test approach automation testing

  • UI (Acceptance layer)
  • Integration testing (API layer)
  • Unit testing (by developers)

Related links

Test execution priority

Priority Test that checks The order of the test
High 1. All have to work, work Run first
2. And working properly
3. And everything is available, that is necessary for the rights
4. Use the correct input (with permission).
Normal 1. All have to work, work If possible
And working properly
Use the correct input (with permission)
Tests to "break the program":
search for conditions under which does not work.
Or malfunctioning.
Or works that rights are not supposed to.
Used incorrect input (negative tests)
Low Tests that test rarely used functionality and interface features Performed by the presence of time

Bug tracking

GitHub Workflow

Short: short bug flow

Full: full bug flow

Bug Severity

  • Blocker: The defect affects critical functionality or critical data. It does not have a workaround. Example: Cannot login, cannot upload data,
  • Critical: The defect affects major functionality or major data. It has a workaround but is not obvious and is difficult. Example: All exceptions, security bugs
  • Minor: The defect affects minor functionality or non-critical data. It has an easy workaround. Example: Open unexpected page
  • Trivial: The defect does not affect functionality or data. It does not even need a workaround. It does not impact productivity or efficiency. It is merely an inconvenience. Example: Petty layout discrepancies, spelling/grammatical errors.

Bug template

  1. Open chronozoom
  2. Magnify to any leap year and to February mounth. // Or simply open [Url]
  3. Try to find 29 February

Observed result: Text

Expected result: Text