You are viewing an old version of this page. View the current version.

Compare with Current View Page History

« Previous Version 26 Next »

 

 

Introduction

This page was created after the DCAT Meeting September 2015, where a working group was formed to prepare the test plans for the DSpace 6 Testathon.

Objectives

  • Create two plans, one for XMLUI, one for JSPUI
  • Create these plans as collaborative spreadsheets (Google Docs)
  • Group the different tests per user persona/access level
    • Anonymous - Non-authenticated, anonymous users
    • Submitter - Authenticated user with submission rights
    • Collection Admin - Authenticated user with collection administration rights
    • Admin - Authenticated user with admin rights
  • Create plans to be used on demo.dspace.org
    • As a secondary benefit, it would be great if these plans can work as a model for in-house testing of your own upgrades.
    • It should be very clear what the reference installation on demo.dspace.org entails

Out of scope

  • Specific work on DSpace documentation
    • This will be treated in parallel, related to each new feature for which documentation needs to be reviewed
  • Tests for the other DSpace webapps
    • OAI-PMH
    • REST
    • SWORD
    • LNI
  • Performance and Load tests

 

Official drafts

JSPUI Test plan: TODO

XMLUI Test plan

Different people's own test plans (examples)

Virginia Tech, VTechWorks - https://docs.google.com/spreadsheets/d/1_-20Y06C8dP2VzWxaSBbMnXgvuBOQoc3D0chJGRmyGg/edit?usp=sharing

  • LIKE
    • Bram: column with "expected results".
  • DISLIKE / NOT APPLICABLE
    • ...

Georgetown University Institutional Repository - https://docs.google.com/spreadsheets/d/1QkasjcwhhiAj4HXIDFDeX7TFEocwnqsZGnmDTHNmnCs/edit?usp=sharing

  • LIKE
    • Bram: Section column to group tests together
  • DISLIKE / NOT APPLICABLE
    • ...

 

Atmire - https://docs.google.com/spreadsheets/d/1rbgJk2_NkYWyE9tSIHquiLqN5PdcqRP0PCBHLnVrL_I/edit#gid=409937147

  • LIKE
    • Categorization of what the different types of feedback can be
  • DISLIKE/NOT APPLICABLE HERE
    • Columns for different servers
    • Mix of things that can be tested in the UI & technical things that can only be verified by an admin.

 

Edinburgh - https://wiki.duraspace.org/download/attachments/69834931/DataShare%20Testing%20Results%20Blank%20v4.xlsx?version=1&modificationDate=1441726810758&api=v2 

  • LIKE
  • DISLIKE / NOT APPLICABLE HERE

Mandatory fields (per test)

Ref - a short, unique reference that can be used in related JIRA tickets etc. example VIEW1

Category - an indication of the functional group a test belongs to. example VIEWING

Description - a step by step description of how the test should be executed

Expected outcome - the expected result after execution of the description

Test status - status of the last time this was tested. This column should be filled in with following vocabulary:

  • 2014-09-12 : when a date is entered, the functionality was verified. 

  • TEST UNCLEAR: the description of the test doesn't make it clear either what to test, or how a user can test this.

  • FEATURE MISSING: the described functionality couldn't be verified

  • UI PROBLEM: the feature is there and seems to "work", but there is a UI problem (nature of the problem clarified in comments)

  • NON COMPLIANCE: the test fails the expected results

  • DOCUMENTATION: feature seems to work but docs seem to be missing or incomplete

If you are doing a test later in time, you can overwrite the test status that someone had previously entered

Last tester - name of the person who executed the last test. The main goal of including the person here, is to make sure he or she can be contacted for further information about the test

Comments - free text field for additional comments. If the comments of the previous tester are not yet resolved or are still relevant in another way, you can leave them in and add your own comments in the same cell. When doing this, start your comment with your name, so people can see a different person added the next comment

JIRA tickets - links to JIRA tickets that are related to this test. These links should never be removed, so we still have the entire backlog on what has changed to these tests.  

Structure of the test spreadsheets

The spreadsheets are organized into different tabs. Each tab represents a "role" of a user that is able to execute all tests on the tab. The roles who need a higher level of authorization are more on the right. This means that a role on the right could theoretically also execute all of the tests on the tabs more on the left (but we're not targetting that actively right now).

Requirements for demo.dspace.org installation

Integrations with external systems

Following integrations may not be enabled out of the box. They need to be configured to ensure they are "up and running" on demo.dspace.org.

  • Handle.net
  • DOI
  • ORCID
  • Google Analytics
  • Creative Commons (licensing)
  • Sherpa Romeo Lookup
  • LDAP or other custom authentication methods
    • Many people use this feature but I can't see how we can enable this on demo.dspace.org without also putting up an entire directory infrastructure.
      Emilio: what about LDAP directory "maintained" by Stewart Lewis? Just some need to define the same users at demo.dspace.org.....

Specific credentials

  • Admin email inbox: specific testers should have access to the email inbox that SHOULD receive entries in the repository feedback from, the "Request a copy" requests, ...
  • Google Analytics account

Checklist for features that should be enabled

  • Workflow: Basic workflow or XML workflow ??? Only one of the two can be enabled.
  • PDF cover page?
  • Versioning
  • Collection strength should be shown
  • Legacy stats ???? (Bram: I'm not in favor of still showing these or recommending use of them to anyone, but some people still seem to use them)
  • Thumbnail generation (filter media execution and potential XPDF installation)
  • outgoing email
  • i18n: ideally, all language files we have should be enabled, this makes it clear which messages are wrong/missing in which files.

DRAFT of our procedures that we want to add to the official testathon page

Unless specified otherwise, the credentials that should be used for the tests are the ones listed on the demo.dspace.org/xmlui and /jspui homepages. 

The reference browser for the tests is Google Chrome. However, there are possibilities to report browser compatibility issues with IE, Firefox and Safari.

Testathon page: DSpace Release 6.0 Testathon Page

DSpace 6 Testathon Role Managers

following volunteers are looking after each of the roles, represented by tabs in the XMLUI test plan:

Anonymous user: ADD YOUR NAME HERE

Submitter: ADD YOUR NAME HERE

Collection Administrator: ADD YOUR NAME HERE

Repository Administrator: ADD YOUR NAME HERE

 

 

  • No labels