The purpose of change management is to implement processes for effecting change, controlling change and helping the project teams to adapt to change arising out of various internal and external factors. This document seeks to outline the change management strategy for the HCMP rollout in Mozambique.
All activities that require a change in the agreed scope shall be handled through the change management process. The following activities will be executed as part of change management:
Receipt of a change request (CR) from internal or external entities. This could be CHAI/NMCP or other partners as well as eGov's internal teams.
Impact analysis of the change request by the eGov team.
Negotiation/decision on the change request implementation.
Change implementation
Change tracking
Change version control
Change release and record management.
The initiator of the change request will have to formally raise a request for a change mentioning the relevant details. eGov as the platform provider will establish a change control processes to manage changes to the agreed scope and other dimensions within the project. The process will capture all requests for change while at the same time ensuring that decisions are traceable, made at the correct level, and that their impacts are fully analyzed and absorbed.
We propose to follow the listed change management procedure:
Any request received to make changes in the project scope will be logged as "Change Requests" and compared against the agreed scope and the schedule baselines for the project. The reasons for the change request could be due to one of the following drivers:
Changes in the business environment
Changes in functional requirements
Changes in technical specifications
Changes in system design
Changes in technology
Changes in volume and load projections
Changes in personnel and staffing, and their roles, etc.
Changes in the times and schedule which may necessitate scope changes
All change requests raised will be logged in the ‘Change Register or log’ which is expected to be maintained jointly by the EGOV and CHAI project coordinators. This will help maintain a consistent log of all change requests that are raised. The Change Register will be shared at least on a fortnightly basis with all stakeholders.
The Change Request Note (CRN) will describe the change:
Scope of the change
The rationale for the change
The estimated benefits expected to accrue from the change,
The priority and importance that is attributed to the implementation of the change
The estimated time within which this change is expected to be applied
The purpose of this document is to outline the User Acceptance Testing (UAT) process for the DIGIT-HCM platform rollout in Mozambique as part of the seasonal malaria chemoprevention (SMC) campaign in the Nampula province. This will enable the SMC campaign users to validate that the HCM software meets the agreed upon acceptance criteria for various functionalities. It will ensure that the system satisfies the needs of the business as specified in the functional requirements and provides confidence in its use. It will also help to identify the defects with associated risks and suggest enhancements (change requests), communicate all known defects/enhancements to the project team, and ensure that all issues are addressed in an appropriate manner prior to go-live.
At the end of the UAT, the following results are expected:
Sign-off of the application indicating that the app meet the requirements gathered.
Obtain feedback from the end users (Health Facility coordinators, Community Distributors and Community Distributor Supervisors).
Verify if the application is able to do the functionality required to conclude the task of the Health Facility coordinators, Community Distributors and Community Distributor Supervisors (completeness functionality).
Verify if the application is easy to use for the users (usability).
Understand the areas that end-user may face challenges to improve the training curricula.
Participants
SPS (Provincial Health Service) Health Facility coordinators ( 2 people)
Community Distributor Supervisors (2 people) Community Distributors (3 people)
M&E (2 people)
NMCP - National Malaria Control Program
1 person
MC - Malaria Consortium
1 person
eGov
2 people
Dates
16th and 17th of November
Location
Nampula City
Devices Needed
6 phones for the DPS participants
Macro plan for Day 1
From 9 am to 12 pm: - Session with TWG to go through the UAT Test Cases; From 2 pm to 4:30 pm:
Induction session with Health Facility (HF) Coordinators and Community Distributor (CD) Supervisor;
UAT with Community Distributor Supervisor;
Debrief session with CD Supervisor and HF Coordinator;
Macro plan for Day 2
From 8 am to 11:30 am:
Induction session with community distributor supervisor and -community distributor;
UAT with Community Distributor and Community Distributor Supervisor
Debrief session with CD Supervisor and CD
From 1 pm to 3:30 pm:
Session with TWG to debrief on the UAT
During the testing process, a pre-approved list of test cases/scripts will be executed to verify the behaviour and performance of various functionalities. The observations from the testing will be noted and classified as defects or enhancements. Some of the enhancements may be doable using configuration changes while others (change requests) may have to go through a change management process.
UAT observation classification:
Serial number
Observation Type
Description
Addressing Mechanism
1
Defects
Any observation pertaining to a feature or functionality not working as expected or agreed at the time of scope review, will be classified as issue
The observations classified as defects will be taken up by the eGov program team for further validation, prioritisation and fixing. Minor issues originating due to incorrect configurations or erroneous labels, or translations will be fixed and be made available for re-testing during the next UAT cycle.
2
Change Requests (CR)
Any recommendations to enable or disable a functionality from the initial requirements or a new functionality to be added
These will be handled via Change control Process as per the SoP defined and will be evaluated based on their impact and effort.
The UAT team will execute all the test scripts referenced later in the "UAT Feedback" section. Users may also perform additional tests not detailed in the plan but remain relevant and within the scope of the project.
Users will report the feedback to the eGov team for documentation and escalation using a Google sheet. These defects will be described, prioritised, and tracked by using screen captures and steps necessary for the development team to reproduce the defect. For change requests, the requirements will be described, analysed, prioritised, and tracked.
The various phases of UAT are shown in the below diagram:
UAT will be conducted in 1 phase with the below mentioned scope
Test Phase
Tracks
Scope
UAT
SMC HCM Mobile App
Registration and Service Delivery, Supervision, Inventory from the mobile app.
For the below mentioned functionalities, separate demo and/or training sessions will be conducted for the targeted user groups.
Help desk for complaints and user management
a. User Management module to add/update/ activate/deactivate users and role mapping
b. Handling complaints via inbox functionality
Central, Provincial, District dashboards and reports.
During UAT , the team will validate the end-to-end flow and features of the application as listed below:
End-to-end business flow for each of the identified user flows
All the required fields are present
System is accepting only valid values
User can save after form filling all the mandatory fields
Correctness of label and its translations
Following is the list of prerequisites for conducting the UAT
The HCM Mobile App for UAT deployed in the UAT environment
Mobile phones setup with HCM app.
Configuration of HCM Mobile App in UAT environment with master data from Nampula
Readiness of handouts
- Test Case document
- Mockups
- Defect/Change request reporting template
Availability of teams from SMC TWG (Technical Working Group) and SPS for UA testing
Nomination of participants for the UAT session so that test accounts can be created by eGov.
The UAT environment will be similar to the production environment in terms of specifications, and will be provided by eGov, so that accurate assumptions can be drawn regarding the application’s performance.
Each test participant will be provided with a checklist to verify access to the applications within the defined scope of testing. The tester will then login, perform the prescribed steps, and generate the expected results for each activity. Any feature identified as missing or bug during testing from the UAT environment should be reported back to eGov.
Access to the test data is a vital component in conducting a comprehensive test of the system. All UAT participants will require the usage of test accounts and other pertinent test data which should be provided by SMC TWG upon request. The user roles should fully emulate production in the UAT path. The test accounts and role mapping shall be done for the users identified by eGov. Following are sample test data for UA testing:
Sample Master Data
- Users data for test login creation
- Location master (AP/Locality/Health Facility/Village )
- Micro Plan of a district
Training
P1
Pradipta
13-Feb-23
15-Feb-23
16-Feb-23
21-Feb-23
23-Feb-23
Helpdesk
P1
Pradipta
13-Feb-23
15-Feb-23
16-Feb-23
21-Feb-23
23-Feb-23
User management
P1
Ankit
16-Feb-23
17-Feb-23
17-Feb-23
21-Feb-23
23-Feb-23
UAT
P1
Sagar
16-Feb-23
17-Feb-23
17-Feb-23
21-Feb-23
23-Feb-23
Change Management
P2
Ankit
Master Data management
P2
Sagar
Platform and app monitoring SoP
P2
Elzan
MDM(scale fusion ) SoP
P2
CHAI
Login/logout and sync
P1
Nomination of stakeholders for executing the UAT
Verification of product based on UAT scenarios
Decision regarding date and venue
Providing feedback after the UAT session and sign off
Readiness of HCMP with required features
Demo to SMC TWG before the UAT event
Framing test scenarios
Creation of UAT Plan
Collection and triaging of the UAT feedback
Translation of UAT feedback
Arrangement of logistics-Venue
Addressing costs for all the government personnel involved in the UAT
Providing feedback after the UAT session
Availability of the end users for SMC
Access to the HF
The test team is composed of members who possess a knowledge of the operations and processes on the ground.
These team members will be able to initiate test input, review the results,
Have prior experience/learnings from SMC campaign digitisation in Mozambique
All team members will be presented by the eGov team with an overview of the test process and what their specific role in UAT will be. The eGov team will oversee testing by assigning scripts to testers, providing general support, and serving as the primary UAT contact point throughout the test cycle.
Name of participant
Project Role/Designation
Phone
Entity
(NMCP/DIS/DTIC/MC)
Note: The above table needs to be filled by SMC TWG with the details of the nominated participants of the UAT session.
The following sections detail milestones crucial to the completion of the UAT phase of the project. Once all dependent milestones have been completed, NMCP/DIS/DTIC will formally sign-off on the system’s functionality and distribute an email to all project stakeholders.
All core UAT activities along with the deliverable dates are described in the below table:
Task
Owner
Start Date
Expected Closure date
UAT
Creation of the UAT-1 test plan and sharing with NMCP for review
eGov
18-Oct-23
18-Oct-23
Creation of the UAT test cases/scripts and sharing with NMCP for review
eGov
18-Oct-23
20-Oct-23
Nomination of UAT participants so that user accounts can be created
SMC TWG
16-Oct-23
19-Oct-23
Arrangement of venue and other logistics for UAT session
NMCP
19-Oct-23
20-Oct-23
Readiness of the UAT server instance
eGov
19-Oct-23
23-Oct-23
Installing test version of the HCM application in personal phones
NMCP
15-Nov-23
15-Nov-23
Conducting the UAT testing along with the end users and SMC TWG
SMC TWG
16-Nov-23
17-Nov-23
Collation and triaging of UAT feedback and communication to SMC TWG regarding what can be incorporated in the HCM app
eGov
17-Nov-23
17-Nov-23
Reviewing the test results and providing sign off of UAT
SMC TWG
17-Nov-23
17-Nov-23
The UAT session will be conducted hybrid, in Nampula and remotely using Zoom Meeting. The detailed plan for the UAT session is as follows:
Serial number
Activity
Approximate Duration/Participants
Time / Location
Day 1: 16 Nov 2023
1
Introductions to UAT Session plan
10 minutes | TWG members
9.00 am-9.10 am
Zoom Meeting/ SPS Office
2
Walkthrough of Mobile app and Test cases
Walkthrough of Defect/CR reporting templates
Distribution of test cases
50 Minutes | TWG members
8.10 am - 9.00 am
Zoom Meeting /SPS Office
3
Execution of Test cases
Registration & Distribution
Supervision
Refer
Inventory Management
Capture feedback in google sheet
120 Minutes | TWG members
9.00 am-11.00 am
Zoom Meeting /SPS Office
4
Collation of feedback received
90 minutes | TWG members
11.00am-12.30 pm
Zoom Meeting /SPS Office
5
Walkthrough of Mobile app and Test cases for HF Supervisor and CD Supervisor and
Inventory Management and supervision test Cases
30 Minutes | HF Supervisor and CD Supervisor
2.00pm - 2.30pm
6
Execution of Test cases
Supervision
Inventory Management
Capture feedback in google sheet
90 min | HF Supervisor and CD Supervisor
2.30pm - 4.00pm
7
Collation of feedback received and debrief of the session and results
30 minutes | HF Supervisor and CD Supervisor
4.00pm - 4.30 pm
Day 2: 17 November 2023
1
Induction session with community distributor supervisor and community distributor;
Registration & Distribution
Refer
30 min | CD Supervisor and CD
8:30 am - 9:00 am
2
UAT with Community Distribuitor and Community Distribuitor Supervisor
120 min | CD Supervisor and CD
9.00 am - 11.00 am
3
Debrief session with CD Supervisor and CD
30 min | CD Supervisor and CD
11.00 am - 11.30 am
4
Review of feedback received and clarification of doubts if any (agreement on CRs and bugs)
60 minutes | CD Supervisor and CD
2.00pm-3.00pm
5
UAT Sign off and defining next steps
30 minutes | CD Supervisor and CD
3.00 pm - -3.30pm
For the CRs identified during UAT, the standard change management process will be followed.
The mutually agreed defects/CR from UAT will be discussed in the post- meeting.
The sign-off shall be provided at the end of UAT by SMC TWG via an email communication.
Post sign-off the application will be ready for the deployment in the production environment.
Test cases provide a high-level description of the functionality to be tested. All regression and new functionality test cases are contained in the Excel spreadsheet “UA Test Cases” to be shared before.
The team will leverage relevant QA test cases for project specific functionality. Each test case based on new functionality will reference a specific functional requirement.
Each script contains the test case number, test description, requirement number, action to be performed, test data to be utilized, expected results, error descriptions (if applicable), pass/fail results, date tested, and any additional comments from the UA tester.
The UA test scripts are contained within the UAT test case spreadsheet and can be accessed via hyperlink from each individual test case.
Defects and change requests will be entered and tracked in Google sheet by any member of SMC TWG during the UAT process. Each entry will include detailed information about each defect/CR.
Test team will be provided with instruction on how to effectively execute test scripts, as well identify, capture, and report defects/observations by the eGov team at the beginning of the UAT session. The test team will present their findings during the UAT session.
Defects must be clearly captured and escalated to ensure prompt resolution by development. Each defect submitted by UAT will be assigned a priority, worked by development, resolved, and re-tested by UAT prior to closure. The following is a snapshot of the standard defect lifecycle:
the SMC TWG together will prioritise and classify defects. Defects found in UAT can be assigned one of three (3) levels of severity:
P1, Work Halted – Testing defects that due to the complexity of the function or the scheduled dates are putting the implementation date at risk. No workaround exists.
P2, Work Slowed – Testing defects occurring in a less complex function of the application with sufficient time to resolve before the implementation date – but must be implemented as scheduled. A workaround has been identified and is listed in the defect.
P3 and lower, Work Unaffected – Testing defects occurring in a function that are simple to fix or could be excluded if not resolved by the scheduled implementation date.
Severity
Description
P1, Work Halted
Critical application outage impacting service for which the cause is unknown. No bypass or workaround available.
P2, Work Slowed
A key component of the application is degraded, unusable or unavailable, some users affected.
P3 and lower, Work Unaffected
A component of the application is degraded, which causes a minor inconvenience, but a workaround is available.
As a non-profit, we are unable to make any commitment on how long issues will take to fix and deploy in production. We will endeavour to resolve P1 issues as quickly as possible.
The change request must be clearly captured and reported for analysis to identify effort and Impact to the eGov team. Each CR submitted will be validated and categorised for acceptance and then assigned with a priority. The development team will work on it and will be made available for testing. Following is a snapshot of the standard CR lifecycle:
eGov in consultation with NMCP will decide acceptance and categorisation of change requests. Change requests found in UAT can be assigned one of three (3) levels of category:
Must Have – Change requests that are needed for the success of a campaign. No workaround exists.
Should Have – Change requests that are required for better tracking and monitoring and will increase the ease of use of the system for users. A workaround has been identified and is listed in the CR.
Good to Have – Change requests that are simply for better visualisation and reporting. It could be excluded if not resolved by the scheduled implementation date.
eGov to endeavour to cover the "must have" changes before distribution. Lower priority changes will be taken through the eGov gating process for planning the subsequent releases.