These assessments focus on continuous improvement and learning at regular intervals.
All services within the scope of the Digital Service Standard are assessed against the standard.
Depending on the service, this will follow either the in-flight assessment process or the staged assessment process.
The in-flight assessment process focuses on continuous improvement and learning. It is less about regular tests and more about a mentoring relationship between the assessment team and the delivery team.
Both teams should celebrate moving from one stage to the next.
Your delivery team will meet with the assessment team regularly during the Discovery, Alpha and Beta stages, until the service is launched as a public beta.
As the service moves towards the Live stage, the assessment and delivery teams can move these check-ins to longer intervals. However, check-ins should happen at least monthly.
If you are moving from staged assessment to in-flight, you might spend more time on the first check-in meeting to help the assessment team understand the product.
How the check-ins work
Your check-in meetings should focus on the progress you’ve made against the Digital Service Standard. You should document this using a red, amber, green (RAG) kanban. You can use the Digital Service Standard kanban poster (PDF 117 KB) to help with this.
Both teams work together to make sure the service meets all of the requirements for each criterion in the current stage.
The assessors document what the delivery team has done well since the last meeting and make recommendations for the next period of work. The assessors will also recommend ways the delivery team can work closely to the Digital Service Standard and create a successful service.
The teams should complete their comments and RAG ratings by the end of each in-flight session and make them available for weekly reporting to stakeholders.
Assessors will use a spreadsheet to track the delivery team’s progress and record recommendations.
Moving to the next stage
Both the assessors and the delivery team agree on when the delivery team has met the requirements of the stage and when the service can move to the next stage. They can do this by showing the criteria being assessed as green in the red, amber, green (RAG) rating.
At each of the 3 stages — Alpha, Beta and Live — the assessment team produces a brief report. This explains the achievements and learnings of the previous stage. The assessment report is published on our website.
If you work on a high-volume transactional service, or you started designing/redesigning your service after 6 May 2016, you will need to provide us with your Alpha, Beta and Live reports for publishing.
Services moving to public beta will need to start reporting on the Performance Dashboard.
Maintaining your service
Launching your service is only the beginning. Once a service goes live you will need to keep listening to your users to make sure it continues to meet their needs and you are still following the Digital Service Standard. You will need to continue to update and improve the service on the basis of user feedback, performance data, changes to best practice and service demand.
You will also need to check that:
you maintain high levels of user satisfaction and transaction completion
cost per transaction goes down
digital take-up goes up
you provide assisted digital support to people who need it
there is progress against other measurements you identified when you were developing the service