Smartgrade Features Built for MATs
From trust-wide insights to seamless setup, every feature is shaped around the way MATs operate.

Powermark's Guiding Principles
Highly engaged MATs trust us
Why not become one of them?
Testimonials
“Smartgrade is a key element of our curriculum and assessment strategy across our 33 primaries.”
“Analytical tools enable deep and detailed understanding of what pupils know and smart marking features mean minimal workload for maximum information to inform next steps. We are data-rich without wasting a minute of learning or teacher time!”
Emily Hobson - Oasis Community Learning
“Smartgrade allows us to standardise our assessments across a national cohort, giving teachers, leaders and students in our group and beyond a highly valid and powerful understanding of how their performance compares to their peers.”
Dale Bassett – United Learning
“Smartgrade’s process allows us to celebrate successes and target support where it’s needed, at a curriculum level. It is a crucial tool as part of our KS3 common assessments.”
Nimish Lad - Creative Education Trust
Read about Powermark AI auto-marking
Frequently Asked Questions
As with teacher marking, accuracy varies depending on the type of assessment and the level of subjectivity involved in marking. As a principle, we aim for 94%+ accuracy.
Where we have detailed marking guidance, we find that we can consistently achieve this target of 94%+ accuracy. For example, in our largest ever test, using HeadStart Reading and GPS assessments and evaluating across a sample of 9,500 student responses in multiple primary year groups, our AI achieved 97% accuracy, compared with 94% for teachers.
Where some questions are ambiguous or the marking guidance is less rigorous, accuracy can dip into the 80-90% range. For this reason, we do not currently make automarking available by default for all assessments. Instead, we work with our assessment and MAT partners to ensure that assessments have high-quality marking guidance before we enable automarking.
Where some questions are ambiguous or the marking guidance is less rigorous, accuracy can dip into the 80-90% range. For this reason, we do not currently make automarking available by default for all assessments. Instead, we work with our assessment and MAT partners to ensure that assessments have high-quality marking guidance before we enable automarking.
We continue to evaluate accuracy on an ongoing basis as the system is rolled out more widely, with random sampling and expert review to ensure quality is maintained.
Smartgrade mitigates risks commonly associated with AI systems, such as brittleness, hallucinations, embedded bias, uncertainty, and false positives using the following measures:
Brittleness
We have conducted extensive testing and piloting using a wide range of assessment types, subjects and year groups. This ensures that our approach does not just work within limited, tested parameters, reducing brittleness.
Uncertainty
We are currently implementing an approach which involves the AI assigning a confidence score (1–10) to each mark, where 10 is full confidence and 0 is no confidence. Questions with a confidence score of 7 or below will be flagged as “priority for teacher review” in the product going forward.
Hallucinations
We are experiencing minimal levels of hallucination, partly because we automark using tightly defined marking guidance, which leaves less space for hallucination. That said, we still check for hallucinations in a number of ways. Teachers moderate automarking, allowing them to spot and correct hallucinations. A sample of teacher adjustments are then checked by us to scrutinise for hallucinations, amongst other things.Hallucinations are more likely when the AI is less certain of its answer, so by taking the “confidence score” approach described above we can further mitigate the risk of hallucinations appearing, and make it more likely that teachers will correct them when they do occur.We do periodic random sampling of all marks in our assessments and we get an expert to automark those samples and flag any hallucinations. In the most recent 2,000 expert-marked questions we have discovered no hallucinations.
Embedded Bias
Our approach to automarking involves no passing of personally identifiable information to our marking engine, so no bias can be derived from knowledge of a student’s characteristics. Moreover, we use a prescriptive mark scheme that the AI applies directly, rather than allowing the model to generate open-ended interpretations, which could in theory be subject to bias of some form.
We have conducted extensive testing and piloting using a wide range of assessment types, subjects and year groups. This ensures that our approach does not just work within limited, tested parameters, reducing brittleness.
Uncertainty
We are currently implementing an approach which involves the AI assigning a confidence score (1–10) to each mark, where 10 is full confidence and 0 is no confidence. Questions with a confidence score of 7 or below will be flagged as “priority for teacher review” in the product going forward.
Hallucinations
We are experiencing minimal levels of hallucination, partly because we automark using tightly defined marking guidance, which leaves less space for hallucination. That said, we still check for hallucinations in a number of ways. Teachers moderate automarking, allowing them to spot and correct hallucinations. A sample of teacher adjustments are then checked by us to scrutinise for hallucinations, amongst other things.Hallucinations are more likely when the AI is less certain of its answer, so by taking the “confidence score” approach described above we can further mitigate the risk of hallucinations appearing, and make it more likely that teachers will correct them when they do occur.We do periodic random sampling of all marks in our assessments and we get an expert to automark those samples and flag any hallucinations. In the most recent 2,000 expert-marked questions we have discovered no hallucinations.
Embedded Bias
Our approach to automarking involves no passing of personally identifiable information to our marking engine, so no bias can be derived from knowledge of a student’s characteristics. Moreover, we use a prescriptive mark scheme that the AI applies directly, rather than allowing the model to generate open-ended interpretations, which could in theory be subject to bias of some form.
False Positives
We evaluated false positives using a confusion matrix. While present, they were uncommon. Interestingly, we observed more false negatives than false positives.
We evaluated false positives using a confusion matrix. While present, they were uncommon. Interestingly, we observed more false negatives than false positives.
Automarking of scanned paper assessments is the next phase of Powermark - Smartgrade’s auto-marking project.
We’re currently trialing using AI to mark scanned in paper assessments.
If your school or MAT would like to be part of this pilot project, please get in touch with us on sales@smartgrade.co.uk.
We do not believe there is any legal obligation under UK law to inform students or parents of AI use.
This is because no personally identifiable information is passed to the AI system, and no automated decision-making occurs.
However, if participating schools would like to notify parents, we provide wording that can be used in parent-school communications.
Everything you need for smarter assessments
Book a Demo Today





.png)