Reduce marking time by up to 90%!
Powermark is Smartgrade’s easy-to-use AI-powered automarking tool that helps teachers save valuable time while delivering high marking accuracy and personalised class and cohort feedback.

Assessment marking shouldn’t take hours
With Powermark, Smartgrade’s AI-powered automarking tool, assessments are marked in moments - giving you more time to focus on analysis, trends, and supporting your pupils.
Automatically marks HeadStart Primary Reading and GPS online assessments, or your own custom online assessments, in real-time, with better-than-teacher levels of marking accuracy (typically 94-97%).
With Powermark, your marking workload drops dramatically so you can swiftly moderate assessments and reclaim your evenings. Customers have reported a time saving of anything between 65% and 90% compared to traditional marking.
No personally identifiable information is ever passed to our AI models, ensuring full data protection for your students. We only use AI tools that meet our strict data standards, and your data is never used for model training.
Instantly generate aggregated feedback reports for your class, year group, or even MAT featuring strengths, misconceptions, and areas for improvement.
You’re always in control. Any answers that Powermark is less confident in are automatically flagged for easy teacher review at the top of each report.
How does Powermark work?
Use selected HeadStart Primary online termly assessments or build your own custom assessments to use Powermark AI-powered automarking. They’re simple to use and get started in a matter of clicks.
Our AI automarking principles
Smartgrade has defined six principles that underpin our AI work and guide our product and policy choices.
Testimonials
"My teachers have been skipping along the corridors - we really love this."
Read about Powermark AI auto-marking
Frequently Asked Questions
Where some questions are ambiguous or the marking guidance is less rigorous, accuracy can dip into the 80-90% range. For this reason, we do not currently make automarking available by default for all assessments. Instead, we work with our assessment and MAT partners to ensure that assessments have high-quality marking guidance before we enable automarking.
We have conducted extensive testing and piloting using a wide range of assessment types, subjects and year groups. This ensures that our approach does not just work within limited, tested parameters, reducing brittleness.
Uncertainty
We are currently implementing an approach which involves the AI assigning a confidence score (1–10) to each mark, where 10 is full confidence and 0 is no confidence. Questions with a confidence score of 7 or below will be flagged as “priority for teacher review” in the product going forward.
Hallucinations
We are experiencing minimal levels of hallucination, partly because we automark using tightly defined marking guidance, which leaves less space for hallucination. That said, we still check for hallucinations in a number of ways. Teachers moderate automarking, allowing them to spot and correct hallucinations. A sample of teacher adjustments are then checked by us to scrutinise for hallucinations, amongst other things.Hallucinations are more likely when the AI is less certain of its answer, so by taking the “confidence score” approach described above we can further mitigate the risk of hallucinations appearing, and make it more likely that teachers will correct them when they do occur.We do periodic random sampling of all marks in our assessments and we get an expert to automark those samples and flag any hallucinations. In the most recent 2,000 expert-marked questions we have discovered no hallucinations.
Embedded Bias
Our approach to automarking involves no passing of personally identifiable information to our marking engine, so no bias can be derived from knowledge of a student’s characteristics. Moreover, we use a prescriptive mark scheme that the AI applies directly, rather than allowing the model to generate open-ended interpretations, which could in theory be subject to bias of some form.
We evaluated false positives using a confusion matrix. While present, they were uncommon. Interestingly, we observed more false negatives than false positives.