Norming, Calibration, and the Surprisingly Hard Problem of Fair Grading
Here's a problem I never thought I'd spend months on: making sure that when two different teachers grade the same essay, they give it roughly the same score. Turns out that's really, really hard - and the software that supports it is equally non-trivial.
At Finetune, I developed and maintained the norming and calibration modules that were central to the assessment workflow. If you've never heard of "norming" or "calibration" in this context: welcome to a niche rabbit hole that I found genuinely fascinating.
The problem, simply put
You have a rubric. You have hundreds of graders (teachers). You need them all to apply that rubric consistently. Some graders are naturally lenient, some are harsh, and some are wildly inconsistent depending on whether they've had coffee. The system needs to catch all of this before their scores count.
Calibration is the process of training graders - showing them pre-scored anchor examples and saying "this is what a 3 looks like, this is what a 5 looks like." Norming is the ongoing monitoring - periodically slipping pre-scored responses into their queue to check if they're still on track.
What I built
The implementation required some fairly complex data modeling. I needed to track grader performance over time, manage rubric versioning (because rubrics change mid-cycle, because of course they do), and calculate inter-rater reliability statistics that the administrators could actually understand.
I used Python for the backend - it was a natural fit given the data processing involved. Lots of statistical calculations, agreement matrices, and report generation. The frontend had to be dead simple because the users (educators) were already doing cognitively demanding work. The last thing they needed was a confusing UI on top of it.
The part nobody talks about
The hardest part wasn't technical. It was understanding the domain deeply enough to build the right thing. I sat through way more calibration sessions than I expected, watching how educators actually used the tools, where they got confused, where the flow broke down. You can't just read a spec and build this stuff. You have to watch people use it and feel slightly uncomfortable about all the assumptions you made.
It worked, though. Grading consistency improved measurably, and more importantly, educators trusted the system. In ed-tech, that trust is everything.