Case Study · NoRedInk · Enterprise

Benchmarks.

How redesigning admin-facing views drove clearer data, better decisions, and a 17% lift in teacher adoption of district assessments.

Enterprise District Admin Assessment

Ben Dansby · Product Design

01 / 13
Context

Assessment at the district level.

Benchmarks were NoRedInk's first foray into enterprise assessment. Unlike the teacher-centric model, Benchmarks flowed from district admin down: admins create and distribute assessments, teachers assign them, and the resulting data rolls back up to give admins a picture of student performance across their entire district.

The feature supported series-based comparisons: the same assessment given at different points in time, so that districts could track growth.

But customer feedback made one thing clear: adoption was suffering. The interfaces admins used to track progress and view results weren't delivering the instructional value districts needed to stay engaged.

How Benchmarks work
1

Admin creates a Benchmark Series

Selects content, sets dates, and sends it out to teachers across the district.

2

Teachers assign to students

Each teacher assigns the Benchmark to their classes within the admin's timeline.

3

Data rolls up for comparison

Admins view participation and results across schools, classes, and over time.

02 / 13
Research & Discovery

What admins actually needed.

We ran a multi-pronged research effort: a feature audit, admin interviews, competitive analysis, and a pedagogical review. A consistent picture emerged.

Admins had two core goals: get a reliable snapshot of where students stood relative to grade-level expectations, and use those insights to guide coaching and intervention decisions.

The current experience wasn't delivering. Results didn't lead to clear next steps. Rubrics didn't map cleanly to state expectations. District leaders couldn't quickly identify patterns. The data was there. The story wasn't.

"We need to see which schools are falling behind, not just which students."

Admin interview participant

What we heard

Results don't lead to clear next steps

Can't quickly spot trends across the district

Dashboard offers little visibility into progress

03 / 13
Starting Point

The foundation was there. The usability wasn't.

Admins had a Dashboard listing active assessments grouped by Series, and a Results page with breakdowns by school, teacher, class, and student. The rudiments were right, but the data was buried and the layouts made it hard to act on anything.

Original dashboard
Original results page
Original results detail
04 / 13
Dead End

Not every request can be fulfilled.

Research surfaced a desire for aggregate reports: performance data rolled up by grade and school, independent of any single assessment.

I explored this deeply, but hit a fundamental data problem. A single grade or school can have multiple unrelated Series, so aggregating across them would mix differing dates, content, and student bodies. Meaningless at best, misleading at worst.

We dropped the feature rather than ship something that looked authoritative but wasn't.

Knowing when to say no is as much a design skill as knowing what to build.

05 / 13
Visibility & Insight

More data where admins already looked.

Though aggregate reports were out, the Dashboard was still nearly opaque. I added student participation data alongside teacher participation in a new tabular layout, and surfaced performance data directly on the Dashboard, so admins got a real window into results without drilling into each Series.

These additions gave users the bird's-eye view of data they were looking for while keeping the page focused and uncluttered.

06 / 13
Layout Redesign

Series first. Data forward.

Customer Success told us the distinction between individual Benchmarks and a Series wasn't clear enough, causing admins to create duplicate Series when they meant to add to an existing one.

The new layout put Series at the center rather than the individual Benchmarks within them. Sequence became visually explicit. Button copy was tightened to clarify intent. And with the aggregate report concept gone, the page became simpler and more focused.

More emphasis on Series hierarchy

Clearer sense of sequence within a Series

Revised copy to reduce action confusion

Redesigned dashboard layout
07 / 13
Usability Testing

Testing sharpened the details.

Prototype walkthroughs surfaced four targeted improvements that made the data more legible and the interface easier to act on.

1

Better affordances for status numbers

Made participation counts more legible and easier to interpret at a glance.

2

Numbered each Benchmark in a Series

Reinforced sequence and made it obvious how many had been administered.

3

Updated performance charts

Redesigned to read less like progress bars and more like comparative data.

4

Low participation alerts + email

Admins who rarely logged in would still be alerted to underperforming Benchmarks.

08 / 13
Handoff

Spec'ing complexity.

A thorough handoff gave engineers everything they needed: interaction specs, Figma prototypes covering each Benchmark state and edge case, and the low-participation email copy that admins would rely on when they weren't actively in the product.

I kicked off development with a team presentation, then fielded edge cases on Slack and in GitHub as they arose.

09 / 13
Results Page

Making school comparison effortless.

With the Dashboard in development, we turned to the Results page. Admins needed to compare school performance at a glance, but many didn't even realize the capability existed. I iterated from inline bar charts, to charts above the table, to the final horizontal layout that made comparisons obvious rather than something you had to squint to find.

10 / 13
Data Hierarchy

Tell the story, not just the data.

Customer Success flagged that overall scores were being overshadowed by individual skill scores. Some admins didn't even notice the overall score was available. We were providing data but not telling a story with it.

Research also revealed a need for filtered views: data counted only from students who had taken both Benchmarks in a comparison, ensuring growth numbers were apples-to-apples.

The redesigned Results page led with overall performance, added comparative filters, and was modular enough that efficacy-specific data could be surfaced or hidden based on future product decisions.

11 / 13
Creation Flow

Reducing noise upstream.

Better views only help if Benchmarks are set up correctly. An audit of the creation flow turned up a glut of inappropriate content, unlabeled AI assignments, inconsistent editing, and buggy date selection. Tightening guardrails (required time between Benchmarks, matched grade levels, consistent class assignments) meaningfully improved outcome quality.

12 / 13
Outcomes

What shipped, what moved.

72%

increase year-over-year in Benchmarks using the Grading Assistant feature

17%

increase in teacher adoption of Benchmarks sent out by admins

increased confidence from Customer Success teams in recommending Benchmarks to districts

With the core admin experience stabilized, the next phase focused on expanding Benchmark capabilities to include writing-focused assessments and deeper instructional recommendations.

13 / 13
1 / 13