AI-Enhanced Assessment CRUD

Designing a Scalable, Human-in-the-Loop System for AI-Generated Assessments

UX CASE STUDY

BEFORE

AFTER

Overview

Udacity’s enterprise business relied heavily on third-party assessments and manual, engineering-driven workflows to create and manage in-house assessments. As demand increased—and AI-generated content became viable—these processes failed to scale.

I led the design of an AI-enhanced assessment creation and management system within Udacity’s enterprise management platform, enabling internal teams to quickly create, review, edit, and maintain assessments—using both manual and AI-generated questions—under significant time and technical constraints.

Team
1 Product Manager, 2 Engineers, 1 Product Designer (Me)

Time
30 days

Tools
Figma, ChatGPT, Lovable, Gemini, Jira, confluence, G suite

All collaboration was conducted remotely.

Audience
Udacity Staff

Platform
Web

Key Stakeholders
Sales Solution Architects, Customer Success Operations, Customer Success Managers, Content

Why This Matters

As organizations shift toward skills-based hiring and upskilling, assessments become critical infrastructure—not just content. Yet most learning platforms struggle to scale assessment creation due to heavy engineering dependency, slow workflows, and limited quality control.

This project addressed a foundational problem: how to safely and efficiently scale assessment creation using AI without sacrificing quality, trust, or operational control.

By designing a flexible assessment CRUD system with built-in human review, this work:

  • Reduced reliance on engineering for day-to-day assessment creation

  • Enabled non-technical teams to operate independently at scale

  • Established a durable foundation for AI-generated assessments across multiple future use cases

  • Directly supported enterprise contracts and revenue timelines

At scale, this system shifts assessment creation from a bottleneck into a strategic capability, enabling Udacity to move faster as AI-driven content generation becomes table stakes.

The Challenge

Several systemic issues limited Udacity’s ability to scale assessments:

1. Engineering Dependency

Assessments and questions were manually created or configured by engineers, creating bottlenecks and slowing response to customer needs.

2. Non-Scalable Workflows

Non-technical teams (Customer Success, Content, Ops) lacked tools to independently create or manage assessments, resulting in operational delays.

3. No Quality Control Framework

There was no consistent way to evaluate question quality or performance, limiting confidence in assessment accuracy and fairness.

4. AI Without Governance

As the team explored AI-generated questions, there was no structured human-review process to ensure pedagogical quality, relevance, or trustworthiness.

At the same time, contractual commitments required delivering a functional solution in weeks, not months.


Design Objective

Design a scalable assessment creation system that:

  • Enables full CRUD functionality for assessments and questions

  • Supports both manual and AI-generated content

  • Empowers non-technical staff to operate independently

  • Introduces human-in-the-loop review for AI-generated questions

  • Can be delivered quickly within legacy platform constraints


My Role

Lead Product Designer (IC Lead)
I owned the end-to-end UX strategy and execution for this initiative.

Responsibilities

  • Experience definition under tight constraints

  • Workflow and systems design

  • AI-assisted ideation and prototyping

  • UX/UI design within an existing enterprise platform

  • Cross-functional alignment and handoff


Constraints & Scope

This project was intentionally execution-heavy and time-boxed.

Timeline

  • ~30 days end-to-end

Key Constraints

  • Existing legacy assessment experience partially built

  • New AI APIs and databases under active development

  • Minimal time for foundational refactors

  • Q3–Q4 enterprise contracts dependent on delivery

Scope Expansion
Originally focused on AI-driven question creation, early discovery revealed the core assessment experience was incomplete. To support AI responsibly, we expanded scope to design foundational assessment CRUD—ensuring both manual and AI-generated workflows could scale.

This decision traded visual polish for long-term system viability.

Approach

1. Rapid System Understanding

  • Audited the existing assessment experience and internal workflows

  • Reviewed an engineering-built Python prototype for AI-generated questions

  • Partnered with engineers to understand backend constraints and data models

2. Experience Principles

  • Non-technical users should operate independently

  • AI should accelerate creation, not bypass human judgment

  • CRUD workflows must support iteration, not just creation

  • The system must be extensible to future assessment types

3. AI-Enhanced Ideation

I used AI tools to:

  • Rapidly explore workflow variations

  • Generate UI and copy concepts

  • Stress-test edge cases in AI-generated content review

This allowed faster convergence while maintaining design rigor.


The Solution

I designed a flexible assessment management system embedded within Udacity’s enterprise platform.

Core Capabilities

Assessment CRUD

  • Create, edit, publish, and manage assessments without engineering support

Question Management

  • Add, edit, replace, and remove questions

  • Support both manual and AI-generated content

AI-Generated Question Flow

  • Guided AI prompt inputs

  • Clear distinction between AI-generated and approved content

  • Human review and approval before questions become active

Quality & Control

  • Structured review points for AI-generated questions

  • Ability to iterate on low-performing or low-quality content

This approach balanced speed with governance—critical for trust in AI-generated assessments.

Lovable ideation example screens

Impact

Before

  • Assessments manually created by engineers

  • No scalable workflow for updates or iteration

  • Limited visibility or control for non-technical teams

After

  • Fully functional assessment and question CRUD

  • AI-generated and manual workflows supported

  • Non-technical teams enabled to operate independently

Business Impact

  • Unblocked enterprise assessment contracts for Q3–Q4

  • Reduced operational and engineering load

  • Established a scalable foundation for AI-generated assessment content

Reflection

This project required making intentional tradeoffs:

  • Speed over visual refinement

  • System integrity over incremental feature delivery

  • Governance over unchecked AI automation

The outcome was not a polished surface feature, but a durable internal system—one designed to evolve as AI capabilities mature.

More importantly, it demonstrated how AI can be responsibly integrated into enterprise workflows when paired with strong UX and human oversight.

prototype

Introduce your brand

Take a minute to write an introduction that is short, sweet, and to the point.

Sign up

Make it stand out.

It all begins with an idea. Maybe you want to launch a business. Maybe you want to turn a hobby into something more. Or maybe you have a creative project to share with the world. Whatever it is, the way you tell your story online can make all the difference.

Make It