Clarity in a complex learning system

Leading a research-driven IA redesign to align system structure with user mental models and improve trust in system state

MY ROLE

Product Designer

TIMELINE

March 2026 – Present

TEAM

1 PM, 3 Engineers, QA

IMPACT

Introduced a user research process that informed the tool direction

Context

Workhub's internal Online Training tool is used by Content and Customer Success teams to build, manage and evaluate training courses. Over time, the system became increasingly layered due to its legacy structure and added functionality.

Users were also working across both internal and public-facing tools, depending on which workflows were easier to complete.

Problem

The tool was structured around system and backend logic rather than how users think about building and managing courses.

This created friction across core workflows:

  • Difficulty navigating between courses, lessons, versions and quizzes

  • Buried performance insights like quiz success rates requiring manual investigation

  • Inconsistent understanding of where changes should be made

  • Uncertainty about whether actions like saving or updating had actually succeeded

  • In some cases, users bypassed the internal tool in favour of the public-facing product

Research & insights

I introduced the first user research process for the team, including stakeholder buy-in, interview planning, execution, and synthesis. I leveraged AI to help guide me through user interview plans and draft initial questions, which I refined based on the specific workflows I tested.

Key insights from 6 user interviews:

  • Navigation breakdowns

    • Users were unclear where they were within the system hierarchy

    • Course, lesson, version, and quiz relationships were difficult to follow in practice

  • System state uncertainty

    • Users were often unsure whether actions like saving or updating had succeeded

    • Inconsistent feedback (toasts, errors, save states) reduced confidence in the system

  • Mental model mismatch

    • The system structure reflected backend logic rather than how users think about building courses

    • Terms like “versions” and “quiz assignment” did not match user expectations

  • Hidden performance insights

    • Key metrics like quiz success rates existed but were buried several layers deep

    • Accessing them required manual investigation rather than being visible in context


Quotes from participants:

"Adding quiz questions is confusing. I have to jump back out a layer to add them and then assign them to a lesson version.”

"I'm not even sure why we need lesson versions.”

"The course feedback is buried too deep for me to bother looking at it. I usually only look if a customer reaches out to the customer success team or if we're updating it."

"Sometimes when uploading a file it will give an error but the upload itself always goes through."


The core issue was not just navigation complexity, but a breakdown in system transparency and alignment with user mental models. Users couldn’t reliably understand what the system was doing or whether their actions had succeeded.

Key design decisions

  • Restructured the information architecture (IA) to match user mental models
    Reorganized courses, lessons, versions, and quizzes to reflect how users think about building content, rather than backend system structure.

  • Clarified versioning and quiz logic
    Simplified confusing overlaps where quiz questions could be assigned in multiple places but only edited in one, creating clearer ownership of where changes happen.

  • Introduced variants
    Broke out language and file quality differences from “versions” to reduce overloaded structures and make content states easier to understand and manage.

  • Using AI to pressure-test
    Testing problem framing, refining labels and statuses, helping ensure clarity and consistency.

What I'm testing

I'll be conducting usability tests soon to determine:

  • Whether the new IA reduces navigation confusion

  • Whether surfacing insights earlier improves issue detection

  • Whether clearer system feedback increases confidence in actions

Impact

This is currently in prototyping and validation stages.

Stakeholders are aligned on the redesigned IA direction

Early internal reviews show improved clarity in navigation and structure

Usability testing is planned to validate system understanding and feedback clarity

Key takeaway

This project reinforced that usability issues are often not just navigation problems, but breakdowns in how users understand system structure and state. Designing for clarity required aligning the information architecture with user mental models and ensuring system feedback is always visible and trustworthy.

I'm also incorporating how I use AI in my design process such as refining system structure and challenge my thinking, while mintaing ownership over decisions and design direction.

Amanda Wondergem

amandawondergemdesign@gmail.com

Calgary, Alberta, Canada

© 2026 All right reserved

Amanda Wondergem

amandawondergemdesign@gmail.com

Calgary, Alberta, Canada

© 2026 All right reserved

Amanda Wondergem

amandawondergemdesign@gmail.com

Calgary, Alberta, Canada

© 2026 All right reserved