How can cognitive accessibility simplify workflows?

API deployment was hindered by a complex technical infrastructure with confusing processes and cumbersome documentation. Cognitive accessibility practices provided the design insights to improve usability for everyone.

*Note: anonymized images and content shared with permission of my immediate supervisor.

Product
API Deployment Wizard

Client
Fortune 50 retailer

Timeframe
2 months

My role
Lead UX Designer & Researcher

Research methods
Ethnography, prototyping, usability testing

Participants
20+

Tools
Teams, Figma

The opportunity

For many products and organizations, APIs are the way customers buy and use products, whether those products are books, groceries, clothes, or even data or insights. For example: a shopper on a mobile app searches for a product, and an API retrieves the search results that are displayed to the shopper.

Established processes and reusable assets allowed one organization to accelerate deployment time within an incredibly complex cloud, network, and security infrastructure. However, process and documentation was difficult to navigate, understand, and apply, and frequently outdated. Mistakes were frequent, frustration was high, and the time-to-market was slowing down.

Additionally, tech workers are more likely to be neurodivergent — even beyond 50%. Accessibility practices, particularly cognitive accessibility, could provide the design insights to simplify the experience for not just for engineers that need cognitive accessibility, but for everyone.

objective

Accelerate API deployment with an easy-to-use and easy-to-understand deployment wizard, designed with an accessibility-first approach.

business value

A streamlined and simplified wizard can accelerate API deployment and reduce mistakes, leading to a faster time-to-market, happier customers, and ultimately more revenue.

The work

A kickoff meeting with developers and the project / product lead established the problem statement, project goals, and timelines. I scoped several phases of research based on the timeline and project risks: initial research, design, testing, and refinement.

research questions

  • What makes deploying an API so challenging?
  • What workarounds to engineers use to deploy an API?
  • What parts of the process are confusing?
  • What accessibility challenges, particularly cognitive accessibility, were most prevalently impacting the deployment process and tools?

Interviews

For initial discovery, I conducted 4 semi-scripted interviews with engineers and engineering leads. The participants indicated that the labels and terms were confusing, redundant work required the engineers remember a lot of inputs and selection, and that the documentation was difficult to navigate and understand.

Additionally, participants indicated they would create their own reusable scripts and templates, and share them with other developers. This explained why some of the mistakes in APIs were so prevalent — engineer workarounds had essentially made the errors “go viral” within the organization.

Outlining accessibility requirements

WCAG 2.2 AA conformance was established as acceptance criteria. However, this project needed to rely on a design system with many non-conformant components and patterns.

A plan was established to document WCAG failures resulting from the design system components, and provide those to the design system team as product requests.

Even with the design system limitations, there were many elements of accessibility we could control, particularly cognitive accessibility. In addition to WCAG standards, W3C document Making Content Usable for People with Cognitive and Learning Disabilities provided user stories and design objectives to apply as requirements.

Workflow mapping

The primary step of the design phase was to map out the workflow, and particularly the inputs needed from engineers, such as repo names, workflow spaces, paths, project and team leads, business units, etc.

Once these inputs were identified, they were grouped into logical categories. Redundant fields were removed. Every field’s necessity was questioned, at times requiring discovery across systems and teams to find out if and how the input was used.

About 50% of the “necessary” inputs were able to be removed. The remaining were arranged into a logical, linear flow, with branching paths identified and mapped.

Initial design

The workflow was translated into a Figma prototype with accessibility annotations (using the CVS Figma annotation kit), with the design emphasizing the following goals:

Set expectations and be predictable

A clear introduction to the wizard helped engineers understand what the wizard would do and what they would need to complete the wizard. Additionally, the stepper component provided visibility into the wizard progress and remaining steps.

Focus attention

A content overlay component was used with a stepper workflow. Early design testing found that engineers found the interface simpler with the fields spread across 4 steps plus a final review step.

Simplify language

Business jargon in labels was replaced with terms more familiar to the engineers. Help tooltips provided a shorthand definition along with where the engineer may find the information. Additionally, a definitions page provided an searchable and sortable table with definitions, explanations for how inputs are used, and more details on how to find the information needed.

Reduce and guide choices

Where possible, fields were autopopulated based on user profiles. Where appropriate, fields were defaulted according to best practice options. Dropdown menus helped engineers understand fields by providing options.

Meaningfully group fields

Fields were grouped into steps that represented meaningful categories, such as information about the user, business context, technical details, and information for the end-user of the API. These groupings helped engineers understand the definition and purpose of the fields.

Be consistent

UI consistency from step to step made it easier for engineers to understand how to navigate the wizard and where to find help.

Make help easy

Bringing documentation directly into the workflow limited task switching and made it easier to complete the field.

Prevent lost work

Sometimes engineers may not have the information they needed. Autosave prevented them from needing to re-enter information and remember prior responses.

Assist with errors

Error detection and suggestions made it easy to correct issues. Additionally, the autopopulated and default fields helped reduce some of the most common errors uncovered during the initial discovery.

Clarify next steps

After completion of the wizard, engineers were clearly guided with next steps so predictability and usability continued even beyond the wizard interface.

Usability testing & refinements

Usability testing was both formative and summative. Formative usability testing identified minor enhancements to simplify language, and to improve predictability both before and after the wizard with a clear list of what was built by the wizard. Refinements were implemented into the design prior to launch.

Summative usability collected several key metrics: time-on-task, task success, the UX-Lite, and the Single Ease Questionnaire (S.E.Q.). In order to establish clear benchmarks, the wizard was compared to the existing process to deploy an API using a within-subjects design. Participants were asked to deploy an API using both the existing process and the new wizard (the ordering of the tasks was randomized). Confidence intervals were calculated for the metrics for each process as well as for the difference between the two.

See results below under outcomes.

Accessibility testing & remediation

Accessibility was tested against WCAG 2.2 AA standards manually and assisted with the WAVE and ARC Toolkit testers. A significant number of issues were identified with the design system components and patterns, and were catalogued and submitted to the design system team. The remaining issues were easily addressed within a one-hour pairing session with a developer (for example, empty buttons / links, adding labels to fields, proper heading and list structures, appropriate ARIA).

Final design

A series of Figma prototype screens illustrating the overall flow.
The overall workflow captured error states, autosaving features, and success and failure results screens.
A prototype screen with a bulleted list above a "start the launchpad" button.
The wizard starting screen clearly explains what an engineer will get.
A prototype screen showing a fairly small form with tooltips and help materials, and a stepper progress indicator.
The content overlay focuses attention on the form, and the stepper provides transparency into the overall workflow. Step descriptions, field labels and descriptions, and easy to access help materials make it easy to get started.
Two prototype screens, one showing a form field with business inputs, and the second displaying alternative states that appear based on user inputs.
Engineers described the Business Context section as the biggest improvement through simplified language, field descriptions, and auto-populating fields.
Two prototype screens, one showing a form field with technical and networking inputs, and a second showing new fields and new default selections based on user input.
The build details section simplified networking and security, two of the most complicated (and error-prone) parts of the previous process. This wizard guided engineers toward proper security protocols through progressive disclosure and field defaults.
One screen displaying fields to provide an API name and description for end-users.
The wizard made it easy to set up the documentation with the minimum fields needed for the developer catalog.
Two screens of a review and final submit screen, with the second providing error guidance.
Any good form has an opportunity to review and correct mistakes. Once the wizard began deployment, making changes would be difficult. A final review page summarized selections and provided error guidance for missing fields.
Screen showing a project details page, outlining what's being built and what the user should do next.
One of the biggest challenges with the original process was “What’s next?” The process provided no indication of whether or not everything worked, what all the engineer got, where to access everything, and what to do. The wizard finished by clearly indicating to engineers the status of each thing the wizard was doing, what they needed to do next (including security protocols), where they could find everything, and what to do if they needed changes.
Two screens showing final states for successful or failed deployments, with indicators for what the engineer should do next.
Nothing works perfectly, and the post-deployment page provided clear indications of success and failure, and in the event of a failure what specifically failed and what to do about it.

The outcomes

2x faster

vs. the existing process

66% improved usability

(UX-Lite)

“Excellent” –
“Best imaginable”
SUS rating

(projected from UX-Lite)

55% improved task difficulty

(Single Ease Questionnaire)

Reflections

This project was an exciting way to demonstrate the value of accessibility to an organization.

The end result may just seem like “good UX” — after all, who doesn’t benefit from simpler workflows?

But the approach, getting to the simpler workflows, was what differentiated this project. Cognitive accessibility was the critical practice that made API deployment so much easier for everyone (not to mention a population that likely over-indexes as neurodivergent).

Challenges

One of the biggest challenges was in determining which fields were actually necessary. Finding how the data was used, and where it went, was often a “wild goose chase”, requiring tracing social networks as each person told us that someone else might know the answer.

Also, knowing which accessibility issues we could control and which we couldn’t was a challenge given the design system limitations. It often took collaboration between myself, the wizard developers, and the design system team to identify what we could control and what we couldn’t.

If I could do it again…

While I did provide accessibility annotations, I didn’t do it for every screen because I was pairing so close with the developers building the wizard. However, when I got pulled onto other projects simultaneously and my time to pair became limited, we started seeing some mistakes. If I could do it again, I’d lean more into accessibility annotations even just for the practice of educating of others and demonstrating a status quo.

Back to my portfolio