Log in or create a free Rosenverse account to watch this video.
Log in Create free account100s of community videos are available to free members. Conference talks are generally available to Gold members.
Garbage in, garbage out? Measuring error rates to get ready for AI
Summary
We’re all aware of a big push to implement AI everywhere, including in the services that many of us are working on. It seems only fair to try to give the AI some good quality input in the hope of getting decent output from it. Or, being more pessimistic: we probably expect to get some level of errors from the AI, but what do we know about the error rates in what we’re putting into the AI? In this session, we will compare our ideas on identifying errors and measuring error rates, including thinking about errors in six ways: 1) Problems along the way 2) Wrong result 3) Unnecessary action 4) Delayed-impact problem 5) Non-uptake or over-uptake 6) Technology problem We’ll wrap up with “tips and next steps”: an opportunity to consider what we now need to find out or do differently.
Key Insights
-
•
Errors in data collection and user input are foundational issues that compromise AI and service outcomes.
-
•
Users often 'fudge' answers due to ambiguous questions, privacy concerns, or to achieve a desired outcome.
-
•
Non-uptake, where users abandon a form or process, is a major source of error but is rarely published or measured.
-
•
Mistakes can be categorized as problems along the way, wrong results, unnecessary actions, and delayed impact issues.
-
•
Multiple accounts creation often occurs due to users forgetting existing accounts, leading to data duplicates and service inefficiencies.
-
•
Measuring error rates is complex; different metrics (per person, per attempt, completion vs. start) yield different perspectives.
-
•
Elections provide a useful model for measuring data quality, using turnout, participation, and eligibility rates.
-
•
Data quality deteriorates over time due to changes like moving, name changes, loss of documents, or organizational restructuring.
-
•
AI initiatives can provide a compelling rationale and funding opportunity for improving longstanding data quality problems.
-
•
Frameworks like the UK Government Data Quality Framework help organizations systematically assess and address data issues.
Notable Quotes
"If we get garbage in, we get garbage out — this is true for AI as much as for surveys or forms."
"People can make all sorts of inventive mistakes on their forms that AI struggles to interpret."
"Sometimes a form forces you into a wrong answer by giving inappropriate options."
"I’ve seen people fudge their date of birth so their child can attend a summer camp they aren’t technically eligible for."
"A major error in many services is users creating multiple accounts because they can’t find or reuse existing ones."
"An error might not be immediate; data can be fine when collected but deteriorate over time and cause problems later."
"Completion rates (conversion rates) and dropout rates are simple metrics but often not tracked or shared."
"Organizations rarely know their error rates, which limits their ability to improve user experience or data accuracy."
"Linking data quality efforts to AI initiatives can help secure attention and budget for necessary improvements."
"Data quality involves accuracy, completeness, uniqueness, timeliness, and representativeness—not just error reduction."
Or choose a question:
More Videos
"This is where UX and product strategy bring huge value—defining what good means rather than leaving it to engineers alone."
Peter Van DijckBuilding impactful AI products for design and product leaders, Part 2: Evals are your moat
July 23, 2025
"The use of dovetail helps organize diverse information without losing the author’s voice or connection to original data."
Xenia Adjoubei Sean BruceEmpowering Communities Through the Researcher in Residence Program
March 29, 2023
"If you’re concerned about cybersecurity, buy a used car with fewer electronic features or one that has a strong reliability score."
James RamptonThe Basics of Automotive UX & Why Phones Are a Part of That Future
July 25, 2024
"Professional testers are very crafty; one even changed her name repeatedly to get recruited multiple times."
Lily Aduana Savannah Hobbs Brittany Rutherford5 Reasons to Bring Your Recruiting in-House (and How To Do It)
March 12, 2021
"Agents often have fascinating second or third career backgrounds that unlock deeper research insights."
Kayla Farrell Chelsey Glasson Sean Fitzell Jared LeClercWhat It's Like To Be a User Researcher at Compass
March 12, 2021
"Accessibility research is part of overall user research; including people with accessibility needs can replace some general population participants."
Kate KalcevichIntegrating Accessibility in DesignOps
September 23, 2024
"Machine learning models can see, hear, think, and act but are limited to problem-solving based on past observations."
Ovetta SampsonResearch in the Automated Future
March 11, 2022
"By embedding themselves in environments, the product team underwent a fundamental shift in perspective."
Deanna MitchellDesigning with culture: Unlocking impactful insights for Product and UX
March 12, 2025
"Knowledge systems are enterprise level applications that are repositories of a body of knowledge, usually focused on a single topic or related group of topics."
Ren PopeBuilding Experiences for Knowledge Systems
June 6, 2023