Log in or create a free Rosenverse account to watch this video.
Log in Create free account100s of community videos are available to free members. Conference talks are generally available to Gold members.
Summary
Quantitative instruments are frequently sought because 1) they can be quickly fielded to lots and lots of people, and 2) when carefully sampled, they can be generalizable to the population of users/customers. However, because many times the focus is on speed to launch because decision-makers need results quickly, there is not much depth given to their development, nor an investigation of the validity evidence. In the session, I will share a framework that centers validity and is necessarily a mixed methods approach to research. I will also share ideas on how to scale the research over time so that findings and insights are able to be iteratively delivered to stakeholders, while also iteratively informing one another in a qual-quant research dance that brings more trustworthy, user-centered evidence to decision-makers. Finally, I will share ideas for a course I am developing for supporting qualitative researchers to become more mixed in their approach.
Key Insights
-
•
Validity in surveys is multi-faceted, relying on five evidence sources: test content, response processes, internal structure, relation to other variables, and consequences of testing.
-
•
Mixed methods combining qualitative cognitive interviews and quantitative analysis enhance survey validity and build stakeholder trust.
-
•
Breaking down survey validation efforts across multiple teams makes the process more manageable and effective.
-
•
Iterative survey development over multiple rounds helps improve instrument quality while balancing the need for timely insights.
-
•
Qualitative research plays a vital role even within quantitative validity frameworks by revealing respondent interpretation and cognitive processes.
-
•
Careful stakeholder engagement and communicating rapid but incremental insights increases buy-in for rigorous validity processes.
-
•
Survey validity is closely tied to ethical considerations, including the impact on respondents and responsible data use.
-
•
Significant product or user base changes necessitate revisiting and revising surveys to maintain validity.
-
•
Statistical methods like factor analysis and Rasch modeling help detect underlying constructs and response biases across subpopulations.
-
•
It is often necessary to accept imperfect early versions of surveys, improving them progressively while acknowledging limits to change-over-time comparisons.
Notable Quotes
"Validity is the degree to which evidence and theory support the interpretations of test scores for proposed uses."
"Qualitative research is vital for establishing validity in mixed methods because it helps us understand how respondents interpret questions."
"Surveys are products too—they need to be iteratively tested and refined."
"You don’t know what you don’t know—surveys have blind spots that qualitative techniques can help reveal."
"If you have broader research goals shared across quant and qual teams, then the overlap supports answering difficult validity questions."
"Conversations with stakeholders need to focus on delivering usable information quickly, not just on the validity process itself."
"Consequences of testing include ethical considerations about how survey responses affect user experience and product decisions."
"It’s better to partner with quantitative experts if you don’t have that expertise yourself to understand internal structure analyses."
"If the survey doesn’t work well for subpopulations, focus initially on groups where it’s reliable while you investigate others."
"Improving surveys iteratively can undermine longitudinal measures, so it’s critical to balance validity and tracking over time."
Or choose a question:
More Videos
"Each species lost is a thread untethered from the web of life that supports us all."
Alex Hurworth Bonnie John Fahd Arshad Antoine MarinDesigning a Contact Tracing App for Universal Access
October 23, 2020
"Time commitment for onboarding interns was close to 25%, maybe 50%, of at least two weeks, easing as they became familiar with roles."
Laine Riley Prokay Lisa GordonCarving a Path for Early Career DesignOps Practitioners
September 9, 2022
"Launching a design system is not a sprint, there’s no end, it’s always a continuous process."
Eniola OluwoleLessons From the DesignOps Journey of the World's Largest Travel Site
October 24, 2019
"Most strategy is done pretty poorly; it's misleading, sloppy, and often ignored after it's produced."
Nathan ShedroffDouble Your Mileage: Use Your Research Strategically
March 31, 2020
"Mobile live captions can caption any sounds around the user in real time, making it a powerful accessibility tool."
Sam ProulxMobile Accessibility: Why Moving Accessibility Beyond the Desktop is Critical in a Mobile-first World
November 17, 2022
"Foundational research answers big, nebulous questions; rapid research focuses on specific usability questions."
Feleesha SterlingBuilding a Rapid Research Program (Videoconference)
May 18, 2023
"Starting with specific, meaningful priority communities rather than homogenized personas leads to powerful breakthroughs."
Neil BarrieWidening the Aperture: The Case for Taking a Broader Lens to the Dialogue between Products and Culture
March 25, 2024
"You can’t measure long-term customer relationship value with short-term KPIs."
John DevanneyThe Design Management Office
November 6, 2017
"Small experiments that fit into existing structures make it easier to involve collaborators and reduce resistance."
Katy MogalBut Do Your Insights Scale?
March 12, 2021