top of page
Overview

AI-Detective

This project explores how user research methods support iterative divergence and synthesis to uncover valuable design insights. The final output is a mini-game prototype that helps college students better understand AI bias and use generative AI more mindfully.

MacBook-Pro-16.png

Timeline

Jan 2025 - May 2025

Project

Course project

Responsibilities

UX Research
Prototyping
Interaction Design

User Interface Design

Tools

UX Research methods
Figma

 

Challenge

AI bias was a relatively unfamiliar topic for us as researchers at the start. Additionally, our user research showed that most college students lack strong motivation to learn about it. The challenge was to balance raising awareness while keeping users engaged without causing frustration or boredom.

Opportunity

Although we were new to AI bias, prior user research from our instructor provided valuable insights into the topic and user attitudes. Through iterative ideation and testing, we gained further user insights that helped ensure the practicality of our final design.

OVERVIEW

The Process

1

Preliminary research

Background Research

Data Analysis

Heuristic Evaluation

Usability Testing

2

Synthesis

Walking the wall

Reframe & Define

3

Contextual Research

Preparation

Conduct & Analysis

4

Modeling

Affinity Clustering

Modeling

Research Report

Speed Dating

5

Deliverable Reflection

Assumption Artifacts

Elevator Pitch

Course Relection

PRELIMINARY RESEARCH

Background Research

01. Overall approach

In the background research phase, I conducted 2 experiential and 5 informational studies, from which I synthesized six key insights and completed a Context & Change Worksheet. Click here to view the full research report.

Frame 3.png

02. Insights

- Feedback could partly improve results

Feedback can indeed help optimize the generated results, but the effectiveness and ease of providing feedback are relatively poor.

- Lack of tools

Users need more accessible tools to help identify potential biases during their usage and provide quick feedback.

- Showing AI's thinking and sources can aid in spotting bias

Displaying the thought process and information sources behind generative AI can help users more easily identify potential bias issues.

- Frequent audits reduce ethical problems

Regular audits can help increase awareness of bias issues and offer valuable data to enhance generative AI.

- Marginalized group

Currently, generative AI's focus on bias issues within marginalized groups primarily targets the LGBTQ community, people of color, and women, with a lack of attention to disabled communities.

- Stereotypes in bias

Bias discussions often focus on race and current topics, overlooking other biases such as cultural and language bias.

03. Context & Change Worksheet

Frame 4.png
Frame 5.png

PRELIMINARY RESEARCH

Data Analysis

01. Oveview

We analyzed survey data collected by the WeAudit team, including our instructor Dr. Hong Shen. The survey (N = 2,197) explores how individuals’ identities, experiences, and levels of knowledge influence their ability to detect harmful algorithmic biases, particularly in the context of image search. Our analysis focused on three primary identity dimensions—race, sexuality, and gender—along with intersecting factors such as education level. Click here to view the full data analysis report.

02. Visualization

I was responsible for researching how sexual orientation influences the perception of AI bias within our team. I ultimately chose to use a bar chart for visualization, as it clearly illustrates the differences in opinions among various groups.

Screenshot 2025-07-28 at 4.13.27 PM.png

Overall, regardless of sexual orientation, approximately 50% of people perceive the content as harmful, while the other 50% do not. Additionally, the heterosexual group perceives the least harm from the image, with only 27.73% of them considering the content to be harmful.

03. Takeaways

Insights

  • People are more likely to perceive biases in areas that pertain to themselves.

  • People tend to recognize bias when there are familiar over-representations or under-representations present within the results of a query. 

Hypothesis

  • Certain demographics of people identify specific algorithmic biases (ie. racial, gender, etc.) in GenAI more accurately than their peers.

  • Certain algorithmic biases (ie. sexuality) in GenAI are more difficult for people to identify, necessitating different methods for their identification.

PRELIMINARY RESEARCH

Heuristic Evaluation

01. Oveview

We conducted a heuristic evaluation of the WeAudit TAIGA tool—a platform designed to help users detect algorithmic bias—using Nielsen's Ten Usability Heuristics. Click here to view the whole team heuristic evaluation and click here to read the heuristic report.

02. My Heuristic Evaluation Findings

Frame 6.png

PRELIMINARY RESEARCH

Usability Test

01. Oveview

Following the heuristic evaluation, the goal of our usability testing was to evaluate WeAudit’s intuitiveness for first-time users and the speed at which they understood its core features. We employed the Think-Aloud protocol to capture user feedback. Click here to view the full usability testing plan and click here to see the whole usability test report.

02. My Usability Test Findings

Screenshot 2025-07-28 at 4.47.57 PM.png
Synthesis

SYNTHESIS

Walking The Wall

We conducted Walking the Wall to re-immerse ourselves in the data and analyses we had completed throughout the project. We walked through selected outputs from our background research, findings from our data analysis, and high-level insights from our heuristic evaluation and usability testing. Click here to see the detail.

Screenshot 2025-02-28 at 11.02.00 PM (1).png

SYNTHESIS

Reframe & Define

01. Reframing Activity

We chose the Reverse Assumptions method because we shared similar mental models of the generative AI bias problem space. To break out of our conventional thinking, we believed that intentionally challenging and flipping our assumptions would help us generate a broader and more diverse set of potential approaches for our project.

Frame 7.png

Our future research will focus on enhancing everyday users’ understanding of generative AI bias, increasing transparency within AI systems regarding potential biases, and creating more opportunities for users to provide detailed feedback on bias issues.

02. Define The Project

Frame 8.png
Contextual Research

CONTEXTUAL RESEARCH

Preparation

01. Overview

Before conducting the experiment, we first drafted a research guide, which included the introduction, consent script, consent form, research goals, our chosen methods (selected from think-aloud, contextual inquiry, and directed storytelling), the experimental procedure, and interview questions.

To ensure the feasibility of the study plan, we conducted a pilot test and refined the research guide based on the feedback and results.

Frame 9.png

02. Draft Research Guide

We chose the think-aloud protocol to better understand users’ thought processes when interacting with a gen-AI tool. By placing users in realistic scenarios, we can closely observe how they respond to current gen-AI technologies. The main challenge will be designing effective scenarios and asking timely, insightful questions. Click here to see full draft research guide.

03. Pilot Test

The structure of the experiment is well-designed and facilitates strong participant engagement. However, some questions were too general, resulting in superficial answers, and should be removed. Additionally, allocating more time to interviews rather than the Think-Aloud Protocol is recommended, as interviews tend to provide deeper and more meaningful insights. Click here to see feedback from pilot test.

Frame 11.png

03. Final Research Guide

In the final research plan, the experiment was optimized based on feedback from each participant in the pilot test, with a greater focus on users’ feedback after using the AI tool. Additionally, qualified participants were prepared for the formal experiment.

Our goals are to better understand:

- How familiar users are already with potential bias in gen-AI content

- What aspects of gen-AI bias users may be unaware of

- Users’ ability to label/identify gen-AI bias

- Strategies to improve users’ ability to identify gen-AI bias

Core questions:

- How well can everyday users identify algorithmic bias already?

- What factors affect a user’s ability to detect algorithmic bias?

- What are potential ways to improve users’ ability to identify bias?

CONTEXTUAL RESEARCH

Conduct & Analysis

01. Conducting Session

Based on the planned procedure, I carried out the formal experiment and summarized the following key findings.

1. Users prioritize output accuracy over fairness or diversity

2. Bias is acknowledged but not internalized as a problem

3. Bias perception depends on result, not the prompt itself

4. Limited motivation to change or question the system

5. Curiosity about AI mechanisms could be a potential entry point

02. Interpretation Notes

Upon completing our individual experiments, we held a group discussion to interpret our notes and align our findings. We then synthesized our insights into a shared spreadsheet for further analysis. Click here to see the spreadsheet.

image.png
Modeling

MODELING

Affinity Clustering

01. Create and Group Notes

We used FigJam to synthesize and group the notes from our Interpretation Notes. From this process, we identified different themes and represented them using color-coded labels.

M6.3 _ Deliverable F_ Affinity Clustering-1.png

02. Top-Level Themes

Through iterative clustering, we ultimately organized the notes into three main themes:

    - Initial expectations and beliefs about generative AI systems and algorithmic bias

    - Processing and perceiving algorithmic bias in generative AI

    - Approaches to increasing awareness of bias

MODELING

Modeling

01. Personas

Modeling is another form of research synthesis that allows us to explore our data from different perspectives. We chose to create personas as a way to reframe our notes, because although we had decided to focus on college students as our target users, we observed significant differences in their levels of understanding and attitudes toward generative AI. Developing personas helped us capture these variations and better tailor our design decisions to distinct subgroups within the college student population.

image.png
image.png
image.png

02. Journey Mapping

In addition, we chose to create a journey map because we found that users’ motivation plays a crucial role in reducing AI bias. This tool allowed us to identify how motivation shifts across different stages of the user experience, revealing new design opportunities along the way.

image.png

MODELING

Research Report

After conducting affinity clustering and analyzing our data through personas and journey mapping, we gained more actionable insights and further refined our research goals and methods. Click here to see the research report.

High Level Insights:

1. A users’ ability to identify bias is inherently linked to their understanding of discrimination, bias, and prejudice.

2. There is a common belief that AI tasked with menial or non-creative tasks won’t produce biased consequences.

3. User could benefit from a better understanding of the development and operations behind Gen-AI models when identifying algorithmic biases.

4. Distrust arises in generative AI, especially on human-centric topics, due to lack of transparency, single-perspective responses, and inherent skepticism.

5. Bias detection in generative AI is influenced by external factors and improves when bias is a focal point of discussion.

MODELING

Speed Dating

01. Overview

In the Speed Dating process, we validated user needs and values and identified conceptual risk factors through a series of steps, including Walk the Wall, Crazy 8s, concept selection, storyboard creation, and Speed Dating sessions.

02. Walk the Wall

We conducted another round of Walk the Wall to organize additional notes and experimental data, resulting in a list of insights, user needs, and questions.

image.png

03. Crazy 8s

To generate ideas that focus on the needs and values of future users based on our research, we conducted a Crazy 8s session. We rapidly sketched storyboards to communicate the user insights we uncovered and to promote discussion around people’s desires, values, and needs.

User Needs:

1. Desire to learn more about AI bias

2. Desire to learn inner workings of AI systems (including source of results)

3. Multiple outputs in gen-AI results

4. Examples to demonstrate AI bias

5. Quick, intuitive, and engaging delivery method

image.png
image.png
image.png
image.png
image.png

03. Speed Dating

Starting from the needs identified during the Crazy 8s session, we designed safe, risky, and crazy-risky concepts for each need and visualized them through storyboards. We then conducted Speed Dating sessions with users to gather feedback on these concepts.

Click here to read the full Speed Dating report.

image.png
image.png
image.png

New Design Opportunities:

- More diverse views for displaying AI results

- Provide users with options to modify the details of AI-generated results.

- A simple and quick mini-game about AI bias.

Common Misunderstandings:

- The extent to which users need more bias education

- Distinction between incorrect answers and biased answers

- Bias reported by the user vs. the platform

Deliverable

DELIVERABLE

Assumption Artifacts

01. Assumption Artifact Proposal

Each member of our team individually envisioned an Assumption Artifact to help address our current research question. From our own perspectives, we identified the riskiest assumption, honest signal, measures, as well as the participants and setting.

My proposal is to add a bias-check button within generative AI systems. By clicking this button, users can review their generated content and clearly identify different types of potential bias. I believe this approach strikes a balance between supporting user motivation and ensuring functional effectiveness. Click here to see my proposal.

MacBook Air - 18.png
MacBook Air - 19.png
MacBook Air - 20.png

02. Assumption Artifact & Testing Plan

Our team consolidated all of our individual proposals and selected one direction to move forward with based on a group vote. We then created a prototype to conduct further testing. Click here to read the testing plan.

We created a low-fidelity Figma prototype of a mini-game embedded in ChatGPT. Before accessing the platform, users play as a crocodile investigator assessing fairness in marsh entry decisions. The goal is to spark interest and help users better understand bias in generative AI.

Frame 2.png

03. Test Result

We used both quantitative and qualitative methods to analyze the data. We tested with undergraduate students using a think-aloud protocol and collected follow-up data via a Google Form. Click here to see the test result.

image 6.png
image 7.png
Findings.png

Users found the game easy but lacking in nuance, and some were confused about its context and feedback design. Moving forward, we plan to improve clarity, add feedback regardless of answer correctness, provide game context upfront, and introduce more complex bias scenarios. These insights will guide our next design iteration.

04. Elevator Pitch

College students—especially in the age of generative AI—value speed and efficiency. When learning something new, they often turn to gamified platforms like Duolingo rather than traditional textbooks. To support this behavior, we created a short, engaging educational mini-game embedded within ChatGPT to help everyday college students improve their foundational understanding of generative AI bias.

Screenshot 2025-07-30 at 11.31.01 AM.png
Reflection

REFLECTION

Course Reflection

As a designer, I learned many valuable user research methods through the UCRE course. For example, the think-aloud protocol gave me deeper insights into users’ thoughts on our prototype, while speed dating proved to be an efficient method for both generating and evaluating design solutions.

This class really helped me see the broader value of storyboards. Previously, I only used them to present design ideas, but now I understand how they can also be used in speed dating sessions to effectively evaluate those ideas.

bottom of page