
THE 5 NO’s
Prologue
NO clear method to design for AI.
NO predictions that capture evolving nature of AI.
NO humans in the loop. Mismatch in agency and automation.
NO human-centered AI. Focus is on accuracy of outcomes.
NO emphasis on contextual values (trust, human agency).
Evolving AI has been elusive to design for. We propose a contextual model finding an overlap between human expectations with machine’s capabilities to build trust.
PROBLEM
How do you design for AI considering its probabilistic and adaptive behavior, finding a sync between human expectations and evolving AI capabilities?
SOLUTION
We took a human-centered approach, framing design methods on transparency and trust. Measuring the effectiveness through task load.
IMPACT
After conducting a seminar study, we observed a 11% decrease in workload among our participants, measured via NASA-task load indexing protocol.
KEY DECISIONS
Narrowing down focus on the right problem, brainstorming on what values are key for human-AI interactions. Identifying the nuances of probabilistic AI.
MY ROLE
Lead UX Researcher, synthesized research questions through literature review. Designed research study for concept validation, along with identifying research protocols.
MY TEAM
1 Lead Researcher (Me), 3 Researcher, 1 Designer, 1 Data Scientist, 1 Supervisor
Human Expectations and Evolving Capabilities
Finding the right balance between human expectations and machine capabilities enables designers to find optimal solutions when designing human-AI interactions (for AI). This project focuses on designing for AI.
Design FOR AI
Designing for AI focuses conveying AI capabilities and finding a fit with human expectations.
Designing with AI focuses on using AI’s generative capabilities to expedite the design process.
After breaking down AI complexity, it becomes imperative to understand human expectations of AI. For this study, I focus on level 3 and 4 systems since they are the most complex to design for.
1
Simplest system among the 4 levels of complexity. Probabilistic systems have fixed number of outputs.
2
Adaptive systems, although a bit complex compared than previous, identifying the right demographics enable the designers to prototype effectively.
3
Evolving Probabilistic Systems
Evolving probabilistic systems is where the problem arises. The capability of the system changes due to the inclusion of a dynamic dataset.
4
Evolving Adaptive Systems
Evolving probabilistic systems is where the problem arises. The capability of the system changes due to the inclusion of the dynamic dataset, making it difficult to identify the right scenarios.
Primary Research
I conducted primary research which included interviewing people from academia, primarily professors along with industry professionals actively working with AI.
12
Academia
We interviewed many professors and people working in academia to understand their priorities when it comes to understanding human-AI interactions.
04
Industry professionals
In addition to academicians we also interviewed industry professionals to understand their challenges when designing for AI.
Secondary Research
In order to add richness to our findings, I along with my team also conducted literature review of several research papers spanning across AI, perception, and neural networks.
23
Research papers
We interviewed many professors and people working in academia to understand their priorities when it comes to understanding human-AI interactions.
4 Thematic Diagramming
The results of this study are based on qualitative coding shedding light on principles of human-AI interactions. Transcripts were generated from 6 hours of recorded data.
Emerged Themes
31
Explainability
I identified four emerging themes along with their frequency of occurrence through interview transcripts and secondary research. This helped us in identifying the key factors for AI design.
User Quotes
Furthermore, I have illustrated few user quotes which reflect the sentiment of our participants and their connections with the identified themes.
Trainability
“Its been good for most tasks, but AI still isn't good at some tasks and with heavy usage the AI agent tends to hallucinate.”-P16
Explainability
“The information provided by AI is not always accurate and requires me to go back and review it.”-P23
Deterministic UX Methods
“AI has helped with my workflow by improving and speeding out my research and design flow.”-P8
Improvements
“AI enhances efficiency pushing people's boundaries, but AI being trained with copyrighted content makes it unfair for creators.”-P4
Pre and Post Seminar Survey
We received a good response on our post and pre seminar survey. The surveys were designed to understand the change in workload pre and post the seminar, to understand the effectiveness of the qualitative values (transparency, autonomy, and ethics).
Pre-seminar survey
32
Post seminar survey
12
Conducting the Seminar
I hosted the seminar along with my team to capture the participants’ reaction, and understand how the qualitative values affects the participants trust in AI.
Rule based Simulators: Wekinator
The last phase of the seminar was used to explore the parallels between rule based simulators and AI prototyping, as it highlights the increasing complexity/unpredictability of models post training.
Understanding Task load
Task load focuses on parameters like cognitive engagement, latency, and physical load which directly links to the user engagement with the concerned application.
Increased Engagement
Mean workload changed from 5.4 to 6.4, indicating that that the average user was more engaged when using AI due to the exploration of qualitative factors like transparency, autonomy and AI ethics.
5.4
Pre-study
6.4
Post study
Higher Workload
High Use
Design vs Development
Ease of Understanding
Trust in AI
Trust in AI had a relatively moderate average (4.75) with broad spread (SD = 1.93), indicating mixed perceptions even within a technically literate sample, trust-building mechanisms should not be assumed and may need contextual adaptation.
Emphasis on Contextual Values
In addition to rule-based simulation, contextual values provide richness to human experiences in the design of human-AI interactions, as evident through thematic diagramming. The excerpt frequency is provided below.
agency
51
TRANSPARENCY
18
task load
41
accountability
37
Emphasis on Qualitative Research
The qualitative factors like accountability or transparency are unique to user demographic and characteristics of the task, therefore, qualitative research is imperative, and provides holistic and rich insights in-tandem with quantitative research.
Measuring Trustworthiness
After identifying the change in task load post seminar. It is vital to conduct further study on how qualitative factors like agency and transparency can affect trustworthiness through task load.
Task Load and Trust
Furthermore, as it was identified that a higher task load is associated with an increased trust, but an extensive study is required to validate these exploratory findings.
from data to PATTERNS
I learned how meaningful patterns can be drawn from a large pool of data through qualitative and quantitative methods.
insight synthesis
Patterns helped me identify emerging themes and led to the synthesis of insights, highlighting the role of trust in AI use.
Conducting seminar
Conducting the seminar was an overwhelming experience but it taught me how convey esoteric ideas to a diverse audience.




