Synthesizing Design Methods for Probabilistic AI

THE 5 NO’s

Prologue

NO clear method to design for AI.

NO predictions that capture evolving nature of AI.

NO humans in the loop. Mismatch in agency and automation.

NO human-centered AI. Focus is on accuracy of outcomes.

NO emphasis on contextual values (trust, human agency).


Evolving AI has been elusive to design for. We propose a contextual model finding an overlap between human expectations with machine’s capabilities to build trust.

Summary

PROBLEM

How do you design for AI considering its probabilistic and adaptive behavior, finding a sync between human expectations and evolving AI capabilities?

SOLUTION

We took a human-centered approach, framing design methods on transparency and trust. Measuring the effectiveness through task load.

IMPACT

After conducting a seminar study, we observed a 11% decrease in workload among our participants, measured via NASA-task load indexing protocol.

KEY DECISIONS

Narrowing down focus on the right problem, brainstorming on what values are key for human-AI interactions. Identifying the nuances of probabilistic AI.

MY ROLE

Lead UX Researcher, synthesized research questions through literature review. Designed research study for concept validation, along with identifying research protocols.

MY TEAM

1 Lead Researcher (Me), 3 Researcher, 1 Designer, 1 Data Scientist, 1 Supervisor

with ai vs. for ai

1 Scoping the project

Human Expectations and Evolving Capabilities

Finding the right balance between human expectations and machine capabilities enables designers to find optimal solutions when designing human-AI interactions (for AI). This project focuses on designing for AI.

Design FOR AI

Designing for AI focuses conveying AI capabilities and finding a fit with human expectations.

Design with AI

Designing with AI focuses on using AI’s generative capabilities to expedite the design process.

the four levels

2 Breaking down the problem

After breaking down AI complexity, it becomes imperative to understand human expectations of AI. For this study, I focus on level 3 and 4 systems since they are the most complex to design for.

1

Probabilistic Systems

Simplest system among the 4 levels of complexity. Probabilistic systems have fixed number of outputs.

2

Adaptive Systems

Adaptive systems, although a bit complex compared than previous, identifying the right demographics enable the designers to prototype effectively.

3

Evolving Probabilistic Systems

Evolving probabilistic systems is where the problem arises. The capability of the system changes due to the inclusion of a dynamic dataset.

4

Evolving Adaptive Systems

Evolving probabilistic systems is where the problem arises. The capability of the system changes due to the inclusion of the dynamic dataset, making it difficult to identify the right scenarios.

scoping domain

3 Conducting the Research

Primary Research

I conducted primary research which included interviewing people from academia, primarily professors along with industry professionals actively working with AI.

12

Academia

We interviewed many professors and people working in academia to understand their priorities when it comes to understanding human-AI interactions.

04

Industry professionals

In addition to academicians we also interviewed industry professionals to understand their challenges when designing for AI.

Secondary Research

In order to add richness to our findings, I along with my team also conducted literature review of several research papers spanning across AI, perception, and neural networks.

23

Research papers

We interviewed many professors and people working in academia to understand their priorities when it comes to understanding human-AI interactions.

data analysis

4 Thematic Diagramming

The results of this study are based on qualitative coding shedding light on principles of human-AI interactions. Transcripts were generated from 6 hours of recorded data.

Emerged Themes

60

Trainability

43

Deterministic UX

44

Improvements

31

Explainability

I identified four emerging themes along with their frequency of occurrence through interview transcripts and secondary research. This helped us in identifying the key factors for AI design.

User Quotes

Furthermore, I have illustrated few user quotes which reflect the sentiment of our participants and their connections with the identified themes.

Trainability

“Its been good for most tasks, but AI still isn't good at some tasks and with heavy usage the AI agent tends to hallucinate.”-P16

Explainability

“The information provided by AI is not always accurate and requires me to go back and review it.”-P23

Deterministic UX Methods

“AI has helped with my workflow by improving and speeding out my research and design flow.”-P8

Improvements

“AI enhances efficiency pushing people's boundaries, but AI being trained with copyrighted content makes it unfair for creators.”-P4

ai seminar

5 Validating the findings

Pre and Post Seminar Survey

We received a good response on our post and pre seminar survey. The surveys were designed to understand the change in workload pre and post the seminar, to understand the effectiveness of the qualitative values (transparency, autonomy, and ethics).

Pre-seminar survey

32

Post seminar survey

12

Conducting the Seminar

I hosted the seminar along with my team to capture the participants’ reaction, and understand how the qualitative values affects the participants trust in AI.

Rule based Simulators: Wekinator

The last phase of the seminar was used to explore the parallels between rule based simulators and AI prototyping, as it highlights the increasing complexity/unpredictability of models post training.

nasa-tlx

6 Calculating Task load

Understanding Task load

Task load focuses on parameters like cognitive engagement, latency, and physical load which directly links to the user engagement with the concerned application.

Increased Engagement

Mean workload changed from 5.4 to 6.4, indicating that that the average user was more engaged when using AI due to the exploration of qualitative factors like transparency, autonomy and AI ethics.

5.4

Pre-study

6.4

Post study

‘The’ Answer does not exist

7 Results

Higher Workload

Higher Workload = Higher Engagement

Users who reported higher workload also reported higher engagement (Correlation Matrix, r = 0.944). This suggests a higher engagement effect or that deeper cognitive engagement.

High Use

β = -0.69

Frequent AI users tended to trust AI less. This implies that repeated exposure may increase critical awareness or highlight system limitations.

Design vs Development

p<0.001

Academic program significantly affected perceived workload. Software developers reported the higher NASA-TLX scores, followed by designers.

Ease of Understanding

SD=3.22

Ease of Understanding showed the highest variability (SD = 3.22). This suggests that while some users find AI interactions intuitive, others may experience significant confusion, a priority area for design improvement.

Trust in AI

Trust in AI had a relatively moderate average (4.75) with broad spread (SD = 1.93), indicating mixed perceptions even within a technically literate sample, trust-building mechanisms should not be assumed and may need contextual adaptation.

Synthesis

8 Conclusion

Emphasis on Contextual Values

In addition to rule-based simulation, contextual values provide richness to human experiences in the design of human-AI interactions, as evident through thematic diagramming. The excerpt frequency is provided below.

agency

51

TRANSPARENCY

18

task load

41

accountability

37

Emphasis on Qualitative Research

The qualitative factors like accountability or transparency are unique to user demographic and characteristics of the task, therefore, qualitative research is imperative, and provides holistic and rich insights in-tandem with quantitative research.

Quantitative

9 Future Work

Measuring Trustworthiness

After identifying the change in task load post seminar. It is vital to conduct further study on how qualitative factors like agency and transparency can affect trustworthiness through task load.

Task Load and Trust

Furthermore, as it was identified that a higher task load is associated with an increased trust, but an extensive study is required to validate these exploratory findings.

learnings

Epilogue

from data to PATTERNS

I learned how meaningful patterns can be drawn from a large pool of data through qualitative and quantitative methods.

insight synthesis

Patterns helped me identify emerging themes and led to the synthesis of insights, highlighting the role of trust in AI use.

Conducting seminar

Conducting the seminar was an overwhelming experience but it taught me how convey esoteric ideas to a diverse audience.