Algorithmic decision making has become prevalent in systems such as online search, social media, productivity software, and autonomous vehicles. While these intelligent systems powered by AI offer tremendous benefits, the algorithmic decision making behind them often feels like a mystery.
Without giving much thought to it, I used to blindly follow my friends’ activities on Facebook. I expected all my Facebook friends’ posts to appear in my news feed because I’m connected to them. When things had gone quiet, I had assumed that my friends and family were either not posting, or that they were actively hiding posts from me. That is, until I learned that the content that I see on Facebook is algorithmically curated.
After realizing that Facebook filters and ranks all available posts from my Facebook friends, I started to wonder how Facebook selects and controls the ordering and presentation of posts. Moreover, I began to speculate on how Facebook tracks my actions, so they could use them as signals to determine what content to display on my news feed.
Rather than attributing my lack of awareness about the Facebook’s content-curation algorithm to my own ignorance, I see this as an example of AI-driven products lacking transparency. User uncertainty and lack of knowledge about algorithmic decision making may cause anxiety (e.g., suspecting Facebook deliberately conceals posts). Worse, algorithmic avoidance could develop when users have a lower level of confidence in algorithmic than in human recommendations, even if the algorithmic recommendations are more accurate and insightful. This is why AI-enabled products must clarify the reasoning behind their predictions, recommendations, and other system decisions, so that we, as users, can understand, appropriately trust, and responsibly leverage the outputs from decision-making applications.
UX for AI design principle: Clarify
To enhance the productivity of all users with the power of CLAIRE, Informatica's intelligence engine, our user experience (UX) design approach considers the principle of Clarify when we design for AI-driven experiences. We understand and respect that users need and desire explanations, because the rationale behind algorithmically generated outcomes can be crucial to helping them make informed decisions and assessing the right actions to take.
To that end, when we clarify AI, we consider the following:
This way, users have clear understanding about how the system works, can make the right decisions, make the decisions faster, and most importantly, have greater confidence in the system.
What counts as a good explanation, however, depends on the knowledge that the users already have, as well as what additional information they need to understand. Thus, our approach is rooted in understanding the users’ existing knowledge, needs, and goals. This is our persona-based approach to design.
Below are a few examples of what we aspire to achieve with CLAIRE:
In other blog posts, we explore additional AI design principles that need to be considered when designing AI-enabled products, including the AI UX design principles of Trust , Simplify, Control, and Humanize.