The success of an AI deployment within existing human processes comes down to one thing.
(Hint: It's not Accuracy.)
It's Trust.
Trust is the foundation upon which successful AI integration is built. And here's a critical insight: data alone isn't the way most people build trust. A glance at history reveals why. Look at the climate change communications of the 80s and 90s. Those efforts were overwhelmingly "data-focused," loaded with charts, figures, and scientific rigor—but they failed to change minds. Why? Because trust and persuasion aren't built on numbers alone; they hinge on emotional connection and perceived intent.
In the context of AI, trust isn't about presenting the best metrics. It's about how it feels to use the system and whether users believe in the intent behind its design.
And yes, Trust can be measured.
AI/ML practitioners would do well to borrow from brand marketing's playbook on trust measurement. Brand marketers have spent decades honing tools to assess how people feel about products, services, and experiences. That same "softer side" of business could hold the key to solving some of the hardest challenges in AI adoption.
Ironically, while scientists and engineers often focus on technical metrics, the ultimate success of AI often depends on softer, less tangible qualities. Consider this: stakeholders will realize efficiency gains from AI only if they trust its outputs. If stakeholders spend more time second-guessing and invalidating the AI’s recommendations than they would doing the task manually, the whole purpose of automation collapses.
For many ML practitioners, this is a fundamentally hard shift in perspective. Here’s what they tend to prioritize:
- F1 score: A balanced measure of precision and recall, especially useful in imbalanced datasets.
- Precision: The percentage of relevant results correctly identified.
- Recall: The percentage of all relevant instances retrieved.
- Accuracy: How often the model gets things right.
- Performance: Speed and responsiveness of the system.
- Computational resource utilization: Efficiency in using hardware and energy.
- Scalability: How well the system adapts to larger workloads.
- Data utilization: Effective use of available data.
- Model interpretability: The ability for humans to understand why the model made a certain decision.
These are all important. But without trust, even the most sophisticated metrics and models are meaningless. Trust is the bridge that connects technical achievement to real-world impact.
Take, for example, some well-documented AI failures that eroded trust:
Here’s another real-world example from my consulting experience of how a lack of trust can undermine even a technically sound AI project:
Untrusted Sources
A company I consulted had spent a year with an external team re-architecting their platform, developing machine learning models, and deploying a solution intended to support their operations department with manual categorization tasks. The AI solution claimed measurable improvements over the system’s default capabilities and had "data" to back it up. However, the solution was deployed to a team already in conflict with the internal department sponsoring the project. This existing tension caused the operations team to view the AI solution as untrustworthy and even threatening. Consequently, they saw little to no improvement in efficiency. Instead, the deployment exacerbated interdepartmental friction. The core issue? The AI came from an untrusted source and was therefore untrusted itself. Trust, or the lack thereof, determined the outcome more than the technical performance.
How Marketers Measure Trust and Applying It Internally
By applying these strategies internally, organizations can create a structured approach to fostering trust in AI systems, ensuring smoother adoption and greater efficiency gains.
Conclusions
These failures remind us that trust isn’t just a nice-to-have; it’s essential for long-term success. Building and maintaining trust means acknowledging potential pitfalls and addressing them proactively.
So, how are you measuring and building trust in your AI systems? Are you considering the human experience as deeply as the technical performance? Let’s shift the narrative from "what can AI do" to "how does AI make people feel."
Trust might just be the ultimate metric.