The AI Trust Fall

The AI Trust Fall

If I asked you, “When someone turns in a work assignment, how accurate is it? 80%, 90%, 95% or perhaps 100%?”

We don’t think this way about coworkers’ spreadsheets. But we will probably think this way about AI & this will very likely change the way product managers on-board users.

When was the last time you signed up for a SaaS & wondered : Would the data be accurate? Would the database corrupt my data? Would the report be correct?

But today, with every AI software now tucking a disclaimer at the bottom of the page, we will be wondering. “Gemini may display inaccurate info, including about people, so double-check its responses” & “ChatGPT/Claude can make mistakes. Check important info” are two examples.

In the early days of this epoch, mistakes will be common. Over time, less so, as accuracies improve.

The more important the work, the greater peoples’ need to be confident the AI is correct. We will demand much better than human error rates. Self-driving cars provide an extreme example of this trust fall. Waymo & Cruise have published data arguing self-driving cars are 65-94% safer.

Yet, 2/3 of Americans surveyed by the AAA fear them.

We suffer from a cognitive bias : work performed by a human is likely more trustworthy because we understand the biases & the limitations. AIs are a Schrodinger’s cat stuffed in a black box. We don’t comprehend how the box works (yet), nor can we believe our eyes if the feline is dead or alive when we see it.

New product on-boarding will need to mitigate this bias.

One path may be starting with low-value tasks where the software-maker has tested exhaustively the potential inputs & outputs. Another tactic may be to provide a human-in-the-loop to check the AI’s work. Citations, references, & other forms of fact-checking will be a core part of the product experience. Independent testing might be another path.

As with any new colleague, the first impressions & a series of small wins will determine the person’s trust. Severe errors in the future will erode confidence, that must be rebuilt - likely with the help of human support teams who will explain, develop tests for the future, & assure users.

I recently asked a financial LLM to analyze NVIDIA’s annual report. A question about the company’s increase in dividend amount vaporized its credibility, raising the question : is it less work to do the analysis myself than to check the AI’s work?

That will be the trust fall for AI. Will the software catch us if we trust it?

Jason M. Lemkin

SaaStr AI London is Dec 1-2!! See You There!!

1y

It in essence passes the test by default for jobs no one is willing to do anymore

Kirill Solovjov

AI LQA safely & reliably deployed to any TMS · LLM multilingual evals made easy · TQE automation for Localization teams · Translation Quality Evaluation expert · B2B SaaS founder · 20yrs in tech & int’l

1y

Trust but verify, this is what we always say at ContentQuo! We’ve built the company to help buyers of AI-powered services develop trust towards their vendors based on easy to understand & openly shared quality metrics. In our first vertical, this really helped alleviate trust issue over time and move to constructive data-based vendor performance conversations and procurement processes. Expecting to see more of it as LLM space grows!

Konstantin Ristl

AI & Outreach Enthusiast | Founder/CEO @histack.io - More Leads. Less Effort. | Post-exit investor

1y

 love your post! I’m a big advocate for pointing out that not only AI but also humans make mistakes. When trusting someone with a task for the first time, I’m more skeptical than after seeing them perform it multiple times. As for people: Not every task fits every person, just like not every task fits AI.

Jeff Dodge

Chief Revenue Officer at Nimble Gravity - Building AI Foundations and Solutions on Azure

1y

Great piece. Your point about humans trusting self-driving cars is particularly poignant as the data is pretty overwhelming that lives would be saved if self-driving cars were broadly adopted in suitable circumstances. Ultimately humans do want to know "who is accountible" and "who will be punished if something goes wrong" even if there are less occurances of something going wrong. When it's a human at fault, society punishes that human and it 'feels right' ie justice is served. When AI is wrong, the only 'person' to hold accountable in most circumstances is a corporation, and generally a trillion dollar corporation and they have a lot to lose finincially so the risk/reward for those corporations is tricky and even more so the legal ramification.

Luke Shalom

CEO @ Atticus | AI sales systems for founders | 5–10 deals/month from LinkedIn in <30 mins/day I $125M+ pipeline generated for clients | #4 in Growth & Lead Gen in the UK

1y

Trusting AI without fully knowing it can be worrying.

Like
Reply

To view or add a comment, sign in

More articles by Tomasz Tunguz

  • The Bacon & the Skillet: When Does the AI Market Congeal?

    The AI market today is bacon in a hot skillet. Everything is sizzling, moving, & changing at an incredible pace.

    18 Comments
  • The Scaling Wall Was A Mirage

    Two revelations this week have shaken the narrative in AI : Nvidia’s earnings & this tweet about Gemini. The AI…

    21 Comments
  • What 375 AI Builders Actually Ship

    70% of production AI teams use open source models. 72.

    21 Comments
  • Teaching Local Models to Call Tools Like Claude

    Ten months ago, DeepSeek collapsed AI training costs by 90% using distillation - transferring knowledge from larger…

    30 Comments
  • Running Out of AI

    By Monday lunch, I had burned through my Claude code credits. I’d been warned ; damn the budget, full prompting ahead.

    35 Comments
  • Datadog: As Reliable as Your Golden Retriever

    Datadog is becoming a platform company, & its Q3 2025 results underscore how successful this transition is. If nothing…

    20 Comments
  • Are We Being Railroaded by AI?

    Just how much are we spending on AI? Compared to other massive infrastructure projects, AI is the sixth largest in US…

    7 Comments
  • Are We Being Railroaded by AI?

    Just how much are we spending on AI? Compared to other massive infrastructure projects, AI is the sixth largest in US…

    23 Comments
  • A 1 in 15,787 Chance Blog Post

    I wrote a post titled Congratulations, Robot. You’ve Been Promoted! in which OpenAI declared that their AI coders were…

    31 Comments
  • OpenAI's $1 Trillion Infrastructure Spend

    OpenAI has committed to spending $1.15 trillion on hardware & cloud infrastructure between 2025 & 2035.

    28 Comments

Others also viewed

Explore content categories