What Happens When AI Needs a Therapist?

Key Takeaways

  • Artificial intelligence systems are common but can fail unexpectedly.
  • Figuring out *why* an AI failed is difficult because they often work like complex “black boxes.”
  • Researchers have created a new tool called AI Psychiatry (AIP) to investigate AI failures.
  • AIP recovers the exact AI model from the moment it failed and allows investigators to test it.
  • This helps determine if the failure was due to a bug, bad data, or a cyberattack.
  • The tool can be used for various AIs, not just self-driving cars, and also for preventative audits.

AI is everywhere, from delivery drones to helpful digital assistants. While these tools offer amazing benefits, they aren’t perfect and can sometimes go wrong.

When an AI system fails, understanding the cause is crucial. Was it a design flaw? Biased data? Or even a malicious hack?

The problem is that AI systems are often incredibly complex and opaque, making it hard even for their creators to pinpoint the exact reason for a malfunction. This is a major hurdle for investigators trying to figure out what happened after an incident.

To tackle this challenge, computer scientists at the Georgia Institute of Technology developed a system called AI Psychiatry, or AIP. According to an article republished by Gizmodo from The Conversation, this system provides a new way to perform digital forensics on AI.

Imagine a self-driving car suddenly crashes. AIP works by taking a snapshot of the AI’s digital “brain” – its working memory – at the exact moment of the failure.

Using this snapshot, AIP can reconstruct and essentially “reanimate” the faulty AI model in a safe, controlled lab environment.

Investigators can then carefully test this reanimated AI, feeding it different inputs to see if they can trigger the same failure or uncover hidden vulnerabilities introduced by hackers.

This process helps determine if the crash was caused by a flaw in the AI itself or if investigators need to look for other causes, like a broken sensor.

The research team successfully tested AIP on numerous AI models, including types used for recognizing street signs in autonomous vehicles. It reliably recovered and tested the models, even those intentionally altered to fail under specific conditions.

AIP isn’t limited to self-driving cars. Its core method works on the fundamental parts common to most AI models built with popular tools. This makes it adaptable for investigating almost any AI system, from recommendation bots to drone fleet controllers.

Furthermore, AI Psychiatry can be used proactively for auditing AI systems before they are deployed or cause problems. As more government agencies adopt AI, tools like AIP offer a consistent way to ensure these systems are working safely and fairly.

The researchers have made AI Psychiatry open source, meaning it’s freely available for investigators and auditors to use, helping to build trust and understanding in the AI systems shaping our world.

Independent, No Ads, Supported by Readers

Enjoying ad-free AI news, tools, and use cases?

Buy Me A Coffee

Support me with a coffee for just $5!

 

More from this stream

Recomended