From The Hill.com (Jan. 11, 2024):
Generative artificial intelligence (AI) models frequently produce false legal information, with so-called “hallucinations” occurring between 69 percent to 88 percent of the time, according to a recent study.
Large language models (LLMs) — generative AI models, like ChatGPT, that are trained to understand and produce human language content — have previously been known to “hallucinate” and generate false information.
However, the “pervasive” nature of legal hallucinations raises “significant concerns” about the reliability of using LLMs in the field, the authors from Stanford University’s Institute for Human-Centered AI and Regulation, Evaluation, and Governance Lab noted in a blog post.
When asked direct, verifiable questions about federal court cases, the study found the model behind ChatGPT, GPT-3.5, hallucinated 69 percent of the time, while Google’s PaLM 2 gave incorrect answers 72 percent of the time and Meta’s Llama 2 offered false information 88 percent of the time.
The models performed worse when asked more complex legal questions, such as the core legal question or central holding of a case, or when asked about case law from lower courts, like district courts.
They also frequently failed to contradict false premises in legal queries and tended to overstate their confidence in their responses, the study found.
“Today, there is much excitement that LLMs will democratize access to justice by providing an easy and low-cost way for members of the public to obtain legal advice,” the authors wrote in the blog post published Thursday. “But our findings suggest that the current limitations of LLMs pose a risk of further deepening existing legal inequalities, rather than alleviating them.”
“Ideally, LLMs would excel at providing localized legal information, effectively correct users on misguided queries, and qualify their responses with appropriate levels of confidence,” they added. “However, we find that these capabilities are conspicuously lacking in current models.”
The consequences of such hallucinations have already been seen in the legal field. A federal judge sanctioned two lawyers in June after one used fake case citations that were generated by ChatGPT.
Michael Cohen, former President Trump’s ex-fixer and personal lawyer, also admitted last month to giving his attorney fake case citations after using Google Bard, which ran on PaLM 2 until recently.
In his annual year-end report, Chief Justice John Roberts warned about the potential drawbacks of using AI in the legal field, even as he suggested that the technology could significantly affect judicial work in the future.
“Any use of AI requires caution and humility,” he noted. “One of AI’s prominent applications made headlines this year for a shortcoming known as ‘hallucination,’ which caused the lawyers using the application to submit briefs with citations to non-existent cases. (Always a bad idea.).” [source]
88 percent is a big percentage of fake data. AI should only be used as a tool, not the final say especially concerning legal content.
More articles on AI:
- AI's Hidden Geometry of Thought
- AI models are reprogramming themselves to 'play dumb' or copy their data to other servers to survive
- AI Is Breaking Into a Higher Dimension—Literally—to Mimic the Human Brain and Achieve True Intelligence
- Should AI be Shut Down?
- Resistance Is Futile, But Maybe Not With AI
- AI Lacks the Entrepreneurial Intelligence to Plan an Economy
- AI Can Be Used in Spying, Cost US Jobs, Experts Warn
- AI system can predict the hour an animal gives birth
- AI Misled by Distractions: Tissue Contamination Can Throw Medical Diagnoses into Chaos
- AI falsely accuses, fines artificial intelligence expert of using phone while driving - report
- Scientists say they can use AI to solve a key problem in the quest for near-limitless clean energy
- AI-engineered enzyme eats entire plastic containers
- Could AI predict future earthquakes? Israeli researchers say yes
- Artificial intelligence could soon diagnose illness based on the sound of your voice
- AI will require even more energy than we thought
- AI-powered headphones can tune into a single voice in a crowd
- AI used to predict potential new antibiotics in groundbreaking study
- AI Deep Learning Decodes Hand Gestures from Brain Images
- A Secret AI System Called Sentient Is Capable of “Seeing and Detecting UFOs”
- AI 'resurrects' 54 Roman emperors, in stunningly lifelike images
- Microsoft Launches New AI Service For U.S. Spies
- US Aims to Stay Ahead of China in Using AI to Fly Fighter Jets, Navigate without GPS and More
- Israel unveils IDF's new AI-powered 'Barak' super tank
- Human Vs. Machine: Behavioral Differences Between Expert Humans And Language Models In Wargame Simulations [pdf]
- AI Sweeps Human Pilot in DARPA Dogfight Simulation
- AI 'personality' now running to be American mayor!
- AI Accurately Predicts Psychosis Well Before It Even Occurs
- AI can identify passwords by sound of keys being pressed, study suggests
- Catholic AI chatbot used by over 180K people to ‘ensure fidelity to Church teaching’
- USC Researchers Use AI to Help Translate Bible Into Very Rare Languages
- Rubio Impostor Using AI Voice to Call High-Level Officials
- The Rise of the Machines

No comments:
Post a Comment