AI is not your lawyer
Share- Nishadil
- January 12, 2024
- 0 Comments
- 1 minutes read
- 64 Views
— generative AI models, like ChatGPT, that are trained to understand and produce human language content — have previously been known to “hallucinate” and generate false information. However, the raises “significant concerns” about the reliability of using LLMs in the field, the authors from Stanford University’s Institute for Human Centered AI and Regulation, Evaluation, and Governance Lab noted in a blog post.
When asked direct, verifiable questions about federal court cases, the study found the model behind ChatGPT, hallucinated of the time. gave incorrect answers to legal queries 72 percent of the time, while offered false information 88 percent of the time. The models performed worse when asked more complex legal questions, such as the core legal question or central holding of a case, or when asked about case law from lower courts, like district courts.
They also frequently in legal queries and tended to overstate their confidence in their responses, the study found. “Today, there is much excitement that LLMs will democratize access to justice by providing an easy and low cost way for members of the public to obtain legal advice,” the authors wrote in the blog post published Thursday.
“But our findings suggest that the current limitations of LLMs pose a risk of further deepening existing legal inequalities, rather than alleviating them,” they added. Thank you for signing up! Subscribe to more newsletters Copyright 2023 Nexstar Media Inc. All rights reserved. This material may not be published, broadcast, rewritten, or redistributed..