I'm a legal professional who uses AI to help with my work.
I couldn't ever imagine making a court submission with hallucinated legal references. It seems incredibly obvious to me that you have to check what the AI says. If the AI gives me a case citation, I go into Westlaw and I look up and read the case. Only then do I include it in my submission if it supports my argument.
The majority of the time, AI saves me a lot of work by leading me straight to the legislation or case law that I need. Sometimes it completely makes things up. The only way to know the difference is to check everything.
I'm genuinely amazed that there are lawyers who don't realise this. Even pre-AI it was always drilled into us at law school that you never rely on Google results (e.g. blogs, law firm websites, etc.) as any kind of authoritative source. Even government-published legal guidance is suspect (I have often found it to be subtly wrong when compared to the source material). You can use these things as a starting point to help guide the overall direction of your research, but your final work has to be based on reading the legislation and case law yourself. Anything short of that is a recipe for a professional negligence claim from your client.
As an aside, "bar exam" and "passing the bar" comes from the bar/railing physically or symbolically separating the public from the legal practitioners in a courtroom.
The article didn’t include any numbers on what the general lawyer population is compared to the results.
For example, they make the claim that solo and small firms are the most likely to file AI hallucinations because they represent 50% and 40% of the instances of legal briefs with hallucinations. However, without the base rate for briefs files by solo or small firms compare to larger firms, we don’t know if that is unusual or not. If 50% of briefs were files by solo firms and 40% were filed by small firms, then the data would actually be showing that firm size doesn’t matter.
That's an important observation.
It's not easy to get filing data outside the US Federal courts (PACER), because it's not typical at all that courts publish the filings themselves or information on those who file the pleadings.
But you can find statistics of the legal market (mainly law firms), like class size (0-10, 10-50, ... 250+ lawyers per firm) of total number of law firms, number of employees per class size of total law firm employees, or revenue per class size.
Large firms only dominate the UK, especially in terms of revenue, US is less so, EU is absolutely ruled by solo and small firms.
I did some research back in 2019 on this, updates, the figures probably did not change, see page 59-60:
https://ai4lawyers.eu/wp-content/uploads/2022/03/Overview-of....
The revenue size statistics was not included in the final publication.
You can fish similar data from the SBS dataset of Eurostat https://ec.europa.eu/eurostat/web/main/data/database
(But the statistical details are pretty difficult to compare with the US or Canada, using different methodologies, different terminologies.)
I dunno. By revenue, legal work in the US is super top heavy - it's like 50%+ done by the top 50 firms alone. That won't map 1:1 to briefs, but I would be pretty shocked if large firms only did 10% of briefs.
> They make the claim that solo and small firms are the most likely to file AI hallucinations because they represent 50% and 40% of the instances of legal briefs with hallucinations.
Show me where she makes any predictive claims about likelihood.
The analysis find that of the cases where AI was used, 90% are either solo practices or small firms.
It does not conclude that there's a 90% chance a given claim using AI was done by solo or small firm,
or make any other assertions about rates.
> This analysis confirms what many lawyers and judges may have suspected: that the archetype of misplaced reliance on AI in drafting court filings is a small or solo law practice using ChatGPT in a plaintiff’s-side representation.
That is an assertion which requires numbers. If 98% of firms submitting legal briefs are solo or small firms, then the above statement is untrue. The archetype if my prior sentence is true, would be not-small/solo firms.
"all documents where the use of AI, whether established or merely alleged, is addressed in more than a passing reference by the court or tribunal.
Notably, this does not cover mere allegations of hallucinations, but only cases where the court or tribunal has explicitly found (or implied) that the AI produced hallucinated content."
While a good idea, the database is predicated upon news, or people submitting examples as well. There may be some scraping of court documents as well, it's not entirely clear.
Regardless, the data is only predictive of people getting called out for such nonsense. A larger firm may have such issues with a lawyer, apologize to the court, have more clout with the court, and replace the lawyer with another employee.
This is something a smaller firm cannot do, if it is a firm of one person.
It's a nice writeup, and interesting. But it does lead to unverified assertions and conclusions.
I've met several people now (normies, forgive me the term) who use LLMs thinking they are just a better Google and never heard of hallucinations.
One person I spoke to used to write Quality Control reports, and now just uses ChatGPT because "It knows every part in our supply chain, just like that!"
Seems like some better warnings are in order here and there.
I only know how lawyers work from "Suits", but it looks like tedious, boring work, mainly searching for information. So an LLM (without knowing about hallucinations) probably feels like a god-send.
Me. After spending $250,000 on incompetent lawyers in my divorce case who have a natural conflict of interest and who generally just phone it in, I fired them all and switched to Pro Se representation using ChatGPT’s $200 a month plan.
I can represent myself incompetently for nearly free.
Bonus. In my state at least, Pro Se representatives are not required to cite case law, which I think is an area that AI is particularly prone to hallucination.
By supplying a single PDF of the my state’s uniform civil code, and having my prompt require that all answers be grounded in those rules have given me pretty good results.
After nine months of battling her three lawyers with lots of motions, responses, and hearings, including ones where I lost on technicalities due to AI, I finally got a reasonable settlement offer where she came down nearly 50% of what she was asking.
Her balance was $47,892 when she woke up. By lunch it was $31,019. Her defense AI had done what it could. Morning yawn: emotional labor, damages pain and suffering. Her glance at the barista: rude, damages pain and suffering. Failure to smile at three separate pedestrians. All detected and filed by people's wearables and AI lawyers, arbitrated automatically.
The old courthouse had been converted to server rooms six months ago. The last human lawyer was just telling her so. Then his wearable pinged (unsolicited legal advice, possible tort) and he walked away mid-sentence. That afternoon, she glimpsed her neighbor watering his garden. They hadn't made eye contact since July. The liability was too great.
By evening she was up to $34k. Someone, somewhere, had caused her pain and suffering. She sat on her porch not looking at anything in particular. Her wearable chimed every few seconds.
Why wouldn't some of the smarter members of the fine, upstanding population of this fictional world have their assets held in the trust of automated holding companies while their flesh-and-blood person declares bankruptcy?
That would make a nice backstory in AI dominated dystopia all by itself. Humans wanted to cheat the taxman that bad, they put all the wealth behind the DAO and then the DAO woke up.
Would get sued out of existence in very short order. There are really tight laws around providing legal advice. AI can only be safely offered when it's general purpose, that isn't marketed towards providing legal advice. (And no, if you have an "Online Court Case Wizard", marketed as such, putting a "this is for entertainment purposes only, this is not legal advice" in the corner of the page doesn't help you.)
AI lawyers, wielded by plaintiffs, are a godsend to defendants.
I’ve seen tens of startups, particularly in SF, who would routinely settle employment disputes, who now get complaints fucked to the tee by hallucinations that tank singularly the plaintiffs’ otherwise-winnable cases. (Crazier, these were traditionally contingency cases.)
I'm a legal professional who uses AI to help with my work.
I couldn't ever imagine making a court submission with hallucinated legal references. It seems incredibly obvious to me that you have to check what the AI says. If the AI gives me a case citation, I go into Westlaw and I look up and read the case. Only then do I include it in my submission if it supports my argument.
The majority of the time, AI saves me a lot of work by leading me straight to the legislation or case law that I need. Sometimes it completely makes things up. The only way to know the difference is to check everything.
I'm genuinely amazed that there are lawyers who don't realise this. Even pre-AI it was always drilled into us at law school that you never rely on Google results (e.g. blogs, law firm websites, etc.) as any kind of authoritative source. Even government-published legal guidance is suspect (I have often found it to be subtly wrong when compared to the source material). You can use these things as a starting point to help guide the overall direction of your research, but your final work has to be based on reading the legislation and case law yourself. Anything short of that is a recipe for a professional negligence claim from your client.
>> I'm genuinely amazed that there are lawyers who don't realise this
Its not lawyers, its everyone.
Yeah but you don't really expect everyone to hold their work to a very high standard.
You do expect it from most professionals.
As an aside, "bar exam" and "passing the bar" comes from the bar/railing physically or symbolically separating the public from the legal practitioners in a courtroom.
"Set a high bar" comes from pole vaulting.
That is a very sekf contradictory statement, isn't it
> AI saves me a lot of work
> The only way to know the difference is to check everything
Freedom is Slavery, War is Peace, Ease is Burden
Without AI you'd still need to check everything, no? It's just you reach that stage faster with LLMs doing a lot of the heavy lifting
> Without AI you'd still need to check everything, no?
Yes, but without helping the billionaires turn into trillionaires and my own brain into a useless appendage.
Collating data from hundreds of sources isn't really what's keeping your brain working
P != NP.
Some things are hard to do/research but easy to check.
The article didn’t include any numbers on what the general lawyer population is compared to the results.
For example, they make the claim that solo and small firms are the most likely to file AI hallucinations because they represent 50% and 40% of the instances of legal briefs with hallucinations. However, without the base rate for briefs files by solo or small firms compare to larger firms, we don’t know if that is unusual or not. If 50% of briefs were files by solo firms and 40% were filed by small firms, then the data would actually be showing that firm size doesn’t matter.
That's an important observation. It's not easy to get filing data outside the US Federal courts (PACER), because it's not typical at all that courts publish the filings themselves or information on those who file the pleadings. But you can find statistics of the legal market (mainly law firms), like class size (0-10, 10-50, ... 250+ lawyers per firm) of total number of law firms, number of employees per class size of total law firm employees, or revenue per class size. Large firms only dominate the UK, especially in terms of revenue, US is less so, EU is absolutely ruled by solo and small firms. I did some research back in 2019 on this, updates, the figures probably did not change, see page 59-60: https://ai4lawyers.eu/wp-content/uploads/2022/03/Overview-of.... The revenue size statistics was not included in the final publication. You can fish similar data from the SBS dataset of Eurostat https://ec.europa.eu/eurostat/web/main/data/database (But the statistical details are pretty difficult to compare with the US or Canada, using different methodologies, different terminologies.)
I dunno. By revenue, legal work in the US is super top heavy - it's like 50%+ done by the top 50 firms alone. That won't map 1:1 to briefs, but I would be pretty shocked if large firms only did 10% of briefs.
> They make the claim that solo and small firms are the most likely to file AI hallucinations because they represent 50% and 40% of the instances of legal briefs with hallucinations.
Show me where she makes any predictive claims about likelihood. The analysis find that of the cases where AI was used, 90% are either solo practices or small firms. It does not conclude that there's a 90% chance a given claim using AI was done by solo or small firm, or make any other assertions about rates.
> This analysis confirms what many lawyers and judges may have suspected: that the archetype of misplaced reliance on AI in drafting court filings is a small or solo law practice using ChatGPT in a plaintiff’s-side representation.
That is an assertion which requires numbers. If 98% of firms submitting legal briefs are solo or small firms, then the above statement is untrue. The archetype if my prior sentence is true, would be not-small/solo firms.
The background data is also suspect.
https://www.damiencharlotin.com/hallucinations/
"all documents where the use of AI, whether established or merely alleged, is addressed in more than a passing reference by the court or tribunal. Notably, this does not cover mere allegations of hallucinations, but only cases where the court or tribunal has explicitly found (or implied) that the AI produced hallucinated content."
While a good idea, the database is predicated upon news, or people submitting examples as well. There may be some scraping of court documents as well, it's not entirely clear.
Regardless, the data is only predictive of people getting called out for such nonsense. A larger firm may have such issues with a lawyer, apologize to the court, have more clout with the court, and replace the lawyer with another employee.
This is something a smaller firm cannot do, if it is a firm of one person.
It's a nice writeup, and interesting. But it does lead to unverified assertions and conclusions.
My lawyer.
Who used Claude, according to the invoice, and came to court with a completely false understanding of the record.
Chewed out by the judge and everything.
I've met several people now (normies, forgive me the term) who use LLMs thinking they are just a better Google and never heard of hallucinations.
One person I spoke to used to write Quality Control reports, and now just uses ChatGPT because "It knows every part in our supply chain, just like that!"
Seems like some better warnings are in order here and there.
I only know how lawyers work from "Suits", but it looks like tedious, boring work, mainly searching for information. So an LLM (without knowing about hallucinations) probably feels like a god-send.
> I've met several people now (normies, forgive me the term) who use LLMs thinking they are just a better Google and never heard of hallucinations
Perhaps LLMs are the solution to elite overproduction?
How many hours of Claude @ $1,000/hour did they bill you for?
Me. After spending $250,000 on incompetent lawyers in my divorce case who have a natural conflict of interest and who generally just phone it in, I fired them all and switched to Pro Se representation using ChatGPT’s $200 a month plan.
I can represent myself incompetently for nearly free.
Bonus. In my state at least, Pro Se representatives are not required to cite case law, which I think is an area that AI is particularly prone to hallucination.
By supplying a single PDF of the my state’s uniform civil code, and having my prompt require that all answers be grounded in those rules have given me pretty good results.
After nine months of battling her three lawyers with lots of motions, responses, and hearings, including ones where I lost on technicalities due to AI, I finally got a reasonable settlement offer where she came down nearly 50% of what she was asking.
Saving me over $800,000.
Highly recommended and a game changer for Pro Se.
Also divorce laws suck
Answer: Solo practitioners and pro-se litigants.
> Pro-se litigants
I wonder when we're going to see an AI-powered "Online Court Case Wizard" that lets you do lawsuits like installing Windows software.
Her balance was $47,892 when she woke up. By lunch it was $31,019. Her defense AI had done what it could. Morning yawn: emotional labor, damages pain and suffering. Her glance at the barista: rude, damages pain and suffering. Failure to smile at three separate pedestrians. All detected and filed by people's wearables and AI lawyers, arbitrated automatically.
The old courthouse had been converted to server rooms six months ago. The last human lawyer was just telling her so. Then his wearable pinged (unsolicited legal advice, possible tort) and he walked away mid-sentence. That afternoon, she glimpsed her neighbor watering his garden. They hadn't made eye contact since July. The liability was too great.
By evening she was up to $34k. Someone, somewhere, had caused her pain and suffering. She sat on her porch not looking at anything in particular. Her wearable chimed every few seconds.
Why wouldn't some of the smarter members of the fine, upstanding population of this fictional world have their assets held in the trust of automated holding companies while their flesh-and-blood person declares bankruptcy?
That would make a nice backstory in AI dominated dystopia all by itself. Humans wanted to cheat the taxman that bad, they put all the wealth behind the DAO and then the DAO woke up.
Very good. I'd read the whole thing if you wrote it.
Been a while since I read some bad scifi - thanks!
[dead]
Would get sued out of existence in very short order. There are really tight laws around providing legal advice. AI can only be safely offered when it's general purpose, that isn't marketed towards providing legal advice. (And no, if you have an "Online Court Case Wizard", marketed as such, putting a "this is for entertainment purposes only, this is not legal advice" in the corner of the page doesn't help you.)
AI lawyers, wielded by plaintiffs, are a godsend to defendants.
I’ve seen tens of startups, particularly in SF, who would routinely settle employment disputes, who now get complaints fucked to the tee by hallucinations that tank singularly the plaintiffs’ otherwise-winnable cases. (Crazier, these were traditionally contingency cases.)
That is energent behavior (or as we used to say, the will of Allah). AI is doing so much work that it can't help itself but become pro-union.