Comment by tsoukase
If an LLM hallucinates in 1% of occasions and gives subpar output in 5%, this kills his effectiveness to replace anyone. Imagine a support guy on the other side of the phone to speak gibberish 10 times a day. Now, imagine a doctor. These will never lose their jobs.
The models don't need to be perfect. They only need to be as reliable as humans.
Emergency department doctors misdiagnose about 5% of patients [1], so replacing them with an LLM that hallucinates on 1% of cases would actually be a significant improvement.
1: https://effectivehealthcare.ahrq.gov/products/diagnostic-err...