> Something to keep in mind the next time someone tries to sell you a large language model for expert advice.

just finished reading The Intelligence Illusion. Really informative read and very accessible. For me the most important parts came at the end, but I will never have to manage implementation. It gave me a good understanding of where things stand. I plan to follow up on some of your reading recommendations to further deepen my knowledge.

it’s a good book. Informative in just the ways you meant it to be. Interesting side note. My brother-in-law is head toxicologist and VP for a relatively major pharmaceutical company. They are experimenting with incorporating AI into their research work flow. He told me that they task it with, for example, with something like summarizing/assessing the research literature on topic x, y, or z. Their experience is that it gets the summary/assessment 100% right but the list of citations it built the summary/assessment on 100% wrong.

@mbkriegh Thanks. 🙂
Interesting to hear about the research papers. The benchmarks for language model summarisation are usually collections of research papers, so it would stand to reason that their results there would be more accurate than with most other papers. And it would make sense that the citations were 100% wrong as that's exactly where they're weak.
The worry I would have is, what are the consequences if it's not 100% right, but 100% right 98% of the time and 100% wrong 2% of the time?
Because that's the dynamic with these models. You hit the long tail or an edge case that's just a little bit too novel to it and it goes bonkers, but because of all the other times it worked, you've come to trust it. I'm glad that it gets the citations 100% wrong. That should make people trust it less and they need to be distrustful, because its the 98% right use cases where adopting these tools can do the most damage.