Plenty of progress in models that can use tools and search. Would love to see how one of these tool/search-enabled models do at this kind of a task. In my experience, they don't fabricate things anymore, just sometimes occasionally misrepresent the content of citations (put a citation somewhere where it doesn't actually support what is written).
A few days ago I asked GPT 5 for links to news on the Charlotte murder before the story got reported by the mainstream media. It gave me five different links, including AP and Reuters. Every one, five out of five, was a hallucination.
403 for me - which makes me wonder how anyone else is commenting on the actual content of the link, rather than just recycling general comments. without knowing the details.
Plenty of progress in models that can use tools and search. Would love to see how one of these tool/search-enabled models do at this kind of a task. In my experience, they don't fabricate things anymore, just sometimes occasionally misrepresent the content of citations (put a citation somewhere where it doesn't actually support what is written).
A few days ago I asked GPT 5 for links to news on the Charlotte murder before the story got reported by the mainstream media. It gave me five different links, including AP and Reuters. Every one, five out of five, was a hallucination.
403 for me - which makes me wonder how anyone else is commenting on the actual content of the link, rather than just recycling general comments. without knowing the details.
Me too. https://web.archive.org/web/20250914073627/https://www4.cour... works
The title needs some punctuation, but the link works fine for me.
Two of them were real? That's a state-of-the-art model, compared to what I've seen…
A PhD-level degree in fabrication.