Apple study exposes deep cracks in LLMs’ “reasoning” capabilities
Apple study exposes deep cracks in LLMs’ “reasoning” capabilities
arstechnica.com Apple study exposes deep cracks in LLMs’ “reasoning” capabilities
Irrelevant red herrings lead to “catastrophic” failure of logical inference.
You're viewing a single thread.
View all comments
109
comments
Are the uncensored models more capable tho?
5 12 ReplyGiven the use cases they were benchmarking I would be very surprised if they were any better.
11 0 Reply
You've viewed 109 comments.
Scroll to top