So, what explains the shocking failures of basic reasoning that Simple Bench exposes? Let’s return to the vegetables — or fruits and vegetables, if you wish. Here’s my take on the first of my two questions: Why do models fail the question you saw above? … The clue is in their name: language models. They model language.…
src: https://www.linkedin.com/posts/denis-o-b61a379a_genai-activity-7237647652052488192-CwHF?utm_source=share&utm_medium=member_desktop
With new headlines cropping up all the time showing how poorly a job Tesla Autopilot actually does at piloting cars, it’s not surprising that the people tasked with building that system’s knowledge out are underpaid and heavily surveilled — or that when they tried to fight back against what they felt was unfair treatment, they were summarily cast aside. https://futurism.com/the-byte/tesla-autopilot-workers-treatment
Martin Bernklau is a German journalist who reported for decades on criminal trials. He looked himself up on Bing, which suggests you use its Copilot AI. Copilot then listed a string of crimes Bernklau had supposedly committed — saying that he was an abusive undertaker exploiting widows, a child abuser, an escaped criminal mental patient.
“The web is becoming increasingly a dangerous place to look for your data,” Rice University graduate student Sina Alemohammad, who coauthored a 2023 paper that coined the term “MAD” — short for “Model Autophagy Disorder” — to describe the effects of AI self-consumption, told the NYT. https://futurism.com/ai-slowly-killing-itself