Tags
Aktuelle Nachrichten
America
Aus Aller Welt
Breaking News
Canada
DE
Deutsch
Deutschsprechenden
Europa
Europe
Global News
Internationale Nachrichten aus aller Welt
Japan
Japan News
Kanada
Konflikt
Korea
Krieg in der Ukraine
Latest news
Map
Nachrichten
News
News Japan
Russischer Überfall auf die Ukraine seit 2022
Science
South Korea
Ukraine
UkraineWarVideoReport
Ukraine War Video Report
Ukrainian Conflict
UkrainianConflict
United Kingdom
United States
United States of America
US
USA
USA Politics
Vereinigte Königreich Großbritannien und Nordirland
Vereinigtes Königreich
Welt
Welt-Nachrichten
Weltnachrichten
Wissenschaft
World
World News
9 Comments
They never could reason and the only people who believed this were laymen unfamiliar with how GPTs actually work.
At their core, they are very fancy prediction and probability engines. Thats it. They either predict the next word in a sentence or the next pixel in an image. Most times they are right, sometimes they are laughably wrong. Even calling them AI is a huge stretch.
[deleted]
I don’t disagree with the premise of the article, but when you’re testing an LLM “with a given math question” you’re unlikely to get good results.
Uh. Duh? No shit. New to LLMs?
Article makes no mention of GPT-4o1. I wonder if the study included the latest preview model from OpenAI which aims to solve this.
So apple intelligence is bullshit? Got it.
This is a significant problem, because as someone who works effectively in tech support, I can say the vast majority of humans do not have the ability to parse down what they want, or what problem they are having, into concise questions with only the relevant info.
It’s usually either “my phone isn’t working” or it’s a story so meandering that even Luis from *Ant-Man* would be saying “Get to the point!!!”
This will be a more important thing for AI researchers to figure out.
>
Hence why LLM’s are called *predictive* models, and not *reasoning* models
There was already a paper on this called “ChatGPT is Bullshit”: https://link.springer.com/article/10.1007/s10676-024-09775-5