LLMs don’t do formal reasoning - and that is a HUGE problem
LLMs don’t do formal reasoning - and that is a HUGE problem
Important new study from AppleGary Marcus (Marcus on AI)
like this
Adam Tooze: Bidenomics is Maga for thinking people.
Facing war in the Middle East and Ukraine, the US looks feeble. But is it just an act?
The idea that all Biden is doing to trying to avoid a third world war isn’t convincing. Look closely and his foreign policy has been as radical as Trump’s, says history professor Adam ToozeAdam Tooze (The Guardian)
Until then, you know, you just have to be true to yourself because, if you're not being true to yourself, you'll be living a lie.
In the end you've just got to remember that it is what is and you've got to what you've got to do. You gotta do your thing you, know?
So just be you but a you that's true to yourself while going with the flow and bossing it your way, all the way.
Most of all, be lucky.
Future antiquities researchers
like this
Despite Chrystia Freeland’s denials, her grandfather was complicit in the Nazi genocide
Despite Chrystia Freeland’s denials, her grandfather was complicit in the Nazi genocide ⋆ The Breach
A new book provides the most authoritative study of Mykhailo Chomiak and the history of Ukrainian Nazis in CanadaPeter McFarlane (The Breach)
3 New Wayland Protocols about to drop (commit & presentation timing) - Needed for 3rd Protocol (FIFO)
wayland/wayland-protocols!248
Needs 2 Acks + Review
wayland/wayland-protocols!320
Just got Completed:
wayland/wayland-protocols!256
FIFO Just got completed too.
like this
A ‘dark period’ of repression: Jordanian authorities arrest thousands in year since October 7
A ‘dark period’ of repression: Jordanian authorities arrest thousands in year since October 7
Jordan has witnessed increasing popular protests expressing solidarity with Gaza and demanding an end to normalization with Israel. The Jordanian government has responded with an unprecedented crackdown on protests and free expression.Synne Furnes Bjerkestrand (Mondoweiss)
New Age Weekly No 40. October 06–12, 2024
Japan’s new prime minister: Dreaming of an Asian version of NATO?
Japan’s new prime minister: Dreaming of an Asian version of NATO?
Japan's new prime minister envisions further militarizing the country and the region through a military alliance aimed at China and Russia.Midori Ogasawara (rabble)
like this
c-pipes: draw pipes in terminal window
gitlab.com/christosangel/c-pip…
This program written in the C language will render random coloured
zigzag lines in the terminal, while the font, speed, density and
number of lines are fully customizable.
Each line stops once it reaches the edge of the window, only for
a new line to begin.
This program was inspired by this bash script:
github.com/pipeseroni/pipes.sh
Screenshots:
Feel free to discover the endless possibilities of customization.
like this
On 8 October, the European Data Protection Board (EDPB) issued guidelines on the processing of personal data on the basis of Article 6(1)(f) of the EU General Data Protection Regulation (GDPR). This Note is a quick immediate response to the EDPB comments in that document relating to the processing of certain special categories of personal data that enjoy special protection under the GDPR, commonly referred to as “sensitive data”. Specifically, the EDPB appears to suggest that such data can be processed on the basis of the “legitimate interest” legal basis set out in Article 6(1)(f) of the GDPR, provided certain “additional conditions” for processing of sensitive data contained in Article 9(2) GDPR are met. In this note, I explain why this is not clear enough.
KORFF – GDPR – sensitive data and the legitimate interest legal basis – 241011Download
Estimated Russian army spending is between $85-$105 billion USD. (This has likely skyrocketed since that that estimate was taken as Russia has transitioned to a wartime economy.)
Chinese? ~$212-$230 billion USD.
Spending on military is better put in context of GDP, and actual spending is going to be very different than published or even estimated numbers. (It's likely much more, is what I am implying.)
I actually agree that this money is better spent on social welfare. It's a stupid situation across the board and many countries are guilty of this disparity.
For better or for worse, much of that money goes back into the overall economy of the country supplying the aid. Not all, but most. (This can get complicated due to the lifespan of specific types of munitions.)
What I am saying is that there is a ton of blame to pass around and poking at one country or another is an agenda, not a solution.
OHSU study uses imaging in neurosurgery patients to show how brain’s glymphatic system clears waste; lifestyle measures can keep system sharp
Brain’s waste-clearance pathways revealed for the first time
OHSU study uses imaging in neurosurgery patients to show how brain’s glymphatic system clears waste; lifestyle measures can keep system sharp.OHSU News
like this
like this
Ukraine’s Best Fighting Vehicles Attacked Past Veseloe In Western Russia—And Got Caught In A Brutal Ambush
Ukraine’s Best Fighting Vehicles Attacked Past Veseloe In Western Russia—And Got Caught In A Brutal Ambush
The Ukrainians lost at least one CV90, a Marder, an M-2 and a Stryker.David Axe (Forbes)
Ukraine’s Best Fighting Vehicles Attacked Past Veseloe In Western Russia—And Got Caught In A Brutal Ambush
Ukraine’s Best Fighting Vehicles Attacked Past Veseloe In Western Russia—And Got Caught In A Brutal Ambush
The Ukrainians lost at least one CV90, a Marder, an M-2 and a Stryker.David Axe (Forbes)
Did it not seem strange to them that Russia had set up no defenses along that part of the border, where the land is sparsely populated and not especially militarily strategic? Not even mine fields or an anti-tank trenches as elsewhere?
World’s oldest known (representational) artwork in Indonesian cave dated using lasers
Laser-induced imaging of radioactive elements was used to work out the age of an ancient cave painting on the Indonesian island of Sulawesi. The results reveal that the narrative scene is 51,200 years old, making it the earliest known example of representational art. This study challenges previous dating methods and suggests a deeper origin for human image-making and storytelling.
TL;DR or if you don't have access to the article: the researchers invented a faster, less-destructive and more-accurate rock art dating method & applied it to humanity's oldest known rock art in Sulawesi, Indonesia. The art is at least 51,200 years old (authors' lower estimate)!
Edit: contrary to what the news title original stated: this is the oldest representational art, not the literal oldest human-created art.
The paper itself (open access): doi.org/10.1038/s41586-024-075…
like this
Cool.
Title might be a bit clickbait though.
It’s oldest known representational art. Not oldest known art.
For example the carvings in the Blomos cave in South Africa are atleast 75’000 years old.
Edit: Thank you for editing the title! That’s pretty weird mistake by Nature I thought they had high standards. Well they have peer reviewed and approved some dodgy research in my field recently so maybe I should be more skeptical.
I love hearing stuff like this. 51, 000 years old is already insane. 75,000 years old is 24,000 years older than that. I can't even imagine 24,000 years older than today.
Why can't we get movies about this shit instead of another Marvel sequel? I want some scientifically accurate adventure about life in 73,000 BC.
Well the problem is we know very little. So a movie like that would be complete guesswork.
You might enjoy the youtube channel “Stephan Milo” though. His videos are well sourced and have a lot of expert interviews. And he focuses on this kind of stuff.
m_f
in reply to ☆ Yσɠƚԋσʂ ☆ • • •☆ Yσɠƚԋσʂ ☆
in reply to m_f • • •m_f
in reply to ☆ Yσɠƚԋσʂ ☆ • • •Gary Marcus should be disregarded because he's emotionally invested in The Bitter Lesson being wrong. He really wants LLMs to not be as good as they already are. He'll find some interesting research about "here's a limitation that we found" and turn that into "LLMS BTFO IT'S SO OVER".
The research is interesting for helping improve LLMs, but that's the extent of it. I would not be worried about the limitations the paper found for a number of reasons:
- There doesn't seem to be any reason to believe that there's a ceiling on scaling up
- LLM's reasoning abilities improve with scale (notice that the example they use for kiwis they included the answers from
- Techniques such as RAG and Chain of Thought help immensely on many probl
... show moreo1-miniandllama3-8B, which are much smaller models with much more limited capabilities. GPT-4o got the problem correct when I tested it, without any special prompting techniques or anything)Gary Marcus should be disregarded because he's emotionally invested in The Bitter Lesson being wrong. He really wants LLMs to not be as good as they already are. He'll find some interesting research about "here's a limitation that we found" and turn that into "LLMS BTFO IT'S SO OVER".
The research is interesting for helping improve LLMs, but that's the extent of it. I would not be worried about the limitations the paper found for a number of reasons:
o1-miniandllama3-8B, which are much smaller models with much more limited capabilities. GPT-4o got the problem correct when I tested it, without any special prompting techniques or anything)Until we hit a wall and really can't find a way around it for several years, this sort of research falls into the "huh, interesting" territory for anybody that isn't a researcher.
like this
SaltySalamander likes this.
☆ Yσɠƚԋσʂ ☆
in reply to m_f • • •Actually we do know that there are diminishing returns from scaling already. Furthermore, I would argue that there are inherent limits in simply using correlations in text as the basis for the model. Human reasoning isn't primarily based on language, we create an internal model of the world that acts as a shared context. The language is rooted in that model and that's what allows us to communicate effectively and understand the actual meaning behind words. Skipping that step leads to the problems we're seeing with LLMs.
That said, I agree they are a tool, and they obviously have uses. I just think that they're going to be a part of a bigger tool set going forward. Right now there's an incredible amount of hype associated with LLMs. Once the hype settles we'll know what use cases are most appropriate for them.
m_f
in reply to ☆ Yσɠƚԋσʂ ☆ • • •The whole "it's just autocomplete" is just a comforting mantra. A sufficiently advanced autocomplete is indistinguishable from intelligence. LLMs provably have a world model, just like humans do. They build that model by experiencing the universe via the medium of human-generated text, which is much more limited than human sensory input, but has allowed for some very surprising behavior already.
We're not seeing diminishing returns yet, and in fact we're going to see some interesting stuff happen as we start hooking up sensors and cameras as direct input, instead of these models building their world model indirectly through purely text. Let's see what happens in 5 years or so before saying that there's any diminishing returns.
☆ Yσɠƚԋσʂ ☆
in reply to m_f • • •I'm saying that the medium of text is not a good way to create a world model, and the problems LLMs have stem directly from people trying to do that. Just because autocomplete produces results that look fancy doesn't make it actually meaningful. These things are great for scenarios where you just want to produce something aesthetically pleasing like an image or generate some text. However, this quickly falls apart when it comes to problems where there is a specific correct answer.
Furthermore, there is plenty of progress being made with DNNs and CNNs using embodiment which looks to be far more promising than LLMs in actually producing machines that can interact with the world meaningfully. This idea that GPT is some holy grail of AI seems rather misguided to me. It's a useful tool, but there are plenty of other approaches being explored, and it's most likely that future systems will use a combination of these techniques.
JackGreenEarth
in reply to ☆ Yσɠƚԋσʂ ☆ • • •like this
SaltySalamander likes this.
Hazzard
in reply to JackGreenEarth • • •I think it is a problem. Maybe not for people like us, that understand the concept and its limitations, but "formal reasoning" is exactly how this technology is being pitched to the masses. "Take a picture of your homework and OpenAI will solve it", "have it reply to your emails", "have it write code for you". All reasoning-heavy tasks.
On top of that, Google/Bing have it answering user questions directly, it's commonly pitched as a "tutor", or an "assistant", the OpenAI API is being shoved everywhere under the sun for anything you can imagine for all kinds of tasks, and nobody is attempting to clarify it's weaknesses in their marketing.
As it becomes more and more common, more and more users who don't understand it's fundamentally incapable of reliably doing these things will crop up.
like this
missingno likes this.
☆ Yσɠƚԋσʂ ☆
in reply to JackGreenEarth • • •vrighter
in reply to ☆ Yσɠƚԋσʂ ☆ • • •☆ Yσɠƚԋσʂ ☆
in reply to vrighter • • •geekwithsoul
in reply to JackGreenEarth • • •like this
missingno likes this.
Letstakealook
in reply to ☆ Yσɠƚԋσʂ ☆ • • •slacktoid
in reply to ☆ Yσɠƚԋσʂ ☆ • • •