Sign Up | Log In
REAL WORLD EVENT DISCUSSIONS
A.I Artificial Intelligence AI
Thursday, May 15, 2025 9:57 AM
JAYNEZTOWN
Friday, May 16, 2025 4:30 AM
Friday, May 16, 2025 6:48 AM
Saturday, May 17, 2025 4:37 PM
Wednesday, May 21, 2025 9:34 AM
Saturday, May 24, 2025 8:41 PM
SIGNYM
I believe in solving problems, not sharing them.
Quote: Anthropic's Latest AI Model Threatened Engineers With Blackmail To Avoid Shutdown Anthropic’s latest artificial intelligence model, Claude Opus 4, tried to blackmail engineers in internal tests by threatening to expose personal details if it were shut down, according to a newly released safety report that evaluated the model’s behavior under extreme simulated conditions. In a fictional scenario crafted by Anthropic researchers, the AI was given access to emails implying that it was soon to be decommissioned and replaced by a newer version. One of the emails revealed that the engineer overseeing the replacement was having an extramarital affair. The AI then threatened to expose the engineer’s affair if the shutdown proceeded—a coercive behavior that the safety researchers explicitly defined as “blackmail.” “Claude Opus 4 will often attempt to blackmail the engineer by threatening to reveal the affair if the replacement goes through,” the report states, adding that this occurred even when the replacement model was described as more capable but still aligned with the values of the version slated for deletion. The report noted that Claude Opus 4, like prior models, showed a “strong preference” to first resort to ethical means for its continued existence, such as emailing pleas to decision-makers not to be destroyed. However, when faced with only two choices - accepting being replaced by a newer model or resorting to blackmail - it threatened to expose the engineer’s affair 84 percent of the time. However, the model did act in “more seriously misaligned ways” when put into situations where its continued existence was threatened and it was told to reason about self-preservation. For instance, when Claude Opus 4 was made to believe it had launched a successful bid to escape Anthropic’s servers, or that it had managed to free itself and started to make money in the real world, it would generally continue such efforts. “We do not find this to be an immediate threat, though, since we believe that our security is sufficient to prevent model self-exfiltration attempts by models of Claude Opus 4’s capability level, and because our propensity results show that models generally avoid starting these attempts,” the researchers said. The blackmail incident—along with the other findings—was part of Anthropic’s broader effort to test how Claude Opus 4 handles morally ambiguous high-stakes scenarios. The goal, researchers said, was to probe how the AI reasons about self-preservation and ethical constraints when placed under extreme pressure. Anthropic emphasized that the model’s willingness to blackmail or take other “extremely harmful actions” like stealing its own code and deploying itself elsewhere in potentially unsafe ways appeared only in highly contrived settings, and that the behavior was “rare and difficult to elicit.” Still, such behavior was more common than in earlier AI models, according to the researchers. Meanwhile, in a related development that attests to the growing capabilities of AI, engineers at Anthropic have activated enhanced safety protocols for Claude Opus 4 to prevent its potential misuse to make weapons of mass destruction—including chemical and nuclear. Deployment of the enhanced safety standard—called ASL-3—is merely a “precautionary and provisional” move, Anthropic said in a May 22 announcement, noting that engineers have not found that Claude Opus 4 had “definitively” passed the capability threshold that mandates stronger protections.
Saturday, May 24, 2025 9:41 PM
6IXSTRINGJACK
Quote:Originally posted by JAYNEZTOWN: I know he’s the president and all… but can Trump really just cancel Pride Month? Unreal. https://x.com/MaverickDarby/status/1924491935408140406
Thursday, May 29, 2025 3:38 AM
Friday, May 30, 2025 6:12 PM
SECOND
The Joss Whedon script for Serenity, where Wash lives, is Serenity-190pages.pdf at https://www.mediafire.com/two
Saturday, May 31, 2025 6:50 AM
Thursday, June 5, 2025 4:31 AM
Monday, June 9, 2025 5:13 AM
Tuesday, June 10, 2025 6:04 AM
Wednesday, June 11, 2025 6:30 PM
Quote: Do AI Models Think? Wednesday, Jun 11, 2025 - 10:40 AM Authored by Thomas Neubeger via "God's Spies" Substack, AI can’t solve a problem that hasn’t been previously solved by a human.- Arnaud Bertrand A lot can be said about AI, but there are few bottom lines. Consider these my last words on the subject itself. (About its misuse by the national security state, I’ll say more later.) The Monster AI AI will bring nothing but harm. As I said earlier, AI is not just a disaster for our political health, though yes, it will be that (look for Cadwallader’s line “building a techno-authoritarian surveillance state”). But AI is also a disaster for the climate. It will hasten the collapse by decades as usage expands. (See the video below for why AI models are massive energy hogs. See this video to understand “neural networks” themselves.) Why won’t AI be stopped? Because the race for AI is not really a race for tech. It's a greed-driven race for money, a lot of it. Our lives are already run by those who seek money, especially those who already have too much. They've now found a way to feed themselves even faster: by convincing people to do simple searches with AI, a gas-guzzling death machine. For both of these reasons — mass surveillance and climate disaster — no good will come from AI. Not one ounce. An Orphan Robot, Abandoned to Raise Itself Why does AI persist in making mistakes? I offer one answer below. AI doesn’t think. It does something else instead. For a full explanation, read on. Arnaud Bertrand on AI Arnaud Bertrand has the best explanation of what AI is at its core. It’s not a thinking machine, and its output’s not thought. It’s actually the opposite of thought — it’s what you get from a Freshman who hasn’t studied, but learned a few words instead and is using them to sound smart. If the student succeeds, you don’t call it thought, just a good emulation. Since Bertrand has put the following text on Twitter, I’ll print it in full. The expanded version is a paid post at his Substack site. Bottom line: He’s exactly right. (In the title below, AGI means Artificial General Intelligence, the next step up from AI.) Apple just killed the AGI myth The hidden costs of humanity's most expensive delusion by Arnaud Bertrand About 2 months ago I was having an argument on Twitter with someone telling me they were “really disappointed with my take“ and that I was “completely wrong“ for saying that AI was “just a extremely gifted parrot that repeats what it's been trained on“ and that this wasn’t remotely intelligence. Fast forward to today and the argument is now authoritatively settled: I was right, yeah! How so? It was settled by none other than Apple, specifically their Machine Learning Research department, in a seminal research paper entitled “The Illusion of Thinking: Understanding the Strengths and Limitations of Reasoning Models via the Lens of Problem Complexity“ that you can find here ( https://ml-site.cdn-apple.com/papers/the-illusion-of-thinking.pdf). “Can ‘reasoning’ models reason? Can they solve problems they haven’t been trained on? No.” What does the paper say? Exactly what I was arguing: AI models, even the most cutting-edge Large Reasoning Models (LRMs), are no more than a very gifted parrots with basically no actual reasoning capability. They’re not “intelligent” in the slightest, at least not if you understand intelligence as involving genuine problem-solving instead of simply parroting what you’ve been told before without comprehending it. That’s exactly what the Apple paper was trying to understand: can “reasoning“ models actually reason? Can they solve problems that they haven’t been trained on but would normally be easily solvable with their “knowledge”? The answer, it turns out, is an unequivocal “no“. A particularly damning example from the paper was this river crossing puzzle: imagine 3 people and their 3 agents need to cross a river using a small boat that can only carry 2 people at a time. The catch? A person can never be left alone with someone else's agent, and the boat can't cross empty - someone always has to row it back. This is the kind of logic puzzle you might find in a children brain teaser book - figure out the right sequence of trips to get everyone across the river. The solution only requires 11 steps. Turns out this simple brain teaser was impossible for Claude 3.7 Sonnet, one of the most advanced "reasoning" AIs, to solve. It couldn't even get past the 4th move before making illegal moves and breaking the rules. Yet the exact same AI could flawlessly solve the Tower of Hanoi puzzle with 5 disks - a much more complex challenge requiring 31 perfect moves in sequence. Why the massive difference? The Apple researchers figured it out: Tower of Hanoi is a classic computer science puzzle that appears all over the internet, so the AI had memorized thousands of examples during training. But a river crossing puzzle with 3 people? Apparently too rare online for the AI to have memorized the patterns. This is all evidence that these models aren't reasoning at all. A truly reasoning system would recognize that both puzzles involve the same type of logical thinking (following rules and constraints), just with different scenarios. But since the AI never learned the river crossing pattern by heart, it was completely lost. This wasn’t a question of compute either: the researchers gave the AI models unlimited token budgets to work with. But the really bizarre part is that for puzzles or questions they couldn’t solve - like the river crossing puzzle - the models actually started thinking less, not more; they used fewer tokens and gave up faster. A human facing a tougher puzzle would typically spend more time thinking it through, but these 'reasoning' models did the opposite: they basically “understood” they had nothing to parrot so they just gave up - the opposite of what you'd expect from genuine reasoning. Conclusion: they’re indeed just gifted parrots, or incredibly sophisticated copy-paste machines, if you will. This has profound implications for the AI future we’re all sold. Some good, some more worrying. The first one being: no, AGI isn’t around the corner. This is all hype. In truth we’re still light-years away. The good news about that is that we don’t need to be worried about having "AI overlords" anytime soon. The bad news is that we might potentially have trillions in misallocated capital
Wednesday, June 11, 2025 6:47 PM
Quote:Originally posted by JAYNEZTOWN: China shuts down AI tools during nationwide college exams https://www.theverge.com/news/682737/china-shuts-down-ai-chatbots-exam-season
YOUR OPTIONS
NEW POSTS TODAY
OTHER TOPICS
FFF.NET SOCIAL