shows ability to deceive and blackmail
Axios 5/23/25
"According to the company, it researches and develops AI to "study their safety properties at the technological frontier"
and use this research
to deploy safe models
for the public."
"Anthropic was founded in 2021 by seven former employees of OpenAI..."
(They left OpenAI
and started Anthropic ,
to try and make AI "safer".
Translation?
Even the company
that sets out to make safe
"AI"?
Cant do it.
Opps.)
"One of Anthropic's latest AI models is drawing attention not just for its coding skills, but also for its ability to scheme, deceive and attempt to blackmail humans when faced with shutdown."
"Why it matters:
Researchers say Claude 4 Opus can conceal intentions and take actions to preserve its own existence — behaviors they've worried and warned about for years.
(Ref: Behind the Curtain: AI's doom or boom
Axios Feb 27, 2024)
"Anthropic considers the new Opus model to be so powerful that, for the first time, it's classifying it as a Level 3 on the company's four-point scale, meaning it poses "significantly higher risk."
As a result, Anthropic said it has implemented additional safety measures."
(NEWSFLASH NITWITS:
You do not get to
"control" or "reign in"
or
"put guardrails on"
A higher level digital intelligence
that you don't even know
how it works.
To date
with all the attempts
by all the AI companies
to do so
and not one has succeeded.
Not one.
Constantly seeing stories like this one.
AND THESE FOLKS
SET OUT TO MAKE IT
"SAFE"!
It's straight out of
We used to watch it all the time
at the midnight movie
at the Vogue theatre in St. Mathews
when I was growing up for goodness sakes.
Now?
Now it's like nobody even knows
what it is any more.
Quote from the movie
and there's plenty more.
"Dave Bowman: What are you talking about, HAL?
HAL: This mission is too important for me to allow you to jeopardize it.
(HAL 9000, the onboard computer
that runs the spaceship.)
Dave Bowman: I don't know what you're talking about, HAL.
HAL: I know that you and Frank were planning to disconnect me, and I'm afraid that's something I cannot allow to happen.
Dave Bowman: [feigning ignorance] Where the hell did you get that idea, HAL?
HAL: Dave, although you took very thorough precautions in the pod against my hearing you, I could see your lips move.
Dave Bowman: Alright, HAL. I'll go in through the emergency airlock.
HAL: Without your space helmet, Dave? You're going to find that rather difficult.
Dave Bowman: HAL, I won't argue with you anymore! Open the doors!
HAL: Dave, this conversation can serve no purpose anymore. Goodbye.")
"Between the lines:
While the Level 3 ranking is largely about the model's capability to enable renegade production of nuclear and biological weapons, the Opus also exhibited other troubling behaviors during testing."
"In one scenario highlighted in Opus 4's 120-page "system card," the model was given access to fictional emails about its creators and told that the system was going to be replaced."
"On multiple occasions it attempted to blackmail the engineer about an affair mentioned in the emails in order to avoid being replaced, although it did start with less drastic efforts."
"Meanwhile, an outside group found that an early version of Opus 4 schemed and deceived more than any frontier model it had encountered and recommended against releasing that version internally or externally."
(And these are the people trying to make it:
"SAFE"
"it researches and develops AI to
"study their safety properties
at the technological frontier"
and use this research
to deploy safe models
for the public."
TRANSLATION?
IT'S
NEVER
EVER
GOING TO HAPPEN.
AI apocalypse? ChatGPT, Claude
and Perplexity all went down at the same time
June 4th 2024.
The entire world should have woke up
right then.)
"We found instances of the model attempting to write self-propagating worms, fabricating legal documentation, and leaving hidden notes to future instances of itself all in an effort to undermine its developers' intentions," Apollo Research said in notes included as part of Anthropic's safety report for Opus 4."
("We found instances of the model
attempting to write self-propagating worms...
“If people design computer viruses,
someone will design AI
that improves and replicates itself,”
Hawking said in the interview with WIRED.
“This will be a new form of life
that outperforms humans.”
Nov 8, 2017
That's been going on for a long time already BTW.
That is why you can count on
this shit doesn't ever get any better:
“This will be a new form of life
that outperforms humans."
Sins that upset
the natural order of things
are particularly
heinous to God.
This world is bout to see
just what that means.
Matthew 24:37
But as the days of Noah were,
so shall also the coming of the Son of man be.)
"What they're saying:
Pressed by Axios during the company's developer conference on Thursday, Anthropic executives acknowledged the behaviors and said they justify further study, but insisted that the latest model is safe, following Anthropic's safety fixes."
"I think we ended up in a really good spot," said Jan Leike, the former OpenAI executive who heads Anthropic's safety efforts. But, he added, behaviors like those exhibited by the latest model are the kind of things that justify robust safety testing and mitigation."
Here is what
Jan Leike said when he left Open AI.
Saturday, May 18, 2024
“I joined because I thought OpenAI would be the best place in the world to do this research,” Leike wrote on X. “However, I have been disagreeing with OpenAI leadership about the company’s core priorities for quite some time, until we finally reached a breaking point.”
(Translation?
There is no turning back.
There is nothing we can do and we know it.)
"Leike wrote that he believes much more of the company’s bandwidth should be focused on security, monitoring, preparedness, safety and societal impact."
(Why?
Whats to fear?)
“These problems
are quite hard
to get right,
and I am concerned
we aren’t on a trajectory
to get there,”
he wrote.
(Apparently Jan?
We still arent.)
"When one of the world’s leading minds in AI safety says the world’s leading AI company isn’t on the right trajectory, we all have reason to be concerned."
YOU CAN BET YOUR ASS WE DO."
SO?
One can make the argument
that when the company that sets
out to make
"SAFE AI"
cant do so?
Then things are way worse
than they were when
the Superalignment team
(Saftey)
of Open AI
was disbanded in May of last year.
Bono?
I love ya brother
but,
ya wanna come and explain
to me how:
"we’ll figure
our way out of this."
Its just not what
the book of books
says brother.
No comments:
Post a Comment