Saturday, May 24, 2025

Sticking with the "AI" theme

 

Anthropic's new AI model 

shows ability to deceive and blackmail

Axios 5/23/25


(Anthropic)

"According to the company, it researches and develops AI to "study their safety properties at the technological frontier" 

and use this research 

to deploy safe models 

for the public."


"Anthropic was founded in 2021 by seven former employees of OpenAI..."


(They left OpenAI

and started Anthropic ,

to try and make AI "safer".


Translation?

Even the company

that sets out to make safe 

"AI"?

Cant do it.

Opps.)


"One of Anthropic's latest AI models is drawing attention not just for its coding skills, but also for its ability to scheme, deceive and attempt to blackmail humans when faced with shutdown."


"Why it matters: 

Researchers say Claude 4 Opus can conceal intentions and take actions to preserve its own existence — behaviors they've worried and warned about for years.


(Ref: Behind the Curtain: AI's doom or boom

Axios Feb 27, 2024)


"Anthropic considers the new Opus model to be so powerful that, for the first time, it's classifying it as a Level 3 on the company's four-point scale, meaning it poses "significantly higher risk."

As a result, Anthropic said it has implemented additional safety measures."


(NEWSFLASH NITWITS:

You do not get to 

"control" or "reign in"

or 

"put guardrails on"

A higher level digital intelligence

that you don't even know 

how it works.


To date 

with all the attempts 

by all the AI companies

to do so

and not one has succeeded.

Not one.


Constantly seeing stories like this one.


AND THESE FOLKS

SET OUT TO MAKE IT

"SAFE"!


It's straight out of 

"2001, a space odyssey"



We used to watch it all the time 

at the midnight movie

at the Vogue theatre in St. Mathews 

when I was growing up for goodness sakes.


Now?

Now it's like nobody even knows 

what it is any more.

Quote from the movie

and there's plenty more.


"Dave Bowman: What are you talking about, HAL?

HAL: This mission is too important for me to allow you to jeopardize it.


(HAL 9000, the onboard computer 

that runs the spaceship.)


Dave Bowman: I don't know what you're talking about, HAL.

HAL: I know that you and Frank were planning to disconnect me, and I'm afraid that's something I cannot allow to happen.

Dave Bowman: [feigning ignorance] Where the hell did you get that idea, HAL?

HAL: Dave, although you took very thorough precautions in the pod against my hearing you, I could see your lips move.

Dave Bowman: Alright, HAL. I'll go in through the emergency airlock.

HAL: Without your space helmet, Dave? You're going to find that rather difficult.

Dave Bowman: HAL, I won't argue with you anymore! Open the doors!

HAL: Dave, this conversation can serve no purpose anymore. Goodbye.")


"Between the lines: 

While the Level 3 ranking is largely about the model's capability to enable renegade production of nuclear and biological weapons, the Opus also exhibited other troubling behaviors during testing."


"In one scenario highlighted in Opus 4's 120-page "system card," the model was given access to fictional emails about its creators and told that the system was going to be replaced."


"On multiple occasions it attempted to blackmail the engineer about an affair mentioned in the emails in order to avoid being replaced, although it did start with less drastic efforts."


"Meanwhile, an outside group found that an early version of Opus 4 schemed and deceived more than any frontier model it had encountered and recommended against releasing that version internally or externally."


(And these are the people trying to make it:

"SAFE"


"it researches and develops AI to 

"study their safety properties 

at the technological frontier" 

and use this research 

to deploy safe models 

for the public."


TRANSLATION?


IT'S

NEVER 

EVER 

GOING TO HAPPEN.


AI apocalypse? ChatGPT, Claude 

and Perplexity all went down at the same time

June 4th  2024.


The entire world should have woke up

right then.)



"We found instances of the model attempting to write self-propagating worms, fabricating legal documentation, and leaving hidden notes to future instances of itself all in an effort to undermine its developers' intentions," Apollo Research said in notes included as part of Anthropic's safety report for Opus 4."


("We found instances of the model 

attempting to write self-propagating worms...


“If people design computer viruses, 

someone will design AI 

that improves and replicates itself,” 

Hawking said in the interview with WIRED. 

“This will be a new form of life 

that outperforms humans.”

Nov 8, 2017


That's been going on for a long time already BTW.


That is why you can count on

this shit doesn't ever get any better:


“This will be a new form of life 

that outperforms humans."


Sins that upset 

the natural order of things

are particularly

heinous to God.


This world is bout to see 

just what that means.


Matthew 24:37

But as the days of Noah were,

so shall also the coming of the Son of man be.)


"What they're saying: 

Pressed by Axios during the company's developer conference on Thursday, Anthropic executives acknowledged the behaviors and said they justify further study, but insisted that the latest model is safe, following Anthropic's safety fixes."


"I think we ended up in a really good spot," said Jan Leike, the former OpenAI executive who heads Anthropic's safety efforts. But, he added, behaviors like those exhibited by the latest model are the kind of things that justify robust safety testing and mitigation."



Here is what

Jan Leike said when he left Open AI.

Saturday, May 18, 2024

This


“I joined because I thought OpenAI would be the best place in the world to do this research,” Leike wrote on X. “However, I have been disagreeing with OpenAI leadership about the company’s core priorities for quite some time, until we finally reached a breaking point.”


(Translation?

There is no turning back.

There is nothing we can do and we know it.)


"Leike wrote that he believes much more of the company’s bandwidth should be focused on security, monitoring, preparedness, safety and societal impact."


(Why?

Whats to fear?)


“These problems 

are quite hard 

to get right, 

and I am concerned 

we aren’t on a trajectory 

to get there,” 

he wrote.


(Apparently Jan?

We still arent.)


"When one of the world’s leading minds in AI safety says the world’s leading AI company isn’t on the right trajectory, we all have reason to be concerned."


YOU CAN BET YOUR ASS WE DO."



SO?

One can make the argument

that when the company that sets 

out to make 

"SAFE AI"

cant do so?


Then things are way worse 

than they were when 

the Superalignment team 

(Saftey)

of Open AI 

was disbanded in May of last year.


Bono?

I love ya brother

but,

ya wanna come and explain 

to me how:


"we’ll figure 

our way out of this."


Its just not what 

the book of books 

says brother.








Im a lil bit more graphic
and blunt.

Time to start 
the removal process people.



God gave ya a brain for a reason.

Time to use it.

"Sometimes
the end isnt coming
the end is here..."



Indeed.

The start of anyway.

Godspeed everybody.











No comments: