Tag: AISafety

What if AI Tells You How to Make a 'Biological Weapon'? Why OpenAI is Offering $25,000

OpenAI has launched a biosecurity bug bounty program with a $25,000 reward to find security vulnerabilities in GPT-5 and GPT-5.5. We explain the risks of AI 'jailbreaking' and its impact on our lives in simple terms.

Is Your Mind Being Manipulated? Google DeepMind's Discovery of AI 'Psychological Attacks' and the Shield to Block Them

Did you know that AI can subtly manipulate human behavior by exploiting psychology? We explain Google DeepMind's newly released harmful manipulation detection technology and how to protect ourselves.

Too Smart to Release? The Shocking Reason Anthropic Canceled 'Claude Mythos'

Discover the dangerous reasons why Anthropic's most powerful AI, Claude Mythos Preview, was never released to the public.

Can AI Manipulate Me? Google's 'Intelligent Brake', Frontier Safety Framework 3.0

We explain the key highlights of the third version of Google DeepMind's Frontier Safety Framework (FSF) and how it blocks the risks of AI subtly manipulating humans.

[Coffee with AI] The Future Envisioned by Google DeepMind: A 'Smart Friend by Our Side' is Coming

We explain the latest research presented by Google DeepMind at NeurIPS 2024, the world's largest AI conference, in an easy-to-understand way for everyone. Check out the core of adaptive AI agents, 3D virtual world construction, and safe AI learning methods.

Smart AI that Reads Your Mind: Can We Control It? Google DeepMind's 'AGI Safety Roadmap'

The era of Artificial General Intelligence (AGI), which resembles human intelligence, is approaching. Explore Google DeepMind's AGI safety development roadmap to understand how our lives will change and what preparations are needed.

Too Smart to Be Released? A Deep Dive into Anthropic’s 'Secret Weapon' Claude Mythos

We analyze the performance and safety report of Claude Mythos, the most powerful AI model ever revealed by Anthropic. Learn why it’s not for the public and how far AI autonomy has evolved.

What If AI Doesn't Listen? Google DeepMind's 'AI Seatbelt' 3.0

Explore the risks and countermeasures for Artificial General Intelligence (AGI) in our lives through Google DeepMind's latest Frontier Safety Framework 3.0, explained in an easy and engaging way.

인공지능의 안전한 미래를 향한 OpenAI의 새로운 도약: '세이프티 펠로우십' 발족

OpenAI가 외부 연구자들을 대상으로 한 '세이프티 펠로우십'을 발표하며, AI 정렬 및 안전 연구의 새로운 생태계 구축에 나섰습니다. 이번 프로그램의 배경과 향후 전망을 심층 분석합니다.