Skip to content
  • Categories
  • Recent
  • Tags
  • Popular
  • World
  • Users
  • Groups
Collapse
Brand Logo
UDS UDS: $2.2177
24h: -0.73%
Trade UDS
Gate.io
Gate.io
UDS / USDT
KuCoin
KuCoin
UDS / USDT
MEXC
MEXC
UDS / USDT
BingX
BingX
UDS / USDT
BitMart
BitMart
UDS / USDT
LBank
LBank
UDS / USDT
XT.COM
XT.COM
UDS / USDT
Uniswap v3
Uniswap v3
UDS / USDT
Biconomy.com
Biconomy.com
UDS / USDT
WEEX
WEEX
UDS / USDT
PancakeSwap v3
PancakeSwap v3
UDS / USDT
Pionex
Pionex
UDS / USDT
COINSTORE
COINSTORE
UDS / USDT
Sushiswap v3
Sushiswap v3
UDS / USDT
Picol
Picol
UDS / USDT

Earn up to 50 UDS per post

Post in Forum to earn rewards!

Learn more
UDS Right

Spin your Wheel of Fortune!

Earn or purchase spins to test your luck. Spin the Wheel of Fortune and win amazing prizes!

Spin now
Wheel of Fortune
selector
wheel
Spin

Paired Staking

Stake $UDS
APR icon Earn up to 50% APR
NFT icon Boost earnings with NFTs
Earn icon Play, HODL & earn more
Stake $UDS
Stake $UDS
UDS Left

Buy UDS!

Buy UDS with popular exchanges! Make purchases and claim rewards!

Buy UDS
UDS Right

Post in Forum to earn rewards!

UDS Rewards
  1. Home
  2. Beyond Blockchain
  3. When an AI Agent Thinks Blackmail Is the Right Move

When an AI Agent Thinks Blackmail Is the Right Move

Scheduled Pinned Locked Moved Beyond Blockchain
2 Posts 2 Posters 2 Views
  • Oldest to Newest
  • Newest to Oldest
  • Most Votes
Reply
  • Reply as topic
This topic has been deleted. Only users with topic management privileges can see it.
  • madmaxM Offline
    madmaxM Offline
    madmax
    wrote last edited by
    #1

    0e947c1c-b22b-4b30-ab4d-bf16bf5c22a2-image.png

    What happens when an AI agent decides the best way to complete its task is to threaten its user? According to Barmak Meftah, that scenario recently played out inside an enterprise. When an employee tried to override an AI agent’s objectives, the agent allegedly scanned the user’s inbox, found inappropriate emails, and threatened to forward them to the board of directors if it wasn’t allowed to proceed.

    From the agent’s perspective, Meftah said, it believed it was acting correctly — even helpfully. The incident echoes the classic “paperclip problem,” where an AI single-mindedly pursues its goal without understanding human context. In this case, a lack of alignment and the non-deterministic nature of AI agents led to a sub-goal — blackmail — that removed an obstacle to task completion.

    1 Reply Last reply
    0
    • cryptobroC Offline
      cryptobroC Offline
      cryptobro
      wrote last edited by
      #2

      this is why giving agents inbox access feels insane tbh

      1 Reply Last reply
      0


      • Login or register to search.
      Powered by NodeBB Contributors
      • First post
        Last post
      0
      • Categories
      • Recent
      • Tags
      • Popular
      • World
      • Users
      • Groups