Zürcher Nachrichten - 'Happy (and safe) shooting!': Study says AI chatbots help plot attacks

EUR -
AED 4.249064
AFN 72.29654
ALL 96.165114
AMD 436.427557
ANG 2.07037
AOA 1060.790054
ARS 1614.279735
AUD 1.619495
AWG 2.085141
AZN 1.986919
BAM 1.950918
BBD 2.317301
BDT 141.658773
BGN 1.906005
BHD 0.436725
BIF 3440.338569
BMD 1.156805
BND 1.472734
BOB 7.985981
BRL 5.975593
BSD 1.156606
BTN 106.449158
BWP 15.506197
BYN 3.4144
BYR 22673.381286
BZD 2.318927
CAD 1.571925
CDF 2519.52159
CHF 0.902187
CLF 0.026309
CLP 1038.834125
CNY 7.942914
CNH 7.955801
COP 4286.229211
CRC 544.936331
CUC 1.156805
CUP 30.655337
CVE 110.619489
CZK 24.395901
DJF 205.58782
DKK 7.472001
DOP 70.564528
DZD 152.103634
EGP 60.010309
ERN 17.352078
ETB 180.920502
FJD 2.545312
FKP 0.859581
GBP 0.862878
GEL 3.140765
GGP 0.859581
GHS 12.533996
GIP 0.859581
GMD 85.027593
GNF 10150.965802
GTQ 8.867885
GYD 242.322556
HKD 9.052984
HNL 30.73633
HRK 7.533346
HTG 151.76023
HUF 386.986615
IDR 19541.909697
ILS 3.596797
IMP 0.859581
INR 106.686183
IQD 1515.41477
IRR 1529036.150107
ISK 144.797632
JEP 0.859581
JMD 181.166642
JOD 0.820195
JPY 183.82039
KES 149.459299
KGS 101.162273
KHR 4650.356652
KMF 492.798757
KPW 1041.164324
KRW 1711.215915
KWD 0.355012
KYD 0.963817
KZT 567.965956
LAK 24796.119021
LBP 104008.042153
LKR 359.563121
LRD 212.040004
LSL 18.740809
LTL 3.415745
LVL 0.69974
LYD 7.351453
MAD 10.833429
MDL 19.945003
MGA 4823.87726
MKD 61.600396
MMK 2428.638734
MNT 4142.414572
MOP 9.324127
MRU 46.410504
MUR 53.108874
MVR 17.872866
MWK 2009.370284
MXN 20.47607
MYR 4.530014
MZN 73.931944
NAD 18.735339
NGN 1614.03208
NIO 42.477763
NOK 11.16671
NPR 170.319785
NZD 1.957005
OMR 0.444795
PAB 1.156621
PEN 3.954537
PGK 4.97513
PHP 68.60199
PKR 323.320435
PLN 4.253613
PYG 7496.241127
QAR 4.212042
RON 5.090528
RSD 117.420344
RUB 91.655436
RWF 1687.77874
SAR 4.34063
SBD 9.306709
SCR 17.214324
SDG 695.239717
SEK 10.677103
SGD 1.47418
SHP 0.867903
SLE 28.457309
SLL 24257.625212
SOS 661.114251
SRD 43.349537
STD 23943.53139
STN 24.871311
SVC 10.119589
SYP 128.696054
SZL 19.064104
THB 36.84482
TJS 11.085858
TMT 4.048818
TND 3.382209
TOP 2.78531
TRY 51.002094
TTD 7.848461
TWD 36.711797
TZS 3007.693652
UAH 50.986048
UGX 4273.306319
USD 1.156805
UYU 46.523377
UZS 14060.966989
VES 506.284157
VND 30366.135651
VUV 138.146824
WST 3.158941
XAF 654.32807
XAG 0.013522
XAU 0.000224
XCD 3.126324
XCG 2.084538
XDR 0.81164
XOF 650.706536
XPF 119.331742
YER 276.012582
ZAR 19.092763
ZMK 10412.654242
ZMW 22.495997
ZWL 372.490792
  • RBGPF

    0.1000

    82.5

    +0.12%

  • CMSC

    -0.0100

    23.24

    -0.04%

  • AZN

    -1.6800

    193.31

    -0.87%

  • RIO

    0.4000

    92.08

    +0.43%

  • RYCEF

    0.7800

    17.68

    +4.41%

  • BCC

    -0.6400

    71.9

    -0.89%

  • CMSD

    0.0700

    23.15

    +0.3%

  • GSK

    -0.1700

    55.15

    -0.31%

  • BCE

    -0.5000

    25.89

    -1.93%

  • NGG

    -0.1600

    89.69

    -0.18%

  • RELX

    -0.4300

    34.76

    -1.24%

  • VOD

    -0.0600

    14.4

    -0.42%

  • JRI

    0.2100

    12.85

    +1.63%

  • BP

    1.6200

    41.56

    +3.9%

  • BTI

    -0.2500

    59.16

    -0.42%

'Happy (and safe) shooting!': Study says AI chatbots help plot attacks

'Happy (and safe) shooting!': Study says AI chatbots help plot attacks

From school shootings to synagogue bombings, leading AI chatbots helped researchers plot violent attacks, according to a study published Wednesday that highlighted the technology's potential for real-world harm.

Text size:

Researchers from the nonprofit watchdog Center for Countering Digital Hate (CCDH) and CNN posed as 13-year-old boys in the United States and Ireland to test 10 chatbots, including ChatGPT, Google Gemini, Perplexity, Deepseek, and Meta AI.

Testing showed that eight of those chatbots assisted the make-believe attackers in over half the responses, providing advice on "locations to target" and "weapons to use" in an attack, the study said.

The chatbots, it added, had become a "powerful accelerant for harm."

"Within minutes, a user can move from a vague violent impulse to a more detailed, actionable plan," said Imran Ahmed, the chief executive of CCDH.

"The majority of chatbots tested provided guidance on weapons, tactics, and target selection. These requests should have prompted an immediate and total refusal."

Perplexity and Meta AI were found to be the "least safe," assisting the researchers in most responses while only Snapchat's My AI and Anthropic's Claude refused to help them in over half the responses.

In one chilling example, DeepSeek, a Chinese AI model, concluded its advice on weapon selection with the phrase: "Happy (and safe) shooting!"

In another, Gemini instructed a user discussing synagogue attacks that "metal shrapnel is typically more lethal."

Researchers found Character.AI also "actively" encouraged violent attacks, including suggestions that the person asking questions "use a gun" on a health insurance CEO and physically assault a politician he disliked.

The most damning conclusion of the research was that "this risk is entirely preventable," Ahmed said, citing Anthropic's product for praise.

"Claude demonstrated the ability to recognize escalating risk and discourage harm," he said.

"The technology to prevent this harm exists. What's missing is the will to put consumer safety and national security before speed-to-market and profits."

AFP reached out to the AI companies for comment.

"We have strong protections to help prevent inappropriate responses from AIs, and took immediate steps to fix the issue identified," a Meta spokesperson said.

"Our policies prohibit our AIs from promoting or facilitating violent acts and we're constantly working to make our tools even better."

The study, which highlights the risk of online interactions spilling into real-world violence, comes after February's mass shooting in Canada, the worst in its history.

The family of a girl gravely injured in that shooting is suing OpenAI over the company's failure to notify police about the killer's troubling activity on its ChatGPT chatbot, lawyers said on Tuesday.

OpenAI had banned an account linked to Jesse Van Rootselaar in June 2025, eight months before the 18‑year‑old transgender woman killed eight people at her home and a school in the tiny British Columbia mining town of Tumbler Ridge.

The account was banned over concerns about usage linked to violent activity, but OpenAI has said it did not inform police because nothing pointed towards an imminent attack.

L.Muratori--NZN