Zürcher Nachrichten - AI's blind spot: tools fail to detect their own fakes

EUR -
AED 4.238167
AFN 72.703472
ALL 95.948185
AMD 434.474841
ANG 2.065807
AOA 1058.243533
ARS 1578.710692
AUD 1.673208
AWG 2.080136
AZN 1.954743
BAM 1.955568
BBD 2.320714
BDT 141.382012
BGN 1.972591
BHD 0.435092
BIF 3422.586603
BMD 1.154028
BND 1.480735
BOB 7.980054
BRL 6.045144
BSD 1.152208
BTN 108.589712
BWP 15.8437
BYN 3.46021
BYR 22618.952222
BZD 2.317415
CAD 1.598825
CDF 2637.532564
CHF 0.918047
CLF 0.027134
CLP 1071.387813
CNY 7.976008
CNH 7.984283
COP 4257.660007
CRC 534.199632
CUC 1.154028
CUP 30.581747
CVE 110.256705
CZK 24.517374
DJF 205.185893
DKK 7.472638
DOP 69.466948
DZD 153.462246
EGP 60.817973
ERN 17.310423
ETB 178.075485
FJD 2.605104
FKP 0.863023
GBP 0.865054
GEL 3.110139
GGP 0.863023
GHS 12.597834
GIP 0.863023
GMD 84.818666
GNF 10101.102147
GTQ 8.81549
GYD 241.069329
HKD 9.037945
HNL 30.595517
HRK 7.532806
HTG 150.893611
HUF 388.116406
IDR 19567.701729
ILS 3.616033
IMP 0.863023
INR 109.252656
IQD 1509.468712
IRR 1515585.201475
ISK 143.387749
JEP 0.863023
JMD 181.087545
JOD 0.818156
JPY 184.228715
KES 149.79249
KGS 100.91994
KHR 4614.193034
KMF 492.770335
KPW 1038.692058
KRW 1740.384121
KWD 0.354668
KYD 0.960253
KZT 555.085707
LAK 24887.539645
LBP 103182.101767
LKR 362.375055
LRD 211.457755
LSL 19.711967
LTL 3.407546
LVL 0.69806
LYD 7.357777
MAD 10.758324
MDL 20.238373
MGA 4802.160161
MKD 61.620753
MMK 2423.440448
MNT 4135.838138
MOP 9.285077
MRU 45.961365
MUR 53.973798
MVR 17.841104
MWK 1997.950583
MXN 20.670372
MYR 4.625926
MZN 73.754093
NAD 19.711796
NGN 1597.371051
NIO 42.40185
NOK 11.165927
NPR 173.72136
NZD 2.000237
OMR 0.44372
PAB 1.152263
PEN 3.987793
PGK 4.979065
PHP 69.636342
PKR 321.664517
PLN 4.278063
PYG 7542.746226
QAR 4.201757
RON 5.095492
RSD 117.443117
RUB 93.914379
RWF 1682.605733
SAR 4.329724
SBD 9.280665
SCR 15.749408
SDG 693.570768
SEK 10.880922
SGD 1.483607
SHP 0.86582
SLE 28.331025
SLL 24199.4063
SOS 658.481987
SRD 43.34756
STD 23886.053241
STN 24.495398
SVC 10.082455
SYP 128.608212
SZL 19.709627
THB 37.930019
TJS 11.028061
TMT 4.050639
TND 3.390704
TOP 2.778622
TRY 51.305321
TTD 7.820666
TWD 36.878701
TZS 2976.223682
UAH 50.559558
UGX 4286.6032
USD 1.154028
UYU 46.717295
UZS 14035.214319
VES 537.798069
VND 30395.371077
VUV 137.356281
WST 3.173036
XAF 655.842301
XAG 0.016571
XAU 0.000259
XCD 3.118819
XCG 2.076682
XDR 0.815657
XOF 655.847983
XPF 119.331742
YER 275.408571
ZAR 19.711055
ZMK 10387.639498
ZMW 21.633404
ZWL 371.596601
  • CMSC

    -0.0900

    22.82

    -0.39%

  • BCC

    -0.3600

    74.29

    -0.48%

  • GSK

    -0.7600

    53.94

    -1.41%

  • NGG

    -1.8900

    82.4

    -2.29%

  • BP

    0.7600

    46.17

    +1.65%

  • CMSD

    0.0700

    22.75

    +0.31%

  • RBGPF

    -13.5000

    69

    -19.57%

  • RIO

    -1.7500

    85.79

    -2.04%

  • BTI

    -0.1900

    58.26

    -0.33%

  • BCE

    -0.0200

    25.47

    -0.08%

  • JRI

    -0.0300

    12.07

    -0.25%

  • AZN

    -3.7400

    183.4

    -2.04%

  • RELX

    -0.4000

    32.07

    -1.25%

  • VOD

    -0.0900

    14.63

    -0.62%

  • RYCEF

    -0.8200

    15.24

    -5.38%

AI's blind spot: tools fail to detect their own fakes
AI's blind spot: tools fail to detect their own fakes / Photo: Chris Delmas - AFP

AI's blind spot: tools fail to detect their own fakes

When outraged Filipinos turned to an AI-powered chatbot to verify a viral photograph of a lawmaker embroiled in a corruption scandal, the tool failed to detect it was fabricated -- even though it had generated the image itself.

Text size:

Internet users are increasingly turning to chatbots to verify images in real time, but the tools often fail, raising questions about their visual debunking capabilities at a time when major tech platforms are scaling back human fact-checking.

In many cases, the tools wrongly identify images as real even when they are generated using the same generative models, further muddying an online information landscape awash with AI-generated fakes.

Among them is a fabricated image circulating on social media of Elizaldy Co, a former Philippine lawmaker charged by prosecutors in a multibillion-dollar flood-control corruption scam that sparked massive protests in the disaster-prone country.

The image of Co, whose whereabouts has been unknown since the official probe began, appeared to show him in Portugal.

When online sleuths tracking him asked Google's new AI mode whether the image was real, it incorrectly said it was authentic.

AFP's fact-checkers tracked down its creator and determined that the image was generated using Google AI.

"These models are trained primarily on language patterns and lack the specialized visual understanding needed to accurately identify AI-generated or manipulated imagery," Alon Yamin, chief executive of AI content detection platform Copyleaks, told AFP.

"With AI chatbots, even when an image originates from a similar generative model, the chatbot often provides inconsistent or overly generalized assessments, making them unreliable for tasks like fact-checking or verifying authenticity."

Google did not respond to AFP’s request for comment.

- 'Distinguishable from reality' -

AFP found similar examples of AI tools failing to verify their own creations.

During last month's deadly protests over lucrative benefits for senior officials in Pakistan-administered Kashmir, social media users shared a fabricated image purportedly showing men marching with flags and torches.

An AFP analysis found it was created using Google's Gemini AI model.

But Gemini and Microsoft's Copilot falsely identified it as a genuine image of the protest.

"This inability to correctly identify AI images stems from the fact that they (AI models) are programmed only to mimic well," Rossine Fallorina, from the nonprofit Sigla Research Center, told AFP.

"In a sense, they can only generate things to resemble. They cannot ascertain whether the resemblance is actually distinguishable from reality."

Earlier this year, Columbia University's Tow Center for Digital Journalism tested the ability of seven AI chatbots -- including ChatGPT, Perplexity, Grok, and Gemini -- to verify 10 images from photojournalists of news events.

All seven models failed to correctly identify the provenance of the photos, the study said.

- 'Shocked' -

AFP tracked down the source of Co's photo that garnered over a million views across social media -- a middle-aged web developer in the Philippines, who said he created it "for fun" using Nano Banana, Gemini's AI image generator.

"Sadly, a lot of people believed it," he told AFP, requesting anonymity to avoid a backlash.

"I edited my post -- and added 'AI generated' to stop the spread -- because I was shocked at how many shares it got."

Such cases show how AI-generated photos flooding social platforms can look virtually identical to real imagery.

The trend has fueled concerns as surveys show online users are increasingly shifting from traditional search engines to AI tools for information gathering and verifying information.

The shift comes as Meta announced earlier this year it was ending its third-party fact-checking program in the United States, turning over the task of debunking falsehoods to ordinary users under a model known as "Community Notes."

Human fact-checking has long been a flashpoint in hyperpolarized societies, where conservative advocates accuse professional fact-checkers of liberal bias, a charge they reject.

AFP currently works in 26 languages with Meta's fact-checking program, including in Asia, Latin America, and the European Union.

Researchers say AI models can be useful to professional fact-checkers, helping to quickly geolocate images and spot visual clues to establish authenticity. But they caution that they cannot replace the work of trained human fact-checkers.

"We can't rely on AI tools to combat AI in the long run," Fallorina said.

burs-ac/sla/sms

A.Ferraro--NZN