Zürcher Nachrichten - AI's blind spot: tools fail to detect their own fakes

EUR -
AED 4.35335
AFN 77.050797
ALL 96.66512
AMD 452.977132
ANG 2.121943
AOA 1087.00321
ARS 1715.259993
AUD 1.706088
AWG 2.136666
AZN 2.019869
BAM 1.955701
BBD 2.406579
BDT 146.012629
BGN 1.990709
BHD 0.449077
BIF 3539.921292
BMD 1.18539
BND 1.513224
BOB 8.256583
BRL 6.231008
BSD 1.19484
BTN 109.724461
BWP 15.634211
BYN 3.403228
BYR 23233.647084
BZD 2.403079
CAD 1.614917
CDF 2684.909135
CHF 0.911322
CLF 0.026011
CLP 1027.058063
CNY 8.240537
CNH 8.248946
COP 4350.080393
CRC 591.67013
CUC 1.18539
CUP 31.412839
CVE 110.259434
CZK 24.334287
DJF 212.769259
DKK 7.470097
DOP 75.226202
DZD 154.463202
EGP 55.903178
ERN 17.780852
ETB 185.61503
FJD 2.613371
FKP 0.865849
GBP 0.861444
GEL 3.194674
GGP 0.865849
GHS 13.089339
GIP 0.865849
GMD 86.533903
GNF 10484.470707
GTQ 9.164537
GYD 249.97738
HKD 9.259024
HNL 31.537408
HRK 7.536597
HTG 156.372106
HUF 381.328619
IDR 19883.141804
ILS 3.663335
IMP 0.865849
INR 108.693763
IQD 1565.320977
IRR 49934.560565
ISK 144.985527
JEP 0.865849
JMD 187.240547
JOD 0.840489
JPY 183.456955
KES 154.262212
KGS 103.662825
KHR 4804.757439
KMF 491.93733
KPW 1066.851144
KRW 1719.768532
KWD 0.36382
KYD 0.99575
KZT 600.939662
LAK 25713.701882
LBP 106998.998316
LKR 369.511346
LRD 215.369127
LSL 18.971842
LTL 3.500149
LVL 0.717031
LYD 7.497621
MAD 10.838453
MDL 20.096985
MGA 5339.730432
MKD 61.636888
MMK 2489.708718
MNT 4227.553379
MOP 9.608515
MRU 47.674593
MUR 53.852723
MVR 18.32658
MWK 2071.895403
MXN 20.70407
MYR 4.672854
MZN 75.580924
NAD 18.971842
NGN 1643.520192
NIO 43.96778
NOK 11.437875
NPR 175.559137
NZD 1.964681
OMR 0.458017
PAB 1.19484
PEN 3.994898
PGK 5.114742
PHP 69.837307
PKR 334.289724
PLN 4.215189
PYG 8003.59595
QAR 4.35638
RON 5.097064
RSD 117.394074
RUB 90.535429
RWF 1743.311992
SAR 4.447217
SBD 9.544303
SCR 17.203132
SDG 713.016537
SEK 10.580086
SGD 1.506161
SHP 0.88935
SLE 28.834661
SLL 24857.038036
SOS 682.865527
SRD 45.104693
STD 24535.182964
STN 24.498763
SVC 10.454472
SYP 13109.911225
SZL 18.966043
THB 37.225573
TJS 11.153937
TMT 4.148866
TND 3.433027
TOP 2.854135
TRY 51.401485
TTD 8.11259
TWD 37.456003
TZS 3076.744675
UAH 51.211415
UGX 4271.784345
USD 1.18539
UYU 46.367659
UZS 14607.262574
VES 410.075543
VND 30749.020682
VUV 140.814221
WST 3.213333
XAF 655.923887
XAG 0.014004
XAU 0.000244
XCD 3.203577
XCG 2.153391
XDR 0.815759
XOF 655.923887
XPF 119.331742
YER 282.508153
ZAR 19.134414
ZMK 10669.938133
ZMW 23.448816
ZWL 381.695147
  • RBGPF

    1.3800

    83.78

    +1.65%

  • SCS

    0.0200

    16.14

    +0.12%

  • RELX

    -0.3700

    35.8

    -1.03%

  • CMSC

    0.0500

    23.76

    +0.21%

  • RYCEF

    -0.4300

    16

    -2.69%

  • AZN

    0.1800

    92.77

    +0.19%

  • RIO

    -4.1000

    91.03

    -4.5%

  • BP

    -0.1600

    37.88

    -0.42%

  • GSK

    0.9400

    51.6

    +1.82%

  • NGG

    0.2000

    85.27

    +0.23%

  • BTI

    0.4600

    60.68

    +0.76%

  • BCC

    0.5100

    80.81

    +0.63%

  • CMSD

    -0.0400

    24.05

    -0.17%

  • VOD

    -0.0600

    14.65

    -0.41%

  • BCE

    0.3700

    25.86

    +1.43%

  • JRI

    0.1400

    13.08

    +1.07%

AI's blind spot: tools fail to detect their own fakes
AI's blind spot: tools fail to detect their own fakes / Photo: Chris Delmas - AFP

AI's blind spot: tools fail to detect their own fakes

When outraged Filipinos turned to an AI-powered chatbot to verify a viral photograph of a lawmaker embroiled in a corruption scandal, the tool failed to detect it was fabricated -- even though it had generated the image itself.

Text size:

Internet users are increasingly turning to chatbots to verify images in real time, but the tools often fail, raising questions about their visual debunking capabilities at a time when major tech platforms are scaling back human fact-checking.

In many cases, the tools wrongly identify images as real even when they are generated using the same generative models, further muddying an online information landscape awash with AI-generated fakes.

Among them is a fabricated image circulating on social media of Elizaldy Co, a former Philippine lawmaker charged by prosecutors in a multibillion-dollar flood-control corruption scam that sparked massive protests in the disaster-prone country.

The image of Co, whose whereabouts has been unknown since the official probe began, appeared to show him in Portugal.

When online sleuths tracking him asked Google's new AI mode whether the image was real, it incorrectly said it was authentic.

AFP's fact-checkers tracked down its creator and determined that the image was generated using Google AI.

"These models are trained primarily on language patterns and lack the specialized visual understanding needed to accurately identify AI-generated or manipulated imagery," Alon Yamin, chief executive of AI content detection platform Copyleaks, told AFP.

"With AI chatbots, even when an image originates from a similar generative model, the chatbot often provides inconsistent or overly generalized assessments, making them unreliable for tasks like fact-checking or verifying authenticity."

Google did not respond to AFP’s request for comment.

- 'Distinguishable from reality' -

AFP found similar examples of AI tools failing to verify their own creations.

During last month's deadly protests over lucrative benefits for senior officials in Pakistan-administered Kashmir, social media users shared a fabricated image purportedly showing men marching with flags and torches.

An AFP analysis found it was created using Google's Gemini AI model.

But Gemini and Microsoft's Copilot falsely identified it as a genuine image of the protest.

"This inability to correctly identify AI images stems from the fact that they (AI models) are programmed only to mimic well," Rossine Fallorina, from the nonprofit Sigla Research Center, told AFP.

"In a sense, they can only generate things to resemble. They cannot ascertain whether the resemblance is actually distinguishable from reality."

Earlier this year, Columbia University's Tow Center for Digital Journalism tested the ability of seven AI chatbots -- including ChatGPT, Perplexity, Grok, and Gemini -- to verify 10 images from photojournalists of news events.

All seven models failed to correctly identify the provenance of the photos, the study said.

- 'Shocked' -

AFP tracked down the source of Co's photo that garnered over a million views across social media -- a middle-aged web developer in the Philippines, who said he created it "for fun" using Nano Banana, Gemini's AI image generator.

"Sadly, a lot of people believed it," he told AFP, requesting anonymity to avoid a backlash.

"I edited my post -- and added 'AI generated' to stop the spread -- because I was shocked at how many shares it got."

Such cases show how AI-generated photos flooding social platforms can look virtually identical to real imagery.

The trend has fueled concerns as surveys show online users are increasingly shifting from traditional search engines to AI tools for information gathering and verifying information.

The shift comes as Meta announced earlier this year it was ending its third-party fact-checking program in the United States, turning over the task of debunking falsehoods to ordinary users under a model known as "Community Notes."

Human fact-checking has long been a flashpoint in hyperpolarized societies, where conservative advocates accuse professional fact-checkers of liberal bias, a charge they reject.

AFP currently works in 26 languages with Meta's fact-checking program, including in Asia, Latin America, and the European Union.

Researchers say AI models can be useful to professional fact-checkers, helping to quickly geolocate images and spot visual clues to establish authenticity. But they caution that they cannot replace the work of trained human fact-checkers.

"We can't rely on AI tools to combat AI in the long run," Fallorina said.

burs-ac/sla/sms

A.Ferraro--NZN