Google's AI Detection Tool Can't Decide If Its Own AI Made Doctored Photo of Crying Activist.
When the White House X account posted an image depicting activist Nekima Levy Armstrong in tears during her arrest, there were telltale signs that the image had been altered. Less than an hour before, Homeland Security Secretary Kristi Noem had posted a photo of the exact same scene, but in Noem’s version Levy Armstrong appeared composed, not crying in the least.
To determine if the White House version of the photo had been altered using artificial intelligence tools, researchers turned to Google's SynthID - a detection mechanism that Google claims is able to discern whether an image or video was generated using Google's own AI. They followed Google's instructions and used its AI chatbot, Gemini, to see if the image contained SynthID forensic markers.
The results were clear: The White House image had been manipulated with Google’s AI. Researchers published a story about it.
However, after posting the article, subsequent attempts to use Gemini to authenticate the image with SynthID produced different outcomes. In their second test, Gemini concluded that the image of Levy Armstrong crying was actually authentic.
But in their third test, SynthID determined that the image was not made with Google's AI, directly contradicting its first response. This inconsistency raises serious questions about SynthID’s reliability to tell fact from fiction at a time when AI-manipulated photos and videos are growing increasingly prevalent.
Aside from Google's proprietary tool, there is no easy way for users to test whether an image contains a SynthID watermark. That makes it difficult in this case to determine whether Google's system initially detected the presence of a SynthID watermark in an image without one or if subsequent tests missed a SynthID watermark in an image that actually contains one.
If AI-detection technology fails to produce consistent responses, though, there’s reason to wonder who will call bullshit on the bullshit detector. The incident highlights concerns about the reliability of AI-generated tools and their potential misuse in spreading misinformation.
Supporters of the technology argue tools that can detect if something is AI will play a critical role establishing the common truth amid the pending flood of media generated or manipulated by AI. However, if AI-detection technology fails to deliver accurate results, the very notion of fact-checking becomes meaningless.
When the White House X account posted an image depicting activist Nekima Levy Armstrong in tears during her arrest, there were telltale signs that the image had been altered. Less than an hour before, Homeland Security Secretary Kristi Noem had posted a photo of the exact same scene, but in Noem’s version Levy Armstrong appeared composed, not crying in the least.
To determine if the White House version of the photo had been altered using artificial intelligence tools, researchers turned to Google's SynthID - a detection mechanism that Google claims is able to discern whether an image or video was generated using Google's own AI. They followed Google's instructions and used its AI chatbot, Gemini, to see if the image contained SynthID forensic markers.
The results were clear: The White House image had been manipulated with Google’s AI. Researchers published a story about it.
However, after posting the article, subsequent attempts to use Gemini to authenticate the image with SynthID produced different outcomes. In their second test, Gemini concluded that the image of Levy Armstrong crying was actually authentic.
But in their third test, SynthID determined that the image was not made with Google's AI, directly contradicting its first response. This inconsistency raises serious questions about SynthID’s reliability to tell fact from fiction at a time when AI-manipulated photos and videos are growing increasingly prevalent.
Aside from Google's proprietary tool, there is no easy way for users to test whether an image contains a SynthID watermark. That makes it difficult in this case to determine whether Google's system initially detected the presence of a SynthID watermark in an image without one or if subsequent tests missed a SynthID watermark in an image that actually contains one.
If AI-detection technology fails to produce consistent responses, though, there’s reason to wonder who will call bullshit on the bullshit detector. The incident highlights concerns about the reliability of AI-generated tools and their potential misuse in spreading misinformation.
Supporters of the technology argue tools that can detect if something is AI will play a critical role establishing the common truth amid the pending flood of media generated or manipulated by AI. However, if AI-detection technology fails to deliver accurate results, the very notion of fact-checking becomes meaningless.