Overview
- MIT Technology Review confirmed that the Department of Homeland Security is using Google and Adobe AI video generators to produce public-facing content.
- The White House shared a digitally altered photo of a woman arrested at an ICE protest and did not clarify whether the manipulation was intentional.
- Adobe’s Content Authenticity Initiative applies labels primarily to fully AI‑generated content, and platforms can remove or fail to display these labels, as seen on X and the Pentagon’s DVIDS site.
- MIT professor David Karger argues for standardized provenance and a broader set of trusted verifiers, with tools and potential regulation enabling users to prioritize signals from sources they choose.
- A study in Communications Psychology found that people still relied on a deepfake confession even after being told it was fabricated, indicating transparency alone does not neutralize influence.