GenAI: Problems Introduced and Countermeasures
Deepfake
Accessible, scalable and performant
Disinformation
- Influence democratic process [4, 8, 11]
- Mislead public perception and belief through social media [4, 6, 8]
- Erode trust in media, democracy, authority, artwork, archival and legal evidence; “truth decay” [3, 9–11]
Countermeasures
- C2PA to attest content is real [1, 4, 11]
- 👍 Source and history [3, 6, 7, 10]
- Traceability and trust for not being tempered [5]
- 👎 Can be spoofed by taking photo with photo [12]
- 👎 Privacy: disclose sensitive information, but mechanisms exist [5]
- 👎 Can be stripped, is usually stripped [4]
- 👎 Need wide adoption [1, 4]
- 👎 Provenance does not imply trust [7]
- 👍 Source and history [3, 6, 7, 10]
- Preserve content on searchable blockchain [1, 2, 6, 8, 9]
- 👎 complex to implement; need wide adoption
- Watermarking by content creator [6]
- 👍 detect tempering, protect copyright
- JPEG Trust: “trust report” for user to evaluate trust [7]
- 👍 Directly address trust
- 👎 Implementation details unclear; privacy
Cybercrime
- Video conference scam targeting business [6, 9]
- Video promotion scam targeting individual [6, 9]
- Video romance scam [9]
- Personal attack: porn, fake speech, cyberbullying [6, 7]
- Identity theft [3, 7, 9]
Countermeasures
- Cloud provider limit bad actor [6]
Generation not for faking
- Bias in generation [3, 7]
- Intellectual property laundry [1, 3, 4]
- Not transparent or explainable, thus problem for regulation [3]
- Privacy and consent of training data [1, 3, 4]
Countermeasures
- Explainable AI (XAI) [3, 9]
- Bias mitigation and de-identification of training data [3]
- NFT of generation with authenticity metadata, ORA (Ownership, Rights, Attribution) [1, 4]
- Visual matching for generation [1]
General countermeasures
- Public awareness and education [6, 12]
- Detection
- Human technique [3]
- ML detection model, e.g., GAN
- 👍 Much more accurate than human [3]
- 👍 Can be improved constantly [12]
- 👎 Cannot outcompete generator [6, 9]
- Physiological signal analysis for video, e.g., blood flow [9]
- Multi-modal detection [9]
- Detection (cont.)
- Watermarking training data [3, 4]
- Watermarking generated content [9, 10]
- 👎 Can be stripped, tempered or faked [4, 10]
- Fingerprinting: search by perceived hash [1, 4]
- 👍 Recover stripped metadata; some resilience to tempering [4]
- 👎 Sensitive to editing [4]
- 👎 Need access to content database [4]
- 👎 Scalability and privacy concern [4]
- Similarity search (correlation) [4]
- 👎 May not reflect causality [4]
- Organizational effort: partnership [6]
- Legal framework and global standard [6, 9]
References
[1] Balan, K., Agarwal, S., Jenni, S., Parsons, A., Gilbert, A. and Collomosse, J. 2023. EKILA: Synthetic media provenance and attribution for generative art. Proceedings of the IEEE/CVF conference on computer vision and pattern recognition (2023), 913–922.
[2] Bureacă, E. and Aciobăniței, I. 2024. A blockchain blockchain-based framework for content provenance and authenticity. 2024 16th international conference on electronics, computers and artificial intelligence (ECAI) (2024), 1–5.
[3] Bushey, J. 2023. AI-generated images as an emergent record format. 2023 IEEE international conference on big data (BigData) (2023), 2020–2031.
[4] Collomosse, J. and Parsons, A. 2024. To authenticity, and beyond! Building safe and fair generative AI upon the three pillars of provenance. IEEE Computer Graphics and Applications. 44, 3 (2024), 82–90.
[5] Fotos, N. and Delgado, J. 2023. Ensuring privacy in provenance information for images. 2023 24th international conference on digital signal processing (DSP) (2023), 1–5.
[6] Kharvi, P.L. 2024. Understanding the impact of AI-generated deepfakes on public opinion, political discourse, and personal security in social media. IEEE Security & Privacy. (2024).
[7] Mo, J., Kang, X., Hu, Z., ZHou, H., Li, T. and Gu, X. 2023. Towards trustworthy digital media in the aigc era: An introduction to the upcoming IsoJpegTrust standard. IEEE Communications Standards Magazine. 7, 4 (2023), 2–5.
[8] Rainey, J., Elawady, M., Abhayartne, C. and Bhowmik, D. 2023. TRAIT: A trusted media distribution framework. 2023 24th international conference on digital signal processing (DSP) (2023), 1–5.
[9] Romero-Moreno, F. Deepfake fraud detection: Safeguarding trust in generative ai. Available at SSRN 5031627.
[10] Shoker, S., Reddie, A., Barrington, S., Booth, R., Brundage, M., Chahal, H., Depp, M., Drexel, B., Gupta, R., Favaro, M., et al. 2023. Confidence-building measures for artificial intelligence: Workshop proceedings. arXiv preprint arXiv:2308.00862. (2023).
[11] Strickland, E. 2024. This election year, look for content credentials: Media organizations combat deepfakes and disinformation with digital manifests. IEEE Spectrum. 61, 01 (2024), 24–27.
[12] Vilesov, A., Tian, Y., Sehatbakhsh, N. and Kadambi, A. 2024. Solutions to deepfakes: Can camera hardware, cryptography, and deep learning verify real images? arXiv preprint arXiv:2407.04169. (2024).