We have to be able to hold tech platforms accountable for fraud
Algorithms ensure that people who click on scams are likely to see more of them
IN APRIL this year, I wrote a column on a “deepfake” on Instagram, one of Meta’s platforms. A former colleague had brought it to my attention in March, because the fake purported to be me. But this Martin Wolf gave investment advice, which I would never do. The Financial Times persuaded Meta to take it down. But it soon reappeared. We were playing “whack-a-mole” with scammers.
In the end, I was enrolled in a new Meta system that uses facial recognition technology to combat such scams. This worked: the deepfakes disappeared. My conclusion is that Meta can stop them if it is determined to do so.
Unfortunately, this was not before they had been widely circulated. A colleague told me there were at least three different deepfake videos and multiple Photoshopped images running over 1,700 advertisements with slight variations across Facebook and Instagram. Data from Meta’s Ad Library showed these ads reached over 970,000 users in the European Union alone – where regulations require platforms to report such figures. Worldwide, the number exposed to the ads must have been a multiple of this.
Decoding Asia newsletter: your guide to navigating Asia in a new global order. Sign up here to get Decoding Asia newsletter. Delivered to your inbox. Free.
Share with us your feedback on BT's products and services