ONSITE Overnight $100 - 1 Hour Minimum | Same Day $80 - 1 Hour Minimum | Next Day $70 - 1 Hour Minimum | MORE INFO...
Computer Repair fix IT services Business & Residential Toll Free: 1-800-432-1475 | Direct: (438)488-3296 - (347)389-2546

[ad_1]

In late 2017, Motherboard reported on an AI technology that could swap faces in videos. At the time, the tech—later called deepfakes—produced crude, grainy results and was mostly used to create fake porn videos featuring celebrities and politicians.

Two years later, the technology has advanced tremendously and is harder to detect with the naked eye. Along with fake news, forged videos have become a national security concern, especially as the 2020 presidential elections draw near.

Since deepfakes emerged, several organizations and companies have developed technologies to detect AI-tampered videos. But there’s a fear that one day, the deepfakes technology will be impossible to detect.

Researchers at the University of Surrey developed a solution that might solve the problem: instead of detecting what’s false, it’ll prove what’s true. Scheduled to be presented at the upcoming Conference on Computer Vision and Pattern Recognition (CVPR), the technology, called Archangel, uses AI and blockchain to create and register a tamper-proof digital fingerprint for authentic videos. The fingerprint can be used as a point of reference for verifying the validity of media being distributed online or broadcasted on television.

Using AI to Sign Videos

The classic way to prove the authenticity of a binary document is to use a digital signature. Publishers run their document through a cryptographic algorithm such as SHA256, MD5, or Blowfish, which produces a “hash,” a short string of bytes that represents the content of that file and becomes its digital signature. Running the same file through the hashing algorithm at any time will produce the same hash if its contents haven’t changed.

Hashes are supersensitive to changes in the binary structure of the source file. When you modify a single byte in the hashed file and run it through the algorithm again, it produces a totally different result.

But while hashes work well for text files and applications, they present challenges for videos, which can be stored in different formats, according to John Collomosse, professor of computer vision at the University of Surrey and project lead for Archangel.

“We wanted the signature to be the same regardless of the codec the video is being compressed with,” Collomosse says. “If I take my video and convert it from, say, MPEG-2 to MPEG-4, then that file will be of a totally different length, and the bits will have completely changed, which will produce a different hash. What we needed was a content-aware hashing algorithm.”

To solve this problem, Collomosse and his colleagues developed a deep neural network that is sensitive to the content contained in the video. Deep neural networks are a type of AI construction that develops its behavior through the analysis of vast amounts of examples. Interestingly, neural networks are also the technology at the heart of deepfakes.

When creating deepfakes, the developer feeds the network with pictures of a subject’s face. The neural network learns the features of the face and, with enough training, becomes capable of finding and swapping faces in other videos with the subject’s face.

Archangel’s neural network is trained on the video it’s fingerprinting. “The network is looking at the content of the video rather than its underlying bits and bytes,” Collomosse says.

After training, when you run a new video through the network, it will validate it when it contains the same content as the source video regardless of its format and will reject it when it’s a different video or has been tampered with or edited.

According to Collomosse, the technology can detect both spatial and temporal tampering. Spatial tamperings are changes made to individual frames, such as the face-swapping edits done in…

[ad_2]
Source link

No Comment

Comments are closed.