At Twine, we’re fortunate to work with some of the biggest artificial intelligence research institutions out there.
Every day, freelancers in the Twine community are helping to make the world a better place by providing video research material to researchers at the cutting edge of AI.
In this article, we’re going to take you through some of the most exciting AI research projects going on in the world right now.
Fakecatcher
You might have heard of the concept of ‘deep fakes‘: a form of synthetic media in which a person in an existing image or video is replaced with someone else’s likeness. As it’s been all over the media in 2020, it’s a pretty exciting trend.
Although deep fakes are mostly just harmless fun, using them for political purposes, or to spread misinformation, can have much more sinister consequences.
To fight this threat, researchers from Binghamton University have teamed up with Intel Corp. to develop a tool called FakeCatcher. FakeCatcher works by analyzing the subtle differences in skin color caused by the human heartbeat.
Incredible, right?
This technique, known as Photoplethysmography, (abbreviated as PPG) boasts an accuracy rate above 90% and uses the same technique used in Apple Watches and wearable fitness tracking devices that measure your heartbeat during exercise.
Ilke Demir, a senior research scientist at Intel, said in an interview:
We extract several PPG signals from different parts of the face and look at the spatial and temporal consistency of those signals. In deepfakes, there is no consistency for heartbeats and there is no pulse information. For real videos, the blood flow in someone’s left cheek and right cheek—to oversimplify it—agree that they have the same pulse.
The potential impact of Fakecatcher is huge. We’re really excited to see advances in this program will help in the fight against fake media.
Deblur your videos in one click with Adobe’s Project Sharp Shots
At their MAX user conference, Adobe dropped a bombshell with the announcement of Project Sharp Shots.
Powered by Adobe’s Sensei AI platform, Sharp Shots is a research project that uses AI to deblur videos — with a single click.
In the event demos, the impact of the AI was shown via a video of a woman playing the ukulele, filmed on a fast-moving motorcycle.
Shubhi Gupta, the engineer behind the project, told Tech Crunch “there’s no parameter tuning and adjustment like we used to do in our traditional methods. This one is just a one-click thing. It’s not magic. This is simple deep learning and AI working in the background, extracting each frame, deblurring it, and producing high-quality deblurred photos and videos.”
Multi-modal Video Content Analysis for Content Recommendation
What if AI could recommend new videos to you not just based on meta data but on the actual content in the video itself?
Visual Display Intelligence Lab have found an answer with multi modal AI video content analysis.
They’ve built an AI agent which processes data in multiple modalities (e.g., video, images, audio, language) and learns how to recommend new multi-media content to an user. The AI agent can process different characteristics of a video, and videos in their user history, and recommends new videos of interest to the user based on alignment of characteristics between the videos.
Their architecture vastly out-performs current state-of-the-art models in performance and we’re excited to see what impact it will have on the future of content recommendation!
ECCV2020 Pose Research
What if you could visualise the position and pose of people in videos where only a portion of their body is visible in the shot?
Researchers at the University of Michigan have trained neural network models to identify to do exactly that. Their breakthrough has opened up a vast array of new innovations for AI video analysis – even enabling machines to learn the meaning behind people’s poses and their interactions with their environment.
Their research pushes the boundaries of a field of study called human pose estimation.
Prepare yourself for profound changes to video editing software and creatives using this technology to enable really complex interactions of characters and environments within film – exciting stuff!
Are there any projects we’ve missed?
Let us know in the comments below.