首页 500强 活动 榜单 商业 科技 领导力 专题 品牌中心
杂志订阅

深伪视频,躺枪的为什么总是汤姆·克鲁斯?

Jeremy Kahn
2021-03-03

汤姆·克鲁斯是世界上被拍照次数最多的名人之一,有关他的数据更容易训练出一个他的极其出色的图像生成器。

文本设置
小号
默认
大号
Plus(0条)

据称,有三条疯传的视频拍摄了汤姆·克鲁斯表演魔术、讲一个不太搞笑的笑话和练习高尔夫球挥杆动作等场面。据数字图像法医分析专家表示,这三条视频是深度伪造视频最逼真的例子。深度伪造视频是利用人工智能技术创作的具有高度迷惑性的虚假视频。

上周,一个名为@deeptomcruise的账号在社交媒体TikTok上发布了这三条视频,累计观看次数达到约1,100万次。该账号有34.2万粉丝,获得该平台的其他用户100万次点赞。

@deeptomcruise账号背后的运营人员和运营团队尚无法确定,但汤姆·克鲁斯的模仿者埃文·费兰迪在上周末告诉Mic网站,他认为这些视频应该出自演员米尔斯·费舍尔之手。费舍尔与克鲁斯长相酷似,以前曾经模仿过克鲁斯。社交媒体网站上也有许多人认为是费舍尔在视频中模仿了克鲁斯,并使用深度伪造技术对脸部进行了修改。

加州大学伯克利分校(University of California at Berkeley)的汉尼·法立德教授专门从事数字图像分析。他表示,他相信这些是深度伪造视频,但他们“做得非常逼真”。

根据法立德和他的一名研究生施卢蒂·阿加瓦尔的分析,有一些微小的证据能够证明这些视频是人工智能生成的伪造视频。在一条视频中,克鲁斯似乎正在用一枚硬币表演魔术,在视频最后克鲁斯的眼睛颜色和眼型有轻微变化。克鲁斯的虹膜上还可以看到两个异常的小白点,表面看是光线反射。但法立德表示,这两个小白点的变化较真实视频的变化更加明显。

使用人工智能方法创建深度伪造视频,通常会在他们创建的图像和视频中留下微妙的视觉异常,比如眼睛颜色或眼型不一致、耳朵轮廓古怪或发际线异常等。

深度伪造视频常被用于换头或换脸,而不是生成整个身体,而且法立德指出,表演硬币魔术的双手看起来像是克鲁斯本人的双手。或许它们属于正在表演硬币魔术的另一位演员,但将脸换成了汤姆·克鲁斯。

法立德还表示,虽然深度伪造视频通常会涉及到全脸更换,但使用人工智能技术只生成脸部的一部分,有时候能够达到更逼真的效果。他和阿加瓦尔怀疑,这三条视频就是这种情况。他们认为,嘴部可能是真实的,但眼睛区域使用了深度伪造技术。

法立德说:“如果视频中是真人在模仿克鲁斯,可能通过化妆达到某种良好的效果,这或许有一定的道理,而且换掉独特的眼睛足以让脸部肖像达到逼真的效果。也有可能是在拍摄完视频之后进行了一些编辑。”

深度伪造视频使用一种名为GAN(生成对抗网络)的机器学习技术创作而成,该技术会对两个深度神经网络进行轮流训练。这种技术是大致基于人脑运行模式的一种机器学习技术。一个网络通过克鲁斯本人的图片或视频进行训练,用于生成克鲁斯处在不同环境中或摆出不同姿势的新图像,其逼真程度足以骗过另外一个网络,后者接受的训练是从许多人的图像中挑选出汤姆·克鲁斯。

与大部分人工智能方法一样,数据的数量和质量决定了系统的效果。这解释了为什么克鲁斯经常成为深度伪造视频模仿的对象:他是世界上被拍照次数最多的名人之一。这些数据更容易训练出一个极其出色的汤姆·克鲁斯图像生成器。

法立德表示,克鲁斯有独特的嗓音和一些特殊的动作,这并没有坏处,反而增加了与他有关的深度伪造视频的娱乐价值和社交媒体传播力。

在这三条克鲁斯视频之前,有一条最广为流传并且最神秘的深度伪造视频同样涉及到克鲁斯。这条视频的发布者网名为Ctrl Shift Face,已经创作了大批高度逼真的深度伪造视频。在其去年发布的这条视频中,喜剧演员比尔·哈德正在2008年大卫·莱特曼的节目中模仿克鲁斯。发布者利用深度伪造技术对这条视频进行了修改,让哈德在模仿克鲁斯的时候几乎与本人一模一样。在GAN技术诞生大约三年之后,即2017年,首次出现了深度伪造视频。最早的视频是将名人的头换到色情影片女演员的身体上。但在那之后,该技术被用于创作了不同名人在不同环境下的视频。现在已经有一款现成的软件,支持用户创作同样令人难以辨别的深度伪造视频,并且由于深度伪造视频可能被用于复杂的政治虚假信息宣传的情况,因此引起了安全研究人员的高度警惕。不过到目前为止,虽然专家仍然在就几种可能的情况进行争论,但深度伪造视频并不是虚假信息传播的主要原因。

虽然今天的深度伪造视频通常可以使用数字法医分析进行识别,但这个过程耗时漫长,而且需要具备丰富的专业知识。研究人员正在研究创建能够自动识别虚假伪造视频的人工智能系统。2019年,Facebook发起了年度竞赛,希望找到最佳的系统。但在现有系统中,表现最好的系统只有65%的时间可以准确识别。

阿加瓦尔表示,使用专门的商业软件,有可能创作出与这三条克鲁斯视频同等质量的深度伪造视频。但这需要具备一定的技能,相关人工智能系统还需要有大量数据和训练时间,而且训练时间成本高昂。所以,为了一条TikTok爆款视频付出这么多努力和时间是否值得,我们仍然无法确定。(财富中文网)

翻译:刘进龙

审校:汪皓

据称,有三条疯传的视频拍摄了汤姆·克鲁斯表演魔术、讲一个不太搞笑的笑话和练习高尔夫球挥杆动作等场面。据数字图像法医分析专家表示,这三条视频是深度伪造视频最逼真的例子。深度伪造视频是利用人工智能技术创作的具有高度迷惑性的虚假视频。

上周,一个名为@deeptomcruise的账号在社交媒体TikTok上发布了这三条视频,累计观看次数达到约1,100万次。该账号有34.2万粉丝,获得该平台的其他用户100万次点赞。

@deeptomcruise账号背后的运营人员和运营团队尚无法确定,但汤姆·克鲁斯的模仿者埃文·费兰迪在上周末告诉Mic网站,他认为这些视频应该出自演员米尔斯·费舍尔之手。费舍尔与克鲁斯长相酷似,以前曾经模仿过克鲁斯。社交媒体网站上也有许多人认为是费舍尔在视频中模仿了克鲁斯,并使用深度伪造技术对脸部进行了修改。

加州大学伯克利分校(University of California at Berkeley)的汉尼·法立德教授专门从事数字图像分析。他表示,他相信这些是深度伪造视频,但他们“做得非常逼真”。

根据法立德和他的一名研究生施卢蒂·阿加瓦尔的分析,有一些微小的证据能够证明这些视频是人工智能生成的伪造视频。在一条视频中,克鲁斯似乎正在用一枚硬币表演魔术,在视频最后克鲁斯的眼睛颜色和眼型有轻微变化。克鲁斯的虹膜上还可以看到两个异常的小白点,表面看是光线反射。但法立德表示,这两个小白点的变化较真实视频的变化更加明显。

使用人工智能方法创建深度伪造视频,通常会在他们创建的图像和视频中留下微妙的视觉异常,比如眼睛颜色或眼型不一致、耳朵轮廓古怪或发际线异常等。

深度伪造视频常被用于换头或换脸,而不是生成整个身体,而且法立德指出,表演硬币魔术的双手看起来像是克鲁斯本人的双手。或许它们属于正在表演硬币魔术的另一位演员,但将脸换成了汤姆·克鲁斯。

法立德还表示,虽然深度伪造视频通常会涉及到全脸更换,但使用人工智能技术只生成脸部的一部分,有时候能够达到更逼真的效果。他和阿加瓦尔怀疑,这三条视频就是这种情况。他们认为,嘴部可能是真实的,但眼睛区域使用了深度伪造技术。

法立德说:“如果视频中是真人在模仿克鲁斯,可能通过化妆达到某种良好的效果,这或许有一定的道理,而且换掉独特的眼睛足以让脸部肖像达到逼真的效果。也有可能是在拍摄完视频之后进行了一些编辑。”

深度伪造视频使用一种名为GAN(生成对抗网络)的机器学习技术创作而成,该技术会对两个深度神经网络进行轮流训练。这种技术是大致基于人脑运行模式的一种机器学习技术。一个网络通过克鲁斯本人的图片或视频进行训练,用于生成克鲁斯处在不同环境中或摆出不同姿势的新图像,其逼真程度足以骗过另外一个网络,后者接受的训练是从许多人的图像中挑选出汤姆·克鲁斯。

与大部分人工智能方法一样,数据的数量和质量决定了系统的效果。这解释了为什么克鲁斯经常成为深度伪造视频模仿的对象:他是世界上被拍照次数最多的名人之一。这些数据更容易训练出一个极其出色的汤姆·克鲁斯图像生成器。

法立德表示,克鲁斯有独特的嗓音和一些特殊的动作,这并没有坏处,反而增加了与他有关的深度伪造视频的娱乐价值和社交媒体传播力。

在这三条克鲁斯视频之前,有一条最广为流传并且最神秘的深度伪造视频同样涉及到克鲁斯。这条视频的发布者网名为Ctrl Shift Face,已经创作了大批高度逼真的深度伪造视频。在其去年发布的这条视频中,喜剧演员比尔·哈德正在2008年大卫·莱特曼的节目中模仿克鲁斯。发布者利用深度伪造技术对这条视频进行了修改,让哈德在模仿克鲁斯的时候几乎与本人一模一样。在GAN技术诞生大约三年之后,即2017年,首次出现了深度伪造视频。最早的视频是将名人的头换到色情影片女演员的身体上。但在那之后,该技术被用于创作了不同名人在不同环境下的视频。现在已经有一款现成的软件,支持用户创作同样令人难以辨别的深度伪造视频,并且由于深度伪造视频可能被用于复杂的政治虚假信息宣传的情况,因此引起了安全研究人员的高度警惕。不过到目前为止,虽然专家仍然在就几种可能的情况进行争论,但深度伪造视频并不是虚假信息传播的主要原因。

虽然今天的深度伪造视频通常可以使用数字法医分析进行识别,但这个过程耗时漫长,而且需要具备丰富的专业知识。研究人员正在研究创建能够自动识别虚假伪造视频的人工智能系统。2019年,Facebook发起了年度竞赛,希望找到最佳的系统。但在现有系统中,表现最好的系统只有65%的时间可以准确识别。

阿加瓦尔表示,使用专门的商业软件,有可能创作出与这三条克鲁斯视频同等质量的深度伪造视频。但这需要具备一定的技能,相关人工智能系统还需要有大量数据和训练时间,而且训练时间成本高昂。所以,为了一条TikTok爆款视频付出这么多努力和时间是否值得,我们仍然无法确定。(财富中文网)

翻译:刘进龙

审校:汪皓

A trio of viral videos allegedly depicting the actor Tom Cruise performing a magic trick, telling a not-so-funny joke, and practicing his golf swing are some of the most sophisticated examples yet seen of deepfakes, highly convincing fake videos created using A.I. technology, according to experts in the forensic analysis of digital images.

The three videos, which were posted last week on the social media platform TikTok from an account called @deeptomcruise, have collectively been viewed about 11 million times. The account has garnered more than 342,000 followers and 1 million likes from other users of the social media platform.

The person or people behind @deeptomcruise have not yet been definitely identified, but Cruise impersonator Evan Ferrante told website Mic over the weekend that he believed the videos were the work of an actor named Miles Fisher, who resembles Cruise and has done impressions of him in the past. Several people on social media sites also said they believed Fisher is depicting Cruise in the videos, with his face modified using deepfake technology.

Hany Farid, a professor at the University of California at Berkeley who specializes in the analysis of digital images, says he is convinced that the videos are deepfakes but that they are “incredibly well done.”

According to an analysis by Farid and one of his graduate students, Shruti Agarwal, there are a few tiny pieces of evidence that give away the fact that the videos are A.I.-generated fakes. In one video, in which Cruise seems to perform a magic trick with a coin, Cruise’s eye color and eye shape change slightly at the end of the video. There are also two unusual small white dots seen in Cruise’s iris—ostensibly reflected light—that Farid says change more than would be expected in an authentic video.

The A.I. methods used to create deepfakes often leave subtle visual oddities in imagery and videos they create—inconsistencies in eye color or shape, or strange ear contours or anomalies around the hairline.

Deepfakes are most often used to swap one person’s head or face for another’s as opposed to generating the entire body, and Farid notes that the hands performing the coin trick don’t look like the real Cruise’s hands. Presumably they belong to an actor, who was filmed performing the coin trick and then had Cruise’s face substituted for his.

Farid also says that while a true deepfake often involves a full-face swap, a more convincing result can sometimes be obtained by using the A.I. technique to generate only a portion of the face. He and Agarwal suspect that this is the case with the three Cruise videos. They think that the mouth is probably real, but that the eye region has been created with deepfake technology.

“This would make sense if the actual person in the video resembles Cruise, did some good work with makeup perhaps, and the swapping of the distinct eyes is enough to finalize a compelling likeness,” Farid says. “It is also possible that there was some postproduction video editing.”

Deepfakes are created using a machine-learning technique called a GAN (generative adversarial network), in which two deep neural networks—a type of machine learning loosely based on the way the human brain works—are trained in tandem. One network is trained from pictures or videos of the real Cruise to generate new images of Cruise in different settings or poses that are realistic enough to fool the other network, which is trained to pick out images of Tom Cruise from those of other people.

As with most A.I. methods, the amount and quality of the data help determine how good the system is. That goes to explain why Cruise has been a frequent target for deepfakes: He is among one of the most photographed celebrities on the planet. All that data makes it easier to train a very good Tom Cruise image generator.

Farid says it also doesn’t hurt that Cruise has a distinctive voice and mannerisms that add to the entertainment value and social media virality of deepfakes involving him.

Prior to the current trio of Cruise videos, one of the most wildly circulated and uncanny examples of a deepfake also involved Cruise. Released last year by a person who goes by the Internet handle Ctrl Shift Face, who has created a number of highly realistic deepfakes, it involves a video of the comedian Bill Hader doing an impersonation of Cruise on the David Letterman show in 2008. Deepfake technology is used to modify the video so that Hader’s face seamlessly morphs into Cruise’s as he does the impression. Deepfakes first surfaced in 2017, about three years after GANs were invented. Some of the earliest examples were videos in which the head of a celebrity was swapped with the body of an actress in a pornographic film. But since then they have been used to create fake videos of a lot of different celebrities in different settings. There is now off-the-shelf software that enables users to create fairly convincing deepfakes, and security researchers have become increasingly alarmed that deepfakes could be used for sophisticated political disinformation campaigns. But so far, despite a couple of possible examples that are still being debated by experts, deepfakes have not become a major factor in disinformation efforts.

While today’s deepfakes are usually identifiable with careful digital forensic analysis, this process is time-consuming and requires a certain amount of expertise. Researchers are working to create A.I. systems that would be able to automatically identify deepfakes, and Facebook in 2019 launched an annual competition to find the best of these. But in its inaugural running, the top performing system was accurate only 65% of the time.

Agarwal says it is possible to create deepfakes of the quality seen in the three Cruise videos using commercial software for deepfake generation. But doing so requires some skill, as well as a significant amount of data and training time for the A.I. system involved—and that training time can be expensive. So whether it would have been worth that sort of effort and cost for a viral TikTok video remains uncertain.

财富中文网所刊载内容之知识产权为财富媒体知识产权有限公司及/或相关权利人专属所有或持有。未经许可,禁止进行转载、摘编、复制及建立镜像等任何使用。
0条Plus
精彩评论
评论

撰写或查看更多评论

请打开财富Plus APP

前往打开