亚洲免费av电影一区二区三区,日韩爱爱视频,51精品视频一区二区三区,91视频爱爱,日韩欧美在线播放视频,中文字幕少妇AV,亚洲电影中文字幕,久久久久亚洲av成人网址,久久综合视频网站,国产在线不卡免费播放

        ?

        Media Enhanced by Artificial Intelligence: Can We Believe Anything Anymore?

        2020-10-20 08:19:22RaminSkibba
        Engineering 2020年7期

        Ramin Skibba

        Senior Technology Writer

        While messages, letters, or photographs could be altered to change both content and intent, people generally assumed they were not because it was not easily done. But this is no longer true in today’s digital world. Tools for altering photos and other media quickly became available with the advent of computers, the internet, and, most recently, smartphones and social media. Now, artificial intelligence (AI) is further transforming digital media, with far more sophisticated programs that can be used to almost seamlessly manipulate video,photographs,audio,and text for a variety of purposes.

        ‘‘Manipulating photographs is as old as photography itself,”said Siwei Lyu, professor of computer science and director of the Computer Vision and Machine Learning Laboratory at the State University of New York at Albany. ‘‘The recent twist is the application of AI, where you can scale up the manipulation. It used to be something that required a lot of time,effort,special training and equipment.” With a powerful computer and enough knowledge to run the algorithms, Lyu said, manipulating video can now be done on a much greater scale.

        Computer engineers are also working to perfect AI systems for‘‘natural language processing” that can generate text and speech that closely approximates human language. For example, in early 2019 the San Francisco-based research laboratory OpenAI announced they had developed a state-of-the-art text generator called GPT-2, which could write coherent sentences in English and even short stories and poetry with just a few prompts. The researchers initially avoided releasing the full model because they feared the software was good enough that it could be used for malicious purposes such as for generating‘‘fake news”[1].But they relented in November 2019 after seeing ‘‘no strong evidence of misuse” [2]. Nonetheless, in this and other media, the old proverb‘‘seeing is believing” appears itself to be becoming fake news.

        Software like Photoshop for modifying photographs has existed for some time (Fig. 1), and now video sequences can be manipulated with similar ease. The most common manipulations, known as ‘‘deepfakes,” typically involve swapping the face of a person(the target) with that of someone else (the donor). Another type of deepfake, a ‘‘lip-sync,” involves modifying the source video so that the mouth movements of a speaker are changed to be consistent with a different audio recording. Done well, the resulting video looks realistic to the viewer,with the effect of them seeming to say something they actually have not. Such deceptive videos could be—and have been—used to manipulate public opinion,commit fraud, and wrongly discredit people [3].

        In practice,deepfake generation depends on feeding data—a significant number of images or text—into machine learning tools known as generative adversarial networks (GANs). In the simplest version, two such neural networks are trained to develop and improve a model for turning input data into new images or video.Early algorithms were trained with massive datasets,derived from easily accessible images of people like politicians and celebrities.While the process used to require some degree of supervision by programmers, the latest programs are almost entirely automated.

        Fig. 1. Adobe Photoshop software was used to create this fanciful but realistic-looking landscape from 16 different photographs. Software powered by AI algorithms now provides tools to create realistic but manipulated and/or simulated video, text, and speech with perhaps even greater ease. Credit: Wikimedia Commons (CC BY-SA 3.0).

        ‘‘You do not need huge amounts of training data. Even just a ten-second clip can be enough,”said Subbarao Kambhampati,professor of computer science and engineering at Arizona State University in Tempe and an expert on human-aware AI. But training the model with longer clips,with source video that has at least 1000 high-quality frames, will generate a better final product. For each frame in a video,current algorithms can map out‘‘landmarks”on a person’s head,as well as their head pose,eye gaze,in addition to the shape of more detailed features,including the eyebrows,eye blinks,eye lids,upper and lower lips,cheeks,chin,and dimples[4].

        The motion in the resulting video can look fluid, like human vision expects.But if not done carefully,the output video can contain‘‘tells”that could make a perceptive viewer suspect it has been modified. ‘‘Sometimes you can see odd things, like a stretching or distortion of facial features, which do not match exactly,” said Doug Goodwin, Fletcher Jones Scholar in computation and a visiting professor in media studies at Scripps College in Claremont,CA,USA. For example, if the training data have insufficient resolution,the output video might have blurry areas, with white strips in the mouth rather than individual teeth, or facial hair that does not move the way it should. The algorithms work better when they are trained with a diversity of facial expressions and words,Goodwin said.

        The advances in manipulation have prompted computer scientists and engineers to develop AI algorithms—forensic software—for detecting altered video and audio[5].‘‘Forensic tools can detect synthesized media and tell whether it is generated by a machine or a human.But if these tools are not kept secret,media can always be crafted to bypass them,” said Paarth Neekhara, a doctoral student in computer science at the University of California, San Diego,who studies audio and video deepfakes.

        The back and forth between manipulation and detection resembles the computer security arms race of viruses and antivirus software, wherein fixes stop hackers and hackers find ways to overcome the fixes [6]. Experts find a flaw, allowing them to spot manipulated media, and then the makers adapt to generate even more realistic deepfakes.For example,when first generation deepfakes showed faces that did not blink periodically, making them easy to detect,the next generation of deepfake software fixed that problem. In another example, a video of then US President Barack Obama was manipulated to make it look like he said something he had not, but his eyebrow movement did not match his lip movement, Kambhampati said. But then in subsequent deepfakes his eyebrows moved as expected. Because AI can be trained to detect and fix such discrepancies, the latest generation deepfakes have very few faults.

        Many negative applications have arisen [3,7], but there are many positive applications as well, and these have also driven advances in the technology. Examples include improving a video or audio recording of someone with a speech impediment, adding more realistic dubbing of language in movies, and even recreating a character in a movie played by an actor who has died, such as Princess Leia, played by the late Carrie Fisher, in the Star Wars movie Rogue One [8]. The application to virtual reality for games or other entertainment seems especially promising and likely [9].

        As exemplified by the OpenAI software mentioned above,computer scientists are also using AI programs to generate credible text and speech [1]. Like modified video, this modeling also now uses GANs to produce realistic sentences [10]. Google Translate, for example, now runs on such AI algorithms [11]. The algorithms are sophisticated enough to generate text in the style of a particular person,to produce,for example,a new story seemingly written by the long-deceased author Jane Austen [12]. Programmers have created chatbots, such as on social media platforms, that read and sound real enough that potential customers interact with them as if they were actual people.And in perhaps the most widely used commercial applications of AI-powered communication,Amazon’s Alexa and Apple’s Siri cloud-based voice services are programmed to mimic real conversations with customers. Alexa and Siri may not be real people, but they do seem to give truthful answers to questions.

        To date, programmers have made more headway with realistic video and still images, Goodwin said. But if current trends continue, he said, it may soon be possible to set up AI algorithms to write and digitally create completely new and credible speech and then meld it with simulated audio and video,in a mostly automated process. This prospect, and its potential use for deception,has prompted researchers to develop code to automatically spot deepfakes, and calls for social media sites to identify such media as manipulated [13]. In December 2020, Facebook launched a Deepfake Detection Challenge in a collaboration with Microsoft,Amazon, and academic computer scientists including Lyu, to recruit researchers to submit their own automated detection tools with a chance to win 1 million USD in prizes[14].Engineers at the US Defense Advanced Research Projects Agency are also working on tools to automatically determine whether a video or photo has been manipulated [15].

        视频女同久久久一区二区| 香蕉色香蕉在线视频| 日韩av中出在线免费播放网站| 国产精品亚洲一区二区三区正片| 国内自拍偷国视频系列| av中文字幕潮喷人妻系列| 国产精品后入内射日本在线观看| 亚洲无线码一区在线观看| 日本av一区二区在线| 午夜不卡无码中文字幕影院| 亚洲 暴爽 av人人爽日日碰| 亚洲精品成人网线在线播放va| 亚洲一区二区三区在线高清中文 | 成视频年人黄网站免费视频| 亚洲 小说区 图片区 都市| 免费中文熟妇在线影片| 中文字幕在线观看国产双飞高清| 一级二级三一片内射视频| 综合图区亚洲另类偷窥| 亚洲av无码不卡久久| 在线观看国产内射视频| 久久精品蜜桃美女av| 亚洲国产一二三精品无码| 四虎影视一区二区精品| 亚洲av乱码国产精品色| 免费亚洲老熟熟女熟女熟女| 首页 综合国产 亚洲 丝袜| 99热门精品一区二区三区无码| 在线亚洲精品免费视频| 婷婷伊人久久大香线蕉av| 欧美国产日本高清不卡| 一级无码啪啪| 青青草在线这里只有精品| 中出人妻中文字幕无码| 免费毛片性天堂| 美女视频在线观看一区二区三区| 色婷婷综合久久久中文字幕| 亚洲av成人无码网站…| 亚洲 暴爽 AV人人爽日日碰| 三级国产高清在线观看| 少妇粉嫩小泬喷水视频www|