Android Q Live Caption Feature Enables Real-Time Subtitles

During Google’s I/O 2019 developers conference this week, the company demonstrated an impressive new feature for mobile operating system Android Q. Called Live Caption, the feature enables real-time transcription for any video or audio that users play on their smartphones. No matter if they’re listening or watching via YouTube, Skype, Instagram, Pocket Casts, or other applications, Live Caption overlays the text on top of whatever is being used. Additionally, Live Caption will work on top of original video or audio recordings on users’ phones.

Continue reading Android Q Live Caption Feature Enables Real-Time Subtitles

Real-Time Virtual Production Moves into Television and Film

To be able to bring real-time feedback to every step of the filmmaking workflow has been both the Holy Grail of production and an impossible dream. When Bradley Weiers began to work in film production, he chafed at the delayed feedback and found that the real-time ecosystem of game production was a better fit. Now head of immersive storytelling at Unity Technologies, Weiers said that for the first time, he believes he can connect his first love, film, and the preferable tools of gaming.  “There’s a bridge to cross over,” he suggested during a panel at NAB 2019. Continue reading Real-Time Virtual Production Moves into Television and Film

DeepRay Uses AI to Reconstruct Distorted Video in Real Time

Cambridge Consultants has trained artificial intelligence to reconstruct images from a highly distorted image or video feed in real time that is the AI’s best guess of what the source image originally looked like. The unique approach of their DeepRay product involves recognizing and correcting for a wide universe of distortion patterns. Cambridge Consultants’ first clients are self-driving car companies concerned with accurate object and risk detection, but DeepRay could also have a number of entertainment industry technical and story applications. Continue reading DeepRay Uses AI to Reconstruct Distorted Video in Real Time

Nvidia Reveals Use of Neural Networks to Create Virtual City

Nvidia used processing power and neural networks to create a very convincing virtual city, which will be open for tours by attendees to this year’s NeurIPS AI conference in Montreal. Nvidia’s system, which uses existing videos of scenery and objects to create these interactive environments, also makes it easier for artists to create similar virtual worlds. Nvidia vice president of applied deep learning Bryan Catanzaro said generative models are key to making the process of creating virtual worlds cost effective. Continue reading Nvidia Reveals Use of Neural Networks to Create Virtual City

5G Could Enable Interactive Video But Raise Privacy Issues

People typically associate 5G with ultra-fast high-bandwidth Internet connections, but few realize it will also impact how we watch video and could lead to a range of privacy concerns. With 5G, truly interactive television programming can become a reality, with minimal latency enabling content to respond quickly to the viewer’s emotional and physical responses. According to interactive video company Wirewax co-founder Dan Garraway, the video becomes “a two-way conversation.” In other words, while we watch 5G content, it watches back. Continue reading 5G Could Enable Interactive Video But Raise Privacy Issues

Upgraded Google Lens to Be Featured in Top Android Phones

During this week’s Google I/O conference, the importance of Google Lens to chief executive Sundar Pichai’s AI-first strategy became apparent. Google Lens combines computer vision and natural language processing with Google Search, for a solution aimed at consumers. Lens, described as “Google’s engine for seeing, understanding, and augmenting the real world,” resides in the camera viewfinder of Assistant and, soon, its top-end Android smartphones. Lens recognizes people, animals, objects, environments and text. Continue reading Upgraded Google Lens to Be Featured in Top Android Phones

NAB 2018: Artificial Intelligence Tools for Animation and VFX

Tools powered by artificial intelligence and machine learning can also be used in animation and visual effects. Nvidia senior solutions architect Rick Grandy noted that the benefit of such tools is that artists don’t have to replicate their own work. That includes deep learning used for realistic character motion created in real-time via game engines and AI, as well as a phase-functioned neural network for character control, whereby the network can be trained by motion capture or animation. Continue reading NAB 2018: Artificial Intelligence Tools for Animation and VFX

New Camera-First Twitter Feature Could Threaten Snap’s Ads

After its first profitable quarter on record, Twitter is focusing on new business and advertising opportunities. For example, sources say the social media company is currently working on a camera-first feature that could compete with Snap and potentially threaten its advertising opportunities. The new feature would combine videos and photos with the Twitter Moments feature, creating more real-time content around events and enabling companies to sponsor events or place ads between tweeted content.

Continue reading New Camera-First Twitter Feature Could Threaten Snap’s Ads

Google Maps Helps Develop Real-World Gaming Experiences

Google is enabling game developers to create “Pokémon Go”-like experiences in which game elements are embedded into real-world maps using the new Google Maps API and the Unity game engine. Over 200 games are already in development. Next Games is developing a game based on the popular TV series “The Walking Dead,” and NBCUniversal and Ludia will release a “Jurassic World Alive” location-based game for mobile. Because Google Maps updates in real time, developers can create gaming experiences with a sharp eye on reality.

Continue reading Google Maps Helps Develop Real-World Gaming Experiences

Google Debuts YouTube Shared Viewing App For iOS Users

Google’s Uptime app, that lets users watch YouTube videos in real time with friends, is now freely available to iOS users following its invite-only limited beta test. The beta, which started in March, was updated a few times, adding the ability to play and share music videos, as well as connect to Facebook to find friends to watch videos with. Users can also chat, leave comments and place emoji on top of the video. A replay of a shared video will include the ability to see comments at the time in the video they were made. Continue reading Google Debuts YouTube Shared Viewing App For iOS Users

Text-to-Speech System Quickly Mimics Hundreds of Accents

As another example of the significant advances we have been following in artificial intelligence and deep learning, Chinese search giant Baidu has introduced Deep Voice 2, the second iteration of its compelling text-to-speech system. The company introduced Deep Voice just three months ago, with the ability to produce speech “in near real time” that was “nearly indistinguishable from an actual human voice,” according to The Verge. While the first system was limited to learning one voice at a time, “and required many hours of audio or more from which to build a sample,” the updated version “can learn the nuances of a person’s voice with just half an hour of audio, and a single system can learn to imitate hundreds of different speakers.” Continue reading Text-to-Speech System Quickly Mimics Hundreds of Accents

Google Reveals More of Its Fuchsia OS for ‘Modern Phones’

Google is working on its third operating system after Android and Chrome OS, this one dubbed Fuchsia, an open-source, real-time OS that first appeared in August last year as a command line. Rather than being based on Linux, Fuchsia relies on a microkernel called Magenta that was developed by Google and incorporates other Google-developed software. According to Google, Magenta is aimed at “modern phones” and PCs with fast processors and large amounts of RAM with “arbitrary peripherals doing open-ended computation.” Continue reading Google Reveals More of Its Fuchsia OS for ‘Modern Phones’

Google Pursues Bringing Machine Learning to Mobile Devices

Google is partnering with Movidius, a semiconductor startup, to begin developing technology that would allow mobile devices to do more heavy computing such as machine learning and image recognition. Movidius specializes in computer vision, so Google has licensed its MA2450 chip to help build image-recognition capabilities. The chip has 12 cores and was chosen for its low power consumption and hefty processing power. Image recognition could be used in all sorts of apps to identify objects and people. Continue reading Google Pursues Bringing Machine Learning to Mobile Devices

Data Analytics: AI Could Assist Coaches in Professional Sports

Football coaches are already using artificial intelligence to help them on the sidelines. Students at North Carolina State University built an AI that could predict whether an NFL team would pass or run the ball. The AI called the plays correctly 91.6 percent of the time during an NFL game. The technology may not yet be ready for a real-time game situation, but with more data and research into machine learning and game theory, AI could become a big league contender. Continue reading Data Analytics: AI Could Assist Coaches in Professional Sports

SMPTE HDR Report Offers Recommendations for Standards

The Society of Motion Picture and Television Engineers has published a study group report on the high dynamic range imaging ecosystem, now available for download on the SMPTE site. The report provides a detailed summary regarding “the implementation of HDR in professional media workflows and addresses key questions that will arise for industry members as they move forward in taking advantage of HDR technology.” In addition, it provides recommendations for standardization efforts. According to SMPTE, the report reflects the efforts of 170 international experts. Continue reading SMPTE HDR Report Offers Recommendations for Standards