CES: Sony Introduces Interactive Experience – ‘The Last of Us’

During CES this week, Sony demonstrated a proof-of-concept experience based on the popular HBO post-apocalyptic drama “The Last of Us.” We were dropped into a six-person pod of newly enlisted defenders and assigned to a hardened defender who needed new recruits to combat a serious surge of zombie assaults that she was convinced could be overcome with our assistance. Armed with LED-enabled shotgun-like devices and tracked flashlights to assist our leader in discovering the concealed attackers, our combat leader led us with sharp and direct commands as she guided us through the terrors of the attack. Continue reading CES: Sony Introduces Interactive Experience – ‘The Last of Us’

CES: Nvidia Unveils New GeForce RTX 50, AI Video Rendering

Nvidia founder and CEO Jensen Huang kicked off CES 2025 with a keynote that was filled with new product announcements and visionary demonstrations of how the company plans to advance the field of AI. The first product that Huang unveiled was the GeForce RTX 50 series of consumer graphics processing units (GPUs). The series is also called RTX Blackwell because it is based on Nvidia’s latest Blackwell microarchitecture design for next generation data center and gaming applications. To showcase RTX Blackwell’s prowess, Huang played an impressively photorealistic video sequence of rich imagery under contrasting light ranges — all rendered in real time. Continue reading CES: Nvidia Unveils New GeForce RTX 50, AI Video Rendering

Hume AI Introduces Voice Control and Claude Interoperability

Artificial voice startup Hume AI has had a busy Q4, introducing Voice Control, a no-code artificial speech interface that gives users control over 10 voice dimensions ranging from “assertiveness” to “buoyancy” and “nasality.” The company also debuted an interface that “creates emotionally intelligent voice interactions” with Anthropic’s foundation model Claude that has prompted one observer to ponder the possibility that keyboards will become a thing of the past when it comes to controlling computers. Both advances expand on Hume’s work with its own foundation model, Empathic Voice Interface 2 (EVI 2), which adds emotional timbre to AI voices. Continue reading Hume AI Introduces Voice Control and Claude Interoperability

D-ID’s New Business-Use Avatars Can Converse in Real Time

D-ID has launched two new types of AI-powered avatars: Premium+ and Express. The company’s video-to-video avatar tools aim to provide personal look-alikes that can sub for their creators in uses ranging from instructional videos to business presentations, offloading on-camera duties in areas including sales, marketing and customer support. “Premium+ Avatars can generate hyper-realistic digital humans that are indistinguishable from real people and will serve as the foundation for fully interactive digital agents revolutionizing how brands communicate,” while Express Avatars can rapidly generate serviceable avatars “from just one minute of source footage.” Continue reading D-ID’s New Business-Use Avatars Can Converse in Real Time

OpenAI to Expand Data Indexing, Analysis with Rockset Tech

OpenAI has acquired Rockset, a database firm that provides real-time analytics, indexing and search capabilities. Rockset will help OpenAI enable its customers to better leverage their own data as they build and utilize intelligent applications. Rockset technology will be integrated into the retrieval infrastructure across OpenAI products, with members of Rockset’s San Mateo, California-based team joining the staff of OpenAI, which is headquartered in San Francisco. This is the second major purchase for OpenAI, following last year’s acquisition of New York-based AI design studio Global Illumination. Financial terms of the deal were not disclosed. Continue reading OpenAI to Expand Data Indexing, Analysis with Rockset Tech

Snapchat Previews Instant AR Filters, GenAI Developer Tools

Snap Inc. teased a new on-device AI model capable of real-time filter creation in-app using Snapchat. At last week’s Augmented World Expo in Long Beach, California, Snap co-founder and CTO Bobby Murphy explained that the model, which runs on smartphones, can re-render frames on the fly guided by text prompts. Snap’s unnamed prototype model “can instantly bring your imagination to life in AR,” Snap says, explaining “this early prototype makes it possible to type in an idea for a transformation and generate vivid AR experiences in real time.” Continue reading Snapchat Previews Instant AR Filters, GenAI Developer Tools

Meta AI Seamless Translator Converts Nearly 100 Languages

The research division of Meta AI has developed Seamless Communication, a suite of artificial intelligence models that generate what the company says is natural and authentic communication across languages, facilitating what amounts to real-time universal speech translation. The models were released with accompanying research papers and data. The flagship model, Seamless, merges capabilities from a trio of models — SeamlessExpressive, SeamlessStreaming and SeamlessM4T v2 — into a single system that can translate between almost 100 spoken and written languages, preserving idioms, emotion and the speaker’s vocal style, Meta says. Continue reading Meta AI Seamless Translator Converts Nearly 100 Languages

Stability AI Intros Real-Time Text-to-Image Generation Model

Stability AI, developer of Stable Diffusion (one of the leading visual content generators, alongside Midjourney and DALL-E), has introduced SDXL Turbo — a new AI model that demonstrates more of the latent possibilities of the common diffusion generation approach: images that update in real time as the user’s prompt updates. This feature was always a possibility even with previous diffusion models given text and images are comprehended differently across linear time, but increased efficiency of generation algorithms and the steady accretion of GPUs and TPUs in a developer’s data center makes the experience more magical. Continue reading Stability AI Intros Real-Time Text-to-Image Generation Model

Twitter Users Vote in Favor of Musk Stepping Down as CEO

Facing backlash against his executive leadership, Twitter’s new owner and CEO, billionaire Elon Musk, conducted an informal 12-hour poll over the weekend asking users of the popular social media platform whether he should keep his new position. “Should I step down as head of Twitter?” the controversial executive asked. “I will abide by the results of this poll.” After more than 17.5 million responses, the results indicate that a majority of users believe Musk should step down from his post (57.5 percent voted in the affirmative). As of press time, it remains unclear what action Musk may take in light of the poll results. Continue reading Twitter Users Vote in Favor of Musk Stepping Down as CEO

Facebook Rolls Out New Messenger Feature, Watch Together

Facebook launched Watch Together, a feature for Messenger and videoconferencing platform Messenger Rooms, to allow users to watch videos in real time with family and friends on Apple and Android mobile devices. Users choose videos to view through Facebook’s video hub, Facebook Watch. The push to promote yet more video comes at a time when, due largely to COVID-19, more people than ever are at home watching content. Facebook Messenger allows up to eight people on a video call, and Messenger Rooms tops out at 50 people. Continue reading Facebook Rolls Out New Messenger Feature, Watch Together

Google Bypasses Cloud to Offer AI to Enterprise Customers

AI can enable many important tasks from manufacturing to medicine, but only if the applications are speedy and secure. Communication via the cloud adds latency and risks privacy, which is why Google worked on a solution — dubbed Coral — that avoids centralized data centers. Coral product manager Vikram Tank described Coral as a “platform of [Google] hardware and software components … that help you build devices with local AI — providing hardware acceleration for neural networks … right on the edge device.” Continue reading Google Bypasses Cloud to Offer AI to Enterprise Customers

Android Q Live Caption Feature Enables Real-Time Subtitles

During Google’s I/O 2019 developers conference this week, the company demonstrated an impressive new feature for mobile operating system Android Q. Called Live Caption, the feature enables real-time transcription for any video or audio that users play on their smartphones. No matter if they’re listening or watching via YouTube, Skype, Instagram, Pocket Casts, or other applications, Live Caption overlays the text on top of whatever is being used. Additionally, Live Caption will work on top of original video or audio recordings on users’ phones.

Continue reading Android Q Live Caption Feature Enables Real-Time Subtitles

Real-Time Virtual Production Moves into Television and Film

To be able to bring real-time feedback to every step of the filmmaking workflow has been both the Holy Grail of production and an impossible dream. When Bradley Weiers began to work in film production, he chafed at the delayed feedback and found that the real-time ecosystem of game production was a better fit. Now head of immersive storytelling at Unity Technologies, Weiers said that for the first time, he believes he can connect his first love, film, and the preferable tools of gaming.  “There’s a bridge to cross over,” he suggested during a panel at NAB 2019. Continue reading Real-Time Virtual Production Moves into Television and Film

DeepRay Uses AI to Reconstruct Distorted Video in Real Time

Cambridge Consultants has trained artificial intelligence to reconstruct images from a highly distorted image or video feed in real time that is the AI’s best guess of what the source image originally looked like. The unique approach of their DeepRay product involves recognizing and correcting for a wide universe of distortion patterns. Cambridge Consultants’ first clients are self-driving car companies concerned with accurate object and risk detection, but DeepRay could also have a number of entertainment industry technical and story applications. Continue reading DeepRay Uses AI to Reconstruct Distorted Video in Real Time

Nvidia Reveals Use of Neural Networks to Create Virtual City

Nvidia used processing power and neural networks to create a very convincing virtual city, which will be open for tours by attendees to this year’s NeurIPS AI conference in Montreal. Nvidia’s system, which uses existing videos of scenery and objects to create these interactive environments, also makes it easier for artists to create similar virtual worlds. Nvidia vice president of applied deep learning Bryan Catanzaro said generative models are key to making the process of creating virtual worlds cost effective. Continue reading Nvidia Reveals Use of Neural Networks to Create Virtual City