Samsung Set to Release Glasses-Free Odyssey 3D Monitors

Samsung Electronics, which teased a glasses-free 3D gaming monitor at CES in January, officially announced the scheduled release of two versions at Gamescom last week. Both sizes employ light field display (LFD) technology to create what Samsung calls “lifelike 3D images” from 2D content by using a lenticular lens on the front panel. “Combined with Eye Tracking and View Mapping technology, Odyssey 3D ensures an optimized 3D experience without the need for separate 3D glasses,” according to Samsung. A built-in stereo camera monitors the movement of both eyes while proprietary View Mapping continuously adjusts the image to fuel depth perception. Continue reading Samsung Set to Release Glasses-Free Odyssey 3D Monitors

Meta, Oxford Advance 3D Object Generation with VFusion3D

VFusion3D is the latest AI model unveiled by Meta Platforms, which developed it in conjunction with the University of Oxford. The powerful model, which uses single-perspective images or text prompts to generate high-quality 3D objects, is being hailed as a breakthrough in scalable 3D AI that can potentially transform sectors including VR, gaming and digital design. The platform tackles the challenge of scarce 3D training data in a world teeming with 2D images and text descriptions. The VFusion3D approach leverages what the developers call “a novel method for building scalable 3D generative models utilizing pre-trained video diffusion models.” Continue reading Meta, Oxford Advance 3D Object Generation with VFusion3D

Nvidia Debuts New Products to Accelerate Adoption of GenAI

After 50 years of SIGGRAPH, the conference has come full circle, from high-tech for PhDs to AI for everyone. That was Nvidia founder and CEO Jensen Huang’s message in back-to-back keynote sessions, including a Q&A with Meta CEO Mark Zuckerberg. Huang touted Universal Scene Description (OpenUSD), discussing developments aiming to speed adoption of the universal 3D data interchange framework for use in everything from robotics to the creation of “highly accurate virtual worlds for the next evolution of AI.” As Zuckerberg’s interlocutor, he prompted the Facebook founder to share a vision of AI’s personalization of social media. Continue reading Nvidia Debuts New Products to Accelerate Adoption of GenAI

Stable Video 4D Adds Time Dimension to Generative Imagery

Stability AI has unveiled an experimental new model, Stable Video 4D, which generates photorealistic 3D video. Building on what it created with Stable Video Diffusion, released in November, this latest model can take moving image data of an object and iterate it from multiple angles — generating up to eight different perspectives. Stable Video 4D can generate five frames across eight views in about 40 seconds using a single inference, according to the company, which says the model has “future applications in game development, video editing, and virtual reality.” Users begin by uploading a single video and specifying desired 3D camera poses. Continue reading Stable Video 4D Adds Time Dimension to Generative Imagery

Captions: Generative Video Startup Raises $60 Million in NYC

Generative video creation and editing platform Captions has raised $60 million in Series C funding. Founded in 2021 by former Microsoft engineer Gaurav Misra and Goldman Sachs alum Dwight Churchill, the company’s technologies — Lipdub, AI Edit and the 3D avatar app AI Creator — have amassed more than 10 million downloads for mobile, the firm says. The C round brings its total raise to $100 million for a stated market valuation of $500 million. With the new funding, Captions plans to expand its presence in New York City, which is “emerging as the epicenter for AI research,” according to Misra. Continue reading Captions: Generative Video Startup Raises $60 Million in NYC

Meta’s 3D Gen Bridges Gap from AI to Production Workflow

Meta Platforms has introduced an AI model it says can generate 3D images from text prompts in under one minute. The new model, called 3D Gen, is billed as a “state-of-the-art, fast pipeline” for turning text input into high-resolution 3D images quickly. The app also adds textures to AI output or existing images through text prompts, and “supports physically-based rendering (PBR), necessary for 3D asset relighting in real-world applications,” Meta explains, adding that in internal tests, 3D Gen outperforms industry baselines on “prompt fidelity and visual quality” and for speed. Continue reading Meta’s 3D Gen Bridges Gap from AI to Production Workflow

Apple Launches Public Demo of Its Multimodal 4M AI Model

Apple has released a public demo of the 4M AI model it developed in collaboration with the Swiss Federal Institute of Technology Lausanne (EPFL). The technology debuts seven months after the model was first open-sourced, allowing informed observers the opportunity to interact with it and assess its capabilities. Apple says 4M was built by applying masked modeling to a single unified Transformer encoder-decoder “across a wide range of input/output modalities — including text, images, geometric and semantic modalities, as well as neural network feature maps.” Continue reading Apple Launches Public Demo of Its Multimodal 4M AI Model

Nokia Makes the First-Ever 3D Spatial Audio Cell Phone Call

Nokia made what it claims is “the world’s first immersive voice and audio call” using cell phones, made possible by the new 3GPP Immersive Voice and Audio Services (IVAS) codec that lets consumers hear 3D spatial sound in real-time. The codec — which Nokia participated in crafting — is a major leap from today’s standard monophonic smartphone voice call experience and is part of the upcoming 5G Advanced standard. The innovation paves the way towards enhanced immersive spatial communications, extended reality and metaverse applications, says Nokia, explaining that it works across “any connected device,” including smartphones, tablets and PCs. Continue reading Nokia Makes the First-Ever 3D Spatial Audio Cell Phone Call

Vision Pro Adds Dual 4K Virtual Display, URSA Cine Immersive

Apple is previewing visionOS 2, the next-gen operating system coming this fall for its Vision Pro mixed-reality headset. The new system includes a Mac Virtual Display that creates the ultra-wide equivalent of two side-by-side 4K monitors. The new OS updates navigational hand gestures and adds a Photos app feature that turns existing 2D pictures into spatial images. At WWDC Apple also announced that Blackmagic Design will release the URSA Cine Immersive, the first commercial camera system designed to capture images for the Vision Pro, and Canon unveiled a dual-lens optical setup for APS-C cameras. Continue reading Vision Pro Adds Dual 4K Virtual Display, URSA Cine Immersive

Acer 3D Camera Makes Glasses-Free Content for Its Displays

Acer has extended its SpatialLabs branding from glasses-free 3D laptops to a 3D camera coming to market in Q3 starting at $549. The Acer SpatialLabs Eyes Stereo Camera has 8MP of resolution per eye and can live stream in 3D to YouTube and enable high-resolution 3D video calls through Zoom, Microsoft Teams and Google Meet. It has a built-in selfie mirror, auto and touch focus capabilities and electronic image stabilization (EIS). It is fully compatible with the Acer Aspire 3D 15 SpatialLabs Edition laptop, released in February, and will also work with other 3D displays, projectors and VR headsets. Continue reading Acer 3D Camera Makes Glasses-Free Content for Its Displays

Autodesk Buys Wonder Dynamics, AI VFX App Wonder Studio

Autodesk is going all-in on artificial intelligence with the acquisition of AI startup Wonder Dynamics, maker of the Wonder Studio VFX tool. Autodesk — whose products include Maya, 3ds Max and Flame — worked with Wonder on a Maya plug-in last year and appears to have been impressed. Wonder Studio was purpose-built to be compatible with 3D tools like Maya, largely automating the process of putting 3D characters within live-action scenes. Terms of the deal were not disclosed, and Autodesk did not detail plans for integrating Wonder Dynamics, but it’s likely the company’s AI expertise will make itself felt across the portfolio. Continue reading Autodesk Buys Wonder Dynamics, AI VFX App Wonder Studio

Looking Glass Debuts Two New Headset-Free Spatial Displays

Looking Glass has launched a new 32-inch, glasses-free spatial display and an OLED version of its 16-inch model. The screens come in both landscape and portrait orientations and are aimed at XR professionals requiring visualization for 3D digital images, video and applications in real time. The 3D displays broadcast 45-100 views for what the company says is an uncompromised group-view experience. Sensors for touchless gesture control are available and the devices support a wide variety of software, including plugins for Unity, Unreal, Blender and WebXR. The 16-inch OLED lists for $4,000 but is offered at $3,000 for a limited time. Continue reading Looking Glass Debuts Two New Headset-Free Spatial Displays

Adobe Considers Sora, Pika and Runway AI for Premiere Pro

Adobe plans to add generative AI capabilities to its Premiere Pro editing platform and is exploring the update with third-party AI technologies including OpenAI’s Sora, as well as models from Runway and Pika Labs, making it easier “to draw on the strengths of different models” within everyday workflows, according to Adobe. Editors will gain the ability to generate and add objects into scenes or shots, remove unwanted elements with a click, and even extend frames and footage length. The company is also developing a video model for its own Firefly AI for video and audio work in Premiere Pro. Continue reading Adobe Considers Sora, Pika and Runway AI for Premiere Pro

Stable Video 3D Generates Orbital Animation from One Image

Stability AI has released Stable Video 3D, a generative video model based on the company’s foundation model Stable Video Diffusion. SV3D, as it’s called,  comes in two versions. Both can generate and animate multi-view 3D meshes from a single image. The more advanced version also let users set “specified camera paths” for a “filmed” look to the video generation. “By adapting our Stable Video Diffusion image-to-video diffusion model with the addition of camera path conditioning, Stable Video 3D is able to generate multi-view videos of an object,” the company explains. Continue reading Stable Video 3D Generates Orbital Animation from One Image

Alibaba’s EMO Can Generate Performance Video from Images

Alibaba is touting a new artificial intelligence system that can animate portraits, making people sing and talk in realistic fashion. Researchers at the Alibaba Group’s Institute for Intelligent Computing developed the generative video framework, calling it EMO, short for Emote Portrait Alive. Input a single reference image along with “vocal audio,” as in talking or singing, and “our method can generate vocal avatar videos with expressive facial expressions and various head poses,” the researchers say, adding that EMO can generate videos of any duration, “depending on the length of video input.” Continue reading Alibaba’s EMO Can Generate Performance Video from Images