By
Paula ParisiDecember 18, 2024
Blackmagic Design is live with URSA Cine Immersive pre-orders. If it meets its late Q1 2025 ship date the $29,995 the camera will be the first on the market optimized for the Apple Immersive Video (AIV) format compatible with the Apple Vision Pro mixed-reality headset. Currently, there isn’t much content that takes advantage of the Vision Pro’s immersive features. The Cine Immersive captures 3D footage in resolution of 8160 x 7200 per eye at 90 fps. The package includes a fixed-distance lens and 8TB of onboard network storage. Also in Q1, DaVinci Resolve Studio will be updated to support AIV editing. Continue reading URSA Cine Immersive for Apple Vision Pro Set for Q1 at $30K
By
Hank GerbaDecember 16, 2024
Google has introduced Gemini 2.0, the latest version of its multimodal AI model, signaling a shift toward what the company is calling “the agentic era.” The upgraded model promises not only to outperform previous iterations on standard benchmarks but also introduces more proactive, or agentic, functions. The company announced that “Project Astra,” its experimental assistant, would receive updates that allow it to use Google Search, Lens, and Maps, and that “Project Mariner,” a Chrome extension, would enable Gemini 2.0 to navigate a user’s web browser to complete tasks autonomously. Continue reading Google Releases Gemini 2.0 in Shift Toward Agentic Era of AI
By
Paula ParisiDecember 13, 2024
With AI powering a range of new world-building apps, 2025 could be the year the metaverse finally makes an impact. Midjourney joins the world-building club with Patchwork, a collaborate canvas for creating “infinite” fictional worlds. Now in research preview, the tool is being developed as a standalone app, though preview access requires a Midjourney Discord account linked to a Google account. Users are able to connect characters and worlds, and “share” their developing world — evolving as a “board” — with up to 100 collaborative partners on Midjourney (though the company recommends fewer participants for a more focused experience). Continue reading Midjourney Touts Collaborative World-Building App Patchwork
By
Paula ParisiDecember 12, 2024
World Labs, the AI startup co-founded by Stanford AI pioneer Fei-Fei Li, has debuted a “spatial intelligence” system that can generate 3D worlds from a single image. Although the output is not photorealistic, the tech could be a breakthrough for animation companies and video game developers. Deploying what it calls Large World Models (LWMs), World Labs is focused on transforming 2D images into turnkey 3D environments with which users can interact. Observers say that reciprocity is what sets World Labs’ technology apart from offerings by other AI companies that transform 2D to 3D. Continue reading World Labs AI Lets Users Create 3D Worlds from Single Photo
By
Paula ParisiDecember 9, 2024
James Cameron’s Lightstorm Entertainment has debuted a new unit, Lightstorm Vision, which launches with a deal that sees Meta Quest become its exclusive mixed reality platform. The multiyear deal will see the Hollywood heavyweight partnering with Meta Platforms “to scale the creation of world-class 3D entertainment experiences spanning live sports and concerts, feature films, and TV series featuring big-name IP on Meta Quest,” Meta announced. In addition to producing original content for the Meta Quest, Lightstorm Vision will use its tech to help third parties create new programming and potentially conform existing work through MR conversion. Continue reading Cameron’s Lightstorm Vision Partners with Meta to Advance MR
By
Paula ParisiDecember 6, 2024
Google DeepMind’s new Genie 2 is a large foundation world model that generates interactive 3D worlds that are being likened to video games. “Games play a key role in the world of artificial intelligence research,” says Google DeepMind, noting “their engaging nature, challenges and measurable progress make them ideal environments to safely test and advance AI capabilities.” Based on a simple prompt image, Genie 2 is capable of producing “an endless variety of action-controllable, playable 3D environments” — suitable for training and evaluating embodied agents — that can be played by a human or AI agent using keyboard and mouse inputs. Continue reading DeepMind Genie 2 Creates Worlds That Emulate Video Games
By
Paula ParisiNovember 8, 2024
Wonder Animation is the latest tool from Wonder Dynamics, the AI startup founded by actor Tye Sheridan and VFX artist Nikola Todorovic in 2017 that Autodesk purchased in May. Now in beta, Wonder Animation can automatically transpose live-action footage into stylized 3D animation. Creators can shoot using any camera, on any set or location, and easily convert to 3D CGI. Matching the camera position and movement to the characters and environment, Wonder Animation lets you film using any camera system and lenses, edit those shots using Maya, Blender or Unreal, and then reconstruct the result as 3D animation using AI. Continue reading Autodesk’s AI Tool Turns Live-Action Video into 3D Animation
By
Paula ParisiNovember 6, 2024
New York-based AI firm Runway has added 3D video camera controls to Gen-3 Alpha Turbo, giving users the ability to manipulate granular aspects of the scene they are generating using effects whether originating from text prompts, uploaded images or self-created video. Users can zoom in and out on a subject or scene, moving around an AI-generated character or form in 3D as if on a real set or actual location. The new feature, available now, lets creators “choose both the direction and intensity of how you move through your scenes for even more intention in every shot,” Runway explains. Continue reading Runway Adds 3D Video Cam Controls to Gen-3 Alpha Turbo
By
Paula ParisiOctober 25, 2024
Runway is launching Act-One motion capture system that uses video and voice recordings to map human facial expressions onto characters using the company’s latest model, Gen-3 Alpha. Runway calls it “a significant step forward in using generative models for expressive live action and animated content.” Compared to past facial capture techniques — which typically require complex rigging — Act-One is driven directly and only by the performance of an actor, requiring “no extra equipment,” making it more likely to capture and preserve an authentic, nuanced performance, according to the company. Continue reading Runway’s Act-One Facial Capture Could Be a ‘Game Changer’
By
Paula ParisiOctober 16, 2024
Adobe has launched a public beta of its Generate Video app, part of the Firefly Video model, which users can try for free on a dedicated website. Login is required, and there is still a waitlist for unfettered access, but the Web app facilitates up to five seconds of video generation using text and image prompts. It can turn 2D pictures into 3D animation and is also capable of producing video with dynamic text. The company has also added an AI feature called “Extend Video” to Premiere Pro to lengthen existing footage by two seconds. The news has the media lauding Adobe for beating OpenAI’s Sora and Google’s Veo to market. Continue reading Adobe Promos AI in Premiere Pro, ‘Generate Video’ and More
By
Paula ParisiOctober 8, 2024
Apple has released a new AI model called Depth Pro that can create a 3D depth map from a 2D image in under a second. The system is being hailed as a breakthrough that could potentially revolutionize how machines perceive depth, with transformative impact on industries from augmented reality to self-driving vehicles. “The predictions are metric, with absolute scale” without relying on the camera metadata typically required for such mapping, according to Apple. Using a consumer-grade GPU, the model can produce a 2.25-megapixel depth map using a single image in only 0.3 seconds. Continue reading Apple Advances Computer Vision with Its Depth Pro AI Model
By
Paula ParisiSeptember 20, 2024
A newly redesigned Snapchat experience is built around a three-tab user interface called Simple Snapchat. As part of that effort, the social platform is launching more generative video features, including text-to-video as part of the app’s Lens Studio AR authoring tool. Easy Lens allows the quick generation of Lenses by typing text prompts, making it possible to do things like experiment with Halloween costumes or explore looks for back to school. Launching in beta for select creators, Snap says the new features are designed for all ability levels. The company is also updating its GenAI Suite and adding an Animation Library of “hundreds of high-quality movements.” Continue reading Snapchat Is Getting a Redesign and Generative Text-to-Video
By
Paula ParisiSeptember 10, 2024
During the 10th annual Roblox Developers Conference (RDC 2024) in San Jose, the gaming platform announced it is opening to global currencies in addition to its own Robux, which generates billions in virtual transactions each year. Starting later this year, a small test bed of developers will be able to charge real money for games that charge fees, with a program expected to open “to all eligible creators by mid-2025.” The massively multiplayer online platform that lets users build online game worlds also discussed a project to develop its own AI foundation model to power generative 3D creation on the platform. Continue reading Roblox Adds Real Currency, Teases Its Coming Generative AI
By
Paula ParisiAugust 29, 2024
Canadian generative video startup Viggle AI, which specializes in character motion, has raised $19 million in Series A funding. Viggle was founded in 2022 on the premise of providing a simplified process “to create lifelike animations using simple text-to-video or image-to-video prompts.” The result has been robust adoption among meme creators, with many viral videos circulating among social media platforms powered by Viggle, including one featuring Joaquin Phoenix as the Joker mimicking the movements of rapper Lil Yachty. Viggle’s Discord community has four million members including “both novice and experienced animators,” according to the company. Continue reading Viggle AI Raises $19 Million on the Power of Memes and More
By
Paula ParisiAugust 26, 2024
Samsung Electronics, which teased a glasses-free 3D gaming monitor at CES in January, officially announced the scheduled release of two versions at Gamescom last week. Both sizes employ light field display (LFD) technology to create what Samsung calls “lifelike 3D images” from 2D content by using a lenticular lens on the front panel. “Combined with Eye Tracking and View Mapping technology, Odyssey 3D ensures an optimized 3D experience without the need for separate 3D glasses,” according to Samsung. A built-in stereo camera monitors the movement of both eyes while proprietary View Mapping continuously adjusts the image to fuel depth perception. Continue reading Samsung Set to Release Glasses-Free Odyssey 3D Monitors