By
Paula ParisiJune 20, 2024
Meta Platforms is publicly releasing five new AI models from its Fundamental AI Research (FAIR) team, which has been experimenting with artificial intelligence since 2013. These models including image-to-text, text-to-music generation, and multi-token prediction tools. Meta is introducing a new technique called AudioSeal, an audio watermarking technique designed for the localized detection of AI-generated speech. “AudioSeal makes it possible to pinpoint AI-generated segments within a longer audio snippet,” according to Meta. The feature is timely in light of concern about potential misinformation surrounding the fall presidential election. Read more
By
Paula ParisiJune 20, 2024
YouTube is experimenting with a feature that allows viewers to add contextual “Notes” under videos, similar to what X does with its Community Notes. The Google-owned company says the intent is to provide clarity around things like “when a song is meant to be a parody,” when newly reviewed products are available for purchase, or “when older footage is mistakenly portrayed as a current event.” However, the timing preceding a pivotal U.S. presidential election and facing concerns about deepfakes and misinformation is no doubt intentional. The pilot will initially be available on mobile in the United States. Read more
By
Paula ParisiJune 20, 2024
Netflix has selected Dallas, Texas and Pennsylvania’s King of Prussia as the first two cities for its massive location-based entertainment venues, set to open in 2025. Specifying “they’re not exactly theme parks,” the Netflix House destinations will feature shopping, eateries and “regularly updated” experiential activities that tie-in with major franchises like “Bridgerton,” “Stranger Things” and “Squid Game.” Netflix CMO Marian Lee says the company has already launched more than 50 experiences in 25 cities, with Netflix House representing “the next generation” venue for the streaming giant. Read more
By
Paula ParisiJune 19, 2024
Runway ML has introduced a new foundation model, Gen-3 Alpha, which the company says can generate high-quality, realistic scenes of up to 10 seconds long from text prompts, still images or a video sample. Offering a variety of camera movements, Gen-3 Alpha will initially roll out to Runway’s paid subscribers, but the company plans to add a free version in the future. Runway says Gen-3 Alpha is the first of a new series of models trained on the company’s new large-scale multimodal infrastructure, which offers improvements “in fidelity, consistency, and motion over Gen-2,” released last year. Read more
By
Paula ParisiJune 19, 2024
Google DeepMind has unveiled new research on AI tech it calls V2A (“video-to-audio”) that can generate soundtracks for videos. The initiative complements the wave of AI video generators from companies ranging from biggies like OpenAI and Alibaba to startups such as Luma and Runway, all of which require a separate app to add sound. V2A technology “makes synchronized audiovisual generation possible” by combining video pixels with natural language text prompts “to generate rich soundscapes for the on-screen action,” DeepMind writes, explaining that it can “create shots with a dramatic score, realistic sound effects or dialogue.” Read more