Adobe Promos AI in Premiere Pro, ‘Generate Video’ and More

Adobe has launched a public beta of its Generate Video app, part of the Firefly Video model, which users can try for free on a dedicated website. Login is required, and there is still a waitlist for unfettered access, but the Web app facilitates up to five seconds of video generation using text and image prompts. It can turn 2D pictures into 3D animation and is also capable of producing video with dynamic text.  The company has also added an AI feature called “Extend Video” to Premiere Pro to lengthen existing footage by two seconds. The news has the media lauding Adobe for beating OpenAI’s Sora and Google’s Veo to market. Continue reading Adobe Promos AI in Premiere Pro, ‘Generate Video’ and More

Meta Announces New GenAI Video Tools at Advertising Week

Meta is rolling out new generative AI advertising tools for video creation on Facebook and Instagram. The expansion to the Advantage+ creative ad suite will become widely available to advertisers in early 2025. The announcement, made at Advertising Week in New York last week, was positioned as a way for marketers to improve campaign performance on Meta’s social platforms. The new tools will allow brands to convert static images into video ads. The company also announced a new full screen video tab for Facebook that feeds short-form Reels with long-form and live-stream content. Continue reading Meta Announces New GenAI Video Tools at Advertising Week

MiniMax’s Hailuo AI Rolls Out New Image-to-Video Capability

Hailuo, the free text-to-video generator released last month by the Alibaba-backed company MiniMax, has delivered its promised image-to-video feature. Founded by AI researcher Yan Junjie, the Shanghai-based MiniMax also has backing from Tencent. The model earned high marks for what has been called “ultra realistic” video, and MiniMax says the new image-to-video feature will improve output across the board as a result of “text-and-image joint instruction following,” which means Hailuo now “seamlessly integrates both text and image command inputs, enhancing your visuals while precisely adhering to your prompts.” Continue reading MiniMax’s Hailuo AI Rolls Out New Image-to-Video Capability

Free Adobe Content Authenticity Web App Shields Against AI

Adobe is introducing a free, web-based Content Authenticity app that lets creators “sign” their work with the aim of protecting rights and controlling attribution against unwanted AI attention. It allows creators to assign “do not train” tags to images, video or audio. Batch designation is another convenience those with voluminous output will appreciate as a time saver. Users can select the Generative AI Training and Usage Preference options in the Adobe Content Authenticity app to set preferences, whether or not the work was created using Adobe Creative Cloud apps. Continue reading Free Adobe Content Authenticity Web App Shields Against AI

Samsung Developer Conference Emphasizes AI, One UI 7 UX

Samsung heralded the world of personalized AI at its 10th annual developer conference, where Samsung Electronics Vice Chairman, CEO and Head of Device eXperience Jong-Hee Han said those who own the company’s top of the line TVs will soon have generative AI, ChatGPT and a more responsive relationship with Bixby, Samsung’s smart assistant. The company introduced AI Cast, making it simpler to get intelligence from Galaxy phones to Samsung TVs. The Galaxy S24 series, released early this year, has native AI that will soon generate content that can be beamed to a sprawling TV screen. Continue reading Samsung Developer Conference Emphasizes AI, One UI 7 UX

Meta’s Movie Gen Model is a Powerful Content Creation Tool

Meta Platforms has unveiled Movie Gen, a new family of AI models that generates video and audio content. Coming to Instagram next year, Movie Gen also allows a high degree of editing and effects customization using text prompts. Meta CEO Mark Zuckerberg demonstrated its abilities last week in an example shared on his Instagram account, where he sends a leg press machine at the gym through transformations as a steam punk machine and one made of molten gold. The models have been trained on a combination of licensed and publicly available datasets. Continue reading Meta’s Movie Gen Model is a Powerful Content Creation Tool

Apple Advances Computer Vision with Its Depth Pro AI Model

Apple has released a new AI model called Depth Pro that can create a 3D depth map from a 2D image in under a second. The system is being hailed as a breakthrough that could potentially revolutionize how machines perceive depth, with transformative impact on industries from augmented reality to self-driving vehicles. “The predictions are metric, with absolute scale” without relying on the camera metadata typically required for such mapping, according to Apple. Using a consumer-grade GPU, the model can produce a 2.25-megapixel depth map using a single image in only 0.3 seconds. Continue reading Apple Advances Computer Vision with Its Depth Pro AI Model

Google Serving Ads in AI Overviews and Lens Search Results

Having demonstrated how advertisements in its AI Overviews would work back in May at its Google Marketing Live event, the search giant is now adding the feature for U.S. mobile users and plans to include Google Lens shopping ads “above and alongside visual search results by the end of the year.” “The ways people ask questions today have expanded beyond the search box,” notes Google, explaining the move as a response to that evolution, as artificial intelligence technology has helped consumers use their voice and cameras “to explore the world around them.” Continue reading Google Serving Ads in AI Overviews and Lens Search Results

Gen AI Among Pinterest Performance+ Advertising Upgrades

Pinterest is unveiling AI and automation tools to support advertising campaigns on the platform. At its annual Pinterest Presents global advertiser summit the social platform known for inspiration boards and shopping unveiled a generative background feature for its Pinterest Performance+ suite, allowing advertisers to replace flat backgrounds with “lifestyle imagery” using AI. Launched earlier this year, Pinterest says Performance+ lets advertisers create campaigns in about half the time and is already producing positive results. Now it is rolling out improvements to the feature in time for the holiday shopping season. Continue reading Gen AI Among Pinterest Performance+ Advertising Upgrades

MIT Spinoff Liquid Eschews GPTs for Its Fluid Approach to AI

AI startup Liquid, founded by alums of MIT’s Computer Science and Artificial Intelligence Laboratory (CSAIL), has released its first models. Called Liquid Foundation Models, or LFMs, the multimodal family approaches “intelligence” differently than the pre-trained transformer models that dominate the field. Instead, the LFMs take a path of “first principles,” which MIT describes as “the same way engineers build engines, cars, and airplanes,” explaining that the models are large neural networks with computational units “steeped in theories of dynamic systems, signal processing and numeric linear algebra.” Continue reading MIT Spinoff Liquid Eschews GPTs for Its Fluid Approach to AI

Snapchat: My AI Goes Multimodal with Google Cloud, Gemini

Snap Inc. is leveraging its relationship with Google Cloud to use Gemini for powering generative AI experiences within Snapchat’s My AI chatbot. The multimodal capabilities of Gemini on Vertex AI will greatly increase the My AI chatbot’s ability to understand and operate across different types of information such as text, audio, image, video and code. Snapchatters can use My AI to take advantage of Google Lens-like features, including asking the chatbot “to translate a photo of a street sign while traveling abroad, or take a video of different snack offerings to ask which one is the healthiest option.” Continue reading Snapchat: My AI Goes Multimodal with Google Cloud, Gemini

Runway Launches $5M AI Film Fund as Open Call to Creators

Artificial intelligence platform Runway has launched The Hundred Film Fund to help finance 100 projects that use its AI to tell stories. Created by the company through its Runway Studios, the Fund is starting with $5 million, “with the potential to grow to $10 million.” Runway is presenting the Fund as “an open call to all creative professionals who have AI-augmented film projects in the pre- or post-production phases and are in need of funding.” Directors, producers and screenwriters are among those invited to apply. The program will consider all formats, from features to shorts, documentaries, experimental projects, music videos and more. Continue reading Runway Launches $5M AI Film Fund as Open Call to Creators

Digital Domain Leverages AWS for Its Virtual Human Initiative

Visual effects studio Digital Domain has brought its Autonomous Virtual Human project to Amazon Web Services, which will provide generative AI and machine learning tools and provide Digital Domain’s creations and processes a home in the global cloud. The collaboration “aims to propel the evolution and global reach of Digital Domain’s AVH technology and expand its use for multiple industries, including entertainment, gaming, healthcare, hospitality, and commercial applications,” Amazon said in a statement that emphasizes “AWS cloud services, particularly Amazon Bedrock,” as providing the infrastructure and adaptability “to drive AVH’s growth.” Continue reading Digital Domain Leverages AWS for Its Virtual Human Initiative

Meta Unveils New Open-Source Multimodal Model Llama 3.2

Meta’s Llama 3.2 release includes two new multimodal LLMs, one with 11 billion parameters and one with 90 billion — considered small- and medium-sized — and two lightweight, text-only models (1B and 3B) that fit onto edge and mobile devices. Included are pre-trained and instruction-tuned versions. In addition to text, the multimodal models can interpret images, supporting apps that require visual understanding. Meta says the models are free and open source. Alongside them, the company is releasing “the first official Llama Stack distributions,” enabling “turnkey deployment” with integrated safety. Continue reading Meta Unveils New Open-Source Multimodal Model Llama 3.2

OpenAI Rolls Out Advanced Voice Mode Feature for ChatGPT

As OpenAI gears up to become a for-profit company next year, it is releasing ChatGPT Advanced Voice Mode, which brings a humanlike conversation mode to ChatGPT 4o. All U.S. subscribers to ChatGPT Plus and Team plans will gain access to the new feature, which will also be made available to those paying for ChatGPT Edu and Enterprise plans in the coming weeks. The firm is also adding five new voices and allowing customers to save personalized instructions for the voice assistant, including memory behaviors. Concurrently, executives including CTO Mira Murati have resigned as the company pivots to commerciality. Continue reading OpenAI Rolls Out Advanced Voice Mode Feature for ChatGPT