Sam Altman Reveals Plans to Simplify OpenAI’s Product Line

OpenAI has decided to simplify its product offerings. A month after announcing the in-development GPT-o3 as its next frontier model, the company has canceled it as a standalone release, explaining that it would be integrated into the upcoming GPT-5 instead. “A top goal for us is to unify o-series models and GPT-series models by creating systems that can use all our tools, know when to think for a long time or not, and generally be useful for a very wide range of tasks,” OpenAI co-founder and CEO Sam Altman wrote in a social media post this week. Expected to ship later this year, the GPT-5 models will incorporate voice, canvas, search, deep research and more, OpenAI says. Continue reading Sam Altman Reveals Plans to Simplify OpenAI’s Product Line

Anthropic Launches Economic Index to Track the AI Economy

Anthropic is launching an AI Economic Index aimed at understanding the technology’s effects on labor markets and the economy over time. The Anthropic Economic Index kicks off with a report that indicates the most concentrated uses of AI today are in software development and technical writing tasks. The information was culled by analyzing data from millions of anonymized conversations on Claude.ai with the goal of revealing how AI is being incorporated into real-world tasks across global markets. It also found that AI usage trends more toward augmenting human capabilities (57 percent), compared to using it for automation (43 percent). Continue reading Anthropic Launches Economic Index to Track the AI Economy

Meta Adds Indigenous Languages to Speech and Translation AI

Meta is seeking to make AI more inclusive with a program to support underserved languages “and help bring their speakers into the digital conversation.” Meta’s Fundamental AI Research (FAIR) unit has teamed with UNESCO to launch the Language Technology Partner Program, which is looking for people who can provide more than 10 hours of speech recordings (with transcriptions) and chunks of written text (200+ sentences, with translation) in diverse languages. “Partners will work with our teams to help integrate these languages into AI-driven speech recognition and machine translation models, which when released will be open sourced,” Meta said. Continue reading Meta Adds Indigenous Languages to Speech and Translation AI

Hugging Face Has Developed Tiny Yet Powerful Vision Models

Most people know Hugging Face as a resource-sharing community, but it also builds open-source applications and tools for machine learning. Its recent release of vision-language models small enough to run on smartphones while outperforming competitors that rely on massive data centers is being hailed as “a remarkable breakthrough in AI.”  The new models — SmolVLM-256M and SmolVLM-500M — are optimized for “constrained devices” with less than around 1GB of RAM, making them ideal for mobile devices including laptops and also convenient for those interested in processing large amounts of data cheaply and with a low-energy footprint. Continue reading Hugging Face Has Developed Tiny Yet Powerful Vision Models

Codename Goose: Block Unveils Open-Source AI Agent Builder

Jack Dorsey’s financial tech and media firm Block (formerly Square) has released a platform for building AI agents: Codename Goose. Previously available in beta, Goose is primarily designed to build agents for coding and software development, but Block built in many basic features that could be applied to general purpose pursuits. Because it is open source and offered under Apache License 2.0, the hope is that developers will apply it to varied use cases. A leading feature of Codename Goose is its flexibility. It can integrate a wide range of large language models, letting developers use it with their preferred model. Continue reading Codename Goose: Block Unveils Open-Source AI Agent Builder

Facebook, Instagram, WhatsApp Get Meta AI Memory Boost

Meta is rolling out personalization updates to its Meta AI personal assistant. At the end of last year, the company introduced a feature that lets Meta AI remember what you’ve shared with it in one-on-one chats on WhatsApp and Messenger so it could produce more relevant responses. That feature will now be available to Meta AI on Facebook, Messenger and WhatsApp for iOS and Android in the U.S. and Canada. “Meta AI will only remember certain things you tell it in 1:1 conversations (not group chats), and you can delete its memories at any time,” explains the company. Continue reading Facebook, Instagram, WhatsApp Get Meta AI Memory Boost

Chinese AI Startup DeepSeek Disrupting the U.S. Tech Sector

Hangzhou-based AI firm DeepSeek is roiling the U.S. tech sector and upending financial markets. The startup has managed to become competitive with Silicon Valley’s deep learning firms despite U.S. sanctions that prevent Chinese technology companies from buying premium chips. DeepSeek has made it into the global top 10 in terms of model performance, and as of this week had the top-ranked free AI assistant at the Apple App Store. DeepSeek’s new R1 model has drawn attention for using less computing power than competing systems, while performing comparably, despite having been developed using older Nvidia chips. Continue reading Chinese AI Startup DeepSeek Disrupting the U.S. Tech Sector

CES: Nvidia’s Cosmos Models Teach AI About Physical World

Nvidia Cosmos, a platform of generative world foundation models (WFMs) and related tools to advance the development of physical AI systems like autonomous vehicles and robots, was introduced at CES 2025. Cosmos WFMs are designed to provide developers a way to generate massive amounts of photo-real, physics-based synthetic data to train and evaluate their existing models. The goal is to reduce costs by streamlining real-world testing with a ready data pipeline. Developers can also build custom models by fine-tuning Cosmos WFMs. Cosmos integrates Nvidia Omniverse, a physics simulation tool used for entertainment world-building. Continue reading CES: Nvidia’s Cosmos Models Teach AI About Physical World

CES: Samsung and Google Team on Spatial Audio Standard

Samsung Electronics has teamed with Google on a new spatial sound standard, Eclipsa Audio, that could emerge as a free alternative to Dolby Atmos. On display at CES 2025 in Las Vegas this week, the format is rolling out across Samsung’s line of 2025 TVs and soundbars, and Google will support it on the content side by enabling Eclipsa 3D audio on some YouTube videos this year. Samsung has been a notable holdout on Dolby Vision HDR embracing instead the competing HDR10+. Now the South Korean electronics giant seems to be staking out its own turf in 3D audio, advocating for open source. Continue reading CES: Samsung and Google Team on Spatial Audio Standard

Meta Rolls Out Watermarking, Behavioral and Concept Models

Meta’s FAIR (Fundamental AI Research) team has unveiled recent work in areas ranging from transparency and safety to agents, and architectures for machine learning. The projects include Meta Motivo, a foundation model for controlling the behavior of virtual embodied agents, and Video Seal, an open-source model for video watermarking. All were developed in the unit’s pursuit of advanced machine intelligence, helping “models to learn new information more effectively and scale beyond current limits.” Meta announced it is sharing the new FAIR research, code, models and datasets so the research community can build upon its work. Continue reading Meta Rolls Out Watermarking, Behavioral and Concept Models

Meta’s Llama 3.3 Delivers More Processing for Less Compute

Meta Platforms has packed more artificial intelligence into a smaller package with Llama 3.3, which the company released last week. The open-source large language model (LLM) “improves core performance at a significantly lower cost, making it even more accessible to the entire open-source community,” Meta VP of Generative AI Ahmad Al-Dahle wrote on X social. The 70 billion parameter text-only Llama 3.3 is said to perform on par with the 405 billion parameter model that was part of Meta’s Llama 3.1 release in July, with less computing power required, significantly lowering its operational costs. Continue reading Meta’s Llama 3.3 Delivers More Processing for Less Compute

Hume AI Introduces Voice Control and Claude Interoperability

Artificial voice startup Hume AI has had a busy Q4, introducing Voice Control, a no-code artificial speech interface that gives users control over 10 voice dimensions ranging from “assertiveness” to “buoyancy” and “nasality.” The company also debuted an interface that “creates emotionally intelligent voice interactions” with Anthropic’s foundation model Claude that has prompted one observer to ponder the possibility that keyboards will become a thing of the past when it comes to controlling computers. Both advances expand on Hume’s work with its own foundation model, Empathic Voice Interface 2 (EVI 2), which adds emotional timbre to AI voices. Continue reading Hume AI Introduces Voice Control and Claude Interoperability

Qwen with Questions: Alibaba Previews New Reasoning Model

Alibaba Cloud has released the latest entry in its growing Qwen family of large language models. The new Qwen with Questions (QwQ) is an open-source competitor to OpenAI’s o1 reasoning model. As with competing large reasoning models (LRMs), QwQ can correct its own mistakes, relying on extra compute cycles during inference to assess its responses, making it well suited for reasoning tasks like math and coding. Described as an “experimental research model,” this preview version of QwQ has 32-billion-parameters and a 32,000-token context, leading to speculation that a more powerful iteration is in the offing. Continue reading Qwen with Questions: Alibaba Previews New Reasoning Model

Lightricks LTX Video Model Impresses with Speed and Motion

Lightricks has released an AI model called LTX Video (LTXV) it says generates five seconds of 768 x 512 resolution video (121 frames) in just four seconds, outputting in less time than it takes to watch. The model can run on consumer-grade hardware and is open source, positioning Lightricks as a mass market challenger to firms like Adobe, OpenAI, Google and their proprietary systems. “It’s time for an open-sourced video model that the global academic and developer community can build on and help shape the future of AI video,” Lightricks co-founder and CEO Zeev Farbman said. Continue reading Lightricks LTX Video Model Impresses with Speed and Motion

Anthropic Protocol Intends to Standardize AI Data Integration

Anthropic is releasing what it hopes will be a new standard in data integration for AI. Called the Model Context Protocol (MCP), its goal is to eliminate the need to customize each integration by having code written each time a company’s data is connected to a model. The open-source MCP tool could become a universal way to link data sources to AI. The aim is to have models querying databases directly. MCP is “a new standard for connecting AI assistants to the systems where data lives, including content repositories, business tools, and development environments,” according to Anthropic. Continue reading Anthropic Protocol Intends to Standardize AI Data Integration