By
Paula ParisiMay 15, 2024
IBM has released a family of its Granite AI models to the open-source community. The series of decoder-only Granite code models are purpose-built to write computer code for enterprise developers, with training in 116 programming languages. These Granite models range in size from 3 to 34 billion parameters in base model and instruction-tuned variants. They offer a range of uses, from modernizing older code with new languages to optimizing programs for on-device memory constraints, such as might be experienced when conforming for mobile gadgets. In addition to generation, the models can repair and explain code. Continue reading IBM Introduces Granite LLMs for Enterprise Code Developers
Meta Platforms announced an expanded collection of generative AI features, tools and services for advertisers and businesses. The enhanced AI features include full image and text generation, text overlay capabilities, and image expansion for Reels and the Feed in Facebook and Instagram. The updated tools will be available via Meta Ads Manager through Advantage+ creative. According to Meta: “Our goal is to help you at every step of your journey, whether that’s improving ad performance by helping you develop creative variations, automating certain parts of the ad creation process, or increasing your credibility and engagement through Meta Verified.” Continue reading Meta Launches Enhanced Generative AI Tools for Advertisers
By
Paula ParisiMay 8, 2024
Google introduced Threat Intelligence at the RSA Conference in San Francisco this week. Claiming actionable information at “visibility only Google can deliver, based on billions of signals across devices and emails,” Threat Intelligence draws on the capabilities of the company’s Gemini LLMs, Mandiant cybersecurity arm, and cloud-based VirusTotal tool. An AI-powered Gemini agent “provides conversational search” across the repository of Threat Intelligence, “enabling customers to gain insights and protect themselves from threats faster than ever before,” Google says in a move to empower even small teams without IT departments with threat protection. Continue reading Google Taps AI for Its ‘Threat Intelligence’ Cybersecurity Plan
Social platform X, formerly Twitter, has added news summaries generated by Grok, the chatbot of Elon Musk’s xAI, exclusively for Premium subscribers. The personalized trending “Stories” appear in the For You tab of the app’s Explore section. The feature is available on the web and for iOS users for now. For You consists of trending items shared across the platform and within individual networks, along with recommended posts. Grok essentially becomes the voice of X’s news aggregator, teasing articles with short blurbs that link to what X users are saying on the subject. Critics are concerned about Grok’s tone, interpretation of information, and potential for creating fake news. Continue reading X Premium Intros News Blurbs Generated by Musk’s Grok AI
As part of its spring product release, Yelp has added Yelp Assistant, an AI feature designed to connect consumers with relevant business professionals. The chatbot leverages OpenAI language models along with Yelp data to find the right fit and can interactively quiz users to learn more about their project and needs. Available on iOS (with plans for Android later this year), Yelp’s move is part of a trend using artificial intelligence to provide operational help managing households and life tasks. For example, the California-based energy intelligence firm Bidgely is now using AI to generate individualized consumer profiles on energy usage. Continue reading Yelp Assistant Joins Movement to Add AI Consumer Services
By
ETCentric StaffApril 26, 2024
The trend toward small language models that can efficiently run on a single device instead of requiring cloud connectivity has emerged as a focus for Big Tech companies involved in artificial intelligence. Apple has released the OpenELM family of open-source models as its entry in that field. OpenELM uses “a layer-wise scaling strategy” to efficiently allocate parameters within each layer of the transformer model, resulting in what Apple claims is “enhanced accuracy.” The “ELM” stands for “Efficient Language Models,” and one media outlet couches it as “the future of AI on the iPhone.” Continue reading Apple Unveils OpenELM Tech Optimized for Local Applications
By
ETCentric StaffApril 25, 2024
Microsoft, which has been developing small language models (SLMs) for some time, has announced its most-capable SLM family, Phi-3. SLMs can accomplish some of the same functions as LLMs, but are smaller and trained on less data. That smaller footprint makes them well suited to run in a local environment, which means they’re ideal for smartphones, where in theory they would not even need an Internet connection to run. Microsoft claims the Phi-3 open models can outperform “models of the same size and next size up across a variety of benchmarks that evaluate language, coding and math capabilities.” Continue reading Microsoft Small Language Models Are Ideal for Smartphones
By
ETCentric StaffApril 17, 2024
Meta is testing a new large language chatbot, Meta AI, on social platforms in parts of India and Africa. The chatbot was introduced in late 2023, and began testing on U.S. WhatApp users in March. The test is expanding to include more territories and the addition of Instagram and Facebook Messenger. India is reported to be Meta’s largest social market, with more than 500 million Facebook and WhatsApp users, and has big implications as the company scales up its AI plans to compete against OpenAI and others. The Meta AI chatbot answers questions and generates photorealistic images. Continue reading Meta Tests Image-Generating Social Chatbot on Its Platforms
By
ETCentric StaffApril 11, 2024
Google is moving its most powerful artificial intelligence model, Gemini 1.5 Pro, into public preview for developers and Google Cloud customers. Gemini 1.5 Pro includes what Google claims is a breakthrough in long context understanding, with the ability to run 1 million tokens of information “opening up new possibilities for enterprises to create, discover and build using AI.” Gemini’s multimodal capabilities allow it to process audio, video, text, code and more, which when combined with long context, “enables enterprises to do things that just weren’t possible with AI before,” according to Google. Continue reading Google Offers Public Preview of Gemini Pro for Cloud Clients
By
ETCentric StaffApril 9, 2024
Opera has become the first browser to add support for large language models (LLMs). At this point the feature is experimental, and available only on the Opera One Developer browser as part of the AI Feature Drops program. The update offers about 150 LLMs from more than 50 different families, including Meta’s LLaMA, Google’s Gemma, Mixtral and Vicuna. Opera had previously only offered local support for its own Aria AI, a competitor to Microsoft Copilot and OpenAI’s ChatGPT. The local LLMs are being offered for testing as a complimentary addition to Opera’s online Aria service. Continue reading Opera Browser Is Experimenting with Local Support for LLMs
By
ETCentric StaffApril 4, 2024
Apple has developed a large language model it says has advanced screen-reading and comprehension capabilities. ReALM (Reference Resolution as Language Modeling) is artificial intelligence that can see and read computer screens in context, according to Apple, which says it advances technology essential for a true AI assistant “that aims to allow a user to naturally communicate their requirements to an agent, or to have a conversation with it.” Apple claims that in a benchmark against GPT-3.5 and GPT-4, the smallest ReALM model performed “comparable” to GPT-4, with its “larger models substantially outperforming it.” Continue reading Apple’s ReALM AI Advances the Science of Digital Assistants
By
ETCentric StaffApril 4, 2024
Microsoft and OpenAI are contemplating an AI supercomputer data center that may cost as much as $100 billion. Called Stargate, the aim would be to have it operational by 2008 to drive OpenAI’s next generation of artificial intelligence. According to reports, the Stargate complex would span hundreds of U.S. acres and use up to 5 gigawatts of power, which is massive (the equivalent of a substantial metropolitan power grid). In light of those power needs, a nuclear power source is said to be under consideration. The project is not yet green-lit, and no U.S. location has been selected. Continue reading Microsoft, OpenAI Considering a Supercomputer Data Center
By
ETCentric StaffMarch 29, 2024
Databricks, a San Francisco-based company focused on cloud data and artificial intelligence, has released a generative AI model called DBRX that it says sets new standards for performance and efficiency in the open source category. The mixture-of-experts (MoE) architecture contains 132 billion parameters and was pre-trained on 12T tokens of text and code data. Databricks says it provides the open community and enterprises who want to build their own LLMs with capabilities previously limited to closed model APIs. Compared to other open models, Databricks claims it outperforms alternatives including Llama 2-70B and Mixtral on certain benchmarks. Continue reading Databricks DBRX Model Offers High Performance at Low Cost
By
ETCentric StaffMarch 21, 2024
Deepgram’s new Aura software turns text into generative audio with a “human-like voice.” The 9-year-old voice recognition company has raised nearly $86 million to date on the strength of its Voice AI platform. Aura is an extremely low-latency text-to-speech voice AI that can be used for voice AI agents, the company says. Paired with Deepgram’s Nova-2 speech-to-text API, developers can use it to “easily (and quickly) exchange real-time information between humans and LLMs to build responsive, high-throughput AI agents and conversational AI applications,” according to Deepgram. Continue reading Deepgram’s Speech Portfolio Now Includes Human-Like Aura
By
ETCentric StaffMarch 20, 2024
Nvidia unveiled what it is calling the world’s most powerful AI processing system, the Blackwell GPU, purpose built to power real-time generative AI on trillion-parameter large language models at what the company says will be up to 25x less cost and energy consumption than its predecessors. Blackwell’s capabilities will usher in what the company promises will be a new era in generative AI computing. News from Nvidia’s GTC 2024 developer conference included the NIM software platform, purpose built to streamline the setup of custom and pre-trained AI models in a production environment, and the DGX SuperPOD server, powered by Blackwell. Continue reading GTC: Nvidia Unveils Blackwell GPU for Trillion-Parameter LLMs