By
Douglas ChanJanuary 8, 2025
Nvidia founder and CEO Jensen Huang kicked off CES 2025 with a keynote that was filled with new product announcements and visionary demonstrations of how the company plans to advance the field of AI. The first product that Huang unveiled was the GeForce RTX 50 series of consumer graphics processing units (GPUs). The series is also called RTX Blackwell because it is based on Nvidia’s latest Blackwell microarchitecture design for next generation data center and gaming applications. To showcase RTX Blackwell’s prowess, Huang played an impressively photorealistic video sequence of rich imagery under contrasting light ranges — all rendered in real time. Continue reading CES: Nvidia Unveils New GeForce RTX 50, AI Video Rendering
By
Paula ParisiDecember 13, 2024
Ayar Labs, which develops optical interconnect chips for large-scale AI workloads, has secured $155 million in financing, including from competing processor companies Nvidia, Intel and AMD. Founded in 2017, the Silicon Valley-based company is pursuing a different processing path — combining photonic elements with electronic circuits on each chip for what it says provides faster, more efficient processing for artificial intelligence and high-performance computing. “This brings the company’s total funding to $370 million and raises the company’s valuation to above $1 billion,” Ayar notes, adding that the new funding allows the company to scale its optical I/O tech. Continue reading Nvidia, Intel and AMD Invest in AI Chiplet Developer Ayar Labs
By
Paula ParisiDecember 10, 2024
Meta Platforms has packed more artificial intelligence into a smaller package with Llama 3.3, which the company released last week. The open-source large language model (LLM) “improves core performance at a significantly lower cost, making it even more accessible to the entire open-source community,” Meta VP of Generative AI Ahmad Al-Dahle wrote on X social. The 70 billion parameter text-only Llama 3.3 is said to perform on par with the 405 billion parameter model that was part of Meta’s Llama 3.1 release in July, with less computing power required, significantly lowering its operational costs. Continue reading Meta’s Llama 3.3 Delivers More Processing for Less Compute
By
Paula ParisiDecember 5, 2024
Amazon Web Services is building a supercomputer in collaboration with Anthropic, the AI startup in which the e-commerce giant has an $8 billion minority stake. Hundreds of thousands of AWS’s flagship Trainium chips will be amassed in an “Ultracluster” that when it is completed in 2025 will be one of the largest supercomputers in the world for model training, Amazon says. The company announced the general availability of AWS Trainium2-powered Amazon Elastic Compute Cloud (EC2) virtual servers as well as Trn2 UltraServers designed to train and deploy AI models and teased next-generation Trainium3 chips. Continue reading AWS Building Trainium-Powered Supercomputer with Anthropic
By
Paula ParisiDecember 2, 2024
Lightricks has released an AI model called LTX Video (LTXV) it says generates five seconds of 768 x 512 resolution video (121 frames) in just four seconds, outputting in less time than it takes to watch. The model can run on consumer-grade hardware and is open source, positioning Lightricks as a mass market challenger to firms like Adobe, OpenAI, Google and their proprietary systems. “It’s time for an open-sourced video model that the global academic and developer community can build on and help shape the future of AI video,” Lightricks co-founder and CEO Zeev Farbman said. Continue reading Lightricks LTX Video Model Impresses with Speed and Motion
By
Paula ParisiOctober 28, 2024
OpenAI is taking a new approach to generating media that it says is 50 times faster than the models commonly used today. Called sCM, the approach is a “consistency model,” a variation on the diffusion method used by many leading systems. OpenAI claims its new model is ideal for training for large scale datasets and generating video, audio and images that are of “comparable sample quality to leading diffusion models.” Such models often require hundreds of steps, creating challenges when it comes to real-time applications. OpenAI aims to change this with a faster system that requires less power. Continue reading OpenAI: sCM Generates Media 50x Faster Than Other Models
By
Paula ParisiOctober 21, 2024
Nvidia has debuted a new AI model, Llama-3.1-Nemotron-70B-Instruct, that it claims is outperforming competitors GPT-4o from OpenAI and Anthropic’s Claude 3.5 Sonnet. The impressive showing has prompted speculation of an AI shakeup and a significant shift in Nividia’s AI strategy, which has thus far been focused primarily on chipmaking. The model was quietly released on Hugging Face, and Nvidia says as of October 1 it ranked first on three top automatic alignment benchmarks, “edging out strong frontier models” and vaulting Nvidia to the forefront of the LLM field in areas like comprehension, context and generation. Continue reading Nvidia’s Impressive AI Model Could Compete with Top Brands
By
Paula ParisiOctober 8, 2024
Apple has released a new AI model called Depth Pro that can create a 3D depth map from a 2D image in under a second. The system is being hailed as a breakthrough that could potentially revolutionize how machines perceive depth, with transformative impact on industries from augmented reality to self-driving vehicles. “The predictions are metric, with absolute scale” without relying on the camera metadata typically required for such mapping, according to Apple. Using a consumer-grade GPU, the model can produce a 2.25-megapixel depth map using a single image in only 0.3 seconds. Continue reading Apple Advances Computer Vision with Its Depth Pro AI Model
By
Paula ParisiSeptember 23, 2024
BlackRock has joined forces with Microsoft to launch what will initially be a $30 billion investment fund to finance AI infrastructure — concentrating primarily on building data centers and developing energy projects. The amount could quickly scale to about $100 billion. Abu Dhabi-based tech investment firm MGX is also participating, as is Global Infrastructure Partners (GIP), which owns, operates and invests across energy, transport, digital and waste management. BlackRock announced it is in the process of acquiring GIP, and says a deal expected to close next month. The new fund is called Global AI Infrastructure Investment Partnership (GAIIP). Continue reading BlackRock Teams with Microsoft to Advance AI Infrastructure
By
Paula ParisiSeptember 13, 2024
Sony’s $700 PlayStation 5 Pro promises improved graphics and gameplay. The midcycle upgrade, releasing next month, aims to keep the console competitive with — or better than — ever-evolving game PCs. Sony says the new model supports 8K gaming, an upgrade to the native 4K available with the PS5. For those gamers interested in a disc drive for the PS5 Pro, an $80 model is available. Sony Interactive Entertainment’s popular “Gran Turismo 7” racing simulation game is poised to be the first 8K title for the PS5 Pro. Sony says the GPU on the PS5 Pro has 67 percent more compute and 28 percent faster memory, for rendering that is 45 percent faster than its predecessor. Continue reading New Sony PS5 Pro Supports 8K and Improves 4K Ray Tracing
By
Paula ParisiSeptember 11, 2024
IBM is the first cloud customer for Intel’s Gaudi 3 AI accelerator chip, which it will make available in early 2025. The Gaudi 3 will be available for hybrid and on-site environments via the IBM Cloud, as part of Watsonx AI and on IBM data platforms. Gaudi 3, which began shipping in Q2 and is expected to go into mass production later this year, is IBM’s AI challenger to GPU accelerators from Nvidia and AMD, the latter having in January begun shipping its own HPC solution, the MI300X. Unlike that chip and Nvidia’s Hopper H100 and more recent Blackwell B200, the Gaudi 3 is not a GPU, but built on an architecture specifically for inference and deep learning. Continue reading IBM Cloud Is First to Widely Implement Intel Gaudi 3 AI Chips
By
Paula ParisiAugust 30, 2024
Nvidia has had another impressive quarter. Record revenue of $30 billion in Q2 was up 122 percent from a year ago, while data center revenue of $26.3 billion marked a 154 percent increase from the same period in 2023. The performance was seen by many as an assurance of AI’s staying power, although others raised concern that if the AI companies buying chips do not start generating profits soon, the sugar high of the two-year AI boom could precede a crash. Nvidia took the occasion to tout its next-generation Blackwell chips, reassuring investors that a mid-production “tweak” would not delay release. Continue reading AI Boom Continues to Drive Strong Nvidia Revenue and Profit
By
Paula ParisiAugust 21, 2024
California-based semiconductor manufacturer AMD is looking to take on Nvidia for a bigger share of business from the artificial intelligence boom. AMD plans to purchase data center equipment maker ZT Systems in a cash and stock deal that values the company at $4.9 billion. The deal, which is subject to regulatory approval, is part of AMD’s goal of offering a wider selection of chips, software and system designs to big data enterprise clients such as Microsoft, Google, Meta Platforms and Apple. Privately held ZT Systems, based in New Jersey, makes gear and server solutions for cloud computing and related infrastructure. Continue reading AMD Buying ZT Systems to Expand Data Center Capabilities
By
Paula ParisiJuly 25, 2024
In April, Meta Platforms revealed that it was working on an open-source AI model that performed as well as proprietary models from top AI companies such as OpenAI and Anthropic. Now, Meta CEO Mark Zuckerberg says that model has arrived in the form of Llama 3.1 405B, “the first frontier-level open-source AI model.” The company is also releasing “new and improved” Llama 3.1 70B and 8B models. In addition to general cost and performance benefits, the fact that the Llama 3.1 405B model is open source “will make it the best choice for fine-tuning and distilling smaller models,” according to Meta. Continue reading Meta Calls New Llama the First Open-Source Frontier Model
By
Paula ParisiJune 3, 2024
Big Tech players have joined forces to develop a new industry standard to advance high-speed and low latency communication among data centers by coordinating component development. AMD, Broadcom, Cisco, Google, Hewlett Packard Enterprise (HPE), Intel, Meta Platforms and Microsoft are backing the Ultra Accelerator Link (UALink) promoter group. The group plans to define and establish an open industry standard that will enable AI accelerators to communicate more effectively. The UALink aims to create a pathway for system OEMs, IT professionals and system integrators to connect and scale their AI-connected data centers. Continue reading Big Tech Forms a Group to Develop AI Connectivity Standard