Artificial Intelligence News - Page 24
Samsung establishes dream team of engineers to win AI chip orders from NVIDIA
Samsung has established a new dream team of engineers to secure HBM memory chip deals for AI GPUs from NVIDIA.
The news is coming from South Korean outlet KED Global, which reports Samsung's new task force features about 100 "excellent engineers" who have been working on improving manufacturing yields and quality with the first objective being passing NVIDIA's tests.
NVIDIA CEO Jensen Huang asked Samsung to raise the yields and quality of its 8-layer and 12-layer HBM3E memory chips for supply according to industry insiders on Monday. HBM3E memory is the cornerstone of NVIDIA's next-gen Blackwell B200 AI GPUs, as well as the new beefed-up Hopper H200 AI GPU, each with HBM3E memory mostly from Samsung's South Korean HBM rival: SK hynix.
TSMC to expand advanced packaging capacity at 3 plants: CoWoS, SoIC, SoW for AI GPU demand
TSMC is focused on getting its advanced packaging production capacity to new levels, with the company boosting advanced packaging capacity at its Zhonghe, Nanka, and Jiake fabs, which are "all in the process of expanding production".
The news is coming from Taiwanese media outlet Ctee, reporting that the Chiayi Science Park was finalized this year and is expected to build two advanced packaging plants first. The first phase of Jiake will see the breaking ground ceremony in the coming weeks, and into operation in the second half of the year.
The second phase of Jiake is expected to begin construction in Q2 2024, with operation in Q1 2027 which is still a few years away, which Ctee reports will continue to expand the market share of AI and HPC.
NVIDIA and AMD have reserved all of TSMC's CoWoS and SoIC advanced packaging for 2024 and 2025
AMD and NVIDIA are both pumping out new generations of AI hardware, with the two GPU giants eating all of TSMC's advanced packaging capacity for not just 2024, but also 2025.
In a new report from UDN, the site reports that TSMC will contract advanced packaging production capacity of its in-house CoWoS and SoIC for this and next year, as AI GPU hardware orders aren't slowing down. TSMC expects revenue contributed by server AI processors to more than double this year, and once NVIDIA's new Blackwell B200 AI GPUs are being pumped out later this year, 2025 is going to be a bananas year for TSMC.
In response to the insatiable AI demand, TSMC is now actively expanding its advanced packaging production capacity, with industry estimates that TSMC's CoWoS monthly production capacity will be between 45,000 and 50,000 by the end of 2024. This is a gigantic 3x increase from the 15,000 in 2023, and it's expected to hit 50,000 in 2025.
Nscale becomes first AI GPU cloud provider to use AMD's new Instinct MI300X accelerators
AMD has announced that Nscale is the world's first AMD technology-focused cloud provider, using the company's new Instinct MI300X AI accelerators.
Nscale is a vertically integrated GPU cloud company spun out of Arkon Energy, with a 300MW data center and hosting business in the United States. Nscale is based in Northern Norway, with its N1 Glomfjord site featuring some of the cheapest renewable energy on the planet, making the N1 Glomfjord site one of the most cost-effective LLM and AI training hubs in the world.
Some of the key benefits and features of the Nscale cloud:
Details leak on how Apple is bringing AI is the next iPhone
For quite some time we have been hearing how iOS 18 is expected to be Apple's biggest update ever to iOS, with reports even stating that internally Apple is referring to iOS 18 as "biggest update ever".
The main reason iOS 18 is so monumental is because it will be the time Apple implements artificial intelligence into its operating system. Rumors of this have been circulating for months and well documented by Bloomberg's Mark Gurman, who has reported Apple will be introducing a suite of new AI features that will include a revamped version of Siri, heavy AI integration into Apple's apps such as Apple Music, iWork apps like Keynote and Pages, and more overall customization.
Now, AppleInsider has learned more details about the AI implementation, with "people familiar with the software" telling the publication about the generative AI capabilities of the Ajax Large Language Model (LLM), Apple's generative AI model. The insiders said Ajax will power a feature that enables text summarization, which covers anything from webpages to messages. The feature will quickly read the on-screen text and provide summarized key points on the analyzed text.
Continue reading: Details leak on how Apple is bringing AI is the next iPhone (full post)
X users will now get AI-generated news summaries
X has rolled out an update that enables its Premium users to receive AI-generated summaries of news and topics trending on the platform.
The new feature called "Stories on X" has become available to Premium subscribers, and according to a post from the company's engineering account, the new feature will appear within the Explore tab and is entirely curated by X's GrokAI tool. So, how does it work? Grok has been leveraged to identify the most popular content trending on X, which can be news stories and any general public discussion that has gained a lot of attention.
Grok then digests the popular content and generates a summary. Users exploring the new feature have posted screenshots to X showcasing the design and layout, with some screenshots showing AI-generated summaries for stories such as Apple's earnings report, aid to Ukraine, and "Musk, Experts Debate National Debt," which was a summary of an online discussion between Musk and other prominent X users.
Continue reading: X users will now get AI-generated news summaries (full post)
SK hynix expects HBM memory chip revenues of over $10 billion in 2024
SK hynix has had an absolutely stellar last 12 months riding the ever-growing AI wave, with the South Korean memory giant expecting over $10 billion in revenue from HBM alone by the end of 2024.
The news is coming from South Korean outlet TheElec, which sums up that SK hynix has sold out of its 2024 supply of HBM memory, and is already close to selling out its 2025 supply of HBM. NVIDIA's current H100 uses HBM3, while its new H200 and next-gen B200 AI GPUs both use HBM3E memory, provided by SK hynix.
SK hynix is staying ahead of its HBM competitors in Samsung and Micron, with plans to provide samples of its new 12-stack HBM3E this month, with mass production of the new HBM memory chips expected in Q3 2024 according to SK hynix CEO Kwak Noh-jung at a press conference on Thursday.
Continue reading: SK hynix expects HBM memory chip revenues of over $10 billion in 2024 (full post)
NVIDIA DGX GB200 AI servers expected to sell 40,000 servers in 2025, mass production in 2H 2024
NVIDIA's new DGX GB200 AI server is expected to enter mass production in the second half of this year, with the volume expected to hit 40,000 units in 2025.
NVIDIA and Quanta are the two suppliers of NVL72 and NVL36 cabinets, respectively, with the NVL72 packing 72 GPUs and 36 Grace CPUs. Each of the AI server cabinets cost 96 million NTD (around $3 million USD or so).
The new NVIDIA DGX NVL72 is the AI server with the most computing power, and thus, the highest unit price. Inside, the DGX NVL72 features 72 built-in Blackwell-based B200 AI GPUs and 36 Grace CPUs (18 servers in total with dual Grace CPUs and 36 x B200 AI GPUs per server) with 9 switches. The entire cabinet is designed by NVIDIA in-house, and cannot be modified, it is 100% made, tested, and provided by NVIDIA.
SK hynix says most of its HBM for 2025 is sold out already, 16-Hi HBM4 coming in 2028
SK hynix has announced that almost all of its HBM volume for 2025 has been sold out, as demand for AI GPUs continues to skyrocket.
During a recent press conference, the South Korean memory giant announced plans to invest on its new M15X lab in the Cheongju and Yongin Semiconductor Cluster in Korea with advanced packaging plants in the US.
SK hynix selling out of most its 2025 HBM volume is pretty crazy, as we're not even half way through the year, and NVIDIA's beefed-up H200 AI GPU with HBM3E isn't quite here yet, and its next-gen Blackwell B200 AI GPUs with HBM3E will be launching later this year... yet SK hynix is selling HBM like they're hotcakes.
NVIDIA ChatRTX updated: new models, voice recognition, media search, and more with AI
NVIDIA has just updated its ChatRTX AI chatbot with support for new LLMs, new media search abilities, and speech recognition technology. Check it out:
The latest version of ChatRTX supports more LLMs including Gemma, the latest open, local LLM trained by Google. Gemma was developed by the same research and technology that Google used to create Gemini models, and is built for responsible AI development.
ChatRTX now supports ChatGLM3, an open, bilingual (English and Chinese) LLM based on the general language model framework. The updated version of ChatRTX now lets users interact with image data through Contrastive Language-Image Pre-training from OpenAI. CLIP is a neural network that, as NVIDIA explains, through training and refinement will learn visual concepts from natural lanage supervsion -- a model that recognizes what the AI is "seeing" in image collections.