GPU

Google has unveiled nextโ€‘generation TPU accelerators with up to a 4x performance boost and secured a multiyear Anthropic commitment reportedly worth billions, signaling a new phase in AI infrastructure competition. Google introduced new Tensor Processing Units that deliver roughly four times the performance of prior generations for training and inference of large models. Beyond speed, the design targets better performance-per-watt, a critical lever as AI energy costs surge. Anthropic has secured access to Google Cloud TPU capacity at massive scale, with reports citing availability up to one million TPU chips over the term of the agreement.
Octoberโ€™s job-cut announcements surged, with AI and cost control reshaping staffing plans across technology and adjacent sectors. Planned layoffs spiked to roughly 153,000 in October, up more than 180% from September and about 175% from a year ago, according to the latest Challenger job-cuts tally. Year-to-date announcements for 2025 have crossed 1.09 million, the highest October-through-period since the pandemic shock of 2020 and above comparable 2009 levels. The cuts reflect a pivot from growth-at-any-cost to profitability, with AI rebalancing roles and budgets across the stack. Across reasons given, cost reduction led by a wide margin, and AI adoption was the second-largest driver, underscoring both macro pressure and structural transformation.
Snap and Perplexity are joining forces to embed a conversational AI search experience directly into Snapchatโ€™s chat interface, signaling a new distribution model for AI and a fresh monetization path for social platforms. Perplexity will integrate its AI-powered answer engine natively into Snapchat, beginning a global rollout in early 2026. Under the agreement, Perplexity will pay Snap $400 million over one year, via a mix of cash and equity, as the integration scales. Snap expects revenue contribution from the partnership to begin in 2026. The move is notable as Snapโ€™s first large-scale integration of an external AI partner directly in-app.
At SK AI Summit 2025, CEO Jung Jaihun outlined plans to expand the Ulsan artificial intelligence data center (AIDC) to 1GW-class capacity, stand up a nationwide trio of hubs (Gasan in the Seoul metro, Ulsan in the south, and a new southwest site), and take the model into Southeast Asia starting with Vietnam. The operator is also deepening technology collaborations with Amazon Web Services (AWS) on Edge AI and with NVIDIA on AI-RAN and a Manufacturing AI Cloud; it intends to buy more than 2,000 NVIDIA RTX PRO 6000 Blackwell GPUs and scale Koreaโ€™s largest GPU cluster, Haein, as core compute for industrial AI workloads.
CrowdStrike and NVIDIA are aligning open models, edge inference, and agentic tooling to push real-time, autonomous cyber defense into data centers, clouds, and MEC sites where telecom and enterprise workloads actually live. By pairing CrowdStrikeโ€™s Charlotte AI AgentWorks with NVIDIAโ€™s Nemotron open models, NeMo Data Designer, NeMo Agent Toolkit, and NIM microservices, the partners aim to shrink detection-to-response windows from minutes to milliseconds, and to do so where latency is lowestโ€”at the edge. The companies expanded their collaboration to deliver always-on, continuously learning AI agents that defend cloud, data center, and edge environments using open and enterprise-grade NVIDIA AI components integrated with CrowdStrikeโ€™s Agentic Security Platform.
AI buildouts have flipped a decade of flat U.S. electricity growth into a structural uptrend, with consumer price concerns rising in parallel. After years of steady demand, U.S. load is climbing as commercial and industrial users tap more power, and hyperscale data centers are now a central driver of the shift. Data centers are estimated to consume roughly 4% of U.S. electricity todayโ€”more than twice their share in 2018โ€”and some credible scenarios place that figure in the high single digits to low teens by 2028, depending on the trajectory of AI training and inference footprints.
Samsung and NVIDIA are scaling a 25-year alliance into an AI-driven manufacturing platform that fuses memory, foundry, robotics and networks on a backbone of accelerated computing. Samsung plans to deploy more than 50,000 NVIDIA GPUs to infuse AI across the companyโ€™s manufacturing lifecycleโ€”from chip design and lithography to equipment operations, logistics and quality control. The โ€œAI factoryโ€ is designed as a unified, data-rich fabric where models continuously analyze and optimize processes in real time, shrinking development cycles and improving yield and uptime. The scope goes beyond semiconductors to include mobile devices and robotics, signaling a company-wide digital transformation anchored in accelerated computing.
SoftBank and NVIDIA have validated a fully software-defined, GPU-accelerated AI-RAN that delivers 16-layer massive MU-MIMO outdoorsโ€”an inflection point for vRAN performance, Open RAN scalability, and AI-native RAN design. SoftBankโ€™s AI-RAN product, AITRAS, executed the entire 5G physical layer on NVIDIA GPUs at the Distributed Unit and demonstrated stable 16-layer multi-user MIMO downlink in an outdoor trial at NVIDIAโ€™s Santa Clara campus. The system connected to O-RAN-compliant radios via Split 7.2x and achieved roughly three times the spectral efficiency and throughput of a conventional 4-layer setup while maintaining per-user rates under high load. The field results show that software-only massive MIMO on GPUs can meet macro-radio conditions without bespoke silicon.
A renewed, three-year collaboration between Magic Leap and Google signals a pragmatic path to AI-capable AR glasses that prioritize visual quality, comfort, and manufacturability. Magic Leap is pivoting from building end-user headsets to becoming an ecosystem partner, offering waveguides, optics, device services, and manufacturing know-how to companies pursuing glasses form factors. The companies are aligning around Android XR, positioning the prototype showcased on stage at the Future Investment Initiative in Riyadh as a reference for future designs. The prototype highlights advances in see-through clarity, low-power displays, and an industrial design that approximates everyday eyewear.
Qualcomm is moving from mobile NPUs into rack-scale AI infrastructure, positioning its AI200 (2026) and AI250 (2027) to challenge Nvidia/AMD on the economics of large-scale inference. The company is translating its Hexagon neural processing unit heritageโ€”refined across phones and PCsโ€”into data center accelerators tuned for inferencing, not training. AI200 and AI250 will ship in liquid-cooled, rack-scale configurations designed to operate as a single logical system. Qualcomm is leaning into that constraint with a redesigned memory subsystem and high-capacity cards supporting up to 768 GB of onboard memoryโ€”positioning that as a differentiator versus current GPU offerings.
SoftBank has reportedly approved the final $22.5 billion tranche of a planned $30 billion commitment to OpenAI, tied to the AI firmโ€™s shift to a conventional forโ€‘profit structure and a path to IPO. The investment completes a massive $41 billion financing round for OpenAI that began in April, making it one of the largest private capital raises in tech history. This funding and restructuring signal faster enterprise AI adoption, heavier infrastructure demand, and new platform dynamics that will ripple across networks, cloud, and edge. OpenAI is pushing deeper into enterprise tools, security features, and domainโ€‘specific assistants.
Nokia delivered a stronger-than-expected third quarter, with comparable operating profit reaching โ‚ฌ435 million against consensus of about โ‚ฌ342 million. Group net sales rose 12% to โ‚ฌ4.83 billion, above forecasts, driven by Optical Networks and cloud-related demand tied to AI data centers. The stock jumped double digits intraday and added billions in market value, reflecting newfound confidence after a challenging first half. The recovery now is concentrated in network infrastructure rather than mobile RAN, underscoring where customers are actually spending to handle AI-era traffic patterns. Nokia nudged its full-year operating profit outlook to โ‚ฌ1.7โ€“2.2 billion, with a reporting change related to scaling down passive venture investments partly in play.

TeckNexus Newsletters

I acknowledge and agree to receive TeckNexus communications in line with the T&C and privacy policy.ย 

Tech News & Insight
Enterprises adopting private 5G, LTE, or CBRS networks need more than encryption to stay secure. This article explains the 4 pillars of private network security: core controls, device visibility, real-time threat detection, and orchestration. Learn how to protect SIM and device identities, isolate traffic, secure OT and IoT, and choose...

Sponsored by: OneLayer

     
Whitepaper
Telecom networks are facing unprecedented complexity with 5G, IoT, and cloud services. Traditional service assurance methods are becoming obsolete, making AI-driven, real-time analytics essential for competitive advantage. This independent industry whitepaper explores how DPUs, GPUs, and Generative AI (GenAI) are enabling predictive automation, reducing operational costs, and improving service quality....
Whitepaper
Explore how Generative AI is transforming telecom infrastructure by solving critical industry challenges like massive data management, network optimization, and personalized customer experiences. This whitepaper offers in-depth insights into AI and Gen AI's role in boosting operational efficiency while ensuring security and regulatory compliance. Telecom operators can harness these AI-driven...
Supermicro and Nvidia Logo
Scroll to Top

Feature Your Brand in Private Network Magazines

With Award-Winning Deployments & Industry Leaders
Sponsorship placements open until Nov 10, 2025