LLM

Apple is reportedly nearing a deal to license Googleโ€™s Gemini for Siri, a move that would reshape assistant architectures and near-term AI roadmaps across devices and networks. Multiple reports indicate Apple is close to licensing a custom version of Googleโ€™s Gemini model, reportedly at a scale of around 1.2 trillion parameters, for roughly $1 billion per year. The model would power a major Siri upgrade while Apple continues building its own foundation models. The objective is clear: boost Siriโ€™s reasoning and task execution in the near term without ceding control over Appleโ€™s system-level integrations or search defaults.
Octoberโ€™s job-cut announcements surged, with AI and cost control reshaping staffing plans across technology and adjacent sectors. Planned layoffs spiked to roughly 153,000 in October, up more than 180% from September and about 175% from a year ago, according to the latest Challenger job-cuts tally. Year-to-date announcements for 2025 have crossed 1.09 million, the highest October-through-period since the pandemic shock of 2020 and above comparable 2009 levels. The cuts reflect a pivot from growth-at-any-cost to profitability, with AI rebalancing roles and budgets across the stack. Across reasons given, cost reduction led by a wide margin, and AI adoption was the second-largest driver, underscoring both macro pressure and structural transformation.
Qualcomm is moving from mobile NPUs into rack-scale AI infrastructure, positioning its AI200 (2026) and AI250 (2027) to challenge Nvidia/AMD on the economics of large-scale inference. The company is translating its Hexagon neural processing unit heritageโ€”refined across phones and PCsโ€”into data center accelerators tuned for inferencing, not training. AI200 and AI250 will ship in liquid-cooled, rack-scale configurations designed to operate as a single logical system. Qualcomm is leaning into that constraint with a redesigned memory subsystem and high-capacity cards supporting up to 768 GB of onboard memoryโ€”positioning that as a differentiator versus current GPU offerings.
The G4 family is built on NVIDIA RTX PRO 6000 Blackwell Server Edition GPUs and targets high-throughput inference, visual computing, and simulation. Each VM can be configured with 1, 2, 4, or 8 GPUs, delivering up to 768 GB of GDDR7 memory in total. Fifth-generation Tensor Cores introduce FP4 precision to drive efficient multimodal and LLM inference, while fourth-generation RT Cores double real-time ray-tracing performance over the prior generation for photorealistic rendering. Google cites up to 9x throughput over G2 instances, positioning G4 as a universal GPU platform spanning AI inference, content creation, CAD/CAE acceleration, and robotics simulation.
OpenAI has launched ChatGPT Atlas, a MacOS AI browser built around its chatbot, positioning agentic browsing and LLM-native search as the next front in the browser wars. Atlas reframes the browser as a conversational interface. It removes the traditional address bar and orients the experience around ChatGPT, with natural language as the primary way to navigate, retrieve, and summarize information. The initial release targets Appleโ€™s MacOS, with OpenAI emphasizing a paid โ€œagent modeโ€ that can autonomously search, read, and act on the userโ€™s behalf using the live browsing context. Agent mode will be available to paying ChatGPT subscribers, extending OpenAIโ€™s monetization beyond API usage and premium chatbot tiers.
Arm and Meta have inked a multi-year partnership to scale AI efficiency from hyperscale data centers to on-device inference, aligning Armโ€™s performance-per-watt strengths with Metaโ€™s AI software and infrastructure stack. Meta plans to run its ranking and recommendation workloads on Arm Neoverse-based data center platforms as part of an ongoing infrastructure expansion. The companies are co-optimizing AI software componentsโ€”spanning compilers, libraries, and frameworks like PyTorch, FBGEMM, vLLM, and the ExecuTorch runtimeโ€”so models can execute more efficiently on Arm CPUs in the cloud and on Arm-based devices at the edge. The work includes leveraging Armโ€™s KleidiAI optimizations to improve inference throughput and energy efficiency, with code contributions flowing back to open source.
ChatGPT users in the U.S. can now buy from Etsy sellers without leaving the conversation, with more than a million Shopify merchants โ€œcoming soon.โ€ The feature, called Instant Checkout, is available to logged-in Free, Plus, and Pro users. It supports Apple Pay, Google Pay, Stripe, and credit cards. The flow is simple: ask for ideas, get curated products with images, prices, and reviews, tap Buy, confirm shipping and payment, and the merchant fulfills the order using its existing systems. Brands like Glossier, Skims, Spanx, and Vuori are expected to be part of the broader Shopify rollout.
Hitachi has launched a global AI Factory built on NVIDIAโ€™s reference architecture to speed the development and deployment of โ€œphysical AIโ€ spanning mobility, energy, industrial, and technology domains. Hitachi is standardizing a centralized yet globally distributed AI infrastructure on NVIDIAโ€™s full-stack platform, pairing Hitachi iQ systems with NVIDIA HGX B200 platforms powered by Blackwell GPUs, Hitachi iQ M Series with NVIDIA RTX 6000 Server Edition GPUs, and the NVIDIA Spectrum-X Ethernet AI networking platform. The environment is designed to run production AI with NVIDIA AI Enterprise and support simulation and physically accurate digital twins using NVIDIA Omniverse libraries.
HUMAIN, a Saudi PIF-backed AI company, introduced Horizon Pro, an โ€œagentic AIโ€ PC built on Qualcommโ€™s Snapdragon X Elite, positioning it as a new class of Windows laptop where on-device AI drives workflows, decisions, and user interaction. At Qualcommโ€™s Snapdragon Summit in Maui, HUMAIN CEO Tareq Amin unveiled the Horizon Pro PC and the companyโ€™s agentic software layer, Humain One, which runs on top of Windows 11 and is slated for formal launch at the Future Investment Initiative in Riyadh.
Fujitsuโ€™s latest generative AI breakthrough compresses large language models by 94% using 1-bit quantization, tripling inference speed and retaining 89% accuracy. Combined with brain-inspired knowledge distillation, these enhancements power high-speed, low-cost AI for CRM, image recognition, and edge deployments, all while reducing energy demands.
Fresh polling signals rising public concern that AI could upend employment, destabilize politics, and strain social and energy systems. A recent Reuters/Ipsos survey of 4,446 U.S. adults found that 71% worry AI will permanently displace too many workers. Seventy-seven percent of respondents fear AI will fuel political instability if hostile actors exploit the technology. The poll also shows broad worry about AIs indirect costs: 66% are concerned about AI companions displacing human relationships, and 61% are concerned about the technology’s energy footprint. Bottom line: Public concern is high, and that increases the cost of missteps.
Orange has expanded its partnership with OpenAI to localize AI models for underrepresented African languages like Wolof and Pulaar. These models will run on Orangeโ€™s secure, sovereign infrastructure, ensuring privacy and regulatory compliance. With applications in health, education, and digital equity, Orangeโ€™s Responsible AI strategy aims to make generative AI more accessible for Africaโ€™s rural populations and especially for women, who face digital and language-based barriers.

TeckNexus Newsletters

I acknowledge and agree to receive TeckNexus communications in line with the T&C and privacy policy.ย 

Tech News & Insight
Enterprises adopting private 5G, LTE, or CBRS networks need more than encryption to stay secure. This article explains the 4 pillars of private network security: core controls, device visibility, real-time threat detection, and orchestration. Learn how to protect SIM and device identities, isolate traffic, secure OT and IoT, and choose...

Sponsored by: OneLayer

     
Whitepaper
Telecom networks are facing unprecedented complexity with 5G, IoT, and cloud services. Traditional service assurance methods are becoming obsolete, making AI-driven, real-time analytics essential for competitive advantage. This independent industry whitepaper explores how DPUs, GPUs, and Generative AI (GenAI) are enabling predictive automation, reducing operational costs, and improving service quality....
Whitepaper
Explore how Generative AI is transforming telecom infrastructure by solving critical industry challenges like massive data management, network optimization, and personalized customer experiences. This whitepaper offers in-depth insights into AI and Gen AI's role in boosting operational efficiency while ensuring security and regulatory compliance. Telecom operators can harness these AI-driven...
Supermicro and Nvidia Logo
Scroll to Top

Feature Your Brand in Private Network Magazines

With Award-Winning Deployments & Industry Leaders
Sponsorship placements open until Nov 10, 2025