Private Network Check Readiness - TeckNexus Solutions

Smaller, Faster, Greener: Cutting-Edge AI Models Do More with Less

The demand for electricity and water to power and cool AI servers is ever increasing. Researchers are developing innovative solutions to mitigate the environmental impact. Four promising techniques include model reuse, ReLora, MoE, and quantization. As AI becomes more prevalent, we need to proactively reduce energy and water usage to benefit clients and contribute to a sustainable future.

Overview of AI’s Environmental Impact

As AI continues to advance and expand, the industry is proactively addressing the growing demand for electricity and water to power and cool the servers that make this technology possible. A standard DGX computer, the gold standard for AI work, consumes over 10KW of power. Big Tech will buy millions of these systems this year, using more power than all of New York City, and with this comes a responsibility to find sustainable ways to manage the energy consumption. To mitigate the environmental impact, researchers and engineers are already developing innovative solutions.

The Growing Environmental Challenges in AI Technology


But it is not just the electricity needed to run these computers.  They get hot, really hot, and so they need cooling.  You have to get rid of that heat.  That typically takes up two times more power than the actual computer.  So now that 10KW machine is really using 30KW when running.  These new servers will consume three times more than all of the electricity used in California in 2022! To get around this, server farms are exploring alternative cooling methods, such as using water, to reduce electricity usage. While this shifts the resource burden, it also presents an opportunity to develop more efficient and eco-friendly cooling technologies.

Sustainable Solutions for AI Energy and Water Usage

This saves electricity, but is using our precious fresh water to help cut costs.

Case Studies: Effective AI Sustainability Techniques

AI is hungry for power, and things are going to get worse.  How can we solve this problem?  Fortunately, researchers are already starting to pursue more efficient methods of making and using AI.  Four promising techniques are model reuse, ReLora, MoE, and quantization.

Selecting Technologies for Sustainable AI

Model reuse involves retraining an already trained model for a new purpose, saving time and energy compared to training from scratch. This approach not only conserves resources but also often results in better-performing models.  Both Meta (Facebooks parent) and Mixtral have been good about releasing models that can be reused.

ReLora and Lora reduce the number of calculations needed when retraining models for new uses, further saving energy and enabling the use of smaller, less power-hungry computers. This means that instead of relying on large, energy-intensive systems like NVidia’s DGX, a modest graphics card can often suffice for retraining.

MoE models, such as those recently released by Mistral, have fewer parameters than conventional models, resulting in fewer calculations and reduced energy consumption.

Moreover, MoE models only activate the necessary blocks when in use, much like turning off lights in unused rooms, leading to a 65% reduction in energy usage.

Advantages of Energy-Efficient AI Models

Quantization is an innovative technique that reduces the size of AI models with minimal impact on performance. By quantizing a model, the number of bits required to represent each parameter is reduced. This shrinks the model size, enabling the use of less powerful and more energy-efficient hardware. For instance, a massive 40 billion parameter model would typically require an energy-hungry GPU system like the DGX to run efficiently. But with quantization, this same model can be optimized to run on a low-power consumer GPU, like those found in most laptops. While quantization can slightly reduce model accuracy in some cases, for many practical applications this tradeoff is negligible or unnoticeable. The performance remains excellent while requiring a fraction of the computing resources.

The Impact of Sustainable Practices in AI on Industry

Overall, quantization provides a way to make AI models much more efficient, compact and eco-friendly, minimizing the hardware requirements and energy consumption. This allows state-of-the-art AI to run on ubiquitous consumer devices while maintaining accuracy where it matters most. Quantization represents an important step towards scalable and sustainable AI.

Current Status of Sustainable AI Developments

By combining these four techniques, we have successfully reused a 47 billion parameter MoE model and retrained it for a client using a server that consumes less than 1KW of power, completing the process in just 10 hours. Furthermore, the client can run the model on standard Apple Mac computers with energy-efficient M2 silicon chips. At smartR AI, when developing and training new models, such as our generative AI loyal companion SCOTi® AI, we have been privileged to be able to utilize the super computer at EPCC, Edinburgh University, reducing the time span required for training of models substantially – we trained a model from scratch in nearly one hour.

Timeline of Advances in AI Sustainability

As AI becomes more prevalent, we all need to start thinking more proactively about the energy and water usage.  Research into more efficient training and utilization methods is yielding promising results. But we also need to start using these methods actively; by integrating these new techniques into our tool flows, we not only benefit our clients but also contribute to a more sustainable future for our planet.


Recent Content

Deutsche Telekom is using hardware, pricing, and partnerships to make AI a mainstream feature set across mass-market smartphones and tablets. Deutsche Telekom introduced the T Phone 3 and T Tablet 2, branded as the AI-phone and AI-tablet, with Perplexity as the embedded assistant and a dedicated magenta button for instant access. In Germany, the AI-phone starts at 149 and the AI-tablet at 199, or one euro each when bundled with a tariff, positioning AI features at entry-level price points and shifting value to services and connectivity. The bundle includes an 18-month Perplexity Pro subscription in addition to the embedded assistant, plus three months of Picsart Pro with monthly credits, which lowers the barrier to adopting AI-powered creation and search.
Zayo has secured creditor backing to push major debt maturities to 2030, creating headroom to fund network expansion as AI-driven demand accelerates. Zayo entered into a transaction support agreement dated July 22, 2025, with holders of more than 95% of its term loans, secured notes, and unsecured notes to amend terms and extend maturities to 2030. By extending maturities, Zayo lowers refinancing risk in a higher-for-longer rate environment and preserves cash for growth capex. The move aligns with its pending $4.25 billion acquisition of Crown Castle Fibers assets and follows years of heavy investment in fiber infrastructure.
An unsolicited offer from Perplexity to acquire Googles Chrome raises immediate questions about antitrust remedies, AI distribution, and who controls the internets primary access point. Perplexity has proposed a $34.5 billion cash acquisition of Chrome and says backers are lined up to fund the deal despite the startups significantly smaller balance sheet and an estimated $18 billion valuation in recent fundraising. The bid includes commitments to keep Chromium open source, invest an additional $3 billion in the codebase, and preserve current user defaults including leaving Google as the default search engine. The timing aligns with a U.S. Department of Justice push for structural remedies after a court found Google maintained an illegal search monopoly, with a Chrome divestiture floated as a central remedy.
A new Ciena and Heavy Reading study signals that AI will become a primary source of metro and long-haul traffic within three years while most optical networks remain only partially prepared. AI training and inference are shifting from contained data center domains to distributed, edge-to-core workflows that stress transport capacity, latency, and automation end-to-end. Expectations are even higher for long-haul: 52% see AI surpassing 30% of traffic and 29% expect AI to account for more than half. Yet only 16% of respondents rate their optical networks as very ready for AI workloads, underscoring an execution gap that will shape capex priorities, service roadmaps, and partnership models through 2027.
South Korea’s government and its three national carriers are aligning fresh capital to speed AI and semiconductor competitiveness and to anchor a private-led innovation flywheel. SK Telecom, KT, and LG Uplus will seed a new pool exceeding 300 billion won (about $219 million) via the Korea IT Fund (KIF) to back core and foundational AI, AI transformation (AX), and commercialization in ICT. KIF, formed in 2002 by the carriers, will receive 150 billion won in new commitments, matched by at least an equal amount from external fund managers. The platforms lifespan has been extended to 2040 to sustain long-cycle bets.
NTT DATA and Google Cloud expanded their global partnership to speed the adoption of agentic AI and cloud-native modernization across regulated and dataintensive industries. The push emphasizes sovereign cloud options using Google Distributed Cloud, with both airgapped and connected deployments to meet data residency and regulatory needs without stalling innovation. The partners plan to build industry-specific agentic AI solutions on Google Agent space and Gemini models, underpinned by secure data clean rooms and modernized data platforms. NTT DATA is standing up a dedicated Google Cloud Business Group with thousands of engineers and aims to certify 5,000 practitioners to accelerate delivery, migrations, and managed services.
Whitepaper
Telecom networks are facing unprecedented complexity with 5G, IoT, and cloud services. Traditional service assurance methods are becoming obsolete, making AI-driven, real-time analytics essential for competitive advantage. This independent industry whitepaper explores how DPUs, GPUs, and Generative AI (GenAI) are enabling predictive automation, reducing operational costs, and improving service quality....
Whitepaper
Explore the collaboration between Purdue Research Foundation, Purdue University, Ericsson, and Saab at the Aviation Innovation Hub. Discover how private 5G networks, real-time analytics, and sustainable innovations are shaping the "Airport of the Future" for a smarter, safer, and greener aviation industry....
Article & Insights
This article explores the deployment of 5G NR Transparent Non-Terrestrial Networks (NTNs), detailing the architecture's advantages and challenges. It highlights how this "bent-pipe" NTN approach integrates ground-based gNodeB components with NGSO satellite constellations to expand global connectivity. Key challenges like moving beam management, interference mitigation, and latency are discussed, underscoring...

Download Magazine

With Subscription

Subscribe To Our Newsletter

Private Network Awards 2025 - TeckNexus
Scroll to Top

Private Network Awards

Recognizing excellence in 5G, LTE, CBRS, and connected industries. Nominate your project and gain industry-wide recognition.
Early Bird Deadline: Sept 5, 2025 | Final Deadline: Sept 30, 2025