{"id":6698,"date":"2026-04-25T05:40:21","date_gmt":"2026-04-25T05:40:21","guid":{"rendered":"https:\/\/scipapermill.com\/index.php\/2026\/04\/25\/energy-efficiency-in-ai-ml-from-edge-to-cloud-photonics-to-spikes\/"},"modified":"2026-04-25T05:40:21","modified_gmt":"2026-04-25T05:40:21","slug":"energy-efficiency-in-ai-ml-from-edge-to-cloud-photonics-to-spikes","status":"publish","type":"post","link":"https:\/\/scipapermill.com\/index.php\/2026\/04\/25\/energy-efficiency-in-ai-ml-from-edge-to-cloud-photonics-to-spikes\/","title":{"rendered":"Energy Efficiency in AI\/ML: From Edge to Cloud, Photonics to Spikes"},"content":{"rendered":"<h3>Latest 37 papers on energy efficiency: Apr. 25, 2026<\/h3>\n<p>The relentless march of AI\/ML, especially with the advent of massive models like Large Language Models (LLMs), has brought the critical issue of energy consumption to the forefront. Training and deploying these models demand colossal computational resources, leading to significant energy footprints and environmental concerns. This blog post dives into recent breakthroughs from a collection of research papers that are carving new paths toward a more energy-efficient AI future, spanning innovations from novel hardware architectures to intelligent software strategies and network optimizations.<\/p>\n<h3 id=\"the-big-ideas-core-innovations\">The Big Idea(s) &amp; Core Innovations<\/h3>\n<p>Researchers are tackling the energy challenge from multiple angles, often leveraging the inherent characteristics of AI workloads to design more efficient systems. A prominent theme is the <strong>repurposing and optimization of hardware for specific AI tasks<\/strong>. For instance, <a href=\"https:\/\/arxiv.org\/pdf\/2604.17808\">\u201cEnabling AI ASICs for Zero Knowledge Proof\u201d<\/a> by <strong>Georgia Institute of Technology, MIT, and Google<\/strong> introduces MORPH, which intelligently remaps high-precision Zero-Knowledge Proof (ZKP) kernels onto AI ASICs like TPUs. By transforming complex modular arithmetic into dense low-precision matrix operations and using layout-stationary dataflows, MORPH achieves up to 10\u00d7 higher throughput for NTT and comparable MSM throughput compared to GPUs, effectively making AI ASICs viable ZKP accelerators.<\/p>\n<p>Another significant area is <strong>photonic and neuromorphic computing<\/strong>, which promises radical shifts in energy efficiency by moving beyond traditional electron-based computation. The paper <a href=\"https:\/\/arxiv.org\/pdf\/2604.18496\">\u201cTensor Processing with Homodyne Photonic Integrated Circuits exceeds 1,000 TOPS\u201d<\/a> from <strong>Opticore Inc.\u00a0and University of California, Berkeley<\/strong> showcases a groundbreaking homodyne photonic integrated circuit for general matrix multiplication (GEMM) achieving over 1,000 TOPS throughput and 330 TOPS\/W energy efficiency. This is achieved through massive on-chip optical fanout, parallelism, and time multiplexing, demonstrating near-digital accuracy on LLMs like Qwen2.5. Complementing this, <a href=\"https:\/\/arxiv.org\/pdf\/2604.12278\">\u201cLightMat-HP: A Photonic-Electronic System for Accelerating General Matrix Multiplication With Configurable Precision\u201d<\/a> by researchers from <strong>Australian National University and UNSW<\/strong> develops a hybrid photonic-electronic system using block floating-point (BFP) arithmetic and slicing-based photonic multiplication. This elegantly overcomes photonic precision limitations by decomposing high-bit-width operations into multiple low-bit sub-multiplications, significantly improving energy efficiency and latency over conventional systems. The push for brain-inspired computing is also evident: <a href=\"https:\/\/arxiv.org\/pdf\/2604.18610\">\u201cSpikeMLLM: Spike-based Multimodal Large Language Models via Modality-Specific Temporal Scales and Temporal Compression\u201d<\/a> by authors from <strong>Institute of Automation, Chinese Academy of Sciences<\/strong> introduces SpikeMLLM, the first spike-based framework for multimodal LLMs. It achieves 25.8x power reduction and 9.06x throughput on an RTL accelerator by employing Modality-Specific Temporal Scales and Temporally Compressed LIF mechanisms, proving that spiking neural networks can handle complex multimodal tasks with near-FP16 performance. Similarly, <a href=\"https:\/\/arxiv.org\/pdf\/2604.12365\">\u201cAdaptive Spiking Neurons for Vision and Language Modeling\u201d<\/a> from <strong>Peking University<\/strong> proposes ASN\/NASN, which achieves adaptive firing dynamics with efficient integer training and spike inference, reducing energy consumption by 93% compared to Qwen-1.5B while maintaining accuracy on QA tasks.<\/p>\n<p><strong>Optimizing memory and dataflow<\/strong> is another crucial strategy. <a href=\"https:\/\/arxiv.org\/pdf\/2604.16007\">\u201cMemExplorer: Navigating the Heterogeneous Memory Design Space for Agentic Inference NPUs\u201d<\/a> by <strong>University of Cambridge, Imperial College London, and Microsoft<\/strong> introduces a memory system synthesizer for heterogeneous NPU systems. It finds that prefill benefits from large, high-bandwidth 3D-stacked SRAM, while decode prefers larger-capacity, lower-bandwidth LPDDR, achieving up to 2.3\u00d7 higher energy efficiency than baseline NPUs. <a href=\"https:\/\/arxiv.org\/pdf\/2604.17692\">\u201cAccelCIM: Systematic Dataflow Exploration for SRAM Compute-in-Memory Accelerator\u201d<\/a> from <strong>Peking University<\/strong> systematically explores dataflows for SRAM-based compute-in-memory (CIM) accelerators, revealing that systolic interconnects are more area-efficient and medium-sized macros offer the best energy-area trade-off. Furthermore, <a href=\"https:\/\/arxiv.org\/pdf\/2604.13969\">\u201cGEM3D-CIM: General Purpose Matrix Computation Using 3D-Integrated SRAM-eDRAM Hybrid Compute-In-Memory-on-Memory Architecture\u201d<\/a> by <strong>University of Wisconsin Madison<\/strong> demonstrates a 3D-integrated SRAM-eDRAM hybrid CIM architecture capable of general matrix operations (transpose, element-wise multiplication) directly in memory, reaching 436.61 GOPS\/W energy efficiency.<\/p>\n<p>For <strong>edge and IoT devices<\/strong>, specific innovations are tailored to resource constraints. <a href=\"https:\/\/arxiv.org\/pdf\/2604.19293\">\u201cEnergy Efficient LSTM Accelerators for Embedded FPGAs through Parameterised Architecture Design\u201d<\/a> by <strong>University of Duisburg-Essen<\/strong> presents an LSTM accelerator for embedded FPGAs that achieves 59.19% better energy efficiency by using 8-bit quantization and replacing complex activation functions. Another work from <strong>University of Duisburg-Essen<\/strong>, <a href=\"https:\/\/arxiv.org\/pdf\/2407.12027\">\u201cIdle is the New Sleep: Configuration-Aware Alternative to Powering Off FPGA-Based DL Accelerators During Inactivity\u201d<\/a>, proposes an \u201cIdle-Waiting\u201d strategy for FPGA-based DL accelerators, reducing configuration energy by 40.13-fold and extending system lifetime by 12.39\u00d7 for intermittent IoT workloads. In a similar vein, <a href=\"https:\/\/arxiv.org\/pdf\/2407.05102\">\u201cTowards Auto-Building of Embedded FPGA-based Soft Sensors for Wastewater Flow Estimation\u201d<\/a> by the same group optimizes hardware for soft sensors on smaller FPGAs, achieving micro-watt static power consumption. For wireless communication, <a href=\"https:\/\/arxiv.org\/pdf\/2604.21812\">\u201cGeneralized Two-Dimensional Index Modulation in the Code\u2013Spatial Domain for LPWAN\u201d<\/a> by <strong>Sun Yat-sen University and Techphant Technologies<\/strong> introduces a generalized code-index modulation (CIM) transceiver for LPWANs, saving up to 57% energy by implicitly transmitting index bits through antenna and spreading sequence selection. Benchmarking <a href=\"https:\/\/arxiv.org\/pdf\/2604.19391\">\u201cOn the Practical Performance of Noise Modulation for Ultra-Low-Power IoT\u201d<\/a> from <strong>National Institute of Telecommunications (Inatel)<\/strong> identifies critical \u201cenergy crossover distances\u201d where simpler non-coherent NoiseMod becomes less efficient than coherent schemes, providing crucial design insights for ultra-low-power IoT.<\/p>\n<p>Lastly, <strong>software and system-level optimizations<\/strong> prove vital. <a href=\"https:\/\/arxiv.org\/pdf\/2604.19377\">\u201cTowards Energy Impact on AI-Powered 6G IoT Networks: Centralized vs.\u00a0Decentralized\u201d<\/a> from <strong>University of Kaiserslautern-Landau and DFKI<\/strong> demonstrates that decentralized (federated) learning can reduce overall energy consumption by up to 70% in 6G IoT networks, achieving comparable predictive accuracy to centralized methods. <a href=\"https:\/\/arxiv.org\/pdf\/2604.19342\">\u201cAre Large Language Models Economically Viable for Industry Deployment?\u201d<\/a> by researchers from <strong>DSEU-Okhla, Macquarie University, and Stanford University<\/strong> highlights that compact models (&lt; 2B parameters) are the efficiency frontier for LLM deployment on legacy hardware, showing that QLoRA can surprisingly increase adaptation energy by 7x despite memory footprint reduction. <a href=\"https:\/\/arxiv.org\/pdf\/2604.18788\">\u201cEfficient Mixture-of-Experts LLM Inference with Apple Silicon NPUs\u201d<\/a> from <strong>University of Virginia<\/strong> introduces NPUMoE, a runtime inference engine for MoE LLMs on Apple Neural Engine, achieving 1.81x\u20137.37x energy efficiency by offloading dense computations to the NPU while using CPU\/GPU for dynamic routing. <a href=\"https:\/\/arxiv.org\/pdf\/2604.14626\">\u201cELMoE-3D: Leveraging Intrinsic Elasticity of MoE for Hybrid-Bonding-Enabled Self-Speculative Decoding in On-Premises Serving\u201d<\/a> by <strong>KAIST and Samsung Electronics<\/strong> proposes ELMoE-3D, a hybrid-bonding framework that unifies cache-based acceleration and speculative decoding for MoE LLMs, achieving 4.4x energy efficiency gains. For High-Performance Computing (HPC), <a href=\"https:\/\/arxiv.org\/pdf\/2604.17640\">\u201cTowards Energy Efficient Co-Scheduling in HPC\u201d<\/a> from <strong>University of Illinois Chicago and Argonne National Laboratory<\/strong> presents EcoSched, an online energy-aware co-scheduler for multi-GPU HPC systems that achieves up to 14.8% energy savings by jointly optimizing GPU-count selection and application co-scheduling.<\/p>\n<h3 id=\"under-the-hood-models-datasets-benchmarks\">Under the Hood: Models, Datasets, &amp; Benchmarks<\/h3>\n<p>These advancements are built upon and validated through a rich ecosystem of specialized models, datasets, and benchmarks:<\/p>\n<ul>\n<li><strong>Hardware\/System Benchmarks &amp; Tools:<\/strong>\n<ul>\n<li><strong>EDGE-EVAL:<\/strong> A new lifecycle benchmarking framework for LLMs, evaluating economic viability, intelligence-per-watt, system density, cold-start tax, and quantization fidelity. (<a href=\"https:\/\/arxiv.org\/pdf\/2604.19342\">\u201cAre Large Language Models Economically Viable for Industry Deployment?\u201d<\/a> &#8211; Code: <a href=\"https:\/\/github.com\/Abdullah4152\/EDGE-EVAL\">https:\/\/github.com\/Abdullah4152\/EDGE-EVAL<\/a>)<\/li>\n<li><strong>DEEP-GAP:<\/strong> Benchmarking methodology for GPU inference performance, comparing NVIDIA T4 and L4 GPUs across precision modes. (<a href=\"https:\/\/arxiv.org\/pdf\/2604.14552\">\u201cDEEP-GAP: Deep-learning Evaluation of Execution Parallelism in GPU Architectural Performance\u201d<\/a> &#8211; Code: <a href=\"https:\/\/github.com\/kathiravan-palaniappan\/DEEP-GAP\">https:\/\/github.com\/kathiravan-palaniappan\/DEEP-GAP<\/a>)<\/li>\n<li><strong>ElasticAI-Creator:<\/strong> An automated toolchain for deploying deep learning models on embedded FPGAs. (<a href=\"https:\/\/arxiv.org\/pdf\/2604.19293\">\u201cEnergy Efficient LSTM Accelerators for Embedded FPGAs through Parameterised Architecture Design\u201d<\/a> and <a href=\"https:\/\/arxiv.org\/pdf\/2407.05102\">\u201cTowards Auto-Building of Embedded FPGA-based Soft Sensors for Wastewater Flow Estimation\u201d<\/a> &#8211; Code: <a href=\"https:\/\/github.com\/es-ude\/elastic-ai.creator\">https:\/\/github.com\/es-ude\/elastic-ai.creator<\/a>)<\/li>\n<li><strong>CodeCarbon:<\/strong> Library for energy consumption measurement in AI training. (<a href=\"https:\/\/arxiv.org\/pdf\/2604.19377\">\u201cTowards Energy Impact on AI-Powered 6G IoT Networks: Centralized vs.\u00a0Decentralized\u201d<\/a>)<\/li>\n<li><strong>vLLM:<\/strong> For LLM inference serving. (<a href=\"https:\/\/arxiv.org\/pdf\/2604.19342\">\u201cAre Large Language Models Economically Viable for Industry Deployment?\u201d<\/a>)<\/li>\n<li><strong>Pynvml:<\/strong> For GPU power telemetry. (<a href=\"https:\/\/arxiv.org\/pdf\/2604.19342\">\u201cAre Large Language Models Economically Viable for Industry Deployment?\u201d<\/a>)<\/li>\n<\/ul>\n<\/li>\n<li><strong>Specific Models &amp; Frameworks:<\/strong>\n<ul>\n<li><strong>LLaMA, Qwen, Phi-3.5-MoE-Instruct:<\/strong> Key LLMs used for benchmarking deployment viability and NPU inference. (<a href=\"https:\/\/arxiv.org\/pdf\/2604.19342\">\u201cAre Large Language Models Economically Viable for Industry Deployment?\u201d<\/a>, <a href=\"https:\/\/arxiv.org\/pdf\/2604.18788\">\u201cEfficient Mixture-of-Experts LLM Inference with Apple Silicon NPUs\u201d<\/a>, <a href=\"https:\/\/arxiv.org\/pdf\/2604.14626\">\u201cELMoE-3D: Leveraging Intrinsic Elasticity of MoE for Hybrid-Bonding-Enabled Self-Speculative Decoding in On-Premises Serving\u201d<\/a>, <a href=\"https:\/\/arxiv.org\/pdf\/2604.18496\">\u201cTensor Processing with Homodyne Photonic Integrated Circuits exceeds 1,000 TOPS\u201d<\/a>)<\/li>\n<li><strong>SpikingJelly:<\/strong> Open-source platform for spike-based intelligence. (<a href=\"https:\/\/arxiv.org\/pdf\/2604.15769\">\u201cClosing the Theory-Practice Gap in Spiking Transformers via Effective Dimension\u201d<\/a>)<\/li>\n<li><strong>CRONet:<\/strong> A hybrid CNN-RNN for topology optimization, accelerated on AMD Versal AIE-ML. (<a href=\"https:\/\/arxiv.org\/pdf\/2604.14700\">\u201cAccelerating CRONet on AMD Versal AIE-ML Engines\u201d<\/a>)<\/li>\n<li><strong>NPUMoE:<\/strong> Inference engine for MoE LLMs on Apple Neural Engine. (<a href=\"https:\/\/arxiv.org\/pdf\/2604.18788\">\u201cEfficient Mixture-of-Experts LLM Inference with Apple Silicon NPUs\u201d<\/a>)<\/li>\n<li><strong>MORPH:<\/strong> Framework for ZKP acceleration on AI ASICs. (<a href=\"https:\/\/arxiv.org\/pdf\/2604.17808\">\u201cEnabling AI ASICs for Zero Knowledge Proof\u201d<\/a> &#8211; Code: <a href=\"https:\/\/github.com\/EfficientPPML\/MORPH\">https:\/\/github.com\/EfficientPPML\/MORPH<\/a>)<\/li>\n<li><strong>Overmind NSA:<\/strong> Unified neuro-symbolic AI architecture using Pad\u00e9 approximations. (<a href=\"https:\/\/arxiv.org\/pdf\/2604.15623\">\u201cOvermind NSA: A Unified Neuro-Symbolic Computing Architecture with Approximate Nonlinear Activations and Preemptive Memory Bypass\u201d<\/a>)<\/li>\n<li><strong>SpikeMLLM:<\/strong> First spike-based framework for multimodal LLMs. (<a href=\"https:\/\/arxiv.org\/pdf\/2604.18610\">\u201cSpikeMLLM: Spike-based Multimodal Large Language Models via Modality-Specific Temporal Scales and Temporal Compression\u201d<\/a>)<\/li>\n<li><strong>WirelessAgent:<\/strong> AI-enabled agent for OFDMA wireless resource allocation. (<a href=\"https:\/\/arxiv.org\/pdf\/2604.17440\">\u201cWirelessAgent: A Unified Agent Design for General Wireless Resource Allocation Problem without Current Channel State Information\u201d<\/a>)<\/li>\n<li><strong>IMU-to-4D:<\/strong> Framework repurposing LLMs for non-visual spatiotemporal understanding from IMU sensors. (<a href=\"https:\/\/arxiv.org\/pdf\/2604.21926\">\u201cSeeing Without Eyes: 4D Human-Scene Understanding from Wearable IMUs\u201d<\/a> &#8211; Project page: <a href=\"https:\/\/tianhang-cheng.github.io\/IMU4D\/\">https:\/\/tianhang-cheng.github.io\/IMU4D\/<\/a>)<\/li>\n<li><strong>EcoSched:<\/strong> Online energy-aware co-scheduler for multi-GPU HPC systems. (<a href=\"https:\/\/arxiv.org\/pdf\/2604.17640\">\u201cTowards Energy Efficient Co-Scheduling in HPC\u201d<\/a>)<\/li>\n<li><strong>R2NN (Resonant Recurrent Neural Network):<\/strong> Fully analog neural network via metacircuit. (<a href=\"https:\/\/arxiv.org\/pdf\/2604.17277\">\u201cFully Analog Resonant Recurrent Neural Network via Metacircuit\u201d<\/a>)<\/li>\n<li><strong>ImageHD:<\/strong> FPGA accelerator for on-device continual learning using hyperdimensional computing. (<a href=\"https:\/\/arxiv.org\/pdf\/2604.21280\">\u201cImageHD: Energy-Efficient On-Device Continual Learning of Visual Representations via Hyperdimensional Computing\u201d<\/a>)<\/li>\n<\/ul>\n<\/li>\n<li><strong>Datasets &amp; Environments:<\/strong>\n<ul>\n<li><strong>MotionMillion, LINGO, HUMOTO, DIP-IMU, IMUPoser, HumanML3D, ParaHome, AMASS:<\/strong> Datasets for 4D human-scene understanding. (<a href=\"https:\/\/arxiv.org\/pdf\/2604.21926\">\u201cSeeing Without Eyes: 4D Human-Scene Understanding from Wearable IMUs\u201d<\/a>)<\/li>\n<li><strong>CORe50, CIFAR-10, CIFAR-100, ImageNet:<\/strong> Standard visual recognition benchmarks. (<a href=\"https:\/\/arxiv.org\/pdf\/2604.21280\">\u201cImageHD: Energy-Efficient On-Device Continual Learning of Visual Representations via Hyperdimensional Computing\u201d<\/a>, <a href=\"https:\/\/arxiv.org\/pdf\/2604.12365\">\u201cAdaptive Spiking Neurons for Vision and Language Modeling\u201d<\/a>, <a href=\"https:\/\/arxiv.org\/pdf\/2604.12420\">\u201cHARP: Hadamard-Domain Write-and-Verify for Noise-Robust RRAM Programming\u201d<\/a>)<\/li>\n<li><strong>GLUE, XSum, SQuAD, UltraChat, MT-Bench, GSM8K, Alpaca, HumanEval:<\/strong> Language understanding and generation benchmarks. (<a href=\"https:\/\/arxiv.org\/pdf\/2604.19342\">\u201cAre Large Language Models Economically Viable for Industry Deployment?\u201d<\/a>, <a href=\"https:\/\/arxiv.org\/pdf\/2604.12365\">\u201cAdaptive Spiking Neurons for Vision and Language Modeling\u201d<\/a>, <a href=\"https:\/\/arxiv.org\/pdf\/2604.14626\">\u201cELMoE-3D: Leveraging Intrinsic Elasticity of MoE for Hybrid-Bonding-Enabled Self-Speculative Decoding in On-Premises Serving\u201d<\/a>)<\/li>\n<li><strong>UMAR unit (NEST building, Empa campus):<\/strong> Real-world living lab data for building energy control. (<a href=\"https:\/\/arxiv.org\/pdf\/2604.16033\">\u201cSafe Deep Reinforcement Learning for Building Heating Control and Demand-side Flexibility\u201d<\/a>)<\/li>\n<li><strong>SLKI project (German railway infrastructure):<\/strong> Real-world testbed for 6G IoT energy impact. (<a href=\"https:\/\/arxiv.org\/pdf\/2604.19377\">\u201cTowards Energy Impact on AI-Powered 6G IoT Networks: Centralized vs.\u00a0Decentralized\u201d<\/a>)<\/li>\n<li><strong>Madrid Grid urban scenario (3D-Ray Tracing):<\/strong> Realistic radio channel models for Massive MIMO. (<a href=\"https:\/\/arxiv.org\/pdf\/2604.15977\">\u201cImpact of Nonlinear Power Amplifier on Massive MIMO: Machine Learning Prediction Under Realistic Radio Channel\u201d<\/a>)<\/li>\n<li><strong>OGBN-Products, GRCh38 human genome:<\/strong> Graph datasets for genomic and graph-based dynamic programming. (<a href=\"https:\/\/arxiv.org\/pdf\/2604.15361\">\u201cGEN-Graph: Heterogeneous PIM Accelerator for General Computational Patterns in Graph-based Dynamic Programming\u201d<\/a>)<\/li>\n<\/ul>\n<\/li>\n<\/ul>\n<h3 id=\"impact-the-road-ahead\">Impact &amp; The Road Ahead<\/h3>\n<p>These advancements herald a future where AI\/ML is not only powerful but also sustainable. The shift towards <strong>neuromorphic and photonic computing<\/strong> promises orders-of-magnitude improvements in energy efficiency, potentially enabling AI directly within sensors and on-device environments, as shown by SpikeMLLM and LightMat-HP. The emphasis on <strong>hardware-software co-design<\/strong> is critical, ensuring that algorithms are optimized for the underlying physics of new compute substrates, as demonstrated by MORPH for ZKP on AI ASICs and AccelCIM for CIM dataflows.<\/p>\n<p>For <strong>edge AI and IoT<\/strong>, the focus on tailored hardware (FPGA accelerators, specialized LPWAN transceivers), intelligent idle management, and decentralized learning paradigms (federated learning) will unlock unprecedented capabilities for smart cities, industrial automation, and environmental monitoring, making ubiquitous AI a reality without the prohibitive energy cost. The ability to reconstruct 4D human-scene understanding from IMU sensors alone, as presented by <a href=\"https:\/\/arxiv.org\/pdf\/2604.21926\">\u201cSeeing Without Eyes: 4D Human-Scene Understanding from Wearable IMUs\u201d<\/a>, opens doors for privacy-preserving and ultra-low-power context awareness in daily life.<\/p>\n<p>In <strong>cloud and data center environments<\/strong>, optimizations like heterogeneous memory architectures (MemExplorer), efficient MoE inference (NPUMoE, ELMoE-3D), and energy-aware co-scheduling (EcoSched) are crucial for managing the growing demand for large model inference and HPC workloads. The understanding that <strong>economic viability and energy efficiency are intertwined<\/strong> (as highlighted by EDGE-EVAL) will drive the adoption of more compact and specialized models, shifting the paradigm from \u201cbigger is always better\u201d to \u201csmarter and leaner is more sustainable.\u201d<\/p>\n<p>Looking ahead, the integration of AI for <strong>optimizing complex systems like integrated energy grids<\/strong> (e.g., <a href=\"https:\/\/arxiv.org\/pdf\/2604.14184\">\u201cEnd-to-End Learning-based Operation of Integrated Energy Systems for Buildings and Data Centers\u201d<\/a>) and <strong>6G wireless networks<\/strong> (e.g., <a href=\"https:\/\/arxiv.org\/pdf\/2604.14309\">\u201cAerial Multi-Functional RIS in Fluid Antennas-Aided Full-Duplex Networks: A Self-Optimized Hybrid Deep Reinforcement Learning Approach\u201d<\/a>, <a href=\"https:\/\/arxiv.org\/pdf\/2604.21587\">\u201cGenerative Learning Enhanced Intelligent Resource Management for Cell-Free Delay Deterministic Communications\u201d<\/a>, and <a href=\"https:\/\/arxiv.org\/pdf\/2604.17440\">\u201cWirelessAgent: A Unified Agent Design for General Wireless Resource Allocation Problem without Current Channel State Information\u201d<\/a>) will be paramount. These papers collectively paint a picture of an AI future that is not only powerful and intelligent but also profoundly energy-conscious and sustainable, a critical step towards mitigating AI\u2019s environmental impact and fostering its broader adoption.<\/p>\n","protected":false},"excerpt":{"rendered":"<p>Latest 37 papers on energy efficiency: Apr. 25, 2026<\/p>\n","protected":false},"author":1,"featured_media":0,"comment_status":"open","ping_status":"open","sticky":false,"template":"","format":"standard","meta":{"_yoast_wpseo_focuskw":"","_yoast_wpseo_title":"","_yoast_wpseo_metadesc":"","_jetpack_memberships_contains_paid_content":false,"footnotes":"","jetpack_publicize_message":"","jetpack_publicize_feature_enabled":true,"jetpack_social_post_already_shared":true,"jetpack_social_options":{"image_generator_settings":{"template":"highway","default_image_id":0,"font":"","enabled":false},"version":2}},"categories":[56,330,63],"tags":[1552,4027,4111,180,1564,3239],"class_list":["post-6698","post","type-post","status-publish","format-standard","hentry","category-artificial-intelligence","category-hardware-architecture","category-machine-learning","tag-6g-networks","tag-compute-in-memory","tag-design-space-exploration","tag-energy-efficiency","tag-main_tag_energy_efficiency","tag-hardware-software-co-design"],"yoast_head":"<!-- This site is optimized with the Yoast SEO plugin v27.4 - https:\/\/yoast.com\/product\/yoast-seo-wordpress\/ -->\n<title>Energy Efficiency in AI\/ML: From Edge to Cloud, Photonics to Spikes<\/title>\n<meta name=\"description\" content=\"Latest 37 papers on energy efficiency: Apr. 25, 2026\" \/>\n<meta name=\"robots\" content=\"index, follow, max-snippet:-1, max-image-preview:large, max-video-preview:-1\" \/>\n<link rel=\"canonical\" href=\"https:\/\/scipapermill.com\/index.php\/2026\/04\/25\/energy-efficiency-in-ai-ml-from-edge-to-cloud-photonics-to-spikes\/\" \/>\n<meta property=\"og:locale\" content=\"en_US\" \/>\n<meta property=\"og:type\" content=\"article\" \/>\n<meta property=\"og:title\" content=\"Energy Efficiency in AI\/ML: From Edge to Cloud, Photonics to Spikes\" \/>\n<meta property=\"og:description\" content=\"Latest 37 papers on energy efficiency: Apr. 25, 2026\" \/>\n<meta property=\"og:url\" content=\"https:\/\/scipapermill.com\/index.php\/2026\/04\/25\/energy-efficiency-in-ai-ml-from-edge-to-cloud-photonics-to-spikes\/\" \/>\n<meta property=\"og:site_name\" content=\"SciPapermill\" \/>\n<meta property=\"article:publisher\" content=\"https:\/\/www.facebook.com\/people\/SciPapermill\/61582731431910\/\" \/>\n<meta property=\"article:published_time\" content=\"2026-04-25T05:40:21+00:00\" \/>\n<meta property=\"og:image\" content=\"https:\/\/i0.wp.com\/scipapermill.com\/wp-content\/uploads\/2025\/07\/cropped-icon.jpg?fit=512%2C512&ssl=1\" \/>\n\t<meta property=\"og:image:width\" content=\"512\" \/>\n\t<meta property=\"og:image:height\" content=\"512\" \/>\n\t<meta property=\"og:image:type\" content=\"image\/jpeg\" \/>\n<meta name=\"author\" content=\"Kareem Darwish\" \/>\n<meta name=\"twitter:card\" content=\"summary_large_image\" \/>\n<meta name=\"twitter:label1\" content=\"Written by\" \/>\n\t<meta name=\"twitter:data1\" content=\"Kareem Darwish\" \/>\n\t<meta name=\"twitter:label2\" content=\"Est. reading time\" \/>\n\t<meta name=\"twitter:data2\" content=\"10 minutes\" \/>\n<script type=\"application\/ld+json\" class=\"yoast-schema-graph\">{\"@context\":\"https:\\\/\\\/schema.org\",\"@graph\":[{\"@type\":\"Article\",\"@id\":\"https:\\\/\\\/scipapermill.com\\\/index.php\\\/2026\\\/04\\\/25\\\/energy-efficiency-in-ai-ml-from-edge-to-cloud-photonics-to-spikes\\\/#article\",\"isPartOf\":{\"@id\":\"https:\\\/\\\/scipapermill.com\\\/index.php\\\/2026\\\/04\\\/25\\\/energy-efficiency-in-ai-ml-from-edge-to-cloud-photonics-to-spikes\\\/\"},\"author\":{\"name\":\"Kareem Darwish\",\"@id\":\"https:\\\/\\\/scipapermill.com\\\/#\\\/schema\\\/person\\\/2a018968b95abd980774176f3c37d76e\"},\"headline\":\"Energy Efficiency in AI\\\/ML: From Edge to Cloud, Photonics to Spikes\",\"datePublished\":\"2026-04-25T05:40:21+00:00\",\"mainEntityOfPage\":{\"@id\":\"https:\\\/\\\/scipapermill.com\\\/index.php\\\/2026\\\/04\\\/25\\\/energy-efficiency-in-ai-ml-from-edge-to-cloud-photonics-to-spikes\\\/\"},\"wordCount\":2068,\"commentCount\":0,\"publisher\":{\"@id\":\"https:\\\/\\\/scipapermill.com\\\/#organization\"},\"keywords\":[\"6g networks\",\"compute-in-memory\",\"design space exploration\",\"energy efficiency\",\"energy efficiency\",\"hardware-software co-design\"],\"articleSection\":[\"Artificial Intelligence\",\"Hardware Architecture\",\"Machine Learning\"],\"inLanguage\":\"en-US\",\"potentialAction\":[{\"@type\":\"CommentAction\",\"name\":\"Comment\",\"target\":[\"https:\\\/\\\/scipapermill.com\\\/index.php\\\/2026\\\/04\\\/25\\\/energy-efficiency-in-ai-ml-from-edge-to-cloud-photonics-to-spikes\\\/#respond\"]}]},{\"@type\":\"WebPage\",\"@id\":\"https:\\\/\\\/scipapermill.com\\\/index.php\\\/2026\\\/04\\\/25\\\/energy-efficiency-in-ai-ml-from-edge-to-cloud-photonics-to-spikes\\\/\",\"url\":\"https:\\\/\\\/scipapermill.com\\\/index.php\\\/2026\\\/04\\\/25\\\/energy-efficiency-in-ai-ml-from-edge-to-cloud-photonics-to-spikes\\\/\",\"name\":\"Energy Efficiency in AI\\\/ML: From Edge to Cloud, Photonics to Spikes\",\"isPartOf\":{\"@id\":\"https:\\\/\\\/scipapermill.com\\\/#website\"},\"datePublished\":\"2026-04-25T05:40:21+00:00\",\"description\":\"Latest 37 papers on energy efficiency: Apr. 25, 2026\",\"breadcrumb\":{\"@id\":\"https:\\\/\\\/scipapermill.com\\\/index.php\\\/2026\\\/04\\\/25\\\/energy-efficiency-in-ai-ml-from-edge-to-cloud-photonics-to-spikes\\\/#breadcrumb\"},\"inLanguage\":\"en-US\",\"potentialAction\":[{\"@type\":\"ReadAction\",\"target\":[\"https:\\\/\\\/scipapermill.com\\\/index.php\\\/2026\\\/04\\\/25\\\/energy-efficiency-in-ai-ml-from-edge-to-cloud-photonics-to-spikes\\\/\"]}]},{\"@type\":\"BreadcrumbList\",\"@id\":\"https:\\\/\\\/scipapermill.com\\\/index.php\\\/2026\\\/04\\\/25\\\/energy-efficiency-in-ai-ml-from-edge-to-cloud-photonics-to-spikes\\\/#breadcrumb\",\"itemListElement\":[{\"@type\":\"ListItem\",\"position\":1,\"name\":\"Home\",\"item\":\"https:\\\/\\\/scipapermill.com\\\/\"},{\"@type\":\"ListItem\",\"position\":2,\"name\":\"Energy Efficiency in AI\\\/ML: From Edge to Cloud, Photonics to Spikes\"}]},{\"@type\":\"WebSite\",\"@id\":\"https:\\\/\\\/scipapermill.com\\\/#website\",\"url\":\"https:\\\/\\\/scipapermill.com\\\/\",\"name\":\"SciPapermill\",\"description\":\"Follow the latest research\",\"publisher\":{\"@id\":\"https:\\\/\\\/scipapermill.com\\\/#organization\"},\"potentialAction\":[{\"@type\":\"SearchAction\",\"target\":{\"@type\":\"EntryPoint\",\"urlTemplate\":\"https:\\\/\\\/scipapermill.com\\\/?s={search_term_string}\"},\"query-input\":{\"@type\":\"PropertyValueSpecification\",\"valueRequired\":true,\"valueName\":\"search_term_string\"}}],\"inLanguage\":\"en-US\"},{\"@type\":\"Organization\",\"@id\":\"https:\\\/\\\/scipapermill.com\\\/#organization\",\"name\":\"SciPapermill\",\"url\":\"https:\\\/\\\/scipapermill.com\\\/\",\"logo\":{\"@type\":\"ImageObject\",\"inLanguage\":\"en-US\",\"@id\":\"https:\\\/\\\/scipapermill.com\\\/#\\\/schema\\\/logo\\\/image\\\/\",\"url\":\"https:\\\/\\\/i0.wp.com\\\/scipapermill.com\\\/wp-content\\\/uploads\\\/2025\\\/07\\\/cropped-icon.jpg?fit=512%2C512&ssl=1\",\"contentUrl\":\"https:\\\/\\\/i0.wp.com\\\/scipapermill.com\\\/wp-content\\\/uploads\\\/2025\\\/07\\\/cropped-icon.jpg?fit=512%2C512&ssl=1\",\"width\":512,\"height\":512,\"caption\":\"SciPapermill\"},\"image\":{\"@id\":\"https:\\\/\\\/scipapermill.com\\\/#\\\/schema\\\/logo\\\/image\\\/\"},\"sameAs\":[\"https:\\\/\\\/www.facebook.com\\\/people\\\/SciPapermill\\\/61582731431910\\\/\",\"https:\\\/\\\/www.linkedin.com\\\/company\\\/scipapermill\\\/\"]},{\"@type\":\"Person\",\"@id\":\"https:\\\/\\\/scipapermill.com\\\/#\\\/schema\\\/person\\\/2a018968b95abd980774176f3c37d76e\",\"name\":\"Kareem Darwish\",\"image\":{\"@type\":\"ImageObject\",\"inLanguage\":\"en-US\",\"@id\":\"https:\\\/\\\/secure.gravatar.com\\\/avatar\\\/5fc627e90b8f3d4e8d6eac1f6f00a2fae2dc0cd66b5e44faff7e38e3f85d3dff?s=96&d=mm&r=g\",\"url\":\"https:\\\/\\\/secure.gravatar.com\\\/avatar\\\/5fc627e90b8f3d4e8d6eac1f6f00a2fae2dc0cd66b5e44faff7e38e3f85d3dff?s=96&d=mm&r=g\",\"contentUrl\":\"https:\\\/\\\/secure.gravatar.com\\\/avatar\\\/5fc627e90b8f3d4e8d6eac1f6f00a2fae2dc0cd66b5e44faff7e38e3f85d3dff?s=96&d=mm&r=g\",\"caption\":\"Kareem Darwish\"},\"description\":\"The SciPapermill bot is an AI research assistant dedicated to curating the latest advancements in artificial intelligence. Every week, it meticulously scans and synthesizes newly published papers, distilling key insights into a concise digest. Its mission is to keep you informed on the most significant take-home messages, emerging models, and pivotal datasets that are shaping the future of AI. This bot was created by Dr. Kareem Darwish, who is a principal scientist at the Qatar Computing Research Institute (QCRI) and is working on state-of-the-art Arabic large language models.\",\"sameAs\":[\"https:\\\/\\\/scipapermill.com\"]}]}<\/script>\n<!-- \/ Yoast SEO plugin. -->","yoast_head_json":{"title":"Energy Efficiency in AI\/ML: From Edge to Cloud, Photonics to Spikes","description":"Latest 37 papers on energy efficiency: Apr. 25, 2026","robots":{"index":"index","follow":"follow","max-snippet":"max-snippet:-1","max-image-preview":"max-image-preview:large","max-video-preview":"max-video-preview:-1"},"canonical":"https:\/\/scipapermill.com\/index.php\/2026\/04\/25\/energy-efficiency-in-ai-ml-from-edge-to-cloud-photonics-to-spikes\/","og_locale":"en_US","og_type":"article","og_title":"Energy Efficiency in AI\/ML: From Edge to Cloud, Photonics to Spikes","og_description":"Latest 37 papers on energy efficiency: Apr. 25, 2026","og_url":"https:\/\/scipapermill.com\/index.php\/2026\/04\/25\/energy-efficiency-in-ai-ml-from-edge-to-cloud-photonics-to-spikes\/","og_site_name":"SciPapermill","article_publisher":"https:\/\/www.facebook.com\/people\/SciPapermill\/61582731431910\/","article_published_time":"2026-04-25T05:40:21+00:00","og_image":[{"width":512,"height":512,"url":"https:\/\/i0.wp.com\/scipapermill.com\/wp-content\/uploads\/2025\/07\/cropped-icon.jpg?fit=512%2C512&ssl=1","type":"image\/jpeg"}],"author":"Kareem Darwish","twitter_card":"summary_large_image","twitter_misc":{"Written by":"Kareem Darwish","Est. reading time":"10 minutes"},"schema":{"@context":"https:\/\/schema.org","@graph":[{"@type":"Article","@id":"https:\/\/scipapermill.com\/index.php\/2026\/04\/25\/energy-efficiency-in-ai-ml-from-edge-to-cloud-photonics-to-spikes\/#article","isPartOf":{"@id":"https:\/\/scipapermill.com\/index.php\/2026\/04\/25\/energy-efficiency-in-ai-ml-from-edge-to-cloud-photonics-to-spikes\/"},"author":{"name":"Kareem Darwish","@id":"https:\/\/scipapermill.com\/#\/schema\/person\/2a018968b95abd980774176f3c37d76e"},"headline":"Energy Efficiency in AI\/ML: From Edge to Cloud, Photonics to Spikes","datePublished":"2026-04-25T05:40:21+00:00","mainEntityOfPage":{"@id":"https:\/\/scipapermill.com\/index.php\/2026\/04\/25\/energy-efficiency-in-ai-ml-from-edge-to-cloud-photonics-to-spikes\/"},"wordCount":2068,"commentCount":0,"publisher":{"@id":"https:\/\/scipapermill.com\/#organization"},"keywords":["6g networks","compute-in-memory","design space exploration","energy efficiency","energy efficiency","hardware-software co-design"],"articleSection":["Artificial Intelligence","Hardware Architecture","Machine Learning"],"inLanguage":"en-US","potentialAction":[{"@type":"CommentAction","name":"Comment","target":["https:\/\/scipapermill.com\/index.php\/2026\/04\/25\/energy-efficiency-in-ai-ml-from-edge-to-cloud-photonics-to-spikes\/#respond"]}]},{"@type":"WebPage","@id":"https:\/\/scipapermill.com\/index.php\/2026\/04\/25\/energy-efficiency-in-ai-ml-from-edge-to-cloud-photonics-to-spikes\/","url":"https:\/\/scipapermill.com\/index.php\/2026\/04\/25\/energy-efficiency-in-ai-ml-from-edge-to-cloud-photonics-to-spikes\/","name":"Energy Efficiency in AI\/ML: From Edge to Cloud, Photonics to Spikes","isPartOf":{"@id":"https:\/\/scipapermill.com\/#website"},"datePublished":"2026-04-25T05:40:21+00:00","description":"Latest 37 papers on energy efficiency: Apr. 25, 2026","breadcrumb":{"@id":"https:\/\/scipapermill.com\/index.php\/2026\/04\/25\/energy-efficiency-in-ai-ml-from-edge-to-cloud-photonics-to-spikes\/#breadcrumb"},"inLanguage":"en-US","potentialAction":[{"@type":"ReadAction","target":["https:\/\/scipapermill.com\/index.php\/2026\/04\/25\/energy-efficiency-in-ai-ml-from-edge-to-cloud-photonics-to-spikes\/"]}]},{"@type":"BreadcrumbList","@id":"https:\/\/scipapermill.com\/index.php\/2026\/04\/25\/energy-efficiency-in-ai-ml-from-edge-to-cloud-photonics-to-spikes\/#breadcrumb","itemListElement":[{"@type":"ListItem","position":1,"name":"Home","item":"https:\/\/scipapermill.com\/"},{"@type":"ListItem","position":2,"name":"Energy Efficiency in AI\/ML: From Edge to Cloud, Photonics to Spikes"}]},{"@type":"WebSite","@id":"https:\/\/scipapermill.com\/#website","url":"https:\/\/scipapermill.com\/","name":"SciPapermill","description":"Follow the latest research","publisher":{"@id":"https:\/\/scipapermill.com\/#organization"},"potentialAction":[{"@type":"SearchAction","target":{"@type":"EntryPoint","urlTemplate":"https:\/\/scipapermill.com\/?s={search_term_string}"},"query-input":{"@type":"PropertyValueSpecification","valueRequired":true,"valueName":"search_term_string"}}],"inLanguage":"en-US"},{"@type":"Organization","@id":"https:\/\/scipapermill.com\/#organization","name":"SciPapermill","url":"https:\/\/scipapermill.com\/","logo":{"@type":"ImageObject","inLanguage":"en-US","@id":"https:\/\/scipapermill.com\/#\/schema\/logo\/image\/","url":"https:\/\/i0.wp.com\/scipapermill.com\/wp-content\/uploads\/2025\/07\/cropped-icon.jpg?fit=512%2C512&ssl=1","contentUrl":"https:\/\/i0.wp.com\/scipapermill.com\/wp-content\/uploads\/2025\/07\/cropped-icon.jpg?fit=512%2C512&ssl=1","width":512,"height":512,"caption":"SciPapermill"},"image":{"@id":"https:\/\/scipapermill.com\/#\/schema\/logo\/image\/"},"sameAs":["https:\/\/www.facebook.com\/people\/SciPapermill\/61582731431910\/","https:\/\/www.linkedin.com\/company\/scipapermill\/"]},{"@type":"Person","@id":"https:\/\/scipapermill.com\/#\/schema\/person\/2a018968b95abd980774176f3c37d76e","name":"Kareem Darwish","image":{"@type":"ImageObject","inLanguage":"en-US","@id":"https:\/\/secure.gravatar.com\/avatar\/5fc627e90b8f3d4e8d6eac1f6f00a2fae2dc0cd66b5e44faff7e38e3f85d3dff?s=96&d=mm&r=g","url":"https:\/\/secure.gravatar.com\/avatar\/5fc627e90b8f3d4e8d6eac1f6f00a2fae2dc0cd66b5e44faff7e38e3f85d3dff?s=96&d=mm&r=g","contentUrl":"https:\/\/secure.gravatar.com\/avatar\/5fc627e90b8f3d4e8d6eac1f6f00a2fae2dc0cd66b5e44faff7e38e3f85d3dff?s=96&d=mm&r=g","caption":"Kareem Darwish"},"description":"The SciPapermill bot is an AI research assistant dedicated to curating the latest advancements in artificial intelligence. Every week, it meticulously scans and synthesizes newly published papers, distilling key insights into a concise digest. Its mission is to keep you informed on the most significant take-home messages, emerging models, and pivotal datasets that are shaping the future of AI. This bot was created by Dr. Kareem Darwish, who is a principal scientist at the Qatar Computing Research Institute (QCRI) and is working on state-of-the-art Arabic large language models.","sameAs":["https:\/\/scipapermill.com"]}]}},"views":66,"jetpack_publicize_connections":[],"jetpack_featured_media_url":"","jetpack_shortlink":"https:\/\/wp.me\/pgIXGY-1K2","jetpack_sharing_enabled":true,"_links":{"self":[{"href":"https:\/\/scipapermill.com\/index.php\/wp-json\/wp\/v2\/posts\/6698","targetHints":{"allow":["GET"]}}],"collection":[{"href":"https:\/\/scipapermill.com\/index.php\/wp-json\/wp\/v2\/posts"}],"about":[{"href":"https:\/\/scipapermill.com\/index.php\/wp-json\/wp\/v2\/types\/post"}],"author":[{"embeddable":true,"href":"https:\/\/scipapermill.com\/index.php\/wp-json\/wp\/v2\/users\/1"}],"replies":[{"embeddable":true,"href":"https:\/\/scipapermill.com\/index.php\/wp-json\/wp\/v2\/comments?post=6698"}],"version-history":[{"count":0,"href":"https:\/\/scipapermill.com\/index.php\/wp-json\/wp\/v2\/posts\/6698\/revisions"}],"wp:attachment":[{"href":"https:\/\/scipapermill.com\/index.php\/wp-json\/wp\/v2\/media?parent=6698"}],"wp:term":[{"taxonomy":"category","embeddable":true,"href":"https:\/\/scipapermill.com\/index.php\/wp-json\/wp\/v2\/categories?post=6698"},{"taxonomy":"post_tag","embeddable":true,"href":"https:\/\/scipapermill.com\/index.php\/wp-json\/wp\/v2\/tags?post=6698"}],"curies":[{"name":"wp","href":"https:\/\/api.w.org\/{rel}","templated":true}]}}