­

PSI project: Pervasive and Sustainable AI with Distributed Edge Intelligence

The Edge Intelligence (EI) has emerged as a promising alternative to the traditional artificial intelligence (AI) paradigm. However, due to the exponential upsurge of data generated, the continued expansion of connected devices, and the projected AI annual growth rate of 37.3% from 2023 to 2030, incorporating sustainable design principles into EI platforms is essential to prevent environmental degradation and avoid the rebound effect problem. PSI proposes a novel energy-aware integrated computing and communication platform that will radically transform the current energy-intensive AI paradigm by centralized AI. The focus will be developing fully decentralized peer-to-peer training solutions that integrate continual and knowledge transfer learning, aiming for a massive reduction in energy consumption compared to standard ML models, while maintaining equivalent accuracy and latency. Furthermore, energy-efficient brain-inspired algorithms will be integrated to further minimize energy usage without compromising performance. PSI advocates for a pervasive and liquid AI system, where edge devices can actively contribute to AI service provision. Through this approach, devices will collaboratively process data from the underlying physical systems using lightweight, brain-inspired ML models, regardless of their resource capabilities, enabling the Artificial Intelligence of Things (AIoT). This paradigm will position users and devices as active participants in the AI process, not merely passive data generators. Moreover, AIoT has the potential to democratize AI and reduce the digital gap, shifting control away from large tech companies with expensive AI services and enabling more transparent, user-centered AI systems. The overarching goal of this project is to demonstrate that a sustainable, user-driven AI paradigm is not only possible but necessary – one where users and governmental bodies alike can participate in deploying innovative AI services without costly infrastructure and environmental harm.  

PSI project is developed within Sustainable Artificial Intelligence Research Unit by Marco Miozzo and Matteo Mendula

Energy efficiency in the Information and Communication Technologies (ICT) ecosystem has always been a pillar for reducing the energy used and enabling higher integration. However, despite the continuous increment in efficiency, the energy consumption of ICT has kept increasing during the last decades and it now starts to become a pressing problem that we urgently need to solve. To keep the temperature increment below 2ºC we need to cut the CO2 emissions by half by 2030, which means a decrease of 7% per year on a world scale basis. ICT is part of the problem: recent estimations fix its contribution to global emissions high as 2.1%–3.9% [1]. Moreover, ICT is one of the sectors that is experiencing the highest growth in terms of CO2 emissions, reaching 6% nowadays, and, with the current trend, it will be eating up 15% by 2030. In addition, new services based on AI are becoming more pervasive each day, with models that are exponentially increasing in their complexity. A very recent example from standard Centralized AI (CAI) can be found in OpenAI´s ChatGPT service, whose monthly electricity consumption is estimated to be millions of kWh: “ChatGPT may have consumed as much electricity as 175,000 people in January 2023" [2]. Similarly, Meta used 2.6 million KWh hours of electricity and emitted 1,000 tons of CO2 for the development of their new Large Language Model Meta AI (LLaMA) models [3]. This energy is estimated to have led to the emission of 1,015 tCO2e, roughly the annual carbon footprint of 150 EU citizens [4]. Moreover, CAI suffers from managing data coming from distributed sources in a centralized fashion. In fact, data must be transmitted to the cloud to be processed, with the consequent issues of high network congestion, waste of energy, and high latency. Furthermore, cloud data centers are normally managed by a third party, which may represent a breach in the privacy preservation of sensitive content. The Edge Intelligence (EI) paradigm has emerged as a mitigation to the challenges described above. EI involves performing AI computations directly at the network edge, where most of the data is generated. This approach offers several advantages with respect to CAI, e.g. reduced communication overhead, faster response time, and increased privacy preservation. Current research in EI focuses on distributing learning processes to achieve accuracy levels comparable to CAI [5], regardless of the energy demands of computation and communication. However, considering the exponential upsurge of data generated [6], the continued expansion of connected devices, and the projected AI annual growth rate of 37.3% from 2023 to 2030 [7], incorporating sustainable design principles into EI platforms is essential to prevent environmental degradation and avoid the rebound effect problem [1]. This shift in AI design and application is the central objective of the PSI project. 

PSI proposes a novel energy-aware integrated computing and communication platform that will radically transform the current energy-intensive AI paradigm by centralized CAI. In this framework, traditional data centers will be replaced by far-edge devices, such as cyber-physical systems (CPSs), becoming key enablers of pervasive AI services, also called Artificial Intelligence of Things (AIoT). These devices, often limited in connectivity, computing capabilities, memory, and reliant on batteries, will require the adoption of highly efficient energy-saving, communication, and computational strategies. Moreover, data distribution across CPSs may exhibit varying statistical characteristics, even when sourced from the same physical system [8], leading to non-identical and independent distributions (non-iid), which can negatively impact the performance of traditional ML models. The PSI proposed solution empowers CPSs to compute accurate ML models collaboratively on-device, reducing or eliminating the reliance on cloud data centers. To overcome the limitations of standard CPSs (e.g., sensor nodes), and their constrained computational and energy resources, highly efficient learning architectures are essential. Existing CAI-based learning solutions are excessively complex for these devices, necessitating the development of novel, streamlined algorithms.  

By implementing decentralized techniques, PSI aims to enable direct communication among devices for both training and inference, removing the need for centralized entities and reducing long-range data transmissions. This approach will mitigate potential bottlenecks in local edge nodes and aggregators, enhancing both system scalability and energy efficiency. The same decentralized and energy-efficient AI architecture will be applicable to a wide range of AIoT contexts, including cognitive buildings, where resource-constrained sensors can manage real-time data for energy-efficient building operations. Moreover, continual learning approaches will ensure that models adapt to non-stationary environments, a typical characteristic in real-world conditions. 

PSI proposes two primary technological enablers to achieve this vision: 

Brain-inspired computing: PSI will consider both reservoir and neuromorphic computing to enable lightweight ML models. Reservoir computing is especially well-suited for learning dynamical systems from time-series data, requiring only small training datasets, using linear optimization, and demanding minimal computational resources [9]. This makes it a natural fit for energy-limited CPSs. In neuromorphic computing circuits are made of physical neurons interconnected by physical synapses with in-situ, non-volatile memory, drastically reducing the need for data transfer within the circuit. Thus, it offers substantial improvements in speed and energy efficiency [10], making it a key driver for edge intelligence. The interplay between the reservoir and neuromorphic computing will also be considered to further improve energy savings. 

Distributed and collaborative learning: Techniques such as Federated Learning (FL) and Continual Learning (CL) [11][12][13] enable efficient model sharing across data sources, fostering knowledge distillation and enhancing the system´s generalization capabilities, especially in the presence of non-iid data. These approaches are essential for achieving high-performance ML in distributed environments with limited computational and energy resources. In particular, FL aims to coordinate the nodes to learn from their local data simultaneously and merge the shared knowledge in a parallel fashion. Whereas CL can be used to learn from the distributed source of knowledge sequentially, hybrid solutions that combine the strengths of both FL and CL must be explored to create a truly sustainable AI paradigm tailored to specific applications.  

It is important to note that, to date, these areas have not yet been thoroughly studied from an energy-aware perspective, nor have lightweight ML models (such as those based on reservoir or neuromorphic computing) been considered in distributed learning systems. This represents a significant gap in the state-of-the-art, one that this project seeks to fill. 

In summary, PSI advocates for a pervasive and liquid AI system, where edge devices can actively contribute to AI service provision, the AIoT [14]. Through this approach, CPSs will collaboratively process data from the underlying physical systems using lightweight, brain-inspired ML models, regardless of their resource capabilities. This paradigm will position users and devices as active participants in the AI process, not merely passive data generators. Moreover, it has the potential to democratize AI, shifting control away from large tech companies and enabling more transparent, user-centered AI systems. Finally, this approach also enhances user privacy by keeping personal data at the edge, avoiding the need to transfer it to third-party data centers with limited transparency. 

PSI will apply novel research methodologies to the sustainable design of AIoT devices, platforms, and services, through the interdisciplinary understanding of this timely technology area of key importance for society. In particular, the aim is to persuade that the current narrative that “the larger the AI system the more valuable, powerful, and interesting” can be changed by working with small and collaborative ML solutions. In the last decade, the research community embraced the bigger-is-the-better AI paradigm with the consequence of an explosion in investment in large-scale AI models [15]. From a research perspective, reviewers ask for experiments at a large scale, both in the context of new models and in the context of measuring performance against existing models, which implies running experiments many times to search for the optimal hyperparameter. This pushes the need for computing capabilities beyond the common budgets of most university labs and, in turn, many labs have to be increasingly dependent on industry to secure such access. The final consequence is that investigating AI can be increasingly difficult for anyone outside of large industrial labs. The expense of the computing elements needed to build and operate large AI models is benefiting the actors in AI that have already access to it. This phenomenon concentrates power and influence over AI, perpetuating the bigger-is-better AI paradigm in service of maintaining their market advantage. An example can be found in the recent shortages in chips when the gap between the “GPU rich” and “GPU poor” has increased due to the lack of computing resources. However, choosing the right model architecture for the data at hand is crucial, and Transformer-based models, widely perceived to be SOTA on most or even all ML benchmarks, are not always the most fitting solution. For this, it is useful to consider the trade-off between task performance (e.g. performance on an ML benchmark) and the computing resources used, the so-called Green AI paradigm. PSI wants to persuade that task performance can still play an important role and compete with bigger-is-better AI, especially when considering intrinsic distributed problems, as for AIoT. To do so, PSI wants to show the limitations and expectations of brain-inspired ML solutions when working in a collaborative environment. This would open a more democratic paradigm for doing research on AI, enabling the use of small-scale and less expensive hardware. 

[1] C. Freitag, M. Berners-Lee, K. Widdicks, B. Knowles, G. S. Blair, A. Friday, “The real climate and transformative impact of ICT: A critique of estimates, trends, and regulations”, Patterns, Volume 2, Issue 9, 2021, 100340, ISSN 2666-3899, https://doi.org/10.1016/j.patter.2021.100340. 

[2] ChatGPT’s Electricity Consumption, Kasper Groes Albin Ludvigsen, Toward Data Science,  https://towardsdatascience.com/chatgpts-electricity-consumption-7873483feac4 

[3] H. Touvron et al  “LLaMA: Open and Efficient Foundation Language Models”, 2023, ArXiv, abs/2302.13971. 

[4] Facebook disclose the carbon footprint of their new LLaMA models, Kasper Groes Albin Ludvigsen, Medium, https://kaspergroesludvigsen.medium.com/facebook-disclose-the-carbon-footprint-of-their-new-llama-models-9629a3c5c28b 

[5] Z. Zhou, X. Chen, E. Li, L. Zeng, K. Luo, and J. Zhang, ‘‘Edge intelligence: Paving the last mile of artificial intelligence with edge computing,’’ Proc. IEEE, vol. 107, no. 8, pp. 1738–1762, Aug. 2019 

[6] ACEEE. 2017. Intelligent Efficiency Technology and Market Assessment. Report IE1701. Prepared by E.A. Rogers and E. Junga of the American Council for an Energy Efficient Economy. 

[7] Forbes, “24 Top AI Statistics And Trends In 2024”, https://www.forbes.com/advisor/business/ai-statistics/#sources_section 

[8] Zhu, J. Xu, S. Liu, Y. Jin, “Federated learning on non-IID data: A survey”, Neurocomputing, Volume 465, 2021, Pages 371-390,  ISSN 0925-2312,  https://doi.org/10.1016/j.neucom.2021.07.098. 

[9] D.J. Gauthier, E. Bollt, A. Griffith, et al. Next generation reservoir computing. Nat Commun12, 5564 (2021). https://doi.org/10.1038/s41467-021-25801-2 

[10] Big data needs a hardware revolution. Nature http://www.nature.com/articles/d41586-018-01683-1 (2018) doi:10.1038/d41586-018-01683-1. 

[11] M. Miozzo, Z. Ali, L. Giupponi and P. Dini, "Distributed and Multi-Task Learning at the Edge for Energy Efficient Radio Access Networks," in IEEE Access, vol. 9, pp. 12491-12505, 2021, doi: 10.1109/ACCESS.2021.3050841. 

[12] C. Lanza, E. Angelats, M. Miozzo and P. Dini, "Urban Traffic Forecasting using Federated and Continual Learning," 2023 6th Conference on Cloud and Internet of Things (CIoT), Lisbon, Portugal, 2023, pp. 1-8, doi: 10.1109/CIoT57267.2023.10084875. 

[13] E. Guerra, F. Wilhelmi, M. Miozzo and P. Dini, "The Cost of Training Machine Learning Models Over Distributed Data Sources," in IEEE Open Journal of the Communications Society, vol. 4, pp. 1111-1126, 2023, doi: 10.1109/OJCOMS.2023.3274394. 

[14] Zhang and D. Tao, "Empowering Things With Intelligence: A Survey of the Progress, Challenges, and Opportunities in Artificial Intelligence of Things," in IEEE Internet of Things Journal, vol. 8, no. 10, pp. 7789-7817, 15 May 15, 2021. 

[15] Varoquaux, G., Luccioni, A. S., & Whittaker, M. (2024). Hype, Sustainability, and the Price of the Bigger-is-Better Paradigm in AI. arXiv preprint arXiv:2409.14160. 

­
We use cookies

We use cookies on our website. Some of them are essential for the operation of the site, while others help us to improve this site and the user experience (tracking cookies). You can decide for yourself whether you want to allow cookies or not. Please note that if you reject them, you may not be able to use all the functionalities of the site.