PSI project: Pervasive and Sustainable AI with Distributed Edge Intelligence
The Edge Intelligence (EI) has emerged as a promising alternative to the traditional artificial intelligence (AI) paradigm. However, due to the exponential upsurge of data generated, the continued expansion of connected devices, and the projected AI annual growth rate of 37.3% from 2023 to 2030, incorporating sustainable design principles into EI platforms is essential to prevent environmental degradation and avoid the rebound effect problem. PSI proposes a novel energy-aware integrated computing and communication platform that will radically transform the current energy-intensive AI paradigm by centralized AI. The focus will be developing fully decentralized peer-to-peer training solutions that integrate continual and knowledge transfer learning, aiming for a massive reduction in energy consumption compared to standard ML models, while maintaining equivalent accuracy and latency. Furthermore, energy-efficient brain-inspired algorithms will be integrated to further minimize energy usage without compromising performance. PSI advocates for a pervasive and liquid AI system, where edge devices can actively contribute to AI service provision. Through this approach, devices will collaboratively process data from the underlying physical systems using lightweight, brain-inspired ML models, regardless of their resource capabilities, enabling the Artificial Intelligence of Things (AIoT). This paradigm will position users and devices as active participants in the AI process, not merely passive data generators. Moreover, AIoT has the potential to democratize AI and reduce the digital gap, shifting control away from large tech companies with expensive AI services and enabling more transparent, user-centered AI systems. The overarching goal of this project is to demonstrate that a sustainable, user-driven AI paradigm is not only possible but necessary – one where users and governmental bodies alike can participate in deploying innovative AI services without costly infrastructure and environmental harm.
PSI project is developed within Sustainable Artificial Intelligence Research Unit by Marco Miozzo and Matteo Mendula.

PSI proposes a novel energy-aware integrated computing and communication platform that will radically transform the current energy-intensive AI paradigm by centralized CAI. In this framework, traditional data centers will be replaced by far-edge devices, such as cyber-physical systems (CPSs), becoming key enablers of pervasive AI services, also called Artificial Intelligence of Things (AIoT). These devices, often limited in connectivity, computing capabilities, memory, and reliant on batteries, will require the adoption of highly efficient energy-saving, communication, and computational strategies. Moreover, data distribution across CPSs may exhibit varying statistical characteristics, even when sourced from the same physical system [8], leading to non-identical and independent distributions (non-iid), which can negatively impact the performance of traditional ML models. The PSI proposed solution empowers CPSs to compute accurate ML models collaboratively on-device, reducing or eliminating the reliance on cloud data centers. To overcome the limitations of standard CPSs (e.g., sensor nodes), and their constrained computational and energy resources, highly efficient learning architectures are essential. Existing CAI-based learning solutions are excessively complex for these devices, necessitating the development of novel, streamlined algorithms.
By implementing decentralized techniques, PSI aims to enable direct communication among devices for both training and inference, removing the need for centralized entities and reducing long-range data transmissions. This approach will mitigate potential bottlenecks in local edge nodes and aggregators, enhancing both system scalability and energy efficiency. The same decentralized and energy-efficient AI architecture will be applicable to a wide range of AIoT contexts, including cognitive buildings, where resource-constrained sensors can manage real-time data for energy-efficient building operations. Moreover, continual learning approaches will ensure that models adapt to non-stationary environments, a typical characteristic in real-world conditions.
PSI proposes two primary technological enablers to achieve this vision:
Brain-inspired computing: PSI will consider both reservoir and neuromorphic computing to enable lightweight ML models. Reservoir computing is especially well-suited for learning dynamical systems from time-series data, requiring only small training datasets, using linear optimization, and demanding minimal computational resources [9]. This makes it a natural fit for energy-limited CPSs. In neuromorphic computing circuits are made of physical neurons interconnected by physical synapses with in-situ, non-volatile memory, drastically reducing the need for data transfer within the circuit. Thus, it offers substantial improvements in speed and energy efficiency [10], making it a key driver for edge intelligence. The interplay between the reservoir and neuromorphic computing will also be considered to further improve energy savings.
Distributed and collaborative learning: Techniques such as Federated Learning (FL) and Continual Learning (CL) [11][12][13] enable efficient model sharing across data sources, fostering knowledge distillation and enhancing the system´s generalization capabilities, especially in the presence of non-iid data. These approaches are essential for achieving high-performance ML in distributed environments with limited computational and energy resources. In particular, FL aims to coordinate the nodes to learn from their local data simultaneously and merge the shared knowledge in a parallel fashion. Whereas CL can be used to learn from the distributed source of knowledge sequentially, hybrid solutions that combine the strengths of both FL and CL must be explored to create a truly sustainable AI paradigm tailored to specific applications.
It is important to note that, to date, these areas have not yet been thoroughly studied from an energy-aware perspective, nor have lightweight ML models (such as those based on reservoir or neuromorphic computing) been considered in distributed learning systems. This represents a significant gap in the state-of-the-art, one that this project seeks to fill.
In summary, PSI advocates for a pervasive and liquid AI system, where edge devices can actively contribute to AI service provision, the AIoT [14]. Through this approach, CPSs will collaboratively process data from the underlying physical systems using lightweight, brain-inspired ML models, regardless of their resource capabilities. This paradigm will position users and devices as active participants in the AI process, not merely passive data generators. Moreover, it has the potential to democratize AI, shifting control away from large tech companies and enabling more transparent, user-centered AI systems. Finally, this approach also enhances user privacy by keeping personal data at the edge, avoiding the need to transfer it to third-party data centers with limited transparency.
PSI will apply novel research methodologies to the sustainable design of AIoT devices, platforms, and services, through the interdisciplinary understanding of this timely technology area of key importance for society. In particular, the aim is to persuade that the current narrative that “the larger the AI system the more valuable, powerful, and interesting” can be changed by working with small and collaborative ML solutions. In the last decade, the research community embraced the bigger-is-the-better AI paradigm with the consequence of an explosion in investment in large-scale AI models [15]. From a research perspective, reviewers ask for experiments at a large scale, both in the context of new models and in the context of measuring performance against existing models, which implies running experiments many times to search for the optimal hyperparameter. This pushes the need for computing capabilities beyond the common budgets of most university labs and, in turn, many labs have to be increasingly dependent on industry to secure such access. The final consequence is that investigating AI can be increasingly difficult for anyone outside of large industrial labs. The expense of the computing elements needed to build and operate large AI models is benefiting the actors in AI that have already access to it. This phenomenon concentrates power and influence over AI, perpetuating the bigger-is-better AI paradigm in service of maintaining their market advantage. An example can be found in the recent shortages in chips when the gap between the “GPU rich” and “GPU poor” has increased due to the lack of computing resources. However, choosing the right model architecture for the data at hand is crucial, and Transformer-based models, widely perceived to be SOTA on most or even all ML benchmarks, are not always the most fitting solution. For this, it is useful to consider the trade-off between task performance (e.g. performance on an ML benchmark) and the computing resources used, the so-called Green AI paradigm. PSI wants to persuade that task performance can still play an important role and compete with bigger-is-better AI, especially when considering intrinsic distributed problems, as for AIoT. To do so, PSI wants to show the limitations and expectations of brain-inspired ML solutions when working in a collaborative environment. This would open a more democratic paradigm for doing research on AI, enabling the use of small-scale and less expensive hardware.
[1] C. Freitag, M. Berners-Lee, K. Widdicks, B. Knowles, G. S. Blair, A. Friday, “The real climate and transformative impact of ICT: A critique of estimates, trends, and regulations”, Patterns, Volume 2, Issue 9, 2021, 100340, ISSN 2666-3899, https://doi.org/10.1016/j.patter.2021.100340.
[2] ChatGPT’s Electricity Consumption, Kasper Groes Albin Ludvigsen, Toward Data Science, https://towardsdatascience.com/chatgpts-electricity-consumption-7873483feac4
[3] H. Touvron et al “LLaMA: Open and Efficient Foundation Language Models”, 2023, ArXiv, abs/2302.13971.
[4] Facebook disclose the carbon footprint of their new LLaMA models, Kasper Groes Albin Ludvigsen, Medium, https://kaspergroesludvigsen.medium.com/facebook-disclose-the-carbon-footprint-of-their-new-llama-models-9629a3c5c28b
[5] Z. Zhou, X. Chen, E. Li, L. Zeng, K. Luo, and J. Zhang, ‘‘Edge intelligence: Paving the last mile of artificial intelligence with edge computing,’’ Proc. IEEE, vol. 107, no. 8, pp. 1738–1762, Aug. 2019
[6] ACEEE. 2017. Intelligent Efficiency Technology and Market Assessment. Report IE1701. Prepared by E.A. Rogers and E. Junga of the American Council for an Energy Efficient Economy.
[7] Forbes, “24 Top AI Statistics And Trends In 2024”, https://www.forbes.com/advisor/business/ai-statistics/#sources_section
[8] Zhu, J. Xu, S. Liu, Y. Jin, “Federated learning on non-IID data: A survey”, Neurocomputing, Volume 465, 2021, Pages 371-390, ISSN 0925-2312, https://doi.org/10.1016/j.neucom.2021.07.098.
[9] D.J. Gauthier, E. Bollt, A. Griffith, et al. Next generation reservoir computing. Nat Commun12, 5564 (2021). https://doi.org/10.1038/s41467-021-25801-2
[10] Big data needs a hardware revolution. Nature http://www.nature.com/articles/d41586-018-01683-1 (2018) doi:10.1038/d41586-018-01683-1.
[11] M. Miozzo, Z. Ali, L. Giupponi and P. Dini, "Distributed and Multi-Task Learning at the Edge for Energy Efficient Radio Access Networks," in IEEE Access, vol. 9, pp. 12491-12505, 2021, doi: 10.1109/ACCESS.2021.3050841.
[12] C. Lanza, E. Angelats, M. Miozzo and P. Dini, "Urban Traffic Forecasting using Federated and Continual Learning," 2023 6th Conference on Cloud and Internet of Things (CIoT), Lisbon, Portugal, 2023, pp. 1-8, doi: 10.1109/CIoT57267.2023.10084875.
[13] E. Guerra, F. Wilhelmi, M. Miozzo and P. Dini, "The Cost of Training Machine Learning Models Over Distributed Data Sources," in IEEE Open Journal of the Communications Society, vol. 4, pp. 1111-1126, 2023, doi: 10.1109/OJCOMS.2023.3274394.
[14] Zhang and D. Tao, "Empowering Things With Intelligence: A Survey of the Progress, Challenges, and Opportunities in Artificial Intelligence of Things," in IEEE Internet of Things Journal, vol. 8, no. 10, pp. 7789-7817, 15 May 15, 2021.
[15] Varoquaux, G., Luccioni, A. S., & Whittaker, M. (2024). Hype, Sustainability, and the Price of the Bigger-is-Better Paradigm in AI. arXiv preprint arXiv:2409.14160.