Wednesday, April 03, 2024

 

Computer scientists show the way: AI models need not be SO power hungry



UNIVERSITY OF COPENHAGEN - FACULTY OF SCIENCE
Scatter plot 

IMAGE: 

EACH DOT IN THIS FIGURE IS A CONVOLUTIONAL NEURAL NETWORK MODEL WITH THE ENERGY CONSUMPTION ON HORIZONTAL AXIS AND PERFORMANCE ON VERTICAL AXIS. CONVENTIONALLY, MODELS ARE SELECTED ONLY BASED ON THEIR PERFORMANCE - WITHOUT TAKING THEIR ENERGY CONSUMPTION INTO ACCOUNT - RESULTING IN MODELS IN THE RED ELLIPSE. THIS WORK ENABLES PRACTITIONERS TO SELECT MODELS FROM THE GREEN ELLIPSE WHICH GIVE GOOD TRADE-OFF BETWEEN EFFECTIVENESS AND EFFICIENCY. 

view more 

CREDIT: FIGURE FROM SCIENTIFIC ARTICLE (HTTPS://IEEEXPLORE.IEEE.ORG/DOCUMENT/10448303)




The fact that colossal amounts of energy are needed to Google away, talk to Siri, ask ChatGPT to get something done, or use AI in any sense, has gradually become common knowledge. One study estimates that by 2027, AI servers will consume as much energy as Argentina or Sweden. Indeed, a single ChatGPT prompt is estimated to consume, on average, as much energy as forty mobile phone charges. But the research community and the industry have yet to make the development of AI models that are energy efficient and thus more climate friendly the focus, computer science researchers at the University of Copenhagen point out.

"Today, developers are narrowly focused on building AI models that are effective in terms of the accuracy of their results. It's like saying that a car is effective because it gets you to your destination quickly, without considering the amount of fuel it uses. As a result, AI models are often inefficient in terms of energy consumption," says Assistant Professor Raghavendra Selvan from the Department of Computer Science, whose research looks in to possibilities for reducing AI’s carbon footprint.

But the new study, of which he and computer science student Pedram Bakhtiarifard are two of the authors, demonstrates that it is easy to curb a great deal of CO2e without compromising the precision of an AI model. Doing so demands keeping climate costs in mind from the design and training phases of AI models.

"If you put together a model that is energy efficient from the get-go, you reduce the carbon footprint in each phase of the model's 'life cycle'. This applies both to the model’s training, which is a particularly energy-intensive process that often takes weeks or months, as well as to its application," says Selvan.

Recipe book for the AI industry

In their study, the researchers calculated how much energy it takes to train more than 400,000 convolutional neural network type AI models – this was done without actually training all these models. Among other things, convolutional neural networks are used to analyse medical imagery, for language translation and for object and face recognition – a function you might know from the camera app on your smartphone.

Based on the calculations, the researchers present a benchmark collection of AI models that use less energy to solve a given task, but which perform at approximately the same level. The study shows that by opting for other types of models or by adjusting models, 70-80% energy savings can be achieved during the training and deployment phase, with only a 1% or less decrease in performance. And according to the researchers, this is a conservative estimate.

"Consider our results as a recipe book for the AI professionals. The recipes don’t just describe the performance of different algorithms, but how energy efficient they are. And that by swapping one ingredient with another in the design of a model, one can often achieve the same result. So now, the practitioners can choose the model they want based on both performance and energy consumption, and without needing to train each model first," says Pedram Bakhtiarifard, who continues:

"Oftentimes, many models are trained before finding the one that is suspected of being the most suitable for solving a particular task. This makes the development of AI extremely energy-intensive. Therefore, it would be more climate-friendly to choose the right model from the outset, while choosing one that does not consume too much power during the training phase."

The researchers stress that in some fields, like self-driving cars or certain areas of medicine, model precision can be critical for safety. Here, it is important not to compromise on performance. However, this shouldn’t be a deterrence to striving for high energy efficiency in other domains.

"AI has amazing potential. But if we are to ensure sustainable and responsible AI development, we need a more holistic approach that not only has model performance in mind, but also climate impact. Here, we show that it is possible to find a better trade-off. When AI models are developed for different tasks, energy efficiency ought to be a fixed criterion – just as it is standard in many other industries," concludes Raghavendra Selvan.

The “recipe book” put together in this work is available as an open-source dataset for other researchers to experiment with. The information about all these 423,000 architectures is published on Github which AI practitioners can access using simple Python scripts. 

 

 

[BOX:] EQUALS 46 YEARS OF A DANE’S ENERGY CONSUMPTION

The UCPH researchers estimated how much energy it takes to train 429,000 of the AI subtype models known as convolutional neural networks in this dataset. Among other things, these are used for object detection, language translation and medical image analysis.

It is estimated that the training alone of the 429,000 neural networks the study looked at would require 263,000 kWh. This equals the amount of energy that an average Danish citizen consumes over 46 years. And it would take one computer about 100 years to do the training. The authors in this work did not actually train these models themselves but estimated these using another AI model, and thus saving 99% of the energy it would have taken.

 

 

[BOX:] WHY IS AI’S CARBON FOOTPRINT SO BIG?

Training AI models consumes a lot of energy, and thereby emits a lot of CO2e. This is due to the intensive computations performed while training a model, typically run on powerful computers. This is especially true for large models, like the language model behind ChatGPT. AI tasks are often processed in data centers, which demand significant amounts of power to keep computers running and cool. The energy source for these centers, which may rely on fossil fuels, influences their carbon footprint.

 

 

[BOX:] ABOUT THE STUDY

  • The scientific article about the study will be presented at the International Conference on Acoustics, Speech and Signal Processing (ICASSP-2024).
  • The authors of the article are Pedram Bakhtiarifard, Christian Igel and Raghavendra Selvan from the University of Copenhagen’s Department of Computer Science.

 

A new 'Deep Learning' model predicts with great accuracy water and energy demands in Agriculture


Researchers develop a model based on the ‘Transformer’ architecture to guide irrigation communities' decision-making


UNIVERSITY OF CÓRDOBA

Researchers Emilio Camacho, Juan Antonio Rodríguez and Rafael González 

IMAGE: 

RESEARCHERS EMILIO CAMACHO, JUAN ANTONIO RODRÍGUEZ AND RAFAEL GONZÁLEZ, FROM THE AGRONOMY DEPARTMENT AT THE UNIVERSITY OF CÓRDOBA.

view more 

CREDIT: UNIVERSITY OF CORDOBA




Water scarcity and the high cost of energy represent the main problems for irrigation communities, which manage water for this end, making it available to agriculture. In a context of drought, with a deregulated and changing electricity market, knowing when and how much water crops are going to be irrigated with would allow those who manage them to overcome uncertainty when making decisions and, therefore, guide them towards objectives like economic savings, environmental sustainability and efficiency. For this, data science and Artificial Intelligence are important resources.

Researchers from the Hydraulics and Irrigation group with the María de Maeztu Unit of Excellence in the Agronomy Department at the University of Córdoba (DAUCO) are working to apply this cutting-edge technology to the field of precision agriculture. An example of this is the HOPE project, focused on the development of a holistic precision irrigation model that also involves the application of AI to guide decision-making. Within the framework of this effort, prediction models have been developed that would furnish irrigation communities with rigorous estimates of the amount of water that growers will need to meet their crops' needs. The latest model developed, and the most accurate to date, makes it possible to predict the actual demand for irrigation water one week ahead and with a margin of error of less than 2%, thus make possible the effective management of resources, all without detracting autonomy from its users.

According to researchers Rafael González, Emilio Camacho and Juan Antonio Rodríguez, this advance represents another step in the line of digitization applied to irrigation developed by the AGR 228 "Hydraulics and Irrigation" research group. Now they have applied the revolutionary architecture of Transformer Deep Learning to the field of precision irrigation. Since its appearance in 2017, this has been implemented in various sectors and is at the root of Artificial Intelligence milestones, such as ChatGPT. The ‘Transformer’ architecture stands out for its ability to establish long-term relationships in sequential data through what are known as 'attention mechanisms.' In the case of irrigation, this data architecture allows a lot of information to be processed simultaneously, delegating the selection and extraction of the information necessary for optimal prediction to its artificial neural network.

Daily data from the irrigation campaigns from 2015 to 2022 in the Zujar Canal's Community of Irrigators, in Don Benito (Badajoz), were used to validate the results of this model. In total, more than 1,800 water consumption measurements were used to train the model, combined with data on temperature, precipitation, solar radiation, evapotranspiration, wind speed, humidity, and crop types, etc.
This has reduced the margin of error from previous models from 20% to just 2%. Applied to integrated decision-making support systems, this can be very useful for managers of irrigation communities by offering an accurate forecast of the daily demand for irrigation water for the next seven days in contexts of water scarcity and high energy prices, but also in the framework of a commitment to sustainable resource management.

AI can take over key management roles in scientific research



ESMT BERLIN
AI can take over key management roles in scientific research 

IMAGE: 

PORTRAIT OF HENRY SAUERMANN AND MAXIMILIAN KOEHLER

view more 

CREDIT: ESMT BERLIN




Researchers Maximilian Koehler, PhD candidate at ESMT, and Henry Sauermann, professor of strategy at ESMT, explore the role of AI, not as a “worker” performing specific research tasks such as data collection and analysis, but as a “manager” of human workers performing such tasks. Algorithmic management (AM) suggests a significant shift in the way research projects are conducted and can enable projects to operate at larger scale and efficiency.

With the complexity and scope of scientific research rapidly increasing, the study illustrates that AI can not only replicate but also potentially surpass human managers by leveraging its instantaneous, comprehensive, and interactive capabilities. Investigating algorithmic management in crowd and citizen science, Koehler and Sauermann discuss examples of how AI effectively performs five important managerial functions: task division and allocation, direction, coordination, motivation, and supporting learning.

The researchers investigated projects through online documents; by interviewing organizers, AI developers, and project participants; and by joining some projects as participants. This allowed the researchers to identify projects that use algorithmic management, to understand how AI performs management functions, and to explore when AM might be more effective.

The growing number of use cases suggests that the adoption of AM could be a critical factor in improving research productivity. “The capabilities of artificial intelligence have reached a point where AI can now significantly enhance the scope and efficiency of scientific research by managing complex, large-scale projects,” states Koehler.

In a quantitative comparison with a broader sample of projects, the study also reveals that AM-enabled projects are often larger than projects that do not use AM and are associated with platforms that provide access to shared AI tools. This suggests that AM may enable projects to scale but also requires technical infrastructures that stand-alone projects may find difficult to develop. These patterns point towards changing sources of competitive advantage in research and may have important implications for research funders, digital research platforms, and larger research organizations such as universities or corporate R&D labs.

Although AI can take over important management functions, this does not mean that principal investigators or human managers will become obsolete. Sauermann notes, “If AI can take over some of the more algorithmic and mundane functions of management, human leaders could shift their attention to more strategic and social tasks such as identifying high-value research targets, raising funding, or building an effective organizational culture.”

For more information on this research, please contact Maximilian Koehler. The study “Algorithmic Management in Scientific Research,” published in the journal Research Policy, can be viewed here.

--

About ESMT Berlin

ESMT Berlin is a leading global business school with its campus in the heart of Berlin. Founded by 25 global companies, ESMT offers master, MBA, and PhD programs, as well as executive education on its campus in Berlin, in locations around the world, online, and in online blended format. Focusing on leadership, innovation, and analytics, its diverse faculty publishes outstanding research in top academic journals. Additionally, the international business school provides an interdisciplinary platform for discourse between politics, business, and academia. ESMT is a non-profit private institution of higher education with the right to grant PhDs and is accredited by AACSB, AMBA, EQUIS, and ZEvA. It is committed to diversity, equity, and inclusion across all its activities and communities. esmt.berlin

From data to decisions: AI and IoT for earthquake prediction


KEAI COMMUNICATIONS CO., LTD.
Proposed integrated system architecture with multiple data sources used for AI and ML Earthquake model Prediction. 

IMAGE: 

PROPOSED INTEGRATED SYSTEM ARCHITECTURE WITH MULTIPLE DATA SOURCES USED FOR AI AND ML EARTHQUAKE MODEL PREDICTION.

view more 

CREDIT: PWAVODI JOSHUA, ET AL




The study of earthquake remains a main interest worldwide as it is  one of the least predictable natural disasters. In a new review published in the KeAi journal Artificial Intelligence in Geosciences, a tea, of researchers from France and Turkey explored the role of conventional tools like seismometers and GPS in understanding earthquakes and their aftermath.

“These tools have provided invaluable insights into various seismic parameters, such as ground deformation and displacement waves. However, they face several limitations, including the inability to predict earthquakes in real-time, challenges with temporal data resolution, and uneven spatial coverage,” explains Joshua Pwavodi, lead author of the review. “Despite their historical significance, these tools struggle to distinguish seismic signals from environmental noise.”

Nevertheless, the authors note that recent advancements in AI and IoT have significantly addressed some of these limitations. AI methodologies have proven instrumental in identifying intricate patterns and complex relationships within historical seismic data. By leveraging AI, unique insights into seismic patterns across diverse geological locations have been gained.

“Both classical and advanced machine learning techniques have contributed to the development of robust early warning systems and decentralized prediction models. IoT devices have also played a crucial role by enabling seamless data transmission for real-time monitoring,” adds Pwavodi.

The versatility of IoT devices enhances data accessibility and storage, creating a dynamic network for earthquake prediction. However, challenges such as computational complexity, data quality, and interpretability persist. A major limitation is the integration of primary hydrogeological measurements into AI model training. Monitoring hydrogeological data, including pore-fluid pressures and fluid flow, is often costly. Tools like the Circulation Obviation Retrofit Kits (CORKs) provide in-situ measurements of these parameters, but data transmission is not always in real-time, unlike IoT systems.

“To address these challenges, we proposed a comprehensive approach that integrates diverse datasets, including seismic, GPS, meteorological, and IoT sensor data,” says Pwavodi. “By combining these datasets, researchers can develop more robust earthquake prediction models that account for various contributing factors.”

Specifically, the authors suggest integrating IoT devices with tools like Circulation Obviation Retrofit Kits (CORKs) to enable real-time transmission of hydrogeological measurements influencing earthquakes. This real-time data, combined with other datasets, can be used to construct predictive AI models capable of providing real-time earthquake predictions.

###

Contact the author: Pwavodi Joshua, Hwodye Technology, France, Email: pwavodi@hwodye.com

The publisher KeAi was established by Elsevier and China Science Publishing & Media Ltd to unfold quality research globally. In 2013, our focus shifted to open access publishing. We now proudly publish more than 100 world-class, open access, English language journals, spanning all scientific disciplines. Many of these are titles we publish in partnership with prestigious societies and academic institutions, such as the National Natural Science Foundation of China (NSFC).