Skip to main content

AI for Sustainability and Sustainability in AI

I will be referring to the following 3 papers on this very interesting topic.

(1}  https://link.springer.com/article/10.1007/s43681-021-00043-6

 Sustainable AI: AI for sustainability and the sustainability of AI

A Van Wynsberghe – AI and Ethics, 2021 – Springe

(2) https://www.researchgate.net/publication/342763375_Carbontracker_Tracking_and_Predicting_the_Carbon_Footprint_of_Training_Deep_Learning_Models/link/5f0ef0f2a6fdcc3ed7083852/download

(3)      Lacoste, A., Luccioni, A., Schmidt, V., Dandres T.: Quantifying

the Carbon Emissions of Machine Learning. (2019)

While there is a tremendous push for using new-generation generative AI based on large language models to solve business applications, there are also voices of concern from experts in the community about the dangers and ethical consequences.  A lot has been written about this but one aspect which has not picked up sufficient traction, in my opinion, is Sustainable AI.  

In (1), Wynsberghe defines two disciplines on AI & sustainability.   AI for Sustainability and Sustainable AI.

AI for Sustainability is any business application using AIML technology to solve climate problems.  Use of this new generation technology to help in climate change and CO2 reductions.   Major applications are getting developed for optimal energy distribution across renewable and fossil energy sources. Any % extra use from renewable sources, help in less use of fossil fuels and help in climate change.  Various other applications may include better climate predictions and the use of less water, pesticides, and fertilizers for food production.  Many Industry 4.0 applications to build new smart factories, smart cities, and smart buildings fall into this category.

On the other hand, Sustainable AI measures the massive use of GPU and other computing, storage, and communications energy usage while building the AI models and suggest ways to reduce this.  While digital software development and testing can be done in a few developers’ laptops with minimal use of IT resources, the AIML software development life cycle calls for the use of massive training data and develop deep learning neural networks with multiple millions of nodes.   Some of the new generation Large Language models use billions of parameters beyond the imagination of all of us.  The energy use does not stop here.  Fine Tuning learning for specific domains or relearning is as energy-consuming or sometimes even more than original Training.   Some numbers mentioned in (1) are reproduced here to highlight the point.   One deep-learning NLP model consumed energy equivalent to 600,000 lbs of CO2.  Google Alpha-Go-Zero generated over 90 Tonnes of CO2 over 40 days it took for the initial training.  These numbers are large and at least call for review and discussions.   I hope I have been able to open your eyes and generate some interest in this new dimension of AI & Ethics i.e impact on climate change.

I am sure many of you will ask “Isn’t any next-generation industrialization from horse carriages to automobiles or steam engines to combustion always increased the use of energy and why do we need to worry about this for AI?”.  Or “There has been so much talk on how many light bulbs one can light for the same power used for a simple google search , why worry about this now ?”.  All valid questions.  

However, I will argue that

  1. The current climate change situation is already in a critical stage and any unplanned large-scale usage new of energy can become “the feather that broke the camel’s back!”.
  2. Use of fully data driven life cycle and billions of parameters, deep neural networks are being used for the first time at an industrial scale and industry-wide and there are too many unknowns.

What are the suggestions?

  • Energy consumption measurement and publication must become part of the AI & Ethics practice followed by all AI development organizations.   (2)  Carbon Tracker Tool and (3) Machine learning emission calculator are suggestions for this crucial measurement.  I strongly recommend organizations use their Quality & Metrics departments to research and agree on a measurement acceptable to all within each organization.  More research and discussions need to calculate the net increased use of energy compared to current IT tools to get the right measurement. In some cases, the current IT tools may be using legacy mainframes and expensive dedicated communication lines using up large amounts of energy and the net difference by using AIML may not be that large.
  • Predicting the energy use at the beginning of the AIML project life cycle also is required. (3). 
  • The prediction data of CO2 equivalent emissions need to be used as another cost in approving AIML projects.
  • Emission prediction also will force AIML developers to select the right size training data and use of right models for the application. Avoid the temptation of running the model on billions of data sets just because data is available!. Use the right tools for the right job.  You don’t need a tank to kill an ant!.
  • Ask the question of whether the use of deep learning is appropriate for this business application? For example, a simple HR application used for recruitment or employee loyalty prediction with Deep learning models may turn out to be too expensive in terms of Co2 emissions and need not be considered a viable project.
  • CEOs include this data in their Climate Change Initiatives Report to the Board and shareholders and also deduct carbon credits used up by these AIML applications in the company’s Carbon credit commitments.

More Later,

L Ravichandran

Small talk about Large Language Models

Since its formal launch, ChatGPT has been receiving a lot of press and has also been the topic of – heated – discussions in the recent past.

I had played with generative AI some time back and also shared the result in one of my earlier posts.

Post ChatGPT, the investments in AI – or more specifically generative AI tech – based companies has seen a sharp rise.

There is also a general sense of fear – rising from uncertainty and the dread of the possibility of such technologies taking away specialized jobs and roles has been noticed across industries.

I was talking to an architect a few days ago and she said that in their community, the awe and fear of AI tech is unprecedented.

With just a few words, some of the sketches generated by tools like Dall-E, Craiyon, Stable diffusion etc are apparently so realistic and logical.. for example, when the query was to have the porch door opening out into the garden with a path to the main gate.. the image was generated in less than a couple of minutes..

With all the promise of creating new content quickly, many questions have also come up, without clear answers.

The first – also a topic of interest on aithougts.org – is that of ethics.

Whether it is deep fakes – btw, I had experimented with a technology that could have been used for this – when I was looking for tools to simplify podcast editing – on a platform called Descript – where I could train the model with my voice.. I had to read a predefined text for about 30 minutes – and then, based on written text, it could synthesize that text in my voice.. At that time, the technology was not yet as mature as today and so, I did not pursue.

I digress..

Getting back to the debate on generative AI, ethics of originality [I believe that there are now tools emerging that can check if content was generated by ChatGPT!] that could influence how students create their assignment papers.. or generate more marketing content, all based on content that is already available on the net – and ingested by the ChatGPT transformer.

Another aspect is the explainability of the generated content. The bias in the generated content or when there is a need for an expert opinion to also be factored in, would not be possible unless the source is known. The inherent bias in the training data is difficult to overcome as much of this is historical and if balanced data has not been captured or recorded in the past, would be very difficult to fix, or at least adjust the relevance.

The third aspect is about the ‘originality’ or ‘uniqueness’ of the generated content – let me use the term solution from now on..

There is a lot of work being done in these areas, some in research institutions and some in companies applying them in specific contexts.

I had an opportunity recently to have a conversation with the founder of a startup that is currently in stealth mode, working on a ‘domain aware, large language model based’ generative AI solution.

A very interesting conversation that touches upon many of the points as above.

 

You can listen to this conversation as a podcast in 2 parts here:

https://pm-powerconsulting.com/blog/the-potential-of-large-language-models-with-steven-aberle/

https://pm-powerconsulting.com/blog/episode-221/

 

Or watch the conversation as a video in 2 parts here:

https://www.youtube.com/watch?v=86fGLa9ljso

https://www.youtube.com/watch?v=f9DnDNUwFBs

 

Do share your comments and experiences with the emerging applications of GAN, Transformers etc.

Are You Human? Tale of CAPTCHA

Recently I gave a keynote speech in Mahindra University, Hyderabad as part of a 2-day workshop on “Data Science for the Industry”. Great opportunity to share my thoughts on Data Sciences/AIML technologies and industry use cases.

I talked about various problems to be solved by these rapidly advancing technologies. One of them was “Are You Human?” question. The problem is created by AIML technology and also solutions need to come from AIML technology. Basically, how does any IT system distinguish between humans and machines during transactional interfaces?

Is this problem important enough to worry about? Yes. I will give you both technical and commercial reasons for it.

First Commercial reason.

I am sure all of you heard of the Twitter take-over bid by Elon Musk, CEO Tesla for US$44 Billion. The deal was cancelled by Elon Musk due to inability to determine % of non-human or BOT users on Twitter. Elon Musk accused Twitter of using incorrect algorithms to determine BOT users and under estimating the real BOT numbers. The issue is now in legal dispute.

Many commercial decisions are based on number of customers. For e.g., number of people visiting the website determines cost of advertisements and royalty payments to the website content authors.

Second Technology reason.

The age of Digital has transformed IT landscape across the enterprises and use of web, mobile phones, chatbots and IOT devices are the norm and not exceptions. All of these channels are communicating with the enterprise IT systems and getting business executed i.e. placing orders for products, registering service issues etc. At the same time automation is also become a norm and Robotic Process automation tools are widely used in enterprises. Many cases they use various technologies to simplify data entry by using a single screen input and on the background simulating multiple screens data entry to various enterprise systems. These interactions are legitimate and should be flagged as non-human but legitimate approved interfaces.

I am sure now you are convinced about the importance of the problem.

 

Now let us come to main topic of our blog i.e., CAPTCHA.

All of us have used on-line or mobile Banking to do banking transactions. Most of us would have encountered some thing called CAPTCHA. The system throws a set of characters twisted in a wavy curvy fashion and system expects the interacting person to see the image and do the right interpretation and enter it back to the system for confirmation. Some examples are given below.

 


 

The system generates random sequence of case sensitive alpha-numeric characters such as 263S2V. This is twisted in to an image as you see above and shown back to the interacting agent. It is assumed that automated systems will fail to interpret this correctly and only human can interpret and type back the same set of characters 263S2V.

What is full form of Captcha? “Completely Automated Public Turing Test to tell Computers and Humans Apart”.

When was this invented? Between 1997 to 2003. The most common type of CAPTCHA (displayed as Version 1.0) was first invented in 1997 by two groups working in parallel. In 2000, CMU professors Luis Von Ahn, Manuel Blum and John Langford wrote a paper titled “Telling Humans and Computers Apart (Automatically) or How Lazy Cryptographers do AI”. The term CATCHA was coined in 2003 by Luis von Ahn, Manuel Blum, Nicholas J. Hopper and John Langford.

This form of CAPTCHA requires someone to correctly evaluate and enter a sequence of letters or numbers perceptible in a distorted image displayed on their screen. Because the test is administered by a computer, in contrast to the standard Turing test that is administered by a human, a CAPTCHA is sometimes described as a revised Turing test.

I am sure you are wondering how come a 20-year-old technology is still in use for Hitech Banking industry as a digital security solution?

This technology is very cumbersome and frustrating for all humans. Younger people with sharp 20-20 eye vision may be able to get it right first time but it still adds 10-15 seconds to the transaction. Senior citizens, people with glasses, people with vison disability, people with low quality display units, poor lighting in capital or not with all the wavy and squiggly images. In addition, the systems are badly designed and it forces me to re-enter all the data fields till I get my Captcha right.

In the last 20 years, the AIML technology has improved exponentially. Hand writing recognition and image recognition technologies are very good and they can easily recognize the Captcha transformed images. I can go one step further and say that if a senior citizen customer gets the CAPTCHA first time right, then the bank should assume it a fraud!. Unfortunately, the banks assume the exact opposite, which was the original basis of the CAPTCHA technology.

Various other ideas such as speed of data filling were also considered as part of CAPTCHA. Humans do take time to type the data while automatic BOTS can do it at super-fast speeds. However, RPA based automation systems will always be fast and they are genuine systems interactions. Also, it is so easy for a BOT to slow things down by waiting few seconds before submitting the data and fool the timing algorithm.

We have seen big discussions on evolution race between the prey and the hunter in the biological world. The deer evolves to get stronger legs to out run the tiger. Tiger evolved stronger lungs to sustain long chases. Same way as the AIML technologies evolving so fast to mimic human interactions, we need to get better technologies to solve the “Are You Human?” problem.

 

More later. Do share your views.

 

Regards,

L Ravichandran

AiThoughts.Org

 

 


 

Test your AI Quotient

Take this fun quiz to find ten words related to the world of AI.

These may be acronyms or terms that you would come across while exploring the wide world of Artificial Intellingence, Machine Learning, using techniques, applications etc.

For this puzzle, there may also be phrases – or multiple words – that have been included without a separating space.

Example: Machine learning would be machinelearning

The words may be horiontal, vertical, diagonal or, sometimes, in reverse.

Since this is a dynamic and interactive puzzle, if you play multiple times, the grid may appear differently.

This is the first in many more such interactives you will see on this site.

So, to make it easy, the vocabulary is also shown below the grid – to help you look for them.

If you have any suggestions for other interactive and fun ways to explore the world of AI, do contact us and share your ideas.

All the best!

[h5p id=”1″]

AI becoming Sentinel

Google CEO demonstrated their new Natural Language chatbot LAMDA.  The video is available on youtube. https://www.youtube.com/watch?v=aUSSfo5nCdM

The demo was very impressive. All the planets in the solar system were created as personas and any human can converse with LaMDA and ask questions about that particular planet.  LaMDA responses had sufficient human like qualities. For e.g. If you talk good about the planet then it says thanks for appreciating and when you talk about myths about the planet, it corrects you with human like statements.  Google CEO also mentioned that this is still under R&D but being used internally and this is Google’s efforts to make machines understand and respond as humans using natural language constructs.

Huge controversy was also created by a Google engineer, Blake  Lemoine.  His short interview is available on Youtube. https://www.youtube.com/watch?v=kgCUn4fQTsc&t=556s

Blake was part of testing team of LAMDA and after many question & answer sessions with LAMDA, felt that LaMDA is becoming a real person with feelings, understanding of trick questions and answering with trick or silly answers like a person would do etc.  He asked a philosophical question “Is LaMDA sentinel? “

Google management and many other AI experts have dismissed these claims and questioned him on his motives for over playing his cards.

In simple terms let me summarize both the positions.

  • Google and other big players in the AI space are trying to crack the Artificial General Intelligence ( AGI) area i.e how to make AI/ML models as human as possible. This is their stated purpose and there is no question of denying this.
  • Any progress towards AGI will involve machines to behave in irrational ways as humans do. Machines may not always chose the correct decision all the times ..  may not want to answer the same question many times like humans do ..  may show signs of emotions such as feeling hurt , sad , happy etc. like humans do.
  • This does not mean that AI has become sentinel and has actually become a person demanding its rights as a global citizen!.  All new technologies have rewards and risks and may be we are exaggerating the risks of AI tech too much.
  • Blake gave an example of one test case during his testing role at Google.  He tried various test conversations with LaMDA to identify ethical issues like bias etc.  When he gave a trick question to LaMDA which had no right answer, LaMDA responded back with a real stupid out of the line answer.   Blake reasoned that LaMDA understood that this was a trick question, deliberately being asked to confuse LaMDA and hence gave a out of the line stupid answer. For another question “what are you afraid of”, LaMDA said it is afraid of being turned off.  He felt these answers are way and beyond just conversational intelligence and hence felt that LaMDA has become more of a person.
  • You may refer my earlier Blogs on Turing test for AI.  Prof Turing published this test in 1953 to determine whether an AI machine has full general intelligence.  Blake also wanted Google to run this Turing test on LaMDA and see if LaMDA passes or fails this.  He says Google felt this is not necessary. He also claims that as per Google’s policy, LaMDA is hard coded to fail the Turing test.  If you ask a question “Are you an AI” , LaMDA is hardcoded to say Yes thus failing the Turing test.

Very interesting thoughts and discussions.  Nothing dramatic about this as AGI by its definition very controversial as it gets in to deep human knowledge replication.

What do enterprises who are planning on using AI/ML need to do? 

For enterprise applications of AI/ML, we do not need AGIs and our focused domain specific AI/ML models are sufficient.  Hence no need to worry about these sentinel discussions as yet.

However, the discussions on AI Ethics are still very relevant for all enterprise AIML applications and not to be confused with the AGI sentinel discussions.  

More Later,

L Ravichandran.

EU Artificial Intelligence Act proposal

Lot has been talked about #ResponsibleAI, #ai and #ethics. We also have a brand new filed called #xai Explainable AI with the sole objective of creating new simpler models to interpret more complex original models.  Many tech companies such as Google, Microsoft, IBM etc. have released their #ResponsibleAI guiding principles. 

European Union has circulated a proposal for a “The EU Artificial Intelligence Act”.  As per process this proposal  will be discussed, debated, modified and made in to law by the European parliament soon.

Let me give you a brief summary of the proposal.  

First is the definition of 4 risk categories with different type of checks & balances in each category. 

The categories are  

  1. Unacceptable
  2. High Risk
  3. Limited Risk
  4. Minimal Risk

Category 1 the recommendation is a Big NO.   No company can deploy this category SW within EU for commercial use.

Category 2 consisting of many of the business innovation & productivity improvement applications will be under formal review & certification before put to commercial use.

Category 3 will require full transparency to the end users and option to ask for alternate human in the loop solutions.

Category 4 is not addressed in this proposal.  Expected to be self-governed by companies

Let us look at what kind of applications fall in Category 2

  • Biometric identification and categorization
  • Critical Infrastructure management
  • Education and vocational training
  • Employment
  • Access to public and private services including benefits
  • Law enforcement ( Police & Judiciary)  
  • Border management ( Migration and asylum)
  • Democratic process such as elections & campaigning.

Very clearly EU is worried about the ethical aspects of these complex AI systems with their inbuild biases, lack of explain ability, transparency etc. and also clearly gives very high weightage to human rights and fairness & decency.

 I recommend that all organizations start reviewing this and include this aspect in their AIML deployment plans without waiting for the eventual EU law.