Lorem ipsum dolor sit amet, consectetur adipiscing elit. Suspendisse varius enim in eros elementum tristique. Duis cursus, mi quis viverra ornare, eros dolor interdum nulla, ut commodo diam libero vitae erat.
Lorem ipsum dolor sit amet, consectetur adipiscing elit. Suspendisse varius enim in eros elementum tristique. Duis cursus, mi quis viverra ornare, eros dolor interdum nulla, ut commodo diam libero vitae erat.
Lorem ipsum dolor sit amet, consectetur adipiscing elit. Suspendisse varius enim in eros elementum tristique. Duis cursus, mi quis viverra ornare, eros dolor interdum nulla, ut commodo diam libero vitae erat.
Lorem ipsum dolor sit amet, consectetur adipiscing elit. Suspendisse varius enim in eros elementum tristique. Duis cursus, mi quis viverra ornare.
Lorem ipsum dolor sit amet, consectetur adipiscing elit. Suspendisse varius enim in eros elementum tristique. Duis cursus, mi quis viverra ornare.
Lorem ipsum dolor sit amet, consectetur adipiscing elit. Suspendisse varius enim in eros elementum tristique. Duis cursus, mi quis viverra ornare.
Want to see us in action? Browse our recent projects and discover more about the transformative power of AI.
Lorem ipsum dolor sit amet, consectetur adipiscing elit. Suspendisse varius enim in eros elementum tristique.
Like the monolith from 2021: A Space Odyssey, AI is being touted as something that will herald a new era of computational power and human capability. The things that AI is already able to do are truly mind-blowing - and it’s still very much in its infancy.
But let’s take the rose-tinted glasses off for a moment. We are all staring down the barrel of a full-blown, man-made climate crisis. Technology is helping us mitigate its impacts somewhat, with improvements in fields like renewable energy offering glimmers of hope.
AI holds the potential to make our lives easier and more efficient. However, AI’s relationship with energy consumption is complicated. At worst AI is a huge energy and resource hog - at best, AI professionals are working on mitigating that fact.
Pinning down the exact carbon impacts of AI, and especially of specific Large Language Model (LLM) AIs like ChatGPT, is tough. But energy use - and misuse - is totally under human control. It’s our responsibility as humans who use AI tools to do so ethically and mindfully.
Before we explore the ways we as AI practitioners can redress this much-needed balance, we need to fully examine some of the ecological worries inherent when using AI, especially Large Language Models.
AI - and especially elements like deep learning and neural networks - requires much more computational power than standard computing. The sheer amount of information storage and data crunching inherently needed by AI consumes far more energy than any kind of computing that has ever come before.
The wide-reaching, multi-purpose AI tools that make the headlines, like ChatGPT and Gemini, are called “large language models”. To vastly simplify, these tools maintain access to trillions of data points that they can draw upon when they give a response, and contain countless parameters to help them score and link those data points; hence why their responses seem so natural and human.
But maintaining access to those countless data points, parameters, and the inferred links between them - and growing its access to millions more data points every day - takes a lot of energy, infrastructure, and hardware. And that’s before we get to the energy consumption involved in answering the countless requests they receive every 24 hours.
This becomes especially concerning when you zoom out and think about the vast data centres needed to operate cloud tools of all kinds, not just AI ones. Data centres constantly consume energy. Their servers are running 24/7, as are the routers and switches that get data where it needs to go. IT equipment needs to be cooled, so air conditioners are always running. Their employees have to commute to and from them every day, likely using combustion engines of some kind.
With energy supply around the world still so reliant on fossil fuels, this is understandably a significant draw on carbon resources. (Though it’s worth noting that both Google and Microsoft have made strides in their efforts to become carbon neutral/negative.)
Research from University of Massachusetts, Amherst found that merely training a large language AI model emits approximately the same amount of CO2 as 5 average cars across their whole lifespan. That’s just training the thing - once the LLM is actually put to use, it could end up consuming orders of magnitude more energy than that throughout its existence.
In order to build new computer hardware, manufacturers need to mine rare earth elements (REEs) and minerals. This strips the Earth of its natural finite resources; and once they are mined, they still need to be transported, processed, and then manufactured into whatever board or device they are destined to become - all of which may still rely on fossil fuels and polluting practices.
Also well worth mentioning here is the continued labour and human rights abuses taking place throughout technology supply chains.
Heading back to AI for a moment, computer hardware undergoes quite intense usage when running large language models day in, day out. This all chips away at the hardware’s natural lifespan. When that hardware fails and needs to be replaced, that perpetuates the demand for more resource stripping and potentially sends more e-waste to landfill.
Sadly, many view AI as the latest tech toy. And who can blame them? Interacting with publicly available LLM AI tools like Chat GPT is deceptively simple. With similar effort to a Google search, users can generate new text, stories, art, or code; get answers to questions; research and hash out ideas; and much more.
As such, both individuals and businesses alike are already using LLM AI tools alarmingly frivolously, such as asking their LLM of choice what the weather is up to or what’s the best air fryer to buy - things that would usually be the subject of a normal web search.
And though typing a prompt into a generative AI tool may appear very similar to a web search, each LLM AI query comes at a far higher carbon cost than anything generated by standard computing. Research suggests that a ChatGPT query consumes around 60 times more energy per query than a simple Google search.
But this isn’t just a case of ease - it’s a case of trendiness too. Organisations want to appear ahead of the curve by adopting AI and LLM tools, because it’s the hot new tech. So some end up doing so just for the sake of it; without a meaningful use case that justifies the need for such a complex, heavy, and potentially polluting computational tool.
In computer programming, there's usually more than one way to code a solution to a problem. Sometimes code is kept efficient and lean, only carrying the bare minimum instructions needed to carry out the functions in question. Other times, code becomes bloated and inefficient, full of exceptions and weird sticking plaster workarounds.
When code is poorly optimised, it takes more energy and computing power to navigate the inefficient twists, turns, and dead ends. Code that is well-optimised is far more energy efficient.
This is true of any software, not just tools that rely on AI. But if your software does involve AI, you need to make sure your AI is targeted precisely to your specific use case and the code doesn’t make erroneous or frivolous AI requests.
Simply moving data around also consumes energy. A data packet travelling from LA to London, through various routers and data centres, is going to have a larger carbon impact than one going from Liverpool to Manchester. So think: does a computation have to happen on some distant server somewhere and then be transported to you? Or can that process take place within your own network? Or perhaps directly on the relevant device? There’s no need in making your data move further than it has to!
In fact, it’s already being used to great effect in the fight against climate change. Let’s acknowledge some of the good work that is already happening in the eco-AI space:
There are many sensible ways SMEs can benefit from AI and be as kind as possible to the planet. Here are a few things you can bear in mind before embarking on your next AI tech transformation project:
We get it, AI is trendy. Companies want to boast about how their tools are using the latest technology buzzword as it makes them seem cutting edge. Yet this trendiness can lead to AI solutions being applied in situations where standard computation would have worked equally well.
Our advice? Maintain a single-pointed focus on the specific use cases you need from your new technical solution and what problems you need it to solve. If that involves AI, then great! We’re here to help. But don’t try to shoehorn in a particular kind of tech where it might not be needed.
Not all AIs are created equal. Yes, the large language models make the headlines, but not all AI implementations require such massive, wide-reaching datasets.
Small language models (SLMs) are artificial intelligence tools which rely on much smaller, more targeted data sets. This makes them far narrower in scope, but also often far more carbon efficient than their LLM cousins.
SLMs can be perfect for applications that require some of the more flexible and artificially creative elements of AI, but only need a limited focus; for example an AI-powered website chatbot that only needs to know details about a company’s product catalogue.
In contrast, using a vast LLM like ChatGPT to power a simple website chatbot would be like using a sledgehammer to crack a nut!
Large language, generative artificial intelligence is very different from any kind of computing that has come before. It requires a huge amount of data crunching, which in turn requires a lot of hardware and energy resources to maintain.
Keep this in mind when deciding to use AI, and don’t get carried away with using overpowered LLM AI solutions when a SLM or more conventionally programmed solution would suffice.
But here’s one thing you and your teams can do in the here and now: don’t overuse generative LLM tools for frivolous things where a simple web search would suffice!
Whether your solution uses AI or not, aim to design the most energy and computationally efficient solution. For our more tech savvy readers, this might mean optimising your code so it doesn’t run into errors, exceptions, and dead ends.
But creating an efficient solution truly starts at the ideation phase. Does that data point really need to be fetched from the other side of the planet? Do all of our networked devices need to be set up to carry out this complex type of calculation, or will one or two machines suffice? Does the software really need to speak to an AI in order to do [X]? Do we really need new hardware when the old stuff is still up to spec?
We wholeheartedly welcome the fact that an increasing number of data centre operators are committing to use only renewable energy in their data centres. Alas, despite these noble efforts, it still doesn’t mean that the whole supply chain is rendered magically spotless.
The tech hardware supply chain is reportedly slow to address human rights abuses. Rare earth elements are expensive and polluting to extract - and only available in finite quantities.
Sadly this means that no tech supply chain is ever going to be 100% clean. But here’s another tip you can start right now: research your supply chains and do what you can to ensure suppliers align with your principles.
Right now, AI definitely helps to expand human throughput, which is a great thing. But, despite the hand wringing in the headlines and the shareholder-pleasing statements coming out from "Big AI", machines cannot think like humans yet, and can’t make decisions on what is right or wrong. They’re not even close.
AI certainly magnifies our ability to make positive impacts in the world. But it magnifies our ability to make negative impacts too. Whether those negative impacts are cruelly negative on purpose, or come about from well-meaning misunderstanding or error.
This is why it's so important for us as decision-making humans to carefully consider the impacts of our use of technology before we rush into using AI simply because it's the latest tech trend.
Thankfully, AI is still in its infancy, and we have time to make it a much less polluting, efficiency-driving force for good.
But it’s up to us as humans and decision makers to create that future.
Data has traditionally been collected and saved in databases, often relational databases, which have the capability to store large amounts of data. However, these databases have limitations due to the complex nature of data and its connections in the real world.
To overcome these limitations knowledge graphs are used. Knowledge graphs offer a novel approach to data storage whilst accounting forthe complex relationships in data. This results in easily accessible data, where it is possible to uncover hidden features and find new insights from your data.
Knowledge graphs are models of data about a certain topic. These topics can be anything where data can be collected, such as people across multiple organisations, products for sale in a business or movies, actors,directors and how they are all connected. These models allow us to visualise the way connections are made when the data is used in the real world.
A knowledge graph composes of nodes, edges and properties. Edges and nodes are crucial to a knowledge graph whilst properties provide additional information.
· Nodes are usually entities, such as people, organisations or products.
· Edges are the relationships between nodes. Relationships could be between two nodes describing people such as ‘related to’ or ‘employed by’.
· Properties can be any further information about a node and properties can vary depending on the node type. Properties donot link to the edges.
When we combine nodes, edges and properties we have a knowledge graph!
A simple knowledge graph example is a movie database. This type of database can be shown in a straight forward way whilst still containing the complexities of the relationships involved.
If we consider our knowledge graph components:
· Nodes – People, Movies, Directors,Actors, Genres
· Relationships – ‘Watched’, ‘Directed By’, ‘Acted in’
· Properties – Age, Run time, Release Date, Number of movies directed
An example of a small section of a Movies knowledge graph is visually displayed below. This simple knowledge graph contains the key components previously described.
This example shows which movies Alice and Bob have watched, what genre they are in, who directed them and who acted in them. There are manyreasons why this information is important and how it can be used, but we willget to that later…
Representing data in a knowledge graph provides contextual understanding that may not be possible in a traditional database. The power of a knowledge graph becomes clear when trying to follow connections between data points to retrieve information. Graph queries can take a tiny amount of time to perform this compared to retrieving the same data from a relational database.Not only does the faster search provide huge benefits but the flexibility of a knowledge graph enables the use of complex algorithms to uncover insights in your data and provide real world solutions.
A further benefit gained from knowledge graphs is the flexibility and scalability. These graphs can be edited to include new information easily without affecting other data entries. Knowledge graphs also store data efficiently resulting in a data store that can scale to hold huge amounts of information. For example, one of the most commonly used knowledge graphs can be found on Amazon, linking every product sold in order to improve searchability and recommendations on a huge scale.
Graph databases can be used in a wide variety of use cases, each scenario benefits from a different aspect of a knowledge graph.
We can revisit our previous example of a movie dataset stored in a knowledge graph. The data stored can be used to connect likes, dislikes and other data to provide a complex and effective recommendation system. Our example can be extended to show this.
Knowledge graphs have been used to detect fraudulent transactions between groups of people and organisations. Knowledge graphs arekey to the success of this as anomalies can be traced through the graph to the intended recipient.
Knowledge graphs can enhance search engines by providing context and semantic knowledge. This contextual knowledge results in more accurate and personalised search results, which in return will lead to a better service for the customer.
To conclude, knowledge graphs are a powerful tool that allows the user to uncover previously hidden insights in their data. Representing data in a graph form rather than a traditional table gives improved and additional use cases such as fraud detection, recommendation systems, semantic search and much more! However, in this introduction we have only covered the basics of knowledge graphs. We will have to return to delve deeper into their true potential.
In an era where safeguarding data privacy and security are paramount, the convergence of privacy preserving machine learning and IoT technology presents both immense opportunities and daunting challenges. The need to prioritise trustworthy AI through development that is robust, explainable and transparent is an urgent priority.
Recognising this pivotal moment, OctaiPipe, a pioneer in federated edge AI innovation, has joined forces with industry leaders P&G, SME partners Site Assist, Raiven and Digital Catapult and InnovateUK to deliver a groundbreaking SaferAI concept. groundbreaking concept for safer AI
“Bringing together SMEs, large organisations and research organisations, these novel solutions will demonstrate how trusted AI and machine learning technologies can aid and be incorporated into many of the UK’s industries and sectors.”
Dr Kedar Pandya, UKRI Technology Missions Fund Senior Responsible Owner
“Federated Learning with OctaiPipe is unlocking potential not yet realized in industry. It’s enabling competing business to collaborate on the use of technology and AI for good. Specifically, within SaferAI, I can see significant value in utilizing federated learning to share model results for predicting safety incidents and risks. I can see this changing approaches in industrial safety. Before, we could only cooperate on best practices, technology or processes, but now we can actually prevent incidents with the use of AI and shared data.”
Christoph Wagner-Gillen, P&G Product Supply HSE Governance
SaferAI, or as the projects catchier title suggests, Safety advancing federated estimation of risk using AI, is a 12-month, £1.64m project focused on the development of a highly comprehensive federated learning model for predicting incident occurrences and risk.
SaferAI is an industry first, enabling world leading consumer goods manufacturer, P&G, to collaborate with other interested Consumer Goods Companies to forecast safety incident rates to reduce Health & Safety (H&S) incidents.
Until now, with low event occurrence in any single setting, there has been insufficient data availability to predict future incident events or near-misses. However, using federated learning with OctaiPipe, multiple large-scale parties with similar Health, Safety and Environment (HSE) situations can now collaborate to pool intelligence to achieve scale.
Many organisations already collect high-level operational and HSE incident data intelligence through various Industrial IoT devices and cameras to successfully predict safety incidents. However, analytics based on this is not meaningfully actionable to drive changes that preventatively reduce risks.
For predictions to be meaningfully actionable, they must be made at a sufficient level of granularity within a factory or workgroup.
Whilst we’re fortunate that HSE critical events are rare within single sites or even organisations, it means insufficient data exists to employ ML models capable of predicting when and why H&S incidents might occur so they can be prevented.
To date, progress towards an AI-enabled solution has been impeded by:
a) a requirement for more observations than one organisation can generate alone, so it is imperative to share data, and
b) barriers to sharing data across organisations that, until now, have not been overcome.
Federated learning solves this. SaferAI will enable organisations to combine data to facilitate actionable incident predictions for small work groups.
Federated Learning is a privacy preserving machine learning technique that allows for machine learning models to be trained across multiple distributed edge devices without the need to see or move the data.
It is a key technology in the convergence of AI, IOT and connectivity as it enables edge intelligence whilst safeguarding sensitive or private data.
Federated Learning will be used the project to establish a viable on-device (Edge) AI system that pairs privacy-focused telemetry and computer vision (CV) systems. It will also enable continual collaborative learning in edge environments, combining model results provided by collaborating organisations to yield a highly comprehensive HSE model.
During SaferAI, OctaiPipe are further developing and validating as a high-trustworthiness FL-for-IoT system along with a high-accuracy model that uses observed use case data generated by AIoT camera systems. This solves the challenge of predicting safety critical incidents, near misses and hazardous/non-compliant conditions in industrial settings, enabling organisations to pre-emptively take remedial/mitigating actions to reduce H&S risks.
Michelle Donelan
Secretary of State for Science, Innovation, and Technology
The consortium’s mission is to help solve the challenge of implementing trustworthy AI-in-IoT. This will be achieved by accelerating the development of a federated, secure, privacy-preserving, and auditable AI-for-IoT platform optimised for machine learning in IoT and edge systems. OctaiPipe is a first-of-its-kind innovation that combines privacy-preserving machine learning technology, cyber security, continuous collaborative learning, and AI lifecycle management. This will allow IoT-enabled businesses to build, deploy, and manage machine learning software that guarantees the privacy and security of device data and its use, allowing the user to have a high degree of trust in the AI solutions embedded in them.
Raiven will be building a predictive maintenance model on the OctaiPipe platform with the support of P&G. The model will aim to forecast safety incident rates to reduce Health & Safety (H&S) incidents. Raiven will be exploring new Data Science methodologies that allow organisations to work collaboratively and privately, by exploring data abstraction and alignment techniques.
Whilst Site-Assist will identify non-HSE compliant environments and situations in security-sensitive settings, i.e., critical infrastructure to apply the technology and Digital Catapult will assess the AI/ML explainability and verifiability of the solution; driving optimisation and new standards.
OctaiPipe is a revolutionary Federated Edge AI-for-IoT platform that learns the AI model on the edge device, meaning raw data is always kept at source and private. Captured model parameters are then distributed to centralised Federated Learning (FL) server in real-time (<5ms), where contributions are aggregated into a single global model before being fed back to edge devices to benefit from globally aggregated learnings of 1,000s devices. This delivers a unique market offering connecting the customer to Artificial Intelligence of Things (AIoT) that is more private, secure, efficient, and autonomous—enabling scale through automation of trust.
Procter & Gamble (P&G) is one of the world’s largest fast-moving consumer goods companies and home to iconic, trusted brands, including Ariel, Lenor, Flash, Pampers, Always, Pantene, Herbal Essences, Oral B and Gillette. With a large global footprint of 70 countries and with 5 billion consumers worldwide, the design, development, growth and success of these products is driven by innovation and insight of its employees.
Digital Catapult is the UK authority on advanced digital technology with an existing track record of delivering over 50 collaborative R&D projects looking to advance the adoption of advanced digital technologies. Through collaboration and innovation across our specialist programmes and experimental facilities, we accelerate industry adoption of advanced digital technologies to drive growth and opportunity across the economy, making sure that innovation thrives, and the right solutions make it to the real world.
Raiven was formed to bridge the skills and technology gap in industry, and to enable technology adoption across industries. Raiven’s expertise in Artificial Intelligence and Data Science ensure innovations are closely monitored and continue to deliver business value. Raiven aims to be a trusted expert that your business can turn to for advice, strategy and delivery, with a focus on implementing responsible A.I. solutions that are pragmatic, people-centric, ethical and results-oriented.
Site-Assist (SME) is a software service provider of app-based solutions for collecting HSE and productivity data on-site—enabling risk mitigation, supporting compliance, enhancing time efficiency/productivity, and minimising carbon impacts. Site Assist were incorporated in early 2021 when an opportunity for a market entry product was identified around permitting, which was developed and deployed accordingly. Adopted by Balfour Beatty and HS2. Subsequently, other large players have adopted the software to varying degrees; Babcock, AWE, Hinkley Point C, Morgan Sindall, Emcore inter alia.
No matter your technological know-how, we’re here to help. Send us a message or book a free consultation call today.