Artificial Intelligence is transforming industries at an unprecedented pace, but without robust safety measures, its potential risks could outweigh its benefits. From ethical concerns to system reliability, AI safety is now a critical global conversation. How can businesses and governments ensure AI remains a force for good?
AI safety refers to the ideas, policies, and actions designed to ensure that artificial intelligence technology works in a safe, trustworthy, and useful manner for mankind.
A recent study, “The Critical Conversation on AI Safety and Risk” highlights that aligning AI systems with human values is fundamental to AI safety, as is ensuring that these systems are dependable, scalable, and responsible. This involves technology safeguards, legislative frameworks, industry standards, and ongoing research into AI safety protocols.
In early 2024, AI-generated deepfake audio recordings of UK political leaders surfaced online, falsely depicting them in controversial conversations. These deepfakes, created with sophisticated generative models, spread rapidly on social media, raising concerns over election security and public misinformation.
This incident underscores the need for AI governance, digital ethics, and proactive measures against AI-enabled cyber manipulation. It highlights the challenge of balancing innovation with safety and ensuring AI systems are not misused for harmful or deceptive purposes.
What does this example suggest?
The connection between innovation and safety is the key. It is significant to bridge cutting-edge developments with the ethical imperative to avoid undesirable outcomes.
Effective AI safety plans include active involvement with civil society, ongoing risk assessment, and reaction to developing concerns. Nevertheless, the process of creating and implementing safety standards in this quickly changing industry is quite complex
The Connection Between Innovation and Safety
AI safety in action: Cross-sector strategies
AI safety is critical across various industries, each implementing stringent safeguards:
- Aviation & healthcare: Regulatory bodies enforce rigorous certification processes and clinical trials to ensure AI applications meet high safety standards before deployment.
- Cybersecurity: AI-driven security systems are tested through ‘red teaming’ exercises, where external experts stress-test models to identify vulnerabilities.
- Media & content authentication: Organisations like OpenAI have adopted Coalition on Material Provenance and Authenticity (C2PA) standards to verify digital content authenticity and combat misinformation.
- Smart infrastructure: AI-powered urban planning tools, such as Singapore’s Smart Mobility Initiative, leverage machine learning to enhance road safety and reduce congestion.
The United Nations AI for Good Initiative promotes responsible AI adoption in sectors such as disaster response, healthcare, and climate change mitigation, demonstrating AI’s potential to address global challenges while ensuring safety.
Similarly, AI organisations such as OpenAI have implemented techniques such as incorporating Coalition on Material Provenance and Authenticity (C2PA) standards to authenticate the provenance of digital material, reducing the dangers associated with AI-generated media. Another technique is ‘red teaming,’ in which external specialists stress-test AI systems to identify flaws, yet this is still a developing field.
The Hawaii Department of Transportation’s proactive strategy, which includes AI for activities ranging from traffic control to human resource operations, demonstrates AI’s potential to save lives and increase operational efficiency.
The Driving Transportation Safety Forward with AI report highlights how AI-driven systems such as Traffic Incident Management enabled by large-data innovations (TIMELI) can leverage AI to improve safety by predicting and mitigating traffic incidents.
AI Alliance and EU AI Act
The European AI Alliance, launched in 2018, is a European Commission initiative fostering open discussions on AI policy. It unites diverse stakeholders—including academia, businesses, and civil society—to shape AI regulations aligned with European values.
The Alliance focuses on:
- Ethical AI guidelines: Developed in partnership with the High-Level Expert Group on Artificial Intelligence (AI HLEG), these guidelines ensure AI systems are lawful, ethical, and robust.
- Policy and investment: The Alliance provides insights for AI-related legislation and investment strategies.
- Community engagement: With over 6,000 stakeholders participating in consultations and discussions, the Alliance influences AI policy formation.
EU Artificial Intelligence Act (AI Act)
The AI Act, which came into force on August 1, 2024, stands as the world’s first comprehensive legal framework governing AI. Its primary aim is to ensure that AI systems used within the EU are safe, respect fundamental rights, and align with European values. The Act adopts a risk-based approach, categorising AI applications based on the potential harm they might pose:
- Unacceptable Risk: AI applications that pose a clear threat to safety or fundamental rights are prohibited. This includes systems that manipulate human behavior or enable social scoring by governments.
- High Risk: Applications in critical sectors such as healthcare, education, and law enforcement are subject to stringent obligations, including rigorous risk assessments and adherence to transparency and oversight requirements.
- Limited Risk: Systems that interact with users, like chatbots, must comply with transparency obligations, ensuring users are informed they are engaging with an AI system.
- Minimal Risk: Applications such as AI-enabled video games or spam filters are largely unregulated under the Act, given their low impact on users’ rights or safety.
The AI Act also introduces measures to address general-purpose AI systems, including foundational models like ChatGPT. These systems are subject to specific transparency requirements, especially when they pose significant risks. The Act’s extraterritorial scope means that providers and deployers outside the EU must comply with its provisions if their AI systems impact individuals within the EU.
What are the key pillars of AI safety?
Key aspects for redesigning ethics for times of AI are: Data privacy, fairness, explainability, transparency and accountability.
Data Privacy
Data privacy, often known as “information privacy,” refers to the notion that individuals should have control over their personal data. This principle involves determining how organisations acquire, keep, and use their data.
Data privacy involves providing individuals with control over their personal data, and it has grown in importance as digital technologies have evolved, particularly in the context of artificial intelligence. AI systems sometimes require massive volumes of data to work well, raising questions about how personal information is stored and secured.
Organisations have a legal and ethical need to follow these principles, ensuring that data subjects—the people who own the data—have control over their information.
The General Data Protection Regulation (GDPR), an important piece of legislation in this field, requires businesses to create policies and systems that protect individual rights, even in the absence of official data privacy regulations. Organisations and companies can profit from implementing strong data privacy safeguards, which not only help with compliance but also increase consumer confidence.
Data security solutions such as encryption, automated policy enforcement, and audit monitoring are critical components for organisations to comply with rules and protect personal data.
Fairness
In the context of AI and machine learning, fairness refers to the systems’ impartial and equal treatment of all people, ensuring that no one is discriminated against based on race, gender, age, or sexual orientation. But what is fairness?
In “What does “fairness” mean for machine learning systems?” Smith states that fairness is frequently characterised in machine learning as the characteristic of being fair or unbiased; however, this definition might vary depending on the subject. AI researchers and practitioners try to design models that not only perform effectively but also treat individuals justly.
Fairness must take into consideration how AI systems allocate and withhold resources. Fairness in AI entails detecting possible biases upfront and ensuring that various perspectives, including domain experts, are included in the discussion. The objective is not to construct a completely fair system but rather to detect and minimise fairness-related problems to the greatest extent feasible.
Google’s What-If Tool enables users to investigate model performance across datasets while analysing fairness restrictions, whereas IBM’s AI Fairness 360 Toolkit offers technical solutions via fairness measurements and algorithms to help discover and minimise bias in models. The algorithm’s goal is to properly forecast biases for various individuals.
Transparency
Transparency ensures that AI systems are not only understandable but also responsible to the people with whom they interact. In a Forbes article “Examples that demonstrate why transparency is critical in AI” Bernard believes it is vital to assess the clarity of AI algorithms, the data sources they use, and the decision-making processes they utilise.
For AI to be considered transparent, users and stakeholders must understand how these systems arrive at their findings, ensuring that outcomes are fair, unbiased, and ethical. Various organisations (like Cognisant and others) advocate for the formation of centres of excellence to centralise AI oversight inside a corporation.
This strategy enables the universal use of transparency techniques across all AI efforts, ensuring that AI systems are not only accountable but also intelligible to users and stakeholders, resulting in increased confidence and responsible AI adoption.
When there is a lack of transparency in AI, it can have both positive and negative repercussions. For example, Microsoft’s Python SDK for Azure Machine Learning has a model explainability option that is set to ‘true’ by default in current versions. This feature enables developers to obtain insight into the interpretability of AI choices, ensuring that they are done equitably and ethically.
On the legislative front, the EU AI Act, the first comprehensive regulation on AI by a major regulator anywhere, requires openness for AI systems used in vital applications, with substantial penalties for companies who implement opaque, black-box models.
Explainability
Explainability in artificial intelligence refers to the ability for humans to comprehend AI systems’ processes and judgments. “In “What is Explainable Artificial Intelligence?”, IBM suggests that Explainable AI (XAI) refers to a set of methodologies and procedures that enable users to understand how machine learning (ML) algorithms achieve their results, hence increasing confidence and dependability in these systems.
Unlike standard AI models, which may deliver outcomes without a clear understanding of how they were obtained, XAI guarantees that each choice can be traced and explained. This traceability is critical because it solves the so-called ‘black box’ problem in AI, in which the inner workings of complicated models, particularly deep learning and neural networks, are frequently unknown.
XAI enhances the openness, accuracy, fairness, and accountability of AI models. By making AI more interpretable, XAI encourages responsible AI development by identifying and mitigating biases associated with sensitive traits such as race, gender, and age. As AI systems become more intertwined into numerous parts of society, like healthcare and economics, the significance of explainability grows. It enables organisations to retain trust in AI-driven judgements by providing explicit reasons for actions performed, which is particularly critical in regulated sectors where decisions may have a big influence on people’s lives.
A specific case study demonstrating the value of XAI involves forecasting a patient’s chance of developing diabetes using a machine learning model. Varad Vishwarupe’s case study, “Explainable AI and Interpretable Machine Learning,” used a dataset from a 2021 medical survey and a Random Forest classifier, taking into account clinical factors such as age, skin thickness, insulin levels, and BMI.
Using XAI frameworks and techniques such as SHAP (Shapley Additive Explanations), LIME (Local Interpretable Model-agnostic Explanations), and ELI5, the researchers were able to offer extensive explanations for the model’s predictions. These technologies enabled them to assess the influence of each clinical aspect on prediction, changing a previously opaque decision-making process into one that medical practitioners could readily understand and confirm.
AI Safety Summit, source: Flickr
Explainable AI is important because it bridges the gap between complicated machine learning models and human comprehension, making AI-driven judgements more visible, trustworthy, and actionable.
Accountability
Accountability in artificial intelligence is a complicated but critical issue that entails deciding who is liable when AI systems make bad or destructive choices.
In the article, “Critical Issues About AI Accountability Answered”, the authors suggest that as AI gets more integrated into numerous industries, the subject of responsibility becomes more relevant, especially when these systems have unintended repercussions. Accountability in AI involves more than simply determining who is to blame when anything goes wrong; it also includes ensuring that measures are in place to prevent such occurrences in the first place. This includes creating clear norms and duties for AI developers, consumers, and suppliers.
The qualities of accountability in AI are diverse, including openness, effective human monitoring, and the capacity to challenge AI choices. AI systems are employed in financial organisations to identify potentially fraudulent transactions and assess creditworthiness. However, a lack of transparency in these systems can lead to the unjust treatment of consumers, such as stopped payments or refused credit applications.
Raising awareness: AI’s role in our evolution
We know there is much more than these 5 elements for increased AI safety. And the most important is widespread awareness of the tsunami wave that AI is.
Raising awareness about AI safety is not just a responsibility for policymakers, researchers, or technologists—it is a collective imperative for all of humanity. The rapid acceleration of AI and AGI development demands a proactive approach to governance, ethics, and public engagement. Without widespread understanding and participation, decisions about the future of AI may be shaped by a small few, leaving the majority unprepared for its profound implications.
To navigate this pivotal moment in our evolutionary journey, we must prioritise education, transparency, and inclusive dialogue, ensuring that AI serves as a force for progress rather than an uncontrollable disruption.
AI and AGI are overwriting our evolutionary path as humans and opening new ways, that can be either highways or sinister roads.
Some of us are aware of this but the very truth is that 99% of humanity is more focused on going on with their lives in many cases in survival mode even if in statistics we live at the best stage of human evolutionary social health and economic levels.
Yuval Noah Harari says how “Humans think in stories rather than in facts, numbers, or equations, and the simpler the story, the better.”
But AI models compute, that is, “think” rational facts, numbers, or equations. Is this a good idea then? To just allow rationality to rule our lives? The narrative of Humanity is in our hands.

Dinis Guarda is an author, academic, influencer, serial entrepreneur, and leader in 4IR, AI, Fintech, digital transformation, and Blockchain. Dinis has created various companies such as Ztudium tech platform; founder of global digital platform directory openbusinesscouncil.org; digital transformation platform to empower, guide and index cities citiesabc.com and fashion technology platform fashionabc.org. He is also the publisher of intelligenthq.com, hedgethink.com and tradersdna.com. He has been working with the likes of UN / UNITAR, UNESCO, European Space Agency, Davos WEF, Philips, Saxo Bank, Mastercard, Barclays, and governments all over the world.
With over two decades of experience in international business, C-level positions, and digital transformation, Dinis has worked with new tech, cryptocurrencies, driven ICOs, regulation, compliance, and legal international processes, and has created a bank, and been involved in the inception of some of the top 100 digital currencies.
He creates and helps build ventures focused on global growth, 360 digital strategies, sustainable innovation, Blockchain, Fintech, AI and new emerging business models such as ICOs / tokenomics.
Dinis is the founder/CEO of ztudium that manages blocksdna / lifesdna. These products and platforms offer multiple AI P2P, fintech, blockchain, search engine and PaaS solutions in consumer wellness healthcare and life style with a global team of experts and universities.
He is the founder of coinsdna a new swiss regulated, Swiss based, institutional grade token and cryptocurrencies blockchain exchange. He is founder of DragonBloc a blockchain, AI, Fintech fund and co-founder of Freedomee project.
Dinis is the author of various books. He has published different books such “4IR AI Blockchain Fintech IoT Reinventing a Nation”, “How Businesses and Governments can Prosper with Fintech, Blockchain and AI?”, also the bigger case study and book (400 pages) “Blockchain, AI and Crypto Economics – The Next Tsunami?” last the “Tokenomics and ICOs – How to be good at the new digital world of finance / Crypto” was launched in 2018.
Some of the companies Dinis created or has been involved have reached over 1 USD billions in valuation. Dinis has advised and was responsible for some top financial organisations, 100 cryptocurrencies worldwide and Fortune 500 companies.
Dinis is involved as a strategist, board member and advisor with the payments, lifestyle, blockchain reward community app Glance technologies, for whom he built the blockchain messaging / payment / loyalty software Blockimpact, the seminal Hyperloop Transportations project, Kora, and blockchain cybersecurity Privus.
He is listed in various global fintech, blockchain, AI, social media industry top lists as an influencer in position top 10/20 within 100 rankings: such as Top People In Blockchain | Cointelegraph https://top.cointelegraph.com/ and https://cryptoweekly.co/100/ .
Between 2014 and 2015 he was involved in creating a fabbanking.com a digital bank between Asia and Africa as Chief Commercial Officer and Marketing Officer responsible for all legal, tech and business development. Between 2009 and 2010 he was the founder of one of the world first fintech, social trading platforms tradingfloor.com for Saxo Bank.
He is a shareholder of the fintech social money transfer app Moneymailme and math edutech gamification children’s app Gozoa.
He has been a lecturer at Copenhagen Business School, Groupe INSEEC/Monaco University and other leading world universities.