business resources

AI Safety And The Key Principles For Responsible AI Development

Dinis Guarda Author

20 Feb 2025, 11:40 am GMT

AI safety is more than just a tech issue—it’s a global priority. With global AI investment exceeding £120 billion in 2023 and 77% of businesses adopting AI, ensuring ethical, transparent, and accountable development is critical. Discover the key principles driving responsible AI and why securing its future matters now more than ever.

AI safety refers to the ideas, policies, and actions designed to ensure that artificial intelligence technology works in a safe, trustworthy, and useful manner for mankind. 

recent study,  “The Critical Conversation on AI Safety and Risk” highlights that aligning AI systems with human values is fundamental to AI safety, as is ensuring that these systems are dependable, scalable, and responsible. This involves technology safeguards, legislative frameworks, industry standards, and ongoing research into AI safety protocols. 

In early 2024, AI-generated deepfake images and videos of Taylor Swift circulated widely on social media platforms, depicting her in inappropriate and explicit content. The deepfakes were generated using advanced AI models that can create hyper-realistic media, making it difficult to distinguish them from real footage.

This incident raises questions about AI governance, digital ethics, and the potential for AI to be misused in cyber harassment and defamation cases.

What does this example suggest? 

The connection between innovation and safety is the key. It is significant to bridge cutting-edge developments with the ethical imperative to avoid undesirable outcomes. 

Effective AI safety plans include active involvement with civil society, ongoing risk assessment, and reaction to developing concerns. Nevertheless, the process of creating and implementing safety standards in this quickly changing industry is quite complex

Global AI Safety measures across sectors

Multiple approaches and real-world implementations highlight the necessity of AI safety in a variety of sectors. For example, in aviation and medicine, demanding certification processes and clinical studies guarantee that items satisfy high safety requirements before they reach the market. 

Similarly, AI organisations such as OpenAI have implemented techniques such as incorporating Coalition on Material Provenance and Authenticity (C2PA) standards to authenticate the provenance of digital material, reducing the dangers associated with AI-generated media. Another technique is 'red teaming,' in which external specialists stress-test AI systems to identify flaws, yet this is still a developing field. 

The Hawaii Department of Transportation's proactive strategy, which includes AI for activities ranging from traffic control to human resource operations, demonstrates AI's potential to save lives and increase operational efficiency. 

The Driving Transportation Safety Forward with AI report highlights how AI-driven systems such as Traffic Incident Management enabled by large-data innovations (TIMELI) can leverage AI to improve safety by predicting and mitigating traffic incidents.

AI Alliance and EU AI Act

Launched in 2018, the European AI Alliance is an initiative by the European Commission aimed at fostering an open policy dialogue on AI. This platform brings together a diverse group of stakeholders, including citizens, civil society organisations, businesses, consumer groups, trade unions, academia, public authorities, and experts. The primary objective is to collaboratively shape AI policies that are both innovative and aligned with European values.

The AI Alliance focuses on:

  • Ethics Guidelines for Trustworthy AI: In collaboration with the High-Level Expert Group on Artificial Intelligence (AI HLEG), the Alliance contributed to the formulation of guidelines that emphasise the ethical dimensions of AI, ensuring that AI systems are lawful, ethical, and robust.
  • Policy and Investment Recommendations: The Alliance has provided insights and feedback leading to comprehensive recommendations that guide AI-related policy-making and investment decisions within the EU.
  • Community Engagement: Through regular events, public consultations, and an active online forum, the Alliance has engaged approximately 6,000 stakeholders, facilitating discussions that influence AI policy directions.

EU Artificial Intelligence Act (AI Act)

The AI Act, which came into force on August 1, 2024, stands as the world's first comprehensive legal framework governing AI. Its primary aim is to ensure that AI systems used within the EU are safe, respect fundamental rights, and align with European values. The Act adopts a risk-based approach, categorising AI applications based on the potential harm they might pose:

  • Unacceptable Risk: AI applications that pose a clear threat to safety or fundamental rights are prohibited. This includes systems that manipulate human behavior or enable social scoring by governments.
  • High Risk: Applications in critical sectors such as healthcare, education, and law enforcement are subject to stringent obligations, including rigorous risk assessments and adherence to transparency and oversight requirements.
  • Limited Risk: Systems that interact with users, like chatbots, must comply with transparency obligations, ensuring users are informed they are engaging with an AI system.
  • Minimal Risk: Applications such as AI-enabled video games or spam filters are largely unregulated under the Act, given their low impact on users' rights or safety.

The AI Act also introduces measures to address general-purpose AI systems, including foundational models like ChatGPT. These systems are subject to specific transparency requirements, especially when they pose significant risks. The Act's extraterritorial scope means that providers and deployers outside the EU must comply with its provisions if their AI systems impact individuals within the EU. 

What are the key pillars of AI safety?

Key aspects for redesigning ethics for times of AI are: Data privacy, fairness, explainability,  transparency and accountability.

Data Privacy

Data privacy, often known as "information privacy," refers to the notion that individuals should have control over their personal data. This principle involves determining how organisations acquire, keep, and use their data.

Data privacy involves providing individuals with control over their personal data, and it has grown in importance as digital technologies have evolved, particularly in the context of artificial intelligence. AI systems sometimes require massive volumes of data to work well, raising questions about how personal information is stored and secured. 

Organisations have a legal and ethical need to follow these principles, ensuring that data subjects—the people who own the data—have control over their information. 

The General Data Protection Regulation (GDPR), an important piece of legislation in this field, requires businesses to create policies and systems that protect individual rights, even in the absence of official data privacy regulations. Organisations and companies can profit from implementing strong data privacy safeguards, which not only help with compliance but also increase consumer confidence. 

Data security solutions such as encryption, automated policy enforcement, and audit monitoring are critical components for organisations to comply with rules and protect personal data.

Fairness

In the context of AI and machine learning, fairness refers to the systems' impartial and equal treatment of all people, ensuring that no one is discriminated against based on race, gender, age, or sexual orientation. But what is fairness? 

In "What does “fairness” mean for machine learning systems?"  Smith states that fairness is frequently characterised in machine learning as the characteristic of being fair or unbiased; however, this definition might vary depending on the subject. AI researchers and practitioners try to design models that not only perform effectively but also treat individuals justly. 

Fairness must take into consideration how AI systems allocate and withhold resources. Fairness in AI entails detecting possible biases upfront and ensuring that various perspectives, including domain experts, are included in the discussion. The objective is not to construct a completely fair system but rather to detect and minimise fairness-related problems to the greatest extent feasible.

Google's What-If Tool enables users to investigate model performance across datasets while analysing fairness restrictions, whereas IBM's AI Fairness 360 Toolkit offers technical solutions via fairness measurements and algorithms to help discover and minimise bias in models. The algorithm's goal is to properly forecast biases for various individuals. 

Transparency

Transparency ensures that AI systems are not only understandable but also responsible to the people with whom they interact. In a Forbes article “Examples that demonstrate why transparency is critical in AI” Bernard believes it is vital to assess the clarity of AI algorithms, the data sources they use, and the decision-making processes they utilise. 

For AI to be considered transparent, users and stakeholders must understand how these systems arrive at their findings, ensuring that outcomes are fair, unbiased, and ethical. Various organisations (like Cognisant and others)  advocate for the formation of centres of excellence to centralise AI oversight inside a corporation. 

This strategy enables the universal use of transparency techniques across all AI efforts, ensuring that AI systems are not only accountable but also intelligible to users and stakeholders, resulting in increased confidence and responsible AI adoption.

When there is a lack of transparency in AI, it can have both positive and negative repercussions. For example, Microsoft's Python SDK for Azure Machine Learning has a model explainability option that is set to 'true' by default in current versions. This feature enables developers to obtain insight into the interpretability of AI choices, ensuring that they are done equitably and ethically. 

On the legislative front, the EU AI Act, the first comprehensive regulation on AI by a major regulator anywhere,  requires openness for AI systems used in vital applications, with substantial penalties for companies who implement opaque, black-box models. 

Explainability

Explainability in artificial intelligence refers to the ability for humans to comprehend AI systems' processes and  judgments. “In “What is Explainable Artificial Intelligence?”, IBM suggests that Explainable AI (XAI) refers to a set of methodologies and procedures that enable users to understand how machine learning (ML) algorithms achieve their results, hence increasing confidence and dependability in these systems. 

Unlike standard AI models, which may deliver outcomes without a clear understanding of how they were obtained, XAI guarantees that each choice can be traced and explained. This traceability is critical because it solves the so-called 'black box' problem in AI, in which the inner workings of complicated models, particularly deep learning and neural networks, are frequently unknown. 

XAI enhances the openness, accuracy, fairness, and accountability of AI models. By making AI more interpretable, XAI encourages responsible AI development by identifying and mitigating biases associated with sensitive traits such as race, gender, and age. As AI systems become more intertwined into numerous parts of society, like healthcare and economics, the significance of explainability grows. It enables organisations to retain trust in AI-driven judgements by providing explicit reasons for actions performed, which is particularly critical in regulated sectors where decisions may have a big influence on people's lives.

A specific case study demonstrating the value of XAI involves forecasting a patient's chance of developing diabetes using a machine learning model. Varad Vishwarupe's case study, "Explainable AI and Interpretable Machine Learning," used a dataset from a 2021 medical survey and a Random Forest classifier, taking into account clinical factors such as age, skin thickness, insulin levels, and BMI. 

Using XAI frameworks and techniques such as SHAP (Shapley Additive Explanations), LIME (Local Interpretable Model-agnostic Explanations), and ELI5, the researchers were able to offer extensive explanations for the model's predictions. These technologies enabled them to assess the influence of each clinical aspect on prediction, changing a previously opaque decision-making process into one that medical practitioners could readily understand and confirm. 

AI Safety Summit, source: Flickr

Explainable AI is important because it bridges the gap between complicated machine learning models and human comprehension, making AI-driven judgements more visible, trustworthy, and actionable.

Accountability

Accountability in artificial intelligence is a complicated but critical issue that entails deciding who is liable when AI systems make bad or destructive choices. 

In the article, “Critical Issues About AI Accountability Answered”, the authors suggest that as AI gets more integrated into numerous industries, the subject of responsibility becomes more relevant, especially when these systems have unintended repercussions. Accountability in AI involves more than simply determining who is to blame when anything goes wrong; it also includes ensuring that measures are in place to prevent such occurrences in the first place. This includes creating clear norms and duties for AI developers, consumers, and suppliers. 

The qualities of accountability in AI are diverse, including openness, effective human monitoring, and the capacity to challenge AI choices. AI systems are employed in financial organisations to identify potentially fraudulent transactions and assess creditworthiness. However, a lack of transparency in these systems can lead to the unjust treatment of consumers, such as stopped payments or refused credit applications. 

The simpler the story the better

We know there is much more than these 5 elements for increased AI safety. And the most important is widespread awareness of the tsunami wave that AI is. 

Raising awareness about AI safety is not just a responsibility for policymakers, researchers, or technologists—it is a collective imperative for all of humanity. The rapid acceleration of AI and AGI development demands a proactive approach to governance, ethics, and public engagement. Without widespread understanding and participation, decisions about the future of AI may be shaped by a small few, leaving the majority unprepared for its profound implications. 

To navigate this pivotal moment in our evolutionary journey, we must prioritise education, transparency, and inclusive dialogue, ensuring that AI serves as a force for progress rather than an uncontrollable disruption.

AI and AGI are overwriting our evolutionary path as humans and opening new ways, that can be either highways or sinister roads.

Some of us are aware of this but the very truth is that 99% of humanity is more focused on going on with their lives in many cases in survival mode even if in statistics we live at the best stage of human evolutionary social health and economic levels.

Yuval Noah Harari says how “Humans think in stories rather than in facts, numbers, or equations, and the simpler the story, the better.” 
But AI models compute, that is,  “think” rational facts, numbers, or equations. Is this a good idea then? To just allow rationality to rule our lives?  The narrative of Humanity is in our hands.

Share this

Dinis Guarda

Author

Dinis Guarda is an author, entrepreneur, founder CEO of ztudium, Businessabc, citiesabc.com and Wisdomia.ai. Dinis is an AI leader, researcher and creator who has been building proprietary solutions based on technologies like digital twins, 3D, spatial computing, AR/VR/MR. Dinis is also an author of multiple books, including "4IR AI Blockchain Fintech IoT Reinventing a Nation" and others. Dinis has been collaborating with the likes of  UN / UNITAR, UNESCO, European Space Agency, IBM, Siemens, Mastercard, and governments like USAID, and Malaysia Government to mention a few. He has been a guest lecturer at business schools such as Copenhagen Business School. Dinis is ranked as one of the most influential people and thought leaders in Thinkers360 / Rise Global’s The Artificial Intelligence Power 100, Top 10 Thought leaders in AI, smart cities, metaverse, blockchain, fintech.