Ethical AI in Government
Government agencies face a unique challenge when it comes to regulating AI while simultaneously using it. They are bound by legal and ethical obligations to protect citizens’ rights and privacy; therefore, when employing AI, they must do so responsibly to uphold democratic values and protect individual rights. In the first part of this series, we outlined the ethical AI framework. But what would that look like for government entities? Expert Darrell M. West, Senior Fellow at the Center for Technology Innovation, outlines six key steps:
- Concrete Codes of Conduct: Government agencies need clear codes of conduct that outline major ethical standards, values, and principles. These should include fairness, transparency, privacy, and human safety.
- Operational Tools: Employees involved in AI development must have access to operational tools that promote ethics and fight bias. These tools should be designed with input from ethicists, social scientists, and legal experts to ensure impartial and safe decision-making.
- Evaluation Benchmarks: Clear evaluation benchmarks and metrics should be established to assess AI systems’ performance and adherence to ethical principles. These metrics should consider both substantive and procedural fairness.
- Technical Standards: Governments should adopt technical standards that guide AI development to prevent idiosyncratic designs and ensure consistent safeguards, especially in areas like fairness and equity.
- Pilot Projects and Sandboxes: Government agencies should conduct pilot projects and establish sandboxes for experimenting with AI deployments. This allows testing AI in a controlled environment, minimizing risks, and learning from initial tests.
- Workforce Capacity: A well-trained workforce with a mix of technical and non-technical skills is essential. Government agencies should invest in professional development opportunities to keep their employees updated on emerging technologies.
What are the applications of AI in government?
Recent data shows that over 77% of companies are either using or exploring the use of AI within their business. As we discussed previously, AI has the power to streamline processes – and that is already impacting customer’s expectations.
According to a Salesforce research report, 83% of consumers expect immediate engagement when they contact a company, while 73% expect companies to understand their unique needs and expectations. Nearly 60% of all customers want to avoid customer service altogether, preferring to resolve issues with self-service features. Naturally, it influences the public sector: Citizens today expect seamless digital interactions with government services, similar to their experiences in the private sector.
AI undoubtedly has the power to change government-citizen relations and aid policymakers in their decisions. “Studies have shown that citizen’s digital experience with government services is a large predictor of trust in the government,” says John Weigelt, National Technology Officer at Microsoft Canada. “Artificial Intelligence enabled services delivery, as part of government’s digital transformation, and helps ensure that constituents get the right outcomes to their interactions with governments.”
The possibilities are endless – some of the main applications of AI in government relations are:
- Enhancing Digital Interactions with Public Services – Generative AI is able to not only analyze data but generate content according to the context of a certain interaction. In government, it can ensure a greater coverage of services as well as customization.
- Back Office Automation: AI technologies like robotic process automation (RPA), natural language processing (NLP), and computer vision are digitizing paper documents and accelerating claims processing. This not only reduces paperwork but also enhances the speed and accuracy of service delivery
- Data-based policymaking: AI enables policymakers to make more informed decisions based on data. It offers insights into industry regulation, social and environmental impacts, and citizen perceptions of government policies. This results in more effective and well-informed policymaking across all government sectors.
- Health and Environmental Predictions: AI can be used to help identify patterns, and impacts related to public health and climate change, as well as predict risks of housing and food insecurity. This assists in crafting policies to improve citizens’ quality of life.
There are already successful case studies of AI in governments like Australia, Canada and the United States. For example, Australia’s Taxation Office Chatbot had more than 3 million conversations and was able to resolve 88% of queries on first contact. In the US, Atlanta’s Fire Rescue Department Predictive Analysis was able to accurately predict 73% of fire incidents in a building.
“Empowering employees, finding efficiencies and transforming operations are key pillars of government digital transformation efforts,” says Weigelt. “Artificial intelligence helps employees gain faster and more accurate access to knowledge, speed and streamline decision making and provides a platform to reimagine how government operations are performed.”
Regulating AI: What’s in store
From the makers of GDPR: The proposed EU AI Act
AI has altered the way we interact with technology. Thierry Breton, the EU’s Commissioner for Internal Market, aptly noted, “[AI] has been around for decades but has reached new capacities fueled by computing power.”
Recognizing the transformative potential of AI and the need to mitigate its inherent risks, the EU AI Act represents a pivotal response to the transformative potential and risks of AI. It underscores the global significance of AI regulation and the need for international collaboration to address the challenges and opportunities posed by AI technologies.
First introduced in April 2021 by the European Commission, the EU AI Act’s implications extend far beyond European borders. Just as their General Data Protection Regulation (GDPR) has influenced the development of data protection laws in other countries, as the world’s first comprehensive regulatory framework for AI, it sets a precedent for responsible AI governance worldwide.
However, the Act has not been without its share of criticisms. Some European companies have voiced concerns about its potential impact on competitiveness and technological sovereignty. Nevertheless, the proposed legislation signifies a significant step towards achieving a harmonious balance between innovation, ethics, and accountability in the world of AI.
What it entails
The EU AI Act adopts a risk-based approach to AI regulation, categorizing AI systems based on the level of risk they pose to users. This classification serves as the foundation for imposing varying degrees of regulation on AI technologies. Three primary risk categories emerge:
- Unacceptable Risk: This category encompasses AI systems that pose a direct threat to individuals or specific vulnerable groups. Examples include AI-driven devices that manipulate children into engaging in harmful behaviours or social scoring systems that categorize individuals based on personal characteristics. The EU takes a stringent stance against such AI systems, proposing an outright ban.
- High Risk: AI systems falling under this category negatively impact safety or fundamental rights. They include AI used in products covered by the EU’s product safety legislation, such as toys, aviation, medical devices, and more. Additionally, certain specific areas like biometric identification, critical infrastructure management, and law enforcement require registration in an EU database. All high-risk AI systems undergo a rigorous assessment before market placement and throughout their lifecycle.
- *Limited Risk: AI systems posing limited risk must comply with transparency requirements, ensuring users are informed about AI-generated content. This includes AI systems responsible for generating or manipulating image, audio, or video content, such as deepfakes.
*On the earlier version of the Act. More information is below.
How it affects Generative AI
Generative AI tools, like ChatGPT, are not exempt from the regulations proposed in the EU AI Act. On June 14, 2023, the European Parliament passed a draft law with relevant amendments to the EU AI Act after the immense adoption of ChatGPT by both ordinary customers and organizations. The amendment broadened the EU AI Act scope to encompass new areas of concern, including environmental impact and effects on political campaigns.
Another noteworthy aspect is the introduction of “foundation models” and Article 28b, which exclusively addresses the responsibilities of providers of such models. A “foundation model” refers to an AI model trained on extensive and diverse data, designed to produce a wide range of outputs, and adaptable to various specific tasks – much like ChatGPT.
Providers of foundation models will now have additional transparency requirements, including:
- Disclose that the content was generated by AI.
- Be designed to prevent the generation of illegal content.
- Publish summaries of copyrighted data used for training.
- These measures aim to ensure accountability and transparency in AI-generated content, protecting users and society at large from the potential misuse of AI technologies.
This shift is particularly intriguing considering that chatbots and “deepfakes” were previously considered low-risk and subjected to minimal transparency obligations in earlier versions of the Act.
What about the US?
Following the EU AI Act, The White House Office of Science and Technology Policy has proposed a “Blueprint for an AI Bill of Rights” to protect the American public in the age of artificial intelligence. The Blueprint contains five principles, each of which includes a technical companion that provides guidance for responsible implementation:
- Safe and Effective Systems: The first principle emphasizes the need to protect individuals from AI systems that are unsafe or ineffective.
- Algorithmic Discrimination Protections: This principle aims to prevent discrimination by algorithms, ensuring that AI systems are designed and used in an equitable manner.
- Data Privacy: The third principle addresses the importance of protecting individuals from abusive data practices, giving people agency over how their data is used.
- Notice and Explanation: People should be informed when an automated system is being used and should understand how and why it influences outcomes that affect them.
- Alternative Options: Individuals should have the ability to opt-out when appropriate and access assistance when encountering problems with AI systems.
The release of the Blueprint has generated mixed reactions. Some experts argue that the Blueprint does not go far enough and lacks the checks and balances present in the EU AI Act. On the other hand, others fear that regulation could stifle innovation.
Nevertheless, in December 2020, the US Federal government signed Executive Order 13960, which emphasizes the benefits of AI for government operations, public services, and efficiency while highlighting the need to maintain public trust and protect privacy and civil rights.
The order sets forth principles for AI use in non-national security contexts, stressing accountability, transparency, and adherence to laws and values, and mandates agencies to inventory their AI use cases, ensure consistency with the principles, and share this information with the public.
Both the European Union and the United States’ journey to regulate AI it’s just the beginning, but an important step toward ensuring that AI benefits society while safeguarding individuals’ rights and well-being. The path forward requires a delicate balance between innovation and regulation, with an eye on the evolving global landscape of AI governance.
“AI in Government” series, in partnership with Microsoft Canada.
Part I: Navigating the Future of AI: Responsibility, Regulation, and Generative AI Impact
Part II: AI in Government: The fine balance between applying and regulating AI
Part III: Generative AI Impact on Governments
Additional resources:
Read Brad Smith’s blog “How do we best govern AI?“
Read Diana Parker’s blog to learn 4 steps to advance your AI journey with Microsoft for Government.
Learn more about the EU AI Act
Learn more about the proposed AI Legislation in Canada
Sign up for our newsletter to stay up to date on AI Regulation
Get Private AI on Azure Marketplace.