AI

EU and US lawmakers move to draft AI Code of Conduct fast

Comment

Image Credits: Getty Images under a Jonas EKSTROMER / TT NEWS AGENCY / AFP) / Sweden OUT license.

The European Union has used a transatlantic trade and technology talking shop to commit to moving fast and producing a draft Code of Conduct for artificial intelligence, working with US counterparts and in the hope that governments in other regions — including Indonesia and India — will want to get involved.

What’s planned is a set of standards for applying AI to bridge the gap, ahead of legislation being passed to regulate uses of the tech in respective countries and regions around the world.

Whether AI giants will agree to abide by what will be voluntary (non-legally binding) standards remains to be seen. But the movers and shakers in this space can expect to be encouraged to do so by lawmakers on both sides of the Atlantic — and soon, with the EU calling for the Code to be drafted within weeks. (And, well, given the rising clamour from tech industry CEO screaming for AI regulation it would be pretty hypocritical for leaders in the field to turn their noise up at a voluntary Code.)

Speaking at the close of a panel session on generative AI that was held at the fourth meeting of the US-EU Trade & Tech Council (TTC) which was taking place in Sweden this week — with the panel hearing from stakeholders including Anthropic CEO Dario Amodei and Microsoft president Brad Smith — the European Union’s EVP Margrethe Vestager, who heads up the bloc’s competition and digital strategy, signalled it intends to get to work stat.

“We will be very encouraged to take it from here. To produce a draft. To invite global partners to come on board. To cover as many as possible,” she said. “And we will make this a question of absolute urgency to have such an AI Code of Conduct for a voluntary signup.”

The TTC was established back in 2021, in the wake of the Trump presidency as EU and US lawmakers sought to repair trust and find ways to cooperate on tech governance and trade issues.

Vestager described generative AI as a “seismic change” and “categorical shift” that she said demands a regulatory response in real-time.

“Now technology is accelerating to a completely different degree than what we’ve seen before,” she said. “So obviously, something needs to be done to get the most of this new technology… We’re talking about technology that develops by the month so what we have concluded here at this TTC is that we should take an initiative to get as many other countries on board on an AI Code of Conduct for businesses voluntarily to sign up.”

While Vestager couched industry input as “very welcome”, she indicated that — from the EU side at least — the intent is for lawmakers to draw up safety provisions and companies to agree to get on board and apply the standards, rather than having companies drive things by suggesting a bare minimum on standards (and/or seeking to reframe AI safety to focus on existential future threats rather than extant harms) and lawmakers swallowing the bait.

“We will not forget that there are other sorts of artificial intelligence, obviously,” she went on. “There are a number of things that need to be done. But the thing is that we need to show that democracy is up to speed because legislative procedures they must take their time; that is the nature of legislation. But this is a way for democracies to answer in real time a question that is really, really in our face right now. And I find this very encouraging to do it and I’m looking very much forward to work with as many as possible in depth and very fast.”

The EU is ahead of the regulatory curve on AI since it already has draft legislation on the table. But the risk-based framework the Commission presented, back in April 2021, is still winding its way through in the bloc’s co-legislative — looping in lawmakers in the European Council and Parliament (and for a sense of how live that process is parliamentarians recently proposed amendments targeted at generative AI) — so there’s no immediate prospect of those hard rules applying to generative or any other type of AI.

Even with the most optimistic outlook for the EU adopting the AI Act, Vestager suggested today that it would be two or three years before those hard rules bite. Hence the bloc’s urgency for stop-gap measures.

“… a pace like no other”

Also present at the TTC meeting was Gina Raimondo, US secretary of state for commerce — who indicated the Biden administration’s willingness to engage with discussion toward shaping a voluntary AI Code of Conduct. Although she kept her cards close to her chest on what kind of standards the US might be comfortable to push onto what are predominantly US AI giants.

“[AI] is coming at a pace like no other technology,” observed Raimondo. “Like other technologies, we are already seeing issues with data privacy, misuse, what happens when the models get into the hands of malign actors, misinformation. Unlike other technology, the rate of the pace of innovation is at a breakneck pace, which is different and a hockey stick that doesn’t exist in other technologies.

“In that respect, I think the TTC could play an incredibly relevant role because it will take a little bit of time for the US Congress or the parliament or other regulatory agencies to catch up. Whereas the risk for some of AI is today. And so we are committed to making sure that the TTC provides a forum for stakeholder engagement, engagement of the private sector, engagement of our companies, to figure out what can we do in the here and now to mitigate the risks of AI but also not to stifle the innovation. And that is a real challenge”

“As we figure out the benefits of AI, I hope we’re all really eyes wide open about the costs and do the analysis of whether we should do it,” she also warned. “I think if we all are honest with ourselves about other technologies, including social media, we probably wish we had not done things even though we could’ve. You know, we could have but should have we? And so let’s work together to get this right, because the stakes are a whole lot higher.”

As well as high level lawmakers, the panel discussion heard from a handful of industry and civil society groups, chipping in with perspectives on the imperative for and/or challenge of regulating such a fast moving field of technology.

Anthropic’s Amodei heaped praise on the transatlantic conversation taking place around AI rule-making. Which likely signals relief that the US is actively involving itself in standards-making which might otherwise be exclusively driven by Brussels.

The bulk of his remarks sounded a sceptical note over how to ensure AI systems are truly safe prior to release — implying we don’t yet have techniques for achieving reliable guardrails around such shape-shifting tools. He also suggested there should be a joint commitment from the US and EU to fund the development of “standards and evaluation” for AI — rather than calling for any algorithmic auditing in the here and now.

“When I think about the rate at which this technology is bringing new sources of power into the world, combined with the resurgent threat from autocracies that we’re seeing over the last year, it seems to me that it’s all the more important that we work together to prevent [AI] harms and defend our shared democratic values. And the TTC seems like a critical for forum for doing that,” he said early in his allotted time before going on to predict that developments in AI would continue to come at a steady clip and setting out some of his major concerns — including highlighting “measurement” for AI safety as a challenge.

“What we’re going to be able to do in one to four years are things that seem impossible now. This is, I would say, if there’s a central fact about the field of AI to know, this is the central fact to know. And though there will be many positive opportunities to come from this, I worry greatly about risks — particularly in the domain of cybersecurity, biology, things like disinformation, where I think there’s the potential for great destruction,” he said. “In the longer term, I worry even about the risks of truly autonomous systems. That’s a little further out.

“On measurement, I think we’re very used to — when we think about regulating technologies like automobiles or aeroplanes — we’re measuring safety as a secure field; you have a given set of tests you can run to tell if the system is safe. AI is much more of a wild west than that. You can ask an AI system to do anything at all in natural language and it can answer in any way it chooses to answer.

“You might try to ask a system 10 different ways whether it can conduct, say, a dangerous cyber attack and find that it won’t. But you forgot to ask it an 11th way that would have shown this dangerous behaviour. A phrase I like to use is ‘no one knows what an AI system is capable of until it’s deployed to a million people’. And of course, this is a bad thing, right? We don’t want to deploy these things in this cowboy-ish way. And so this difficulty of detecting dangerous capabilities is a huge impediment to mitigating them.”

The contribution looked intended to lobby against any hard testing of AI capabilities being included in the forthcoming Code of Conduct — by seeking to kick the can down the road.

“This difficulty of detecting dangerous capabilities a huge impediment to mitigating them,” he suggested, while conceding that “some kind of standards or evaluation are a crucial prerequisite for effective AI regulation” but also further muddying the water by saying “both sides of the Atlantic have an interest in developing this science”.

“US and EU have a long tradition of collaborating on [standards and evaluation] which we could extend and then maybe more radically, a commitment to adopt an eventual set of common standards and evaluations as a sort of raw material for the rules of the road in AI,” he added, gazing into the eventual distance.

Microsoft’s Smith used his four minutes’ speaking time to urge regulators to “move forward the innovation and safety standards together” — also amping up the AI hype by lauding the potential benefits for AI to “do good for the world” and “save people’s lives”, such as by detecting or curing cancer or enhancing disaster response capabilities, while conceding safety needs focus with an affirmation that “we do need to be clear eyed about the risks”.

He also welcomed the prospect of transatlantic cooperation on AI standards. But pressed for lawmakers to shoot for broader international coordination on things like product development processes — which he suggested would help drive forward on both AI safety standards and innovation.

“Certain things benefit enormously from international coordination, especially when it comes to product development processes. We’re not going to advance safety or innovation if there’s different approaches to, say, how our red team should work in the safety product process for developing a new AI model,” he said.

“Other things there’s more room for divergence and there will be some because the world — even the countries that share common values — we’ll have some differences. And there’s areas around licensing or usage where one can manage with that divergence. But in short, there’s a lot that we will benefit from learning now and then putting into practice.”

Towards external audits?

No one from OpenAI was speaking during the TTC panel but Vestager had a videoconference meeting with CEO Sam Altman in the afternoon.

In a read-out of the meeting, the Commission said the pair shared ideas for the voluntary AI code of conduct that was launched at the TTC — with discussion touching on how to tackle misinformation; transparency issues, including ensuring users are made aware if they communicate with AI; how to ensure verification (red teaming) and external audits; how to ensure monitoring and feedback loops; and the issue of ensuring compliance while avoiding barriers for startups and SMEs.

The Commission added that there was “a strong overall agreement to advance on the voluntary code of conduct as fast as possible and with G7 and other key partners, as a stopgap measures until regulation is in place”, adding there would be “a continued engagement on the AI Act as the legislative process progresses”.

In a subsequent tweet Vestager said discussions with OpenAI’s Altman and Anthropic’s Amodei had featured talk of external audits, watermarking and “feedback loops”.

In recents days Altman has ruffled feathers in Brussels with some flatfooted lobbying in which he seemingly threatened to pull his tool out of the region if provisions in the EU’s AI Act targeted at generative AI aren’t watered down.

He then quickly withdrew the threat after the bloc’s internal market commissioner tweeted a public dressing down at OpenAI, accusing the company of attempting to blackmail lawmakers. So it will be interesting to see how enthusiastically (or otherwise) Altman engages with the substance of the Code of Conduct for AI.

(For its part, Google has previously indicated it wants to work with the EU on stop-gap AI standards — as part of a so-called “AI Pact” which appears to be a separate EU initiative to the Code of Conduct; per a Commission spokesperson the AI Pact is focused on getting companies to agree to front-load the implementation of key AI Act provisions on a voluntary basis, whereas the Code aims to promote guardrails for the use of generative AI or “advanced GPAI” (general purpose AI) models on a global level.)

While AI giants have been relatively reluctant to focus on current AI risks and how they might be reined in, preferring talk of far-flung fears of non-existent “superintelligent” AIs, the TTC meeting also heard from Dr. Gemma Galdon-Clavell, founder and CEO of Eticas Consulting — a business which runs algorithmic audits for customers to encourage accountability around uses of AI and algorithmic technology — who was eager to school the panel in current-gen accountability techniques.

“I am convinced [algorithmic auditing] is going to be the main tool to understand quantify and mitigate harms in AI,” she said. “We ourselves are hoping to be the first auditing unicorn that puts the tools [on the table] that maximise engineering possibilities while taking into account fundamental rights and societal values.”

She described the EU’s recently adopted overhaul of ecommerce and marketplace rules, aka the Digital Services Act (DSA), as a pioneering piece of legislation in this regard — on account of the law’s push to require transparency from very large online platforms on how their algorithms work, predicting algorithmic audits will become the go-to AI safety tool in the coming years.

“The good news is that audits are informally becoming the consensus on one of the potential ways of regulating AI,” she argued. “We have audit in the wording of the DSA, an absolute pioneer in Europe. We have audits in the New York City regulation of the use of AI hiring systems. The very recent NTIA consultation process is focused on audit. The FTC keeps asking about what are the standards and the thresholds that need to be used in audits. So there’s an emerging consensus that audits — call them inspection mechanisms… validations… the name doesn’t really matter — the thing is that we need inspection mechanisms that allow us to land the concerns of the policymakers, while understanding the technologies that are making change possible.”

“I am convinced, in a few years from now — not many years, three, five years from now — we’ll be amazed there was a time where we released AI systems without audits. We will not believe it,” she added, comparing the current wild west of unregulated AI safety and transparency to the 19th Century when a consumer could walk into a pharmacy and buy cocaine.

Despite spotlighting a regulatory trajectory she suggested is headed towards auditing, Galdon-Clavell was also critical of a tendency for policymakers to lock on to talk of far-flung theoretical harms — what she dubbed “science fiction debates” — which she skewered as a distraction from addressing current AI-driven harms, warning: “There’s an opportunity costs when we talk about science fiction in the long distance impact. What happens to the current impacts that we are seeing right now? How do we protect people today? This generation from the harms that we already understand and know are happening around those systems?”

She also urged lawmakers to get on with passing legislation with “teeth”. “The industry needs to do better and right now the incentives to do better are not there,” she emphasized. “Our clients come to us to be audited because they think that’s what they need to do. And not because they are forced by anyone. And those that don’t audit have no incentive to start doing it if they don’t want to.

“The regulation that needs to come needs to have teeth. So my plea to the TTC would be to please start listening to the emerging consensus that’s already there. It is transatlantic, it is very much global. There’s things on the table they can already help us protect right now — tomorrow, today — the current generation that is suffering the negative impacts of some of these new technological developments.”

Another speaker on the panel, Alexandra Reeve Givens, the president & CEO of the US-based human rights focused not-for-profit, the Center for Democracy and Technology, also urged policymakers to direct their attention on “real” AI risks that she said are “manifesting already”.

“We’re seeing already the professional, reputational and potential physical harms when people rely on generated text results as accurate, unaware of the likelihood of hallucinations or fabricated results,” she warned. “There’s the risk that generative AI tools will supercharge fraud, as tools make it easier to quickly generate personalised scams or to trick people by impersonating a familiar voice. There are risks of deep fakes that misrepresent public figures in a way that threatens elections, national security, or general public order. And there are risks of fake images being used to harass, exploit and extort people. None of these harms are new but they’re made cheaper, faster and more effective by the ease and accessibility of generative AI tools.”

She also urged lawmakers not to limit their focus to only generative AI and ignore harms being generated by less viral flavors of AI — which she said are “directly impacting people’s rights, freedoms and access to opportunity today” — such as AI tools being used to determine who gets a job or receives public benefits, or the use of AI surveillance tools by law enforcement.

“Policymakers cannot lose sight of those core issues, even as they expand their focus to generative AI,” she said, calling for any AI assessment initiatives to be similarly comprehensive and address a full-spectrum of real-world harms.

“Policymakers must be crystal clear that efforts to evaluate and manage AI risk must meaningfully address a full spectrum of real world harms. What I mean by that is policymakers must ensure AI audits and assessments are rigorous, comprehensive, and escape issues of capture. Policymakers must address the danger that frameworks for measuring risks often address only those harms that can be easily measured, which privileges economic and physical harms over equally important harms to people’s privacy, dignity and the right not to be stereotyped or maligned.”

“As policymakers in the US and the EU look to industry efforts or technically oriented standards bodies to consider questions of measuring and managing risk they must ensure these fundamental rights-based concerns are appropriately addressed,” she added.

Reeve Givens also called for transparency plans in the TTC roadmap, which calls for joint tracking of emerging AI risks and incidents of harms, to be expanded to tackle information asymmetries to establish a common foundation for regulators to work from — and for civil society voices and marginalized communities who may face disproportionate risk and harm from AI outputs to be involved in any processes to draw up standards.

This report was updated with additional detail about the difference between the EU’s AI Pact initiative and the AI Code of Conduct 

Sam Altman’s big European tour

Google to work with Europe on stop-gap ‘AI Pact’

Europe spins up AI research hub to apply accountability rules on Big Tech

More TechCrunch

Apple is bringing new accessibility features to iPads and iPhones, designed to cater to a diverse range of user needs.

Apple announces new accessibility features for iPhone and iPad users

TechCrunch Disrupt, our flagship startup event held annually in San Francisco, is back on October 28-30 — and you can expect a bustling crowd of thousands of startup enthusiasts. Exciting…

Startup Blueprint: TC Disrupt 2024 Builders Stage agenda sneak peek!

Mike Krieger, one of the co-founders of Instagram and, more recently, the co-founder of personalized news app Artifact (which TechCrunch corporate parent Yahoo recently acquired), is joining Anthropic as the…

Anthropic hires Instagram co-founder as head of product

Seven orgs so far have signed on to standardize the way data is collected and shared.

Venture orgs form alliance to standardize data collection

As cloud adoption continues to surge towards the $1 trillion mark in annual spend, we’re seeing a wave of enterprise startups gaining traction with customers and investors for tools to…

Alkira connects with $100M for a solution that connects your clouds

Charging has long been the Achilles’ heel of electric vehicles. One startup thinks it has a better way for apartment dwelling EV drivers to charge overnight.

Orange Charger thinks a $750 outlet will solve EV charging for apartment dwellers

So did investors laugh them out of the room when they explained how they wanted to replace Quickbooks? Kind of.

Embedded accounting startup Layer secures $2.3M toward goal of replacing Quickbooks

While an increasing number of companies are investing in AI, many are struggling to get AI-powered projects into production — much less delivering meaningful ROI. The challenges are many. But…

Weka raises $140M as the AI boom bolsters data platforms

PayHOA, a previously bootstrapped Kentucky-based startup that offers software for self-managed homeowner associations (HOAs), is an example of how real-world problems can translate into opportunity. It just raised a $27.5…

Meet PayHOA, a profitable and once-bootstrapped SaaS startup that just landed a $27.5M Series A

Restaurant365, which offers a restaurant management suite, has raised a hot $175M from ICONIQ Growth, KKR and L Catterton.

Restaurant365 orders in $175M at $1B+ valuation to supersize its food service software stack 

Venture firm Shilling has launched a €50M fund to support growth-stage startups in its own portfolio and to invest in startups everywhere else. 

Portuguese VC firm Shilling launches €50M opportunity fund to back growth-stage startups

Chang She, previously the VP of engineering at Tubi and a Cloudera veteran, has years of experience building data tooling and infrastructure. But when She began working in the AI…

LanceDB, which counts Midjourney as a customer, is building databases for multimodal AI

Trawa simplifies energy purchasing and management for SMEs by leveraging an AI-powered platform and downstream data from customers. 

Berlin-based trawa raises €10M to use AI to make buying renewable energy easier for SMEs

Lydia is splitting itself into two apps — Lydia for P2P payments and Sumeria for those looking for a mobile-first bank account.

Lydia, the French payments app with 8 million users, launches mobile banking app Sumeria

Cargo ships docking at a commercial port incur costs called “disbursements” and “port call expenses.” This might be port dues, towage, and pilotage fees. It’s a complex patchwork and all…

Shipping logistics startup Harbor Lab raises $16M Series A led by Atomico

AWS has confirmed its European “sovereign cloud” will go live by the end of 2025, enabling greater data residency for the region.

AWS confirms will launch European ‘sovereign cloud’ in Germany by 2025, plans €7.8B investment over 15 years

Go Digit, an Indian insurance startup, has raised $141 million from investors including Goldman Sachs, ADIA, and Morgan Stanley as part of its IPO.

Indian insurance startup Go Digit raises $141M from anchor investors ahead of IPO

Peakbridge intends to invest in between 16 and 20 companies, investing around $10 million in each company. It has made eight investments so far.

Food VC Peakbridge has new $187M fund to transform future of food, like lab-made cocoa

For over six decades, the nonprofit has been active in the financial services sector.

Accion’s new $152.5M fund will back financial institutions serving small businesses globally

Meta’s newest social network, Threads, is starting its own fact-checking program after piggybacking on Instagram and Facebook’s network for a few months.

Threads finally starts its own fact-checking program

Looking Glass makes trippy-looking mixed-reality screens that make things look 3D without the need of special glasses. Today, it launches a pair of new displays, including a 16-inch mode that…

Looking Glass launches new 3D displays

Replacing Sutskever is Jakub Pachocki, OpenAI’s director of research.

Ilya Sutskever, OpenAI co-founder and longtime chief scientist, departs

Intuitive Machines made history when it became the first private company to land a spacecraft on the moon, so it makes sense to adapt that tech for Mars.

Intuitive Machines wants to help NASA return samples from Mars

As Google revamps itself for the AI era, offering AI overviews within its search results, the company is introducing a new way to filter for just text-based links. With the…

Google adds ‘Web’ search filter for showing old-school text links as AI rolls out

Blue Origin’s New Shepard rocket will take a crew to suborbital space for the first time in nearly two years later this month, the company announced on Tuesday.  The NS-25…

Blue Origin to resume crewed New Shepard launches on May 19

This will enable developers to use the on-device model to power their own AI features.

Google is building its Gemini Nano AI model into Chrome on the desktop

It ran 110 minutes, but Google managed to reference AI a whopping 121 times during Google I/O 2024 (by its own count). CEO Sundar Pichai referenced the figure to wrap…

Google mentioned ‘AI’ 120+ times during its I/O keynote

Firebase Genkit is an open source framework that enables developers to quickly build AI into new and existing applications.

Google launches Firebase Genkit, a new open source framework for building AI-powered apps

In the coming months, Google says it will open up the Gemini Nano model to more developers.

Patreon and Grammarly are already experimenting with Gemini Nano, says Google

As part of the update, Reddit also launched a dedicated AMA tab within the web post composer.

Reddit introduces new tools for ‘Ask Me Anything,’ its Q&A feature