TelcoNews New Zealand - Telecommunications news for ICT decision-makers
Nodero

NZ faces legal and sovereignty risks as EU AI rules take effect - experts

Tue, 17th Mar 2026

New Zealand risks losing control over how artificial intelligence is used in critical sectors such as health, education and justice unless it introduces its own binding AI legislation and builds sovereign capability, according to industry experts.

The warning comes as sweeping new artificial intelligence rules begin taking effect across Europe under the European Union's Artificial Intelligence Act. While some provisions are already in force this year, the most stringent requirements for high-risk systems will apply from August 2, 2026. From that date, organisations operating in the EU must meet strict standards around risk assessment, transparency, documentation, human oversight and accountability.

New Zealand has no equivalent standalone AI law, with technology risks managed through the Algorithm Charter for Aotearoa, the Privacy Act and general human rights protections. There is no dedicated AI regulator and no unified compliance regime tailored specifically to artificial intelligence.

Dr. Athar Imtiaz, Applied AI researcher at Massey University and AI & Data Lead at digital consultancy Nodero, says AI adoption in New Zealand is moving faster than the governance structures designed to oversee it.

He says government agencies, including Te Whatu Ora, are already trialling generative AI tools to streamline document processing and service delivery, including in health care. At the same time, AI uptake in New Zealand's finance and logistics sectors is being driven largely by private firms.

"As our capability continues to accelerate, our governance over the technology, unfortunately, is not keeping pace."

Dr. Imtiaz says AI systems are increasingly influencing real-world decisions rather than operating as experimental tools.

"When AI assists with medical triage, welfare eligibility or justice processes, it becomes part of the decision-making infrastructure. 

"Most modern AI systems are probabilistic in that they generate likelihoods, not certainties. We need clear standards for acceptable error and bias testing, and validation against datasets that properly reflect New Zealand's communities, including Maori, rather than relying solely on international training data."

Dr. Imtiaz says existing legislation was not drafted with machine-learning systems in mind.

"The Privacy Act remains important, but it does not define technical standards for model validation, dataset integrity or audit requirements. Without tailored legislation, accountability becomes fragmented."

Beyond regulation, Dr. Imtiaz says New Zealand must develop its own AI training and adaptation capability if it wants systems that reflect local realities.

"Most advanced AI models are trained on international datasets shaped by larger economies. 

"If we rely entirely on offshore systems, we adapt ourselves to their assumptions rather than shaping them to our context."

Dr. Imtiaz says building sovereign capability would require investment in high-performance computing, secure data environments and specialist expertise. 

He says that level of national AI infrastructure and fine-tuning capability could require several hundred million dollars over the next five years.

"Australia has already committed more than A$100 million in federal funding toward AI capability and regulatory reform, alongside broader digital infrastructure investment. Singapore has invested billions across successive national AI strategies. The United Kingdom continues to fund AI research and governance capacity through its central technology portfolio.

"In contrast, New Zealand has no dedicated AI budget line and no central authority equivalent to the Department for Science, Innovation and Technology in the UK, the Ministry of Digital Development and Information in Singapore, or Japan's Digital Agency."

Mark Easton, chief executive of Nodero, says the issue is not simply regulatory alignment but sovereignty.

He says the 2025 Oxford Insights Government AI Readiness Index ranks New Zealand around 40th globally, behind Australia and several comparable economies.

"When major markets such as Europe set compliance thresholds, vendors design to those standards. 

"If New Zealand does not define its own statutory and institutional expectations, the operating assumptions embedded in AI systems will increasingly be shaped elsewhere."

He says governance in New Zealand must also reflect its bicultural constitutional setting.

"Maori data sovereignty principles emphasise guardianship and collective rights over data. 

"An imported regulatory model will not automatically reflect those obligations."

"We should not simply copy Europe, we have the opportunity to design a framework that reflects New Zealand's legal system, social expectations and Treaty commitments."

Dr. Imtiaz says New Zealand has attracted investment from global cloud providers expanding local data centre capacity. However, he says commercial infrastructure alone does not equate to sovereign control.

He says reliance on offshore infrastructure also introduces physical and geopolitical exposure.

"The key questions are who controls model training, who audits outputs and which jurisdiction ultimately sets the compliance standards. Infrastructure location is only one part of that equation.

"If critical government systems are hosted or processed overseas, New Zealand inherits the resilience profile of that infrastructure. That includes exposure to foreign jurisdiction, regulatory shifts, and in extreme cases, physical attack or disruption to data centres or network routes beyond our control."

He says while such events are unlikely, national infrastructure planning traditionally accounts for low-probability, high-impact risks.

"We would not build an electricity grid or water system without considering resilience. AI systems are increasingly embedded in essential services. The same logic should apply."

Dr. Imtiaz says under the EU regime, high-risk AI systems must undergo conformity assessments and demonstrate ongoing human oversight. He says standards are expected to shape global product design in much the same way Europe's privacy rules reshaped international data protection.