TechnologyUK Freedom of Information Laws Overhaul in 2025

UK Freedom of Information Laws Overhaul in 2025

Key Takeaways

  • The UK government has set a legal precedent for handling interactions with AI chatbots after releasing Peter Kyle’s ChatGPT logs.
  • The release of the chat interactions was a shock to data protection and FOI experts, with many considering it a world first.
  • The UK government has become more cautious around FOI requests, especially concerning AI use, since the story was published.
  • Governments are adopting AI at a rapid pace, making transparency over its use crucial to ensure accountability and understand potential biases.
  • The release of the chat logs has gained international interest, with researchers in other countries seeking to obtain similar information from their governments.

Introduction to the Story
When I fired off an email at the start of 2025, I hadn’t intended to set a legal precedent for how the UK government handles its interactions with AI chatbots, but that is exactly what happened. It all began in January when I read an interview with the then-UK tech secretary Peter Kyle in Politics Home. Trying to suggest he used first-hand the technology his department was set up to regulate, Kyle said that he would often have conversations with ChatGPT. That got me wondering: could I obtain his chat history? Freedom of information (FOI) laws are often deployed to obtain emails and other documents produced by public bodies, but past precedent has suggested that some private data – such as search queries – aren’t eligible for release in this way.

The FOI Request and Its Outcome
I was interested to see which way the chatbot conversations would be categorised. It turned out to be the former: while many of Kyle’s interactions with ChatGPT were considered to be private, and so ineligible to be released under FOI laws, the times when he interacted with the AI chatbot in an official capacity were. So it was that in March, the Department for Science, Industry and Technology (DSIT) provided a handful of conversations that Kyle had had with the chatbot – which became the basis for our exclusive story revealing his conversations. The release of the chat interactions was a shock to data protection and FOI experts. “I’m surprised that you got them,” said Tim Turner, a data protection expert based in Manchester, UK, at the time. Others were less diplomatic in their language: they were stunned.

International Interest and Subsequent Requests
When publishing the story, we explained how the release was a world first – and getting access to AI chatbot conversations went on to gain international interest. Researchers in different countries, including Canada and Australia, got in touch with me to ask for tips on how to craft their own requests to government ministers to try to obtain the same information. For example, a subsequent FOI request in April found that Feryal Clark, then the UK minister for artificial intelligence, hadn’t used ChatGPT at all in her official capacity, despite professing its benefits. But many requests proved unsuccessful, as governments began to rely more on legal exceptions to the free release of information.

The UK Government’s Response and Increased Caution
I have personally found that the UK government has become much cagier around the idea of FOI, especially concerning AI use, since my story for New Scientist. A subsequent request I made via FOI legislation for the response within DSIT to the story – including any emails or Microsoft Teams messages mentioning the story, plus how DSIT arrived at its official response to the article – was rejected. The reason why? It was deemed vexatious, and sorting out valid information that ought to be included from the rest would take too long. I was tempted to ask the government to use ChatGPT to summarise everything relevant, given how much the then-tech secretary had waxed lyrical about its prowess, but decided against it.

The Importance of Transparency in AI Use
Overall, the release mattered because governments are adopting AI at pace. The UK government has already admitted that the civil service is using ChatGPT-like tools in day-to-day processes, claiming to save up to two weeks’ a year through improved efficiency. Yet AI doesn’t impartially summarise information, nor is it perfect: hallucinations exist. That’s why it is important to have transparency over how it is used – for good or ill. As AI becomes more integrated into government processes, it is crucial that there is accountability and understanding of potential biases. The release of Peter Kyle’s ChatGPT logs has set a precedent for how governments handle interactions with AI chatbots, and it will be interesting to see how this develops in the future.

- Advertisement -spot_img

More From UrbanEdge

CISA Mandate: Upgrade & Identify Unsupported Edge Devices for Agencies

CISA mandates federal agencies to replace unsupported edge devices prone to advanced threat actor exploits. Agencies have three months to identify, 12 months to begin upgrades, and 18 months for full remediation to protect network perimeters from cyber threats. SecureEdge Solutions offers assistance in securing network vulnerabilities...

Coinbase Insider Breach: Leaked Support Tool Screenshots

In May 2025, Coinbase experienced a sophisticated insider breach affecting 70,000 users. Hackers bribed support agents to leak sensitive data, resulting in over $2 million in theft through targeted scams. Coinbase responded by refusing ransom, launching a bounty program, and refunding victims...

Sector Impact Overview: Architecting the AI Integration Era

Sector Impact Overview: Architecting the AI Integration Era 1. Introduction:...

The Pulse of the Global Artificial Intelligence Landscape

This collection of news headlines highlights the rapidly evolving landscape...

NSW Police Tighten Protest Rules Ahead of Israeli President’s Visit

Key Takeaways The NSW Police commissioner has announced an extension...

Meet Team USA’s Most Seasoned Athlete: A Midwest Curler Bound for 2026 Olympics

Key Takeaways Rich Ruohonen, a 54-year-old curler from Minnesota, is...

Maddie Hall Inquest: Family Seeks Answers Over Mental Health Failures

Key Takeaways Madeleine Hall, a 16-year-old girl, died by suicide...

Will Arnett Booted Famous Comedian from Podcast After Just 10 Minutes

Key Takeaways: Will Arnett shares a harsh opinion about a...

Insider Threat: How Unhappy Employees Compromise Data Security

Key Takeaways Disgruntled employees pose a significant cybersecurity threat to...
- Advertisement -spot_img