Primary tabs

‘Your bot of choice is not a filing cabinet’

By Eliza.Compton, 26 February, 2026
With GenAI taking a larger role in research and education, Sorin Krammer looks at the data management habits academics can no longer ignore
Article type
Article
Main text

A senior professor toggled a single privacy switch in ChatGPT. In an instant, two years of his carefully structured grant applications, lecture notes, publication drafts and exam analyses disappeared. No warning. No undo option. Just a blank page. 

The story, published as a career column in Nature last month, attracted widespread sympathy, and understandably so. Losing months or years of precious work is genuinely distressing. But sympathy, however well deserved, should not obscure a more important lesson about how researchers and educators, as well as other individuals and businesses, manage their digital work – namely that over-reliance on commercial AI platforms carries serious, foreseeable risks. Well intentioned as it is, this account inadvertently normalises deeply problematic practices among academics and business professionals, early career ones in particular, who are more likely to adopt AI without critical scrutiny.

AI’s contradiction is worth examining

The author acknowledges understanding that ChatGPT produces “seemingly confident but sometimes incorrect statements”, yet they deemed it reliable enough to serve as the sole repository for two years’ worth of academic work. This contradiction is striking. If you don’t trust a system’s core functions, why trust it as your primary data storage? 

Further, this concern deepens when we consider that data is the main input for large language model (LLM) training – meaning that AI companies’ incentives are to harness and use as much as possible – and that there are numerous, well-documented concerns, both ethical and legal, about these companies’ handling of user data.

Legal, institutional and ethical dimensions

Treating ChatGPT Plus as a primary repository for academic content is inappropriate for many reasons. First, ChatGPT Plus is not GDPR-compliant, a non-trivial concern for any researcher operating within the European regulatory context, for whom this is not a matter of preference but of legal obligation. 

Second, exclusive reliance on ChatGPT violates one of the basic principles of data management: having multiple back-ups to avoid critical data loss. Most research institutions have explicit data management policies, many of which have been recently updated to address AI tools usage as well. Typically, these policies require that primary research materials be stored on institutional or otherwise secure and recoverable systems. Sensitive research or academic records should never be deposited on third-party, unvetted platforms. 

Third, given the scope of the Nature account, we could conclude that ChatGPT Plus was an integral part in the creation of these lectures, grant applications and research publications. This casts doubt on the independence of these outputs. When intellectual work is co-developed iteratively inside a proprietary commercial platform, questions of ownership, originality, reproducibility and individual contributions become difficult to resolve. These are not peripheral concerns; they bear directly on research integrity, as illustrated by a growing number of cases that span research, teaching and funding. This problem is already mutating in unforeseen directions, from researchers inserting hidden instructions in manuscripts designed to manipulate AI peer review towards favourable verdicts, to fully AI-generated papers spanning the entire research process from the inception of ideas through to analysis and writing.

Platform accountability versus user responsibility

In the original article, the professor frames this incident as a failure of OpenAI, flagging that there was no warning or option to undo the deletion. Yet OpenAI did nothing wrong. The company implemented exactly what privacy-by-design requires: immediate and permanent deletion upon user request. The author acknowledges this, yet stops short of taking responsibility for his own judgement and digital hygiene. OpenAI’s response was not a malfunction; it was the system working precisely as documented and agreed to by the user prior to use. Subsequently, characterising this incident as a platform failure seems rather tenuous, whether viewed through the lens of responsible AI use or basic common-sense judgement.

What responsible AI use looks like

The broader takeaway is not that AI tools are unsafe for academic and professional use, but that integrating them responsibly requires the same basic hygiene we apply to any system. Several principles deserve wider attention:

  • Commercial AI chat platforms should never serve as primary repositories for academic work. Local, institutional or version-controlled back-ups should always be maintained.
  • GDPR compliance and institutional data governance policies must be consulted before using any AI tool with sensitive or confidential research materials. Beyond data loss, deviations from established institutionalised practices may have serious legal implications. Most universities have now published guidance on exactly this.
  • Privately owned cloud services are governed by various user agreements that often detail data deletion and training use. Researchers that choose such commercial services bear the responsibility of reading and agreeing with their terms. Reliability and data stewardship vary considerably across providers, and a monthly subscription fee is no guarantee of either. 

In sum, commercial platforms should not be used as primary repositories for academic work, and user error cannot be reframed as platform failure. Clearer interface warnings from OpenAI would certainly have helped, and the experience does point to a genuine design gap worth addressing. But the real cautionary tale here is not about AI tools; it is about the fundamental importance of proper data management, regardless of which systems you use.

AI has a legitimate and increasingly important role in academic and business workflows. But that role requires deliberate governance, not just enthusiasm. The question is not whether to use these tools, it is whether we are prepared to use them wisely. Your bot of choice is not a filing cabinet. Treat it accordingly.

Sorin M. S. Krammer is a professor of strategy and international business at the University of Southampton.

If you would like advice and insight from academics and university staff delivered direct to your inbox each week, sign up for the Campus newsletter.

Standfirst
With GenAI taking a larger role in research and education, Sorin Krammer looks at the data management habits academics can no longer ignore

comment