HNNewShowAskJobs
Built with Tanstack Start
When two years of academic work vanished with a single click(nature.com)
40 points by stmw 4 days ago | 65 comments
  • haritha-j2 hours ago

    > to write e-mails, draft course descriptions, structure grant applications, revise publications, prepare lectures, create exams and analyse student responses, and even as an interactive tool as part of my teaching.

    So your grant applications were written by AI, your lectures were written by AI, your publications were written by AI, and your students exams were marked by AI? Buddy what were YOU doing?

    • c0482 hours ago |parent

      To make it even worse, he deemed none of it important enough to backup.

      AI fan or not, that is astoundingly stupid.

  • almusdives3 hours ago

    The lost "work" is two years of ChatGPT logs. Sounds like AI systems had concrete benefits to this researcher in a number of applications, but I'm not sure how I feel that their discussions with AIs are so casually being described as "work". Seems slightly misleading?

    • PunchyHamster3 hours ago |parent

      People really do think that throwing prompts at AI is equivalent quality with actual work.

      • Dansvidania2 hours ago |parent

        I ask with genuine curiosity: have you ever talked about AI as a tool? If so, does it not count at all as work?

        I lean towards no, but I routinely call it a tool. I can’t reconcile the two statements.

    • 2 hours ago |parent
      [deleted]
  • Brajeshwar2 hours ago

    I had had my fair share of data loss lessons. For us, it is easy to say, “Why didn’t you back up?” But most people have an innate trust in tools, especially from big companies such as Apple, Google, and Facebook/Meta. I have heard so many people happily claim, “I won’t worry. I have it on Google.”

    Even for my daughters’ much simpler school homework, projects, and the usual drawings/sketches, I’ve set up Backups so they don’t cry when their work gets lost. I set up the Macs I handed down to them to be backed up to iCloud, and added a cheap HDD for Time Machine. They think I’m a Magician with Computers when I teach them to use the Time Machine and see the flying timeline of their work. The other thing is the Google Workspace for Schools. I have found that having a local copy always available via a tool (such as InSync) does wonders.

    The only sob story now is Games. They sometimes lose points, the game coin thingies, and developer-kids with bugs that reset gameplay earnings. I have no idea how to help them there besides emotional support and how the world works — one step at a time.

    How about if ChatGPT/Claude writes a local Markdown copy of each conversation? Won’t that be Nice?

    • snowe20102 hours ago |parent

      I don’t really think it counts as data loss when you actually choose to delete it and get a prompt confirming as such.

      • perching_aixan hour ago |parent

        I don't think such an idea is consistent with the existence of trashbin features, or the non-insignificant use of data recovery tools on normally operating devices.

        I can definitely see the perspective in clarifying that ChatGPT didn't lose anything, the person did, but that's about it.

  • wodenokoto2 hours ago

    A lot of snark in the comments, but I think author is absolutely right: this should har come with a big warning and that warning should have had 3 options:

    1) go ahead and delete everything 2) back up and then go ahead 3) abort and keep things as they are

    ChatGPT definitely wants to be the copilot of all your work. Guy didn’t just have chats, he had drafts that his virtual assistant helped formulate and proof read. Give how big and used ChatGPT has become, it shouldn’t be a surprise to anyone tech savvy that this is being used for serious work outside of vibecoders.

    • dalmo32 hours ago |parent

      I still don't understand what could've have happened here. I'm not a chatgpt user so I'm not familiar with the UI.

      He starts out saying he "disabled data consent". That wording by itself doesn't mean delete the content at all. The content could theoretically live in local storage etc. He says the data was immediately deleted with no warning.

      Then OpenAI replies that there is a confirmation prompt, but doesn't say what the prompt says. It could still be an opaque message.

      Finally, he admits he "asked them to delete" the data.

    • perching_aix2 hours ago |parent

      It's always interesting to see how hostile and disparaging people can start to act when given the license. Hate AI yourself, or internalize its social stading as hated even just a little, and this article becomes a grand autoexposé, a source for immense shame and schadenfreude.

      • m_rpn2 hours ago |parent

        The shame is not that he was so imbecile to not have appropriate backups, it is that he is basically defrauding his students, his colleagues, and the academic community by nonchalantly admitting that a big portion of his work was ai-based. Did his students consent to have their homework and exams fed to ai? Are his colleagues happy to know that probably most of the data in their co-authored studies where probably spat out by ai? Do you people understand the situation?

        • perching_aix2 hours ago |parent

          > The shame is not that much that he was so imbecile to not have appropriate backups

          These comments vehemently disagree:

          https://news.ycombinator.com/item?id=46726570

          https://news.ycombinator.com/item?id=46777039

          https://news.ycombinator.com/item?id=46777123

          > ... (the rest) ...

          It's not that I don't see or even agree with concerns around the misuse and defrauding angle to this, it's that it's blatantly clear to me that's not why the many snarky comments are so snarky. It's also not as if I was magically immune to such behavioral reflexes either, it's really just regrettable.

          Though I will say, it's also pretty clear to me that many taking issue with the misuse angle do not seem to think that any amount or manner of AI use can be responsible or acceptable, rendering all use of it misuse - that is not something agree with.

    • m_rpn2 hours ago |parent

      The author is an absolute and utter embarrassment for all the good academic professionals out there, and he is also literally admitting to defrauding his students of their precious money, which they thought was going to human-led instruction, he's also put all of his colleagues in an very dodgy position right now. It is preposterous that we are even arguing about it, it is the sign of how much AI-sloppiness is permeating our lives, it is crazy to think that you can be entitled to give years of work to a chatbot without even caring and then write an article like this "uh oh, ai eat my homework".

  • rspoerri3 hours ago

    Never rely on any subscription based service for any data that is important. Never use data formats that lock you in. Especially not online services without (automatic) export options.

    Keep a copy (cloud) and a backup (offline) for all you own data.

    • GJim2 hours ago |parent

      > Especially not online services without (automatic) export options

      GDPR gives the right to data portability.

      https://gdpr-info.eu/art-20-gdpr/

      Any _legitimate_ organisation following the GDPR will allow export of your data; and you shouldn't be stupid enough to trust sensitive or valuable data with some dodgy organisation that doesn't follow the GDPR.

      Obviously this doesn't alleviate the need for proper offline backups of your own valuable data!

    • giik2 hours ago |parent

      cloud is a subscription service too...sadly

      hardcopy and offline copy(ies) is the only way, if one is able to maintain it in a disciplined way

    • Joel_Mckay3 hours ago |parent

      As someone that just lost a OS build project last week, I remind myself the off-site physical copy and optical media do still have advantages.

      1. kernel update on pre-release OS did something weird to the usb driver

      2. heavily corrupted disk enters device hardware into lock-out

      3. Pull ISO snapshot, and reinstall OS bare bone SSH install in Qemu session

      4. verify OS working in VM

      5. install new OS USB device in PC, boot once, and watch the kernel bug nuke the drive again

      6. Boot from last weeks backup boot disk image, and watch the kernel bug nuke the drive again

      7. install old kernel OS, and verify it was not a hardware failure

      8. Remember to Backup, FOSS can byte hard when and not if things go sideways.

      Best regards =3

  • stefanfisk3 hours ago

    I really liked Angela Collier’s rant on this topic https://youtu.be/7pqF90rstZQ.

    ”Dr Flattery always wrong robot” is such a wonderful way to describe ChatGPT and friends when used like this <3

    • Dansvidania2 hours ago |parent

      I giggled when I saw this article quoted here after watching her video.

      I particularly feel close to the “how do you just admit to this and publish it on nature” feeling.

      Are we getting desensitised to the boundary between self and ai?

      On the other side of it, if we speak of AI as a tool, does it not count as work to prompt and converse ?

      • stefanfisk2 hours ago |parent

        I see a modern version of this scene playing itself out in orgs across the world in a few years. ”What would you say you do here?”

        https://youtu.be/z4Xw6WMvfNk

  • d7w3 hours ago

    This issue looks like a situation where one person stored all their files and folders in the Windows Recycle Bin and somebody emptied it.

    It might be my professional deformation, but I never store anything in ChatGPT and Claude for longer than a day or two.

    • d1sxeyes2 hours ago |parent

      That’s not very generous. Keeping files in the Recycle Bin is an incorrect use of the Recycle Bin. Keeping conversations in your ChatGPT history is how it’s supposed to be used.

    • Brajeshwar3 hours ago |parent

      In ChatGPT and Claude, what settings allow data to be stored only for a set period of time/day?

  • muppetman3 hours ago

    Are we supposed to feel sorry for this person, or chuckle at them? This is like storing all your data on a floppy disk you never back up and then accidentally dropping it in the toilet.

    • duskdozer2 hours ago |parent

      As much as I want to dunk because it's an AI user, I do think it's really frustrating and bad design when actions don't clearly indicate they will make permanent deletions. I've been bitten by similar things before because the effects weren't obvious to me, in part because when I design things I automatically think to give a warning before doing things that may be unintuitive. Even if I do have backups, it's usually a big annoyance for me to restore, and I'd rather never need to.

    • perching_aix2 hours ago |parent

      I have a hard time chuckling at data loss. Espcially given that exporting and backing up your data from online services has an even smaller tradition than taking backups of one's local data, which is sadly rare in itself on the individual level.

    • kranner3 hours ago |parent

      What's the expression? Kein Backup, kein Mitleid.

      Not that I've ever had the heart to say it to a friend who has shown up with a computer that won't boot and data they must recover. Sometimes it's the same friend a second time.

    • voidUpdate3 hours ago |parent

      I think even a toileted floppy would be fine if you dried it off, surely? You can't exactly flush it unless you have one of those tiny floppies

  • bekoeppel27 minutes ago

    I've built a tool to turn an export of the ChatGPT conversations into Markdown files, ideal for saving them e.g. to Notion: https://chat-capsule.com/

  • storystarling4 days ago

    This is exactly why I don't rely on the web UI for anything critical. It seems like a mistake to treat a chat log as a durable filesystem. I just hit the API and store the request/response pairs in a local Postgres DB. It's a bit of extra boilerplate to manage the context, but at least I own the data and can back it up properly.

  • BrenBarn2 hours ago

    I'm pretty anti-AI but this isn't really anything to do with AI. The same problem would arise with any online service that you use to hold important data. And it's pretty evil for any such service to have a trap "delete all my stuff with no warning" button.

    • perching_aixan hour ago |parent

      I wonder what would be a better flow.

      One big reason I can think of that would make one want a permanent data purge feature, is that the data is not on their premises but on the service provider's. I think GDPR might even require such a feature under a similar rationale.

      So maybe a better formulation would be to force the user to transfer out a copy of their data before allowing deletion? That way, the service provider could well and truly wash their hands of this issue.

      • BrenBarn43 minutes ago |parent

        Forcing an export is an interesting idea. But, like, from the article it sounds like almost anything would be a better flow. It didn't even warn that any data would be deleted at all.

        • perching_aix3 minutes ago |parent

          Yeah, that's true.

          One further refinement I can think of is bundling in a deletion code with the export archive, e.g. a UUID. Then they could request the user to put in that code into a confirmation box, thereby "guaranteeing" they did indeed downloaded the thing.

          Wouldn't really be a guarantee in technical actuality, but one really needs to go out of their way to violate it.

      • duskdozeran hour ago |parent

        >This will delete all the data we have stored for you. Would you like to [export your data](...) first?

  • orwinan hour ago

    The question i have: did he anonymize or ask his students before putting their name and work into ChatGPT? because that's a RGPD violation if he didn't.

  • flowerthoughts2 hours ago

    Those are rookie numbers.

    This guy [1] (in Swedish) was digitizing a municipal archive. 25 years later, the IT department (allegedly) accidentally deleted his entire work. With no backup.

    Translated:

    > For at least 25 years, work was underway to create a digital, searchable list of what was in the central archive in Åstorp municipality. Then everything was deleted by the IT department.

    > “It felt empty and meaningless,” says Rasko Jovanovic.

    > He saw his nearly 18 years of work in the archive destroyed. HD was the first to report on it.

    > “I was close, so close to taking sick leave. I couldn't cope,” he says. The digital catalog showed what was in the archive, which dates back to the 19th century, and where it could be found.

    > "If you ask me something today, I can't find it easily, I have to go down and go through everything.

    > “Extremely unfortunate”

    > Last fall, the IT department found a system that had no owner or administrator. They shut down the system. After seven months, no one had reported the system missing, so they deleted everything. It was only in September that Åstorp discovered that the archive system was gone.

    > “It's obviously very unfortunate,” says Thomas Nilsson, IT manager. Did you make a mistake when you deleted the system?

    > “No. In hindsight, it's clear that we should have had different procedures in place, but the technician who did this followed our internal procedures.”

    In typical Swedish fashion, there cannot have been a mistake made, because procedures were followed! Or to put it in words that accurately reflect having 25 years of work removed: "Own it, you heartless bastard."

    Translated with DeepL.com (free version) [1] https://www.svt.se/nyheter/lokalt/helsingborg/rasko-digitali...

  • arjie3 hours ago

    A typical example of Hyrum's Law: ...all observable behaviors of your system will be depended on by somebody. It's like how your draft folder feature will be used as a secret messaging app by a general and his mistress, or as Don Norman points out, your flat topped parapet will be used as a table for used cups, or your reliable data store of chats will be used as academic research storage.

    But I have to say, quite an incredible choice! ChatGPT released in Nov 2022. This scientist was an early adopter and immediately started putting his stuff in there with the assumption that it would live there forever. Wow, quite the appetite for risk.

    But I can't call him too many names. I have a similar story of my own: one thing I once did was ETL a bunch of advertising data into a centralized data lake. We did this through the use of a Facebook App that customers would log in to and authorize ads insights access to. One of the things you need to do is certify that you are definitely not going to do bad things with the data. All we were doing was calculating ROAS and stuff like that: aggregate data. We were clean.

    But you do have to certify that you are clean if you even go close to user data, which means answer a questionnaire (periodically). I did answer the questionnaire, but for everyone who has used anything near Meta's business and advertising programs (at the control plane, the ad delivery plane must be stupendous) you know they are anything but reliable. The flaky thing popped up an alert the next day that I had to certify again and it wouldn't go away. Okay, fine, I do need the one field but how about I just turn off the permission and try to work without it. I don't want anyone thinking I'm doing shady stuff when I'm not.

    Only problem? If you have an outstanding questionnaire and you want to remove a permission you have to switch from Live to Development. That's fine too, normally, it's a 5 second toggle. Works every time. Except if you have an outstanding questionnaire you cannot switch from Development to Live. We were suddenly stuck, no data, nothing and every client is getting this page about app not approved. And there's nothing to be done but to beg Meta Support who will ignore you. I just resubmitted the app and we waited 24 hours and through the love of God it all came back.

    But I was oh-so-cavalier clicking that bloody button! The kind of mistake you make once before you treat any Data Privacy Questionnaire like it's the Demon Core.

  • ifh-hn3 hours ago

    Is it deleted though? Last I heard there was a court case or some such that required them retaining all data for a lawsuit, did that go away?

    • CjHuber3 hours ago |parent

      It was because of the NYT OpenAI case, however since mid October they are no longer under that legal order. What they keep retaining now and what not, nobody knows but even if they still had the date they surely wouldn't blow their cover

  • xvxvx4 days ago

    The 2nd comma he uses is incorrect. Did he also use ChatGPT for this article?

    I frown when people currently trust AI, let alone have been doing so for 2 years already.

    • hastily31143 hours ago |parent

      If using commas incorrectly means that you are AI, then I must be a bot.

  • thanzex3 hours ago

    > [...] two years of carefully structured academic work disappeared [...]

    > [...] but large parts of my work were lost forever [...]

    I wouldn't really say parts of his work were lost. At most the output of an AI agent, nothing more.

    If somehow e-mails, course descriptions, lectures, grant applications, exams and other tools, over the period of two years disappeared in an instant, they did not really exist to begin with.

    For once, the actual important stuff is the deliverable of these chats, meaning these documents should exist somewhere. If we're being honest everything should be able to be recreated in an instant, given the outputs and if the actual intellectual work was being done by Mr. Bucher.

    Does it suck to lose data? Even if just some AI tokens we developed an attachment to? Sure.

    Would I have outed myself and my work shamelessly, to the point that clicking a "don't retain my data" option undermines your work like this? Not really.

  • wewewedxfgdf2 hours ago

    ChatGPT did not lost it.

    The user without backups lost their own work.

    Simple as that, no argument.

    No backups, you the loser.

    You might WANT someone else to be responsible but that doesn't change anything.

  • GuestFAUniverse3 hours ago

    Ever heard of backup? Multiple copies? Offline?

    How can you loose "important work" of multiple years? -- can't be important and how can somebody _expected to become management_ be so incompetent?

    "...two years of carefully structured academic work disappeared. No warning appeared. There was no undo option. Just a blank page. Fortunately, I had saved partial copies of some conversations and materials, but large parts of my work were lost forever." -- stupid: that drive could have died, the building could have burned down, the machine could have been stolen, the data could have been accidentally deleted... and all there was: "a partial" backup.

    I mean, that isn't even a scenario where he didn't know about the data ("carefully structured") and discovered it wasn't covered by the backup schema (that would be a _real_ problem) Another problem would be of your churn is so high that backing up becomes a real issue (bandwidth, latency, money, ...). None of that applies.

    And yet they reserve a spot in "nature" for such whining and incompetence?

    • rossant2 hours ago |parent

      Genuine question: how do you backup (and reimport) a bunch of ChatGPT conversations?

      • perching_aix2 hours ago |parent

        Not aware of reimporting, but there's an export all data option in the settings, which works pretty well.

  • jstummbillig2 hours ago

    Hot take: Actual "irreversibly delete x stuff with the next action" is simply too powerful and bad design for most people, and has probably caused considerably more harm than good in the world. It's particularly silly with software, where few reasons exist for this to be an actual thing.

    What the average human needs is laws and enforcement, and trust in both.

  • 2 hours ago
    [deleted]
  • KnuthIsGod4 days ago

    A plant science academic who can't be bothered to back up their work...

    If that was the intellectual calibre of the person, I wonder how truly worthwhile the lost work was.

    • thunfischbrot3 hours ago |parent

      How fair would you think if a potential employer assumed from your (hypothetical) incompetence in picking a suitable hairstyle and outfit for an interview that you were not fit for a non-customer facing role?

      While it absolutely makes sense to keep your important data backed up, I know people who were great academics in their field and yet managed to delete all their PhD work (before services like Dropbox and OneDrive became common).

    • actionfromafar4 days ago |parent

      That's rough to assume IMHO. I don't assume either way, but I think most have met someone who is briliant in one domain but just hopeless on a lot of other things.

  • bell-cot3 hours ago

    I recall hearing "she saved everything for her doctoral thesis on that one floppy disk, and then..." warning stories back in the 1980's USA.

    Is anyone familiar with current academic culture in Germany, to comment on how (or if) it warns its members about such risks?

    • slow_typist3 hours ago |parent

      From my perspective as an employee of a German academic institution, administrations are still figuring out if and how to regulate the use of AI, while some professors rely pretty heavily on AI tools, so the story is completely believable. However the double naïveté demonstrated here is strange.

      Generation of boilerplate prose for grant applications was the beginning around 2023, which is absolutely understandable. The DFG recently allowed the use of AI for reviewers, too, to read an summarise the AI generated applications.

      Researchers using qualitative methods seem (in general) to be more sceptical.

      I wish we had an open debate about the implications instead of half assed institutional rules that will and must be widely ignored by researchers.

  • lazylizarda day ago

    once upon a time i had a boss who asked for a "super admin" account to "trump" the domain administrators..and a "master key" to decrypt any file , in case the user lost their key.

    • aleph_minus_one2 hours ago |parent

      > once upon a time i had a boss who asked for a "super admin" account to "trump" the domain administrators..and a "master key" to decrypt any file , in case the user lost their key.

      Key escrow is a well-known concept in cryptography:

      > https://en.wikipedia.org/wiki/Key_escrow

      It's just that these "master keys" are super-dangerous to handle for obvious reasons.

  • m_rpn2 hours ago

    The issue is not backup, the issue is that he is publicly and nonchalantly admitting that most of his work for the past years was ai-based, which might or might not constitute fraud given his professional position. Imagine being a student paying thousands over thousands expecting an expert human led instruction just to get this, imagine being a fellow researcher and suddenly being in a situation of not being able to trust this guy's current and past work.

    The worst thing is all the people looking at this behaviour as normal and totally acceptable, this is where ai-sloppiness is taking us guys. I hope it's just the ai bros talking in the comments, otherwise we are screwed.

  • nice_byte3 hours ago

    relevant: https://www.youtube.com/watch?v=7pqF90rstZQ

    • bigiain3 hours ago |parent

      Came to post that.

      I like her meta observation, that using ChatGPT for 2 years rots your brain so badly you somehow think it's a good idea to write an article like this, with your real name and professional/academic reputation attached to it, and get it published somewhere as high profile as Nature.

      Someone on my Mastodon field commented that if they'd done that "you wouldn't be able to torture it out of me" and that they'd never admit it to anyone.

    • onli2 hours ago |parent

      Good commentary, good video. She is a little bit too harsh about the data loss though. The author did not realize that disabling data sharing would delete the history of the already occurred interactions, probably not realizing that everything was stored on an external server. And it's quite possible there was no proper warning about that.

      I feel that makes her point weaker. Because she is apart from that completely right: The work practice admitted to here is horrible. It likely includes leaking private emails to an US company and in every case meant the job of teaching and publishing wasn't don't properly, not even close.

    • jdhendrickson3 hours ago |parent

      Such a great response to this article. Scathing but enjoyable to watch. Angela is a treasure.

  • hacker_homie3 hours ago

    they are training you conditioning you to never push a button like that ever again.

    How dare you not let us steal your data.

  • olya_pllkh2 hours ago

    [dead]