17 comments

  • krisoft 1 hour ago
    What a mess.

    > One author of a case report was surprised to learn of the correction — because the case described in her article is true.

    So they managed to mess up even the correction of their giant mess.

    > correcting the correction "would be difficult."

    I bet. That's why they should have got it right in the first place. I would be absolutely ballistic if they would be libelling my work like that.

    • SiempreViernes 46 minutes ago
      Yeah, they seem to have been quite sloppy with these vignettes.

      Thought note that in the situation of the mislabeled real case, the formal solution is could be a retraction of the entire highlight article since it is against the (poorly implemented) policy to have a real case study.

      Don't know how patient consent for being used in a case study works, did this author get a perpetual license, did they just copy something from another article they wrote, or from an article someone else wrote?

  • sekuraai 0 minutes ago
    They had access to ChatGPT for last 25 years!
  • programmertote 1 hour ago
    Speaking this as a spouse of a medical doctor -- case reports are sometimes a good way to increase the bullet point count in your CV if you are a medical resident. A lot of residents do that just for the sake of beefing up their CVs (to apply for fellowship for example).
  • damnesian 7 minutes ago
    Too late, it's already in the bloodstream, LLMs will be recommending things to pediatric doctors and families from fabricated archives for years, probably.
    • Towaway69 3 minutes ago
      It’s all an hallucination.
  • TomMasz 8 minutes ago
    This is fine, though somewhat belated. But it does nothing to deal with the public's growing distrust of science in general, and medical science in particular.
  • SiempreViernes 30 minutes ago
    I think this is mainly a case of the common "didn't notice when crucial literature for own published content was retracted, get caught with pants down when the replication police come knocking".

    Obviously the poor labelling is bad, but 9 bad citations per year isn't the end of science and better labelling wouldn't discourage all the lazy authors who chose to cite these highlight articles, it'll just shift whos is to blame.

    The real problem is hosting a review article about research that was retracted, and it sounds like they aren't moving very quickly on taking that piece down.

  • helsinkiandrew 1 hour ago
    > The articles usually start with a case description followed by “learning points” that include statistics, clinical observations and data from CPSP.

    I can see the reason where fictional cases could be used here as teaching aid - based on real cases/ilnesses but simplified to make the learning points succinctly, but surely if the cases are being cited elsewhere someone should have raised the issue earlier?

    • SiempreViernes 56 minutes ago
      Since it was for teaching I expect the case studies were always showing typical features of real cases, so there's nothing in the case vignette itself to give it away unless the author picks a funny name or something like that.

      Rather it would be the entire form of these short highlight articles that would make you keep searching for a proper citation, unless you're lazy or pressed for time.

    • ultropolis 31 minutes ago
      Wouldn't citing actual cases be a HIPAA violation? I can see why they would invent example cases, based on real ones, especially if they are fairly pedestrian cases.

      I mean. Except if your pedestrian example does not reflect reality, then that is bad.

  • sourcegrift 1 hour ago
    In the era of GitHub etc, if you're not giving out every single data point of your research, it should be assumed it's fake.
    • Towaway69 7 minutes ago
      And then there be large amounts of fake data for the next generation of AIs to learn from.

      What is stopping anyone from faking the data they use in their research papers?

      Sure it might be verifiable but if the data was made to give the desired results, i.e. faked to be what is required for the paper.

    • fsh 1 hour ago
      The article is about case reports, not about empirical studies. Putting a fake case report on GitHub wouldn't make it any less fake.
      • qwertox 53 minutes ago
        > Putting a fake case report on GitHub wouldn't make it any less fake.

        Much easier to review for whomever wants to review it.

        • drivingmenuts 27 minutes ago
          Would it be easier, though? Medical records (in the US) are covered by HIPAA and, to my knowledge, there is no anonymized canonical record, similar to what we have for legal decision. Without that, how difficult would it be to just "make shit up"?
        • NewsaHackO 29 minutes ago
          Do you know what a case report is?
        • SiempreViernes 42 minutes ago
          Obviously just sending it via email to the reviewers works just fine in practice anyway, the problem is really that they published a summary piece about research that was later retracted, but didn't take down their own article.
    • avs733 1 hour ago
      out of context that makes sense...but in the context of a case report how do you implement that? The patients have privacy rights and the authors/doctors have a responsibliity to protect them. That doesn't justify this but it does force a conversation about what 'every single data point' means. Does it mean the patient's real name and social security number? their complete medical chart?

      Case reports are descriptive not determinative and should be treated as such by other scholars. They are 'I saw this' not 'this is generalizably true'. They can (and often are) replicated or countered but they are not per se research as you are thinking about it. Whether it is fictitious or not, other scholars should be cautious in citing them as proof/evidence in papers that fit into the 'research' mold.

      • nradov 43 minutes ago
        From a legal perspective, journal article authors can implement this by following the official HHS guidance for de-identification. This applies to any use of protected health information (PHI), not just case reports.

        https://www.hhs.gov/hipaa/for-professionals/special-topics/d...

        The IRB for a particular organization can impose additional restrictions.

  • snapetom 17 minutes ago
    Original HN discussion about the case:

    https://news.ycombinator.com/item?id=46789205

  • insane_dreamer 19 minutes ago
    I don't mind the fact that the case reports were fictional -- actual cases can be problematic in terms of privacy as it may be easy to ascertain the patient's identity from the details -- but not putting a notice that it was fictional (or altered from a real case for privacy), for teaching purposes, is pretty bad.
  • kittikitti 45 minutes ago
    https://onlinelibrary.wiley.com/doi/10.1111/jpc.14206

    Maybe we should revisit the routine practice of infant male genital mutilation?

  • learingsci 1 hour ago
    “Pics or it didn’t happen,” goes a long way in my book.
  • newzino 1 hour ago
    The detail that makes this more than a labeling error: the fictional nature appeared in the journal's author guidelines, not in the published articles. Researchers who cited these 61 papers had no way to distinguish them from genuine case reports. 218 citations later, the fiction is embedded in secondary analyses and literature reviews written by people who had no idea.

    The "Baby Boy Blue" (2010) case is the clearest example of the harm. An infant allegedly exposed to opioids through breast milk. That case influenced clinical guidance on codeine safety in nursing for years. The CARE guidelines (Consensus-based Clinical Case Reporting Guidelines) exist specifically to create transparency in case reporting. They're voluntary, which is how a journal can run a 25-year undisclosed fiction program and technically say the authors knew.

    • SiempreViernes 1 hour ago
      Doesn't sound like these works were "full" articles, but rather something more like short review articles.
  • october8140 2 hours ago
    I think research should be assumed fiction until it’s peer reviewed.
    • contubernio 1 hour ago
      There is not good evidence that peer review improves quality and there is perhaps some to the contrary (many predatory journals are peer reviewed). The arxiv (unreviewed) is among the most reliable sources available.
      • observationist 1 hour ago
        Yeah, it's almost like science is better when the scientific method is applied to everything, instead of delegating validation to some third party based on credentials or authority or social status.
      • ranger_danger 1 hour ago
        What do you suggest instead? Certainly not giving up I hope.
    • Rallen89 1 hour ago
      I think it's a bit different considering the goal was a teaching tool of well recognised conditions

      >all or almost all were cases of very well recognized conditions [...] where a single case report would not generate any interest or ever be cited.

    • readthenotes1 1 hour ago
      That is an ironic proviso given that the article clearly states

      "The peer-reviewed articles don’t state anywhere the cases described are fictional."

      Peer review by peers who are trained by non-replicable science is not helpful...

    • moi2388 1 hour ago
      Independently replicated. Reviewed says pretty much nothing.
      • kergonath 1 hour ago
        Peer review is a sniff test. It cannot guarantee that the results are correct and the conclusions are right. It is just designed to limit some kinds of errors. Replication is important.
      • roywiggins 56 minutes ago
        Case studies can't be replicated. They aren't experiments.
      • ambicapter 44 minutes ago
        Tough to replicate an isolated case study?