Responses

Download PDFPDF

Original research
ChatGPT (GPT-4) versus doctors on complex cases of the Swedish family medicine specialist examination: an observational comparative study
Compose Response

Plain text

  • No HTML tags allowed.
  • Web page addresses and e-mail addresses turn into links automatically.
  • Lines and paragraphs break automatically.
Author Information
First or given name, e.g. 'Peter'.
Your last, or family, name, e.g. 'MacMoody'.
Your email address, e.g. higgs-boson@gmail.com
Your role and/or occupation, e.g. 'Orthopedic Surgeon'.
Your organization or institution (if applicable), e.g. 'Royal Free Hospital'.
Statement of Competing Interests

PLEASE NOTE:

  • A rapid response is a moderated but not peer reviewed online response to a published article in a BMJ journal; it will not receive a DOI and will not be indexed unless it is also republished as a Letter, Correspondence or as other content. Find out more about rapid responses.
  • We intend to post all responses which are approved by the Editor, within 14 days (BMJ Journals) or 24 hours (The BMJ), however timeframes cannot be guaranteed. Responses must comply with our requirements and should contribute substantially to the topic, but it is at our absolute discretion whether we publish a response, and we reserve the right to edit or remove responses before and after publication and also republish some or all in other BMJ publications, including third party local editions in other countries and languages
  • Our requirements are stated in our rapid response terms and conditions and must be read. These include ensuring that: i) you do not include any illustrative content including tables and graphs, ii) you do not include any information that includes specifics about any patients,iii) you do not include any original data, unless it has already been published in a peer reviewed journal and you have included a reference, iv) your response is lawful, not defamatory, original and accurate, v) you declare any competing interests, vi) you understand that your name and other personal details set out in our rapid response terms and conditions will be published with any responses we publish and vii) you understand that once a response is published, we may continue to publish your response and/or edit or remove it in the future.
  • By submitting this rapid response you are agreeing to our terms and conditions for rapid responses and understand that your personal data will be processed in accordance with those terms and our privacy notice.
CAPTCHA
This question is for testing whether or not you are a human visitor and to prevent automated spam submissions.

Vertical Tabs

Other responses

Jump to comment:

  • Published on:
    Comment on: ChatGPT (GPT-4) versus doctors on complex cases of the Swedish family medicine specialist examination: an observational comparative study.
    • Denise E. Hilling, Surgeon, CSO Erasmus MC Datahub Erasmus MC, University Medical Center Rotterdam
    • Other Contributors:
      • Michel E. van Genderen, Internist-Intensivist, Director Erasmus MC Datahub
      • Jan A.J.G. van den Brand, CTO Erasmus MC Datahub

    With great interest we read the recently published article titled "ChatGPT (GPT-4) versus doctors on complex cases of the Swedish family medicine specialist examination: an observational comparative study" by Arvidsson et al. [1] The study provides valuable insights into the capabilities and limitations of generative AI models like GPT-4 in complex medical decision-making scenarios. However, the study's approach, which relies on GPT-4 as a general-purpose model without any domain-specific fine-tuning or optimised prompting strategies, presents an inherent limitation. Deploying an AI system in such a manner is fundamentally inferior and does not align with best practices in any industry. In real-world applications, AI models are typically customised, fine-tuned, or integrated with structured knowledge bases to enhance their relevance and reliability in specific domains. The zero-shot prompting approach used in this study, while convenient for initial evaluation, does not reflect the practical implementation of AI solutions in healthcare or other high-stakes industries.

    In the medical field, AI applications must be trained and validated within a well-defined context, leveraging domain-specific data, tailored prompts, and reinforcement learning with human feedback to improve performance over time. Successful AI implementation in healthcare involves collaboration with medical professionals to refine model outputs, ensuring that the AI system aligns with c...

    Show More
    Conflict of Interest:
    None declared.
  • Published on:
    Declaring the context helps

    Hi,

    I did also try this - started from DoctorAI, and noticed that answers were to generic to fit in Swedish perspective, so I added
    "I Svensk primärvårdskontext" (In Swedish primary care context) and got better results.

    Conflict of Interest:
    None declared.