Mon. Jul 8th, 2024

ChatGPT generates cancer treatment plans that are full of errors, study shows<!-- wp:html --><p>Artificial intelligence chatbot ChatGPT has been one of the most talked about technologies of the year.</p> <p class="copyright">Jakub Porzycki/NurPhoto via Getty Images</p> <p>Study finds that ChatGPT provided false information when asked to design cancer treatment plans.<br /> The chatbot mixed correct and incorrect information together, making it harder to decipher. <br /> Accuracy issues with Generative AI means it's unlikely to be taking over from doctors any time soon.</p> <p>ChatGPT may be taking the world by storm – but a new study suggests there is one key area where it is unlikely to be used any time soon. </p> <p>Researchers at Brigham and Women's Hospital – a teaching hospital of Harvard Medical School in Boston, Massachusetts – found that cancer treatment plans generated by OpenAI's revolutionary chatbot were full of errors.</p> <p>According to <a href="https://jamanetwork.com/journals/jamaoncology/fullarticle/2808731?guestAccessKey=669ffd57-d6a1-4f10-afee-e4f81d445b9f&utm_source=For_The_Media&utm_medium=referral&utm_campaign=ftm_links&utm_content=tfl&utm_term=082423" target="_blank" rel="noopener">the study</a>, which was published in the journal JAMA Oncology and initially reported by <a href="https://www.bloomberg.com/news/articles/2023-08-24/chatgpt-fails-at-recommending-cancer-treatment-study-finds?srnd=technology-vp" target="_blank" rel="noopener">Bloomberg</a> – when asked to generate treatment plans for a variety of cancer cases, one-third of the large language model's responses contained incorrect information. </p> <p>The study also noted that the chatbot had a tendency to mix correct and incorrect information together, in a way that made it difficult to identify what was accurate. Out of a total of 104 queries, around 98% of ChatGPT's responses included at least one treatment recommendation that met the National Comprehensive Cancer Network guidelines, the report said.</p> <p>The chatbot "speaks oftentimes in a very sure way that seems to make sense, and the way that it can mix incorrect and correct information is potentially dangerous," study coauthor Danielle Bitterman said, <a href="https://www.bloomberg.com/news/articles/2023-08-24/chatgpt-fails-at-recommending-cancer-treatment-study-finds?srnd=technology-vp" target="_blank" rel="noopener">per Bloomberg</a>. </p> <p>"It's hard even for an expert to identify which is the incorrect recommendation," she added. </p> <p>Insider reached out to Bitterman for further comment but didn't immediately hear back.</p> <p>ChatGPT became an overnight sensation when it launched in November 2022, reaching 100 million active users <a href="https://www.reuters.com/technology/chatgpt-sets-record-fastest-growing-user-base-analyst-note-2023-02-01/" target="_blank" rel="noopener">two months</a> after its debut. The chatbot sparked a <a href="https://markets.businessinsider.com/news/stocks/chatgpt-openai-artificial-intelligence-stocks-nvidia-mobileye-c3ai-ambarella-alteryx-2023-1">rush to invest</a> in AI companies and an intense debate over the long-term impact of artificial intelligence; Goldman Sachs research found it could affect 3<a href="https://www.businessinsider.com/generative-ai-chatpgt-300-million-full-time-jobs-goldman-sachs-2023-3">00 million jobs </a>globally. </p> <p>Despite ChatGPT's success, generative AI models are still <a href="https://www.businessinsider.com/ai-chatbot-chatgpt-google-microsofty-lying-search-belief-2023-2">prone to "hallucinations,</a>" where they confidently present information that is misleading or wildly incorrect. Famously, Google's ChatGPT rival Bard <a href="https://www.telegraph.co.uk/technology/2023/02/08/googles-bard-ai-chatbot-gives-wrong-answer-launch-event/" target="_blank" rel="noopener">wiped $120 billion</a> off the company's stock value when it gave an inaccurate answer to a question about the <a href="https://www.businessinsider.com/google-ad-ai-chatgpt-rival-bard-gives-inaccurate-answer-2023-2">James Webb space telescope</a>.</p> <p>Efforts to integrate AI into healthcare, primarily to streamline administrative tasks, are already underway. Earlier this month, a major study found that using AI to <a href="https://www.theguardian.com/society/2023/aug/02/ai-use-breast-cancer-screening-study-preliminary-results" target="_blank" rel="noopener">screen for breast cancer</a> was safe, and suggested it could almost halve the workload of radiologists. </p> <p>A computer scientist at Harvard recently found that GPT-4, the latest version of the model, could <a href="https://www.insider.com/chatgpt-passes-medical-exam-diagnoses-rare-condition-2023-4">pass the US medical licensing exam</a> with flying colors – and suggested it had better clinical judgment than some doctors.</p> <p>Despite this, accuracy issues with generative models such as ChatGPT mean they are unlikely to be taking over from doctors any time soon.</p> <p>The <a href="https://jamanetwork.com/journals/jamaoncology/fullarticle/2808731?guestAccessKey=669ffd57-d6a1-4f10-afee-e4f81d445b9f&utm_source=For_The_Media&utm_medium=referral&utm_campaign=ftm_links&utm_content=tfl&utm_term=082423" target="_blank" rel="noopener">JAMA study</a> found that 12.5% of ChatGPT's responses were "hallucinated," and that the chatbot was most likely to present incorrect information when asked about localized treatment for advanced diseases or immunotherapy.   </p> <p>OpenAI has acknowledged that <a href="https://openai.com/blog/chatgpt" target="_blank" rel="noopener">ChatGPT can be unreliable</a>. The company's <a href="https://openai.com/policies/usage-policies" target="_blank" rel="noopener">terms of usage</a> warn that their models are not designed to provide medical information, and should not be used to "provide diagnostic or treatment services for serious medical conditions."</p> <p>OpenAI did not immediately respond to Insider's request for comment. </p> <div class="read-original">Read the original article on <a href="https://www.businessinsider.com/chatgpt-generates-error-filled-cancer-treatment-plans-study-2023-8">Business Insider</a></div><!-- /wp:html -->

Artificial intelligence chatbot ChatGPT has been one of the most talked about technologies of the year.

Study finds that ChatGPT provided false information when asked to design cancer treatment plans.
The chatbot mixed correct and incorrect information together, making it harder to decipher. 
Accuracy issues with Generative AI means it’s unlikely to be taking over from doctors any time soon.

ChatGPT may be taking the world by storm – but a new study suggests there is one key area where it is unlikely to be used any time soon. 

Researchers at Brigham and Women’s Hospital – a teaching hospital of Harvard Medical School in Boston, Massachusetts – found that cancer treatment plans generated by OpenAI’s revolutionary chatbot were full of errors.

According to the study, which was published in the journal JAMA Oncology and initially reported by Bloomberg – when asked to generate treatment plans for a variety of cancer cases, one-third of the large language model’s responses contained incorrect information. 

The study also noted that the chatbot had a tendency to mix correct and incorrect information together, in a way that made it difficult to identify what was accurate. Out of a total of 104 queries, around 98% of ChatGPT’s responses included at least one treatment recommendation that met the National Comprehensive Cancer Network guidelines, the report said.

The chatbot “speaks oftentimes in a very sure way that seems to make sense, and the way that it can mix incorrect and correct information is potentially dangerous,” study coauthor Danielle Bitterman said, per Bloomberg

“It’s hard even for an expert to identify which is the incorrect recommendation,” she added. 

Insider reached out to Bitterman for further comment but didn’t immediately hear back.

ChatGPT became an overnight sensation when it launched in November 2022, reaching 100 million active users two months after its debut. The chatbot sparked a rush to invest in AI companies and an intense debate over the long-term impact of artificial intelligence; Goldman Sachs research found it could affect 300 million jobs globally. 

Despite ChatGPT’s success, generative AI models are still prone to “hallucinations,” where they confidently present information that is misleading or wildly incorrect. Famously, Google’s ChatGPT rival Bard wiped $120 billion off the company’s stock value when it gave an inaccurate answer to a question about the James Webb space telescope.

Efforts to integrate AI into healthcare, primarily to streamline administrative tasks, are already underway. Earlier this month, a major study found that using AI to screen for breast cancer was safe, and suggested it could almost halve the workload of radiologists. 

A computer scientist at Harvard recently found that GPT-4, the latest version of the model, could pass the US medical licensing exam with flying colors – and suggested it had better clinical judgment than some doctors.

Despite this, accuracy issues with generative models such as ChatGPT mean they are unlikely to be taking over from doctors any time soon.

The JAMA study found that 12.5% of ChatGPT’s responses were “hallucinated,” and that the chatbot was most likely to present incorrect information when asked about localized treatment for advanced diseases or immunotherapy.   

OpenAI has acknowledged that ChatGPT can be unreliable. The company’s terms of usage warn that their models are not designed to provide medical information, and should not be used to “provide diagnostic or treatment services for serious medical conditions.”

OpenAI did not immediately respond to Insider’s request for comment. 

Read the original article on Business Insider

By