This article has been reviewed according to Science X's editorial process and policies. Editors have highlighted the following attributes while ensuring the content's credibility:

fact-checked

trusted source

proofread

Private and secure generative AI tool supports operations and research in a cancer center

ai
Credit: CC0 Public Domain

Dana-Farber Cancer Institute has implemented an artificial intelligence (AI) application intended for general use in a medical center or hospital. The system, called GPT4DFCI, is permitted for operations, administrative, and research uses but prohibited in direct clinical care. The system is deployed within the Dana-Farber digital premises, so all operations, prompts, and responses occur inside a private network. The application is private, secure, HIPAA-compliant, and auditable.

The application rolled out in phases over the last year to increasingly more users. GPT4DFCI was rolled out with detailed guidance; e.g., users were reminded that they are directly responsible for the completeness, veracity, and fairness of any final work products and must verify GPT-generated content because it might be incomplete, biased, or factually false.

The rollout of this tool and associated policy has been guided by the Dana-Farber Generative AI Governance Committee, which includes broad representation of DFCI constituencies, including research, operations, legal, privacy, , bioethics, compliance, intellectual property, and patients.

Once clinical care use was ruled out, a survey of initial users showed that the most reported primary uses were extracting or searching for information in notes, reports, or other files and answering general knowledge questions. Other reported uses included summarizing documents or and drafting or editing letters, meeting minutes, or presentations. The most common concerns reported were inaccurate or false output and ethics and/or compliance with policies.

There is significant potential for generative AI to aid in , coupled with significant risks of bias, inaccuracy, incompleteness, and misuse. Despite these risks, the Dana-Farber team decided that the broad prohibition of generative AI tools would inhibit learning and innovation, which are central to the Dana-Farber mission.

To manage risk and advance discovery, a broadly representative, multidisciplinary governance body guided the technical, ethical, and policy decisions behind this implementation. The experience and technical material have been shared to inform other health care institutions considering similar efforts.

The findings are published in the journal NEJM AI.

More information: Renato Umeton et al, GPT-4 in a Cancer Center — Institute-Wide Deployment Challenges and Lessons Learned, NEJM AI (2024). DOI: 10.1056/AIcs2300191

Citation: Private and secure generative AI tool supports operations and research in a cancer center (2024, March 29) retrieved 27 April 2024 from https://medicalxpress.com/news/2024-03-private-generative-ai-tool-cancer.html
This document is subject to copyright. Apart from any fair dealing for the purpose of private study or research, no part may be reproduced without the written permission. The content is provided for information purposes only.

Explore further

WHO weighs up AI risks and benefits for health care

12 shares

Feedback to editors