Blog

Continue Reading

  • Journal of Medical Internet Research

    Journal of Medical Internet Research

    The application of large language models (LLMs) within the medical domain is undergoing rapid growth [,]. Key areas of investigation include assisted diagnosis [] and the structured representation of electronic health records []. These models exhibit considerable potential, with preliminary findings from research and practical implementations showing promising results. LLMs through pretraining and fine-tuning on extensive datasets containing medical literature, clinical records, and biomedical knowledge bases, leverage deep learning methodologies to develop rich linguistic representations and demonstrate robust contextual understanding and knowledge integration []. This results in significantly enhanced performance compared with traditional language processing tools across natural language understanding, pattern recognition, and correlation analysis, with notable advantages in processing intricate medical data and facilitating cross-domain knowledge transfer []. While the pursuit of enhanced model performance remains crucial, ensuring robust protection of patient privacy and data security remains a paramount concern and a fundamental requirement for the responsible and sustainable advancement of LLM applications in health care. This presents a complex challenge that necessitates both technical and regulatory solutions [-].

    A substantial portion of current research on LLMs for health care applications uses publicly available resources [], such as the Medical Information Mart for Intensive Care-IV []. However, some studies are conducted using internal patient health information (PHI) repositories, which often contain highly sensitive personally identifiable information (PII) [], including patient names, medical record numbers, age, zip code, admission date, and so on. This practice necessitates robust data governance frameworks to ensure patient privacy and data security. In conventional approaches to developing computer models using patient data, researchers typically processed and trained data within local or strictly regulated environments. This practice inherently reduced the risk of sensitive data compromise during transmission and storage [,]. However, deploying LLMs on the cloud often requires uploading vast amounts of raw medical data directly to remote servers. These servers may be distributed across different regions and are frequently not entirely under the control of health care institutions or researchers [,]. Beyond these conventional risks, LLMs introduce additional privacy concerns [] due to their generative nature; they may inadvertently reproduce sensitive information learned during training, and their large scale and complexity expose them to attacks such as model inversion or prompt injection. Together, these factors make the protection of medical data in LLM deployment especially challenging. Recent studies have highlighted concrete privacy threats associated with LLM use: fine-tuning can significantly increase PII memorization and vulnerability to leakage attacks, especially when tuning specific layers of the model []; novel mitigation strategies like “Whispered Tuning,” which combine PII redaction, differential privacy, and output filtering, can markedly reduce leakage while preserving performance []; and beyond memorization, pretrained LLMs can infer personal attributes (eg, location, income, and gender) from seemingly innocuous text with high accuracy []. Consequently, data transmission and storage processes may encounter heightened risks of security vulnerabilities and unauthorized access. Whether through misuse by internal personnel or external cyberattacks, PHI is at risk of improper use or malicious disclosure [].

    Therefore, reconciling the potential of LLMs to enhance health care quality and efficiency with the imperative to protect patient privacy represents a significant challenge requiring careful consideration. The regulatory oversight of LLMs processing PHI is governed by a complex patchwork of national and international privacy laws, including the General Data Protection Regulation in the European Union and the Health Insurance Portability and Accountability Act (HIPAA) in the United States []. Besides, Transparent Reporting of a Multivariable Prediction Model for Individual Prognosis or Diagnosis (TRIPOD)-LLM [] is an extension of the TRIPOD+AI statement, addressing the unique challenges of LLMs in biomedical and health care applications. There is a growing consensus within the academic community regarding the paramount importance of patient privacy protection in LLM research, with numerous concerns and queries being raised concerning the potential risks to sensitive data [,]. A range of privacy-preserving techniques is being widely considered and adopted within the health care domain to ensure data security and regulatory compliance. These include established methods such as deidentification [,], differential privacy [], federated learning [], and homomorphic encryption [,].

    This leads us to our core research question: What measures are being used to protect patient privacy in the PHI-LLMs in the health care field, and are these measures sufficient? Although there are some systematic reviews or scope reviews of LLM research in health care, no scoping review has been published on this critical issue. The primary objective of this study is to conduct a scoping review of the existing literature on PHI-LLMs in health care, evaluate the adequacy of current approaches, and identify areas in need of improvement to ensure robust patient privacy protection.

    Study Design

    This scoping review was guided by the framework for scoping studies by Arksey and O’Malley []. Besides, the study reporting followed the Preferred Reporting Items for Systematic Reviews and Meta-Analyses extension for Scoping Reviews (PRISMA-ScR; checklist provided in ) []. We focused on the following three research questions: (1) What studies on the development and application of LLMs using PHI currently exist within the health care domain? (2) What patient privacy considerations are addressed in existing PHI-LLMs research, and are these measures sufficient? (3) How can future research on the development and application of LLMs using PHI better protect patient privacy?

    Eligibility Criteria

    The inclusion and exclusion criteria for studies are shown in .

    Textbox 1. Inclusion and exclusion criteria.

    Inclusion criteria

    • Studies were included if they focused on the development and application of large language models within health care using personal health information, encompassing activities such as model construction, fine-tuning, optimization, testing, and performance comparison.
    • Eligible literature comprised original research articles written in English.

    Exclusion criteria

    • Studies used publicly available datasets, under the assumption that such data have been adequately deidentified.
    • Studies that were not reviews, abstracts, incomplete reports, or preprints were excluded from the review due to the lack of rigorous peer review.

    Data Sources, Search Strategy, and Study Selection

    We searched PubMed and Embase for studies published between January 1, 2022, and July 20, 2025. This timeframe was chosen to coincide with the release and rapid adoption of advanced LLMs (eg, GPT-3.5 and ChatGPT) and the subsequent surge of their applications in health care. Earlier studies (published before 2022), which primarily investigated transformer-based or nongenerative models, were excluded as they fell outside the scope of this review. The search strategies were drafted by ZXY and further refined through team discussion. The final search strategy can be found in . The final search results were exported into EndNote X9, and duplicates were removed by a library technician.

    First, the titles and abstracts of identified studies were independently screened by 2 researchers (ZXY and LSY) based on the inclusion and exclusion criteria. Any disagreements between the reviewers were resolved through group discussions involving at least 2 researchers to ensure consensus and maintain the integrity of the selection process. The full-text review was also conducted by 2 researchers (ZXY and LSY) independently, with conflicts resolved through whole-group discussion.

    Data Extraction

    The data extraction form was initially drafted by ZXY based on the study objectives and research questions. Following the draft, the form was refined through group discussions to develop a preliminary extraction template. To ensure consistency in the definitions and extraction criteria, 10 articles were randomly selected for a pilot extraction. Feedback from this pilot phase was used to finalize the extraction form. Subsequently, ZXY and LSY independently extracted data from the included studies. Any conflicts or discrepancies encountered during the extraction process were resolved through comprehensive group discussions involving all researchers to maintain the integrity and consistency of the data extraction.

    The extracted data encompassed three main categories:

    1. General characteristics of included studies: this included the first author’s name, publication year, country, the name and type of the LLMs used or developed, the disease domain as classified by the World Health Organization, and the type of tasks based on the classification outlined in the previous article [].
    2. General characteristics of clinical design: this section captured the sample size, the number of centers, the type of data used, the data collection method, the reporting statement they followed, and whether the study protocol was registered or not. The name of the LLM used or developed, the research objective, and the deployment type.
    3. Patient privacy protection considerations: this part focused on the data source, the purpose of using PHI, whether the study underwent ethical review or approval, the declaration of data availability, whether patient consent was obtained, and the PHI protection techniques used.

    All extracts are based on reports from the included study itself.

    Data Analysis

    Descriptive analyses were performed to summarize the characteristics of the included studies. Categorical variables were summarized as frequencies and percentages. Sample size was further categorized into predefined ranges and presented as frequencies and percentages, while other continuous or count-based variables were directly summarized as reported. All results were presented in tables or figures according to the main domains of study characteristics, data characteristics, and privacy protection measures.

    Patient and Public Involvement

    As this was a scoping review of previously published research, no patients or the public were involved in the design of this study.

    Selection of Sources of Evidence

    After removing duplicates, a total of 6174 citations were identified through searches of electronic databases and references in review articles. Based on their titles and abstracts, 3181 full-text articles were retrieved and assessed for eligibility. Of these, 2993 were excluded for the following reasons: 4 were not written in English, 48 were preprint papers, 126 were unrelated to LLM research, 1387 were reviews, comments, or letters, 38 were protocols, and 1390 did not involve relevant patient data. Following the eligibility assessment, a total of 2717 records were excluded for the following reasons: 647 used only public databases or previously published cases, 439 simulated clinical scenarios or patients rather than using real-world data, 1587 focused on medical knowledge quizzes or examinations, and 44 represented secondary analyses. Ultimately, 464 studies were deemed eligible for this review () []. The specific references included in the analysis are listed in the .

    Figure 1. PRISMA (Preferred Reporting Items for Systematic reviews and Meta-Analyses) study selection diagram.

    General Characteristics of the Included Studies

    To provide an overview of the studies included in this review, we summarize their general characteristics in . This review encompasses studies published between January 1, 2022, and July 20, 2025. Of the 464 studies included in this review, the majority were published in 2025 (n=256, 55.2%), with a substantial proportion appearing in 2024 (n=188, 40.5%). In contrast, only a small number of studies were published in 2023 (19/464, 4.1%) and 2022 (1/464, 0.2%). This temporal distribution underscores the rapid and recent surge of research on LLM applications in health care, with most evidence emerging within the past 2 years. Based on the institutions of the first authors, the largest proportion of studies originated from the United States (153/464, 33.0%), followed by China (94/464, 20.3%) and Germany (37/464, 8.0%). Moderate contributions were observed from Turkey (31/464, 6.7%), Italy (17/464, 3.7%), Israel (13/464, 2.8%), South Korea (12/464, 2.6%), and the United Kingdom (12/464, 2.6%). Studies from other countries collectively accounted for 20.5% (n=95).

    Table 1. General characteristics of included studies.
    Characteristic Values
    Publication year, n (%)
    2025 256 (55.2)
    2024 188 (40.5)
    2023 19 (4.1)
    2022 1 (0.2)
    Country, n (%)
    United States 153 (33.0)
    China 94 (20.3)
    Germany 37 (8.0)
    Turkey 31 (6.7)
    Italy 17 (3.7)
    Israel 13 (2.8)
    South Korea 12 (2.6)
    United Kingdom 12 (2.6)
    Other countriesa 95 (20.5)
    Model type, n (%)
    Existing model 356 (76.7)
    Fine-tuning model + existing model 49 (10.6)
    Fine-tuning model 40 (8.6)
    Self-developed model 17 (3.7)
    Self-developed model + existing model 2 (0.4)
    Location of the large language model, n (%)
    Cloud deployment 72 (15.5)
    Local deployment 70 (15.1)
    External cloud service 69 (14.9)
    Othersb 8 (1.7)
    Cannot judge 57 (12.3)
    Not report 188 (40.5)
    Task type, n (%)
    Making diagnoses 134 (28.9)
    Clinical note-taking 62 (13.4)
    Making treatment recommendations 45 (9.7)
    Generating medical reports 43 (9.3)
    Biomedical data mining 38 (8.2)
    Prognostic predictive model 22 (4.7)
    Communicating with patients 20 (4.3)
    Making diagnoses + Making treatment recommendations 19 (4.1)
    Other tasksc 81 (17.5)
    Sample sized, n (%)
    Patients (n=342, 73.7%)
    <100 125 (36.5)
    100-1000 161 (47.1)
    1000-10,000 35 (10.2)
    ≥10,000 19 (5.6)
    Not report 2 (0.6)
    Token (n=16, 3.4%)
    100-1000 10 (62.5)
    1000-10,000 3 (18.8)
    ≥10,000 2 (12.5)
    Not report 1 (6.3)
    Notes (n=94, 20.3%)
    <100 12 (12.8)
    100-1000 19 (20.2)
    1000-10,000 23 (24.5)
    ≥10,000 38 (40.4)
    Not report 2 (2.1)
    Images (n=12, 2.6%)
    <100 2 (16.7)
    100-1000 7 (58.3)
    1000-10,000 1 (8.3)
    ≥10,000 2 (16.7)
    Number of centerse, n (%)
    1 377 (81.3)
    2 41 (8.8)
    3 16 (3.4)
    4 7 (1.5)
    ≥5 20 (4.3)
    Not report 3 (0.6)
    Summarize the large language models used in the research, n (%)
    ChatGPT 341 (73.5)
    Llama 74 (15.9)
    Mistral 27 (5.8)
    Flan 13 (2.8)
    Claude 19 (4.1)
    Gemini 32 (6.9)
    Gemma 9 (1.9)
    GLM 9 (1.9)
    Deepseek 6 (1.3)
    Qwen 13 (2.8)
    Fine-tuning 49 (10.6)
    Others 73 (15.7)
    Type of data, n (%)
    Text 366 (78.9)
    Text + Image 35 (7.5)
    Text + Audio 4 (0.9)
    Image 56 (12.1)
    Audio 2 (0.4)
    Text + Image + Audio 1 (0.2)
    Data collection method, n (%)
    Retrospective 409 (88.1)
    Prospective 53 (11.4)
    Prospective + retrospective 2 (0.4)
    Follow the statement, n (%)
    Not report 435 (93.8)
    STROBEf 17 (3.7)
    STARDg 3 (0.6)
    TRIPODh 7 (1.5)
    CLAIMi 2 (0.4)
    Registered, n (%)
    No 451 (97.2)
    Yes 13 (2.8)

    aOther countries: Japan (n=11), Australia (n=10), France (n=9), Spain (n=9), Canada (n=6), Switzerland (n=5), India (n=5), Singapore (n=4), Belgium (n=3), Brazil (n=3), Croatia (n=3), the Netherlands (n=3), Pakistan (n=3), Saudi Arabia (n=3), Ireland (n=2), Mexico (n=2), Portugal (n=2), Romania (n=2), Thailand (n=2), Burkina Faso (n=1), Finland (n=1), Iran (n=1), Jordan (n=1), Poland (n=1), Ukraine (n=1), United Arab Emirates (n=1), and Vietnam (n=1).

    bOther location of LLMs: local deployment + cloud deployment (n=5); local deployment + external cloud service (n=2); cloud deployment + external cloud service (n=1).

    cOther tasks: synthesizing data for research (n=15), translation (n=12), triaging patients (n=11), conducting medical research (n=7), making diagnoses + triaging patients + Making treatment recommendations (n=5), making diagnoses + generating medical reports (n=4), generating billing codes (n=3), writing prescriptions (n=3), making diagnoses + triaging patients (n=2), making diagnoses + biomedical data mining (n=2); educating patients (n=2), making treatment recommendations + triaging patients (n=2), communicating with patients + making treatment recommendations (n=2), clinical note-taking + making treatment recommendations (n=2); enhancing medical knowledge (n=1), educating patients + making treatment recommendations + making diagnoses (n=1), communicating with patients + making diagnoses + making treatment recommendations (n=1), triaging patients + prognostic (n=1), generating medical reports + making treatment recommendations (n=1), generating medical reports + prognostic (n=1), clinical note-taking + generating medical reports (n=1), clinical note-taking + prognostic (n=1), and clinical note-taking + translation (n=1).

    dSample size was defined according to the primary data modality: number of patients (clinical studies), number of images (imaging studies), or number of clinical notes/documents (text-based studies), number of tokens (referring to the unit of original studies).

    eReferring to the number of clinical sites contributing patient data.

    fSTROBE: Strengthening the Reporting of Observational Studies in Epidemiology.

    gSTARD: Standards for Reporting of Diagnostic Accuracy.

    hTRIPOD: Transparent Reporting of a Multivariable Prediction Model for Individual Prognosis or Diagnosis.

    iCLAIM: Checklist for Artificial Intelligence in Medical Imaging.

    With respect to model type, 356 (76.7%) studies used existing models, 49 (10.6%) studies combined fine-tuning with existing models, 40 (8.6%) studies applied fine-tuning alone, 17 (3.7%) studies developed their own models, and 2 (0.4%) studies reported using both self-developed and existing models. Regarding the location of LLM deployment, 72 (15.5%) studies reported cloud deployment, 70 (15.1%) studies reported local deployment, and 69 (14.9%) studies reported the use of external cloud services. Eight (1.7%) studies used other deployment approaches, 57 (12.3%) studies could not be judged from the report, and 188 (40.5%) studies did not provide deployment information.

    A total of 134 (28.9%) studies focused on making diagnoses, 62 (13.4%) studies focused on clinical note-taking, 45 (9.7%) studies focused on making treatment recommendations, 43 (9.3%) studies focused on generating medical reports, and 38 (8.2%) studies focused on biomedical data mining. Prognostic predictive model tasks were examined in 22 (4.7%) studies, and communication with patients in 20 (4.3%) studies. In addition, 19 (4.1%) studies combined diagnostic and treatment recommendation tasks, while other tasks were reported in 81 (17.5%) studies.

    Regarding sample size, the distribution was varied: 125 (36.5%) studies enrolled fewer than 100 patients, 161 (47.1%) studies included 100-1000 patients, 35 (10.2%) studies included 1000-10,000 patients, 19 (5.6%) studies included more than 10,000 patients, and 2 (0.6%) studies did not report sample size. For token-based datasets (n=16), 10 (62.5%) studies used 100-1000 tokens, 3 (18.8%) studies used 1000-10,000, 2 (12.5%) studies used more than 10,000, and 1 (6.3%) study did not report. Regarding note-based datasets (n=94), 12 (12.8%) studies analyzed fewer than 100 notes, 19 (20.2%) studies used 100-1000, 23 (24.5%) studies used 1000-10,000, 38 (40.4%) studies used more than 10,000, and 2 (2.1%) studies did not report. For image-based datasets (n=12), 2 (16.7%) studies included fewer than 100 images, 7 (58.3%) studies included 100-1000, 1 (8.3%) studies included 1000-10,000, and 2 (16.7%) studies included more than 10,000.

    Regarding the number of centers, 377 (81.3%) studies were conducted in a single center, 41 (8.8%) studies were conducted in 2 centers, 16 (3.4%) studies were conducted in 3 centers, and 7 (1.5%) studies were conducted in 4 centers. Twenty (4.3%) studies involved 5 or more centers, while 3 (0.6%) studies did not report this information. As for LLMs used in the research, ChatGPT was the most frequently used model, accounting for 341 (73.5%) studies. This was followed by 74 (15.9%) studies used Llama, 32 (6.9%) studies used Gemini, 27 (5.8%) studies used Mistral, 19 (4.1%) studies used Claude, and 13 (2.8%) studies used Qwen. Additionally, 49 (10.6%) studies applied fine-tuning techniques, while 73 (15.7%) studies reported using other models not specifically listed. Regarding data type, 366 (78.9%) studies used text; 56 (12.1%) studies used image data; 35 (7.5%) studies combined text and image; 4 (0.9%) studies combined text and audio; 2 (0.4%) studies used audio only; and 1 (0.2%) study combined text, image, and audio. With respect to data collection method, 409 studies (88.1%) were retrospective, 53 (11.4%) studies were prospective, and 2 (0.4%) studies combined both prospective and retrospective approaches.

    With respect to reporting standards, 435 (93.8%) studies did not specify adherence to any guideline/statement, while 17 (3.7%) studies followed Strengthening the Reporting of Observational Studies in Epidemiology (STROBE) statement, 3 (0.6%) studies followed the Standards for Reporting of Diagnostic Accuracy (STARD) statement, 7 (1.5%) studies followed TRIPOD statement, and 2 (0.4%) studies followed Checklist for Artificial Intelligence in Medical Imaging. Regarding study registration, 451 (97.2%) studies were not registered, and only 13 (2.8%) studies reported registration.

    The Sankey diagram () shows that the most frequent disease-task pairs included tumors with making diagnoses (n=35), tumors with making treatment recommendations (n=23), and tumors with clinical note-taking (n=16). In addition, studies categorized as “not special” also frequently addressed making diagnoses (n=24) and clinical note-taking (n=15). Other notable pairs included musculoskeletal disorders with making diagnoses (n=17), neurological disorders with making diagnoses (n=13), and circulatory diseases with making diagnoses (n=11). Tasks such as generating medical reports and biomedical data mining were also commonly associated with tumors and “not special” categories. The summary table could be found in .

    Figure 2. Sankey diagram of disease categories and task types.

    Characteristics of Privacy Protection

    outlines the characteristics of privacy protection measures implemented in the included studies. Regarding ethical oversight, 419 (90.3%) studies reported approval from an ethics committee, whereas 45 (9.7%) studies did not. With respect to patient consent, 224 (48.3%) studies reported a waiver of informed consent, 92 (19.8%) studies indicated that informed consent had been obtained, and 148 (31.9%) studies did not report consent information. For data availability, 203 (43.8%) studies did not provide a statement, 160 (34.5%) studies declared that data were available from the corresponding author upon reasonable request, 66 (14.2%) studies stated that data were not available, and 35 (7.5%) studies made data publicly accessible.

    Table 2. Characteristics of privacy protection. Our categorization of privacy protection methods is based on terminology as reported by the original studies. However, the definitions of “de-identification” and “anonymization” vary across contexts; thus, the risk implications should be interpreted with caution.
    Characteristics Values, n (%)
    Ethical review
    Yes 419 (90.3)
    No 45 (9.7)
    Patient consent
    Waiver of informed consent 224 (48.3)
    Not report 148 (31.9)
    Informed consent has been obtained 92 (19.8)
    Data availability declaration
    Not report 203 (43.8)
    Corresponding author on reasonable request 160 (34.5)
    Not open 66 (14.2)
    Public 35 (7.5)
    Privacy protection technology
    Not report 178 (38.4)
    Deidentification 158 (34.1)
    Cannot judge from report 116 (73.4)
    Based on manual 17 (10.8)
    Based on rule matching 13 (8.2)
    Othersa 12 (7.6)
    Anonymization 91 (19.6)
    Deidentification+Anonymization 23 (5.0)
    Othersb 14 (3.0)
    Is there a statement to remove any personally identifiable information?
    No 363 (78.2)
    Yes 101 (21.8)
    Were direct identifiers or indirect identifiers removed?
    Direct identifiers 166 (35.8)
    Indirect identifiers 9 (1.9)
    Cannot judge 107 (23.1)
    Not report 182 (39.2)
    Whether the degree of deidentification is assessed?
    No 458 (98.7)
    Yes 6 (1.3)
    Reidentification protection technology used?
    No 455 (98.1)
    Yes 9 (1.9)
    Declaration of compliance with safety standards
    Health Insurance Portability and Accountability Act 44 (9.5)
    General Data Protection Regulation 6 (1.3)
    Both 2 (0.4)
    Not report 412 (88.8)

    aBased on rule matching + machine learning + deep learning (n=3), based on LLMs (n=2), based on rule matching + manual (n=2), based on rule matching + machine learning (n=1), based on synthetic data (n=1), based on postprocessing (n=1), based on machine learning (n=1), and based on deep learning+ postprocessing (n=1).

    bData hosting (n=5), anonymization + data hosting (n=3), federated learning (n=1), anonymization + data hosting + homomorphic encryption (n=1), anonymization + homomorphic encryption (n=1), deidentification + data hosting (n=1), data augmentation (likely referred to synthetic data generation; n=1), and homomorphic encryption (n=1).

    When examining privacy protection technologies, 178 (38.4%) studies did not report the methods used, while 158 (34.1%) studies applied deidentification, 91 (19.6%) studies used anonymization, 23 (5.0%) studies reported combining both, and 14 (3.0%) studies used other technologies. Among those providing more detail about deidentification, 17 (10.8%) studies used manual methods, 13 (8.2%) studies applied rule-based matching, 12 (7.6%) studies reported other approaches, and 116 (73.4%) studies cannot be judged from the reports.

    Concerning statements on the removal of personally identifiable information, 363 (78.2%) studies did not provide such a statement, while 101 (21.8%) studies explicitly reported it. About the type of identifiers removed, 166 (35.8%) studies specified the removal of direct identifiers, 9 (1.9%) studies reported the removal of indirect identifiers, 107 (23.1%) studies could not be judged from the report, and 182 (39.2%) studies did not provide this information. Regarding assessment of the degree of deidentification, 458 (98.7%) studies did not report such an assessment, while 6 (1.3%) studies did. Concerning reidentification protection technologies, 455 (98.1%) studies did not use them, and 9 (1.9%) studies reported their use. With respect to compliance with safety standards, 44 (9.5%) studies declared adherence to HIPAA, 6 (1.3%) studies to the General Data Protection Regulation, and 2 (0.4%) studies to both, whereas 412 (88.8%) studies did not provide such information.

    shows the characteristics of privacy protection technology for different data types. Among text-based studies (n=366), 131 (35.8%) studies applied deidentification, 74 (20.2%) studies used anonymization, 15 (4.1%) studies combined both, 4 (1.1%) studies reported data hosting, 2 (0.5%) studies reported anonymization with data hosting, 6 (1.6%) studies used other methods, and 134 (36.6%) studies did not report. For image-based studies (n=56), 12 (21.4%) studies reported deidentification, 10 (17.9%) studies reported anonymization, 5 (8.9%) studies reported both, 1 (1.8%) study reported anonymization with data hosting, 1 (1.8%) study reported data hosting, and 27 (48.2%) studies did not report. For audio-based studies (n=2), one (50.0%) study reported anonymization and one (50.0%) study did not. For combined text and image studies (n=35), 14 (40.0%) studies used deidentification, 6 (17.1%) studies used anonymization, 3 (8.6%) studies used both, and 12 (34.3%) studies did not report. For text and audio studies (n=4), one (25.0%) study reported deidentification, and 3 (75.0%) studies did not. The single study using text, image, and audio data did not report its privacy protection method.

    Table 3. Characteristics of privacy protection technology for different data types.
    Type of data and privacy protection technologya Values, n (%)
    Text (n=366)
    Deidentification 131 (35.8)
    Anonymization 74 (20.2)
    Deidentification + anonymization 15 (4.1)
    Data hosting 4 (1.1)
    Anonymization + data hosting 2 (0.5)
    Othersb 6 (1.6)
    Not reported 134 (36.6)
    Images (n=56)
    Deidentification 12 (21.4)
    Anonymization 10 (17.9)
    Deidentification + anonymization 5 (8.9)
    Anonymization + data hosting 1 (1.8)
    Data hosting 1 (1.8)
    Not reported 27 (48.2)
    Audio (n=2)
    Anonymization 1 (50.0)
    Not reported 1 (50.0)
    Text+Images (n=35)
    Deidentification 14 (40.0)
    Anonymization 6 (17.1)
    Deidentification + anonymization 3 (8.6)
    Not reported 12 (34.3)
    Text+Audio (n=4)
    Deidentification 1 (25.0)
    Not reported 3 (75.0)
    Text + Images + Audio (n=1)
    Not reported 1 (100.0)

    aClassify according to the conditions reported in the original articles of the included studies.

    bAnonymization + data hosting + homomorphic encryption (n=1), anonymization + homomorphic encryption (n=1), federated learning (n=1), data augmentation (n=1), deidentification + data hosting (n=1), and homomorphic encryption (n=1).

    Main Findings

    In this scoping review, we identified 464 studies published between 2022 and 2025 that focus on the development and application of LLMs in health care using PHI. Strikingly, 256 (55.2%) of these studies were published in 2025 alone, compared with 188 (40.5%) studies in 2024, 19 (4.1%) studies in 2023, and only one (0.2%) study in 2022. This sharp increase highlights the extremely rapid pace of research in this emerging field and reflects the growing recognition of both the opportunities and challenges associated with LLM deployment in health care. These studies encompass a variety of countries, disease domains, and task types. Overall, the ethical reviews of these studies have been largely satisfactory. The vast majority of studies have reported on the approval of the ethics committee, ensuring that their procedures meet the relevant ethical standards. Nevertheless, there remains a shortfall in reporting on informed consent in certain prospective studies. It is concerning that a small number of LLM research projects using imaging data or retrospective data from electronic medical records fail to adequately report ethical review processes and the consideration of patient-informed consent, including whether consent was obtained or formally waived. Even when research involves only patient imaging data or retrospective data, it must still undergo rigorous ethical review [,]. Strict adherence to ethical review processes is essential to ensure the fairness and scientific integrity of medical research and to safeguard patient rights.

    According to our findings, more than half of LLMs use cloud deployments, and the generative nature of LLMs can accidentally expose private data learned during training [], including PII []. And the vast parameter counts and extensive training corpora of LLMs enlarge their memorization footprint, leakage vectors, and prompt-manipulation surface, thereby exposing critical vulnerabilities that render these models prime targets for prompt-injection and data-poisoning attacks. Therefore, it is necessary to enhance the privacy protection of PHI. Cloud deployment offers the advantages of cost-effectiveness and scalability, as users are not required to invest in expensive hardware and can dynamically scale computing resources as needed. Additionally, cloud services provide established tools and global accessibility, facilitating rapid iteration and collaboration among distributed teams. However, cloud deployments also face significant data privacy risks, as data must be uploaded to third-party servers, which can lead to risks such as data breaches, unauthorized access, compliance issues, service disruptions, and model exploitation. These risks can be effectively mitigated through the integration of technical countermeasures like zero-trust architecture [], edge computing [], data encryption [], and access control.

    Given the inherently high sensitivity of medical data and the significant risk of irreversible harm upon disclosure, the development of privacy-preserving safeguards for LLMs has become imperative. This requirement is not merely a technical prerequisite but also reflects an uncompromising mandate at the ethical and regulatory levels. In current research on the application of LLMs to health care, approximately one-third of the studies fail to mention any techniques for effectively protecting PHI. PHI includes PII such as patient names, addresses, Social Security numbers, and medical record numbers. Once compromised, such information can lead to serious privacy breaches and security risks. Therefore, removing PII is the first crucial step in safeguarding patient privacy [,]. By effectively eliminating or obfuscating these direct identifiers, it becomes possible to mitigate the risk of unauthorized access to PHI and thereby reduce the negative consequences of potential data breaches. The failure to prioritize PII protection in the deployment of LLMs within health care poses significant risks that extend beyond immediate privacy concerns. It threatens the integrity of patient-provider relationships, exposes individuals to financial and identity-related crimes, stifles technological and scientific progress, and raises critical ethical issues. Among the included studies, only one mentioned the use of federated learning techniques, and none used homomorphic encryption. Despite this, federated learning and homomorphic encryption are emerging as pivotal techniques for privacy-preserving in LLMs. Federated learning, due to its distributed training and decentralized model architecture, has gained significant traction in health care []. Future research should prioritize the development of comprehensive data privacy protection to facilitate the broader adoption of federated learning and homomorphic encryption in health care. Synthetic data generation has emerged as a promising solution to address privacy concerns in health care research and LLMs []. This approach uses AI models to create realistic, anonymized patient data that preserves privacy while enabling data access for secondary purposes []. While synthetic data offers benefits in promoting privacy, equity, safety, and continual learning, challenges remain, including the potential introduction of flaws and biases. Further research is needed to develop unified quality assessment metrics and address the current deficiency in longitudinal synthetic data generation [].

    In studies involving PHI deidentification techniques or the use of anonymized data, the related descriptions are often very vague. Although these studies mention the removal of PII or the use of anonymized data, they typically do not specify which PII elements were removed or merely use ambiguous phrases such as “removing PII” or “using anonymized data,” lacking detailed technical explanations and transparency. First, the absence of clear technical descriptions undermines the credibility and reproducibility of the research, making it difficult for other researchers to replicate the results under the same conditions and thereby affecting the scientific validity and the effectiveness of subsequent applications. Second, privacy protection may be at risk because the lack of transparency in the deidentification process can result in PII not being fully removed, increasing the risk of data breaches. Finally, ethical issues arise as well. If the data processing for deidentification and anonymization is not sufficiently transparent, it can lead to ethical disputes regarding privacy protection and informed consent, especially when assessing whether there remains a risk of reidentifying individuals from the data. This lack of transparency may result in failing to meet the requirements of ethical reviews. According to the Food and Drug Administration, even deidentified data must retain traceability to meet regulatory requirements. Traceability after deidentification is not only a critical component of privacy protection but also essential for ensuring data availability, compliance, and credibility. It plays an indispensable role in data sharing, research transparency, emergency response, and other areas, providing a robust foundation for data-driven decision-making and innovation. When designing and implementing deidentification schemes, the need for traceability must be carefully considered to achieve a balance between privacy protection and data utility. In addition, the descriptions of “de-identification assessment” provided in current studies often lack transparency, leaving their scope and rigor unclear. For instance, it remains uncertain whether such assessments involved quantitative estimation of reidentification risk, reference to regulatory standards such as HIPAA Safe Harbor, or validation by external independent experts. The absence or ambiguity of these elements makes it difficult to determine the effectiveness and compliance of deidentification practices. Future studies should explicitly report these aspects in both methodology and results to ensure greater rigor and credibility in protecting patient privacy.

    The HIPAA establishes standards and practices for deidentifying PHI. According to this rule, there are 2 methods for deidentifying PHI: expert determination, which requires a formal assessment by a qualified expert; and Safe Harbor, which involves the removal of specified identifiers so that covered entities and business associates cannot identify individuals from the remaining information. Although HIPAA does not explicitly use the term “anonymization,” anonymization is often considered an irreversible process that ensures data can no longer identify individuals. Anonymization requirements are more stringent than deidentification, as they guarantee that the data cannot be reidentified under any circumstances. While this study retains the original terminology used in the articles reviewed, most studies are vague in their descriptions of deidentification and anonymization, making it difficult to determine which specific methods were used. Researchers should clearly specify the approaches used to protect data privacy to ensure transparency and accuracy.

    Medical big data exhibits unique multimodal characteristics. The term “multimodal” refers to the diverse sources and forms of medical data, which include laboratory data (eg, laboratory results), imaging data (eg, computed tomographic scans, x-rays, ultrasounds, and electrocardiograms), and video data containing audio (eg, fetal ultrasounds). Depending on the specific data type, its confidentiality, integrity, and availability are ensured through various methods. In our research, we found that current medical data privacy protection primarily relies on deidentification and anonymization techniques. However, in the context of multimodal medical data, a single privacy protection method is often insufficient to effectively prevent data leakage, tampering, and misuse. Therefore, designing multimodal data privacy protection technologies represents a critical direction for future research.

    Strengths and Limitations

    While existing literature reviews predominantly focus on the applications of LLMs in health care, there remains a notable gap in comprehensive scoping reviews that specifically evaluate privacy protection measures for PHI within LLM implementations (PHI-LLMs). Previous analyses addressing privacy concerns have primarily examined broader contexts rather than focusing specifically on patient information protection in language model applications, resulting in insufficient coverage of both technical safeguards and systemic compliance aspects within health care ecosystems.

    A limitation of this study is that the evaluation of privacy protection measures relies solely on the information reported in published papers. Therefore, if certain studies have implemented privacy protection methods but did not disclose them in detail within their articles, we are unable to identify them. This situation may affect the comprehensiveness and accuracy of our evaluation. While we catalogued the adoption of different privacy protection methods, our review did not evaluate their security levels, implementation quality, or practical trade-offs. Future research should systematically assess the effectiveness and applicability of these techniques in health care-specific settings.

    Implications for Research and Practice

    Based on the 3 key findings outlined above, we offer the following additional recommendations for protecting patient privacy in health care–related LLM research, structured around 3 phases: study design, implementation, and reporting. When conducting research reports, the key terms can be referred to for a standardized design. The glossary of key terms () can be found in the .

    Textbox 2. Glossary of key terms.
    • Protected health information: Individually identifiable health information relating to health status, care, or payment that is protected under the Health Insurance Portability and Accountability Act.
    • Personally identifiable information: It refers to any information that can directly or indirectly identify a specific individual, including name, ID number, address, contact information, and data that can identify identity when combined with other information. Personally identifiable information is a broad concept that encompasses all data that can identify an individual.
    • Deidentification (Safe Harbor, Health Insurance Portability and Accountability Act): Removal of 18 specified identifiers (eg, name, address, phone number, and dates) such that the data is no longer considered protected health information.
    • Deidentification (Ontario Guidance): A risk-based statistical approach that quantifies reidentification risk and determines whether it is acceptably small.
    • Anonymization (General Data Protection Regulation): Data is processed in such a way that reidentification is no longer possible by any means “reasonably likely to be used.”
    • Rule-based matching: Algorithmic detection and removal of direct identifiers (eg, names, addresses) using predefined rules or dictionaries.
    • Federated learning: A decentralized machine learning paradigm where models are trained collaboratively without exchanging raw data.
    • Reidentification protection technology: Technical and organizational measures (eg, k-anonymity and differential privacy) that reduce the likelihood of reidentifying individuals, acknowledging that zero risk is unattainable but very low residual risk is acceptable under most regulations.
    • Direct identifiers: Variables that can uniquely and directly identify an individual, such as name, social security number, phone number, full address, and medical record number.
    • Indirect identifiers (quasi-identifiers): Variables that cannot identify an individual alone but may enable reidentification when combined with other data, such as age, gender, zip code, and admission date.

    Considerations in Research Design

    In the design phase of LLM research, patient privacy protection must be prioritized. First, the data minimization principle should be strictly adhered to, meaning only the minimum necessary PHI required to achieve the research objectives should be collected and used []. Second, a clear definition of research purpose and usage scope is essential, ensuring that all PII used has a well-defined purpose and is not repurposed for unauthorized studies or commercial applications. Additionally, ethical approval and informed consent are critical components of the design phase. Researchers must submit detailed research plans to an Institutional Review Board or Ethics Committee, outlining how PII will be obtained, used, and protected. Where applicable, obtaining informed consent from patients is necessary to ensure they are aware of how their data will be used and safeguarded.

    Considerations in Research Implementation

    In the implementation phase, priority should be given to deploying the LLM locally. During model training, multiple patient privacy protection strategies such as deidentification, anonymization, federated learning, synthetic data, and differential privacy should be used. Due to the potential risk of reidentification, continuous security monitoring and auditing are indispensable. The research team should conduct regular security assessments and vulnerability scans to promptly identify and address potential security vulnerabilities. Postbreach responses also constitute an indispensable part of a comprehensive privacy protection framework. Effective incident response should include rapid detection, containment, patient notification, and remediation strategies, which are increasingly emphasized in health care data governance guidelines.

    Considerations in Research Reporting

    Research reports should fully embody the principles of transparency and reproducibility. Researchers should disclose in detail the data sources, ethical approval processes, informed consent procedures, and privacy protection techniques used. Selecting appropriate reporting guidelines (such as STROBE [], STARD [], CONSORT-AI [Consolidated Standards of Reporting Trials–Artificial Intelligence] [], and TRIPOD-LLM []) can improve report quality and provide a reference for other researchers.

    Conclusions

    Our scoping review sounds an alarm on the inadequately addressed imperative of patient privacy protection in medical research using LLMs. In response, we formulate comprehensive recommendations for the study design, implementation, and reporting phases to fortify PHI protection and foster transparency in PHI-LLM research. Our findings compellingly argue for the urgent development of stricter regulatory frameworks and the integration of advanced privacy-preserving technologies to safeguard PHI. It is anticipated that such measures will enable future health care applications of LLMs to achieve a balance between innovation and rigorous patient privacy protection, thus elevating ethical standards and scientific credibility.

    The data generated during the scoping review are available from the corresponding author on reasonable request. However, the majority of such data have been presented on paper in tables, figures, and text.

    This study is supported by the National Key Research and Development Program of China (numbers 2022YFC3502300 and 2022YFC3502305). The funders had no role in considering the study design or in the collection, analysis, interpretation of data, writing of the report, or decision to submit the article for publication.

    ZXY and LSY conceptualized the study, defined the methodology, performed the database searches, and managed the screening process. ZXY and LSY also performed data extraction and authored the original draft. CZ, YDD, and WSJ reviewed the literature, charted the data, and conducted the analyses. GL, YLZ, and SHC provided supervision and contributed to the writing review. All authors contributed to the discussion of literature screening, data extraction, and writing of the paper. All authors reviewed and approved the final version of the paper. The corresponding author attests that all listed authors meet authorship criteria and that no others meeting the criteria have been omitted. All authors affirm that the manuscript is an honest, accurate, and transparent account of the study being reported; that no important aspects of the study have been omitted; and that any discrepancies from the study as planned (and, if relevant, registered) have been explained.

    None declared.

    Edited by A Coristine; submitted 26.Apr.2025; peer-reviewed by P Han, Z Hou, L Pilgram; comments to author 16.Jul.2025; revised version received 02.Nov.2025; accepted 03.Nov.2025; published 21.Nov.2025.

    ©Xiaoying Zhong, Siyi Li, Zhao Chen, Long Ge, Dongdong Yu, Shijia Wang, Liangzhen You, Hongcai Shang. Originally published in the Journal of Medical Internet Research (https://www.jmir.org), 21.Nov.2025.

    This is an open-access article distributed under the terms of the Creative Commons Attribution License (https://creativecommons.org/licenses/by/4.0/), which permits unrestricted use, distribution, and reproduction in any medium, provided the original work, first published in the Journal of Medical Internet Research (ISSN 1438-8871), is properly cited. The complete bibliographic information, a link to the original publication on https://www.jmir.org/, as well as this copyright and license information must be included.

    Continue Reading

  • Bitcoin and other crypto assets sink in flight from risk

    Bitcoin and other crypto assets sink in flight from risk

    Bitcoin and ether slumped to multi-month lows on Friday, with cryptocurrencies swept up in a broader flight from riskier assets as investors worried about lofty tech valuations and bets on near-term U.S. interest rate cuts faded.

    Bitcoin, the world’s largest cryptocurrency, fell 5.5% to a seven-month low of $81,668. Ether slid more than 6% to $2,661.37, its lowest in four months.

    Both tokens are down roughly 12% so far this week.

    Cryptocurrencies are often viewed as a barometer of risk appetite and their slide highlights how fragile the mood in markets has turned in recent days, with high-flying artificial intelligence stocks tumbling and volatility spiking VIX.

    “If it’s telling a story about risk sentiment as a whole, then things could start to get really, really ugly, and that’s the concern now,” Tony Sycamore, a market analyst at IG, said of the fall in bitcoin.

    About $1.2 trillion has been wiped off the market value of all cryptocurrencies in the past six weeks, according to market tracker CoinGecko.

    Bitcoin’s slide follows a stellar run this year that propelled it to a record high above $120,000 in October, buoyed by favourable regulatory changes towards crypto assets globally.

    But analysts say the market remains scarred by a record single-day slump last month that saw more than $19 billion of positions liquidated.

    “The market feels a little bit dislocated, a bit fractured, a bit broken, really, since we had that selloff,” said Sycamore.

    Bitcoin has since erased all its year-to-date gains and is now down 12% for the year, while ether has lost close to 19%.

    Citi analyst Alex Saunders said $80,000 would be an important level as it is around the average level of bitcoin holdings in ETFs.

    The selloff has also hurt share prices of crypto stockpilers, following a boom in public digital asset treasury companies this year as corporates took advantage of rising prices to buy and hold cryptocurrencies on their balance sheets.

    Shares of Strategy, once the poster child for corporate bitcoin accumulation, have fallen 11% this week and were down nearly 4% in premarket trade, languishing at one-year lows.

    JP Morgan said in a note this week that the company could be excluded from some MSCI equity indexes, which could spark forced selling by funds that track them.

    Its Japanese peer Metaplanet has tumbled about 80% from a June peak.

    Crypto exchange Coinbase was down 1.9% in premarket trade and is on course for its longest losing streak in more than a month.

    Crypto miners MARA Holdings and CleanSpark were down 2.4% and 3.6%, respectively, while the Winklevoss twins’ newly-listed Gemini has plunged 62% from its listing price.

    “Bitcoin market conditions are the most bearish they have been since the current bull cycle started in January 2023,” said digital asset research firm CryptoQuant in its weekly crypto report on Wednesday.

    “We are highly likely to have seen most of this cycle’s demand wave pass.”

    Continue Reading

  • Why some volcanoes don’t explode

    Why some volcanoes don’t explode

    Poiseuille number in volcanic systems. Credit: Science (2025). DOI: 10.1126/science.adw8543

    The explosiveness of a volcanic eruption depends on how…

    Continue Reading

  • Earth just got hit by a stealth solar storm no one saw coming

    Earth just got hit by a stealth solar storm no one saw coming

    A stealth solar storm hit Earth on Nov. 20. It arrived without warning and may have contributed to auroras visible at mid latitudes.

    This event didn’t trigger a geomagnetic storm; however, such solar storms, or coronal mass ejections (CMEs), are…

    Continue Reading

  • Kia MVP Ladder: Cade Cunningham’s early-season surge nets Top 5 spot

    Kia MVP Ladder: Cade Cunningham’s early-season surge nets Top 5 spot

    Cade Cunningham is looking to do what no Detroit player has ever done: win a regular-season MVP.

    In the rich history of the Detroit Pistons, which includes a trio of NBA championships and many Hall of Famers who have played for the franchise,…

    Continue Reading

  • Europe’s economy is geared towards a disappearing world, says ECB’s Lagarde | European Central Bank

    Europe’s economy is geared towards a disappearing world, says ECB’s Lagarde | European Central Bank

    Europe’s economy is “geared towards a world that is gradually disappearing”, according to a warning from Christine Lagarde that the EU needs reforms to spur growth.

    The president of the European Central Bank (ECB) said the EU’s dependence on international trade had left it vulnerable, as major partners had turned away from the trade that made the bloc’s exporters wealthy.

    Donald Trump has led a global turn towards protectionism and against globalisation, with steep tariffs imposed on almost every trading partner. At the same time, China has used its dominance of production of certain critical materials and products to exert pressure.

    Lagarde argued that Europe was vulnerable because of a “dependency on third countries for our security and the supply of critical raw materials”. She cited China’s control of the supply of rare earth metals that are crucial in electric motors and wind turbines, as well as the “choke point” of power chips made by Nexperia in China that threatened to shut down production across the global car industry.

    Speaking at the European Banking Congress in Frankfurt, Germany, Lagarde said Europe had failed to address its own problems. Policymakers had instead allowed its weaknesses to “erode growth quietly, as each new shock nudges us onto a slightly lower trajectory”.

    “Our internal market has stood still, especially in the areas that will shape future growth, like digital technology and artificial intelligence, as well as the areas that will finance it, such as capital markets,” she said.

    Europe also faced a “vicious circle” of its own savers allocating money to US stocks, helping the American economy to advance faster than the EU and leaving “stagnating productivity at home and growing dependence on others”, she said.

    Lagarde did highlight some European strengths as well, including a resilient labour market, increasing digital investment, and government spending, particularly on defence in response to Russia’s invasion of Ukraine, that has counteracted economic slowdown.

    Part of Lagarde’s prescription for recovery was lowering barriers to services and goods trade between EU countries. Those barriers are equivalent to a 100% tariff on services and 65% on goods, according to ECB analysis. Lowering those barriers to the same level as the Netherlands – a relatively open economy – would fully make up for the hit from US tariffs, she said.

    skip past newsletter promotion

    She called for mutual recognition of regulated companies, allowing them to sell across Europe when authorised by any one country. She also said the EU should adopt qualified majority voting on tax, preventing any single member state from vetoing changes.

    She argued that benefits would include allowing the harmonisation of VAT, making it easier for smaller European companies to gain access to the whole EU market without having to comply with 27 different tax regimes.

    Continue Reading

  • Citius Oncology to Advance Commercial Launch of LYMPHIR™ with Verix AI Integration

    Leading edge artificial intelligence and machine learning platform supports enhanced salesforce targeting and engagement for cutaneous T-cell lymphoma immunotherapy launch

    CRANFORD, N.J., Nov. 21, 2025 /PRNewswire/ — Citius Oncology, Inc. (“Citius Oncology”) (Nasdaq: CTOR), the oncology-focused subsidiary of Citius Pharmaceuticals, Inc. (“Citius Pharma”) (Nasdaq: CTXR), a late-stage biopharmaceutical company developing and commercializing first-in-class critical care products, today announced a deeper collaboration with Verix, a leader in AI-powered commercial optimization technology for the life sciences sector. Citius Oncology’s commercial team intends to further leverage Verix’s innovative Tovana platform to support the anticipated fourth quarter 2025 U.S. commercialization of LYMPHIR™ (denileukin diftitox-cxdl), a novel immunotherapy approved by the U.S. Food and Drug Administration (FDA) for the treatment of adult patients with relapsed or refractory Stage I-III cutaneous T-cell lymphoma (CTCL) after at least one prior systemic therapy.

    Verix’s Tovana platform integrates advanced analytics, real-world claims data, and machine learning to help inform Citius Oncology’s commercial strategy and enable real-time field execution. The platform refines targeting over time and supports data-driven decisions, allowing the Citius Oncology commercial team to embed predictive intelligence into sales and marketing plans, and to prioritize engagement with high-impact healthcare providers (HCPs). Using company-defined criteria, the system identifies patterns in treatment and diagnosis, enabling Citius Oncology’s commercial team to promptly engage prescribers whose patients may benefit from LYMPHIR.

    “We are committed to leveraging leading-edge technologies to maximize the commercial impact of LYMPHIR and look forward to using this innovative AI platform to amplify the precision and impact of our experienced commercial and marketing teams,” stated Leonard Mazur, Chairman and CEO of Citius Oncology and Citius Pharma. “Through our collaboration with Verix, we are able to enhance our salesforce’s experience with the ability to identify key treatment patterns, personalize provider engagement, and allocate commercial resources efficiently. With LYMPHIR poised to become a meaningful new option for patients with relapsed or refractory CTCL, it’s critical that our commercial organization has precision tools that accelerate physician engagement and product uptake. This data-driven approach positions us to execute a focused and impactful launch while maintaining a lean infrastructure,” added Mazur.

    “We are proud to partner with Citius Oncology at this critical stage in their commercialization journey,” said Doron Aspitz, CEO of Verix. “The Tovana platform is designed to empower life sciences commercial organizations to move from retrospective analysis to real-time action. With Citius Oncology, we see a strong alignment in using AI to bridge data with execution, and we are excited to help deliver meaningful insights that can accelerate the reach of an innovative therapy like LYMPHIR.”

    This initiative builds upon foundational launch activities including the establishment of distribution partnerships, permanent reimbursement codes (J-code: J9161), and inclusion in the National Comprehensive Cancer Network (NCCN) Guidelines. Commercial availability of LYMPHIR in the U.S. is expected in the fourth quarter of 2025.

    Citius Oncology believes that integrating Verix’s AI capabilities will not only enable faster market penetration but also deliver durable advantages as the company builds its oncology franchise. This initiative reflects Citius Oncology’s disciplined focus on capital efficiency, executional excellence, and sustainable value creation for shareholders. Most importantly, it reflects the company’s commitment to facilitating access to care for patients living with cutaneous T-cell lymphoma.

    About LYMPHIR™ (denileukin diftitox-cxdl)

    LYMPHIR is a targeted immune therapy for relapsed or refractory cutaneous T-cell lymphoma (CTCL) indicated for use in Stage I-III disease after at least one prior systemic therapy. It is a recombinant fusion protein that combines the IL-2 receptor binding domain with diphtheria toxin (DT) fragments. The agent specifically binds to IL-2 receptors on the cell surface, causing diphtheria toxin fragments that have entered cells to inhibit protein synthesis. After uptake into the cell, the DT fragment is cleaved and the free DT fragments inhibit protein synthesis, resulting in cell death. Denileukin diftitox-cxdl demonstrated the ability to deplete immunosuppressive regulatory T lymphocytes (Tregs) and antitumor activity through a direct cytocidal action on IL-2R-expressing tumors.

    In 2021, denileukin diftitox received regulatory approval in Japan for the treatment of relapsed or refractory CTCL and peripheral T-cell lymphoma (PTCL). Subsequently, in 2021, Citius acquired an exclusive license with rights to develop and commercialize denileukin diftitox in all markets except for India, Japan and certain parts of Asia. LYMPHIR (denileukin diftitox-cxdl) was approved by the FDA in August 2024.

    About Cutaneous T-cell Lymphoma

    Cutaneous T-cell lymphoma is a type of cutaneous non-Hodgkin lymphoma (NHL) that comes in a variety of forms and is the most common type of cutaneous lymphoma. In CTCL, T-cells, a type of lymphocyte that plays a role in the immune system, become cancerous and develop into skin lesions, leading to a decrease in the quality of life of patients with this disease due to severe pain and pruritus. Mycosis Fungoides (MF) and Sézary Syndrome (SS) comprise the majority of CTCL cases.  Depending on the type of CTCL, the disease may progress slowly and can take anywhere from several years to upwards of ten to potentially reach tumor stage. However, once the disease reaches this stage, the cancer is highly malignant and can spread to the lymph nodes and internal organs, resulting in a poor prognosis. Given the duration of the disease, patients typically cycle through multiple agents to control disease progression. CTCL affects men twice as often as women and is typically first diagnosed in patients between the ages of 50 and 60 years of age. Other than allogeneic stem cell transplantation, for which only a small fraction of patients qualify, there is currently no curative therapy for advanced CTCL.           

    INDICATION

    LYMPHIR is an IL2-receptor-directed cytotoxin indicated for the treatment of adult patients with r/r Stage I-III cutaneous T-cell lymphoma (CTCL) after at least one prior systemic therapy.

    IMPORTANT SAFETY INFORMATION

    BOXED WARNING: CAPILLARY LEAK SYNDROME

    Capillary leak syndrome (CLS), including life-threatening or fatal reactions, can occur in patients receiving LYMPHIR. Monitor patients for signs and symptoms of CLS during treatment. Withhold LYMPHIR until CLS resolves, or permanently discontinue based on severity.

    WARNINGS AND PRECAUTIONS

    Capillary Leak Syndrome

    LYMPHIR can cause capillary leak syndrome (CLS), including life-threatening or fatal reactions. CLS was defined in the clinical trials as the occurrence of at least 2 of the following symptoms at any time during LYMPHIR therapy: hypotension, edema, and serum albumin <3 g/dL. These symptoms were not required to occur simultaneously to be characterized as capillary leak syndrome.

    As defined, CLS occurred in 27% of patients in the pooled population across 3 clinical trials, including 8% with Grade 3. There was one (0.8%) fatal occurrence of CLS. Of the patients with CLS, 22% had recurrence. The majority of CLS events (81%) occurred within the first 2 cycles of treatment. The median time to onset from Cycle 1, Day 1 was 6.5 days (range: 1 to 77), the median duration of CLS was 14 days (range: 2 to 40), and 75% of patients had resolution. The most common symptoms included edema, hypoalbuminemia, and hypotension. Pleural effusion, pericardial effusion, and dehydration also occurred.

    Regularly assess patients for weight gain, new onset or worsening of edema, dyspnea, and hypotension (including orthostatic changes). Monitor serum albumin levels prior to the initiation of each cycle of therapy and more often as clinically indicated.

    Withhold, reduce dose, or permanently discontinue based on severity. If LYMPHIR is withheld, resume LYMPHIR following resolution of CLS and when serum albumin is greater than or equal to 3 g/dL.

    Visual Impairment

    LYMPHIR can cause serious visual impairment, including changes in visual acuity and color vision. In the pooled population across 3 clinical trials, visual impairment occurred in 9%, with Grade 1 in 8% and Grade 2 in 1%. The most commonly reported symptom was blurred vision. Of the patients with visual impairment, 67% had resolution of their visual impairment.

    Perform baseline ophthalmic examination and monitor as clinically indicated. If patients experience symptoms of visual impairment, such as changes in visual acuity, changes in color vision, or blurred vision, refer for ophthalmologic evaluation.

    Withhold LYMPHIR until visual impairment resolves or permanently discontinue based on severity.

    Infusion-Related Reactions

    LYMPHIR can cause serious infusion-related reactions. Infusion-related reactions were reported in 69% of patients in the pooled population across 3 clinical trials of patients who received LYMPHIR, with Grade 3 infusion-related reactions in 3.4%. Eighty-three percent of infusion-related reactions occurred in Cycles 1 and 2. The most common symptoms included nausea, fatigue, chills, musculoskeletal pain, vomiting, fever, and arthralgia.

    Premedicate patients for the first three cycles prior to starting a LYMPHIR infusion. Monitor patients frequently during infusion. For Grade 2 or higher infusion reactions, premedicate at least 30 minutes prior to each subsequent infusion with a systemic steroid for at least 3 cycles.

    Interrupt or discontinue LYMPHIR based on severity. Institute appropriate medical management.

    Hepatotoxicity

    LYMPHIR can cause hepatotoxicity. In the pooled safety population, elevated ALT occurred in 70% of patients, with Grade 3 ALT occurring in 22%; elevated AST occurred in 64% of patients, with Grade 3 AST elevation occurring in 9%. For Grade 3 events, median time to onset was 8 days (range: 1 to 15 days); median time to resolution was 15 days (range: 7 to 50 days); all cases of Grade 3 ALT or AST elevations resolved. Elevated total bilirubin occurred in 5% of patients, with Grade 3 occurring in 0.9%.

    Monitor liver enzymes and bilirubin at baseline and during treatment as clinically indicated. Withhold, reduce dose, or permanently discontinue LYMPHIR based on severity.

    Embryo-Fetal Toxicity

    Based on its mechanism of action, LYMPHIR can cause fetal harm when administered to a pregnant woman. Verify the pregnancy status of females of reproductive potential prior to the initiation of LYMPHIR. Advise pregnant women of the potential risk to the fetus. Advise females of reproductive potential to use effective contraception during treatment and for 7 days following the last dose of LYMPHIR.

    ADVERSE REACTIONS

    The most common adverse reactions (≥20%), including laboratory abnormalities, are increased transaminases, albumin decreased, nausea, edema, hemoglobin decreased, fatigue, musculoskeletal pain, rash, chills, constipation, pyrexia, and capillary leak syndrome.

    USE IN SPECIFIC POPULATIONS

    Pregnancy

    Risk Summary
    Based on its mechanism of action, LYMPHIR can cause fetal harm when administered to a pregnant woman. There are no available data on the use of LYMPHIR in pregnant women to evaluate for a drug-associated risk. No animal reproductive and developmental toxicity studies have been conducted with denileukin diftitox.

    Denileukin diftitox-cxdl causes depletion of regulatory T lymphocytes (Treg), immune activation, and capillary leak syndrome, compromising pregnancy maintenance. Advise pregnant women of the potential risk to a fetus.

    In the U.S. general population, the estimated background risk of major birth defects and miscarriage in clinically recognized pregnancies are 2-4% and 15-20%, respectively.

    Lactation

    Risk Summary
    No data are available regarding the presence of denileukin diftitox-cxdl in human milk, the effects on the breastfed child, or on milk production. Because of the potential for serious adverse reactions in breastfed children, advise women not to breastfeed during treatment with LYMPHIR and for 7 days after the last dose.

    Females and Males of Reproductive Potential

    Based on its mechanism of action, LYMPHIR can cause fetal harm when administered to a pregnant woman.

    Pregnancy Testing
    Verify the pregnancy status of females of reproductive potential prior to initiating LYMPHIR.

    Contraception

    Females
    Advise females of reproductive potential to use effective contraception during treatment with LYMPHIR and for 7 days after the last dose.

    Infertility

    Males
    Based on findings in rats, male fertility may be compromised by treatment with LYMPHIR. The reversibility of the effect on fertility is unknown.

    Pediatric Use
    Safety and effectiveness of LYMPHIR in pediatric patients have not been established.

    Geriatric Use
    Of the 69 patients with Stage I-III r/r CTCL who received LYMPHIR, 34 patients (49%) were 65 years of age and older and 10 patients (14%) were 75 years of age and older. Clinical studies of LYMPHIR did not include sufficient numbers of patients 65 years of age and older to determine whether they respond differently from younger adult patients. 

    You may report side effects to the FDA at 1-800-FDA-1088 or www.fda.gov/medwatch. You may also report side effects to Citius Oncology at 1-844-459-6744.

    Please read Important Safety Information and full Prescribing Information, including Boxed WARNING, for LYMPHIR. 

    About Verix

    Verix is a leading provider of advanced commercial optimization solutions for the life sciences industry. Its flagship product, Tovana™, is a patented, AI-powered SaaS platform that enables life science companies to optimize commercial strategy and execution. Tovana integrates cutting-edge technology, advanced machine learning and deep domain expertise into a single, intuitive platform, to transform how life science companies deploy data-driven strategies. The platform analyzes large volumes of data, enabling companies to continuously and consistently automate key commercial functions such as healthcare provider (HCP) targeting, precision forecasting, omnichannel engagement, and patient identification.

    Trusted by Fortune 500 pharmaceutical companies around the world, Verix’s innovative Tovana platform helps teams turn complex data into clear, actionable insights that drive brand performance. Tovana’s agile and easy-to-use platform accelerates decision-making, improves launch execution, and enhances commercial outcomes. For more information, visit www.verix.com.

    About Citius Oncology, Inc.

    Citius Oncology, Inc. (Nasdaq: CTOR) is a platform to develop and commercialize novel targeted oncology therapies. In August 2024, its primary asset, LYMPHIR, was approved by the FDA for the treatment of adults with relapsed or refractory Stage I–III CTCL who had had at least one prior systemic therapy. Management estimates the initial market for LYMPHIR currently exceeds $400 million, is growing, and is underserved by existing therapies. Robust intellectual property protections that span orphan drug designation, complex technology, trade secrets and pending patents for immuno-oncology use as a combination therapy with checkpoint inhibitors would further support Citius Oncology’s competitive positioning. For more information, please visit www.citiusonc.com.

    About Citius Pharmaceuticals, Inc.

    Citius Pharmaceuticals, Inc. (Nasdaq: CTXR) is a biopharmaceutical company dedicated to the development and commercialization of first-in-class critical care products. In August 2024, the FDA approved LYMPHIR, a targeted immunotherapy for an initial indication in the treatment of adults with relapsed or refractory Stage I–III CTCL who had had at least one prior systemic therapy. Citius Pharma’s late-stage pipeline also includes Mino-Lok®, a catheter lock solution to salvage catheters in patients with catheter-related bloodstream infections, and CITI-002 (Halo-Lido), a topical formulation for the relief of hemorrhoids. A pivotal Phase 3 trial for Mino-Lok and a Phase 2b trial for Halo-Lido were completed in 2023. Mino-Lok met primary and secondary endpoints of its Phase 3 trial. Citius is actively engaged with the FDA to outline next steps for both programs. Citius Pharmaceuticals owns 79% of Citius Oncology. For more information, please visit www.citiuspharma.com.

    Forward-Looking Statements

    This press release may contain “forward-looking statements” within the meaning of Section 27A of the Securities Act of 1933 and Section 21E of the Securities Exchange Act of 1934. Such statements are made based on our expectations and beliefs concerning future events impacting Citius Pharma or Citius Oncology. You can identify these statements by the fact that they use words such as “will,” “anticipate,” “estimate,” “expect,” “plan,” “should,” and “may” and other words and terms of similar meaning or use of future dates. Forward-looking statements are based on management’s current expectations and are subject to risks and uncertainties that could negatively affect our business, operating results, financial condition and stock price.  Factors that could cause actual results to differ materially from those currently anticipated, and, unless noted otherwise, that apply to Citius Pharma and Citius Oncology, are: our need for substantial additional funds and our ability to raise additional money to fund our operations for at least the next 12 months as a going concern; our ability to commercialize LYMPHIR and any of our other product candidates that may be approved by the FDA; our ability to use the latest technology to support our commercialization efforts; our ability to maintain Nasdaq’s continued listing standards; our ability to successfully implement and maintain distribution agreements with current or other future distribution partners; potential disruptions or performance issues involving third-party logistics providers; the estimated markets for our product candidates and the acceptance thereof by any market; the ability of our product candidates to impact the quality of life of our target patient populations; risks relating to the results of research and development activities, including those from our existing and any new pipeline assets; our dependence on third-party suppliers; our ability to procure cGMP commercial-scale supply; our ability to obtain, perform under and maintain financing and strategic agreements and relationships; uncertainties relating to preclinical and clinical testing; the early stage of products under development; market and other conditions; risks related to our growth strategy; patent and intellectual property matters; our ability to identify, acquire, close and integrate product candidates and companies successfully and on a timely basis; government regulation; competition; as well as other risks described in our Securities and Exchange Commission (“SEC”) filings. These risks have been and may be further impacted by any future public health risks. Accordingly, these forward-looking statements do not constitute guarantees of future performance, and you are cautioned not to place undue reliance on these forward-looking statements. Risks regarding our business are described in detail in our SEC filings which are available on the SEC’s website at www.sec.gov, including in Citius Oncology’s and Citius Pharma’s Annual Reports on Forms 10-K for the year ended September 30, 2024, filed with the SEC on December 27, 2024, each as amended on January 27, 2025, as updated by our subsequent filings with the SEC. These forward-looking statements speak only as of the date hereof, and we expressly disclaim any obligation or undertaking to release publicly any updates or revisions to any forward-looking statements contained herein to reflect any change in our expectations or any changes in events, conditions or circumstances on which any such statement is based, except as required by law.

    Investor Contact:

    Ilanit Allen
    [email protected] 
    908-967-6677 x113

    Media Contact:

    STiR-communications
    Greg Salsburg
    [email protected] 

    SOURCE Citius Oncology, Inc.

    Continue Reading

  • The Queen hosts a reception for winners of The Queen’s Commonwealth Essay Competition

    The Queen hosts a reception for winners of The Queen’s Commonwealth Essay Competition

    Her Majesty The Queen, Vice-Patron of the Royal Commonwealth Society, hosted a reception for winners and runners-up of The Queen’s Commonwealth Essay Competition 2025.

    At a reception at St James’s Palace, The Queen celebrated the winners of…

    Continue Reading

  • WNBA mock draft 2026: New No. 1 replaces UCLA’s Betts at top

    WNBA mock draft 2026: New No. 1 replaces UCLA’s Betts at top

    While the…

    Continue Reading