Jupiter’s icy moon Europa is an ocean world that is a prime candidate in the search for potential extraterrestrial habitability and life beyond Earth. Europa’s surface hosts many features proposed to originate from brine sources within its…
Blog
-

Busta Rhymes (not Tracy Morgan) checks troll in Miami
Busta Rhymes got a young TikTok creator all in check over the weekend after he boldly and incorrectly identified the rapper as “Saturday Night Live” alum Tracy Morgan.
The “Woo-Hah!! Got You All in Check” and “Calm Down” rap veteran,…
Continue Reading
-

Google Patches AI Flaw That Turned Gemini Into a Spy
Artificial Intelligence & Machine Learning
,
Next-Generation…Continue Reading
-

Samsung Wallet to gain support for digital Porsche keys
Owners of the Porsche Macan and upcoming Porsche Cayenne Electric will be able to unlock and start their cars with their Galaxy phones, Samsung has announced. The cars, alongside other Porsche models, are gaining support for Samsung Wallet’s…
Continue Reading
-

‘Harry Potter’ Is Back On Top
Broadway grosses fell back down to earth last week after the highs of Thanksgiving week.
Overall industry gross dropped 12 percent from the prior week with average ticket price down $16 and attendance down just 2 percent, pointing to the…
Continue Reading
-

Journal of Medical Internet Research
Introduction
Machine learning (ML) has significant potential to enhance diagnostic accuracy, support clinical decision-making, and facilitate early disease detection [-]. Additionally, contemporary ML technologies that streamline administrative tasks, refine billing and coding workflows, and enhance patient flow may substantially benefit hospital operations, particularly in underserved areas facing workforce shortages [-]. As recent ML advancements shift the focus from development to deployment, US health care providers have increasingly embraced these technologies [-], especially through integration with electronic health records (EHRs) [-]. Integrating ML techniques into EHRs offers multiple advantages such as simplifying data sharing, reducing analytical time, and mitigating the risk of data leakage []. These techniques enable the efficient processing and analysis of vast amounts of unstructured EHR data, which is critical for research but remains largely underutilized []. Additionally, ML-driven automation can help alleviate provider burnout by reducing manual data entry, minimizing human errors, and shortening documentation time [,].
Despite significant benefits, several technical and ethical challenges (eg, patient privacy, model accuracy, and data reliability) associated with ML adoption may result in unanticipated adverse consequences for patients [,-]. Moreover, the implementation of new health technology is often complicated by various social and organizational factors, posing potential risks to care delivery [,]. For instance, hospitals serving more vulnerable populations may lack adequate technical support or sufficient staff when implementing ML in EHRs []. Thus, both the benefits and challenges warrant further exploration of the factors influencing ML adoption, as hospitals continue to enhance clinical outcomes and operational efficiency by embracing new health technologies.
Although prior work has examined the implementation of ML in hospital settings [-], most studies concentrated on overall adoption, typically measured as the presence of any ML functions or the number of ML functions adopted in EHRs. Such an oversimplified classification treats ML-enabled tools as a homogeneous technology and obscures substantial variation across ML functions. For instance, ML for patient scheduling and ML for inpatient risk prediction differ markedly in data requirements, implementation complexity, and regulatory oversight [-]. Prior studies may mask meaningful heterogeneity in real-world ML implementation. Moreover, hospital adoption patterns may vary across particular ML domains or functions depending on distinct organizational needs, resource capacities, and technological benefits (eg, a large academic hospital may prioritize ML for clinical decision support, while a smaller facility may favor administrative automation). Prior research focusing solely on overall adoption may fail to capture this nuance. To address these gaps and better reflect the complexity of ML diffusion in real-world hospitals, our study differentiates 3 dimensions of adoption: (1) overall ML adoption, (2) domain-specific ML adoption at different levels, and (3) adoption of individual ML functions. This multidimensional approach yields a more granular understanding of how hospitals implement ML within EHRs and how adoption patterns differ by institutional context.
In this study, our goals are to (1) describe the status quo of ML adoption in EHRs across US general acute hospitals and (2) identify hospitals and characteristics associated with ML adoption in EHRs, guided by the Technology-Organization-Environment (TOE) framework. Leveraging a national hospital sample and regression analysis, our study offers valuable insights for policymakers, health care administrators, and clinicians regarding the current landscape of ML adoption and for identifying strategies to navigate the opportunities and challenges posed by emerging technologies.
Methods
Conceptual Framework
In this study, we integrated the TOE framework with our regression analysis to explain the characteristics that are associated with the ML adoption in EHRs [,]. TOE suggests that ML adoption decisions are shaped by 3 contexts: technology, organization, and external environment. The technological context reflects technical benefits and costs of the current ML innovation, such as ML relative advantage, implementation complexity, compatibility with existing EHR systems and workflows, and technical maturity. The organizational context captures hospital internal readiness and capability, including hospital size and structure, leadership support, IT staffing, available financial resources, and IT-supportive organizational culture. Environmental context encompasses all external enablers or barriers outside hospitals such as market competition, artificial intelligence (AI)–related regulations and policy, or EHR vendor support. In health care settings, TOE has been widely used to explain health technology adoption in hospitals [-]. These characteristics in TOE offer inevitable clues of understanding who adopts, who benefits, and where digital capability gaps persist. Recognizing these factors in the adoption of ML technology is critical for designing future policies and practice interventions that promote equitable and effective ML implementation across the US health care system.
Data and Sample
We utilized data from the 2022‐2023 American Hospital Association (AHA) Annual Survey and the 2023‐2024 AHA IT Supplement Survey to examine the relationship between hospital characteristics and ML adoption into EHR []. Each year, the AHA surveys health care administrators from 6200 US hospitals and 400 health care systems, collecting information on hospital demographics, operations, and financial metrics. In addition to the main survey, the AHA conducts the IT Supplement Survey, which gathers data on health IT adoption, tools, and barriers across several domains, including patient engagement, social determinants of health, health information exchange, EHR systems, and IT vendors. The IT Supplement Survey is typically completed by the chief information officer, and participation is voluntary. Beginning in 2023, the IT Supplement introduced a series of questions on ML and other predictive models, enabling the investigation of various aspects of ML adoption, such as functionality, developers, and model evaluation practices.
Our study cohort comprises US general and acute care hospitals operating across all 50 states and the District of Columbia that successfully responded to the 2023‐2024 AHA IT Supplement Survey. After applying pairwise deletion for missing data and implementing other exclusion criteria, the final analytic sample includes 2562 unique hospitals across multiple years, resulting in a total of 4055 hospital-year observations.
ML Measures
ML adoption is measured based on 3 sets of outcomes using the survey questions from the AHA IT Supplement. The first outcome is a binary indicator for any ML adoption, which equals 1 if a hospital uses any ML or other predictive models. Second, to capture the specific uses of this ML, we used a follow-up question to create a 4-category measure. This question asked respondents to select from a list of specific ML applications, which we grouped into 2 domains. Clinical functions included (1) predicting health trajectories or risks for inpatients, such as early detection of conditions like sepsis or in-hospital fall risk; (2) identifying high-risk outpatients to inform follow-up care (eg, readmission risk); (3) monitoring health through integration with wearable devices; and (4) recommending treatments by identifying similar patients and their outcomes. Operational functions included (5) simplifying or automating billing procedures and (6) facilitating scheduling, such as predicting no-shows or optimizing block utilization. Based on both domains, we classified hospitals’ ML use into 1 of 4 mutually exclusive categories: (1) no ML adoption, (2) adoption of clinical functions only, (3) adoption of operational functions only, and (4) adoption of both. This categorical outcome allows for a more nuanced analysis of a hospital’s implementation strategy and provides an interpretable proxy for its organizational readiness. Open-ended responses for other functions were excluded due to their low frequency. Finally, we also analyzed the adoption of individual ML functions listed above.
Hospital and Environmental Factors
Inspired by the TOE framework and prior studies of health IT adoption [], we selected the following hospital characteristics as explanatory variables. For organizational context, we included hospital ownership (defined as nonfederal governmental, not-for-profit, and for-profit), bed size (classified as small [0‐99 beds], medium [100‐399 beds], and large [400 or more beds]), critical access hospital (CAH) status, and teaching hospital status. Regarding environmental context, we included health system affiliation (defined as whether hospitals are affiliated with health systems), whether to contract with leading EHR vendors which is identified by EHR market share [], and metropolitan location (ie, if a hospital is located in a county identified by Rural-Urban Continuum Codes Category 1‐3, it is defined as a metropolitan hospital, otherwise a nonmetropolitan hospital). Due to the lack of relevant technical context information in the AHA survey, this study did not include the technological context.
Statistical Analysis
First, we reported descriptive statistics for ML adoption, specific ML functions adopted in EHR, who developed ML in EHR, and how hospitals evaluated ML models (ie, accuracy, bias, and postimplementation). Second, we summarized ML adoption status by different hospital characteristics. Chi-squared tests were performed to determine statistically significant differences between hospitals with and without ML adoption.
We further assessed the associations between hospital characteristics and ML adoption using multivariate regressions. We used logistic regressions for ML adoption and multinomial logistic regression for a 4-category measure of ML adoption type. Besides explanatory variables listed above, all regressions controlled for year and geographic region based on Census Divisions (ie, New England, Middle Atlantic, East North Central, West North Central, South Atlantic, East South Central, West South Central, Mountain, and Pacific). For all outcomes, we report marginal effects with 95% CIs, which can be directly interpreted as changes in percentage points in the likelihood for a specific category of ML adoption [,]. SEs are clustered at the hospital level to account for a within-hospital correlation over time. We consider estimates to be statistically significant at P<.05. All statistical analyses are performed using StataNow/SE 18.5 version [].
Given high nonresponse rates in the AHA IT Supplement, we compared the characteristics of respondents versus nonrespondents and found that nonrespondent hospitals were more likely to be smaller, less system affiliated, and located in rural areas—factors which are potentially associated with lower adoption rates (Table S1 in ). To address potential nonresponse bias and produce nationally representative estimates, we constructed inverse probability weights (IPWs) from the regression that estimated each hospital’s probability of responding to the AHA IT Supplement (propensity score) using the same set of characteristics described above as predictors [,]. IPW mitigates selection bias by reweighting observations based on their propensity to respond to the AHA IT Supplement, thereby creating a pseudo-population where the distribution of observed covariates is balanced between respondents and nonrespondents. Specifically, for each year, we fitted a separate logistic regression model and predicted the probability of responding to the AHA IT Supplement. We then constructed the analysis weights as the inverse of this predicted response probability and applied these weights in all our statistical analyses. After adjusting by IPWs, the previously observed discrepancies between respondents and nonrespondents became smaller (Table S1 in ), which suggests mitigating the nonresponse bias. All descriptive analyses and regressions are weighted using the IPWs. It is important to note that this IPW approach assumes that nonresponse depends only on observables; accordingly, it corrects for bias related to the measured hospital characteristics. However, any unobserved factors correlated with both survey nonresponse and ML adoption could still bias our estimates, which could not be fully addressed in our study.
Ethical Considerations
Our study is a secondary analysis of organization-level data whose original survey design and administration are described in AHA technical documentation [], and does not involve human participants. Under the US Common Rule (45 CFR §46), research using deidentified, organization-level data does not constitute human subjects research and therefore does not require Institutional Review Board oversight []. We accessed the AHA data via Wharton Research Data Services (WRDS), under data sharing agreements between the AHA and WRDS and between WRDS and the University of Alabama at Birmingham. The results are reported in accordance with the Strengthening the Reporting of Observational Studies in Epidemiology (STROBE) guidelines.
Results
Descriptive Analysis
Overall ML Adoption
The flowchart of our sample selection is presented in Figure S1 in . Among our 2023‐2024 analytical sample, 73%‐76% of the hospitals reported any ML functionality within their EHR system (Figure S2 in ). The majority of US hospitals tend to adopt both clinical and operative ML functions, while the share increased by 10.7 percentage points from 2023 to 2024.
As summarized in , 83.8% of the metropolitan hospitals adopted ML in EHRs, while only 61.3% of the nonmetropolitan hospitals adopted MLs (all subsequent percentages follow the same interpretation). Adoption rates were higher among large hospitals (94.4% vs 65.0% for small), those contracted with leading EHR vendors (92.1% vs 56.7% for nonleading EHR), not-for-profit hospitals (82.7% vs 69.8% for for-profit), non-CAHs (82.2% vs 57.4% for CAHs), those affiliated with health systems (87.9% vs 40.1% for non-health-system), and teaching hospitals (92.1% vs 56.7% for non-teaching). The geographic distribution of any ML adoption is exhibited in Figure S3 in .
Table 1. ML adoption in electronic health record systems by hospital characteristics. Without ML (%) With ML (%) P value (χ2 test) Organizational context Hospital type <.001 Nonfederal, governmental 48.6 51.4 Not-for-profit 17.4 82.7 For profit 30.2 69.8 Hospital size <.001 Small, 0‐99 beds 35.0 65.0 Medium, 100‐399 beds 17.9 82.1 Large, 400 or more beds 5.6 94.4 Critical access hospital <.001 No 17.8 82.2 Yes 42.6 57.4 Teaching status <.001 No 25.8 74.2 Yes 6.7 93.3 Environmental context Health system <.001 No 59.9 40.1 Yes 12.1 87.9 Leading EHR, <.001 No 43.3 56.7 Yes 7.9 92.1 Metropolitan <.001 No 38.7 61.3 Yes 16.2 83.8 Region <.001 New England 27.3 72.7 Middle Atlantic 20.3 79.7 East North Central 23.8 76.2 West North Central 26.3 73.7 South Atlantic 11.6 88.4 East South Central 21.8 78.6 West South Central 40.7 59.3 Mountain 26.6 73.4 Number of hospital-year observations 825 3230 aML: machine learning.
bThe data are from the 2022‐2023 American Hospital Association (AHA) Annual Survey and 2023‐2024 AHA IT Supplement. Percentages are weighted using propensity-score inverse-probability weights to adjust for nonresponse to the IT Supplement.
cLeading EHR vendors are identified by market share.
dEHR: electronic health record.
eMetropolitan status is categorized based on the 2023 Rural-Urban Continuum Codes: 1‐3 as metropolitan counties and 4‐9 as nonmetropolitan counties.
Specific ML Function
The most widely adopted functions (Figure S4 in ) are predicting health trajectories or inpatient risks and identifying high-risk outpatients for follow-up care. From 2023 to 2024, the adoption of clinical ML functions remained stable, whereas operational functions rose markedly, specifically by 19.9 percentage points for simplifying or automating billing procedures and by 14.3 percentage points for facilitating scheduling.
ML Development
Hospitals leveraged multiple resources to develop ML or other predictive models (Figure S5 in ). About 70%‐80% of hospitals reported that their ML was developed by their EHR vendors. The third-party developers and in-house IT teams continued to play meaningful roles in ML development, which likely reflects hospitals’ needs in customizing features to meet varying clinical and administrative needs. Notably, the share of hospitals unsure of the source rose sharply from 1.0% in 2023 to 16.8% in 2024.
ML Model Evaluation
Among hospitals adopting ML in 2023, 62.6% reported evaluating model accuracy for all or most models, while only 45% assessed model bias (Figure S6 in ). These rates of model accuracy and bias evaluation increased to 70.7% and 56.9% in 2024. In 2024, 58.1% of the hospitals conducted postimplementation evaluation and monitoring.
Regression Analysis Results
Whether to Adopt Any ML
The IPW weighted regression estimates of the associations between ML adoption and hospital characteristics are reported in . Compared to nonfederal governmental hospitals, not-for-profit hospitals were 4.4 (95% CI 0.6-8.2) percentage points more likely to adopt any ML within EHR systems (P=.02), while for-profit hospitals were 8.5 (95% CI −15 to −2.1) percentage points (P=.009) less likely to adopt. Large hospitals were 15.2 (95% CI 9.4-21) percentage points more likely than small hospitals to adopt ML into EHRs (P<.001). Hospitals affiliated with a health system had a 26.8 (95% CI 22.4-31.3) percentage point higher likelihood of ML adoption (P<.001). In contrast, CAHs were 8.4 (95% CI –12.4 to –4.4) percentage points less likely to adopt ML (P<.001). Contracting with leading EHR vendors was associated with a 20.6 (95% CI 17.1-24) percentage points increase in ML adoption likelihood (P<.001). Hospitals located in metropolitan areas were 4.3 (95% CI 0.8-7.8) percentage points more likely to adopt ML technology compared to those in nonmetropolitan areas (P=.02).
Table 2. Associations between hospital characteristics and machine learning adoption. Marginal effects 95% CI P value Organizational context Hospital type Nonfederal, governmental Reference — — Not-for-profit 4.4 0.6 to 8.2 .02 For profit –8.5 –14.9 to –2.1 .009 Hospital size Small, 0‐99 beds Reference — — Medium, 100‐399 beds 3.2 –0.7 to 7.2 .11 Large, 400 or more beds 15.2 9.4 to 21.0 <.001 Critical access hospital –8.4 –12.4 to –4.4 <.001 Teaching status –2.1 –12.2 to 7.9 .68 Environmental context Health system 26.8 22.4 to 31.3 <.001 Leading EHR, 20.6 17.1 to 24.0 <.001 Metropolitan 4.3 0.8 to 7.8 .02 Census division New England Reference — — Middle Atlantic –2.7 –10.8 to 5.4 .51 East North Central 0.7 –6.9 to 8.2 .86 West North Central 6.8 –0.5 to 14.1 .07 South Atlantic 9.4 2.0 to 16.8 .01 East South Central 2.8 –5.9 to 11.4 .53 West South Central –1.8 –9.4 to 5.8 .65 Mountain 6.3 –1.2 to 13.8 .10 Pacific –2.0 –10.1 to 6.1 .62 Year 2024 1.9 0.04 to 3.7 .05 aAuthors’ analysis of data from the 2022‐2023 American Hospital Association (AHA) Annual Survey linked with the 2023‐2024 AHA IT Supplement. The sample includes 4055 hospital-year observations. The table reports average marginal effects, which represent the change in the probability of any machine learning adoption associated with each hospital characteristic. Marginal effects are scaled by 100 and can be interpreted as percentage point changes. Estimates are derived from a weighted logistic regression model using inverse probability weights (derived from propensity scores) to account for IT Supplement nonresponse. SEs are clustered at the hospital level to account for within-hospital correlation over time.
bNot applicable
cP<.05.
dP<.01.
eP<.001.
fLeading EHR vendors are identified by market share.
gEHR: electronic health record.
hMetropolitan status is categorized based on 2023 Rural-Urban Continuum Codes: 1‐3 as metropolitan counties and 4‐9 as nonmetropolitan counties.
Type of ML Adoption
The multinomial logistic regression results in illustrate the association between hospital characteristics and 4 types of ML adoption. The results of both clinical and operational ML adoption are overall consistent with the estimates on any ML adoption: not-for-profit, large size, health system affiliation, non-CAH, leading EHR contract, and metropolitan location were related to both ML adoption. Regarding other types of ML adoption, for-profit (−15.1; 95% CI −20 to −10.2; P<.001) and teaching (−6.2; 95% CI −11 to −1.4; P<.001) hospitals were less likely to adopt only clinical ML. Large size (−1; 95% CI −2 to −0.1; P=.03) and health system affiliation (−1.8; 95% CI −3.4 to −0.2; P=.03) were related to a lower likelihood of only operational ML adoption.
Table 3. Associations between hospital characteristics and types of machine learning adoption in electronic health records. Marginal effects 95% CI P value Panel A: clinical ML only Organizational context Hospital type Nonfederal, governmental Reference — — Not-for-profit –1.2 –5.7 to 3.3 .60 For profit –15.1 –20.0 to –10.2 <.001 Hospital size Small, 0‐99 beds Reference — — Medium, 100‐399 beds –0.1 –3.8 to 3.5 .95 Large, 400 or more beds 2.7 –3.3 to 8.7 .38 Critical access hospital 1.9 –2.0 to 5.9 .34 Teaching status –6.2 –11.0 to –1.4 .01 Environmental context Health system 2.3 –1.8 to 6.5 .28 Leading EHR, –0.4 –3.8 to 2.9 .80 Metropolitan 0 –3.4 to 3.4 .98 Census division New England Reference — — Middle Atlantic –7.8 –16.6 to 1.0 .08 East North Central –16.8 –24.6 to –9.0 <.001 West North Central –8.4 –16.7 to –0.1 .05 South Atlantic –15.5 –23.5 to –7.5 <.001 East South Central 1.2 –8.7 to 11.1 .82 West South Central –13.7 –21.9 to –5.6 <.001 Mountain –12.5 –21.0 to –4.1 .004 Pacific –17.9 –26.1 to –9.8 <.001 Year 2024 –8.8 –10.9 to –6.7 <.001 Panel B: operational ML only Organizational context Hospital type Nonfederal, governmental Reference — — Not-for-profit 0.4 –0.7 to 1.6 .45 For profit –0.4 –1.9 to 1.1 .59 Hospital size Small, 0‐99 beds Reference — — Medium, 100‐399 beds 0.3 –0.7 to 1.4 .52 Large, 400 or more beds –1.0 –2.0 to –0.1 .03 Critical access hospital –0.5 –1.7 to 0.7 .40 Teaching status 1.6 –1.7 to 4.9 .34 Environmental context Health system –1.8 –3.4 to –0.2 .03 Leading EHR 0.6 –0.1 to 1.3 .11 Metropolitan –0.2 –1.3 to 0.9 .72 Census division New England Reference — — Middle Atlantic –0.7 –2.0 to 0.6 .31 East North Central 1.6 –0.1 to 3.3 .06 West North Central 0.5 –1.4 to 2.3 .61 South Atlantic 2.0 0.2 to 3.8 .03 East South Central 0.1 –1.8 to 1.9 .94 West South Central 0.2 –1.6 to 2.0 .80 Mountain 0 –1.6 to 1.6 .97 Pacific –0.9 –2.1 to 0.4 .17 Year 2024 1.6 0.9 to 2.3 <.001 Panel C: both types Organizational context Hospital type Nonfederal, governmental Reference — — Not-for-profit 6.0 0.8 to 11.1 .02 For profit 7.6 0.4 to 14.8 .04 Hospital size Small, 0‐99 beds Reference — — Medium, 100‐399 beds 2.8 –1.7 to 7.3 .22 Large, 400 or more beds 13.5 6.7 to 20.2 <.001 Critical access hospital –9.9 –14.8 to –5.0 <.001 Teaching status 3.2 –5.7 to 12.0 .48 Environmental context Health system 26.5 21.7 to 31.3 <.001 Leading EHR 20.5 16.6 to 24.3 <.001 Metropolitan 4.6 0.3 to 8.9 .03 Census division New England Reference — — Middle Atlantic 5.9 –3.8 to 15.7 .23 East North Central 16.2 6.9 to 25.5 <.001 West North Central 14.7 5.4 to 24.0 .002 South Atlantic 23.0 13.7 to 32.3 <.001 East South Central 1.7 –9.0 to 12.4 .75 West South Central 11.9 2.3 to 21.5 .02 Mountain 19.0 9.0 to 28.9 <.001 Pacific 17.3 7.4 to 27.1 <.001 Year 2024 8.8 6.5 to 11.1 <.001 aAuthors’ analysis of data from the 2022‐2023 American Hospital Association (AHA) Annual Survey linked with the 2023‐2024 AHA Information Technology (IT) Supplement. The sample includes 4055 hospital-year observations. The table presents average marginal effects (MEs) from a single multinomial logistic regression, where the outcome is a 4-category measure of ML adoption type: clinical only (Panel A), operational only (Panel B), both (Panel C), and no ML adoption (the base outcome). Each ME represents the change in the probability of being in a specific category associated with a hospital characteristic. The model is weighted using inverse probability weights (derived from propensity scores) to account for the IT Supplement nonresponse. MEs are scaled by 100 and can be interpreted as percentage point changes. SEs are clustered at the hospital level to account for within-hospital correlation over time.
bML: machine learning.
cP<.001.
dP<.05.
eLeading EHR vendors are identified by market share.
fEHR: electronic health record.
gMetropolitan status is categorized based on the 2023 Rural-Urban Continuum Codes: 1‐3 as metropolitan counties and 4‐9 as nonmetropolitan counties.
hP<.01.
iNot applicable.
Individual ML Functions
We further explored the relationship between hospital features and 6 individual ML functions in Tables S2 and S3 in . Overall, the patterns were consistent with the main findings in , but some exceptions emerged. Nonprofit hospitals showed a stronger emphasis on clinical applications, including predicting health risks for inpatients (7.5; 95% CI 3.6-11.4; P<.001) and outpatients (7.7; 95% CI 3.4-12; P<.001). In contrast, for-profit hospitals (26.4; 95% CI 19.3-33.6; P<.001) and large hospitals (11.6; 95% CI 5.3-18; P<.001) were more likely to adopt scheduling-related ML functions. To enable clearer review and a comparison of evidence across various sets of outcomes, we provided a summary overview in .
Table 4. Summary of all regression results. Overall Types of adoption Adoption of individual function Clinical only Operational only Both types Predict inpatient risk Predict outpatient risk Monitor health Recommend treatments Billing Scheduling Organizational context Hospital type (reference: nonfederal, governmental) Not-for-profit + NS NS + + + NS NS NS NS For profit – – NS + NS – – – NS + Hospital size (reference: small, 0-99 beds) Medium, 100‐399 beds NS NS NS NS NS NS NS NS NS + Large, 400 or more beds + NS – + + + NS NS + + Critical access hospital – NS NS – – – NS NS NS – Teaching status NS – NS NS NS NS NS NS NS NS Environmental context Health system + NS NS + + + + + + + Leading EHR, + NS NS + + + + + + + Metropolitan + NS NS + NS NS + + NS NS a+ means a significant positive association at the .05 level.
bNS means no significant association.
c– means a significant negative association at the .05 level.
dLeading EHR vendors are identified by market share.
eEHR: electronic health record.
fMetropolitan status is categorized based on 2023 Rural-Urban Continuum Codes: 1‐3 as metropolitan counties and 4‐9 as nonmetropolitan counties.
Additional Analysis Results
To check the robustness of our findings, we estimated models without IPWs and compared them with our main estimates (Tables S4 and S5 in ). We found both weighted and nonweighted estimates converge. The estimates are consistent with prior research on hospital AI adoption for workforce optimization [,], lending credibility to our findings.
We also conducted the subgroup analysis to detect heterogeneity patterns in different hospital sizes (Table S6 in ). Health system affiliation and leading EHR vendors significantly increased the probability of ML adoption, and both effects are larger in small hospitals than in medium hospitals.
Discussion
Principal Results
This study provides timely, national-level evidence on ML adoption within hospital EHR systems, revealing substantial variation in specific functions, model evaluation practices, and hospital characteristics. Consistent with prior research on overall ML adoption, organizational and environmental factors such as hospital type, size, CAH status, health system membership, and metropolitan location remain key determinants of adoption decisions, supporting the organizational and environmental contexts in the TOE framework. However, our findings extend beyond existing studies by identifying marked heterogeneity in the adoption of individual ML domains and functions, highlighting that hospitals differ not only in whether they adopt ML but also in which types of ML applications they prioritize. Such heterogeneity aligns well with the technological context in the TOE framework, suggesting that ML adoption may be influenced by each hospital’s assessment of the relative benefits and costs. Understanding this heterogeneity is essential for designing further policies and implementation strategies that promote equitable and effective ML diffusion across diverse health care settings.
Resource-Based Explanation
Within the TOE framework, the organizational and environmental contexts offer a resource-based explanation for our findings, suggesting that both internal and external resources facilitate or restrict hospital technology adoption. Prior literature has identified such resources as key determinants of EHR adoption [,]. Our findings similarly suggest that hospitals with characteristics commonly associated with greater organizational capacity (eg, larger size, urban location, and system affiliation) were more likely to report ML adoption. Although we did not directly document the decision-making process or measure organizational resources, these characteristics may serve as proxies for financial resources, human capital, IT capabilities, and leadership support. Large, urban, and system-affiliated hospitals typically possess the financial flexibility to purchase new add-on functions from EHR vendors or develop custom ML algorithms tailored to their specific needs. Their established IT infrastructure facilitates efficient ML adoption within EHRs, and their resource-rich environments often have sufficient staffing and training programs to support implementation. Executive leadership and strategic vision also play a critical role in advancing ML adoption [].
Further, our findings suggest that the unequal distribution of health care resources acts as a significant barrier to ML adoption for hospitals in rural and underserved areas [-] and that external resource support could facilitate ML adoption in such hospitals. Our subgroup analysis findings showed that health system affiliation and leading EHR contract had stronger effects on small hospitals, partly due to the greater availability of financial and technical support.
Heterogeneous ML Adoption
Although most hospitals tend to adopt both clinical and operational ML functions simultaneously, specific hospitals select particular ML applications in alignment with their technical needs, as the TOE framework suggests. Most notably, for-profit hospitals demonstrated a strong interest in adopting ML regarding scheduling functions, which may be explained by staffing structures and institutional objectives. They more frequently experience staffing shortages [] and higher nursing turnover rates []. In this context, ML-enabled scheduling tools offer a practical solution for maintaining service quality at a lower cost by automating front-office functions and reducing reliance on staff.
These findings suggest that hospitals vary in their perceptions, motivations, and priorities when adopting ML technologies, which are often shaped by their resources and strategic goals. This variability highlights the need for qualitative research and targeted surveys on the motives, facilitators, and barriers to implementing emerging ML technologies among health care providers and health system or hospital managers. For instance, researchers may explore how hospital managers perceive and evaluate specific ML technology, enablers and challenges during implementation, and whether ML models meet their expectations.
Limited Model Evaluation
Our findings highlight a notable gap in current hospital AI practice: limited evaluation of models using local hospital or health system data [-]. Given that ML performance is highly dependent on the distribution and quality of the training data in specific contexts, its real-world effectiveness and validity in dynamic health care environments are often unclear and may degrade for specific patients. Moreover, if the training data contain inherent biases, it may perpetuate or even exacerbate existing systemic disparities among vulnerable populations [-].
Researchers in venues such as the Machine Learning for Healthcare conference and the Conference on Human Factors in Computing Systems have emphasized practical frameworks for evaluating ML quality in care delivery, including real-world effectiveness testing, model actionability, model lifecycle management, and performance tracking across sites and settings [-]. To mitigate risk and support safe integration, developers and scholars recommend tailoring evaluation to the clinical context, engaging clinicians and patients, increasing transparency, labeling bias across patient groups, and fostering a supportive organizational culture [-]. This gap also underscores the need for governmental and industrial initiatives and regulations to promote the safe, effective, and trustworthy use of ML technologies []. Although nascent [-], these efforts could provide critical oversight, standards, and incentives that help minimize risks and encourage responsible implementation.
Policy or Practice Suggestion
Multi-level policy interventions are necessary to ensure equitable ML adoption and bridge the digital divide across health care settings. The federal government may introduce an initiative similar to the Health Information Technology for Economic and Clinical Health Act of 2009, which significantly accelerated the adoption of meaningful EHR use among rural and small hospitals [,]. These interventions could include expanding IT broadband infrastructure in underserved areas or providing direct financial incentives to support ML adoption. Meanwhile, health systems and EHR vendors should consider providing additional financial and technical support specifically for rural and small hospitals. Especially, EHR vendors should prioritize integrating foundational ML functions within the existing EHR systems to reduce the technical barriers of meaningful ML usage. Also, offering discounted or complimentary updates for ML features may be a valuable strategy to support hospitals with limited resources.
Limitation
Our study has several limitations. First, our findings should be interpreted as correlational rather than causal, due to unobserved factors. Second, our analysis relies on secondary data from the AHA, which may be subject to self-report bias. Although our design cannot fully address these biases, the AHA data are widely used in health services research as reliable data sources, and they provide unique and detailed information on hospitals’ adoption of ML in EHR unavailable elsewhere. Future studies could validate the AHA data through primary data collection or cross-validation with objective measures. Third, our measures of ML adoption are self-reported by hospital managers, which may be subject to recall error, social-desirability bias, or heterogeneous interpretations of ML functionality. As a result, hospitals may overreport or misreport their level of ML adoption. In addition, we lack several granular contextual variables such as vendor-specific implementation details, interoperability maturity, IT workforce perspectives, and organizational culture, which provide necessary scenarios to understand ML implementation in practice. To address these gaps, future researchers could leverage qualitative and mixed methods designs (eg, interview or focus groups with clinicians, hospital managers, IT staff, and vendor representatives) to capture nuanced adoption processes, identify facilitators and barriers, and document practical implementation strategies. Fourth, the AHA IT Supplement Survey began collecting ML implementation in 2023, which limited our ability to capture early efforts in ML adoption within EHR. Finally, 44.7% of the AHA hospitals did not complete the IT Supplement Survey, raising concerns about potential nonresponse bias. Although we applied inverse propensity score weighting to balance observed characteristics between respondents and nonrespondents, which could strengthen our findings, the nonresponse bias remains if unobserved characteristics are correlated with both survey response and ML adoption.
Conclusion
In this retrospective study of US hospitals from 2022 to 2024, we observed a high rate of ML adoption within EHR systems and considerable variation across specific ML functions. We also identified several hospital characteristics associated with ML adoption within EHRs, including ownership, hospital size, health system affiliation, CAH status, and metropolitan location. Given that resource availability significantly influences a hospital’s capacity to implement ML technologies within EHRs, our findings highlight the need for policy interventions that provide financial and technical support. Such efforts are essential to ensure that resource-constrained hospitals can adopt emerging health IT innovations and prevent the widening of digital and health disparities across the health care system.
To ensure accuracy, fairness, and safety, health care administrators and policymakers must prioritize not only the adoption of ML technologies but also the ongoing monitoring and evaluation of these tools. Equitable access to these technologies must be a key focus, with targeted support for hospitals facing barriers to adoption. By fostering inclusive and transparent approaches to ML adoption, we can maximize its potential to improve care delivery, reduce disparities, and enhance health care outcomes for all patients.
The view and content are solely the responsibility of the authors. The authors used ChatGPT (OpenAI) to assist in improving the readability and clarity of limited sections. All artificial intelligence–assisted edits were reviewed and revised by the authors to preserve the manuscript style and ensure scientific accuracy, and the authors take full responsibility for the final content.
This study was funded by Augusta University Open Access Article Processing Charges Fund.
The datasets generated or analyzed during this study are not publicly available due to data user agreements. Interested parties may seek to obtain data licenses directly from the American Hospital Association [] or Wharton Research Data Services [].
None declared.
Edited by Javad Sarvestan; submitted 16.Apr.2025; peer-reviewed by Gabriela Morgenshtern, Horng-Ruey Chua, Michael Kanter; final revised version received 16.Nov.2025; accepted 17.Nov.2025; published 09.Dec.2025.
© Huang Huang, Wei Lyu, Md Mahmud Hasan, Shannon H Houser. Originally published in the Journal of Medical Internet Research (https://www.jmir.org), 9.Dec.2025.
This is an open-access article distributed under the terms of the Creative Commons Attribution License (https://creativecommons.org/licenses/by/4.0/), which permits unrestricted use, distribution, and reproduction in any medium, provided the original work, first published in the Journal of Medical Internet Research (ISSN 1438-8871), is properly cited. The complete bibliographic information, a link to the original publication on https://www.jmir.org/, as well as this copyright and license information must be included.
Continue Reading
-
Kirkland Represents ProAmpac on its Acquisition of TC Transcontinental Packaging from TC Transcontinental | News
Kirkland & Ellis advised ProAmpac, a global innovator in flexible packaging and material science, on its definitive agreement to acquire TC Transcontinental Packaging from TC Transcontinental (TSX: TCL.A TCL.B) for US$1.51 billion (approximately CAD$2.1 billion), subject to customary adjustments for debt and debt-like items, cash and net working capital. The transaction is expected to close in the first quarter of calendar 2026, subject to shareholder approval, regulatory approvals and other customary conditions.
Read ProAmpac’s press release here
The Kirkland team included corporate lawyers Adam Wexner, Kevin Stocks and Alexander Romano; debt finance lawyers Michelle Kilkenney and Carolyn Aiken; capital markets lawyer Alborz Tolou; and tax lawyers Rachel Cantor and Rebecca Fine.
Continue Reading
-

Turner Prize 2025: Winner to be revealed for UK’s top art award
What is the Turner Prize – and how much is the prize money?published at 19:00 GMT
Image source, PA MediaImage caption, Madonna presented the Turner Prize in 2001
It’s arguably the UK’s most famous art prize.
Since its debut in…
Continue Reading


