Author: admin

  • Move Over, ‘Marty Supreme’: Charli XCX’s Custom Puffers Are the Movie Merch Moment

    Move Over, ‘Marty Supreme’: Charli XCX’s Custom Puffers Are the Movie Merch Moment

    The premiere of Charli xcx’s The Moment was always going to be, well, a moment. Out in Park City, Utah at the Sundance Film Festival, fashion leans plushy and cozy, maybe even a little GORP-y. Trust, though, that Charli xcx would inject the…

    Continue Reading

  • adidas Originals x Wily Chavarria release the Chavarria Megaride following its debut at Fall/Winter 2026 Paris Fashion Week – adidas News

    1. adidas Originals x Wily Chavarria release the Chavarria Megaride following its debut at Fall/Winter 2026 Paris Fashion Week  adidas News
    2. Willy Chavarria’s New Adidas Megaride AG Finally Has a Release Date  WWD
    3. Willy Chavarría Takes adidas’…

    Continue Reading

  • All the Caribbean nations at Milano Cortina 2026

    All the Caribbean nations at Milano Cortina 2026

    Athletes from four* Caribbean delegations are in line to compete at the Olympic Winter Games Milano Cortina 2026, spanning Alpine skiing, cross-country skiing, men’s bobsleigh, monobob and women’s skeleton.

    Trinidad and Tobago is set for a…

    Continue Reading

  • Gathering medieval French prayerbook, Kabuki in America, Sylvia Plath’s thoughts — Harvard Gazette

    Gathering medieval French prayerbook, Kabuki in America, Sylvia Plath’s thoughts — Harvard Gazette

    Harvard’s Houghton Library is home to more than half a million books, manuscripts, letters, photos and other ephemera spanning centuries and capturing swaths of human experience. And every year that collection grows even more vast.

    To…

    Continue Reading

  • Samsung Galaxy S26 will have ‘pixel level’ privacy feature, Samsung confirms

    Samsung Galaxy S26 will have ‘pixel level’ privacy feature, Samsung confirms

    Samsung is cooking up a fresh new privacy feature for the next Galaxy phones.

    The company announced in a pretty vague blog post that it’s getting ready to show off new security functions…

    Continue Reading

  • Journal of Medical Internet Research

    Journal of Medical Internet Research

    Background

    When people look for health information today, they no longer only consult physicians, pharmacists, or search engines. Increasingly, they also encounter generative artificial intelligence (AI) tools such as ChatGPT or the World Health Organization (WHO)’s chatbot Sarah, which simulate human-like conversations and provide instant responses. These tools promise a new way of accessing medical knowledge: fast, convenient, and interactive. At first glance, this accessibility seems to hold great potential for reducing barriers to health information, therefore directly impacting digital health equity—defined as equitable access to and use of digital health information technology that supports informed decision-making and enhances health [].

    However, the picture is more complex. On the one hand, generative AI can offer cost-free entry points (eg, basic versions of ChatGPT or automatically displayed answers in Google search via Google’s Gemini), deliver content in multiple languages, and rephrase complex medical concepts into more understandable terms. In doing so, it could strengthen patient education, address health inequalities, and help bridge communication gaps between citizens and health care providers [,]. On the other hand, effective use still depends on internet-enabled devices and adequate digital skills, which are not equally distributed. As a result, the very technology that appears open and inclusive may also risk exacerbating existing digital divides [].

    Moreover, unlike other types of information, health-related questions are often sensitive and personal. At the same time, the inner workings of generative AI remain opaque, and the accuracy of its outputs is not guaranteed []. All these tensions raise important questions about adoption: Who is most likely to turn to generative AI for health information, and what factors shape this intention? Moreover, since health communication practices and digital infrastructures differ across countries, cross-national research is urgently needed.

    To address these questions, this study draws on the Unified Theory of Acceptance and Use of Technology 2 (UTAUT2) []. The model proposes that performance expectancy, effort expectancy, facilitating conditions, social influence, habit, and hedonic motivation shape technology use. We extend this framework by also examining the roles of health literacy and health status in predicting intention to use generative AI for health information. Using cross-national survey data from Austria, Denmark, France, and Serbia, we investigate the drivers of adoption to shed light on both individual and contextual factors that may guide the diffusion of generative AI in health contexts.

    Generative AI as Novel Health Information Source

    Generative AI constitutes a potentially disruptive force in the health information ecosystem []. However, despite its rapid advancement and widespread availability, empirical research on its role in health information–seeking remains limited []. At the same time, broader trends highlight the ongoing digitalization of health-related knowledge acquisition. A representative survey conducted in Germany in 2019 revealed that only 48% of respondents consulted a medical professional for their most recent health issue, while 1 in 3 turned first to the internet []. Similar findings show that online sources—particularly search engines—are the primary means of accessing health information, both for caregivers and the general population [,]. Family and friends and traditional mass media (eg, print media and health-related TV programming) rank behind medical professionals and online sources [].

    The introduction of generative AI tools like ChatGPT may shift these established hierarchies. Unlike static web content or conventional search engines, generative AI enables dialogic, personalized interactions that simulate human conversation. These features may position generative AI as a compelling alternative to established online and offline health information sources. However, current evidence suggests that trust in generative AI—especially regarding complex health-related issues—is still limited [], which might restrict its present adoption potential to early adopters []. This raises questions about how generative AI integrates into the broader ecosystem of health information sources.

    To address this gap, we first explore: How does the use of generative AI for health information–seeking compare to that of more established health information sources?

    Explaining Predictors of Technology Adoption: UTAUT2

    The UTAUT2 [] is one of the most popular models to explain technology adoption. It builds on the technology acceptance model [], emphasizing perceived usefulness and ease of use, and the initial UTAUT model [], which added performance expectancy, effort expectancy, facilitating conditions, and social influence as predictors of adoption behavior. UTAUT2 extends these frameworks to consumer contexts by incorporating hedonic motivation and habit [,] . The UTAUT2 model has demonstrated its versatility in explaining the adoption of diverse eHealth technologies, such as wearable devices [], health websites [], and health apps []. Additionally, recent studies have highlighted its relevance in understanding the uptake of generative AI technologies [-], showcasing its capacity to extend beyond traditional eHealth domains. However, so far, studies on predictors of usage intentions in the context of AI health information–seeking are lacking.

    Performance expectancy, a central construct in the UTAUT2 framework, reflects the belief that using technology will lead to performance benefits []. In the context of health information–seeking using generative AI, performance expectancy is shaped by users’ perceptions of how effectively these tools can enhance their own life, including aspects such as health–decision-making and task efficiency []. Consequently, as users anticipate greater usefulness from adopting generative AI as a health information source, their intention to use such technologies strengthens [-,]. Based on this, we propose the following hypothesis: “the higher the performance expectancy, the stronger the intention to use generative AI for health information–seeking” (H1). Effort expectancy, closely tied to ease of use, emphasizes simplicity in technology adoption []. Generative AI tools like ChatGPT benefit from high effort expectancy when users find them intuitive and easy to integrate into their workflows, particularly during the early adoption phase [,,,]. Addressing usability concerns early can reduce resistance and build user confidence, strengthening behavioral intention [,]. Therefore, we propose that “the higher the effort expectancy, the stronger the intention to use generative AI for health information–seeking” (H2). Facilitating conditions refer to the resources, skills, and support necessary for using technology []. These include training, knowledge, technical assistance, and system compatibility, which significantly enhance behavioral intention and usage [,]. In technologically mature settings, facilitating conditions are critical for sustained adoption and user satisfaction []. In line with the UTAUT2, we hypothesize that “the better the facilitating conditions, the stronger the intention to use generative AI for health information–seeking” (H3). Social influence indicates the perception that peers, such as family, friends, or colleagues, believe one should adopt a technology []. It plays a crucial role in early adoption, where external validation often outweighs personal experience []. Positive reinforcement within social or professional networks can normalize usage [,,,]: If people perceive that their peers already use generative AI for health information–seeking, their own intention to do so might increase as well. We, therefore, propose that “the greater the perceived social influence, the stronger the intention to use generative AI for health information–seeking” (H4). Habit specifies the extent to which behavior becomes automatic through repetition and prior use []. It strongly influences behavioral intention and long-term adoption, emphasizing the importance of regular engagement with technology [,]. For generative AI as a health information source, fostering habitual use can solidify its integration into daily routines and enhance sustained adoption []. This leads us to state, “the more it is a habit to use generative AI, the stronger the intention to use generative AI for health information–seeking” (H5). Hedonic motivation refers to the enjoyment or pleasure derived from using technology, particularly relevant in consumer contexts []. It directly impacts behavioral intention, especially for technologies involving entertainment or leisure []. For generative AI like ChatGPT, it can be expected that the interaction is perceived as fun or entertaining, which can boost user engagement and drive adoption []. Accordingly, we suggest the following hypothesis: “the higher the hedonic motivation, the stronger the intention to use generative AI for health information–seeking” (H6).

    Influence of Health Literacy and Health Status

    With the growing integration of digital tools into everyday lives, the role of health literacy in online health information–seeking has garnered increasing attention. Health literacy has been conceptualized as an individual’s capacity to search, access, comprehend, and critically evaluate health information, as well as to use the acquired knowledge to effectively address health-related issues [,]. Digital health literacy refers to these abilities in the context of digital environments [-]. Generally, low health literacy scores have been associated with undesirable health outcomes [].

    Research suggests that low levels of health literacy are associated with decreased trust in online health resources [], including the outputs of AI tools [], and lower overall adoption of online health technologies []. Furthermore, initial studies indicate a positive association between health literacy levels and attitudes toward the use of AI tools for medical consultations [].

    On the other hand, individuals with higher levels of health literacy are generally better equipped to critically evaluate online health information and scrutinize it in greater detail [,]. This heightened evaluative capacity could make them more aware of the limitations and potential risks of generative AI outputs, such as inaccurate information, bias, data privacy concerns, or oversimplified medical advice []. Moreover, individuals with higher health literacy are more likely to trust and use high-quality medical online resources, whereas those with limited health literacy prefer accessible but potentially less reliable sources []. In this context, outputs from generative AI might be perceived as lower-quality sources by highly digital health–literate individuals. As a result, while higher health literacy could foster openness to using generative AI for health purposes, it might also lead to greater skepticism or hesitancy in relying on these tools. Nonetheless, there is not enough research in the context of generative AI specifically to make conclusive predictions.

    Another well-established factor in online health information–seeking, yet underexplored in the context of AI, is individuals’ health status: Studies suggest that people with poor health are significantly more likely to consult the internet for health information compared to those with good health [,]. Being chronically ill has also been associated with increased reliance on internet-based technologies for health-related purposes []. This relationship can be explained by the fact that individuals in poor health often experience greater health-related concerns, which in turn heightens their motivation to seek information online.

    Given these complex relationships, we propose the second research question: How does health literacy and health status influence the intention to seek health information using generative AI?

    Cross-National Comparison

    In this study, we investigate the predictors of generative AI adoption for health information–seeking across 4 European countries: Austria, Denmark, France, and Serbia. While these countries share certain similarities, they also display notable differences that could shape the strength of the UTAUT2 predictors on the intention to use generative AI for health purposes. Thus, this cross-national approach ensures that the observed effects are generalizable and not confined to specific national contexts or unique country conditions.

    The selected countries share two key characteristics. First, all 4 countries provide universal health coverage, ensuring broad access to health care services for their populations. Second, a significant portion of health care expenditure in these countries is publicly funded [-].

    Despite these commonalities, there are also critical factors that differ among the countries and may shape the predictors of generative AI adoption. On the one hand, variations in digital infrastructure could significantly impact facilitating conditions, effort expectancy, and social influence as predictors of generative AI use. Denmark consistently ranks among Europe’s most digitally advanced nations, boasting high internet penetration and widespread adoption of e-health solutions []. This strong digital ecosystem likely enhances the perceived ease of use and social endorsement of generative AI. In contrast, Austria, France, and Serbia exhibit more moderate levels of digital adoption in the context of health information, which may limit the perceived use and social norms regarding such technologies [].

    On the other hand, access to and trust in health care providers vary significantly across these countries, potentially influencing performance expectancy and social influence. In nations with robust health care systems—characterized by a high availability of medical professionals and easy access to care—individuals are more likely to rely on doctors for health advice, as they are often viewed as the most trusted source of health information []. Denmark exemplifies this with its high levels of public trust in the health care system [], which may reduce the perceived benefits and social norms around using generative AI for health purposes. Conversely, in western Balkan countries like Serbia, studies report generally low levels of trust in the health care system []. In such contexts, individuals may be more inclined to seek alternative information sources, potentially amplifying the perceived benefits of generative AI use.

    By examining these diverse national contexts, this study not only tests the universality of the UTAUT2 model but also deepens our understanding of the contextual factors that shape generative AI adoption for health purposes. We ask: How do the predictors of generative AI use for health information–seeking differ across Austria, Denmark, France, and Serbia?

    Ethical Considerations

    Before data collection, the study received ethical approval from the institutional review board of the Department of Communication at the University of Vienna (approval ID: 1205). All participants provided written informed consent prior to participating in the study. The data were collected in anonymized form and no personal identifiers were recorded or stored. Participants received a compensation of €1.50 (US $1.74) for completing the study through the panel provider.

    Recruitment

    Recruitment of participants occurred during September 2024 via Bilendi, an international panel provider. Bilendi recruited the participants via email. The panel is checked for quality and attendance on a regular basis. The study was conducted in Austria, France, Denmark, and Serbia, with participants randomly selected to achieve samples representative of age, gender, and educational background. The provider’s panel sizes in the respective countries were as follows: Austria: n=60,000; Denmark: n=90,000; France: n=815,000; and Serbia: n=15,100. Per country, the study aimed to reach 500 participants.

    Inclusion criteria required participants to be aged between 16 and 74 years. Additionally, respondents who completed the survey in less than one-third of the median completion time (speeders) were excluded. Completion rates (excluding screened-out participants) were high across all 4 countries, ranging from 84.3% to 89.8%. Further details on survey design, administration, and response rates are provided in the CHERRIES (Checklist for Reporting Results of Internet E-Surveys) checklist ().

    Procedure and Measures

    Overview

    The study consisted of two components. The first component, a survey, investigated predictors of the intention to use generative AI for health information–seeking. The second component, an experimental study, explored the influence of disease-related factors on these intentions []. To ensure respondents shared a common understanding of the concept, the survey began with a short definition of “generative AI,” describing it as technologies that engage in natural language conversations with users and generate responses in real-time. Examples such as ChatGPT, Google’s Gemini, and Microsoft Copilot were provided.

    The original questionnaire was developed in English and subsequently translated into German, French, Danish, and Serbian. Each translation was performed by a bilingual team member and back-translated into English by a different native speaker to ensure conceptual equivalence with the original items.

    This study focused on the following constructs (a complete item list with descriptive analysis and construct reliability values can be found in ).

    Dependent Variables
    Sources of Health Information–Seeking

    Participants rated the frequency with which they use 8 health information [] sources on a 7-point Likert scale (1 = “never” to 7 = “very often”). The sources included conversations with medical professionals, pharmacists, and family or friends, as well as books, mass media, internet search engines (eg, Google and Ecosia), and generative AI.

    AI Usage Intentions

    Participants’ willingness to use generative AI [] for health information–seeking was assessed using 3 items (eg, “I intend to use generative AI for health information seeking”), rated on a 7-point Likert scale (1 = “strongly disagree” to 7 = “strongly agree”).

    UTAUT2 Predictor Variables

    All UTAUT2 model [,,,] predictors were measured on a 7-point Likert scale (1 = “strongly disagree” to 7 = “strongly agree”). The predictors have been described below.

    Performance Expectancy

    Perceived benefits of using generative AI for health information–seeking were measured with 4 items (eg, “Using generative AI would save me time when researching health topics”).

    Effort Expectancy

    The perceived ease of using generative AI as a health information source was assessed with 4 items (eg, “Learning to use generative AI for health information–seeking seems easy for me”).

    Social Influence

    Three items measured the extent to which participants felt that important others encouraged their use of generative AI for health information–seeking (eg, “People who are important to me think that I should use generative AI for health-information seeking”).

    Hedonic Motivation

    The enjoyment of using generative AI for health information–seeking was assessed with 3 items (eg, “I think using generative AI for health-information seeking could be fun”).

    Facilitating Conditions

    Participants’ perceptions of available resources and support for using generative AI to seek health information were measured with 4 items (eg, access to devices and reliable internet, and knowledge).

    Habit

    The extent to which turning to generative AI when seeking health information had become a habitual behavior was measured with 3 items (eg, “I automatically turn to generative AI whenever I have questions about my health”).

    Model Extension Variables

    Health Literacy

    Health literacy [] was assessed with 10 items, asking participants to rate their confidence in tasks such as finding understandable health information. Responses were recorded on a 4-point Likert scale (1 = “not at all true” to 4 = “absolutely true”).

    Health Status

    We measured participants’ health status using 1 item (“How would you describe your current health status?”; 1 = “very poor” to 7 = “very good”).

    Control Variables: Sociodemographic Variables

    We further measured participants’ age, gender, and educational level.

    Statistical Analysis

    Power

    An a priori power analysis was conducted to determine the required sample size for structural equation modeling. Assuming an anticipated effect size of 0.25, a desired statistical power of 0.95, and a significance level of .05, the analysis indicated that a minimum of 391 participants per country would be necessary to detect the hypothesized effects [].

    Analytical Plan

    We used AMOS version 26 (IBM Corp) to run a latent variable, multigroup structural equation model using a maximum-likelihood estimator with full information. We computed the comparative fit index, the Tucker-Lewis Index, the chi-square to degrees of freedom ratio (χ2/df), and the root mean square error of approximation. We also secured metric measurement invariance to be able to compare the paths across countries. We controlled for age, gender, and education (binary coded).

    User Statistics

    In total, data were collected from 1990 respondents, comprising 502 from Austria, 507 from Denmark, 498 from France, and 483 from Serbia. The overall mean age of participants was 45.1 (SD 15.7) years, with 50.2% (n=998) identifying as female participants. In terms of educational attainment, 83.8% (n=1634) of the sample reported completing at least a medium or higher level of education (secondary level II or higher). Furthermore, 87.4% (n=787) of respondents indicated prior use of generative AI for health information–seeking (at least rarely). Detailed demographic and background characteristics of the sample are summarized in .

    Table 1. Descriptive characteristics of survey respondents from Austria, Denmark, France, and Serbia (N=1990; September 2024).
    Demographic characteristics Overall, n (%) Austria, n (%) Denmark, n (%) France, n (%) Serbia, n (%)
    Education
    Secondary I or lower 356 (18.24) 93 (18.6) 136 (26.9) 109 (21.9) 18 (3.7)
    Secondary II 1080 (55.36) 303 (60.3) 179 (35.3) 224 (45.0) 374 (77.4)
    Tertiary 554 (28.39) 106 (21.1) 192 (37.9) 165 (33.1) 91 (18.8)
    Gender
    Female 998 (50.15) 252 (49.8) 251 (49.5) 256 (51.4) 239 (49.5)
    Male 992 (49.85) 250 (50.2) 256 (50.5) 242 (48.6) 244 (50.5)
    Prior experience
    No 1203 (60.45) 316 (62.9) 328 (64.7) 326 (65.5) 233 (48.2)
    Yes 787 (39.54) 186 (37.1) 179 (35.3) 172 (34.5) 250 (51.8)

    aEducational attainment was categorized as low (secondary level I or below) and medium or high (secondary level II or higher). In Serbia, however, representativeness was achieved by grouping educational levels into low or medium (secondary level II or below) and high (tertiary education) due to sampling limitations.

    bPrior experience: no = “I have never used Generative AI for health information seeking” and yes = “I have used Generative AI for health information seeking at least rarely.”

    Statistical tests revealed no significant differences in gender distribution across countries (χ²3=0.48; P=.92) and no significant differences in age (Kruskal-Wallis χ²3=2.15; P=.54). In contrast, educational attainment varied significantly between countries (χ²3=550.76; P<.001), reflecting sampling-related imbalances in Serbia where low versus medium or high education was assessed differently than in the other countries. Finally, prior experience with health information–seeking showed significant country differences (Kruskal-Wallis χ²3=30.95; P<.001), with higher levels reported in Serbia.

    Evaluation Outcomes

    Descriptive Analysis

    In our first research question, we explored how generative AI compares to more established health information sources in terms of usage frequency across countries. As illustrated in , generative AI ranks last among all measured sources, indicating that, as of autumn 2024, it is rarely used for health information–seeking (mean 2.08, SD 1.66). In stark contrast, online search engines like Google are highly used, ranking second with a mean usage frequency of 4.57 (SD 1.88), following conversations with physicians, which hold the top position (4.77, SD 1.70). Family and friends also play a significant role, ranking third (4.27, SD 1.73), alongside pharmacists (3.52, SD 1.81). In comparison, traditional mass media such as TV, newspapers, and magazines are used less frequently (2.74, SD 1.68), as are books (2.68, SD 1.70) and free magazines provided by pharmacies or health insurance companies (2.60, SD 1.65). The relative ranking of information sources was consistent across all 4 countries, with physicians, internet search engines, and family or friends occupying the top positions and generative AI ranking last. However, some variation in mean usage frequencies was observed between countries; detailed country-level results are presented in .

    Figure 1. Mean usage frequency of different health information sources among survey respondents (N=1990) in Austria, Denmark, France, and Serbia (95% CI; September 2024). AI: artificial intelligence.
    Model Evaluation

    For the hypothesis tests, the results are shown in . Model fit was good (comparative fit index=0.95; Tucker-Lewis-Index=.93; χ2/df =2.47, P<.001; root mean square error of approximation=0.03, 95% CI 0.03-0.03). We examined the metric measurement invariance of all latent variables by constraining all factor loadings as equal for all 4 countries. When comparing the constrained model to the unconstrained model, we found no significant difference in model fit (P=.16). Thus, metric invariance across countries was established.

    Table 2. Structural equation model predicting the intention to use generative artificial intelligence (AI) for health information among survey respondents in Austria, Denmark, France, and Serbia (N=1990; September 2024).
    Predictor variables Austria Denmark France Serbia
    b SE P value b SE P value b SE P value b SE P value
    Performance expectancy 0.47 0.05 <.001 0.52 0.05 <.001 .053 0.05 <.001 0.44 0.05 <.001
    Effort expectancy −0.07 0.05 .20 0.03 0.05 .54 −0.11 0.05 .04 -0.02 0.06 .77
    Facilitating conditions 0.12 0.04 .01 0.17 0.05 <.001 0.22 0.05 <.001 0.24 0.05 <.001
    Social influence −0.05 0.04 .29 −0.08 0.05 .17 −0.09 0.05 .10 −0.05 0.04 .27
    Habit 0.29 0.04 <.001 0.32 0.04 <.001 0.28 0.05 <.001 0.28 0.04 <.001
    Hedonic motivation 0.45 0.06 <.001 0.22 0.05 <.001 0.33 0.05 <.001 0.23 0.05 <.001
    Health literacy −0.004 0.09 .97 0.04 0.10 .67 −0.02 0.08 .08 0.08 0.10 .40
    Health status −0.002 0.03 .95 0.02 0.03 .61 −0.09 0.03 .01 −0.05 0.03 .13

    aExplained variance=0.84.

    bExplained variance=0.80.

    cExplained variance=0.86.

    dExplained variance=0.79.

    eThe different subscripts in each row indicate a significant difference between paths (P<.05)

    fThe different subscripts in each row indicate a significant difference between paths (P<.05)

    In line with H1, we found a highly significant positive association between performance expectancy and the intention to use generative AI for health information–seeking across all 4 countries (Austria: b=0.47, P<.001; Denmark: b=0.52, P<.001; France: b=0.53, P<.001; Serbia: b=0.44, P<.001). In contrast, H2 was not supported: effort expectancy showed no significant association with behavioral intention in any of the countries. Turning to H3, results revealed a positive association between facilitating conditions and the intention to use generative AI as a health information source, consistently observed across all 4 contexts (Austria: b=0.12, P=.005; Denmark: b=0.17, P<.001; France: b=0.22, P<.001; Serbia: b=0.24, P<.001). By contrast, no support was found for H4: perceived social influence was unrelated to behavioral intention in any of the countries. As predicted in H5, habit was positively associated with behavioral intention to use generative AI for health information–seeking throughout the sample (Austria: b=0.29, P<.001; Denmark: b=0.32, P<.001; France: b=0.28, P<.001; Serbia: b=0.28, P<.001). A similar pattern emerged for H6: hedonic motivation was significantly positively related to behavioral intention in all countries (Austria: b=0.45, P<.001; Denmark: b=0.22, P<.001; France: b=0.33, P<.001; Serbia: b=0.23, P<.001).

    Finally, with regard to our second research question—which examined whether health literacy and health status predict the intention to seek health information using generative AI—we found no substantial associations. Only in France did health status show a marginal negative effect (b=−0.09; P=.007).

    Principal Results

    This study investigated the predictors of intention to use generative AI for health information–seeking, drawing on the UTAUT2 framework and expanding it with health literacy and health status. Using cross-national survey data from Austria, Denmark, France, and Serbia, our findings show that generative AI is still only rarely used for health information–seeking. At the same time, performance expectancy, facilitating conditions, habit, and hedonic motivation consistently emerged as significant predictors of behavioral intention, whereas effort expectancy, social influence, health literacy, and health status were not related to intention. These patterns were consistent across all 4 countries, suggesting a robust set of psychological drivers underlying the early adoption of generative AI in health contexts. A detailed examination of these findings is provided as follows.

    First, with regard to overall usage patterns, the data shows that generative AI currently plays only a minor role in health information–seeking: 60% of the respondents reported never having used a generative AI tool for health-related questions. This result lends itself to 2 contrasting interpretations.

    On the one hand, it challenges the popular narrative that generative AI is rapidly transforming health information–seeking behavior. Instead, the findings align with previous studies, showing that generative AI is currently infrequently used in the context of health information []. Traditional sources—such as medical professionals and search engines—continue to dominate [], underscoring that generative AI has yet to achieve mainstream adoption.

    On the other hand, despite persistent concerns about data privacy, algorithmic bias, and accuracy, it is noteworthy that 40% of the respondents have already experimented with generative AI for health purposes. Given that this technology only became widely accessible relatively recently, such early uptake is remarkable. From the perspective of technology adoption models, such as the Rogers Diffusion of Innovations [], this pattern is characteristic of early adopters. It is therefore plausible to assume that the use of generative AI for health information–seeking will increase further as the technology matures and moves toward mainstream adoption.

    To better understand the drivers of future uptake, we applied an extended version of the UTAUT2 model. Our findings confirmed the predictive power of performance expectancy, facilitating conditions, habit, and hedonic motivation. This aligns with prior research on digital health tools, indicating that users value usefulness, access, familiarity, and enjoyment [,,].

    In detail, the results show that performance expectancy—the perceived usefulness of the technology—had a strong positive effect on behavioral intention in all four countries. This finding suggests that the more respondents believe generative AI is useful to manage health-related questions, the more they will use it. Thus, if public health stakeholders or developers aim to encourage responsible AI use, they should emphasize the tangible benefits of generative AI, such as 24/7 availability, rapid response times, and the potential for personalized information. Perceived usefulness may also be fostered when individuals try out generative AI for the first time, that is, they learn that they can benefit from the technology.

    At the same time, our study challenges established UTAUT2 assumptions. Effort expectancy, often seen as central to technology adoption, was not a relevant factor—possibly due to the intuitive nature of generative AI tools and the ubiquity of basic digital skills []. Using generative AI does not require any specific background knowledge beyond opening a webpage. Since online search engines are already the most frequently used health source, the basic skills needed for generative AI are widely present, potentially rendering effort expectancy less decisive.

    Taken together, this emerging pattern—the strong effect of performance expectancy and the null effect of effort expectancy—underscores the distinction between usefulness and usability, which are closely related but not identical []. Usability refers to the ease of interacting with a system (eg, ease of learning and error prevention), whereas usefulness (utility) captures whether the system provides the functions and information that users actually need. Our findings suggest that in health contexts, utility is the decisive factor: people intend to use generative AI if its outputs are perceived as useful, while usability-related aspects appear less influential.

    Importantly, this does not mean that barriers to adoption are absent. Rather, our findings show that they lie not in usability but in facilitating conditions—the structural and contextual resources that enable technology use. Across all countries, the availability of digital infrastructure, devices, and basic knowledge significantly shaped behavioral intention. In other words, while generative AI may be easy to use once accessed, unequal access to the necessary resources continues to pose a substantial adoption barrier. Consequently, facilitating conditions emerge as a key digital health equity concern []. Without adequate access, disadvantaged populations may be excluded from benefiting from generative AI, meaning that the technology risks widening rather than narrowing the digital divide in health information–seeking.

    We also found that social influence—an important predictor in other studies on AI uptake [,]—did not play a meaningful role in shaping behavioral intention. This suggests that health-related information search is a rather personal topic, and that individuals may not always be willing to disclose what kind of information they are looking for. As a result, the intention to use generative AI for health information–seeking is largely independent of peer opinions or social norms.

    In contrast, habit consistently predicted behavioral intention across all countries. From this finding, we may conclude that generative AI use for health information is likely to occur automatically, similar to how people use search engines. When individuals feel familiar with a technology, they are more likely to rely on it without conscious deliberation. However, this finding should be interpreted with caution, as the majority of participants had never used generative AI for health purposes. Much of the variance in habit may therefore reflect mere use versus nonuse. Accordingly, variables capturing initial adoption should be clearly distinguished from those measuring habit.

    By including health literacy and health status as additional predictors, our study adds a novel dimension to existing research. In contrast to studies showing direct paths between these constructs and online health information–seeking [,,], we found no such association for AI health information–seeking. However, each of these findings carries different implications. First, the absence of a significant association between health literacy and intention indicates that individuals’ ability to understand and evaluate health information was not related to whether they reported turning to generative AI. This finding may suggest that the use of such tools is driven less by informed decision-making and more by general curiosity or interest in new technologies. Importantly, this raises concerns: people with lower health literacy may be just as likely to consult generative AI as those with higher health literacy—despite being less equipped to critically assess its outputs. Given the known risk of AI hallucinations—fabricated or inaccurate information presented in a confident tone []—this could lead to misinformation and, in the worst case, harmful health decisions, as users with limited health literacy might find it difficult to distinguish between reliable and misleading content.

    Second, the lack of an association between self-reported health status and intention suggests that the current use of generative AI is not primarily driven by medical need or urgency. People do not seem more likely to consult generative AI when facing a health problem; rather, usage may occur proactively or even recreationally. This challenges assumptions that such tools are primarily used in response to a health issue, and it underscores the importance of understanding user motivations beyond immediate health concerns.

    Importantly, these patterns were largely consistent across all 4 countries, as confirmed by the measurement-invariant structural model. This cross-national consistency suggests that the psychological drivers of generative AI adoption in health contexts may transcend national boundaries and cultural differences, pointing to a universal set of adoption mechanisms.

    Limitations

    Several limitations should be acknowledged. First, due to the cross-sectional nature of this study, no causal conclusions can be drawn. Future research should therefore aim to replicate these findings using experimental or longitudinal designs. Second, we relied on self-reported data, which may be subject to social desirability bias. The use of behavioral data is thus warranted to validate these findings. Third, including additional predictors—such as individual differences or specific concerns—could provide deeper insights into the use of generative AI. Fourth, our comparative findings are based on data from only 4 countries, which limits the ability to conduct multilevel analyses. Also, as in all cross-sectional research, there is a risk of unmeasured third variables. In particular, we did not include AI trust and perceived AI risk. However, these constructs are conceptually close to performance expectancy, as trust reduces uncertainty about the system’s outputs and thereby enhances expected performance gains, whereas perceived risk erodes expected utility. In this sense, they are likely to be partially reflected in the performance expectancy construct already included in our model. That said, and highlighting that our model explains around 80% of the variance, trust and perceived risk could still suppress some of the predictors we have modeled. Thus, future research should include additional constructs outside the UTAUT2 framework []. Finally, health status was measured with a single self-rated item. While single-item measures of subjective health may not capture the full complexity of an individual’s medical condition, this approach is widely used in demographic and population health research. Prior work has demonstrated that the self-rated health item is a valid and reliable indicator, predicting key outcomes such as mortality, use of health services, and health expenditures in large-scale surveys []. Nevertheless, we acknowledge that a more fine-grained measure (eg, including specific chronic conditions or severity indices) could have provided additional insights, and future studies may benefit from applying such extended health measures.

    Conclusions

    This study applied the UTAUT2 model to investigate the factors that drive the use of generative AI for health information–seeking. Although overall usage remains limited, our findings show that performance expectancy, facilitating conditions, habit, and hedonic motivation are positively associated with behavioral intentions. These patterns, observed across all 4 countries—Austria, Denmark, France, and Serbia—suggest that current users of generative AI are likely to be early adopters: individuals who are tech-savvy, curious, and open to innovation. This aligns with the Rogers Diffusion of Innovations theory, which conceptualizes adoption as a gradual process beginning with a small, innovation-oriented segment of the population.

    The lack of significant effects for effort expectancy and social influence across all countries reinforces this interpretation: early adopters tend to base their decisions on personal evaluations rather than external opinions and are rarely deterred by usability concerns. Furthermore, the fact that behavioral intention was unrelated to health status or health literacy underscores that current usage is not driven by acute medical need or advanced health literacy, but rather by interest, convenience, and technological exploration.

    The cross-national consistency of these findings is particularly striking. Despite differences in health care systems, digital infrastructures, and culture, the same psychological and contextual factors influenced generative AI use in all countries surveyed. This suggests a shared adoption logic that transcends national boundaries—at least in the early stages of diffusion.

    Looking ahead, these insights help illuminate how generative AI might transition from a niche tool to a widely used resource. As the technology becomes more embedded in everyday life, broader segments of the population—the so-called early and late majority—will likely demand stronger assurances of trustworthiness, safety, and added value. To enable responsible and inclusive adoption, it is therefore crucial to reduce digital access barriers, enhance transparency, and implement safeguards against health misinformation, especially for users with limited health literacy.

    From a practical perspective, our findings suggest that communication strategies aiming to promote generative AI for health purposes should emphasize concrete benefits, ease of access, and even enjoyment. Rather than exclusively targeting individuals with chronic or urgent health needs, positioning generative AI as an engaging, low-barrier tool may broaden its appeal—reaching users who might otherwise be disengaged from traditional health information sources.

    In sum, generative AI holds significant potential as a future health information resource—but its trajectory will depend on how well we understand and support the evolving needs of its users across different adoption phases and contexts.

    The authors used ChatGPT (OpenAI) to support language editing of prewritten text sections (eg, to improve grammar and phrasing). All suggestions from the artificial intelligence tool were reviewed by the authors and revised or rejected as necessary. No content was generated solely by the artificial intelligence, and the authors remain fully responsible for the final text.

    The work was supported by Circle U [2024-09 – AIHEALTH].

    All data and supplementary materials related to this study are openly accessible via the Open Science Framework (OSF) at the following link [].

    None declared.

    Edited by Amaryllis Mavragani; submitted 08.Apr.2025; peer-reviewed by Armaun Rouhi, Dillon Chrimes, Sonish Sivarajkumar; accepted 27.Oct.2025; published 28.Jan.2026.

    © Jörg Matthes, Anne Reinhardt, Selma Hodzic, Jaroslava Kaňková, Alice Binder, Ljubisa Bojic, Helle Terkildsen Maindal, Corina Paraschiv, Knud Ryom. Originally published in the Journal of Medical Internet Research (https://www.jmir.org), 28.Jan.2026.

    This is an open-access article distributed under the terms of the Creative Commons Attribution License (https://creativecommons.org/licenses/by/4.0/), which permits unrestricted use, distribution, and reproduction in any medium, provided the original work, first published in the Journal of Medical Internet Research (ISSN 1438-8871), is properly cited. The complete bibliographic information, a link to the original publication on https://www.jmir.org/, as well as this copyright and license information must be included.

    Continue Reading

  • Is America about to attack Iran? – The Economist

    1. Is America about to attack Iran?  The Economist
    2. Iran unveils underwater missile tunnels, warns Hormuz ‘not safe’  The New Arab
    3. Pakistan alarmed by renewed Iran-US tensions  The Express Tribune
    4. US attack on Iran imminent?  eurotopics.net
    5. Trump is…

    Continue Reading

  • February’s PS Plus Monthly Games include Undisputed and Subnautica: Below Zero

    February’s PS Plus Monthly Games include Undisputed and Subnautica: Below Zero

    We’ve somehow almost reached the end of January already, which means it’s time for Sony to divulge the list of PlayStation Plus Monthly Games for February. They are Undisputed, Subnautica: Below Zero, Ultros and Ace Combat 7: Skies Unknown….

    Continue Reading