Klarna has claimed that AI-related savings have allowed the buy now, pay later company to increase staff salaries by nearly 60%, but hinted it could slash more jobs after nearly halving its workforce over the past three years.
Chief executive Sebastian Siemiatkowski said headcount had dropped from 5,527 to 2,907 since 2022, mostly as a result of natural attrition, with departing staff replaced by technology rather than by new staff members.
The figures add to the impact of an internal artificial intelligence programme, which had steadily reduced its use of outsourced workers including those in customer service, with technology now carrying out the work of 853 full-time staff, up from 700 earlier this year.
It meant the company, which was founded in Sweden in 2005, had managed to increase revenues by 108% while keeping operating costs flat. Siemiatkowski told analysts on an earnings call on Tuesday that it was “pretty remarkable, and unheard of as a number, among businesses”.
He explained that Klarna has not hired “for a few years”. However, some of the resulting cost-savings had been used to increase pay for remaining staff, with average compensation – including employee-related taxes and pension contributions – rising by 60% over the past three years.
“We have made a commitment to our employees that all of these efficiency gains, and especially the applications of AI, should also, to some degree, come back in their pay cheques so that they are fully … incentivised [and] aligned with the investors, to drive these changes through the company.”
Average compensation for each employee has jumped from $126,000 (£96,000) in 2022 to $203,000 today, Klarna said.
Siemiatkowski, who is a shareholder in a number of AI firms including OpenAI and Perplexity through his family investment firm Flat Capital, said he hoped to continue increasing a metric measuring revenue per employee, suggesting a further reduction in staff numbers in the years ahead.
“We’re now at $1.1m per employee, and we hope to continue to do that acceleration.”
skip past newsletter promotion
after newsletter promotion
Siemiatkowski warned this week against costly investments in datacentres to power AI, telling the Financial Times that he expected the technology would become more efficient over time.
The comments came as Klarna reported a 26% jump in revenues in the three months to the end of September to $903m, beating analysts’ expectations of $882m.
But the Swedish business reported a $95m loss over the period, significantly higher than the $4m loss last year. Klarna said this was primarily driven by changes to accounting standards that it had to follow in the US, after its decision to list its shares on the New York stock exchange in September.
The announcement underscores AI industry’s insatiable appetite for computing power as companies race to build systems that can rival or surpass human intelligence.
Published On 18 Nov 202518 Nov 2025
Share
Microsoft and Nvidia plan to invest in Anthropic under a new tie-up that includes a $30bn commitment by the Claude maker to use Microsoft’s cloud services, the latest high-profile deal binding together major players in the AI industry.
Nvidia will commit up to $10bn to Anthropic and Microsoft up to $5bn, the companies said on Tuesday, without sharing more details.
Recommended Stories
list of 4 itemsend of list
A person familiar with the matter said both the companies have committed to investing in Anthropic’s next funding round.
The announcement underscores the AI industry’s insatiable appetite for computing power as companies race to build systems that can rival or surpass human intelligence. It also ties major OpenAI-backer Microsoft, as well as key AI chip supplier Nvidia, closer to one of the ChatGPT maker’s biggest rivals.
“We’re increasingly going to be customers of each other. We will use Anthropic models, they will use our infrastructure and we’ll go to market together,” Microsoft CEO Satya Nadella said in a video. He added that OpenAI “remains a critical partner”.
The move comes weeks after OpenAI unveiled a sweeping restructuring that moved it further away from its non-profit roots, giving it greater operational and financial freedom.
The startup has since then announced a $38bn deal to buy cloud services from Amazon.com as it reduces reliance on Microsoft. Its CEO, Sam Altman, has said OpenAI is committed to spending $1.4 trillion to develop 30 gigawatts of computing resources – enough to roughly power 25 million US homes.
Still, three years after ChatGPT’s debut, investors are increasingly uneasy that the AI boom has outrun fundamentals. Some business leaders have noted that circular deals – in which one partner props up another’s revenue – add to the bubble risk.
“The main feature of the partnership is to reduce the AI economy’s reliance on OpenAI,” D A Davidson analyst Gil Luria said of Tuesday’s announcement.
“Microsoft has decided not to rely on one frontier model company. Nvidia was also somewhat dependent on OpenAI’s success and is now helping generating broader demand.
AI industry consolidating
Founded in 2021 by former OpenAI staff, Anthropic was recently valued at $183bn and has become a major rival to the ChatGPT maker, driven by the strong adoption of its services by enterprise customers.
The Reuters news agency reported last month that Anthropic was projecting to more than double and potentially nearly triple its annualised revenue run rate to around $26bn next year. It has more than 300,000 business and enterprise customers.
As part of Tuesday’s move, Anthropic will work with Nvidia on chips and models to improve performance and commit up to 1 gigawatt of compute using Nvidia’s Grace Blackwell and Vera Rubin hardware. Industry executives estimate that one gigawatt of AI computing can cost between $20bn and $25bn.
Microsoft will also give Azure AI Foundry customers access to the latest Claude models, making Claude the only frontier model offered across all three major cloud providers.
“These investments reflect how the AI industry is consolidating around a few key players,” eMarketer analyst Jacob Bourne said.
Despite the looming deal, Microsoft shares are down 3.2 percent in midday trading. Nvidia is also trading 1.9 percent lower than at the market open, and Amazon has fallen 4 percent. Tech stocks remain under pressure after a cloud services outage earlier on Tuesday. Neither OpenAI nor Anthropic is publicly traded.
Scientific advances have significantly influenced the evolution of education and training in recent decades. Emerging technologies such as technology-enhanced learning and simulation-based training have played a crucial role in improving the learning experience of practitioners and have become essential in modern education systems [].
Traditionally, surgical training has mainly focused on gaining experience through a significant number of surgeries and direct involvement, in which trainees receive less supervision from experienced surgeons as they gain competence and eventually become capable of doing surgeries independently []. This model embodies the “see one, do one, teach one” approach []. An experienced surgeon first executes a procedure, which the trainee observes. Then, under supervision, the trainee replicates the process. Finally, upon achieving competence, the trainee is expected to instruct others on how to perform it. This approach underscores the importance of direct observation, practical experience, and the ability to transmit information and expertise to future generations of medical practitioners. However, it also raises inquiries regarding the diversity of learning experiences, the consistency of the skills acquired, and the stress that it places on seasoned surgeons and trainees to quickly comprehend and transmit complex procedures involving inherent risks [,]. Acquiring and improving skills in the field of medicine are complex processes that last throughout a physician’s career. Since the 1990s, ongoing discussions have focused on enhancing teaching practices [].
Researchers have developed various simulators and training platforms to address these challenges and the demands of an expanding spectrum of surgical operations []. These tools enable trainees to develop expertise in different surgical procedures and provide the benefit of unlimited practice opportunities, customizable difficulty levels, and cost-effective solutions that emulate the difficulties of actual surgery procedures [,]. Furthermore, these platforms offer a secure and interactive setting that promotes learning through experimentation, enabling risk-free practice. Nevertheless, there remains considerable potential to improve the effectiveness of these training setups [,].
As technological advancements continue, interest in incorporating artificial intelligence (AI) into medical training has also increased []. AI, with its capacity to emulate certain aspects of human cognition, has the potential to enhance educational outcomes and transform traditional methods of training and teaching []. It enables the creation of the next generation of autonomous systems to execute tasks usually performed by individuals, representing a substantial advancement in computer science. Furthermore, AI algorithms could assist in enhancing conceptual understanding, facilitating virtual practice, and offering analytical feedback on performance. Through the use of data-driven insights and predictive analytics, AI has the potential to revolutionize surgical training, offering customized and efficient learning pathways.
This scoping review aims to map and analyze current applications of AI in surgical training, assessment, and evaluation, identifying the most common surgical procedures, AI techniques, and training setups while highlighting gaps and opportunities for future research. The following research questions guided this study:
What are the specific surgical procedures where AI algorithms are most frequently applied in surgical training?
Which AI techniques have been used in surgical training and evaluation?
How are AI techniques being used to assess and improve surgical training?
How do AI applications in surgical training affect the learning curve of surgical residents and fellows?
The paper is organized as follows: the “Methods” section outlines the methodology used to carry out this scoping review. The “Results” section provides a comprehensive overview of the findings, shows additional findings, and identifies potential areas for opportunity. The “Discussion” section presents an outline of the research questions, shows additional findings, identifies potential areas for opportunity, acknowledges the limits of the current review, and concludes with final thoughts and directions for future research in the realm of AI in surgical education.
Although there are different definitions and approaches to what AI is, this study is particularly interested in Russell and Norvig’s [] approach to systems that act rationally, that is, systems that act intelligently and rationally, ideally in the best possible way given the available information. AI is a disruptive technology that is reshaping education, facilitating a shift toward more efficient teaching protocols []. It enables machines to imitate various complex human skills, and AI-based techniques are typically employed in the following areas:
Expert systems “emulate the behavior of a human expert within a well-defined, narrow domain of knowledge” [].
Intelligent tutoring systems emulate “model learners’ psychological states to provide individualized instruction. They… help learners acquire domain-specific, cognitive, and metacognitive knowledge” [].
AI can be subdivided into machine learning (ML), which further includes deep learning (DL). ML aims to “perform intelligent predictions based on a data set” []. It uses statistical, data mining, and optimization methods to design models that can identify patterns and make predictions with higher precision than human experts. In this field, there are 3 fundamental ML paradigms:
Supervised learning uses input data and their matching labeled output to train models []. A labeled output is data that has been assigned labels to add context; consequently, the objective of supervised learning is to learn and predict outputs for unseen data based on the initial input-output pairs.
Unsupervised learning involves working with unlabeled data []. The algorithms autonomously attempt to discern patterns and relationships within the data.
Reinforcement learning uses an autonomous entity known as an agent, which learns to make decisions by performing activities inside an environment to reach a specific objective []. The feedback the agent receives in the form of rewards or penalties serves as a guide as it iteratively refines its strategy to achieve optimal performance.
Finally, DL is a branch of machine learning that uses artificial neural networks to replicate the sophisticated processes of the human brain []. Algorithms in this category learn to identify patterns and comprehend large datasets. DL is highly efficient because it can automatically extract and learn high-level characteristics from data, reducing the need for manual feature selection. It excels at handling complex tasks such as image and audio recognition, natural language processing, image generation, and data-driven prediction.
Numerous models have been developed within AI to address challenging problems and tasks across different sectors and research fields. Each approach provides certain advantages specific to the type of data to be processed and the analytical needs (see ).
Textbox 1. Approaches and advantages specific to the type of data to be processed and analytical needs.
Regression analysis forecasts a continuous output by considering one or more predictor variables [].
Cluster analysis methods group similar items based on shared characteristics. These algorithms help identify patterns within the data [].
Support vector machine (SVM) categorizes data by identifying the optimal boundary that divides distinct groups [].
Decision trees analyze data by using a series of questions and rules, resulting in the generation of predictions or classifications [].
Random forest (RF) uses a set of decision trees to enhance predictive precision and mitigate overfitting, a phenomenon in which predictions are accurate for training data but not for new data [].
Bayesian networks model the relationships and dependencies among variables using probability theory []. They are represented through a directed acyclic graph. This approach facilitates the prediction of outcomes based on established conditions.
Markov models represent the transitions between states in a system using probabilities []. They are characterized by the Markov property, where the future state depends only on the current state and not on the sequence of events that preceded it.
Fuzzy systems are based on fuzzy logic, which extends classical Boolean logic to handle the concept of partial truth, where truth values can range between completely true and completely false [].
Neural networks (NNs) are inspired by the human brain. They rely on interconnected nodes to process data and detect connections []. This model can be subdivided based on its specific use.
Convolutional neural networks (CNNs) process data that displays a grid-like structure, such as images [].
Recurrent neural networks (RNNs) predict sequences []. They use their internal state (memory) to process sequences of inputs, such as language or time series data.
Long short-term memory (LSTM) networks are a type of RNN that can learn long-term dependencies []. They are ideal for activities that require comprehension of long sequences.
Deep neural networks (DNNs) consist of multiple interconnected layers of neurons []. These networks can learn from extensive amounts of data and detect complex patterns.
Transformers are a type of network that relies on self-attention mechanisms, allowing it to weigh the importance of different parts of the input data [].
Large language models (LLMs) are advanced types of networks that have been trained on vast datasets of words and sentences []. They produce coherent, human-like responses to written text by selecting the most probable next words.
These AI models highlight the potential of this technology in educational contexts. The United Nations Educational, Scientific and Cultural Organization indicates that digital technologies have the potential to complement, enrich, and transform education, aligning with the United Nations’ Sustainable Development Goal 4 (SDG 4) for education and providing universal access to learning []. Consequently, the integration of AI in surgical training could boost independence, self-study, engagement, and motivation.
Methods
Overview
This review adheres to the PRISMA-ScR (Preferred Reporting Items for Systematic Reviews and Meta-Analyses extension for Scoping Reviews; see ) statement, designed for publications in the health and medical sciences []. The review process was organized following a structured protocol consisting of four stages: (1) planning, which involved establishing the criteria for the search and databases to be used; (2) conducting, which entailed performing the search and applying filters for the scoping review; (3) reporting, which included compiling the studies that met the criteria and were included in the review. During stages 1 and 2, the research papers were compiled, and the initial screening process was conducted, focusing solely on papers that fall within the scope of the review and were published in peer-reviewed scientific journals. Stage 3 consists of identifying the main characteristics that distinguish the contributions and unique features of each article that has passed the initial screening process. Subsequently, the necessary analysis was performed to present the summary of the research and compile tables and figures. The starting date for stages 1 and 2 of the scoping review was February 27, 2024, and it concluded on March 18, 2024.
Information Sources
A total of 3 databases were selected to search for relevant studies: PubMed, Scopus, and Web of Science. The inclusion of Web of Science and Scopus databases consolidates information from other sources, such as IEEE Xplore, ScienceDirect, and SpringerLink. Therefore, they expand the scope of accessible academic literature. These platforms also provide search and analytical tools, making it easier to find pertinent studies and analyze trends. By using the 3 databases, the review considered articles with different AI models beyond the limitation of just focusing on clinical trials. By implementing this procedure, the scope of the review is expanded, enabling the identification of significant manuscripts to identify areas of opportunity in the field.
Search Strategy
A total of 4 keywords related to AI concepts and 4 keywords related to surgical training were selected based on the research questions. The selected keywords were converted into search strings and processed to be compatible with the advanced search tool of each database. shows the search strings used in this scoping review.
Table 1. Search strings used in the advanced search tools of PubMed, Web of Science, and Scopus.
Database
String of keywords
PubMed
(“Artificial Intelligence”[MeSH] OR “AI” OR “machine learning” OR “deep learning”) AND (“Surgical Training” OR “surgical education” OR “surgical assessment” OR “surgical evaluation”)
Web of Science
TS = ((“artificial intelligence” OR “AI” OR “machine learning” OR “deep learning”) AND (“surgical training” OR “surgical education” OR “surgical assessment” OR “surgical evaluation”))
Scopus
(TITLE-ABS-KEY(“artificial intelligence” OR “AI” OR “machine learning” OR “deep learning”) AND TITLE-ABS-KEY (“surgical training” OR “surgical education” OR “surgical assessment” OR “surgical evaluation”))
Eligibility Criteria
Records retrieved from the initial search were examined to verify their compliance with the eligibility criteria and their alignment with the research questions ().
Textbox 2. Eligibility criteria.
The inclusion criteria for this review were limited to:
Studies published from January 2020 to March 2024 were reviewed to ensure the review covers the most recent advancements in artificial intelligence (AI) applications in surgical training.
Full-text articles available in English to allow thorough review and analysis.
Studies that focus on the application of AI in surgical training and evaluation, aligning with the research questions.
For the exclusion criteria, this review applied the following criteria:
Studies not centered on the application of AI to assess or evaluate surgical training.
Nonscientific journal publications, non–full-text articles available online, and preprints.
Data Charting and Synthesis
After the inclusion and exclusion criteria had been applied during screening, data were charted for each included study covering three dimensions: (1) the surgical procedure (eg, laparoscopy, minimally invasive surgery, neurosurgery, and arthroscopy), (2) the AI model (eg, support vector machine [SVM], convolutional neural network [CNN], deep neural network [DNN], long short-term memory [LSTM], and transformers), and (3) the training setup (eg, simulation platforms, box trainers, surgical video analysis, in-vivo settings, virtual reality, and da Vinci system). These variables structured the subsequent evidence synthesis and guided the organization of results by procedure, technique, and setup. In addition, bibliographic fields, including year of publication and type of publication, were also charted to support descriptive reporting in the Results section. This structured approach enabled a descriptive and narrative synthesis aimed at elucidating how AI contributes to educational outcomes and skill acquisition in surgical training.
Results
Search Results and Study Selection
presents the PRISMA-ScR flow diagram illustrating the complete selection process. The initial search identified 1400 records: 545 from PubMed, 288 from Web of Science, and 567 from Scopus, obtained using the search strings described in . After applying the publication date range from January 2020 to March 2024, a total of 461 records were excluded, leaving 939 for further screening. Duplicate removal eliminated 363 records, yielding 576 unique studies.
Figure 1. Flow diagram of the scoping review process, illustrating the inclusion and exclusion criteria. AI: artificial intelligence; LLM: large language model.
Subsequent filtering was conducted in stages to ensure methodological rigor and relevance. Database parameters were adjusted to retain only peer-reviewed journal articles and conference proceedings, excluding 260 reviews and 36 editorials that did not meet the inclusion criteria. A total of 280 records proceeded to qualitative screening. During this stage, the relevance of each article to the review objectives was reassessed. This process excluded 76 studies that, despite meeting database filters, were secondary reviews, surveys, or editorials; 9 non-English papers, 7 papers focused on nonsurgical training, and 18 papers described simulator development or validation without AI integration. Additional exclusions comprised 1 duplicate, 3 studies addressing “Data Collection Systems,” 11 centered on “LLMs in Non-Surgical Education,” and 99 that did not provide sufficient information about AI-enhanced surgical training. This filtering process excluded 224 additional studies, leaving 56 studies for the final synthesis and analysis.
The characteristics of the 56 included studies are summarized in , organized across five domains: (1) surgical procedure (eg, laparoscopy, minimally invasive surgery [MIS], neurosurgery, and arthroscopy), (2) year of publication, (3) type of publication, (4) AI technique or model used (eg, SVM, CNN, DNN, LSTM, and transformers), and (5) training setup (eg, simulation platforms, box trainers, da Vinci system, surgical video analysis, and in vivo or virtual-reality environments). This structure enables direct comparison across specialties and methodological approaches, while supporting a descriptive and narrative synthesis of cross-cutting trends.
Across the included studies, MIS, neurosurgery, and laparoscopy represented the majority of AI applications. ML and DL techniques were the most frequently used computational approaches, while simulation environments and box trainers constituted the primary training configurations. Collectively, these trends indicate a primary emphasis on risk-managed training environments that leverage accessible kinematic and video data. However, heterogeneity in studies and limited standardization of outcome measures remain persistent challenges, underscoring the need for unified evaluation frameworks in the future.
Table 2. Characteristics of included studies: surgical procedures, artificial intelligence (AI) techniques, and training setups.
Classification and references
Year
Type
AI model
Setup
MISa skills
Rashidi et al []
2023
Journal
Fuzzy systems
Box trainer
Fathabadi et al []
2022
Conference
Fuzzy systems
Box trainer
Deng et al []
2021
Conference
CNNb
Box trainer
Kulkarni et al []
2023
Journal
Clustering
Box trainer
Wu et al []
2021
Journal
MLc (unspecified)
da Vinci system
Brown and Kuchenbecker []
2023
Journal
Regression analysis
da Vinci system
Keles et al []
2021
Journal
ML (unspecified)
Box trainer
Koskinen et al []
2020
Journal
SVMd
Box trainer
Kasa et al []
2022
Journal
DLe (unspecified)
Box trainer
Gao et al []
2020
Journal
Clustering
Box trainer
Baghdadi et al []
2020
Journal
Clustering
Box trainer
Benmansour et al []
2023
Journal
CNNf+LSTMg
da Vinci system
Yanik et al []
2023
Journal
CNN
Box trainer
Lee et al []
2024
Journal
Markov chains
Simulation training
Hung et al []
2023
Journal
CNN+LSTM
Simulation training
Neurosurgery
Ledwos et al []
2022
Journal
Clustering
Simulation training
Mirchi et al []
2020
Journal
SVM
Simulation training
Yilmaz et al []
2024
Journal
AI (unspecified)
Simulation training
Siyar et al []
2020
Journal
SVM
Simulation training
Reich et al []
2022
Journal
NNh
Simulation training
Natheir et al []
2023
Journal
ML (unspecified)
Simulation training
Siyar et al []
2020
Journal
Clustering
Simulation training
Yilmaz et al []
2022
Journal
DNNi
Simulation training
Fazlollahi et al []
2022
Journal
Tutoring system (unspecified)
Simulation training
Du et al []
2023
Journal
SVM
Simulation training
Dhanakshirur et al []
2023
Conference
CNN
Training station
Laparoscopy
Kuo et al []
2022
Journal
DL (unspecified)
Box trainer
Shafiei et al []
2023
Journal
ML (unspecified)
da Vinci system
Lavanchy et al []
2021
Journal
CNN
In-vivo setting
Ryder et al []
2024
Journal
ML (unspecified)
In-vivo setting
Halperin et al []
2024
Journal
DL (unspecified)
Box trainer
Ebina et al []
2022
Journal
SVM
Box trainer
Hamilton et al []
2023
Journal
AI (unspecified)
Training station
Adrales et al []
2024
Journal
ML (unspecified)
Surgical video
Wang et al []
2023
Conference
AI (unspecified)
Surgical video
Arthroscopy
Mirchi et al []
2020
Journal
NN
Simulation training
Alkadri et al []
2021
Journal
NN
Simulation training
Shedage et al []
2021
Conference
Clustering
Simulation training
Ophthalmology
Tabuchi et al []
2022
Journal
AI (unspecified)
Surgical video
Wang et al []
2022
Journal
DNN
Surgical video
Dong et al []
2021
Journal
ML (unspecified)
Surgical video
Robotic-assisted surgery
Simmonds et al []
2021
Journal
Clustering
Simulation training
Kocielnik et al []
2023
Conference
DL (unspecified)
da Vinci system
Wang et al []
2023
Journal
Bayesian network
da Vinci system
Open surgery
Bkheet et al []
2023
Journal
DL (unspecified)
Surgical video
Kadkhodamohammadi et al []
2021
Journal
CNN
Surgical video
Surgery
Papagiannakis et al []
2020
Conference
ML (unspecified)
Simulation training
Thanawala et al []
2022
Journal
ML (unspecified)
Case logs
Surgery skills
Sung et al []
2020
Journal
CNN
Simulation training
Khan et al []
2021
Journal
ML (unspecified)
Motion data
Otolaryngology
Lamtara et al []
2020
Conference
ML (unspecified)
Simulation training
Orthopedics
Sun et al []
2021
Journal
ML (unspecified)
Surgical video
Plastic surgery
Kim et al []
2020
Conference
DL (unspecified)
Medical images
Radiology
Saricilar et al []
2023
Journal
NN
Simulation training
Urology
Kiyasseh et al []
2023
Journal
Transformer
Surgical video
Vascular surgery
Guo et al []
2020
Journal
SVM+RFj
Slave controller
aMIS: minimally invasive surgery.
bCNN: convolutional neural network.
cML: machine learning.
dSVM: support vector machine.
eDL: deep learning.
fCNN: convolutional neural network.
gLSTM: long short-term memory.
hNN: neural network.
iDNN: deep neural network.
jRF: random forest.
Findings and Interpretation
Specific Surgical Procedures
The scoping review reveals the range of surgical procedures where AI algorithms are being used (see ). The analysis emphasizes the integration of AI in MIS skills (27%, 15/56) [-], neurosurgery (20%, 11/56) [-], and laparoscopy (16%, 9/56) [-] (see ). Moderate representation was observed in arthroscopy (5%, 3/56) [-], ophthalmology (5%, 3/56) [-], and robot-assisted surgery (5%,3/56) [-]. Several other domains appeared less frequently, including open surgery (4%, 2/56) [,], general surgery (4%, 2/56) [,], and surgery skills (4%, 2/56) [,]. Finally, isolated studies were identified in otolaryngology (2%, 1/56) [], orthopedics (2%, 1/56) [], plastic surgery (2%, 1/56) [], radiology (2%, 1/56) [], urology (2%, 1/56) [], and vascular surgery (2%, 1/56) [].
Table 3. Frequency of medical fields in the included articles (N=56).
Specialty
Included articles, n (%)
MISa skills
15 (27)
Neurosurgery
11 (20)
Laparoscopy
9 (16)
Arthroscopy
3 (5)
Ophthalmology
3 (5)
Robot-assisted surgery
3 (5)
Open surgery
2 (4)
Surgery
2 (4)
Surgery skills
2 (4)
Otolaryngology
1 (2)
Orthopedy
1 (2)
Plastic surgery
1 (2)
Radiology
1 (2)
Urology
1 (2)
Vascular surgery
1 (2)
aMIS: minimally invasive surgery.
Functionally, most studies focused on automated skill assessment and learning-curve analysis, while comparatively few examined procedure guidance, workflow recognition, or decision support. This trend was especially evident in MIS and laparoscopy, which relied heavily on video-centric datasets and computer-vision models [-,-], and in neurosurgery, where virtual reality simulators provided standardized training environments and feedback mechanisms [-]. The specialty distribution appears to be driven by the availability of high-quality labeled data. Overall, the distribution of specialties indicates that AI integration aligns strongly with domains that generate structured, labeled, and reproducible data, such as endoscopic or robotic procedures. By contrast, open and specialty surgeries remain underrepresented, constrained by the limited standardization of datasets and variability in operative workflows. Future progress will depend on developing shared, procedure-specific repositories, cross-institutional benchmarks, and multimodal data capture beyond video and kinematic streams to enhance generalizability and educational impact [-].
AI Techniques Used
The scoping review identified a diverse set of AI techniques in surgical training (see ). The most frequent were ML (unspecified; 21%, 12/56) [,,,,,,,,,-], clustering (13%, 7/56) [,,,,,,], and CNNs (11%, 6/56) [,,,,,]. We also observed DL (unspecified; 11%, 6/56) [,,,,,] and SVMs (9%, 5/56) [,,,,], followed by neural networks (NNs; 7%, 4/56) [,,,] and AI (unspecified; 7%; 4/56) [,,,]. Additional categories included CNN+LSTM (4%, 2/56) [,], DNNs (4%, 2/56) [,], and fuzzy systems (4%, 2/56) [,]. Single-study categories (2%, 1/56) included regression analysis [], Markov chains [], tutoring system (unspecified) [], Bayesian network [], transformer [], and SVM+RF [].
Table 4. Application of artificial intelligence (AI) techniques in the included articles (N=56).
AI technique
Included articles, n (%)
MLa (unspecified)
12 (21)
Clustering
7 (13)
CNNsb
6 (11)
DLc (unspecified)
6 (11)
SVMsd
5 (9)
NNse
4 (7)
AI (unspecified)
4 (7)
CNN+LSTMf
2 (4)
DNNsg
2 (4)
Fuzzy systems
2 (4)
Regression analysis
1 (2)
Markov chains
1 (2)
Tutoring system (unspecified)
1 (2)
Bayesian network
1 (2)
SVM+RFh
1 (2)
Transformer
1 (2)
aML: machine learning.
bCNN: convolutional neural network.
cDL: deep learning.
dSVM: support vector machine.
eNN: neural network.
fLSTM: long short-term memory.
gDNN: deep neural network.
hRF: random forest.
From 2020 to 2024 (see ), ML (unspecified) appears every year, CNNs strengthen in 2021 and 2023, and DL (unspecified) is present in 2020 and 2022-2024. Sequential and hybrid models (CNN+LSTM and DNNs) clusters in 2022-2023. AI (unspecified) emerges from 2022 onward. Probabilistic and rule-based approaches (Bayesian networks, fuzzy systems, and Markov chains) and transformer/SVM+RF appear as single-study categories. Overall, the technique mix tracks data modality and availability (video and kinematics), reinforcing the need for shared multimodal repositories and standardized evaluation metrics to compare methods fairly and improve external validity.
Table 5. Temporal distribution of artificial intelligence (AI) models in the included articles (2020-2024).
AI model
2020, n (%)
2021, n (%)
2022, n (%)
2023, n (%)
2024, n (%)
Total, n (%)
MLa (unspecified)
2 (17)
5 (42)
1 (8)
2 (17)
2 (17)
12 (100)
CNNb
1 (17)
3 (50)
0 (0)
2 (33)
0 (0)
6 (100)
Clustering
3 (43)
2 (28)
1 (14)
1 (14)
0 (0)
7 (100)
SVMc
3 (60)
0 (0)
1 (20)
1 (20)
0 (0)
5 (100)
DLd (unspecified)
1 (17)
0 (0)
2 (33)
2 (33)
1 (17)
6 (100)
NNe
1 (25)
1 (25)
1 (25)
1 (25)
0 (0)
4 (100)
AI (unspecified)
0 (0)
0 (0)
1 (25)
2 (50)
1 (25)
4 (100)
DNNf
0 (0)
0 (0)
2 (100)
0 (0)
0 (0)
2 (100)
CNN+LSTMg
0 (0)
0 (0)
0 (0)
2 (100)
0 (0)
2 (100)
Fuzzy systems
0 (0)
0 (0)
1 (50)
1 (50)
0 (0)
2 (100)
Bayesian network
0 (0)
0 (0)
0 (0)
1 (100)
0 (0)
1 (100)
Markov chains
0 (0)
0 (0)
0 (0)
0 (0)
1 (100)
1 (100)
Regression analysis
0 (0)
0 (0)
0 (0)
1 (100)
0 (0)
1 (100)
SVM+RFh
1 (100)
0 (0)
0 (0)
0 (0)
0 (0)
1 (100)
Transformer
0 (0)
0 (0)
0 (0)
1 (100)
0 (0)
1 (100)
Tutoring system (unspecified)
0 (0)
0 (0)
1 (100)
0 (0)
0 (0)
1 (100)
Total per year
12 (21)
11 (20)
11(20)
17 (30)
5 (9)
56 (100)
aML: machine learning.
bCNN: convolutional neural network.
cSVM: support vector machine.
dDL: deep learning.
eNN: neural network.
fDNN: deep neural network.
gLSTM: long short-term memory.
hRF: random forest.
In the analyzed studies, the number of publications increased from 12 in 2020 to 17 in 2023, with 11 in both 2021 and 2022, and 5 in 2024. The literature search concluded on March 18, 2024, which likely accounts for the lower count in 2024. These totals are summarized in the “Total per year” row of .
Application of AI Techniques
AI techniques have been applied across diverse training setups, enhancing both learning experiences and performance assessment in surgical procedures (see ). The most frequent environments were simulation training (36%, 20/56) [-,-,,,,,] and box trainers (23%, 13/56) [40–43,46-50,52,66,70-71], followed by surgical video analysis (16%, 9/56) [,,-,,,,] and robotic systems using the da Vinci platform (11%, 6/56) [,,,,,]. Less frequent configurations included training stations (4%, 2/56) [,] and in-vivo settings (4%, 2/56) [,], with single-study setups for case logs [], motion data [], medical images [], and a slave controller [] (each 2%, 1/56). Across these settings, studies reported the use of automated skill assessment, formative feedback, and adaptive progression, supported by video, kinematic, and performance-metric streams.
Over time, setup diversity increased, peaking in 2023 (see ). Simulation training and box trainers were consistently present, while surgical video and da Vinci deployments clustered in 2021-2023. These patterns mirror data availability and standardization in risk-managed environments, where AI can be trained and evaluated reliably.
Table 6. Distribution of training setups in the included articles (N=56).
Training setup
Included articles, n (%)
Simulation training
20 (36)
Box trainer
13 (23)
Surgical video
9 (16)
da Vinci System
6 (11)
Training station
2 (4)
In-vivo setting
2 (4)
Case logs
1 (2)
Motion data
1 (2)
Medical images
1 (2)
Slave controller
1 (2)
Figure 2. Appearance of setups over the years in the included articles.
Discussion
Principal Findings
This section discusses the study’s implications and contributions to the field. The review maps and analyzes current applications of AI in surgical training, assessment, and evaluation, identifying the most common surgical procedures, AI techniques, training setups, and highlighting gaps and opportunities for future research. The results show that AI is most frequently reported in data-rich, risk-mitigated environments, notably simulation training and box-trainer setups, and that ML (unspecified) and DL (unspecified) approaches dominate model choices.
Within these settings, many studies report models that leverage synchronized inputs, for example, kinematics, video, and other performance metrics, to classify technical skill using consistent criteria, to characterize learning trajectories across repeated attempts, and to localize performance-limiting behaviors at the level of gestures, steps, or procedural phases. When embedded in iterative practice, these capabilities may enable individualized training pathways that adjust task parameters and feedback density to a trainee’s evolving competence, with the potential to shorten time to proficiency and to reduce instructor workload. These implications are consistent with the results, in which simulation training accounted for 36% (20/56) and box trainer setups for 23% (13/56) of the included studies.
Findings in Relation to the Research Questions
Regarding the first research question aimed at identifying the specific surgical procedures where AI algorithms are most frequently applied in surgical training, AI use concentrates on MIS skills [-], neurosurgery [-], and laparoscopy [-]. Rather than simple frequency, the common thread across these areas is structured, high-signal data capture and well-specified tasks. Endoscopic and robotic workflows generate synchronized video, robotic kinematics, and simulator logs, which enable reproducible labels such as phase boundaries, gesture events, and Objective Structured Assessment of Technical Skills–aligned rubrics. This ecosystem lowers barriers to annotation and validation, thereby accelerating method development. Beyond these clusters, activity in ophthalmology [-], open surgery [,], robot-assisted surgery [-], and single-study specialties including radiology [], urology [], and vascular surgery [] signals a widening scope. However, these domains often face less standardized capture or a more variable field-of-view, which complicates model training and external validation. The overall distribution, therefore, appears to reflect data tractability and curricular formalization more than inherent differences in educational need.
The second research question investigated which AI techniques have been used in surgical training and evaluation. Studies use ML (unspecified) [,,,,,,,,,-] and DL (unspecified) [,,,,,] as broad families, with task-appropriate specializations such as CNNs for video [,,,,,] and SVMs for lower-dimensional kinematics or hand-crafted features [,,,,]. NNs [,,,] support competency modeling when feature engineering is feasible, and CNN+LSTM hybrids [,] target temporal dynamics for suturing and task segmentation. DNNs are explicitly mentioned in [,]. Single-study categories (fuzzy systems [,], regression analysis [], Markov chains [], tutoring system (unspecified) [], Bayesian network [], transformers [], and SVM+RF []) illustrate exploratory breadth rather than established consensus. Consistent with coding, CNN+LSTM is treated as a distinct class and not double-counted under CNNs. No single approach emerges as universally optimal; instead, methods align with task structure (classification vs sequence prediction), signal characteristics (video and kinematics), and assessment granularity (summative scores versus frame- or gesture-level feedback).
The third research question investigated how AI techniques are being used to assess and improve surgical training. Across setups, a common pattern is the move from retrospective, manual scoring to prospective, automated analytics that are both standardized and timely. In simulation training, synchronized streams enable immediate feedback and progression gating, which supports deliberate practice cycles grounded in objective metrics. This aligns with the preponderance of simulation studies in the dataset and the consistent application of ML and DL to transform kinematics and video into competency-linked outputs. In box trainers, models quantify motion economy, tool path quality, and task efficiency, enabling skill stratification and targeted coaching [-,-,,,,]. In robotic systems on the da Vinci platform, studies demonstrate automated assessment, uncertainty-aware feedback, and domain adaptation for cross-site or cross-task transfer [,,,,,]. In surgical video pipelines, investigators focus on procedural understanding, ergonomics, and fine-grained performance analytics [,,-,,,,]. The unifying mechanism across these contexts is measurement at scale that reduces feedback latency, increases consistency, and enables adaptive progression rules without displacing instructor oversight.
Finally, the last research question investigated the way in which AI applications in surgical training affect the learning curve of surgical residents and fellows. Multiple studies report outcomes consistent with accelerated learning and improved technical performance under AI-enabled training. This includes predictive modeling of progression [], metric selection and learning-curve characterization in simulation [], a randomized comparison of feedback modalities [], competency-based training backed by neural models [], continuous monitoring of bimanual expertise with deep models [], and competency estimation in laparoscopic training []. Evidence from robotic contexts shows that automated assessment can structure practice with short feedback loops []. That said, effect sizes remain difficult to aggregate due to heterogeneous study designs, small sample sizes, nonstandard outcome measures, and limited external validation. The most defensible interpretation is that personalized, data-driven feedback and objective, repeated measurement are plausible mechanisms for the observed gains, with further multicenter validation needed to establish generalizability and durability.
The findings suggest that current AI deployment in surgical training follows data availability and standardization, that ML/DL with video and kinematics are dominant because they best match that data, and that automated, timely feedback is the primary lever through which AI influences performance and learning. Where capture is less standardized or external validation is sparse, adoption tends to lag. This synthesis directly motivates the recommendations presented later in the Discussion section on common benchmarks, transparent reporting, and SDG 4–aligned scalability.
Comparison With Previous Work
Systematic literature reviews in surgical training found in the literature have focused on specific training methods (eg, simulation-based training) or on specific types of surgery (eg, plastic surgery and orthopedic surgery) rather than providing a cross-specialty map of AI methods for training, assessment, and evaluation. Reviews focused on simulation-based training within specific domains underscore this pattern. Lawaetz et al [] examined simulation-based training and assessment in open vascular surgery, cataloguing common methods and commenting on effectiveness within that context. Abelleyra Lastoria et al [] surveyed simulation-based tools in plastic surgery and concluded that the validity of many approaches requires further investigation. Woodward et al [] reached a similar conclusion in orthopedic surgery, noting concerns about the construct validity and methodological rigor of simulation studies. Reviews centered on robotic-assisted surgery also reflect divergent emphases: Rahimi et al [] provided a descriptive overview of training modalities and assessment practices, whereas Boal et al [] explicitly scrutinized AI methods for technical skills in robotic surgery and highlighted that both manual and automated assessment tools are often insufficiently validated.
Closer to the scope of the present scoping review, several analyses have examined automation and AI across surgical training tasks. Levin et al [] identified families of automated technical skill assessment methods, including computer vision, motion tracking, ML and DL, and performance classification, but did not synthesize evidence on educational effectiveness. Lam et al [] focused specifically on ML methods and reported accuracy rates that generally exceeded 80 percent across included studies, offering a performance-oriented view rather than a training-context analysis. Pedrett et al [] emphasized the central role of video-derived motion and robotic kinematic data as inputs to AI models for technical skill assessment in minimally invasive surgery, reinforcing the importance of structured, high-signal data streams.
Findings from the present review are consistent with these previous observations in several respects. First, the centrality of simulation and other risk-managed environments recurs across literature, reflecting where ground truth is tractable and measurement can be standardized. Second, many reviews identify validation gaps, noting that reported metrics, dataset partitions, and labeling practices vary widely, which complicates comparison across sites and inhibits external generalizability [-]. Third, there is broad agreement that AI-assisted assessment is advancing rapidly in robotic and minimally invasive settings; yet, many frameworks remain descriptive or single-center, and their educational impact is not consistently established with robust designs [-].
At the same time, this review differs from earlier work in several ways. The scope extends across specialties and across training setups, linking procedures, techniques, and use cases in a single comparative framework. Rather than isolating a single algorithm family or specialty, the analysis connects the dominant AI techniques to the data modalities they exploit and to the assessment functions they serve. This mapping clarifies why ML and DL approaches, particularly CNN-based and hybrid temporal models, are prevalent where high-quality video and kinematics are available, and why adoption is slower where capture is less standardized. In addition, the review integrates signals relevant to learning curves, highlighting studies that associate AI-enabled feedback with improvements in proficiency trajectories, while also acknowledging heterogeneity and the need for external validation. By taking this comparative perspective, the review identifies shared deficiencies that cut across specialties, including nonstandard outcome measures, limited transparency in algorithmic reporting, and sparse multicenter testing, and points toward future work on benchmarks, interoperable data schemas, and scalable deployment aligned with SDG 4.
Whereas previous reviews have been primarily domain-specific or method-specific, this scoping review offers a cross-specialty synthesis that links where AI is used, which techniques are used, and how they are used to support training and assessment. This perspective complements existing literature by emphasizing comparability across contexts, illuminating mechanisms by which AI influences learning, and articulating the methodological steps needed to translate promising prototypes into reproducible, generalizable, and educationally meaningful tools.
Strengths and Limitations
This scoping review offers a broad, cross-specialty perspective on the application of AI in surgical training, assessment, and evaluation. It maps procedures, techniques, and training setups within a single comparative framework, which supports interpretation across contexts rather than within a single specialty. The review adheres to PRISMA-ScR guidance, applies explicit inclusion and exclusion criteria, and uses transparent counting rules that assign each study a primary AI technique and a primary setup to avoid double-counting. Results are presented as both narrative synthesis and structured summaries. The Discussion integrates an SDG 4 perspective, offering concrete implementation considerations related to access, scalability, and equity. Together, these elements provide a panoramic view of where AI is currently deployed, why certain methods dominate in specific data environments, and how these choices influence assessment and feedback in practice.
Several constraints should be considered. First, the search was limited to English-language publications and to the period ending March 18, 2024, which may omit relevant work outside this window. Second, many articles describe methods only at a general label level (AI, ML, and DL) without specifying architectures or training details, which limits interpretability and reproducibility. Third, the evidence base is concentrated in simulation, box-trainer, and video-centric settings, which may not fully capture transfer to live clinical performance, patient outcomes, or longer-term retention. Fourth, external validation is limited, as relatively few studies report multicenter testing, performance under domain shift, subgroup analyses, or calibration, which constrains confidence in portability.
To address these limitations, educational outcomes should also be mapped to recognized competency frameworks and reported with standardized metrics that enable replication and meta-synthesis. When multisetup or multi-technique pipelines are used, authors should specify proportional attribution. Reporting on access, resource requirements, and cost per trainee hour will support the deployment and equity assessment of SDG 4. Multicenter collaborations that release shared benchmarks and interoperable datasets will be necessary to improve reproducibility and to allow fair comparisons across techniques and settings.
Future Work Recommendations
This scoping review identified current applications of AI in surgical education and highlighted priority areas for further work. As summarized in and visualized in , a large proportion of studies focus on simulation training [-,-,,,,,], representing 36% (20/56) of the included articles. This concentration reflects the suitability of simulation for controlled data capture and iterative practice. Building on this foundation, AI can enhance simulation-based training with realistic, adaptive, and personalized learning experiences [,], while also enabling standardized and rapid feedback that supports deliberate practice.
Advances in computer vision are particularly significant where high-quality video and kinematic data are accessible, which aligns with the prevalence of simulation and box-trainer studies in the included literature. In these regulated, risk-mitigated environments, AI systems can produce timely and structured feedback linked to defined competency frameworks, including economy of motion, bimanual coordination, camera control, tissue handling, and ergonomics, thereby facilitating deliberate practice. Although natural language processing technologies are less represented in the current review, their growing maturity suggests near-term opportunities to integrate narrative guidance, rubric-based feedback, and reflective prompts alongside quantitative metrics, provided such outputs are aligned with curricular objectives and are appropriately validated.
Future efforts should pursue 5 complementary directions.
First, strengthen external validity. Studies should include multi-institution cohorts, predefined external test sets, and reporting of performance under domain shift, including different camera views, instruments, and case difficulty. Where feasible, researchers should evaluate the transfer from simulation or bench-top tasks to higher-fidelity or clinical settings with clearly specified outcome measures and follow-up intervals.
Second, standardize educational outcomes. Investigators should map AI outputs to recognized competency frameworks and report validity, reliability, learning curve parameters, and time to competency with consistent definitions. Agreement on core outcome sets will enable comparison across techniques and facilitate meta-synthesis.
Third, expand the breadth and transparency of data. New work should prioritize multimodal capture that combines video, kinematics, tool telemetry, where appropriate, eye tracking or physiological signals. Public or data-sharing consortia should release interoperable schemas, labeling protocols, and benchmark tasks that are specific to procedures and skill elements. Clear descriptions of models and training and validation splits will improve reproducibility.
Fourth, improve usability, equity, and scalability in alignment with SDG 4. Models should operate on standard hardware, interoperate with existing simulators and video platforms, and function reliably in low-bandwidth or offline environments. Reporting of access, installation steps, resource needs, and cost per trainee hour will support adoption in diverse settings. Interfaces should disclose uncertainty, make feedback interpretable, and integrate into educator workflows without adding undue burden.
Fifth, broaden methodological scope responsibly. There is an opportunity to study natural language technologies for rubric-based guidance, structured debriefs, and reflective prompts, provided outputs are aligned with curricular objectives and validated for educational use. Prospective trials that compare feedback modalities and density, and that measure downstream retention and transfer, will clarify how AI should be integrated pedagogically.
Together, these directions could move the field from promising prototypes toward reproducible, generalizable, and educationally meaningful tools that improve surgeon training while supporting equitable access to high-quality education.
Conclusions
This scoping review maps current applications of AI in surgical training, assessment, and evaluation across procedures, techniques, and training setups. From 1400 records, 56 studies met the inclusion criteria, with activity concentrated in minimally invasive surgery, neurosurgery, and laparoscopy. AI is most frequently deployed in data-rich, risk-mitigated environments, particularly simulation training and box trainers, where synchronized video and kinematic streams support objective measurement and timely feedback. Technique choices reflect these data conditions, with ML (unspecified) and DL (unspecified) methods predominating and task-specific variants, such as CNNs and hybrid temporal models, applied to video-centric problems.
Across settings, studies describe automated skill assessment, structured formative feedback, and adaptive progression, with several reporting improvements consistent with accelerated learning curves. At the same time, heterogeneity in study design, small samples, nonstandard outcome measures, and limited external validation constrain strong inferences about effect sizes and generalizability. The evidence, therefore, supports cautious optimism that AI-enabled feedback can enhance skill acquisition, while underscoring the need for more rigorous evaluation.
Future work should prioritize precise reporting of models and datasets, multicenter validation, and standardized educational outcomes linked to recognized competency frameworks. Interoperable data schemes, shared benchmarks, and transparent methods will be essential to enable comparison across sites and techniques. Attention to scalability, access, and usability will support alignment with SDG 4, ensuring that benefits extend beyond well-resourced centers. With these elements in place, AI has the potential to deliver reproducible, equitable, and educationally meaningful gains in surgical training.
We thank the Engineering Faculty, the Research Group NexEd Hub, and the Computing Department of Universidad Panamericana, Mexico City Campus. Finally, we would like to thank Rodrigo González Serna and Monserrat Villacampa Espinosa de los Monteros for their assistance during the design and creation of the flow diagram and the graphs, respectively. Generative AI was used to improve the grammar, style, and clarity of some sentences and paragraphs after initial human drafting. The authors verified all output for factual accuracy and scientific integrity. The model was not used to generate paragraphs, summaries, display charts or tables, or to analyze or interpret data. The model used was ChatGPT based on GPT-4-turbo (“omni”), the vendor is OpenAI, over the web app (chat.openai.com). There were no external funding sources for this study. Consequently, funders had no influence on the design of the study, the collection, analysis, or interpretation of data, the writing of the manuscript, or the decision to publish the results.
We would also like to thank the Academy of Medical Sciences (AMS) for their support (NIF0041018), as this study originated from this award.
The datasets generated or analyzed during this study are available in the AI Review – Selected Zotero group library [].
None declared.
Edited by T Leung, G Eysenbach; submitted 29.Mar.2024; peer-reviewed by R Yin, M Pojskic; comments to author 13.Jul.2024; revised version received 20.Oct.2025; accepted 23.Oct.2025; published 18.Nov.2025.
This is an open-access article distributed under the terms of the Creative Commons Attribution License (https://creativecommons.org/licenses/by/4.0/), which permits unrestricted use, distribution, and reproduction in any medium, provided the original work, first published in the Journal of Medical Internet Research (ISSN 1438-8871), is properly cited. The complete bibliographic information, a link to the original publication on https://www.jmir.org/, as well as this copyright and license information must be included.
Sue Harley was diagnosed with Myeloma in 2017 and after running out of treatment options took Talquetamab
A Birmingham grandmother said she can live her life as “if I haven’t got cancer”, after the breakthrough drug she is taking became available on the NHS.
Sue Harley was diagnosed with Myeloma, an incurable blood cancer, in 2017 and after running out of treatment options took Talquetamab two years ago, direct from the drug manufacturer.
She has since been in remission and after a long campaign, commissioners have now approved the drug’s general use, with up to 800 patients potentially benefitting each year.
“For now, I feel like I can live my life like a normal person. This is my sixth line of treatment and there’s that hope that this is the thing that’s going to work for me forever,” she said.
Myeloma is the third most common type of blood cancer, but can be difficult to detect as symptoms are often linked to general ageing or minor conditions, charity Myeloma UK said.
When Ms Harley was diagnosed she had debilitating back pain and could barely walk.
She tried several treatments, but nothing kept her in remission for long until she was offered the new drug via a compassionate use scheme, set up for patients who have no other suitable options so they can access treatments which are not available on the NHS.
Sue Harley
For the last two years Ms Harley has run the Birmingham Half Marathon raising money for Myeloma UK
“It’s made it that I can just my life almost as if I haven’t got cancer,” she said.
“So I would want to do the things that I would want to do and I can plan to do things as well and that makes a difference.”
She was part of a campaign for the drug to be approved by The National Institute for Health and Care Excellence (NICE) and said she was now so pleased it had been made available from this week.
“I wasn’t really aware before that NICE listened to patients in that way,” she added.
“The fact that I was at the committee meetings and they did actually ask my opinion and how I felt this would impact patients, that was really important.
“For patients, what we really want is to have choices and options and some hope for the future.”
Guy Pratt, consultant haematologist at University’s Hospital’s Birmingham, said the drug was “another example of the advances we’re making in treating blood cancers”.
“It’s allowed her to live a normal life at home with a really good quality of life and I’m really grateful that we now have access to it for all our patients with multiple myeloma,” he said.
SAN FRANCISCO (AP) — Meta has prevailed over an existential challenge to its business that could have forced the tech giant to spin off Instagram and WhatsApp after a judge ruled that the company does not hold a monopoly in social networking.
U.S. District Judge James Boasberg issued his ruling Tuesday after the historic antitrust trial wrapped up in late May. His decision follows two separate rulings that branded Google an illegal monopoly in both search and online advertising, dealing yet another regulatory blow to the tech industry that for years enjoyed nearly unbridled growth.
READ MORE: Mark Zuckerberg takes the stand in historic antitrust trial that could force breakup of Meta
The FTC “continues to insist that Meta competes with the same old rivals it has for the last decade, that the company holds a monopoly among that small set, and that it maintained that monopoly through anticompetitive acquisitions,” Boasberg wrote in his ruling. “Whether or not Meta enjoyed monopoly power in the past, though, the agency must show that it continues to hold such power now. The Court’s verdict today determines that the FTC has not done so.”
A free press is a cornerstone of a healthy democracy.
Samsung Recognized for Commitment to Employee Growth, Well-being, and Innovation
Samsung Electronics Co., Ltd. is proud to announce it’s been recognized as Canada’s Top 100 Employers for 2026, by Mediacorp Canada Inc. This recognition underscores Samsung’s dedication to fostering a forward-thinking workplace that prioritizes the growth, well-being, and success of its employees.
“We are incredibly proud of this recognition, which is a true testament to the dedication and hard work of our team,” said Brian Shin, Samsung Electronics Canada Inc., President & CEO. “We are committed to building a resilient and forward-looking organization that empowers our employees to thrive both professionally and personally.”
Samsung has strengthened it focus on talent investments, ensuring employees receive the support and skills necessary to drive innovation and success in the years ahead. From comprehensive health and wellness programs to training and development opportunities, Samsung continues to strive towards high engagement and satisfaction.
Pecans, America’s only native major nut, have a storied history in the United States. Today, American trees produce hundreds of million of pounds of pecans – 80% of the world’s pecan crop. Most of that crop stays here. Pecans are used to produce pecan milk, butter and oil, but many of the nuts end up in pecan pies.
Throughout history, pecans have been overlooked, poached, cultivated and improved. As they have spread throughout the United States, they have been eaten raw and in recipes. Pecans have grown more popular over the decades, and you will probably encounter them in some form this holiday season.
I’m an extension specialist in Oklahoma, a state consistently ranked fifth in pecan production, behind Georgia, New Mexico, Arizona and Texas. I’ll admit that I am not a fan of the taste of pecans, which leaves more for the squirrels, crows and enthusiastic pecan lovers.
The spread of pecans
The pecan is a nut related to the hickory. Actually, though we call them nuts, pecans are actually a type of fruit called a drupe. Drupes have pits, like the peach and cherry.
Three pecan fruits, which ripen and split open to release pecan nuts, clustered on a pecan tree. IAISI/Moment via Getty Images
The pecan nuts that look like little brown footballs are actually the seed that starts inside the pecan fruit – until the fruit ripens and splits open to release the pecan. They are usually the size of your thumb, and you may need a nutcracker to open them. You can eat them raw or as part of a cooked dish.
The pecan derives its name from the Algonquin “pakani,” which means “a nut too hard to crack by hand.” Rich in fat and easy to transport, pecans traveled with Native Americans throughout what is now the southern United States. They were used for food, medicine and trade as early as 8,000 years ago.
Pecans are native to the southern United States. Elbert L. Little Jr. of the U.S. Department of Agriculture, Forest Service
Pecans are native to the southern United States, and while they had previously spread along travel and trade routes, the first documented purposeful planting of a pecan tree was in New York in 1722. Three years later, George Washington’s estate, Mount Vernon, had some planted pecans. Washington loved pecans, and Revolutionary War soldiers said he was constantly eating them.
Meanwhile, no one needed to plant pecans in the South, since they naturally grew along riverbanks and in groves. Pecan trees are alternate bearing: They will have a very large crop one year, followed by one or two very small crops. But because they naturally produced a harvest with no input from farmers, people did not need to actively cultivate them. Locals would harvest nuts for themselves but otherwise ignored the self-sufficient trees.
It wasn’t until the late 1800s that people in the pecan’s native range realized the pecan’s potential worth for income and trade. Harvesting pecans became competitive, and young boys would climb onto precarious tree branches. One girl was lifted by a hot air balloon so she could beat on the upper branches of trees and let them fall to collectors below. Pecan poaching was a problem in natural groves on private property.
Pecan cultivation begins
Even with so obvious a demand, cultivated orchards in the South were still rare into the 1900s. Pecan trees don’t produce nuts for several years after planting, so their future quality is unknown.
An orchard of pecan trees. Jon Frederick/iStock via Getty Images
To guarantee quality nuts, farmers began using a technique called grafting; they’d join branches from quality trees to another pecan tree’s trunk. The first attempt at grafting pecans was in 1822, but the attempts weren’t very successful.
Grafting pecans became popular after an enslaved man named Antoine who lived on a Louisiana plantation successfully produced large pecans with tender shells by grafting, around 1846. His pecans became the first widely available improved pecan variety.
Grafting is a technique that involves connecting the branch of one tree to the trunk of another. Orest Lyzhechka/iStock via Getty Images
The variety was named Centennial because it was introduced to the public 30 years later at the Philadelphia Centennial Expedition in 1876, alongside the telephone, Heinz ketchup and the right arm of the Statue of Liberty.
This technique also sped up the production process. To keep pecan quality up and produce consistent annual harvests, today’s pecan growers shake the trees while the nuts are still growing, until about half of the pecans fall off. This reduces the number of nuts so that the tree can put more energy into fewer pecans, which leads to better quality. Shaking also evens out the yield, so that the alternate-bearing characteristic doesn’t create a boom-bust cycle.
US pecan consumption
The French brought praline dessert with them when they immigrated to Louisiana in the early 1700s. A praline is a flat, creamy candy made with nuts, sugar, butter and cream. Their original recipe used almonds, but at the time, the only nut available in America was the pecan, so pecan pralines were born.
Pralines were originally a French dessert, but Americans began making them with pecans. Jupiterimages/The Image Bank via Getty Images
During the Civil War and world wars, Americans consumed pecans in large quantities because they were a protein-packed alternative when meat was expensive and scarce. One ounce of pecans has the same amount of protein as 2 ounces of meat.
After the wars, pecan demand declined, resulting in millions of excess pounds at harvest. One effort to increase demand was a national pecan recipe contest in 1924. Over 21,000 submissions came from over 5,000 cooks, with 800 of them published in a book.
Pecan consumption went up with the inclusion of pecans in commercially prepared foods and the start of the mail-order industry in the 1870s, as pecans can be shipped and stored at room temperature. That characteristic also put them on some Apollo missions. Small amounts of pecans contain many vitamins and minerals. They became commonplace in cereals, which touted their health benefits.
In 1938, the federal government published the pamphlet Nuts and How to Use Them, which touted pecans’ nutritional value and came with recipes. Food writers suggested using pecans as shortening because they are composed mostly of fat.
The government even put a price ceiling on pecans to encourage consumption, but consumers weren’t buying them. The government ended up buying the surplus pecans and integrating them into the National School Lunch Program.
Today, pecan producers use machines called tree shakers to shake pecans out of the trees. Christine_Kohler/iStock via Getty Images
While you are sitting around the Thanksgiving table this year, you can discuss one of the biggest controversies in the pecan industry: Are they PEE-cans or puh-KAHNS?
Housing bosses at a Surrey council have said they need another £107,000 and more staff to fix deep-rooted problems in the service.
In September, the Regulator of Social Housing (RSH) gave Tandridge District Council a rating of C4 – meaning there were very serious failings and potential for government intervention.
Of the extra money, £87,000 would be spent on salaries for extra staff to help the department and £20,000 on “service costs”, according to the Local Democracy Reporting Service.
Head of housing at the authority, Jane Rochelle, said: “We’re working at a tremendous pace and I’m putting my whole team under pressure.”
She added: “I don’t intend to take my foot off the gas this side of Christmas at least.”
The council had already carved out £420,000 from the housing revenue accounts operating surplus to kickstart the housing improvement plan before the inspection results came in.
The scale of the work under way was outlined at the council’s housing committee on 11 November.
Housing officers are trying to catch up with national standards right across the service, from rewriting policies to overhauling IT systems and carrying out thousands of overdue tenancy audits.
Housing leaders have said they are focusing on what the RSH calls the “big six” safety areas – things like gas, electric, fire, asbestos and water safety.
One of the main issues is a backlog of about 2,000 tenancy audits, which are basic checks that confirm who lives in each property, identify vulnerabilities and pick up risks like fuel poverty or damp.
Savills is currently inspecting every council home, according to a council report, and said of the 710 properties it had already surveyed – about 30% of the stock – most windows and doors would need replacing “sooner rather than later”.
It also said many homes would need insulation upgrades, and many boilers would require associated pipework and radiators to be replaced.
But officers found kitchens and bathrooms to be generally in a fair condition, and said the stock overall is not in a poor condition, but would be “hungry for investment” in the next decade.
The new housing boss said she was “fairly comfortable” with the results and hoped there would not be any more nasty surprises.
Tandridge’s improvement plan will continue into 2026/27 with progress reported back to both the regulator and councillors.
Long Beach, California. Rocket Lab Corporation (Nasdaq: RKLB) (“Rocket Lab” or “the Company”), a global leader in launch services and space systems, today announced it successfully launched a suborbital mission with its HASTE launch vehicle for the Defense Innovation Unit (DIU) and Missile Defense Agency (MDA) – advancing national interests in safeguarding the homeland through the testing of advanced technologies for missile defense.
The launch on HASTE – Rocket Lab’s commercial launch vehicle for regular and reliable hypersonic test flights – took place from Rocket Lab Launch Complex 2 on Wallops Island, Virginia, at 13:00 UTC/08:00 a.m. ET on November 18, 2025. Led by MDA, the mission deployed a government-provided primary payload developed by the John Hopkins University Applied Physics Laboratory, and multiple secondary payloads by federal and industry partners, which tested key technologies for missile defense applications.
The mission was contracted to Rocket Lab through the DIU’s Hypersonic and High-Cadence Airborne Testing Capabilities (HyCAT) program, an initiative supporting test and evaluation of new and emerging hypersonic technologies through low cost, responsive and long endurance flight testing. The mission launched within 14 months of contract signing, demonstrating streamlined operational benefits for government customers through Rocket Lab’s commercial speed, innovation, and efficiency. The mission also exemplified the cost and schedule savings that commercial liquid launch vehicles can bring to the MDA test community for developmental testing, non-traditional targets testing, and risk-reduction payload testing activities.
Rocket Lab’s Vice President Global Launch Services, Brian Rogers, says: “HASTE is an important platform for accelerating hypersonic technology readiness for the nation, and we’re proud to be delivering this mission for DIU and MDA.”
LtCol Nicholas Estep, Director of DIU’s Emerging Technology Portfolio, says: “Accessing the commercial and non-traditional ecosystem is a key enabler to accelerating progress in the hypersonics community of interest, particularly for closing mission timelines and driving towards mass and affordability. Working with MDA to demonstrate commercially-focused sub-orbital launch services is a great example of that axiom.”
The mission was Rocket Lab’s sixth launch of its HASTE rocket since the launch vehicle’s debut in 2023. A suborbital variant of Electron – the world’s most frequently launched small orbital rocket – HASTE includes much of the same innovative technology as Electron, including carbon fiber composite structures and 3D printed rocket engines, but has a modified upper Kick Stage tailored for hypersonic technology tests and a larger payload capacity. HASTE can deploy technologies at speeds of more than 7.5km per second to test air-breathing, glide, and ballistic payloads, as well as technologies to re-enter Earth’s atmosphere from space. Combined, the HASTE and Electron launch vehicles have deployed 200+ payloads for government and commercial customers to date.
Media Contact Murielle Baker media@rocketlabusa.com
About Rocket Lab Rocket Lab is a leading space company that provides launch services, spacecraft, payloads and satellite components serving commercial, government, and national security markets. Rocket Lab’s Electron rocket is the world’s most frequently launched orbital small rocket; its HASTE rocket provides hypersonic test launch capability for the U.S. government and allied nations; and its Neutron launch vehicle in development will unlock medium launch for constellation deployment, national security and exploration missions. Rocket Lab’s spacecraft and satellite components have enabled more than 1,700 missions spanning commercial, defense and national security missions including GPS, constellations, and exploration missions to the Moon, Mars, and Venus. Rocket Lab is a publicly listed company on the Nasdaq stock exchange (RKLB). Learn more at www.rocketlabcorp.com
Forward Looking Statements This press release contains forward-looking statements within the meaning of the Private Securities Litigation Reform Act of 1995. We intend such forward-looking statements to be covered by the safe harbor provisions for forward looking statements contained in Section 27A of the Securities Act of 1933, as amended (the “Securities Act”) and Section 21E of the Securities Exchange Act of 1934, as amended (the “Exchange Act”). All statements contained in this press release other than statements of historical fact, including, without limitation, statements regarding our launch and space systems operations, launch schedule and window, safe and repeatable access to space, Neutron development, operational expansion and business strategy are forward-looking statements. The words “believe,” “may,” “will,” “estimate,” “potential,” “continue,” “anticipate,” “intend,” “expect,” “strategy,” “future,” “could,” “would,” “project,” “plan,” “target,” and similar expressions are intended to identify forward-looking statements, though not all forward-looking statements use these words or expressions. These statements are neither promises nor guarantees, but involve known and unknown risks, uncertainties and other important factors that may cause our actual results, performance or achievements to be materially different from any future results, performance or achievements expressed or implied by the forward-looking statements, including but not limited to the factors, risks and uncertainties included in our Annual Report on Form 10-K for the fiscal year ended December 31, 2024, as such factors may be updated from time to time in our other filings with the Securities and Exchange Commission (the “SEC”), accessible on the SEC’s website at www.sec.gov and the Investor Relations section of our website at www.rocketlabcorp.com, which could cause our actual results to differ materially from those indicated by the forward-looking statements made in this press release. Any such forward-looking statements represent management’s estimates as of the date of this press release. While we may elect to update such forward-looking statements at some point in the future, we disclaim any obligation to do so, even if subsequent events cause our views to change.
Capgemini deepens partnership with SAP to bolster Europe’s digital sovereignty and accelerate time to value for AI-powered enterprise innovation and transformation – Capgemini
Capgemini deepens partnership with SAP to bolster Europe’s digital sovereignty and accelerate time to value for AI-powered enterprise innovation and transformation – Capgemini