New long-term extension data show approximately 80% of patients achieved or maintained meaningful skin improvement (EASI 75) with EBGLYSS with half the doses compared to approved monthly maintenance dosing
Lilly submitted these data to the FDA for a potential label update for EBGLYSS
If approved, EBGLYSS would be a first-line biologic that offers the option of monotherapy with once every eight-week maintenance dosing in moderate-to-severe atopic dermatitis uncontrolled by topicals
INDIANAPOLIS, Oct. 24, 2025 /PRNewswire/ — New results show Eli Lilly and Company’s (NYSE: LLY) EBGLYSS (lebrikizumab-lbkz) sustained similar levels of skin clearance when administered as a single injection of 250 mg once every eight weeks (Q8W) compared with once every four weeks (Q4W), supporting a potential additional, less frequent maintenance dosing option for more individualized treatment of patients with moderate-to-severe atopic dermatitis. These findings from the Phase 3 ADjoin extension trial will be presented at the 2025 Fall Clinical Dermatology Conference, taking place Oct. 23-26 in Las Vegas.1
“For people managing the persistent symptoms of eczema, hesitancy about frequent injections can add to the already heavy toll of this disease,” said Peter Lio, M.D., author of the ADjoin study and clinical assistant professor of dermatology and pediatrics, Northwestern University. “With as few as six maintenance doses per year, EBGLYSS would give patients and providers more flexibility, which may reduce treatment burden for patients with busy lives.”
EBGLYSS is an interleukin-13 (IL-13) inhibitor that selectively blocks IL-13 signaling with high binding affinity.2,3,4 The cytokine IL-13 is a primary cytokine in atopic dermatitis, driving the type-2 inflammatory cycle in the skin, leading to skin barrier dysfunction, itch, skin thickening and infection.5,6
In the ADjoin extension study, results indicate that maintenance dosing every other month demonstrated similarly high rates of disease control compared to monthly dosing:
79% of patients taking EBGLYSS once every other month and 86% of patients taking EBGLYSS monthly, respectively, achieved or maintained EASI 75.*
62% of patients taking EBGLYSS once every other month and 73% of patients taking EBGLYSS monthly, respectively, achieved or maintained IGA 0,1.
There was no increased risk of immunogenicity (the production of anti-drug-antibodies), and no new safety findings. These data support that once every eight-week EBGLYSS dosing could give HCPs and patients a new treatment option using the lowest effective dose.
“Managing moderate-to-severe atopic dermatitis involves ongoing cycles of flare-ups and itching, which can be difficult for people with eczema,” said Kristin Belleson, President and CEO of the National Eczema Association. “Treatment options that have the potential to reduce the time people spend managing symptoms could give them more time to focus on what matters most.”
Lilly has submitted these data from the ADjoin extension trial among other data to the FDA for a potential label update. A study investigating EBGLYSS maintenance dosing of 500 mg administered once every 12 weeks (Q12W) is underway.
“Lilly continues to optimize dosing frequency to push boundaries that redefine the patient experience. These new findings build on EBGLYSS’ proven efficacy and demonstrate the potential for disease control with even less frequent dosing,” said Mark Genovese M.D., senior vice president of Lilly Immunology development. “We are pursuing an every-eight-week maintenance dosing label update with the FDA. We are also testing every-twelve-week maintenance dosing with our partner Almirall, as well as potentially exploring every-twelve-week dosing in independent Lilly-led studies.”
These data build on existing research for EBGLYSS, which has demonstrated long-term results maintained for up to three years, as well as efficacy data across diverse skin tones. EBGLYSS is the only biologic for moderate-to-severe atopic dermatitis with a strong recommendation and high certainty of evidence that can be used with or without topicals, according to guidelines published by the American Academy of Dermatology (AAD),** which are used as a key consideration for dermatologists and managed care providers.
Lilly continues to raise the standard of care in dermatology and invest in our immunology pipeline, which includes big bets on next-generation modalities and the targeted expansion of small molecules. Lilly’s investigational therapies include novel, oral IL-17 inhibitors such as DICE Therapeutics’ DC-853, which is being studied for psoriasis, and eltrekibart, a novel monoclonal antibody that targets neutrophil-driven inflammation and is being assessed in hidradenitis suppurativa. Lilly is also advancing novel science to explore the potential of incretins in dermatology and has initiated the TOGETHER-PsO trial investigating the efficacy and safety of treating adults with moderate-to-severe plaque psoriasis and obesity with both ixekizumab and an incretin-based therapy.
Lilly has exclusive rights for development and commercialization of EBGLYSS in the U.S. and the rest of the world outside Europe. Lilly’s partner Almirall has licensed the rights to develop and commercialize EBGLYSS for the treatment of dermatology indications, including atopic dermatitis, in Europe.
*EASI=Eczema Area and Severity Index, EASI-75=75% reduction in EASI from baseline; IGA=Investigator’s Global Assessment 0 or 1 (“clear” or “almost clear”). **Inclusion in the Focused Update: AAD Guidelines of Care for the Management of Atopic Dermatitis in Adults does not denote endorsement of product use by the AAD.
About the Q8W ADjoin Extension The Q8W extension of ADjoin (NCT04392154) assessed EBGLYSS given every eight weeks (Q8W) compared to every four weeks (Q4W) and evaluated the long-term safety and efficacy of EBGLYSS treatment in patients with moderate-to-severe atopic dermatitis for 32 weeks, in select countries. Adult and adolescent patients (ages 12–17, weighing ≥40 kg) who completed the 100-week ADjoin study, including participants from the ADvocate 1 and 2 trials (52 weeks), ADore trial (52 weeks) and the ADopt-VA (16 weeks) trial, were eligible to enroll in the Q8W extension. Patients in this analysis received open-label EBGLYSS 250 mg, Q8W or Q4W, regardless of their previous treatment in ADjoin (Q2W or Q4W dose). The approved maintenance dose of EBGLYSS is 250 mg once monthly, after taking EBGLYSS 250 mg every two weeks for 16 weeks or later when adequate clinical response is achieved.7
INDICATION AND SAFETY SUMMARY EBGLYSS™ (EHB-glihs) is an injectable medicine used to treat adults and children 12 years of age and older who weigh at least 88 pounds (40 kg) with moderate-to-severe eczema (atopic dermatitis) that is not well controlled with prescription therapies used on the skin (topical), or who cannot use topical therapies. EBGLYSS can be used with or without topical corticosteroids.
It is not known if EBGLYSS is safe and effective in children less than 12 years of age or in children 12 years to less than 18 years of age who weigh less than 88 pounds (40 kg).
Warnings – Do not useEBGLYSS if you are allergic to lebrikizumab-lbkz or to any of the ingredients in EBGLYSS. See the Patient Information leaflet that comes with EBGLYSS for a complete list of ingredients.
Before using Before using EBGLYSS, tell your healthcare provider about all your medical conditions, including if you:
Have a parasitic (helminth) infection.
Are scheduled to receive any vaccinations. You should not receive a “live vaccine” if you are treated with EBGLYSS.
Are pregnant or plan to become pregnant. It is not known if EBGLYSS will harm your unborn baby. If you become pregnant during treatment with EBGLYSS, you or your healthcare provider can call Eli Lilly and Company at 1-800-LillyRx (1-800-545-5979) to report the pregnancy.
Are breastfeeding or plan to breastfeed. It is not known if EBGLYSS passes into your breast milk.
Tell your healthcare provider about all the medicines you take, including prescription and over-the-counter medicines, vitamins, and herbal supplements.
Possible side effects EBGLYSS can cause serious side effects, including:
Allergic reactions. EBGLYSS can cause allergic reactions that may sometimes be severe. Stop using EBGLYSS and tell your healthcare provider or get emergency help right away if you get any of the following signs or symptoms:
breathing problems or wheezing
swelling of the face, lips, mouth, tongue or throat
hives
itching
fainting, dizziness, feeling lightheaded
skin rash
cramps in your stomach area (abdomen)
Eye problems. Tell your healthcare provider if you have any new or worsening eye problems, including eye pain or changes in vision, such as blurred vision.
The most common side effects of EBGLYSS include:
eye and eyelid inflammation, including redness, swelling, and itching
injection site reactions
shingles (herpes zoster)
These are not all of the possible side effects of EBGLYSS. Call your doctor for medical advice about side effects. You may report side effects to FDA at 1-800-FDA-1088 or www.fda.gov/medwatch.
How to take
See the detailed “Instructions for Use” that comes with EBGLYSS for information about how to prepare and inject EBGLYSS and how to properly store and throw away (dispose of) used EBGLYSS prefilled pens and prefilled syringes.
Use EBGLYSS exactly as prescribed by your healthcare provider.
EBGLYSS is given as an injection under the skin (subcutaneous injection).
If your healthcare provider decides that you or a caregiver can give the injections of EBGLYSS, you or a caregiver should receive training on the right way to prepare and inject EBGLYSS. Do not try to inject EBGLYSS until you have been shown the right way by your healthcare provider. In children 12 years of age and older, EBGLYSS should be given by a caregiver.
If you miss a dose of EBGLYSS, inject the missed dose as soon as possible, then inject your next dose at your regular scheduled time.
Learn more EBGLYSS is a prescription medicine available as a 250 mg/2 mL injection prefilled pen or prefilled syringe. For more information, call 1-800-545-5979 or go to ebglyss.lilly.com
This summary provides basic information about EBGLYSS but does not include all information known about this medicine. Read the information that comes with your prescription each time your prescription is filled. This information does not take the place of talking to your doctor. Be sure to talk to your doctor or other healthcare provider about EBGLYSS and how to take it. Your doctor is the best person to help you decide if EBGLYSS is right for you.
LK CON BS AD APP
EBGLYSS, its delivery device base, and Lilly Support Services are trademarks owned or licensed by Eli Lilly and Company, its subsidiaries, or affiliates.
About EBGLYSS EBGLYSS is a monoclonal antibody that selectively targets and neutralizes IL-13 with high binding affinity and a slow dissociation rate.3,4,7 EBGLYSS binds to the IL-13 cytokine at an area that overlaps with the binding site of the IL-4Rα subunit of the IL-13Rα1/IL-4Rα heterodimer, preventing formation of this receptor complex and inhibiting IL-13 signaling. IL-13 is implicated as a primary cytokine tied to the pathophysiology of eczema, driving the type-2 inflammatory loop in the skin, and EBGLYSS selectively targets IL-13.7
The EBGLYSS Phase 3 program consists of five key global studies evaluating over 1,300 patients, including two monotherapy studies (ADvocate 1 and 2), a combination study with topical corticosteroids (ADhere), as well as long-term extension (ADjoin) and adolescent open label (ADore) studies.
EBGLYSS was approved in the U.S. by the Food and Drug Administration (FDA) in 2024 as a first-line monotherapy biologic treatment with once-monthly maintenance dosing for adults and children 12 years of age and older who weigh at least 88 pounds (40 kg) with moderate-to-severe atopic dermatitis that is not well controlled with topical prescription therapies.7 EBGLYSS was also approved in the European Union in 2023 and in Japan and Canada in 2024.
EBGLYSS 250 mg/2 mL injection is dosed as a single monthly maintenance injection following the initial phase of treatment. The recommended initial starting dose of EBGLYSS is 500 mg (two 250 mg injections) at Week 0 and Week 2, followed by 250 mg every two weeks until Week 16 or later when adequate clinical response is achieved; after this, maintenance dosing is a single monthly injection (250 mg every four weeks).7
Lilly is committed to serving patients living with moderate-to-severe atopic dermatitis and is working to enable broad first-line biologic access to EBGLYSS following topical prescription therapy through commercial insurance and as of October 24, Lilly has coverage with all three major national pharmacy benefit managers and over 90% of people with commercial insurance. We are also pursuing similarly broad Medicaid and Medicare coverage as part of Lilly’s health access and affordability initiative. Through Lilly Support Services, Lilly offers a patient support program including co-pay assistance for eligible, commercially insured patients.
About Lilly Lilly is a medicine company turning science into healing to make life better for people around the world. We’ve been pioneering life-changing discoveries for nearly 150 years, and today our medicines help tens of millions of people across the globe. Harnessing the power of biotechnology, chemistry and genetic medicine, our scientists are urgently advancing new discoveries to solve some of the world’s most significant health challenges: redefining diabetes care; treating obesity and curtailing its most devastating long-term effects; advancing the fight against Alzheimer’s disease; providing solutions to some of the most debilitating immune system disorders; and transforming the most difficult-to-treat cancers into manageable diseases. With each step toward a healthier world, we’re motivated by one thing: making life better for millions more people. That includes delivering innovative clinical trials that reflect the diversity of our world and working to ensure our medicines are accessible and affordable. To learn more, visit Lilly.com and Lilly.com/news, or follow us on Facebook, Instagram and LinkedIn. P-LLY
Trademarks and Trade Names All trademarks or trade names referred to in this press release are the property of the company, or, to the extent trademarks or trade names belonging to other companies are references in this press release, the property of their respective owners. Solely for convenience, the trademarks and trade names in this press release are referred to without the ® and ™ symbols, but such references should not be construed as any indicator that the company or, to the extent applicable, their respective owners will not assert, to the fullest extent under applicable law, the company’s or their rights thereto. We do not intend the use or display of other companies’ trademarks and trade names to imply a relationship with, or endorsement or sponsorship of us by, any other companies.
Cautionary Statement Regarding Forward-Looking Statements This press release contains forward-looking statements (as that term is defined in the Private Securities Litigation Reform Act of 1995) about EBGLYSS (lebrikizumab-lbkz) as a treatment for patients with moderate-to severe atopic dermatitis and the timeline for future readouts, presentations, and other milestones relating to EBGLYSS and its clinical trials and reflects Lilly’s current beliefs and expectations. However, as with any pharmaceutical product, there are substantial risks and uncertainties in the process of drug research, development, and commercialization. Among other things, there is no guarantee that future study results will be consistent with the results to date or that EBGLYSS will receive additional regulatory approvals, or that it will be commercially successful. For further discussion of these and other risks and uncertainties that could cause actual results to differ from Lilly’s expectations, see Lilly’s Form 10-K and Form 10-Q filings with the United States Securities and Exchange Commission. Except as required by law, Lilly undertakes no duty to update forward-looking statements to reflect events after the date of this release.
Silverberg J, et al. Lebrikizumab Dosed Every 8 Weeks as Maintenance Provides Long-Lasting Response in Patients with Moderate-to-Severe Atopic Dermatitis. 2025 Fall Clinical Dermatology Conference. October 24, 2025.
Simpson EL, et al. Efficacy and safety of lebrikizumab (an anti-IL-13 monoclonal antibody) in adults with moderate-to-severe atopic dermatitis inadequately controlled by topical corticosteroids: A randomized, placebo-controlled phase II trial (TREBLE). J Am Acad Dermatol. 2018;78(5):863-871.e11. doi:10.1016/j.jaad.2018.01.017.
Okragly A, et al. Binding, Neutralization and Internalization of the Interleukin-13 Antibody, Lebrikizumab. Dermatol Ther (Heidelb). 2023;13(7):1535-1547. doi:10.1007/s13555-023-00947-7.
Ultsch M, et al. Structural basis of signaling blockade by anti-IL-13 antibody Lebrikizumab. J Mol Biol. 2013;425(8):1330-1339. doi:10.10116/j.jmb.2013.01.024.
Bieber T. Interleukin-13: Targeting an underestimated cytokine in atopic dermatitis. Allergy. 2020;75(1):54–62. doi:10.1111/all.13954.
Tsoi LC, et al. Atopic Dermatitis Is an IL-13-Dominant Disease with Greater Molecular Heterogeneity Compared to Psoriasis. J Invest Dermatol. 2019;139(7):1480-1489. doi:10.1016/j.jid.2018.12.018.
JPMorgan is coming off the sidelines when it comes to Coinbase . The bank upgraded shares of the cryptocurrency exchange to an overweight rating from neutral. It also lifted its price target to $404 per share from $342. Coinbase stock has rallied 30% this year, but the firm said it still has an attractive valuation versus competitors. JPMorgan’s revised forecast calls from an additional upside of 25% from here. COIN YTD mountain COIN YTD chart As a catalyst, analyst Kenneth Worthington pointed to Coinbase’s exploration of launching a potential Base token, which he sees as accelerating the growth of development on the Base blockchain, launched in August 2023. The market capitalization of such a token could eventually swell to between $12 billion and $34 billion, he wrote. “We see a token as a win/ win for shareholders and cryptocurrency participants with a token driving development, increasing community participation and contributing to infrastructure buildout,” the analyst added. Another catalyst comes as Coinbase further exploring with USDC payouts in order to promote Coinbase One, its monthly subscription service. USDC is a stablecoin that is pegged to the dollar. “We see Coinbase experimenting with yields on USDC, both to drive further growth of Coinbase One, but also to better monetize on-platform USDC,” the analyst wrote. “We see Coinbase potentially driving more clients to utilize Coinbase One, which offers subscription pricing for trading as well as other benefits, including an enhanced yield on USDC.” Meantime, Worthington believes that the threat of decentralized exchanges taking away share of cryptocurrency spot trading from Coinbase has largely begun to stabilize. Coinbase’s launch of decentralized trading via Base also helps to mitigate this competitive threat, he added. JPMorgan says when applying a 50x P-E that its competitors such as Circle and Bullish have to Coinbase’s 2027 earnings estimate and including the value of the Base token, it arrives at the $404 target price. ( Learn the best 2026 strategies from inside the NYSE with Josh Brown and others at CNBC PRO Live. Tickets and info here . )
Female C57BL/6 mice (6–8 weeks old) were purchased from the Beijing Experimental Animal Center. Conditional Dnmt1 knockout mice (Dnmt1fl/flItgax-Cre) and their littermate controls (Dnmt1fl/fl) were generated by crossing Dnmt1fl/fl mice (Jackson Laboratory, Stock No. 021429) with Itgax-Cre mice (Jackson Laboratory, Stock No. 008068). All experimental protocols were approved by the Animal Care and Use Committee (IACUC) of Shanxi Medical University (Approval Number: A20230086) and complied with the guidelines of the National Institutes of Health (NIH) for the care and use of laboratory animals.
Induction of airway allergy (AA)
Mice were sensitized to induce airway allergy using dust mite extract (DME). On days 0 and 7, mice received subcutaneous injections of 50 µg DME (Dermatophagoides pteronyssinus; Greer Laboratories, Cat. No. XPB82D3A2.5) emulsified in 2 mg Imject® Alum (Thermo Fisher Scientific). From days 14 to 18, mice were challenged intranasally (i.n.) with 25 µg DME in 40 µL phosphate-buffered saline (PBS); control mice received PBS alone. Following DME challenge, clinical symptoms of AA (sneezing, nasal rubbing, and dyspnea) were scored to assess disease severity.
Drug administration
Mice were treated with 5-azacytidine (5AZA; Sigma-Aldrich, Cat. No. A2385) or PBS (vehicle control) to inhibit DNA methyltransferases. Daily intraperitoneal (i.p.) injections of 5AZA (0.5 mg/kg body weight) or PBS were administered from day 0 to day 18 of the experimental timeline.
Cell isolation and culture
Bone marrow-derived DCs (BMDCs)
Bone marrow cells were flushed from the femurs of mice and cultured in RPMI-1640 medium (Gibco) supplemented with 10% fetal bovine serum (FBS), 20 ng/mL granulocyte-macrophage colony-stimulating factor (GM-CSF; PeproTech), and antibiotics (100 U/mL penicillin and 100 µg/mL streptomycin) for 7 days to generate BMDCs.
BMDC stimulation
Mature BMDCs were treated with DME (10–50 µg/mL) in the presence or absence of lipopolysaccharide (LPS; 1 µg/mL, E. coli O111:B4; Sigma-Aldrich) for 24 h to assess cytokine production and epigenetic changes.
Molecular biology techniques
Reverse transcription quantitative PCR (RT-qPCR)
Total RNA was extracted from cells using TRIzol reagent (Invitrogen), and complementary DNA (cDNA) was synthesized with SuperScript IV Reverse Transcriptase (Thermo Fisher Scientific). Quantitative PCR was performed using SYBR Green Master Mix (Bio-Rad) on a CFX96 Real-Time PCR System (Bio-Rad) to quantify the expression of Il12b, Il12a, Tim4, Il4, and Gapdh (housekeeping gene). Primer sequences are listed in Table 1.
Table 1 Primer sequences for RT-qPCR
Enzyme-linked immunosorbent assay (ELISA)
Cytokine levels (IL-4, IL-5, IL-13, IFN-γ, TNF-α) in bronchoalveolar lavage fluid (BALF) were measured using DuoSet ELISA kits (R&D Systems) according to the manufacturer’s instructions. Absorbance was read at 450 nm, and cytokine concentrations were calculated using standard curves.
Methylation-specific quantitative PCR (MS-qPCR)
Genomic DNA was extracted from DCs and bisulfite-converted using the EZ DNA Methylation-Lightning Kit (Zymo Research) to distinguish methylated from unmethylated cytosines. The methylation status of the Il12b promoter was analyzed by MS-qPCR using primers specific for methylated (MSP) or unmethylated (USP) sequences: MSP-F: 5’-TTGGGATTTTTCGCGATTC-3’, MSP-R: 5’-GAACCCGACGAACTCCG-3’; USP-F: 5’-TTGGGATTTTTTGTGATTTG-3’, USP-R: 5’-CAACCCCACAAACTCCA-3’. Reactions were performed with SYBR Green Master Mix (Bio-Rad) on a CFX96 Real-Time PCR System.
Flow cytometry
Lung immune profiling
Lung tissues were processed into single-cell suspensions, which were stained with fluorescently conjugated antibodies: CD11c (APC), MHC-II (PE-Cy7), Ly6G (FITC), Siglec-F (BV421), CD3 (PerCP), CD4 (AF700), and CD8 (PE) (all from BD Biosciences). Stained cells were acquired on a BD LSRFortessa™ flow cytometer, and data were analyzed using FlowJo software (v10).
Intracellular cytokine staining
DCs were stimulated with 50 ng/mL phorbol 12-myristate 13-acetate (PMA) and 1 µg/mL ionomycin for 4 h. Cells were then fixed and permeabilized using the Cytofix/Cytoperm Kit (BD Biosciences) and stained with antibodies against IL-12b (PE) or IL-12a (AF488) (BD Biosciences) for intracellular cytokine detection.
Histology
Lungs were inflated with 4% paraformaldehyde, fixed overnight, and embedded in paraffin. Paraffin blocks were sectioned into 5-µm slices, which were stained with hematoxylin and eosin (H&E) to assess tissue inflammation. Stained sections were imaged using a Leica DMi8 microscope, and histological changes were evaluated by a blinded pathologist.
Airway hyperresponsiveness (AHR)
AHR was measured using a flexiVent system (SCIREQ) to assess airway resistance. Mice were anesthetized, intubated, and exposed to aerosolized methacholine (0–50 mg/mL) for 3 min. Airway resistance was recorded continuously during and after methacholine exposure, and results were normalized to baseline values.
Co-culture experiments
DME-primed BMDCs were co-cultured with naïve CD4+ T cells (isolated from mouse spleens using MACS columns; Miltenyi Biotec) at a 1:5 (DC: T cell) ratio in RPMI-1640 medium supplemented with 10% FBS and antibiotics. After 72 h of co-culture, Il4 mRNA expression in CD4+ T cells was quantified by RT-qPCR to assess Th2 polarization.
Ubiquitination assays (Cross-ELISA)
BMDCs were treated with 50 ng/mL recombinant IL-12b (rIL-12b; BioLegend) in the presence or absence of 10 µM MG132 (a proteasome inhibitor; Sigma-Aldrich) for 6 h. Cells were lysed in RIPA buffer, and ubiquitinated TIM4 was detected using a cross-ELISA method. Briefly, plates were coated with an anti-TIM4 antibody (Ab1), and bound ubiquitinated TIM4 was detected with an anti-ubiquitin antibody (Ab2); protocols were adapted from Luo et al. [10] and otherwise consistent with standard ELISA procedures.
Chromatin immunoprecipitation (ChIP)
DCs were crosslinked with 1% formaldehyde for 10 min at room temperature, and crosslinking was quenched with 125 mM glycine. Cells were lysed, and chromatin was sheared by sonication to generate 200–500 bp fragments. Immunoprecipitation was performed overnight at 4 °C using antibodies against DNMT1 (Abcam), H3K4me3 (Cell Signaling Technology), or RNA polymerase II (RNA Pol II; Santa Cruz Biotechnology); normal IgG was used as a negative control. Immunoprecipitated DNA was purified and quantified by qPCR using primers targeting the Il12b promoter to assess enrichment of the target proteins.
Statistical analysis
All data are presented as the mean ± standard deviation (SD). Statistical comparisons between two groups were performed using Student’s t-test. For multiple groups, one-way analysis of variance (ANOVA) with Tukey’s post hoc test was used. Analyses were conducted using R (v4.5.1) software. Significance was defined as p < 0.05 (p < 0.05 = *, p < 0.01 = **, p < 0.001 = ***, p < 0.0001 = ****).
A fresh consultation has been launched into proposals for what, if approved, would be one of the UK’s largest solar energy schemes.
Energy developer Island Green Power wants to build the 500MW Light Valley Solar project, consisting of seven solar farms on a 2,520 acre (1,020 hectare) site, between York and Selby.
Annette Lardeur, senior project development manager, said this third consultation concerned changes to the plans which were “minor in nature” and would “improve safety and access”.
However, Louise Billingham, from campaign group Rooftops Not Countryside, said its members remained opposed to the plans due to the site’s “vast scale and inappropriate placement”.
Further land on the development would be used for underground cabling to connect the sites to the national grid at Monk Fryston substation, meanwhile a 500MW battery energy storage system (BESS) was also proposed.
Due to the size of the proposed development, which would be located between the villages of Escrick, Monk Fryston, Hambleton, Chapel Haddlesey and South Milford, the Light Valley Solar scheme has been declared a Nationally Significant Infrastructure Project.
That means planning permission will be decided by the national Planning Inspectorate, rather than North Yorkshire Council.
According to the Local Democracy Reporting Service, Ms Lardeur said she wanted to “encourage people to share their views” as part of the new consultation.
“The changes we are consulting on are minor in nature and have been introduced to improve safety and access if the scheme is consented,” she explained.
“As well as this, they aim to reduce the amount of construction traffic passing through villages. I encourage people to share their views.”
Ms Billingham said the campaign group, which had collected nearly 3,000 signatures against the plans, said if the solar farm went ahead, “our countryside will no longer be countryside, it will be a vast industrial estate”.
“This impromptu round of consultation is for Light Valley Solar to make changes that will remove established hedgerows, trees and nature and also build lay-bys and roads,” she said.
“They plan to build one million 4m high solar panels, numerous substations and BESS sites, as well as cable corridors, high-security fencing and CCTV across 13 miles of countryside between Monk Fryston and Escrick.”
The group had so far raised at least £2,000 through crowdfunding for the campaign against the plans, which would pay for banners and leaflets to raise awareness of the proposed solar farm, Ms Billingham added.
If approved, Light Valley would be bigger than what is currently the largest operational solar scheme, at Cleave Hill in Kent.
The scheme would provide enough power for 115,000 homes a year, according to Island Green Power.
Colombia’s data centre market is expanding at a remarkable pace, fuelled by the rise of artificial intelligence (AI). As of 2025, the country hosts 38 data centres — 29 in Bogotá, 4 in Medellín, 3 in Cali, and 2 in Barranquilla — with several more on under development. Valued at US$442 million in 2024, the sector is expected to nearly triple to US$1.16 billion by 2030, growing at an annual rate of 17.6%.
This boom promises major opportunities in digital infrastructure and foreign investment. Yet it also brings serious sustainability challenges. Data centres are notoriously resource-intensive, consuming vast amounts of water for cooling and electricity to keep servers running.
Recognising both the potential and the risks, the Colombian government launched a US$111.5 million National AI Policyin February 2025. The plan seeks to decentralise data centre infrastructure and promote hyperscale facilities powered by renewable energy and water-efficient technologies.
The pressing challenge now is ensuring that Colombia’s AI-driven data centre boom drives inclusive economic growth while safeguarding the country’s water and energy resources.
Projected Impacts
Water Stress
AI’s rapid growth comes with a heavy water footprint. Globally, meeting demand for AI could require up to 6.6 billion cubic meters of water withdrawals by 2027. Data centre water consumption is driven by cooling (Scope 1), electricity usage (Scope 2), and hardware manufacturing (Scope 3).
To address environmental concerns, some providers are experimenting with air-cooling systems that reduce direct water withdrawals. However, these systems are often more energy-intensive, which can increase Scope 2 water use. The lack of reliable data on Scope 3 impacts further complicates assessments, highlighting the need for more transparent and comprehensive reporting.
Example of a data centre’s operational (scope 1+2) water usage
Source: Li et al. (2025)
Colombia faces particular risks. New data centres are planned in hot regions such as Santa Marta, where cooling demands will be especially high. Even Bogotá, with its milder climate, faces serious water availability concerns. Between 2023 and 2024, the city experienced once of its worst droughts on record, with its reservoir levels falling to just 10.5%.
Adding to the challenge, Colombia is already one of the world’s top consumers of water per capita, using nearly three times the OECD average. This is partly driven by industrial users facing some of the lowest water tariffs in the region. In 2023, Coca-Cola reportedly paid just US$2,605 to extract 64 million litres near Bogotá. At US$0.008 per cubic meter, Colombia’s industrial water tariff is a fraction of Mexico’s (US$0.99) or Brazil’s (US$0.036).
Colombia’s governance framework is not keeping up with these mounting pressures. An evaluation of the National Policy for Integrated Water Resources Management (2010–2022) found that the country lacks adequate data on water demand, while Bogotá’s District Water Plan, due for an update in 2021, has yet to be revised. Both national and local policies urgently need updating to integrate future water use projections tied to data centre growth.
Without stronger governance and more realistic tariffs, water scarcity could not only constrain Colombia’s AI-driven data centre industry but also threaten access to clean water for communities, exacerbating social and environmental stresses across the country.
Energy use
AI is driving a sharp rise in data centres’ energy needs. While traditional facilities consume between 10 and 25 megawatts (MW) annually, hyperscale AI-focused centres can exceed 100 MW — the equivalent of powering 100,000 households. Because these facilities often cluster in specific cities, their concentrated demand puts major stress on local power grids.
To manage both energy security and carbon footprints, many operators are investing in renewable sources. These projects can accelerate growth in the clean energy sector, but much of the power produced is consumed by the centres themselves, limiting the amount of renewable energy available to the wider grid.
Colombia’s relatively clean electricity mix has helped attract investment. In 2023, 64.4% of power came from hydropower, while 31.8% came from fossil fuels, and only 1.2% from solar and wind. The country’s heavy reliance on dams, however, leaves the system vulnerable. Recent droughts forced greater use of fossil fuels to meet demand, undercutting Colombia’s climate goals. Without a rapid scale-up of solar and wind capacity, rising demand from data centres risks locking the country into deeper dependence on fossil energy, with consequences for both emissions targets and long-term energy security.
Electricity generation sources in Colombia, 2023
Press enter or click to view image in full size
Source: IEA (2023)
Over the past three decades, Colombia has modernised its energy sector and welcomed private investment, spurring diversification and infrastructure growth. Still, persistent barriers remain. There are gaps in transmission capacity, delays in environmental licensing, and strong community opposition to new projects. Unless these challenges are resolved, Colombia will struggle to expand renewable energy fast enough to keep pace with the demands of an AI-driven data centre boom.
Case Studies
Because data on Colombia’s data centre impacts remains limited, it is useful to look at two regional leaders, Mexico and Chile, where the sector is more mature. Both cases offer valuable lessons for Colombia as it expands its own market.
Mexico
Mexico’s data centre market is projected to reach US$2.26 billion by 2030, growing at an annual rate of 13.5%. The country already hosts 59 centres, with nearly a third concentrated in Querétaro. Tech companies have pledged thousands of jobs and billions in investments, but these promises come with serious environmental trade-offs.
Querétaro is one of Mexico’s most drought-prone states. Yet industrial water use there skyrocketed from 4% of total consumption in 2003 to 40% by 2018. The surge has fuelled protests from indigenous and rural communities, driven up water costs for residents, and left local reservoirs under severe stress. In April 2025, two of the state’s seven reservoirs were completely dry, and overall capacity dropped to just 32.5%. Energy shortages have compounded the crisis, with frequent power outages as the regional grid struggles to keep up with rising demand.
Despite these risks, investment continues. In January 2025, Amazon Web Services announced a US$5 billion investment in Querétaro to serve all of Latin America. Mexico does have a sustainability norm for data centres introduced in 2014 (NMX-489), but because compliance is voluntary, its impact has been limited.
Querétaro’s experience underscores a key warning for Colombia. Without proactive water and energy management, rapid data centre growth can quickly outpace local resources and spark social unrest.
Planned data centre investments in Mexican states under drought.
Source: Menski et al. (2024)
Chile
Chile’s data centre market is projected to hit US$1.24 billion by 2030, growing at a CAGR of 8.27%. The country currently has 62 centres, 55 of them clustered in Santiago. But Chile is also one of the world’s most water-stressed nations, and public opposition has been strong.
In 2024, Google was forced to suspend a planned US$200 million data centre in Santiago after citizen mobilisation and a partial court reversal of its environmental permits. The company later committed to redesigning the project with stricter sustainability standards and more water-efficient technologies.
To address these challenges, Chile recently launched its National Data Centre Plan (2024–2030), developed through multi-stakeholder consultations. The plan includes a digital tool to guide sustainable site selection, a multi-stakeholder monitoring committee, streamlined permitting processes, and public–private clean production agreements to boost efficiency. It also encourages the creation of AI campuses in regions with abundant renewable energy resources, reducing pressure on water-stressed areas like Santiago.
Chile’s proactive approach highlights the importance of early planning, strong stakeholder engagement, and enforceable sustainability standards; lessons Colombia can apply before its own boom accelerates further. That said, it remains to be seen how effectively the plan will be implemented in practice and whether it can truly balance data centre growth with water and energy sustainability.
Policy Recommendations
Colombia is not yet fully prepared to manage the environmental impacts of rapid data centre expansion. To address this challenge, the following recommendations draw on Professor Mariana Mazzucato’s Entrepreneurial State framework. Unlike traditional models that view the state as a passive actor intervening only after market failures occur, this approach positions the state as a proactive market-shaper capable of driving more inclusive and sustainable outcomes.
The framework rests on four core principles: setting a clear direction for change; building the capabilities necessary to drive that change, including sectoral knowledge, risk-taking, and patient finance; adopting dynamic metrics that capture long-term and spillover effects; and ensuring the equitable sharing of both risks and rewards. The recommendations below are structured around these principles.
Set a Direction
Colombia’s National AI Policy sets a strategic direction by tasking the Ministry of Science, Technology, and Innovation with designing and implementing, between 2025 and 2026, a mechanism to accelerate data centre development with a focus on renewable energy and water-efficient technologies. To turn this vision into reality, the government should:
1. Urgently update water management plans at the national and city levels, particularly in areas expecting new facilities. Plans should integrate real-time data on water supply and demand to guide sustainable growth.
2. Embed environmental conditionalities in licensing and public procurement. These include requiring data centre developers to submit detailed technical specifications and projections for water and energy use and mandating the use of resource-efficient technologies accounting for the full spectrum of water use (Scope 1, 2, and 3).
Build Capabilities
To effectively manage data centre expansion and ensure environmental risks are mitigated, the government should invest in developing local expertise and infrastructure. This means not only regulating the sector but also equipping public institutions with the necessary knowledge and resources. To achieve this, the government should:
3. Strengthen public sector expertise by employing AI and data centre specialists, ensuring government agencies have the necessary knowledge to make informed decisions and effectively evaluate the technical information provided by data centre operators.
4. Establish co-investment mechanisms that leverage public, private, and multilateral funding to expand renewable energies and improve water management systems. This funding should be directed not only towards meeting data centre needs, but also to covering the water and energy requirements of low-income households.
5. Support national research and innovation to develop and test solutions for energy and water efficiency. This aligns with Action 3.1 of the National AI Policy, which funds AI-related R&D projects led by universities, research institutions, and private-sector partners.
Use Dynamic Metrics
To effectively manage the impact of data centres, it is essential to move beyond static metrics and ensure real-time data is continuously available. This will enable more responsive and effective resource management, especially during emergencies. To this end, the government should:
6. Install smart meters in data centres to provide real-time water and energy data to authorities. These data streams can feed into AI-powered tools for land use and resource planning that will be developed under Action 6.12 of the National AI Policy.
7. Develop a comprehensive monitoring and evaluation framework that tracks not only economic returns, but also spillover environmental and social impacts.
Share the Risks and the Rewards
With many data centre operators in Colombia being foreign owned, there is a risk that the sector could perpetuate an extractive model, whereby multinationals extract wealth and resources while the government and local communities bear the brunt of the environmental and social costs. To equitably share both the risks and the rewards of the data centre market, the government should:
8. Urgently increase industrial water prices to better reflect the true costs of water and the externalities its extraction imposes on the environment. Energy prices should also be adjusted to more accurately account for Scope 2 water consumption. The water and energy tariffs for data centre providers and other high-water use industries should remain flexible throughout the year, with the option to increase them during periods of water scarcity. This will help regulate demand while encouraging more efficient water use.
9. Integrate conditionalities in the licensing processes for data centres and related renewable energy projects, requiring a small percentage of their computing and energy capacity be reserved at a reduced cost for use by Colombian universities, research institutions, and private sector partners working on environmental projects.
10. Pilot geographical load balancing, distributing AI workloads across data centres based on regional water and energy availability, reducing pressure during emergencies.
These recommendations carry both risks and costs that must be carefully managed. One concern is a potential slowdown in Colombia’s data centre market. However, this does not mean halting growth. As noted, Chile’s market is expanding at an annual rate of 8.27%, while Colombia’s is growing at 17.6%. A modest slowdown is a reasonable trade-off if it ensures growth that is both sustainable and inclusive.
Another challenge is potential pushback from data centre operators who may resist stricter regulations or engage in lobbying. Yet Colombia holds a strong comparative advantage. The cost of building a data centre in Bogotá costs just US$7.10 per watt, compared with São Paulo (US$10.10), Querétaro (US$10.00), and Santiago (US$8.30). This gives the country room to implement higher standards without losing competitiveness. As Chile’s experience shows, clear environmental regulations can even encourage companies to adopt more efficient technologies.
From a financial perspective, while a detailed analysis is still needed, these recommendations appear economically advantageous. The government has already allocated US$111.5 million for the National AI Policy, which can support research, capacity building, and monitoring. Additional revenue from higher tariffs and licensing can be invested in strengthening water and energy infrastructure, while long-term environmental costs can be mitigated through effective conditionalities.
Conclusion
Colombia’s AI-driven data centre boom presents significant opportunities for digital growth, investment, and job creation. However, these benefits could be undermined without careful management of water and energy resources. The National AI Policy provides a strong foundation, but its success will depend on the state embracing an entrepreneurial role: setting a clear strategic direction, building the necessary capabilities, adopting dynamic metrics, and ensuring that risks and rewards are shared equitably.
If implemented effectively, Colombia can become a regional leader in developing a digital economy that is fast-growing, sustainable, and inclusive. Conversely, failure to act could provoke strong community opposition, as local water and energy resources may be perceived as being diverted to foreign-owned facilities. Transparent public engagement will be essential to build trust and demonstrate that policies protect both people and ecosystems.
Beyond water and energy, data centres also create other environmental and social pressures, including increased demand for critical minerals, greenhouse gas emissions, e-waste generation, and land-use competition. While these fall outside the scope of this blog, they merit further analysis to fully understand AI’s environmental impact in Colombia.
This paper is an adaptation of a paper submitted on 8 May 2025 for Latin American Economics (AMER0017), an elective course of IIPP’s MPA in Innovation, Public Policy, and Public Value.
The final piece of China’s cross-border data transfer framework has now been released with the issuance of the Certification Measures. Effective January 1, 2026, businesses must closely monitor certification institutions, standards, and application procedures. Early preparation and strategic planning will be essential for long-term compliance and risk management.
On October 14, 2025, the Cyberspace Administration of China (CAC) and the State Administration for Market Regulation (SAMR) jointly issued the long-awaited Measures for Certification of Cross-Border Personal Information Transfer (hereinafter referred to as the “Measures”), which will officially take effect on January 1, 2026.
The release of the Measures marks a pivotal moment in China’s data governance landscape. It completes the three-pathway framework for cross-border personal information transfers established under the Personal Information Protection Law (PIPL).
Find Business Support
With the certification method now fully defined, China’s regulatory architecture for cross-border data transfer (CBDT) is considered comprehensive and operational.
The Measures clarify key aspects of the certification process, including scope and applicability, application procedures, certification body obligations, as well as supervision and enforcement.
This final regulatory piece enhances legal certainty for businesses. It offers enterprises another structured compliance pathway and may prove especially beneficial for multinational corporations engaged in frequent or large-scale data transfers.
Explore vital economic, geographic, and regulatory insights for business investors, managers, or expats to navigate China’s business landscape. Our Online Business Guides offer explainer articles, news, useful tools, and videos from on-the-ground advisors who contribute to the Doing Business in China knowledge.
Start exploring
Certification in China’s CBDT framework
Under Article 38 of PIPL, personal information processors in China who need to transfer personal data overseas for business or operational purposes must choose one of three legally prescribed pathways:
Undergo a security assessment organized by the CAC, except where exempted by relevant laws and regulations.
Undergo a certification for cross-border personal information (PI) transfer by a professional institution in accordance with the regulations of the CAC.
Sign a standard contract with a foreign party stipulating the rights and obligations of each party in accordance with standards set by the CAC.
Meet other conditions set by the CAC or relevant laws and regulations.
Since 2022, China has gradually built out this framework through a series of regulatory instruments. The Security Assessment Measures, released in July 2022, laid out detailed procedures for high-risk data transfers. In February 2023, the Standard Contract Measures were issued and came into effect in June, providing a more accessible compliance route for many businesses.
Find Business Support
However, the certification pathway remained incomplete for some time. While several technical standards and guidelines were released – such as the Security Certification Specifications for Cross-Border Processing of Personal Information V2.0 (December 2022) and the Information Security Technology – Certification Requirements for Cross-Border Transmission of Personal Information (March 2023) – a formal regulatory document was still missing.
Now with the release of the Measures, this long-awaited document provides the legal and procedural foundation for certification, aligning with earlier standards and the CAC’s January 2025 draft for public consultation.
Moreover, China’s State Administration for Market Regulation (SAMR) and Standardization Administration of China (SAC) jointly released the Data security technology—Security certification requirements for cross-border processing activity of personal information (GB/T 46068-2025), which will take effect on March 1, 2026.
All these developments signal that the three pathways for CBDT under the PIPL are now fully operational, marking a significant milestone in China’s data governance regime.
Applying for a certification for CBDT
The newly released Measures for Certification of Cross-Border Personal Information Transfer outline the specific conditions under which personal information processors may opt for the certification pathway to legally transfer personal data overseas.
Scope of application
To be eligible for certification, a personal information processor must meet all of the following criteria:
It is not a critical information infrastructure operator (CIIO)
Transfer volume thresholds:
Since January 1 of the current year, it has transferred the PI of between 100,000 people and one million people out of China (excluding sensitive PI).
Since January 1 of the current year, it has transferred the “sensitive” PI of less than 10,000 people out of China.
The data being transferred must not include important data as defined by Chinese regulations.
Importantly, the Measures prohibit data volume splitting or other circumvention tactics to avoid the security assessment requirement. If the data transfer volume exceeds the thresholds for certification, the processor must undergo a security assessment instead.
Pre-application obligations
Before applying for certification, personal information processors must fulfill several legal obligations, including:
The PIPIA must evaluate:
The legitimacy, necessity, and scope of data processing by both the processor and the overseas recipient;
The volume, sensitivity, and type of data being transferred, and potential risks to national security, public interest, or individual rights;
The technical and organizational safeguards promised by the overseas recipient;
The risk of data breaches or misuse, and the effectiveness of redress mechanisms;
The legal environment of the recipient’s jurisdiction and its impact on data protection; and
Any other factors that may affect the security of the data transfer.
The PIPIA report should be retained for at least three years.
Also read: How to Conduct a Personal Information Protection Impact Assessment in China
Application process
Processors must apply for certification through a professional certification institution authorized to conduct personal information protection audits. For overseas processors, the application must be submitted via a designated domestic representative or entity.
Once the application is approved, the institution will issue a certificate, valid for three years. To maintain continuity, processors must reapply six months before the certificate expires.
Supervision and enforcement
The Measures establish a multi-layered oversight mechanism:
National and provincial CAC and market regulation authorities will conduct regular and ad hoc inspections of certification activities and outcomes.
If certified processors are found to pose significant risks or experience data security incidents, authorities may initiate interviews and require corrective actions.
Whistleblowers, including organizations and individuals, may report violations to certification bodies or regulators.
Violations of the Measures may result in penalties under the PIPL, Cybersecurity Law, and Certification and Accreditation Regulations, and may lead to criminal liability where applicable.
Comparing standard contract vs. certification
Under China’s CBDT mechanisms, both the standard contract and certification pathways provide legal mechanisms for cross-border transfers of personal information. While they share many similarities, such as overlapping applicability and similar pre-transfer obligations, their structural differences make them suitable for distinct business scenarios.
The standard contract is a self-managed process in which enterprises sign a fixed-format agreement with the overseas recipient, strictly following the CAC’s template. After conducting a self-assessment, the enterprise submits the contract and related materials for filing with the provincial CAC, which may conduct formal or substantive reviews. As a commercial agreement, the standard contract is not publicly disclosed, and its contents are not subject to public scrutiny or external evaluation.
In contrast, certification is conducted by third-party professional institutions based on CAC-issued rules. It involves a comprehensive review of the enterprise’s technical, organizational, and governance measures. Unlike the standard contract, certification carries a degree of public authority – it reflects, to some extent, administrative recognition of the enterprise’s data protection capabilities. For companies with high reputational stakes in personal information protection, certification offers a credible external endorsement of their compliance posture.
Find Business Support
The compliance focus also differs. The standard contract emphasizes the legal obligations of a specific data transfer, ensuring the overseas recipient agrees to uphold data subject rights and assumes clear data protection responsibilities. Certification, on the other hand, assesses the enterprise’s overall compliance framework, including internal governance, data protection systems, and technical safeguards. It promotes ongoing compliance and dynamic supervision. In practice, the standard contract is more suited to “one-off” or occasional transfers, while certification is better aligned with enterprises engaged in frequent or long-term cross-border data activities.
Post-transfer supervision further distinguishes the two. With the standard contract, enterprises are responsible for monitoring the overseas recipient’s compliance, and CAC oversight is primarily conducted through the filing system. Certification bodies, however, implement continuous monitoring mechanisms. Certificates may be suspended or revoked, and violations are publicly disclosed, creating external compliance pressure.
Given these differences, the standard contract is generally more appropriate for low-volume, low-risk transfers. It is relatively easy to implement but offers limited flexibility due to its fixed format. Certification, by contrast, is better suited for enterprises with frequent, high-risk, or high-profile data transfers, or those seeking to demonstrate a high level of data protection. The certification is valid for three years, helping reduce repetitive compliance efforts.
Enterprises should assess their business scale, data export scenarios, and compliance capacity to select the most appropriate pathway for CBDT.
Aspect
Standard Contract
Certification
Legal nature and review mechanism
Enterprises sign a fixed-format agreement with the overseas recipient, following CAC’s template.
Self-assessment and filing with provincial CAC; subject to formal or substantive review.
Conducted by third-party institutions under CAC rules. Reviews technical, organizational, and governance measures. Carries independent credibility.
Compliance focus and depth
Focuses on contractual obligations. Ensures overseas recipient upholds data subject rights and assumes data protection responsibilities.
Evaluates full compliance framework, including governance, systems, and safeguards.
Emphasizes ongoing and dynamic compliance.
Post-transfer supervision and accountability
Enterprises monitor recipients’ compliance. CAC oversight via the filing system.
Certification bodies provide continuous monitoring (e.g., annual inspection). Certificates may be suspended or revoked; violations are publicly disclosed.
Applicability and flexibility
Best for low-volume, low-risk transfers. Easy to implement, but limited flexibility due to fixed format.
Suited for frequent or high-risk transfers. Valid for three years, reducing repetitive compliance.
Cost
Self-declaration will not incur any cost
The certification body will charge relevant fees.
What to watch next?
As the Measuresprepare to take effect on January 1, 2026, enterprises should go beyond understanding the basic provisions and actively monitor several key developments to ensure compliant and efficient implementation.
Key areas to monitor include:
Certification institutions and qualifications: Certification must be conducted by professional institutions that have obtained official qualifications for personal information protection certification. While the list of authorized institutions has not yet been released, it is expected to be jointly determined by the SAMR and CAC.
Certification standards and technical specifications: The Measures indicate that the CAC, together with relevant data governance authorities, will formulate detailed certification standards and technical specifications, while the market regulation authorities will define certification rules, certificate formats, and usage guidelines.
Application procedures and operational details: While the Measures outline the general process, many practical details remain to be clarified or will be defined by certification institutions, such as required documentation beyond the PIPIA report (such as data inventory, agreements, internal governance policies), online/offline application channels, review timelines and feedback mechanisms, proof of compliance capability for overseas recipients, and so on.
Companies are suggested to consult the official websites of the CAC, SAMR, and the Certification and Accreditation Administration of China (CNCA) for timely updates and implementation guidance.
Key takeaway
Personal information protection certification is a critical mechanism for enabling compliant, trustworthy, and internationally aligned data flows. As the Measures enter into force in 2026, enterprises must pay close attention to the qualifications of certification institutions, evolving standards and procedures, and long-term compliance obligations. For legal, compliance, and information security professionals, now is the time to study regulatory trends, plan ahead, and strengthen internal capabilities to ensure a smooth transition into the new regime.
If you need further interpretation or practical support, our team at Dezan Shira & Associates is here to assist you with tailored guidance and hands-on expertise. For more information, please get in touch with China@dezshira.com.
Prince Andrew was paid tens of thousands of pounds by a British businessman linked to a wealth management company which ripped off pension savers.
The King’s brother, who last week said he would no longer use his titles including the Duke of York, has long faced questions about his finances and how he is funding his lifestyle.
He stepped down as a working royal in 2019 because of his association with the sex offender Jeffrey Epstein and no longer receives any money from the King.
Although his finances remain opaque, details of some of his arrangements and controversial business associates have occasionally emerged from court cases.
Prince Andrew did not respond to requests for comment.
Documents from the High Court in London show that Andrew received £60,500 from a British businessman, Adrian Gleave in December 2019, a few weeks after the BBC Newsnight interview which led to his withdrawal from public life.
The payments came to light in a High Court case brought by an elderly Turkish millionaire, Nebahat Isbilen, who claimed money she had paid to Andrew and his ex-wife Sarah Ferguson had been misappropriated by a business adviser.
That money was funnelled through a British company owned by Mr Gleave, Alphabet Capital Limited.
According to an “agreed statement of facts signed by or on behalf of the Duke and Duchess, Mr Gleave and Alphabet”, Mr Gleave’s company Alphabet had “previously made, and might in the future make, substantial payments to HRH Prince Andrew the Duke of York”.
The payments Prince Andrew received directly from Mr Gleave and his businesses, which were also sent via Alphabet Capital, came months after the businessman had stepped down as a director of SVS Securities – a company which had been ordered to stop trading by the financial regulator over pension mis-selling allegations.
SVS Securities collapsed in August 2019, days after the Financial Conduct Authority (FCA) had ordered it to stop regulated activities.
Clients’ pension funds were found to have been invested in high-risk bonds against their interests in order to generate large commissions for SVS.
Some investments made on the basis of the undisclosed commissions then defaulted, leaving customers with substantial losses. Investors were also charged large fees to withdraw funds in an effort to boost profits, the FCA found.
Mr Gleave, 55, was head of business development at SVS, which he had joined in 2013.
He remained a registered director on the FCA’s register until late July 2019, less than two weeks before the regulator’s intervention.
He had been a company director registered with Companies House until a couple of months earlier, although on LinkedIn he claims to have left the business in November 2018.
Three SVS directors were later banned and fined by the FCA but Mr Gleave was not one of them. Two are appealing the decision.
The Financial Services Compensation Scheme has paid out more than £41m to former SVS customers.
At the time of his dealings with Prince Andrew, Mr Gleave also ran a number of caravan and mobile home parks in Northern Ireland and England.
At one point, he was reported to have worked out of one of the parks, a retirement village for over-55s on the east coast of Northern Ireland.
Ten of the parks have since gone into administration and Mr Gleave, who did not respond to a request for comment, now works for a renewable energy company with a focus on AI and crypto financing.
Neither Prince Andrew nor Mr Gleave have ever explained the reason for the payments or the nature of any contractual relationship between the two men.
Baroness Margaret Hodge, a former chair of the Commons public accounts committee, said Mr Gleave’s business background raised questions for Prince Andrew about his judgement and financial dealings.
“This is yet another instance where a dose of transparency would help answer legitimate questions about the origins of the money and the purpose of the payment,” she said.
“Without those answers any sceptical person would be worried that there might be some financial wrongdoing taking place and this would risk sullying the reputation of the Royal family,” she added.
As well as the payments he made himself, Mr Gleave’s company, Alphabet Capital, was also used to funnel significant sums which had originated from Ms Isbilen to Prince Andrew and his ex-wife Sarah, court documents show.
Ms Ferguson was paid £50,000 by Alphabet Capital in February 2020. It has previously been reported that she was paid £20,000 by Alphabet for a role advising the company and that she also received more than £200,000 to cover work she had done as a brand ambassador for a US solar energy company.
Prince Andrew was separately given £750,000 directly by Ms Isbilen, money which he has repaid.
Another £10,000 was paid from Alphabet Capital to the couple’s daughter, Eugenie. This, along with a £15,000 payment from Ms Isbilen’s business adviser has previously been described by Eugenie as a gift from a “long-standing family friend” which she said was to pay for a surprise birthday party for her mother, Sarah.
Alphabet Capital filed accounts claiming it was a dormant company at the time of the payments. These were later corrected but listed a turnover of just £80,000.
Prince Andrew and Mr Gleave did not respond to requests for comment.
The Human RNome Project aims to ensure consistent and reproducible outcomes in RNA sequencing and modification studies by utilizing standardized cell lines maintained under uniform culture conditions. This standardized approach will facilitate meaningful comparisons across technologies and laboratories. The selected cell lines will be widely accessible, easy to maintain in culture, and highly proliferative, ensuring an adequate supply of RNA for sequencing and characterization experiments. Importantly, these cell lines will exhibit genetic stability, characterized by a well-defined genome with minimal mutations and chromosomal aberrations, to guarantee the reliability and robustness of the generated data.
To maintain genomic integrity, cell lines will be sourced from certified distributors at regular intervals and used at low passage numbers (< 8). Genetic integrity will be independently verified through DNA and cDNA sequencing, with results reported alongside direct RNA sequence data. This ensures that any genomic drift is identified and accounted for in downstream analyses.
Table 1 lists cell lines that meet these criteria. These lines have been extensively characterized by large-scale studies such as the ENCODE Project [25, 26] and the 1000 Genomes Project [17, 18]. For instance, GM12878, a cultured B-cell line from a female donor with ancestry from Northern and Western Europe, has been sequenced as part of the 1000 Genomes Project and characterized by ENCODE. IMR-90 lung fibroblasts, BJ foreskin fibroblasts, and H9 human embryonic stem cells are similarly well-characterized and available through trusted sources like Coriell, ATCC, and WiCell, which will also enforce standardized protocols for culturing and handling. Given the sensitivity of RNA to environmental factors, these standardizations are critical for ensuring data comparability. Repositories will also require users to follow consistent protocols for culturing and RNA extraction, as variations in these processes could influence RNA sequence and modification profiles.
Table 1 Cell lines for initial steps of the Human RNome Project
RNA extraction and quality control: RNA will be extracted using a guanidinium thiocyanate-based method to ensure high purity and integrity. RNA quality will be assessed by absorbance ratio (260/280 and 260/230 nm) and capillary electrophoresis (e.g., Agilent TapeStation), requiring a minimum RNA Integrity Number (RIN) of 9 for RNA extracted from cell lines (as the project advances, and RNA samples are extracted from tissues, a lower RIN threshold such as 8 may be necessary). Aliquots of RNA will be archived for validation and further analyses.
Initial RNA targets for sequencing: The pilot phase of the Human RNome Project will focus on sequencing transfer RNA (tRNA), ribosomal RNA (rRNA), and mRNA, with a focus on selected protein-coding transcripts. These RNA classes are ideal initial targets due to their ubiquity, existing knowledge of their modification profiles, and robust expression across cell types.
tRNA and rRNA
tRNA (~ 250 expressed isodecoders) and rRNA (5S, 5.8S, 18S, 28S) are universally expressed and highly conserved, with well-studied modification types and locations [12,13,14]. Table 2 lists examples of modifications typically found in human mRNA, tRNA, and rRNAs. Both total tRNA and rRNAs can be purified from total RNA using electrophoresis or size-exclusion chromatography [28], while affinity-based methods such as chaplet chromatography [29] or reciprocal circulating chromatography [30] can be used to enrich for specific tRNA sequences. One drawback of all RNA purification methods is co-purification of non-target RNAs due to similar size or hybridization to target RNAs. Mass spectrometric analysis of modified ribonucleosides in purified RNA must always be viewed with suspicion for modifications found in multiple forms of RNA (e.g., m6A, m5C).
Table 2 Known modifications in rRNAs and tRNAs as internal validations
Coding genes and their mRNAs
Selected protein-coding genes include ACTB, CDKN2A, ISG15, and SOD1. These genes were chosen based on their known association with diseases, moderate to high expression levels, relatively short transcript lengths (~ 1 kb), and known modifications. For example, SOD1 is associated with amyotrophic lateral sclerosis [31], while ACTB is widely expressed and associated with dystonia (Table 3).
Table 3 Protein-coding genes selected as representative RNAs to begin direct RNA sequencing
Coding RNA enrichment methods: To detect low-abundance modifications, enriched RNA samples are critical. Initial poly-A RNA enrichment can be achieved using oligo-dT kits from various vendors [33]. For specific RNAs, biotinylated antisense oligonucleotides allow ~ fivefold enrichment [34], while microbead-based antisense oligos are claimed to achieve a 100,000-fold enrichment [35]. DNA nanoswitches offer another option, with ~ 75% recovery and purities exceeding 99.8% for RNA ranging from 22 to 400 nts [36].
Future goals
Short-term
1.
Standardized RNA extraction using guanidinium thiocyanate.
2.
Enrichment of test RNAs using antisense-based methods.
3.
Mass spectrometry-based direct RNA-seq for short-read identification of modifications and nanopore sequencing for long-read sequencing and modification mapping.
Medium-term
1.
Sequence transcriptomes from cell sorting-enriched samples of defined cell types.
2.
Compare data with existing programs (e.g., GTEx).
3.
Expand sequencing to include different cell types and tissues from individuals of all ages and ethnicities.
Long-term
1.
Sequence RNAs from specific subcellular regions (e.g., nucleus, cytoplasm, mitochondria).
2.
Integrate single-cell transcriptomic and subcellular data.
Source molecular resources and standards
The Human RNome Project relies on robust molecular resources and chemical standards to develop and validate sequencing and mass spectrometry (MS) technologies. These resources encompass synthetic and native RNA standards, as well as their building blocks, such as ribonucleosides, ribonucleotide triphosphates (NTPs), and oligoribonucleotides. High-quality standards are essential for ensuring accurate analysis of RNA modifications, their chemistry, and their precise locations within RNA molecules.
Chemical standards
Chemical standards are indispensable for training and validating analytical methods before analyzing native RNA samples. They ensure reproducibility, correct identification of RNA modifications, and calibration of detection systems. Standards are summarized in Fig. 2 and include the following.
Fig. 2
Overview of types of chemical standards needed for the Human RNome Project
Ribonucleosides and dinucleotide caps
Chemical standards for individual ribonucleosides are essential for characterizing RNA modifications and quantifying their abundance. Approximately 90 ribonucleoside standards are commercially available, with additional variants synthesized by academic laboratories. Comprehensive lists of vendors are provided on the RNome website [33], while PubChem offers detailed vendor information and links to chemical resources. Prices for these standards range from $20 to $1500 per milligram, with custom synthesis for rare modifications costing between $10,000 and $20,000. For qualitative analysis, 1 mg of a standard is typically sufficient. For quantitative analysis, we recommend assessing the purity of the standard by quantitative NMR prior to preparing calibration solutions for, as an example, LC–MS analysis. Despite the availability of over 90 modified ribonucleosides, many human-specific RNA modifications remain inaccessible as commercial standards. Furthermore, the chemical stability (shelf life) of ribonucleosides is not well-documented. For example, m1A undergoes Dimroth rearrangement to m6A during RNA processing and storage in aqueous solution [37, 38] highlighting the need for further research into ribonucleoside stability.
Ribonucleotide triphosphates
Ribonucleotide triphosphates (NTPs) are essential for in vitro transcription to synthesize RNA molecules longer than 20 nucleotides with defined modification profiles. Canonical NTPs are widely available from commercial sources, including isotopically labeled variants, while modified NTPs for specific ribonucleosides can also be obtained. However, these modified NTPs require rigorous verification of their chemical identity and purity, typically through techniques such as thin-layer chromatography (TLC) or LC–MS [39, 40]. In vitro transcription allows random, but not site-specific incorporation of modified NTPs [41].
Synthetic oligonucleotides and phosphoramidites
Site-specifically labeled RNA oligonucleotides, ranging from 5 to > 60 nucleotides, are essential for training nanopore base callers and validating LC–MS methods. Solid-phase chemical synthesis is commonly used to produce labeled oligonucleotides and vendors typically provide mass spectra to confirm the overall product length, failure sequences, and impurities. However, comprehensive validation, such as mass spectrometric sequence verification and ribonucleoside LC–MS for modification identification, is rarely included but essential for robust validation. To ensure accuracy, researchers must advocate for detailed validation data, including MS sequence validation and ribonucleoside-specific quantification, alongside the standard mass spectra provided by vendors. Despite these advancements, the site-specific incorporation of modifications into long RNA sequences (> 60 nucleotides) remains a significant challenge [42]. Current approaches, which involve combining chemical synthesis, transcription, and ligation, are labor-intensive, low yielding, and not easily scalable. New approaches to long RNA synthesis are needed to facilitate the generation of site-specifically modified RNAs that mimic biological molecules.
Future goals
Short-term
1.
Stability data for modified ribonucleosides is scarce, highlighting the need for systematic studies on shelf life.
Medium-term
1.
Consistent preparation, validation, and distribution protocols are essential to ensure data comparability over time. Quality control samples must be maintained and shipped with detailed documentation.
2.
Researchers should demand comprehensive validation data (e.g., MS/MS, sequence confirmation) from vendors to avoid errors in downstream analyses.
Long-term
1.
Sequencing and MS methods must be regularly validated using both synthetic and native standards
2.
High-quality library of modifications with comprehensive validation data and shelf-lives. Many RNA modifications lack synthetic standards, necessitating collaboration with organic chemists for their production.
Develop advanced sequencing technologies
Sequencing technologies will be pivotal to the Human RNome Project, much like they were for the Human Genome Project. To evaluate the potential impact on the project, it is essential to analyze the current state and project developments over the next 5 to 10 years. The consortium has hence identified and discussed lead questions that concern the type of currently available sequencing technologies, the necessary developments in the near future, and critical quality controls.
Current state of sequencing technologies
Current methods to map modifications can be classified into direct, such as mass spectrometry or direct RNA sequencing, and indirect, which usually relies on sequencing by synthesis, wherein RNA is converted to cDNA via reverse transcriptase [16]. Both indirect and direct RNA sequencing methods require additional steps to assign modifications. This section is meant as a brief summary and not a comprehensive review of all current variations and developments (for a comprehensive review please refer to the Report by the National Academies of Sciences, Engineering and Medicine [23]).
cDNA-based sequencing
Sequencing of cDNA, acquired through reverse transcription of RNA and analyzed through Illumina (and sometimes PacBio or Nanopore), is currently the most widely used form for indirect RNA sequencing. However, it cannot directly detect non-canonical ribonucleotides. Workarounds to map modifications rely on changing the RNA or cDNA product on a molecular level (“molecular input”) and include reverse transcriptase-based error profiling, chemical or enzymatic derivatization, and modification-specific immunoprecipitation [43, 44]. Molecular input methods utilize computational algorithms that infer RNA modifications from misincorporations, gaps, or reverse transcription arrests or reverse transcription incorporation of structurally similar bases. While powerful, no single molecular input method can comprehensively identify all modifications, necessitating the use of multiple techniques on the same RNA sample.
Direct RNA sequencing
Oxford Nanopore Technologies is the only widely available platform currently providing protocols for direct, long-read sequencing of RNA molecules, eliminating the need for cDNA conversion and preserving endogenous or synthetic exogenous RNA modifications (Fig. 3) [45,46,47,48]. Advances in machine learning models have led to more accurate basecalling and lower error rates for sequencing full-length native RNA transcripts [47]. By analyzing unique changes in electrical currents from the direct RNA sequencing process, RNA modifications can be tentatively identified [48]. Identification of modified nucleotide residues can be achieved by comparison against unmodified control samples [49], with base-calling algorithms or supervised models that have been trained on data with known modifications [47, 48, 50]. The training of such models can be achieved using data from cDNA-based approaches, data from modification-free control samples [49], in vitro transcription-generated data, or data from synthetic RNAs. However, the generation and availability of such data and the lack of RNA modification standards still limit the number of modifications that can currently be confidently detected and identified. Furthermore, not all reads from direct RNA-seq correspond to full-length RNAs and challenges remain to detect RNA modifications that occur at the 5′ ends of RNA molecules. To overcome these barriers, researchers are actively developing “molecular input” approaches—such as introducing chemical or enzymatic treatments which change the RNA molecule—to amplify or clarify the signals associated with RNA modifications [51].
Fig. 3
Overview of the sequencing workflow that will allow end-to-end sequencing of RNA including its modifications
Mass spectrometry (MS)
Mass spectrometry (MS)-based RNA sequencing is an essential complement to these efforts as a means to chemically identify and accurately quantify specific modifications [52,53,54,55]. Unlike the chemically nonspecific interpretation of electrical signals in nanopore sequencing, MS-based sequencing involves high mass accuracy (i.e., exact molecular weight) determinations of modification fragments that allow structural identification of the modification, its location in the RNA sequence at single-nucleotide resolution, and its abundance in the population of RNA sequences. While MS sequencing requires larger quantities of RNA than NGS or nanopore sequencing, advances in sensitivity have moved the application from more abundant non-coding RNAs to mRNAs [55,56,57,58]. The major limitation of MS-based RNA-seq is the short fragment size needed for accurate MS analysis, typically 10–60 nt in length depending upon the mass resolution of the instrument [55]. This precludes mapping modifications in long native RNA molecules, as can be achieved with nanopore. MS-based RNA sequencing and nanopore sequencing are thus complementary tools for RNome analysis.
Quality control: ensuring the accuracy of modification-aware sequencing and analysis
The accuracy of epitranscriptomic analysis is determined by the combined impact of errors introduced during experimental procedures and data processing. To ensure high-quality data, experimental design must include an adequate number of replicates, sufficient sequencing depth, and the incorporation of both positive and negative controls. Method-specific data analysis should employ robust statistical frameworks to evaluate the significance of signals at specific sites, accounting for sample size, signal strength, and their relevance within the broader context of all samples, including replicates and controls. Given the diversity of current modification mapping methods, it is challenging to recommend a universal set of parameters for experimental design and data analysis. Therefore, we outline guidelines based on general principles in the following sections.
Conventional sequencing errors and “molecular input” errors
Base-calling accuracy in Illumina sequencing data typically has an error rate of 0.1–0.5% per nucleotide residue, while nanopore sequencing has only recently reduced its error rates to the single-digit range. While these error rates are not typically a major concern for conventional RNA sequencing, they become critical when using molecular input methods that depend on errors for mapping modifications, as these methods can introduce artifact-based errors, such as false positives and false negatives. To ensure data validity and reliability, it is essential to include a sufficient number of both biological and technical replicates, as well as adequate sequencing depth to optimize the signal-to-noise ratio. The significance of a detected signal is further strengthened by comparisons with positive and negative controls, ideally including at least one of each that represents a “gold standard” or ground truth.
Data interpretation
Quality control (QC) parameters are essential at multiple levels, including raw data (e.g., fastq files used for downstream analysis) and the analytical pipelines used for mapping modified residues. For raw data, QC criteria can often follow established standards for the respective sequencing technology, such as a Q-score > 30 for Illumina sequencing. The thresholds, however, may vary depending on whether short-read or long-read sequencing technologies are employed. Beyond this, a second layer of QC is needed to evaluate the performance of molecular input methods, which introduce their own characteristic errors. A third layer of QC pertains to computational analysis, assessing the reliability of data interpretation across different epitranscriptomics mapping protocols and pipelines. In some cases, it may be valuable to integrate these QC layers into aggregated error rates or composite metrics that encompass both molecular and computational aspects.
To advance the field, it is imperative to establish a universally accepted set of QC parameters for benchmarking methods. Equally important is the determination of standardized threshold values for these parameters, which could become mandatory for the Human RNome Project. The diversity of existing technologies, as well as those that will emerge during the project, complicates the establishment of universal QC criteria at the raw data level. However, any method must undergo rigorous validation before being deemed suitable for modification calling.
Validation should involve the creation of models evaluated with metrics such as receiver operating characteristic (ROC) curves, area under the curve (AUC), sensitivity (true positive rate), and specificity (true negative rate). Particular attention must be given to minimizing false positive and false negative rates, as these directly impact the reliability of modification detection. Another critical input parameter for these models is an accurate estimate of the expected number of residues for a given modification, as this will influence thresholds for modification calling. Such an integrative and standardized approach to QC will ensure robust and reliable results across diverse epitranscriptomic applications.
Establishing clear guidelines for reporting QC metrics in publications and data repositories is essential for fostering reproducibility and confidence in results. Comprehensive reporting of raw data quality, molecular input performance, and computational reliability will enable consistent practices across studies. Such transparency not only ensures accountability but also facilitates meta-analyses and comparisons, accelerating progress in the field.
Vision 2025: strategic steps for the next decade
Advancing modification-aware RNA sequencing on an international scale requires both organizational and technical developments. One of the greatest challenges will be achieving consensus within the field on a mandatory set of QC parameters and, even more challenging, establishing universally applicable threshold values. As highlighted earlier, in addition to maintaining a continuously updated overview of methodologies, the field must identify techniques that either deliver the highest throughput with minimal error rates or enable precise quantification of modification levels at specific RNA sites. With these considerations in mind, we outline the following ongoing and future objectives for the Human RNome Project.
Future goals
Short-term
1.
Continue developing NGS, nanopore, and MS technologies to (a) expand the repertoire of modifications for NGS and nanopore by developing and refining chemical derivatization methods; (b) expand training datasets and algorithms for nanopore; and (c) increase the sensitivity, LC resolution, and data processing algorithms for MS-based sequencing.
2.
Integrate orthogonal technologies (e.g., combinations of methods providing different molecular inputs or alternate sequencing technologies) to confirm RNA modifications with high confidence on native RNAs.
3.
Develop and implement robust quality control (QC) protocols for NGS, nanopore, and MS to (a) minimize artifacts, (b) increase statistical power, (c) increase sequencing depth, and (d) assure inter-laboratory consistency.
4.
Lay the groundwork for scaling and throughput: (a) multiplexing MS-based sequencing; (b) automation of sequencing library preparation, sample analysis, data processing, and data mining; and (c) inter-laboratory validation.
5.
Create user groups to develop, implement, and cross-validate RNA-seq methods. Begin developing or adapting websites and databases for public access to protocols and RNA-seq datasets. Engage international funding bodies to support research and development.
Medium-term
1.
Develop automated systems for RNA extraction, size- or sequence-based RNA purification, library preparation, sequencing, data processing, and data analysis.
2.
Develop and refine computational methods: (a) algorithms to interpret raw sequencing data, distinguish true modification signals from noise, and quantify modifications; (b) standardize modification-calling pipelines with open datasets; and (c) develop rigorous benchmarks to ensure reproducibility.
3.
Expand scale of sequencing efforts: Prioritize high-throughput, automated solutions to handle increasing data demands; integrate methods with lower error rates and reliable quantification into streamlined workflows.
4.
Expand RNA-seq analyses using cell and RNA targets identified in section I. Apply improved workflows to diverse cell types and RNA populations to create comprehensive modification maps.
Long-term
1.
Develop new sequencing technologies: (a) design new nanopore pore systems, (b) develop RNA-customized MS ionization, fragmentation, and detection hardware; (c) innovate platforms capable of directly sequencing full-length RNA molecules with single-base resolution and error rates < 0.1%.
2.
Continue developing AI and automation technologies: Design AI-driven base-calling algorithms for real-time error correction and precise modification detection.
3.
Expand RNA-seq databases and integrate across databases.
4.
Expand application of RNA-seq technologies: (a) cells beyond those initially identified as standards for the Human RNome Project (Section I); (b) tissues from animal models; and (c) human clinical samples.
Guidelines for data analysis, formatting, and storage
A tremendous amount of the epitranscriptome sequencing data generated in the last few years has, on the whole, remained unused, because of limited data accessibility, poor findability and reusability. Addressing these gaps could significantly enhance the utility and impact of these data. In this section, we propose FAIR guidelines [59] for data format specifications, model training standards, and protocols for recording and sharing information related to RNA sequences and modifications, applicable to both indirect and nanopore direct RNA sequencing (Fig. 4).
Fig. 4
Data handling to ensure long-term and reproducible usage of the data acquired for the Human RNome Project
Data format specifications and nomenclature
The identity, position, and frequency of RNA modifications are derived from large volumes of raw data, typically mapped reads. These analyses depend on method-specific technological expertise, which can vary significantly across approaches. Raw data alone are often not practical when only site-specific modification information is required. Moreover, this information has historically been disseminated through a range of incompatible formats, governed by varying standards, and often accompanied by limited access or incomplete metadata. These challenges have hindered reproducibility and the ability to compare results across studies.
While data formats for raw sequencing data are well-established, no such standardization exists for modification information derived from these data. At a minimum, site-specific RNA modification data should be reported in a straightforward format and include:
1.
A standardized naming convention for the modification type.
2.
Stoichiometric information, such as the percentage or frequency of the modification.
3.
Depth of coverage for the modification site.
4.
Quantitative confidence scores indicating the reliability of the modification call.
At the dataset level, metadata should be sufficiently detailed to ensure traceability, reproducibility, and reusability. This requires reliance on standardized nomenclature while maintaining flexibility to include free-text information where necessary. Some of this information has recently been incorporated into the latest SAM/BAM format specifications [60], where nucleotide residue modifications and their quality scores are recorded per-read.
At the per-site level, the recently proposed bedRMod format addresses many of these requirements [32]. This format is analogous to the ENCODE bedMethyl standard [61] and nanopore’s extended bedMethyl format [62] and compatible with the widely used BED (Browser Extensible Data) format. It was developed during the Human Genome Project [60] and approved by the GA4GH Standards Steering Committee and it integrates seamlessly with many command-line tools and genome browsers.
However, a significant barrier to the widespread adoption of the bedRMod format is its dependency on information of nucleotide residue modification from SAM/BAM files, which in turn relies on mapping algorithms. Tools to compile this data into bedRMod format at the site level remain underdeveloped, and current workflows often rely on custom algorithms to extract site-specific information into similar tabulated formats. Addressing this gap with robust, standardized tools will be essential to advancing the use and utility of bedRMod for RNA modification studies.
The RNA modification nomenclature adopted by MODOMICS [12, 13] aligns well with the requirements of the bedRMod format by providing standardized names for RNA modifications. MODOMICS uses a variety of representations, including multi-character alphanumeric codes for single and multiple sequence formats like FASTA, a one-letter Unicode-based code for sequence alignments, and a human-readable alphanumeric code for broader accessibility. Recent updates to MODOMICS have expanded its nomenclature to include synthetic residues, accommodating the growing diversity of RNA modifications in both research and practical applications [12]. This system is instrumental in ensuring compatibility across tools and datasets and holds potential as a foundation for a future standardized nomenclature under IUPAC guidelines.
Data requirements for training, validation, and testing of RNA modification
Development of RNA modification calling algorithms requires independent datasets for model training, (cross-) validation, and additional datasets for testing and benchmarking established methods. The previous sections have outlined potential biological and synthetic sources, as well as sequencing approaches, for generating these datasets. To ensure relevance, datasets may need to align with the specific focus of interest, whether tRNAs, mRNAs, or rRNAs, as these classes differ significantly in their epitranscriptomic properties and sequence characteristics.
Additionally, datasets should accurately represent the real-world distribution of modified versus unmodified nucleotides the model is expected to encounter. They must approximate the size and complexity of existing transcriptomes and include high-quality annotations for modification classes. For example, modification-free transcripts from in vitro transcription of cDNA derived from six immortalized human cell lines have been used as a robust ground-truth dataset for unmodified mRNA transcriptomes [49]. Similarly, Chan et al. [63] employed random ligation of RNA oligos with known modification statuses to construct longer transcripts with sufficient complexity in both nucleotide composition and modification density, representing another valuable resource for RNA modification research.
Towards a central RNA modification database
The Human RNome Project seeks to establish guidelines for formatting and sharing RNA sequences and modifications, while also consolidating and integrating the growing volume of high-throughput epitranscriptome data. This effort aims to enhance data accessibility, facilitate the automated discovery of datasets, and optimize data reuse. Sci-ModoM [32] introduces a novel, quantitative framework supported by the bedRMod format, advancing the adoption of FAIR data principles and fostering the use of common standards. These features position Sci-ModoM as a potential cornerstone database for RNA modifications.
Developed in synergy with MODOMICS [12, 13], Sci-ModoM complements this meta-database by offering high-throughput, high-resolution data in a standardized format. Sci-ModoM serves as a centralized platform where modifications from diverse studies can be accessed and compared, while MODOMICS provides a curated repository of RNA sequences enriched with all known modifications, along with detailed metadata on their reliability and prevalence. Together, these resources enable the visualization of modifications within RNA sequences and broaden the utility of epitranscriptome data for research, therapeutic development, and experimental applications. The integration of Sci-ModoM and MODOMICS represents a significant milestone in achieving comprehensive annotation and effective utilization of RNA modifications.
Future goals
Short-term
1.
To establish bedRMod as the format for sharing RNA modification data, and to develop the necessary infrastructure and tools to improve interoperability, and to facilitate its use by the community.
2.
To establish guidelines and minimal requirements for training, validation, testing, and sharing RNA modification data and software.
3.
To make realistic training and validation data available through Sci-ModoM to support the development of new detection methods and algorithms (cf. Future goals of section III).
Medium-term
1.
To continuously and dynamically annotate novel modifications from the large amount of data available in Sci-ModoM using MODOMICS evidence levels, reliability scores, and prevalence metrics.
2.
To enhance the interpretability of transcriptome-wide data accumulated in Sci-ModoM by contextualizing it with broader biochemical, structural, and functional information available in MODOMICS, bridging experimental findings with mechanistic insights.
3.
To provide global mirroring and public access, e.g., through collaboration with academic institutions, following the open-access model of Sci-ModoM.
Long-term
1.
To build on the synergistic development of Sci-ModoM and MODOMICS to establish a virtual central RNA modification database.
2.
To establish a standardized data flow allowing users to transition seamlessly from experimental data in Sci-ModoM to comprehensive annotations in MODOMICS.
The transformative impact of RNA science
The Human RNome Project is a bold and transformative initiative poised to revolutionize diverse sectors, including biomedicine, agriculture, data storage, and global security. By advancing RNA science, this project will deepen our understanding of RNA biology and catalyze groundbreaking innovations, delivering profound societal benefits.
Biomedicine: RNA research has made critical contributions to healthcare, particularly in understanding the biology of RNA viruses like SARS-CoV-2 and other infectious agents. These insights have accelerated the development of RNA-based therapeutics, including mRNA technologies now being adapted for applications such as influenza [64, 65] and malaria prevention [66]. Beyond infectious diseases, RNA-based therapies are transforming treatment paradigms for various conditions. Nusinersen, an antisense oligonucleotide therapy, has significantly improved outcomes for children with spinal muscular atrophy, enabling them to achieve developmental milestones [67, 68]. Inclisiran, an RNA interference-based drug, provides an effective biannual treatment for lowering LDL cholesterol, improving compliance compared to daily regimens [69, 70]. RNA-based therapies continue to advance in oncology, rare diseases, and other fields. The Human RNome Project will support this progress by improving targeting with highly accurate RNA sequences, reducing costs through the production of affordable, high-quality ribonucleotides, including modified forms, bolstering supply chains, and expanding access to RNA therapeutics.
Agriculture: Global food insecurity is a pressing issue affecting millions worldwide. In the USA alone, over 10 million children face hunger, and globally, malnutrition affected 27 million children in 2022. RNA-based technologies offer innovative solutions to address these challenges. Research shows that RNA modifications can enhance crop yields in staples like rice and potatoes [71], improving resilience and productivity. Additionally, RNA interference (RNAi) delivered via high-pressure sprays provides an effective, non-genetically engineered method to combat plant diseases [72, 73]. RNA sequencing technologies will equip plant scientists with powerful tools to improve crop productivity and combat global hunger.
Data storage: RNA presents a transformative approach to ultra-dense, efficient, and scalable data storage, capitalizing on its structural complexity and extensive chemical diversity. Unlike the binary 0,1 system traditionally used for data encoding, RNA’s repertoire of approximately 180 known ribonucleotide modifications vastly expands the encoding alphabet. While the binary system encodes 1 bit of information per symbol, RNA modifications encode approximately 7.49 bits per symbol, enabling a 649% improvement in compression efficiency.
This groundbreaking technology not only addresses the rapidly growing global demand for storage capacity, projected to surpass available resources in the coming decades, but also offers a sustainable and cost-effective alternative. By leveraging RNA’s ability to store dense information in a biochemically compact format, this innovation has the potential to revolutionize data storage while reducing the environmental and financial costs associated with conventional methods.
The Human RNome Project will be at the forefront of this innovation, establishing the necessary infrastructure to produce standard and modified ribonucleotides as foundational components for RNA-based storage. In parallel, the project will drive advancements in sequencing and synthesis technologies to ensure data integrity, reliability, and affordability. These efforts will transition RNA-based storage from theoretical concept to practical reality, opening a new frontier in data technology. By combining unparalleled storage density with innovative compression strategies, RNA-based systems promise to redefine how we store and access the world’s growing digital archives.
Pandemic and biowarfare preparedness: RNA viruses are responsible for nearly half of infectious diseases, including influenza, Ebola, hepatitis A, and COVID-19. Their high mutation rates, up to five times that of DNA viruses, make early detection and control challenging [74, 75]. The Human RNome Project will revolutionize RNA virus sequencing, enabling rapid and accurate identification of emerging pathogens. This capability will enhance global pandemic preparedness, providing the tools needed to respond swiftly to new threats. Moreover, advances in RNA sequencing will be critical for detecting engineered viruses, strengthening defenses against biowarfare, and ensuring global health and security. By enabling rapid, precise detection, the RNome Project will play a vital role in safeguarding against both natural and human-made threats.
In conclusion, the Human RNome Project will drive transformative progress across a range of fields, from health and agriculture to technology and security. By unlocking the full potential of RNA science, this initiative will deepen our understanding of fundamental biological processes and empower innovative solutions to some of humanity’s most pressing challenges. With its far-reaching applications and societal impact, the Human RNome Project promises to be a cornerstone of twenty-first-century innovation, paving the way for a healthier, more resilient future.