“We’re thrilled to be back on board with Carnival, literally! From an ice-cold Coca-Cola with lunch, a Topo Chico by the pool, or a Powerade after exploring a destination, we’ll have the perfect drink for every moment of the journey,” said Dagmar Boggs, president, Foodservice & On-Premise, The Coca-Cola Company, North American Operating Unit. “This partnership is all about enhancing the guest experience, and we’re thrilled to offer a beverage lineup as diverse and exciting as the adventures Carnival guests embark on.”
Carnival anticipates the partnership will be fully integrated across its North American fleet by the end of September 2025. Carnival ships in Australia already serve Coca-Cola beverages.
For additional information on Carnival Cruise Line and to book a cruise vacation on Carnival, call 1-800-CARNIVAL, visit http://www.carnival.com, or contact your favorite travel advisor or online travel site.
In today’s state of overwhelm, it’s easy to spend more time consumed with the present and the future than contemplating the events of the past. This constant forward motion can, at times, become exhausting and disorienting. We lose our grounding. We miss out on the context and insights that history can provide, and the lessons that may guide us through the tumult of the present. This holds true not just for national history and global history, but also geological history. That is, even events that unfolded millions or billions of years ago can offer insights that remain relevant to our lives and national policies today.
I wrote Strata: Stories from Deep Time to share this geologic lens with readers, and to spotlight the researchers working to untangle some of our planet’s oldest stories. This excerpt explores how scientists first began recognizing that oxygen didn’t billow up into the atmosphere until roughly halfway through Earth‘s existence — and how the arrival of this highly reactive gas fundamentally changed the planet from the seafloor to the stratosphere. By learning how and why oxygen showed up when it did, and how the planet responded to this period of intense global environmental change, we can gain context for the environmental crises unfolding across our planet today — and become better equipped to set ourselves on a more stable path forward.
***
As you read this line, the oxygen you are pulling inside your body makes your body possible. It is allowing you to digest your most recent meal, move your eyes across these words, and think your thoughts. It is the single most important gas to your survival. You share this in common with every other animal on Earth, save for one lone parasite of Chinook salmon that somehow doesn’t need it. Well done, Henneguya salminicola.
Throughout a given day, you fill your lungs with oxygen some 20,000 times. Most of us probably don’t give it much thought. Maybe we assume that this gas has always been here, a given on this highly habitable pale blue dot.
Editor’s picks
But it turns out that this dot has not always been highly habitable, nor, for that matter, has it always been blue. The early Earth’s young magma surface sat gooey and cloaked in steam, too hot to hold liquid seas. It took a long time for continents to rise up and for ocean basins to fill in, and far longer still for oxygen to pool up in the atmosphere.
“And so,” writes Rachel Carson in The Sea Around Us, “the rough outlines of the continents and the empty ocean basins were sculptured out of the surface of the earth in darkness, in a Stygian world of heated rock and swirling clouds and gloom.”
Even in those earliest of gloomy days, oxygen — the element O — was all over the place, bound up in molecules like water vapor and quartz and carbon dioxide. It’s the third most abundant element in the universe, and it has been present on Earth since the beginning. But free oxygen — two atoms of O bound together by a pair of shared electrons, liberated from any other material but itself — didn’t emerge as a gas until more than halfway through Earth’s existence.
If you reach out your arms and imagine Earth’s 4.54-billion-year history as a timeline that extends from the tip of your right hand to the tip of your left, the arrival of oxygen gas falls around your heart, at about 2.4 billion years ago, give or take a couple hundred million years.
The fashionably late arrival of oxygen may sound like a planetary sigh of relief. Finally, the possibility for life larger than one cell, with lungs and lips and all the rest of it. But scientists familiar with oxygen’s highly reactive habits suggest its arrival was more like a nightmare.
Related Content
As is true of all elements, an atom of oxygen contains a cloud of negatively charged electrons that spin in an arrangement of “shells” around a positively charged nucleus. The outermost electron shell constantly seeks stability by filling to its capacity. In oxygen’s case, its outermost shell is two electrons short — comparatively fewer than other elements — and the configuration of those electrons contribute to oxygen’s high reactivity. Oxygen’s electron cloud is also relatively thin compared to other elements. Without much of a barrier between it and the outside world, the positive pull of the nucleus easily seeps out and lures in the negative charges of the two electrons it needs to stabilize. Two atoms of oxygen bound together as oxygen gas have a pull similarly as strong as a single atom on its own.
When oxygen first appeared on Earth, it desperately rooted out and bonded with anything willing to share its electrons, fundamentally changing the materials it bonded with. It weaseled into microbial cells and mutilated their machinery. It sulked into currents and eddies and made arsenic more soluble, it spread hydrogen peroxide poisons into DNA. With all the havoc it wreaked, this gas might have initiated one of the worst mass extinctions in all of Earth history — though it’s hard to know this for sure, since the single-celled beings that would have gone extinct were too squishy to leave behind reliable fossils. Even so, some call this geologic moment the Oxygen Catastrophe.
Over time, molecules from the bottom of the ocean to the top of the atmosphere grew to accept oxygen’s reactivity, and living things evolved ways to cope with this new gas. Their cells grew to tolerate it, and then to depend on it. They used it to break down food and generate energy that allowed them to grow larger and more complex, with multiple cells that communicated across newly sophisticated membranes. These oxygen-fueled innovations expanded and cascaded and eventually led to the evolution of eyeballs and brains and lungs and lips and, over billions of years, the possibility of us.
So what, exactly, happened around 2.4 billion years ago? Why did oxygen arrive when it did? And how can we read this in the rock record?
Courtesty of W. W. Norton & Company
THE SEARCH FOR OXYGEN’S origin began with a problem. When Charles Darwin published On the Origin of Species in 1859, he agonized over the seeming absence of fossils in the planet’s oldest rocks. The ages of rocks at this time were known only in a relative sense — as in, what formed first and what followed. The scientific law of superposition, proposed by Danish geologist Nicolas Steno in the seventeenth century, helped clarify that younger strata always sit atop older strata, since that’s how sediments accumulate in lake beds and seafloors and so on.
As hard as paleontologists of that time looked, they couldn’t find any remnants of ancient life in the oldest, bottom-most strata that they examined. Then bits and bobs appeared in what looked like an explosion of living things in strata above a certain age. This troubled Darwin deeply. Any such explosion of life undermined his theory of natural selection, a process of elimination that he argued should inherently take a very long time to unfold. By his estimations, it could never have taken place as instantaneously as those earliest fossils suggested.
Halfway through On the Origin of Species, he gravely acknowledged the implications of this predicament. “The case at present must remain inexplicable[,]” he wrote, “and may be truly urged as a valid argument against the views here entertained.”
But here we are, still entertaining Darwin’s views more than 150 years later. And that is thanks largely to rocks discovered not long after World War II.
At the end of the war, a wave of mineral exploration arose across the world to meet the needs of rapidly expanding economies. Federal agencies hired geologists to scour continents for oil, gas, and coal to fuel those economies, along with metals like iron and uranium to build up arsenals of defense. This was of national interest, not just private economic interest.
As geologists marched around the globe and sketched up their maps of these resources, they noticed other curious details about the planet’s history. That is, in their search for the materials that humans desired, they found inklings of how we got to be here desiring anything in the first place.
In the summer of 1953, Wisconsin geologist Stanley Tyler was studying iron-rich rocks on the north shore of Lake Superior in Canada when he took a Sunday off to rent a boat and go fishing. While his lure bobbed in the water, he absently noted the shapes and colors along the shore, as any geologist might. One outcrop caught his eye, so he motored over to take a closer look.
Tyler recognized the deposit as an extension of the Gunflint Chert, a rock formation with the texture of tightly packed brown sugar and the contents of ancient seafloor sediments. Cherts can take on a whole range of colors depending on the conditions they form within, from beige to red to green to other hues in between. Most of the chert that Tyler had found on that trip had been maroon, but this outcrop caught his eye for its striking shade of jet black. He knew that the color black in rocks was sometimes indicative of organic material, remnants of ancient life.
He lopped off a chunk, stashed it in his boat, and motored on.
Back at his lab in Madison he placed a sliver of that black chert under a microscope, and found shapes that did not speak the language of minerals. The rods, spheres, and squiggles he found did, as he suspected, look more lifelike than lithic.
Based on geologic maps of the region, he knew these rocks had formed during the allegedly fossil-free epoch that had so troubled Darwin. Tyler’s gut told him he may have just found some of the earliest evidence of life ever discovered, but he was a mineralogist more than a paleontologist and so he needed a second opinion.
That fall, he took photographs of his findings to a geology conference in Boston and shared them with a couple of colleagues. One among them, a Harvard paleobotanist named Elso Barghoorn, agreed that the samples looked rather lifelike, and the two published a short paper describing what they had found.
This publication quadrupled the length of the fossil record. It was groundbreaking, but was brief and preliminary. They needed more time to study the fossils to do justice to the scope of their findings.
For years, they didn’t make progress on a follow-up paper. A decade went by and, in 1963, Tyler passed away at the age of 57 from heart complications, without the satisfaction of sharing his discoveries more completely with the world. By 1965, an impatient colleague named Preston Cloud — a bantamweight boxing champion turned acclaimed Earth historian — threatened to beat Barghoorn to the punch with his own paper on the fossils. That was enough to push Barghoorn into gear. He rushed to complete a manuscript and published it in the journal Science a couple months before Cloud published his.
“For all of time it will probably stand as the most important article ever written in the field . . . ,” writes William Schopf, a graduate student who helped Barghoorn pen that manuscript, but who humbly declined authorship himself because he didn’t feel he had contributed enough.
Spurred by this new paper on the Gunflint Chert, geologists went searching for evidence of ancient life in black cherts around the world. Papers flooded out, claiming to have solved Darwin’s dilemma and showing how fossils had been in those seemingly lifeless rocks all along — they had simply been microscopic. The theory of natural selection persevered, and the lengthy record of our ancient roots began to fill out.
But while those microscopic rods and squiggles resolved one nagging dilemma, they opened up a slew of other questions. What, exactly, were those fossils? What kind of world did they evolve into? And what kind of world did they create with their growth?
Around the same time that these questions began bubbling up, another set of observations from the rock record thickened the plot of the squiggles. Geologists were compiling evidence that, before those lifeforms lived, the planet’s atmosphere had no oxygen gas in it at all. Minerals that disintegrate in the presence of oxygen were found locked up in ancient riverbeds older than a certain age. Then, around the time they believed those squiggles showed up on the scene, those riverbed minerals disappeared and the very first, rusty red fingerprints of oxygen began appearing in strata around the world.
Trending Stories
Perhaps, some reasoned, those squiggles were responsible for painting the world’s soil and seafloor sediments red, by ushering in the very first poofs of oxygen. And perhaps, in their delivery of this gas, they catapulted Earth out of its original barrenness and into the tangle of complex life we know today.
Researchers at the Indian Institute of Science (IISc) have developed a unique luminescent probe that uses terbium, a rare earth metal, to sense the presence of an enzyme called β-glucuronidase, which can potentially aid in the detection of liver cancer.
According to IISc, β-glucuronidase is an evolutionarily conserved enzyme found across life forms – from microbes to plants and animals. Its core function is to break down a sugar acid called glucuronic acid.
“Beyond its biological omnipresence, the enzyme also doubles as a critical biomarker for liver cancer. In fact, an upsurge in β-glucuronidase often accompanies colon, breast, and renal cancers, as well as infections of the urinary tract and AIDS,” IISc said.
Conventional methods
“Conventional methods of colorimetry and fluorescence for detecting such enzymes are often restricted by sensitivity or interference from background signals. The ability of rare earth metals to have long-lived excited states allows us to filter out short-lived background fluorescence, resulting in a much clearer signal,” said Ananya Biswas, former PhD student at IISc and co-first author of the paper published in Chemistry.
The roots of the project trace back nearly a decade, beginning with the team’s experiments on metal ions and their gel-forming properties. The team found that terbium ions couched in a gel matrix derived from bile salts can emit green fluorescence.
Within the same gel matrix, the team added an organic molecule called 2,3-DHN (2,3-Dihydroxynaphthalene) “masked” with glucuronic acid. When β-glucuronidase slices this modified molecule, 2,3-DHN gets released. The researchers then shined UV light on the sample.
“The free 2,3-DHN acts as an ‘antenna’ – absorbing UV light and transferring the energy to the terbium ions nearby, greatly enhancing their green emission. The gel matrix ensures sufficient proximity between the ‘antenna’ and terbium ions, facilitating efficient energy transfer,” said Uday Maitra, honorary professor in the Department of Organic Chemistry, IISc, and corresponding author of the study.
Ease of application
For ease of application, the team designed this assay as a simple paper-based sensor by anchoring the gel matrix onto a paper disc. When β-glucuronidase pre-treated with modified 2,3-DHN is added, the disc exhibits a much stronger green glow under UV light.
IISc said that the uniqueness of this technique is the analysis. “Unlike conventional high-end fluorescence detection systems, these sensors can be analysed using a UV lamp and ImageJ, an open source, freely accessible software, making this technique ideal for resource-limited settings,” it said.
Imaging data from Japan’s Himawari-8 and -9 meteorological satellites have been successfully used to monitor temporal changes in Venus’ cloud-top temperature, revealing unseen patterns in the temperature structure of various waves. A team led by the University of Tokyo collated infrared images from 2015–25 to estimate brightness temperatures on day to year scales. The results demonstrate that meteorological satellites can serve as additional eyes to access the Venusian atmosphere from space and complement future observations from planetary missions and ground-based telescopes.
The Himawari-8 and -9 satellites, launched in 2014 and 2016, respectively, were developed to monitor global atmospheric phenomena through use of their multispectral Advanced Himawari Imagers (AHIs). The University of Tokyo team led by visiting researcher Gaku Nishiyama saw the opportunity to use the cutting-edge sensor data for spaceborne observations of Venus, which is coincidentally captured by the AHIs near the Earth’s rim.
Observing temporal temperature variations in the cloud tops of Venus is essential to understand its atmospheric dynamics and related phenomena, such as thermal tides and planetary-scale waves. Obtaining data for these phenomena presents multiple challenges, as Nishiyama explained. “The atmosphere of Venus has been known to exhibit year-scale variations in reflectance and wind speed; however, no planetary mission has succeeded in continuous observation for longer than 10 years due to their mission lifetimes,” he said. “Ground-based observations can also contribute to long-term monitoring, but their observations generally have limitations due to the Earth’s atmosphere and sunlight during the daytime.”
Meteorological satellites on the other hand appear suited to fill this gap with their longer mission lifetimes (the Himawari-8 and -9 satellites are scheduled for operation until 2029). The AHIs allow multiband infrared coverage, which has been limited in planetary missions to date, essential for retrieving temperature information from different altitudes, along with low-noise and frequent observation. Aiming to demonstrate this potential to contribute to Venus science, the team investigated the observed temporal dynamics of the Venusian atmosphere and provided a comparative analysis with previous datasets. “We believe this method will provide precious data for Venus science because there might not be any other spacecraft orbiting around Venus until the next planetary missions around 2030,” said Nishiyama.
The team first established a data archive by extracting all Venus images from the collected AHI datasets, identifying 437 occurrences in total. Taking into account background noise and apparent size of Venus in the captured images, they were able to track the temporal variation in cloud-top temperature during the periods where the geostationary satellite, Venus and the Earth lined up in a row.
The retrieved temporal variations in brightness temperatures were then analyzed on both year and day scales and compared for all infrared bands to investigate variability of thermal tides and planetary-scale waves. Variation in thermal tide amplitude was confirmed from the obtained dataset. The results also confirmed change in amplitude of planetary waves in the atmosphere with time, appearing to decrease with altitude. While definitive conclusions on the physics behind the detected variations were challenging due to the limited temporal resolution of the AHI data, variations in the thermal tide amplitude appeared possibly linked to decadal variation in the Venus atmosphere structure.
In addition to successfully applying the Himawari data to planetary observations, the team was further able to use the data to identify calibration discrepancies in data from previous planetary missions.
Nishiyama is already looking at implications of the study beyond Venus’ horizon. “I think that our novel approach in this study successfully opened a new avenue for long-term and multiband monitoring of solar system bodies. This includes the moon and Mercury, which I also study at present. Their infrared spectra contain various information on physical and compositional properties of their surface, which are hints at how these rocky bodies have evolved until the present.” The prospect of accessing a range of geometric conditions untethered from the limitations of ground-based observations is clearly an exciting one. “We hope this study will enable us to assess physical and compositional properties, as well as atmospheric dynamics, and contribute to our further understanding of planetary evolution in general.”
###
Journal article: Gaku Nishiyama, Yudai Suzuki, Shinsuke Uno, Shohei Aoki, Tatsuro Iwanaka, Takeshi Imamura, Yuka Fujii, Thomas G. Müller, Makoto Taguchi, Toru Kouyama, Océane Barraud, Mario D’Amore, Jörn Helbert, Solmaz Adeli, Harald Hiesinger, “Temporal variation in the cloud-top temperature of Venus revealed by meteorological satellites”, Earth, Planets and Space, DOI: 10.1186/s40623-025-02223-8
Funding: This work was supported by JSPS KAKENHI Grant Number JP22K21344, 23H00150, and 23H01249, and JSPS Overseas Research Fellowship.
Useful links:
Department of Earth and Planetary Science – https://www.eps.s.u-tokyo.ac.jp/en/ Graduate School of Science – https://www.s.u-tokyo.ac.jp/en/index.html
Research contact: Dr. Gaku Nishiyama Department of Earth and Planetary Science, The University of Tokyo, 7-3-1 Hongo, Bunkyo-ku, Tokyo, 113-0033, Japan gaku.nishiyama@dlr.de
Press contact: Mr. Rohan Mehra Public Relations Group, The University of Tokyo, 7-3-1 Hongo, Bunkyo-ku, Tokyo, 113-8656, Japan press-releases.adm@gs.mail.u-tokyo.ac.jp
About The University of Tokyo:
The University of Tokyo is Japan’s leading university and one of the world’s top research universities. The vast research output of some 6,000 researchers is published in the world’s top journals across the arts and sciences. Our vibrant student body of around 15,000 undergraduate and 15,000 graduate students includes over 5,000 international students. Find out more at www.u-tokyo.ac.jp/en/ or follow us on X (formerly Twitter) at @UTokyo_News_en.
Journal
Earth Planets and Space
Method of Research
Observational study
Subject of Research
Not applicable
Article Title
Temporal variation in the cloud-top temperature of Venus revealed by meteorological satellites
Article Publication Date
30-Jun-2025
Disclaimer: AAAS and EurekAlert! are not responsible for the accuracy of news releases posted to EurekAlert! by contributing institutions or for the use of any information through the EurekAlert system.
While we only use edited and approved content for Azthena
answers, it may on occasions provide incorrect responses.
Please confirm any data provided with the related suppliers or
authors. We do not provide medical advice, if you search for
medical information you must always consult a medical
professional before acting on any information provided.
Your questions, but not your email details will be shared with
OpenAI and retained for 30 days in accordance with their
privacy principles.
Please do not ask questions that use sensitive or confidential
information.
Keeping electricity affordable for consumers is a “formidable challenge” amid projections of declining generation capacity reserves and persistent uncertainty around the scale and pace of future load growth, ICF International Vice President of Energy Markets Maria Scheller said Thursday.
Meanwhile, broad policy uncertainty and an increasingly shaky regulatory environment give utilities and capital markets pause about expensive new infrastructure investments that could become stranded assets, Scheller said in a webinar on ICF’s “Powering the Future: Addressing Surging U.S. Electricity Demand” report.
Policy conversations around import tariffs, federal energy tax credits and permitting reform are unfolding as the balance of electricity demand shifts from residential and business consumers to technology and industrial customers, which tend to require around-the-clock power, Scheller added.
Dive Insight:
The coming shift in U.S. electricity consumption represents less of a new paradigm than a return to the industrial-driven demand the country saw from the 1950s into the 1980s, after which deindustrialization and consumer-centric trends like the widespread adoption of air conditioning, electric resistance heating and personal computing shifted the balance toward the residential segment, Scheller said.
The shift is important because unlike residential loads, which show considerable seasonal and intraday variation, industrial loads are flatter, less weather-dependent and more sensitive to voltage fluctuations, Scheller said.
By 2035, ICF expects nearly 40% of total U.S. load will have a “flat, power-quality-sensitive profile,” and that overall load will grow faster than peak load, she said. In 2030, ICF projects more than 3% annual power consumption growth, compared with less than 2% annual peak load growth, according to a webinar slide.
That’s not to say residential demand won’t also grow in the next few years as consumers electrify home heating and buy more electric vehicles — only that data centers and other industrial demand will “dwarf” it, Scheller said.
The only significant regional exception to that expectation is California, where ICF says light- and heavy-duty transportation electrification will drive most load growth through 2040.
Capacity reserves will quickly dwindle across most of the United States as a result of near-term load growth, regardless of the regional drivers, said Lalit Batra, ICF director of energy markets.
“We expect the capacity reserve across most regions to be absorbed in the next few years,” Batra said.
ICF sees nationwide capacity reserves — currently between 20% and 25% — below the 15% target reserve margin by 2030 and in negative territory by 2035, according to a webinar slide. Without a meaningful acceleration in new generation deployment, some combination of delayed power plant retirements, load flexibility or slower overall load growth will be needed to avoid shortages, Batra said.
Building new power plants fast enough to keep pace with load growth will be more difficult if Republicans in Congress effectively repeal the Inflation Reduction Act, according to ICF’s projections. If the final budget reconciliation bill preserves the rapid phaseout of clean energy tax credits in the version the U.S. House of Representatives passed in May, ICF projects the U.S. would deploy 280 GW fewer renewables and storage capacity and 43 GW more gas and nuclear through 2040, even as cost-competitiveness, supportive state policy and corporate power buyers’ preference for clean electricity supports robust regional markets for those technologies.
Regardless of the policy scenario, ICF expects electricity prices to rise as much as 25% in some regions through 2030 due to necessary grid expansion, wildfire hardening and other infrastructure projects, along with gas price volatility, said Deb Harris, ICF vice president for climate change and sustainability. ICF expects U.S. gas prices to be 7% higher in 2035 and 17% higher in 2045 if the House budget becomes law relative to a status-quo scenario, according to a webinar slide.
Utilities and regulators already have the tools to mitigate some of these changes, Harris said. For example, demand response programs and flexible load interconnection could avoid about 30% of infrastructure investment costs that would otherwise be necessary, Harris said. Large loads are more open than some realize to ramping down load or investing in more efficient processes, such as liquid rather than air cooling of server racks, she added.
“These large load customers do offer a lot of opportunities” for efficiency and demand response, she said. “Energy is the number one [operating] cost they face.”
Permitting reforms like uniform siting standards, incentives for brownfield redevelopment and wider adoption of advanced GIS tools to locate “areas of minimal impact” for energy development could speed up new builds and keep prices in check, Harris added.
The catch, she said, is that while efforts to mitigate rising electricity prices may benefit customers and the politicians who represent them, project developers and their lenders and investors want to see durable price signals before committing to build new generation and transmission.
Dublin, June 30, 2025 (GLOBE NEWSWIRE) — The “Vectorized Antibodies Market Size, Share, Trends, Analysis, and Forecast 2025-2034 | Global Industry Growth, Competitive Landscape, Opportunities, and Challenges” has been added to ResearchAndMarkets.com’s offering.
The Global Vectorized Antibodies Market, valued at USD 9.49 billion in 2025, is poised for robust growth, anticipated to expand at a significant CAGR of 7.7% to reach USD 15.94 billion by 2034.
This market is at the cutting edge of biotherapeutics, employing gene therapy techniques to deliver antibody coding sequences directly into the body via viral or non-viral vectors. This novel approach reduces the need for frequent dosing of protein-based drugs, fostering prolonged in vivo production of therapeutic antibodies with the potential for prolonged efficacy and streamlined manufacturing processes.
This innovation holds substantial promise for chronic conditions such as HIV, cancer, autoimmune disorders, and rare genetic diseases, where continuous antibody presence is crucial. By merging genetic engineering precision with antibody targeting specificity, vectorized antibodies aim to enhance therapeutic outcomes and patient adherence.
As clinical trials continue demonstrating safety and efficacy, the vectorized antibodies market may redefine the landscape of biologic therapies. Manufacturing scalability and batch consistency are critical hurdles, but advances in vector production infrastructure support continued progress. The regulatory landscape is evolving, particularly in classifying gene therapies and combination products.
The report provides comprehensive insights into the vectorized antibodies market, covering market size and growth projections, trends, challenges, competitive landscape, and regional analyses across North America, Europe, Asia-Pacific, Middle East and Africa, and South and Central America. It’s an invaluable resource for top management, investors, and stakeholders to strategize and capitalize on emerging opportunities.
Customization options are available, with up to 10% free customization (up to 10 analyst hours) offered to modify segments, geographies, and analyzed companies. Post-sale support includes four analyst hours, available for up to four weeks, delivered in the latest PDF and Excel formats.
Key Market Insights:
Vectorized antibodies reduce the need for repeated dosing, potentially lowering long-term treatment costs and healthcare burdens.
North America’s leadership is due to its advanced gene therapy infrastructure and supportive funding landscape, while Europe progresses with regulatory and collaborative support.
The focus is on enhancing vector payload capacity and specificity while minimizing immunogenic responses.
Applications range across oncology, infectious, autoimmune, and rare genetic diseases.
Key challenges include vector delivery efficacy, long-term safety considerations, and regulatory complexities.
North America: Leads the market, driven by investments in gene therapy research and a dynamic biotech ecosystem. Strategic collaborations between academia and industry further bolster growth.
Europe: Experiences steady growth, supported by regulatory frameworks that encourage advanced therapy medicinal products. Key players focus on optimizing vector delivery, reducing immunogenicity, and scaling production under GMP conditions.
While the market is in its early stages, it is advancing rapidly with numerous partnerships, licensing agreements, and venture capital inflows driving clinical pipeline development.
Key Attributes:
Report Attribute
Details
No. of Pages
150
Forecast Period
2025 – 2034
Estimated Market Value (USD) in 2025
$9.49 Billion
Forecasted Market Value (USD) by 2034
$18.5 Billion
Compound Annual Growth Rate
7.7%
Regions Covered
Global
Companies Featured
Adagio Therapeutics
ReiThera Srl
Voyager Therapeutics
Spark Therapeutics
Freeline Therapeutics
BioNTech SE
Moderna, Inc.
AskBio (subsidiary of Bayer AG)
Genethon
Vector BioPharma AG
Sangamo Therapeutics
Passage Bio
Gilead Sciences, Inc.
Vir Biotechnology, Inc.
Regeneron Pharmaceuticals, Inc.
For more information about this report visit https://www.researchandmarkets.com/r/osx0kq
About ResearchAndMarkets.com ResearchAndMarkets.com is the world’s leading source for international market research reports and market data. We provide you with the latest data on international and regional markets, key industries, the top companies, new products and the latest trends.
People say Battle reminds them of some really good years for Britain as a country. We were entering a new millennium, everyone was running their own business, making money and the underground record industry was thriving. I wanted to do a UK garage version of Southern Freeez, by the 80s UK funk band Freeez. Initially, Battle was going to be another instrumental, and then Lain, the singer, came in the room and goes: “Let me put something on this.” I was like: “I’m not sure it’s really a vocal song.” But Lain stacked the vocals, and someone compared it to Queen’s Bohemian Rhapsody, with all the harmonies.
At the time, some people believed that I had copied a dubplate by [UK garage figure] DJ EZ. It is similar – his bassline does something like that. But I know I didn’t copy him, because I didn’t go out that much! So I’d never really heard EZ at that stage.
I didn’t know Battle was going to be as big as it was or have the impact it did – it was just another tune. Then it started to float around: a few DJs had it, the A&R people. Ears started to prick up. I think we added a level of sophistication to garage, even though people like MJ Cole and TJ Cases were already doing that. When we were trying to get Battle on the radio, one station said it was too intelligent for their listeners and they wouldn’t play it.
Battle eventually reached No 10 in August 2000. Louise Redknapp got the No 9 spot by 2,000 copies. It allowed me to appear on a TV show that I had watched my whole life as I was growing up: Top of the Pops.
A lot of people say: “Oh, garage is on the comeback.” But for me, it has been for the last 13 years. I started DJing in 2012 and every year I’m working. It’s been well received by a wider audience ever since, younger and younger.
Lain, singer
Jason Chue, AKA Wookie, was knocking about in jungle, drum’n’bass, trying to siphon off that whole energy and then putting songs to it.I remember him saying about Battle: “I’ve done something strange with the intro.” I was like: “No, just play it.” He was almost apologising for it – but it was like a godsend. That intro had such an urgency. It felt like it was piercing your soul. I said: “Give me a minute.” Then I walked out and I don’t even know if it was half an hour, but I came back and I had written all of the vocal. That’s divine. For the song title we wanted one word. There are a lot of three-word titles, but one word is strong. Whether it’s bringing up three kids on your own, or addiction, everyone’s battling through something, every day.
People call Battle a gospel song. Back then I was doing a lot of regular R&B but I really wanted to do something that involved my faith. When Jazzie B [founder of Soul II Soul and mentor to Wookie] called me about working with Jason, I thought he was going to say: “No, we don’t want that.” But he said: “Just go for it.” We had all these record label bosses trying to sign Battle – one guy had a Maserati and he blew out his speakers playing it. Months later he said: “I didn’t realise I was blowing out my speakers to a gospel tune.”
I first realised Battle was going to be big at [seminal UK garage night] Twice As Nice. Jason said: “You should come down to Twice As Nice because I think this tune’s going.” I was a bit worried because I don’t really go out. So they played the tune and everyone started going: “Booooo!” OK, that’s not a good sign. But Jason was like, “No, no, no – that means they really like it!”
Battle has endured because of what it means to the person who hears it. Back then, we would do PAs and people would say: “That song – my mum was going through cancer and that helped me.” And, 25 years later, someone said: “While I was in prison, that song got me through.” That’s everyone. That’s anyone. I remember Jason sending me a picture of someone who tattooed the middle eight of Battle on her forearm: “I can always rely / On my faith to get by.”
A new study from scientists at the Buck Institute for Research on Aging has revealed a surprising player in the battle against Alzheimer’s disease and other forms of dementia: brain sugar metabolism. Published in Nature Metabolism, the research uncovers how breaking down glycogen — a stored form of glucose — in neurons may protect the brain from toxic protein buildup and degeneration.
Glycogen is typically thought of as a reserve energy source stored in the liver and muscles. While small amounts also exist in the brain, particularly in support cells called astrocytes, its role in neurons has long been dismissed as negligible. “This new study challenges that view, and it does so with striking implications,” says Professor Pankaj Kapahi, PhD, senior scientist on the study. “Stored glycogen doesn’t just sit there in the brain; it is involved in pathology.”
The research team, led by postdoc Sudipta Bar, PhD, discovered that in both fly and human models of tauopathy (a group of neurodegenerative diseases including Alzheimer’s), neurons accumulate excessive glycogen. More importantly, this buildup appears to contribute to disease progression. Bar says tau, the infamous protein that clumps into tangles in Alzheimer’s patients, appears to physically bind to glycogen, trapping it and preventing its breakdown.
When glycogen can’t be broken down, the neurons lose an essential mechanism for managing oxidative stress, a key feature in aging and neurodegeneration. By restoring the activity of an enzyme called glycogen phosphorylase (GlyP) — which kicks off the process of glycogen breakdown — the researchers found they could reduce tau-related damage in fruit flies and human stem cell-derived neurons.
Rather than using glycogen as a fuel for energy production, these enzyme-supported neurons rerouted the sugar molecules into the pentose phosphate pathway (PPP) — a critical route for generating NADPH (nicotinamide adenine dinucleotide phosphate) and Glutathione, molecules that protect against oxidative stress. “By increasing GlyP activity, the brain cells could better detoxify harmful reactive oxygen species, thereby reducing damage and even extending the lifespan of tauopathy model flies,” said Bar.
Even more promising, the team demonstrated that dietary restriction (DR) — a well-known intervention to extend lifespan — naturally enhanced GlyP activity and improved tau-related outcomes in flies. They further mimicked these effects pharmacologically using a molecule called 8-Br-cAMP, showing that the benefits of DR might be reproduced through drug-based activation of this sugar-clearing system. “This work could explain why GLP-1 drugs, now widely used for weight loss, show promise against dementia, potentially by mimicking dietary restriction,” said Kapahi.
Researchers also confirmed similar glycogen accumulation and protective effects of GlyP in human neurons derived from patients with frontotemporal dementia (FTD), strengthening the potential for translational therapies. Kapahi says the study emphasizes the power of the fly as a model system in uncovering how metabolic dysregulation impacts neurodegeneration. “Work in this simple animal allowed us to move into human neurons in a much more targeted way,” he said.
Kapahi also acknowledges the Buck’s highly collaborative atmosphere as a major factor in the work. His lab, with expertise in fly aging and neurodegeneration, took advantage of proteomics expertise in the Schilling lab and the Seyfried lab (at Emory University) as well as the Ellerby lab which has expertise in human iPSCs and neurodegeneration.
Kapahi says this study not only highlights glycogen metabolism as an unexpected hero in the brain but also opens up a new direction in the search for treatments against Alzheimer’s and related diseases. “By discovering how neurons manage sugar, we may have unearthed a novel therapeutic strategy: one that targets the cell’s inner chemistry to fight age-related decline,” he says. “As we continue to age as a society, findings like these offer hope that better understanding — and perhaps rebalancing — our brain’s hidden sugar code could unlock powerful tools for combating dementia.”
Coauthors: Additional Buck collaborators include Kenneth A. Wilson, Tyler A.U. Hilsabeck, Sydney Alderfer, Jordan B Burton, Samah Shah, Anja Holtz, Enrique M. Carrera, Jennifer N. Beck, Jackson H Chen, Grant Kauwe, Tara E. Tracy, Birgit Schilling, and Lisa M. Ellerby. Other collaborators include Eric B. Dammer, Fatemeh Seifar and Nicholas T. Seyfried, Emory Center for Neurodegenerative Disease, Emory University School of Medicine, Atlanta, GA as well as Ananth Shantaraman, Department of Biochemistry, Emory University School of Medicine, Atlanta, GA
Acknowledgments: The work was supported by NIH grants R01AG038688, R21AG054121, AG045835, R01AG071995, R01AG070193, T32AG000266-23, R01AG061879, P01AG066591 and 1S10 OD016281. Other support came from the Hevolution Foundation, American Federation of Aging Research, the Larry L. Hillblom Foundation and the CatalystX award from Alex and Bob Griswold