Connect with us
DAPA Banner

Tech

13 Highly Rated Harbor Freight Hand Tools For Your Starter Kit

Published

on





Building a quality selection of hand tools to help support your repair, DIY renovation, or emergency preparedness needs isn’t always easy. Finding great gear is rarely the problem, though. Starter kits demand an odd blend of cost-effectiveness and value that can be difficult to identify. As a beginner creating your first collection of versatile hand tools, you’ll often want to target sets of equipment (like a bundle deal of screwdrivers or a socket set with a wide range of turning tools) while purchasing the best brand options you can afford at a reasonable price point. There’s no need to focus on premium quality everywhere, and even those with experience seeking to reimagine their existing mechanic’s tool kit will want to save in some areas and splurge in others. In fact, many pros suggest a different route, opting for inexpensive gear that’s unlikely to last a long time. The things you break first are naturally going to be the equipment you use the most, giving you a personalized blueprint for where to upgrade over time.

Harbor Freight tools offer a solid blend of both worlds. Harbor Freight’s catalog of in-house brands features plenty of quality implements at surprisingly bargain-friendly prices. The outlet certainly carries its fair share of expensive equipment, but many of the hand tools offered by the tool and home improvement store are inexpensive without sacrificing quality or key build features that users crave. These 13 tools are cost-friendly options with great reviews in their respective categories, providing reliable coverage for numerous jobs you may be gearing up to tackle.

Advertisement

Pittsburgh Comfort Grip Screwdriver Set (70-Piece)

There’s always room in a garage or workshop for a set of screwdrivers. This is one of the baseline tools that every fixer, builder, or renovator needs to have in their collection. No matter the task you’re squaring up against, three functions remain at the top of any job’s requirements: measuring, cutting, and fastening. You won’t get very far without screwdriving tools, and a set of handheld fastening tools can be a true game changer. The Pittsburgh Comfort Grip Screwdriver Set includes Allen keys, nut drivers, precision screwdrivers, and standard screwdriving tools. It’s all contained within a storage rack that makes organization simple.

Advertisement

Each screwdriver features an oil-resistant TPR cushion grip that makes the set useful in demanding environments while also remaining comfortable for use over long periods of time. The screwdrivers feature chrome vanadium steel construction with magnetized tips and flat-sided handles that offer additional gripping power while reducing the risk of rolling. About 98% of buyers recommend the set, and its price is among the primary reasons. It’s offered at Harbor Freight for $30, but Inside Track Club members can save $10 on their purchase before April 2. The tool has been reviewed by over 1,230 buyers and holds a 4.8-star average rating.

Advertisement

Central Machinery 5-Inch Swivel Vise with Anvil

The Central Machinery 5-Inch Swivel Vise with Anvil is an in-store-only tool offered by Harbor Freight. The vise is available for $50 for Inside Track Club members until April 2, and $60 for other buyers. Both prices are favorable for a quality swiveling vise designed for heavy-duty use in your garage or workshop. The tool features 5-inch, heat-treated replaceable jaws and can produce 6,600 pounds of clamping force. The tool’s body is made of cast iron, resulting in a 21-pound tool that can offer portability when necessary but is dense enough for solid workholding when you need to lock down a component.

The vise features a 2-1/8-inch throat depth, and the swiveling base offers a full 360-degree rotational arc. The tool has received over 270 buyer reviews with a 4.7-star average rating. It has a 96% recommendation rate, and the replaceable jaws and multifunctional capability, underpinned by the addition of a 3-1/2-inch by 3-5/8-inch anvil, have a lot to do with this high praise.

Advertisement

Quinn 1/4, 3/8, and 1/2-Inch Drive SAE and Metric Hi-Vis Socket Set (66-Piece)

There are plenty of great mechanic’s tool sets for every budget, including some that feature no additions beyond the ratchet and sockets. The Quinn 1/4, 3/8, and 1/2-Inch Drive SAE and Metric Hi-Vis Socket Set is a quality choice in this regard. It’s listed at Harbor Freight for $60, offering 66 total pieces for a good price per element and a relatively low cost overall. The set comes with a carrying case featuring individual storage slots, allowing you to find exactly what you’re looking for without hassle. The set is built with chrome vanadium steel, offering scratch and rust resistance for long-term durability. Each ratchet features a 72-tooth quick-release head, and the sockets offer high-visibility markings with color coding to indicate drive size.

Customers overwhelmingly give this high praise, with a 4.8-star average rating from over 1,460 reviewers. 98% of buyers recommend the item to others. In addition to standard fastener-turning tools contained within the set, it includes three extensions and two spark plug sockets, allowing buyers to change their own spark plugs without having to go out and buy a new, specialized tool.

Advertisement

Pittsburgh 3/4-Inch Pipe Clamp with Base

Clamps come in a dizzying array of styles, but they all feature a basic commonality. Clamping tools are used to hold workpieces in place, allowing glue to set or keeping a component held firm while you fasten or cut away excess material. While numerous clamps can be of great value to someone building a starter toolkit, perhaps the most versatile solution you’ll find is the pipe clamp. Unlike other options, this tool comes as a pair of cast ends that fit onto a pipe you’ll add to the mix yourself. This allows you to customize the length to your own needs and provides extreme adaptability.

The Pittsburgh 3/4-Inch Pipe Clamp with Base is a great option, listed at Harbor Freight for $12. It’s an in-store-only tool. The clamp features a four-plate clutch to deliver intense clamping pressure, along with a quick-release lever to back it off when the job is finished. It’s built with cast steel and a heavy-duty ACME lead screw. It doesn’t come with a pipe but fits 3/4-inch-diameter pipes and includes pre-drilled holes to attach additional jaws if necessary. 98% of buyers recommend it, and more than 420 reviewers give it a 4.7-star average rating.

Advertisement

Doyle 12-Inch Professional Rafter Square

The Doyle 12-Inch Professional Rafter Square is a layout tool with huge functionality built into its subtle frame. Also known as a speed square, this tool makes measuring and marking boards significantly faster. The tool is built with an anodized aluminum body with high-visibility, laser-etched markings across its face. One end of the triangular tool features a lip that allows you to quickly square it up against the edge of a board, with the 90-degree angle delivering a perfect marking edge to scribe measurements or cut lines onto a workpiece with ease. It also features notches along its edge that allow users to drag a line horizontally across a board for rip cuts.

Advertisement

The tool is essential for getting the angles and markings correct when building roof trusses and other angular constructions on a job site or in your own backyard, but it also serves as a key asset for many other projects. It’s available from Harbor Freight for $17 and has received over 500 reviews with a 4.9-star average rating. Among buyers, about 98% recommend it to others.

Advertisement

Central Forge 15-Pound Rugged Cast Iron Anvil

The anvil is often thought of as a tool that’s only required in metalworking shops. Forging hand tools, bladed instruments, and other cast decorative elements is specialized work, so it’s easy to overlook the tool. However, this addition, which acts in opposition to your hammer or other workpiece manipulation assets, can add significant value to many aspiring home improvers’ collections and workshops of all sorts.

The Central Forge 15-Pound Rugged Cast Iron Anvil is a relatively lightweight solution. It also won’t break the bank. The tool’s $20 price tag and 15-pound weight make it a small-scale investment that can pay huge dividends for users long into the future. In the same way that a vise can provide massively versatile workholding capabilities, the anvil offers a smooth striking surface to work with across many different job requirements.

This tool features a cast iron construction with a milled face. It offers an 8-1/4-inch by 3-inch work surface with a hardy hole that supports punching through material, bending components, or accessory installations. It also offers a rounded horn for shaping and smooth bending. The tool also includes extra-large feet that help keep it firmly planted while in use. The anvil has a 4.6-star average rating from more than 470 buyers, with a 93% recommendation rate.

Advertisement

Pittsburgh 16 oz. Fiberglass Rip Claw Hammer

Another critically important tool: Virtually everyone who uses hand implements, even sparingly, will need a hammer to support striking functions and basic fastener removal tasks. The Pittsburgh 16 oz. Fiberglass Rip Claw Hammer is a $7 purchase at Harbor Freight, offering a fiberglass handle for effective shock absorption. It features a non-slip rubber grip as well, pairing with a drop-forged steel head. The tool features a rip claw construction rather than the more commonly observed curved claw, delivering a straighter swing path and improved striking control. The hammer features a smooth face for a classic finish.

Nearly all customers (99%) recommend the hammer to others, and across 2,460 reviews it has received a 4.8-star average rating. The 16 oz. head weight is an ideal middle-ground solution that delivers more than enough force to drive heavy nails while remaining light enough for lengthy use. The handle is 11 inches long and features a tapered shaft layout to allow users to grip up on the tool when setting a nail and move their hand to the rubberized bottom for more striking force.

Advertisement

Doyle 13-Inch Heavy Duty Professional Hand Riveter

The Doyle 13-Inch Heavy Duty Professional Hand Riveter features a double compound hinge that offers added leverage over the standard riveting tool. This makes it an upgraded option for driving fasteners in even the most demanding applications. It includes five interchangeable nose pieces that are color-coded for easy identification and don’t require additional tools to install or remove. It also utilizes a collection bottle that’s built into the back end of the tool, collecting used mandrels rather than dropping them all over your workshop.

Advertisement

The riveter also features an ergonomic PVC grip construction that helps to reduce fatigue as you work on fastening tasks. The tool is available for $25 at Harbor Freight, making it a cost-effective option for securing all manner of material as you work to complete renovation or construction jobs. Buyers give it high praise as well, delivering a 4.8-star average rating across more than 530 reviews. It also has a 97% recommendation rate from those who have purchased the tool.

Advertisement

Icon 1/2-Inch Drive 18-Inch Professional Breaker Bar

The breaker bar is a tool mechanics and many others lean on in a pinch, both literally and figuratively. The breaker bar is an elongated ratchet-type tool that offers more leverage to create additional torque on a stubborn fastener. It’s a tool that’s all about muscling through a tough turning task, and so a heavy-duty option is always going to be a priority. The Icon 1/2-Inch Drive 18-Inch Professional Breaker Bar delivers a critically important crossover between strength and durability, and an approachable price point. It features more than enough length and strength to tackle seized fasteners, while being listed at Harbor Freight for $30.

The tool runs with a 1/2-inch drive size and features a 180-degree pivot in the head to deliver access to your workpiece from a range of angles. The tool is made from chrome vanadium steel and weighs a little over 2 pounds. It’s also chrome-plated to resist corrosion and rust. Nearly 170 reviewers have given it a 4.9-star average rating, with 99% of them recommending it to others.

Advertisement

Pittsburgh Metric or SAE Ball End Hex Key Set (13-Piece)

Allen wrenches come in many formats, but anyone who has used the standard straight-end devices for long enough will know that the typical unit comes with a notable disadvantage. One of the most useful changes that has taken place in these fastener-driving tools is the ball end. The short side of the L-shaped Pittsburgh Metric or SAE Ball End Hex Key Set features the standard square end for a firm connection with your fastener and plenty of driving force as you turn the tool. The longer side of the hex key features a ball-shaped tip that allows users to turn fasteners while the tool is positioned at an angle.

This enhances the reach you experience while trying to tighten a screw, with the ability to engage a screw head from a 25-degree offset. This set comes with all of the standard sizes you’d expect in either SAE or metric measurements. Both options are listed as in-store-only tools, and both are available for $6. The tool set has been reviewed by nearly 1,800 Harbor Freight buyers and holds a 4.7-star average rating with a 96% recommendation rate. There are other Allen wrench sets out there for cheaper, but there’s really no reason to settle for a standard model when the enhanced variant can be found for such a low price.

Advertisement

Fasten-Pro Hammer Tacker

The Fasten-Pro Hammer Tacker is an update to the classic staple gun. Hammer tackers are designed for faster installation with less physical effort than their alternatives. They fire the same staples that a staple gun can accommodate but achieve that result with significantly less hullabaloo, allowing you to work quicker without nearly as much fatigue. This Fasten-Pro model is available from Harbor Freight for $15 and features a 4.4-star average rating across over 510 reviews. Similarly, 90% of customers recommend it to others, with its price coming in as a key strength.

The tool features a spring-loaded strip magazine that continuously pushes the next staple down into the ready position. The tool is 13-1/2 inches long and weighs a hair over two pounds, making it a mobile solution that’s easy to carry and even easier to deploy. It utilizes a non-slip comfort grip with additional oil-resistant features that make it usable in a wide range of situations.

Advertisement

Pittsburgh Precision Electrical Screwdriver Set (6-Piece)

The precision screwdriver is essential for tackling electrical repair and handling access to small parts within all manner of objects you might be working to service or repair. Everything from battery changes in smoke detectors or children’s toys to delicate work under the hood of your project vehicle can benefit from the addition of a precision screwdriver set. The Pittsburgh Precision Electrical Screwdriver Set is a 6-piece solution that’s available from Harbor Freight for $8. This option features electrical insulation with a protection rating up to 1000V. Failing to use insulated tools while working on the wiring in your home or other electrical projects is an easy mistake to make since the consequences aren’t immediately apparent.

Advertisement

The insulated, soft-grip handles make it more comfortable to use the screwdrivers, and they feature GS and VDE-certified protection that can minimize or even negate the risk of electrical shock. The kit also comes with a storage case, and each screwdriver features a color-coded element to help make identification easier. The set features three flathead screwdrivers and three Phillips models, all in small sizes that are essential for tackling delicate tasks like jewelry or watch repair and much more. The low price tag combines perfectly with a 4.8-star average rating and a 99% recommendation rate across more than 520 reviews.

Advertisement

Pittsburgh Ratcheting Screwdriver Set (26-Piece)

There are a wide range of great multibit screwdrivers that can make a difference in your workflow, and the multibit screwdriver is among the most valuable and useful hand tools you’ll encounter. Therefore, it can feel like a big decision when selecting one. Fortunately, Harbor Freight offers a solid choice in the Pittsburgh Ratcheting Screwdriver Set. It features a $14 price tag and a 4.5-star average rating across over 1,560 reviews. It also has a 93% recommendation rate, underpinning a cost-effective and highly rated solution that can help users tackle a wide range of jobs. The tool comes with 24 driver bits as well as a nylon carrying case to keep everything contained and organized as an on-the-go solution or for effective storage in your workshop between jobs. It includes six nut-driving bits as well as a range of Phillips, Pozidrive, TORX, and slotted bits.

Each bit is constructed from chrome vanadium steel to deliver long-lasting performance across numerous jobs, regardless of how demanding each use might become. The screwdriving head also features a ratcheting function with left and right directional shifts as well as a stationary center position that locks the mechanism for standard use.

Advertisement

Methodology

All of these tools have been reviewed by at least 150 buyers. Most have near-perfect average ratings, with the lowest of the bunch scoring a 4.4-star average. They represent a quality cross-section of hand tool options that can provide plenty of versatility throughout the typical installation, repair, or fabrication tasks you might face. They’re also all inexpensive options that won’t break the bank as you search for good value at fair prices.

Advertisement



Source link

Continue Reading
Click to comment

You must be logged in to post a comment Login

Leave a Reply

Tech

Galway PhD student on what led to her discovery of new exoplanet

Published

on

‘Wispit 2C’ is estimated to be about 5m years old and most likely 10 times the mass of Jupiter.

Galway native Chloe Lawlor has discovered a new planet – the second one to be found forming near an infant star called ‘Wispit 2’, some 437 light years away.

As a child, Lawlor wanted to be an artist, she tells SiliconRepublic.com. However, she changed her mind once she joined university. “I moved into physics because I did like physics in school, so I thought, ‘Oh, maybe I’ll just try this out.’”

The 25-year-old says discoveries such as these feed the innate curiosity humans have in wanting to know how we came to be, how we evolved and why we are here. Lawlor is a PhD student at the University of Galway’s Centre for Astronomy at the School of Natural Sciences and the Ryan Institute.

Advertisement

She is working in collaboration with project lead Richelle van Capelleveen, a PhD student from Leiden Observatory in the Netherlands, and postdoctoral researcher Guillaume Bourdarot from the Max Planck Institute for Extraterrestrial Physics in Germany, to learn more about young planets and how they’re forming.

“Most of the planets that we’ve observed have been much older,” Lawlor says. “We don’t know how they get to those sort of final stages like something like our solar system. This is really key for these formation theories and it’s hopefully going to tell us a lot about these young systems, how they’re forming, and then how they evolve.”

Lawlor’s new discovery, an exoplanet named ‘Wispit 2C’, is thought to be about 5m years old. ‘Wispit 2B’, a nearby planet, was discovered last year by van Capelleveen and Dr Laird Close from the University of Arizona.

Both these exoplanets are at early stages of formation in the disc around Wispit 2, which is located in the Constellation of the Eagle, a prominent equatorial constellation visible in the northern hemisphere summer along the Milky Way.

Advertisement

Lawlor’s discovery makes Wispit 2 the second known young and still forming multi-planet system. The only other system yet discovered with more than one planet developing is PDS 70, some 400 light years away from Earth.

Wispit 2C is a gas giant, likely around 10 times the mass of Jupiter. It is twice as massive as Wispit 2B and orbits four-times closer to its host star, which makes it incredibly difficult to detect with ground-based telescopes.

A mostly black space with a hazy white gaseous looking ring in the middle. A graphic is used to circle the object in the centre of the ring.

Wispit 2B and Wispit 2C forming around Wispit 2. Image: ESO/C Lawlor, R F van Capelleveen et al.

The exoplanet was detected using the European Southern Observatory’s Very Large Telescope in Chile’s Atacama desert. By linking several telescopes together to act as one giant instrument, the research team was able to observe regions very close to the star. In their analysis, the team was able to detect carbon monoxide gas, a chemical commonly found in the atmospheres of young giant planets.

Advertisement

Lawlor said earlier this week: “At first, we weren’t sure if it was a planet or a very large dust clump. We very quickly made follow-up observations using the Very Large Telescope Interferometer, an incredible setup where multiple telescopes can be connected to form a large virtual telescope.

“This allowed us to take what we call a spectrum, which is essentially a chemical fingerprint revealing the elements and molecules in an object’s atmosphere.”

Lawlor led the study, which has been published in The Astrophysical Journal Letters.

Prof Frances Fahy, director of the Ryan Institute, said: “The discovery of the planet Wispit 2C is a remarkable achievement and highlights the world-class astrophysics research taking place at University of Galway.”

Advertisement

The team will continue on with their efforts to hopefully find more planets in the system.

Last year, a study from Scotland’s University of St Andrews showed how giant free-floating planets could make their own miniature planetary systems without needing a star to orbit around. In a different study from 2025, scientists – for the first time – observed the very early stages of the creation of a new solar system around a baby star.

Don’t miss out on the knowledge you need to succeed. Sign up for the Daily Brief, Silicon Republic’s digest of need-to-know sci-tech news.

Advertisement

Source link

Continue Reading

Tech

New Harry Potter TV Series Trailer Reintroduces Hogwarts Magic for HBO

Published

on

Go ahead and get ready to cue your inevitable comparisons: the new trailer for HBO’s Harry Potter series dropped on Wednesday, giving audiences a first look at muggles, magical Hogwarts students and The Boy Who Lived. Due to hit HBO and HBO Max for Christmas 2026, the TV show will be a direct adaptation of the wizarding books, starting off with The Philosopher’s Stone.

To fans familiar with the movie franchise, this may feel like a rediscovery — or reintroduction — to the live-action version of the world of Harry Potter. The trailer shows a young Harry and his signature scar, his tyrant of an aunt and the moment he received his invitation from Hogwarts. Take a look at the first meet with Hagrid, tender moments with Ron and Hermione, and a look at Lucius Malfoy and Snape. 

The series features Dominic McLaughlin as Harry Potter, Arabella Stanton as Hermione Granger and Alastair Stout 
as Ron Weasley, and the expansive cast also includes Nick Frost as Hagrid, John Lithgow as Hogwarts headmaster Dumbledore and Paapa Essiedu as Professor Snape. 

Advertisement

Source link

Continue Reading

Tech

Nintendo might charge less for digital Switch games?

Published

on

Nintendo might finally be doing something gamers have been asking for… forever. The company has officially confirmed that Switch 2 games will have different pricing for digital and physical versions, with digital copies expected to be cheaper.

The change begins in May 2026, starting with titles like Yoshi and the Mysterious Book. For example, early listings on the eShop show the game priced at $59.99 digitally vs $69.99 physically, marking a clear shift in how Nintendo handles game pricing.

Why is Nintendo doing this?

Let’s be real, physical games are comparatively expensive to make. Nintendo says the change reflects the higher costs of manufacturing and distributing cartridges, compared to digital downloads. This aligns with what the industry has been doing for years, except Nintendo has been one of the few holdouts where digital and physical games often cost the same.

There’s also a bigger strategy at play here. By making digital games cheaper, Nintendo could nudge more players toward digital purchases. That essentially translates to higher margins, fewer logistics headaches, and a tighter grip on its ecosystem. In other words, this isn’t just about fairness in pricing… It’s also about where Nintendo wants its future sales to go.

What does this mean for players?

So, does this mean all games going forward will have different prices? Well, not exactly, and this is where things get a little messy. While Nintendo is setting a lower MSRP for digital games, actual pricing can still vary depending on the title and retailer. Plus, not every game will follow the same pattern, so bigger releases could still carry higher price tags, making the gap between digital and physical a bit inconsistent.

For players, though, this is still a win. Digital games are finally getting a clear pricing advantage after years of being oddly equal (or sometimes pricier) than physical copies. That said, the trade-off remains. Physical games can be resold or shared, while digital ones stay locked to your account.

Advertisement

Source link

Continue Reading

Tech

Meta and YouTube Found Negligent in Landmark Social Media Addiction Case

Published

on

A jury found Meta and YouTube negligent in a landmark social media addiction case, ruling that addictive design features such as infinite scroll and algorithmic recommendations harmed a young user and contributed to her mental health distress. The verdict awards $3 million in compensatory damages so far and could pave the way for more lawsuits seeking financial penalties and product changes across the social media industry. “Meta is responsible for 70 percent of that cost and YouTube for the remainder,” notes The New York Times. “TikTok and Snap both settled with the plaintiff for undisclosed terms before the trial started.” From the report: The bellwether case, which was brought by a now 20-year-old woman identified as K.G.M., had accused social media companies of creating products as addictive as cigarettes or digital casinos. K.G.M. sued Meta, which owns Instagram and Facebook, and Google’s YouTube over features like infinite scroll and algorithmic recommendations that she claimed led to anxiety and depression.

The jury of seven women and five men will deliberate further to decide what further punitive damages the companies should pay for malice or fraud. The verdict in K.G.M.’s case — one of thousands of lawsuits filed by teenagers, school districts and state attorneys general against Meta, YouTube, TikTok and Snap, which owns Snapchat — was a major win for the plaintiffs. The finding validates a novel legal theory that social media sites or apps can cause personal injury. It is likely to factor into similar cases expected to go to trial this year, which could expose the internet giants to further financial damages and force changes to their products. The verdict also comes on the heels of a New Mexico jury ruling that found Meta liable for violating state law by failing to protect users of its apps from child predators.

Source link

Continue Reading

Tech

How xMemory cuts token costs and context bloat in AI agents

Published

on

Standard RAG pipelines break when enterprises try to use them for long-term, multi-session LLM agent deployments. This is a critical limitation as demand for persistent AI assistants grows.

xMemory, a new technique developed by researchers at King’s College London and The Alan Turing Institute, solves this by organizing conversations into a searchable hierarchy of semantic themes.

Experiments show that xMemory improves answer quality and long-range reasoning across various LLMs while cutting inference costs. According to the researchers, it drops token usage from over 9,000 to roughly 4,700 tokens per query compared to existing systems on some tasks.

For real-world enterprise applications like personalized AI assistants and multi-session decision support tools, this means organizations can deploy more reliable, context-aware agents capable of maintaining coherent long-term memory without blowing up computational expenses.

Advertisement

RAG wasn’t built for this

In many enterprise LLM applications, a critical expectation is that these systems will maintain coherence and personalization across long, multi-session interactions. To support this long-term reasoning, one common approach is to use standard RAG: store past dialogues and events, retrieve a fixed number of top matches based on embedding similarity, and concatenate them into a context window to generate answers.

However, traditional RAG is built for large databases where the retrieved documents are highly diverse. The main challenge is filtering out entirely irrelevant information. An AI agent’s memory, by contrast, is a bounded and continuous stream of conversation, meaning the stored data chunks are highly correlated and frequently contain near-duplicates.

To understand why simply increasing the context window doesn’t work, consider how standard RAG handles a concept like citrus fruit.

Imagine a user has had many conversations saying things like “I love oranges,” “I like mandarins,” and separately, other conversations about what counts as a citrus fruit. Traditional RAG may treat all of these as semantically close and keep retrieving similar “citrus-like” snippets. 

Advertisement

“If retrieval collapses onto whichever cluster is densest in embedding space, the agent may get many highly similar passages about preference, while missing the category facts needed to answer the actual query,” Lin Gui, co-author of the paper, told VentureBeat. 

A common fix for engineering teams is to apply post-retrieval pruning or compression to filter out the noise. These methods assume that the retrieved passages are highly diverse and that irrelevant noise patterns can be cleanly separated from useful facts.

This approach falls short in conversational agent memory because human dialogue is “temporally entangled,” the researchers write. Conversational memory relies heavily on co-references, ellipsis, and strict timeline dependencies. Because of this interconnectedness, traditional pruning tools often accidentally delete important bits of a conversation, leaving the AI without vital context needed to reason accurately.

Naive RAG vs strucured memory

Naive RAG vs structured memory (source: arXiv)

Advertisement

Why the fix most teams reach for makes things worse

To overcome these limitations, the researchers propose a shift in how agent memory is built and searched, which they describe as “decoupling to aggregation.”

Instead of matching user queries directly against raw, overlapping chat logs, the system organizes the conversation into a hierarchical structure. First it decouples the conversation stream into distinct, standalone semantic components. These individual facts are then aggregated into a higher-level structural hierarchy of themes.

When the AI needs to recall information, it searches top-down through the hierarchy, going from themes to semantics and finally to raw snippets. This approach avoids redundancy. If two dialogue snippets have similar embeddings, the system is unlikely to retrieve them together if they have been assigned to different semantic components.

For this architecture to succeed, it must balance two vital structural properties. The semantic components must be sufficiently differentiated to prevent the AI from retrieving redundant data. At the same time, the higher-level aggregations must remain semantically faithful to the original context to ensure the model can craft accurate answers.

Advertisement

A four-level hierarchy that shrinks the context window

The researchers developed xMemory, a framework that combines structured memory management with an adaptive, top-down search strategy.

xMemory continuously organizes the raw stream of conversation into a structured, four-level hierarchy. At the base are the raw messages, which are first summarized into contiguous blocks called “episodes.” From these episodes, the system distills reusable facts as semantics that disentangle the core, long-term knowledge from repetitive chat logs. Finally, related semantics are grouped together into high-level themes to make them easily searchable.

xmemory

xMemory architecture (source: arXiv)

xMemory uses a special objective function to constantly optimize how it groups these items. This prevents categories from becoming too bloated, which slows down search, or too fragmented, which weakens the model’s ability to aggregate evidence and answer questions.

Advertisement

When it receives a prompt, xMemory performs a top-down retrieval across this hierarchy. It starts at the theme and semantic levels, selecting a diverse, compact set of relevant facts. This is crucial for real-world applications where user queries often require gathering descriptions across multiple topics or chaining connected facts together for complex, multi-hop reasoning.

Once it has this high-level skeleton of facts, the system controls redundancy through what the researchers call “Uncertainty Gating.” It only drills down to pull the finer, raw evidence at the episode or message level if that specific detail measurably decreases the model’s uncertainty.

“Semantic similarity is a candidate-generation signal; uncertainty is a decision signal,” Gui said. “Similarity tells you what is nearby. Uncertainty tells you what is actually worth paying for in the prompt budget.” It stops expanding when it detects that adding more detail no longer helps answer the question.

What are the alternatives?

Existing agent memory systems generally fall into two structural categories: flat designs and structured designs. Both suffer from fundamental limitations.

Advertisement

Flat approaches such as MemGPT log raw dialogue or minimally processed traces. This captures the conversation but accumulates massive redundancy and increases retrieval costs as the history grows longer.

Structured systems such as A-MEM and MemoryOS try to solve this by organizing memories into hierarchies or graphs. However, they still rely on raw or minimally processed text as their primary retrieval unit, often pulling in extensive, bloated contexts. These systems also depend heavily on LLM-generated memory records that have strict schema constraints. If the AI deviates slightly in its formatting, it can cause memory failure.

xMemory addresses these limitations through its optimized memory construction scheme, hierarchical retrieval, and dynamic restructuring of its memory as it grows larger.

When to use xMemory

For enterprise architects, knowing when to adopt this architecture over standard RAG is critical. According to Gui, “xMemory is most compelling where the system needs to stay coherent across weeks or months of interaction.”

Advertisement

Customer support agents, for instance, benefit greatly from this approach because they must remember stable user preferences, past incidents, and account-specific context without repeatedly pulling up near-duplicate support tickets. Personalized coaching is another ideal use case, requiring the AI to separate enduring user traits from episodic, day-to-day details.

Conversely, if an enterprise is building an AI to chat with a repository of files, such as policy manuals or technical documentation, “a simpler RAG stack is still the better engineering choice,” Gui said. In those static, document-centric scenarios, the corpus is diverse enough that standard nearest-neighbor retrieval works perfectly well without the operational overhead of hierarchical memory.

The write tax is worth it

xMemory cuts the latency bottleneck associated with the LLM’s final answer generation. In standard RAG systems, the LLM is forced to read and process a bloated context window full of redundant dialogue. Because xMemory’s precise, top-down retrieval builds a much smaller, highly targeted context window, the reader LLM spends far less compute time analyzing the prompt and generating the final output.

In their experiments on long-context tasks, both open and closed models equipped with xMemory outperformed other baselines, using considerably fewer tokens while increasing task accuracy.

Advertisement
xMemory performance

xMemory increases performance on different tasks while reducing token costs (source: arXiv)

However, this efficient retrieval comes with an upfront cost. For an enterprise deployment, the catch with xMemory is that it trades a massive read tax for an upfront write tax. While it ultimately makes answering user queries faster and cheaper, maintaining its sophisticated architecture requires substantial background processing.

Unlike standard RAG pipelines, which cheaply dump raw text embeddings into a database, xMemory must execute multiple auxiliary LLM calls to detect conversation boundaries, summarize episodes, extract long-term semantic facts, and synthesize overarching themes.

Furthermore, xMemory’s restructuring process adds additional computational requirements as the AI must curate, link, and update its own internal filing system. To manage this operational complexity in production, teams can execute this heavy restructuring asynchronously or in micro-batches rather than synchronously blocking the user’s query.

Advertisement

For developers eager to prototype, the xMemory code is publicly available on GitHub under an MIT license, making it viable for commercial uses. If you are trying to implement this in existing orchestration tools like LangChain, Gui advises focusing on the core innovation first: “The most important thing to build first is not a fancier retriever prompt. It is the memory decomposition layer. If you get only one thing right first, make it the indexing and decomposition logic.”

Retrieval isn’t the last bottleneck

While xMemory offers a powerful solution to today’s context-window limitations, it clears the path for the next generation of challenges in agentic workflows. As AI agents collaborate over longer horizons, simply finding the right information won’t be enough.

“Retrieval is a bottleneck, but once retrieval improves, these systems quickly run into lifecycle management and memory governance as the next bottlenecks,” Gui said. Navigating how data should decay, handling user privacy, and maintaining shared memory across multiple agents is exactly “where I expect a lot of the next wave of work to happen,” he said.

Source link

Advertisement
Continue Reading

Tech

Liquid-cooled AI systems expose the limits of traditional storage architecture

Published

on

Presented by Solidigm


Liquid cooling is rewriting the rules of AI infrastructure, but most deployments have not fully crossed the line. GPUs and CPUs have moved to liquid cooling, while storage has depended on airflow, creating an operationally inefficient hybrid architecture.

What appears to be a pragmatic transition strategy is, in practice, a structural liability.

“A hybrid cooling approach is an operationally inefficient situation,” explains Hardeep Singh, thermal-mechanical hardware team manager at Solidigm. “You’re paying for and maintaining two entirely separate, expensive cooling infrastructures, and could be exposed to the worst-of-both-world’s problems.”

Advertisement

While liquid cooling requires pumps, fluid manifolds, and coolant distribution units (CDUs), air-cooled components require CRAC units, cold aisles, and evaporative cooling towers. Organizations moving to a hybrid solution by just adding some liquid cooling are absorbing the cost premium without capturing the full TCO benefit.

The thermal physics makes things worse. Bulky liquid-cooling cold plates, thick hoses, and manifolds physically obstruct airflow inside the GPU server chassis. This concentrates thermal stress on the remaining air-cooled components, including storage drives, memory, and network cards, because server fans cannot push adequate airflow around the liquid plumbing. The components most reliant on fans end up in the worst possible thermal environment.

Water consumption is an all-but ignored, equally serious problem. Traditional air-cooled components rely on server fans to move heat into ambient air, which is then absorbed by a water loop and pumped to evaporative cooling towers. These systems can consume millions of gallons of water over time. As rack power densities continue to climb to support modern AI workloads, the evaporative water penalty becomes, as Singh puts it, “environmentally and economically indefensible.”

As AI infrastructure evolves toward liquid-cooled and fanless GPU systems, the true constraints on scale are shifting from compute performance to system-level thermal design. Modern AI platforms are no longer built server by server; they are engineered as tightly integrated rack- and pod-level systems where power delivery, cooling distribution, and component placement are inseparable.

Advertisement

In this environment, storage architectures designed for airflow-dependent data centers are becoming a limiting factor. As GPU platforms move fully into shared liquid-cooling domains, anchored by rack-level CDUs, every component in the system must operate natively within the same thermal and mechanical design. Storage can no longer rely on isolated cooling paths or bespoke thermal assumptions without introducing inefficiency, complexity, or density trade-offs at the system level.

Why storage is no longer a passive subsystem

For infrastructure leaders, this marks a fundamental transition. Storage is no longer a passive subsystem attached to compute, but instead an active participant in system-level cooling, serviceability, and GPU utilization. The ability to scale AI now depends on whether storage can integrate cleanly into liquid-cooled GPU systems, without fragmenting cooling architectures or constraining rack-level design.

And the race to scale AI is no longer just about who has the most GPUs, but instead about who can keep them cool, says Scott Shadley, director of leadership narrative and evangelist at Solidigm.

“Finding a way to enable liquid-cooled storage while still making it user serviceable has been one of the biggest challenges in designing fanless system solutions,” Shadley says. “As AI workloads evolve, the pressure on storage will only intensify.”

Advertisement

Techniques like KV cache offload, which move data between GPU memory and high-speed storage during inference, make storage latency and thermal performance directly relevant to model serving efficiency. In these architectures, a storage subsystem that throttles due to poor traditional airflow under thermal load slows down both reads and the model itself.

Moving to integrated liquid cooling

Moving from traditional air-cooled GPU servers to integrated liquid-cooled racks improves power usage efficiency (PUE) and reduces the operational cost for the datacenter. It also replaces the noisy computer room air handler (CRAH) and introduces a modern, efficient liquid CDU with potential scope to eliminate chillers if racks can be cooled to a liquid temperature of 45° Celsius.

When storage is cooled through liquid in absence of fans, it must also support serviceability with no liquid leakage. It also creates a new requirement that many infrastructure teams are only beginning to grapple with: every component in the rack must operate natively within the same cooling architecture.

Storage as an active participant in system design

Storage design is no longer an isolated engineering problem. It is a direct variable in GPU utilization, system reliability, and operational efficiency. The solution is to redesign storage from the ground up for liquid-cooled, fanless environments. This is harder than it sounds. Traditional SSD design assumes airflow for thermal management and places components on both sides of a thermally insulated PCB. Neither assumption holds in a CDU-anchored architecture.

Advertisement

“SSDs need to be designed with a best-in-class thermal solution to specifically conduct heat from internal components efficiently and transfer it to fluid,” says Singh. “The design must include a low-resistance path for heat to transfer to a single cold plate attached on one side.”

At the same time, drives must support serviceability without liquid leakage during insertion and removal, and without degrading the thermal interface between the drive and the cold plate.

Solidigm has worked with NVIDIA to address SSD liquid-cooling challenges, such as hot swap-ability and single-side cooling, reducing the thermal footprint of storage within the shared liquid loop, and ensuring GPUs receive their proportional share of coolant.

“If storage is not designed for a liquid-cooled environment efficiently, it will either throttle to lower performance or require more liquid volume,” he says. “Which directly and indirectly leads to under-utilization of GPU capability.”

Advertisement

Alignment on standards and the path to interoperability

Solidigm is not working on this in isolation. The broader industry is coalescing around standards to ensure liquid-cooled AI systems are interoperable rather than a patchwork of custom solutions. The SNIA and the Open Compute Project (OCP) are the primary bodies driving this work.

Solidigm led the industry standard for liquid cooling in SFF-TA-1006 for the E1.S form factor and is an active participant in OCP work streams covering rack design, thermal management, and sustainability. Custom, bespoke cooling solutions for storage are giving way to standards-aligned, production-ready designs that integrate cleanly into liquid-cooled GPU platforms.

“There are several organizations involved in this work,” says Shadley, who is also a SNIA board member. “They started with component-level solutions, driven heavily by SNIA and the SFF TA TWG. The next level is solution-level work, which is currently being heavily driven by OCP.”

Solidigm’s roadmap is leading the way

The design rules for system level architectures have changed due to the advent of liquid and immersion cooling technologies that allow for more unique design rules and removal of some barriers. The ability for systems to drive NVMe SSD-only platforms also allows for the removal of the platter-based box constraint that exists with HDD solutions, Shadley says.

Advertisement

“Solidigm customers have an active and lead role in roadmap decisions for our products due to their deep technical alignment with the ecosystem,” he says. “We do not simply make and sell products, we integrate, co-design, co-develop, and innovate with and alongside our partners, customers, and their customers.”

Adds Singh: “Solidigm’s key strength is innovation and customer-inspired system level engineering. This will continue to aggressively lead the way for liquid cooling adoption for storage.”


Sponsored articles are content produced by a company that is either paying for the post or has a business relationship with VentureBeat, and they’re always clearly marked. For more information, contact sales@venturebeat.com.

Source link

Advertisement
Continue Reading

Tech

New Emoji, Playlist Generator and More: All the New Features iOS 26.4 Brings to Your iPhone

Published

on

Apple released iOS 26.4 on Tuesday, March 24, about a week after the tech giant released iOS 26.3.1 (a), the company’s first Background Security Improvement update. The most recent update brings a slew of features to your iPhone, including new emoji and video podcasts.

Tech Tips

You can download iOS 26.4 now by going to Settings and tapping General. Next, select Software Update, tap Update Now and follow the prompts on your screen.

Here are some of the new features iOS 26.4 brings to your iPhone.

Advertisement

New emoji

An orca, distorted face and other emoji coming out of a smartphone.

All the new emoji iOS 26.4 brings to your iPhone.

CNET/Apple

With iOS 26.4, your iPhone gets eight new emoji. Those emoji include:

The Unicode Consortium is responsible for creating emoji, and it approved these new emoji in September as part of Unicode 17.0. But this is the first time the emoji are showing up on iPhones. 

Advertisement

Video podcasts come to Apple Podcasts

The iOS 26.4 update also brings video to your Podcasts app. To view these video podcasts, open the Podcasts app and start listening to an episode with the video player icon in the top right corner of the title card. Once you’re listening, open the media player and tap the Turn Video On button near the podcast’s progress bar. The podcast’s artwork will be replaced with the video. To turn the video off again, tap Turn Video Off and the podcast’s artwork will return.

Side-by-side screenshots of the Podcasts app. On the left we see a podcasts artwork and on the right we see that artwork replaced by a video.

Video podcasts are a fun addition to the Podcasts app.

Apple/Screenshots by CNET

Reduce some Liquid Glass effects across your device

Apple’s iOS 26.4 update adds another setting to minimize Liquid Glass effects across your device: Reduce Bright Effects. Here’s where to find this setting.

Advertisement

1. Tap Settings.
2. Tap Accessibility.
3. Tap Display & Text Size.
4. Scroll down the menu to find Reduce Bright Effects.

The Reduce Bright Effects option in the Display and Text Size settings menu.

Reduce Bright Effects can eliminate some Liquid Glass effects.

Apple/Screenshot by CNET

Apple says the setting will minimize highlighting and flashing when interacting with on-screen elements, such as buttons or the keyboard. So if you find certain flash elements annoying, you can now disable them. 

Advertisement

Playlist Playground in Apple Music

The iOS 26.4 update also introduces a new playlist generator for Apple Music subscribers called Playlist Playground. Apple says the feature can create a playlist based on your description. Once you enter your description, it will create a playlist with a title, tracklist and general description.

To access Playlist Playground, first you have to be an Apple Music subscriber. Then, open Apple Music and go to your Library. In your Library, you’ll see a new icon at the top of your screen with a plus and a few lines next to it. Tap this, and you’ll be prompted to describe your playlist.

Apple Music's Playlist Playground which can create a playlist for you based on your own desccription.

Playlist Playground can generate a playlist for you in no time.

Advertisement

Apple/Screenshots by CNET

Apple notes this feature is still in beta, so it might create unexpected results. So you might ask for a good gym mix and end up with some Whitney Houston — but who’s to say Whitney isn’t good gym music?

Find nearby concerts with the aptly named Concerts feature

iOS 26.4 brings a new Concerts feature to your Apple Music app. 

“Concerts helps you discover nearby shows from artists in your library and recommends new artists based on what you listen to,” Apple writes in the update’s description. That way, you can easily find nearby shows.

To find Concerts, tap the magnifying glass icon at the bottom of your Apple Music screen, then tap Concerts. The feature may ask for your location the first time you use it. Then you’ll see popular shows nearby, along with their dates, times and locations. Tapping into any of these shows gives you more information on the show, as well as a link to buy tickets.

Advertisement
The Concerts menu in Apple Music.

The Concerts tab in Apple Music makes it easy to see upcoming shows in your area.

Apple/Screenshot by CNET

Shazam works offline, kind of

With iOS 26.4, your Control Center’s Shazam app can work in more ways. Now, if you aren’t connected to the internet and use the Control Center app to identify a song, the app will eventually tell you the song’s identity once you’re back online. 

Ambient Music home screen widgets

Apple introduced two new Ambient Music widgets for your home screen with iOS 26.4. These widgets let you easily access the four Ambient Music playlists: Sleep, Chill, Productivity, Wellbeing. You can quickly turn on a relaxing playlist to unwind after a long day, or one to help you focus on the task at hand.

Advertisement
An iPhone widget for the Ambient Music feature.

The Ambient Music widget makes it easy to play music for just the right setting.

Apple/Screenshot by CNET

Apple introduced these playlists to your iPhone alongside iOS 18.4 in 2024. However, you could only access those playlists from your Control Center at the time. 

Let other adults in your Family pay for themselves

In iOS 26.4, other adults in your Family sharing group can now use their own payment instead of depending on the group organizer’s payment method. That means if you’re an adult and have a family sharing group with your own parents, siblings or other family members, you can now purchase a new game, movie or something else with your own information instead of using someone else’s information and then paying them back for using their money. 

Advertisement

This can be a helpful feature that allows you to avoid the hassle of paying someone else back for using their payment information. And if you’re the person whose card is always used, it can be a nice way to ensure others pay for their own stuff and don’t freeload off you. 

More caption options when viewing videos

With iOS 26.4, you can easily change the caption style while watching content in certain apps, such as Apple TV. 

To see these options, start playing a video, then tap the speech bubble icon in the bottom-right corner of your screen to open the subtitle menu. Tap Style, and you’ll see the subtitle options Classic (the default setting), Large Text, Outline Text and Transparent Background. So if you and a few others are watching something on your iPhone and want to make sure everyone can see the captions, you might choose Large Text.

Advertisement
The subtitle style menu.

You can adjust the subtitles in some apps thanks to iOS 26.4.

Apple/Screenshot by CNET

More control over wallpaper Collections

The iOS 26.4 update also gives you more control over which wallpaper Collections are on your iPhone. Now, if you go to Settings > Wallpaper > Add New Wallpaper, you can tap Get under Collections like Weather and Astronomy. 

If you want to delete a Collection from your device, tap the check mark to the right of the downloaded Collection, and the option to Remove from Gallery appears. Tap this to delete the Collection from your iPhone, saving you some precious space.

Advertisement
The option to Remove from Gallery is highlighted in the Add New Wallpaper menu.

You can remove wallpaper Collections from your iPhone if you want to save a little more space. 

Apple/Screenshot by CNET

Here are the release notes for iOS 26.4.

Apple Music

  • Playlist Playground (beta) generates a playlist from your description, complete with a title, description and tracklist.
  • Concerts helps you discover nearby shows from artists in your library and recommends new artists based on what you listen to.
  • Offline Music Recognition in Control Center identifies songs without an internet connection and delivers results automatically when you’re back online.
  • Ambient Music widget for Sleep, Chill, Productivity and Wellbeing brings curated playlists to the Home Screen.
  • Full-screen backgrounds give album and playlist pages a more immersive look.

Accessibility

  • Reduce bright effects setting minimizes bright flashes when tapping on elements like buttons.
  • Subtitle and caption settings are available from the captions icon while viewing media, making them easier to find, customize and preview.
  • Reduce Motion setting more reliably reduces the animations of Liquid Glass for users sensitive to on-screen motion.

This update also includes the following enhancements:

  • Support for AirPods Max 2.
  • 8 new emoji, including an orca, trombone, landslide, ballet dancer and distorted face, are available in the emoji keyboard.
  • Freeform gains advanced image creation and editing tools, and a premium content library, joining Apple Creator Studio.
  • Mark reminders as urgent from the Quick Toolbar or by touching and holding, and filter for urgent reminders in your Smart Lists.
  • Purchase Sharing lets adult members in Family Sharing groups use their own payment method when making purchases, without relying on the family organizer.
  • Improved keyboard accuracy when typing quickly.

For more iOS news, check out what features were included in iOS 26.3 and iOS 26.2. You can also take a look at our iOS 26 cheat sheet for other tips and tricks.

Advertisement

Watch this: Don’t Wait: iOS 26.4 Brings New Emoji, Keyboard Fixes, AI Playlists

Source link

Continue Reading

Tech

Upcoming 3% Big Tech tax in Poland may leave Apple some wiggle room

Published

on

Poland is moving ahead with a digital services tax aimed at Big Tech revenue, but the structure leaves enough room for companies like Apple to argue they don’t fully qualify.

Blue European Union flag waving on tall metal pole, yellow stars in a circle surrounding stylized white airplane symbol, with modern glass office building in the background
Poland moves to tax Big Tech companies

The country will draft a digital services tax bill with a 3% levy on revenue from online advertising, user platforms, and data-driven services. Poland’s bill targets companies with over $1.16 billion globally and about $6.8 million within Poland.
Apple and other major U.S. tech firms fall within those thresholds and would be affected.
Continue Reading on AppleInsider | Discuss on our Forums

Source link

Continue Reading

Tech

Cloudflare’s new Dynamic Workers ditch containers to run AI agent code 100x faster

Published

on

Web infrastructure giant Cloudflare is seeking to transform the way enterprises deploy AI agents with the open beta release of Dynamic Workers, a new lightweight, isolate-based sandboxing system that it says starts in milliseconds, uses only a few megabytes of memory, and can run on the same machine — even the same thread — as the request that created it.

Compared with traditional Linux containers, the company says Dynamic Workers is roughly 100x faster to start and between 10x and 100x more memory efficient.

Cloudflare has spent months pushing what it calls “Code Mode,” the idea that large language models often perform better when they are given an API and asked to write code against it, rather than being forced into one tool call after another.

The company says converting an MCP server into a TypeScript API can cut token usage by 81%, and it is now positioning Dynamic Workers as the secure execution layer that makes that approach practical at scale.

Advertisement

For enterprise technical decision makers, that is the bigger story. Cloudflare is trying to turn sandboxing itself into a strategic layer in the AI stack. If agents increasingly generate small pieces of code on the fly to retrieve data, transform files, call services or automate workflows, then the economics and safety of the runtime matter almost as much as the capabilities of the model. Cloudflare’s pitch is that containers and microVMs remain useful, but they are too heavy for a future where millions of users may each have one or more agents writing and executing code constantly.

The history of modern isolated runtime environments

To understand why Cloudflare is doing this, it helps to look at the longer arc of secure code execution. Modern sandboxing has evolved through three main models, each trying to build a better digital box: smaller, faster and more specialized than the one before it.

The first model is the isolate. Google introduced the v8::Isolate API in 2011 so the V8 JavaScript engine could run many separate execution contexts efficiently inside the same process. In effect, a single running program could spin up many small, tightly separated compartments, each with its own code and variables.

In 2017, Cloudflare adapted that browser-born idea for the cloud with Workers, betting that the traditional cloud stack was too slow for instant, globally distributed web tasks. The result was a runtime that could start code in milliseconds and pack many environments onto a single machine. The trade-off is that isolates are not full computers. They are strongest with JavaScript, TypeScript and WebAssembly, and less natural for workloads that expect a traditional machine environment.

Advertisement

The second model is the container. Containers had been technically possible for years through Linux kernel features, but the company Docker turned them into the default software packaging model when it popularized them in 2013.

Containers solved a huge portability problem by letting developers package code, libraries and settings into a predictable unit that could run consistently across systems. That made them foundational to modern cloud infrastructure. But they are relatively heavy for the sort of short-lived tasks Cloudflare is talking about here. The company says containers generally take hundreds of milliseconds to boot and hundreds of megabytes of memory to run, which becomes costly and slow when an AI-generated task only needs to execute for a moment.

The third model is the microVM. Popularized by AWS Firecracker in 2018, microVMs were designed to offer stronger machine-like isolation than containers without the full bulk of a traditional virtual machine. They are attractive for running untrusted code, which is why they have started to show up in newer AI-agent systems such as Docker Sandboxes. But they still sit between the other two models: stronger isolation and more flexibility than an isolate, but slower and heavier as well.

That is the backdrop for Cloudflare’s pitch. The company is not claiming containers disappear, or that microVMs stop mattering. It is claiming that for a growing class of web-scale, short-lived AI-agent workloads, the default box has been too heavy, and the isolate may now be the better fit.

Advertisement

Cloudflare’s case against the container bottleneck

Cloudflare’s argument is blunt: for “consumer-scale” agents, containers are too slow and too expensive. In the company’s framing, a container is fine when a workload persists, but it is a bad fit when an agent needs to run one small computation, return a result and disappear. Developers either keep containers warm, which costs money, or tolerate cold-start delay, which hurts responsiveness. They may also be tempted to reuse a live sandbox across multiple tasks, which weakens isolation.

Dynamic Worker Loader is Cloudflare’s answer. The API allows one Worker to instantiate another Worker at runtime with code provided on the fly, usually by a language model. Because these dynamic Workers are built on isolates, Cloudflare says they can be created on demand, run one snippet of code, and then be thrown away immediately afterward. In many cases, they run on the same machine and even the same thread as the Worker that created them, which removes the need to hunt for a warm sandbox somewhere else on the network.

The company is also pushing hard on scale. It says many container-based sandbox providers limit concurrent sandboxes or the rate at which they can be created, while Dynamic Workers inherit the same platform characteristics that already let Workers scale to millions of requests per second. In Cloudflare’s telling, that makes it possible to imagine a world where every user-facing AI request gets its own fresh, isolated execution environment without collapsing under startup overhead.

Security remains the hardest part

Cloudflare does not pretend this is easy to secure. In fact, the company explicitly says hardening an isolate-based sandbox is trickier than relying on hardware virtual machines, and notes that security bugs in V8 are more common than those in typical hypervisors. That is an important admission, because the entire thesis depends on convincing developers that an ultra-fast software sandbox can also be safe enough for AI-generated code.

Advertisement

Cloudflare’s response is that it has nearly a decade of experience doing exactly that. The company points to automatic rollout of V8 security patches within hours, a custom second-layer sandbox, dynamic cordoning of tenants based on risk, extensions to the V8 sandbox using hardware features like MPK, and research into defenses against Spectre-style side-channel attacks. It also says it scans code for malicious patterns and can block or further sandbox suspicious workloads automatically. Dynamic Workers inherit that broader Workers security model.

That matters because without the security story, the speed story sounds risky. With it, Cloudflare is effectively arguing that it has already spent years making isolate-based multi-tenancy safe enough for the public web, and can now reuse that work for the age of AI agents.

Code Mode: from tool orchestration to generated logic

The release makes the most sense in the context of Cloudflare’s larger Code Mode strategy. The idea is simple: instead of giving an agent a long list of tools and asking it to call them one by one, give it a programming surface and let it write a short TypeScript function that performs the logic itself. That means the model can chain calls together, filter data, manipulate files and return only the final result, rather than filling the context window with every intermediate step. Cloudflare says that cuts both latency and token usage, and improves outcomes especially when the tool surface is large.

The company points to its own Cloudflare MCP server as proof of concept. Rather than exposing the full Cloudflare API as hundreds of individual tools, it says the server exposes the entire API through two tools — search and execute — in under 1,000 tokens because the model writes code against a typed API instead of navigating a long tool catalog.

Advertisement

That is a meaningful architectural shift. It moves the center of gravity from tool orchestration toward code execution. And it makes the execution layer itself far more important.

Why Cloudflare thinks TypeScript beats HTTP for agents

One of the more interesting parts of the launch is that Cloudflare is also arguing for a different interface layer. MCP, the company says, defines schemas for flat tool calls but not for programming APIs. OpenAPI can describe REST APIs, but it is verbose both in schema and in usage. TypeScript, by contrast, is concise, widely represented in model training data, and can communicate an API’s shape in far fewer tokens.

Cloudflare says the Workers runtime can automatically establish a Cap’n Web RPC bridge between the sandbox and the harness code, so a dynamic Worker can call those typed interfaces across the security boundary as if it were using a local library. That lets developers expose only the exact capabilities they want an agent to have, without forcing the model to reason through a sprawling HTTP interface.

The company is not banning HTTP. In fact, it says Dynamic Workers fully support HTTP APIs. But it clearly sees TypeScript RPC as the cleaner long-term interface for machine-generated code, both because it is cheaper in tokens and because it gives developers a narrower, more intentional security surface.

Advertisement

Credential injection and tighter control over outbound access

One of the more practical enterprise features in the release is globalOutbound, which lets developers intercept every outbound HTTP request from a Dynamic Worker. They can inspect it, rewrite it, inject credentials, respond to it directly, or block it entirely. That makes it possible to let an agent reach outside services while never exposing raw secrets to the generated code itself.

Cloudflare positions that as a safer way to connect agents to third-party services requiring authentication. Instead of trusting the model not to mishandle credentials, the developer can add them on the way out and keep them outside the agent’s visible environment. In enterprise settings, that kind of blast-radius control may matter as much as the performance gains.

More than a runtime: the helper libraries matter too

Another reason the announcement lands as more than a low-level runtime primitive is that Cloudflare is shipping a toolkit around it. The @cloudflare/codemode package is designed to simplify running model-generated code against AI tools using Dynamic Workers. At its core is DynamicWorkerExecutor(), which sets up a purpose-built sandbox with code normalization and direct control over outbound fetch behavior. The package also includes utility functions to wrap an MCP server into a single code() tool or generate MCP tooling from an OpenAPI spec.

The @cloudflare/worker-bundler package handles the fact that Dynamic Workers expect pre-bundled modules. It can resolve npm dependencies, bundle them with esbuild, and return the module map the Worker Loader expects. The @cloudflare/shell package adds a virtual filesystem backed by a durable Workspace using SQLite and R2, with higher-level operations like read, write, search, replace, diff and JSON update, plus transactional batch writes.

Advertisement

Taken together, those packages make the launch feel much more complete. Cloudflare is not just exposing a fast sandbox API. It is building the surrounding path from model-generated logic to packaged execution to persistent file manipulation.

Isolates versus microVMs: two different homes for agents

Cloudflare’s launch also highlights a growing split in the AI-agent market. One side emphasizes fast, disposable, web-scale execution. The other emphasizes deeper, more persistent environments with stronger machine-like boundaries.

Docker Sandboxes is a useful contrast. Rather than using standard containers alone, it uses lightweight microVMs to give each agent its own private Docker daemon, allowing the agent to install packages, run commands and modify files without directly exposing the host system. That is a better fit for persistent, local or developer-style environments. Cloudflare is optimizing for something different: short-lived, high-volume execution on the global web.

So the trade-off is not simply security versus speed. It is depth versus velocity. MicroVMs offer a sturdier private fortress and broader flexibility. Isolates offer startup speed, density and lower cost at internet scale. That distinction may become one of the main dividing lines in agent infrastructure over the next year.

Advertisement

Community reaction: hype, rivalry and the JavaScript catch

The release also drew immediate attention from developers on X, with reactions that captured both excitement and skepticism.

Brandon Strittmatter, a Cloudflare product lead and founder of Outerbase, called the move “classic Cloudflare,” praising the company for “changing the current paradigm on containers/sandboxes by reinventing them to be lightweight, less expensive, and ridiculously fast.”

Zephyr Cloud CEO Zack Chapple called the release “worth shouting from the mountain tops.”

But the strongest caveat surfaced quickly too: this system works best when the agent writes JavaScript. Cloudflare says Workers can technically run Python and WebAssembly, but that for small, on-demand snippets, “JavaScript will load and run much faster.”

Advertisement

That prompted criticism from YouTuber and ThursdAI podcast host Alex Volkov, who wrote that he “got excited… until I got here,” reacting to the language constraint.

Cloudflare’s defense is pragmatic and a little provocative. Humans have language loyalties, the company argues, but agents do not. In Cloudflare’s words, “AI will write any language you want it to,” and JavaScript is simply well suited to sandboxed execution on the web. That may be true in the narrow sense the company intends, but it also means the platform is most naturally aligned with teams already comfortable in the JavaScript and TypeScript ecosystem.

The announcement also triggered immediate competitive positioning. Nathan Flurry of Rivet used the moment to contrast his Secure Exec product as an open-source alternative that supports a broader range of platforms including Vercel, Railway and Kubernetes rather than being tied closely to Cloudflare’s own stack.

That reaction is worth noting because it shows how quickly the sandboxing market around agents is already splitting between vertically integrated platforms and more portable approaches.

Advertisement

Early use cases: AI apps, automations and generated platforms

Cloudflare is pitching Dynamic Workers for much more than quick code snippets. The company highlights Code Mode, AI-generated applications, fast development previews, custom automations and user platforms where customers upload or generate code that must run in a secure sandbox.

One example it spotlights is Zite, which Cloudflare says is building an app platform where users interact through chat while the model writes TypeScript behind the scenes to build CRUD apps, connect to services like Stripe, Airtable and Google Calendar, and run backend logic. Cloudflare quotes Zite CTO and co-founder Antony Toron saying Dynamic Workers “hit the mark” on speed, isolation and security, and that the company now handles “millions of execution requests daily” using the system.

Even allowing for vendor framing, that example gets at the company’s ambition. Cloudflare is not just trying to make agents a bit more efficient. It is trying to make AI-generated execution environments cheap and fast enough to sit underneath full products.

Pricing and availability

Dynamic Worker Loader is now in open beta and available to all users on the Workers Paid plan. Cloudflare says dynamically loaded Workers are priced at $0.002 per unique Worker loaded per day, in addition to standard CPU and invocation charges, though that per-Worker fee is waived during the beta period. For one-off code generation use cases, the company says that cost is typically negligible compared with the inference cost of generating the code itself.

Advertisement

That pricing model reinforces the larger thesis behind the product: that execution should become a small, routine part of the agent loop rather than a costly special case.

The bigger picture

Cloudflare’s launch lands at a moment when AI infrastructure is becoming more opinionated. Some vendors are leaning toward long-lived agent environments, persistent memory and machine-like execution. Cloudflare is taking the opposite angle. For many workloads, it argues, the right agent runtime is not a persistent container or a tiny VM, but a fast, disposable isolate that appears instantly, executes one generated program, and vanishes.

That does not mean containers or microVMs go away. It means the market is starting to split by workload. Some enterprises will want deeper, more persistent environments. Others — especially those building high-volume, web-facing AI systems — may want an execution layer that is as ephemeral as the requests it serves.

Cloudflare is betting that this second category gets very large, very quickly. And if that happens, Dynamic Workers may prove to be more than just another Workers feature. They may be Cloudflare’s attempt to define what the default execution layer for internet-scale AI agents looks like.

Advertisement

Source link

Continue Reading

Tech

Razer’s Nikke collab finally lets you arm your rifle-wielding waifu with a cat-eared gamer headset

Published

on


  • Goddess of Victory: Nikke is getting a Razer collaboration
  • It includes the ability to unlock a new character skin featuring the brand’s Razer Kraken Kitty V2 BT headset
  • There will also be pop-up events at some Razer stores

Goddess of Victory: Nikke publisher Level Infinite has revealed a new collaboration with gaming hardware giant Razer that brings one of the brand’s cutest headsets to the mobile game.

Starting on March 26, 2026, players will be able to unlock the new Punky Street skin for the character Viper by working their way through the limited-time Punky Street Pass. The skin decks out Viper in trendy streetwear and a white Razer Kraken Kitty V2 BT wireless gaming headset, complete with cat ears and some custom pink decals.

Source link

Advertisement
Continue Reading

Trending

Copyright © 2025