Ancient Slashdot reader ewhac writes: The maintainers of the Python package `chardet`, which attempts to automatically detect the character encoding of a string, announced the release of version 7 this week, claiming a speedup factor of 43x over version 6. In the release notes, the maintainers claim that version 7 is, “a ground-up, MIT-licensed rewrite of chardet.” Problem: The putative “ground-up rewrite” is actually the result of running the existing copyrighted codebase and test suite through the Claude LLM. In so doing, the maintainers claim that v7 now represents a unique work of authorship, and therefore may be offered under a new license. Version 6 and earlier was licensed under the GNU Lesser General Public License (LGPL). Version 7 claims to be available under the MIT license.
The maintainers appear to be claiming that, under the Oracle v. Google decision, which found that cloning public APIs is fair use, their v7 is a fair use re-implementation of the `chardet` public API. However, there is no evidence to suggest their re-write was under “clean room” conditions, which traditionally has shielded cloners from infringement suits. Further, the copyrightability of LLM output has yet to be settled. Recent court decisions seem to favor the view that LLM output is not copyrightable, as the output is not primarily the result of human creative expression — the endeavor copyright is intended to protect. Spirited discussion has ensued in issue #327 on `chardet`s GitHub repo, raising the question: Can copyrighted source code be laundered through an LLM and come out the other end as a fresh work of authorship, eligible for a new copyright, copyright holder, and license terms? If this is found to be so, it would allow malicious interests to completely strip-mine the Open Source commons, and then sell it back to the users without the community seeing a single dime.
Grayson Shor, far right, at a recent Pacific Northwest Battery Collaborative meet up at a Seattle brewery on Capitol Hill. Shor launched the organization to help the sector build connections. (PNWBC Photo)
The collaborative’s launch in October 2024 was so popular it ran out of chairs and the group now caps RSVPs because venues keep maxing out. The nonprofit has hosted 1,400 attendees at 17 different events in Washington, Oregon and online. Shor’s latest project is helping create a battery-focused mini-series he describes as a hybrid between Anthony Bourdain’s “Parts Unknown” and “Cosmos.”
Who knew that energy storage devices could generate so much enthusiasm?
“Batteries are sexy right now,” Shor said.
Batteries are making electric vehicle adoption more attractive as they’ve become increasingly powerful and quicker to recharge. They’re ubiquitous given the pervasive use of phones and consumer electronics. And as electricity demand is spiking thanks to data centers and other energy users, they’re a relatively quick, affordable way to add more power to the grid.
Advertisement
“We are installing more grid batteries in 2025 than the total amount that existed globally just two years ago,” Shor said. “This isn’t just growth, it’s a total reimagining of how our economy is powered.”
A battery ecosystem emerges
Part of the crowd at the Pacific Northwest Battery Collaborative launch party, with founder Grayson Shor in the front row in a tie. (PNWBC Photo)
Shor has spent nearly a decade working on sustainability, circular economy and battery-related issues for organizations ranging from the U.S. Department of State to Amazon to startups. When the former diplomat landed in Seattle from the other Washington more than two years ago, he was impressed by the region’s battery sector.
That included startups in electric aviation, alternative chemistries such as sodium batteries, and next-generation silicon battery materials, plus R&D resources and support at the University of Washington’s Clean Energy Institute.
But he realized the industry lacked the connections to bring together companies, academics, entrepreneurs and investors, and set out to address it. The sector welcomes his efforts.
“I’ve paid attention to folks trying to knit together community, and for the Northwest battery innovation and application ecosystem, Grayson Shor has been an unrelenting force seeking to build and amplify our unique strengths,” said Dan Schwartz, founding director of the Clean Energy Institute.
Advertisement
Tom Gurski, founder of the plug-in hybrid vehicle startup Blue Dot Motorworks, has attended the group’s functions. “In a region famous for introverted personalities their events and happy hours are invaluable for breaking down silos and getting people to connect,” Gurski said.
Beyond building community, Shor is lobbying for support for local and state policies that promote the industry and get more batteries deployed in the state. The energy storage devices have important societal benefits, he said, including better electrical grid performance and helping meet power needs during peak demand.
‘The Battery Life’
Shor speaking at a Pacific Northwest Battery Collaborative event in Seattle during 2025 PNW Climate Week. (PNBC Photo)
Shor is also the co-founder and chief product officer for Buckstop, an “urban mining” startup helping recover critical minerals from waste electronics. He also volunteers as the policy and government affairs director for the Volta Foundation, the world’s largest battery industry association.
And there’s the TV series, called “The Battery Life.” Crews recently spent three days in the Seattle area filming the first episode, visiting the battery materials company Group14 Technologies and interviewing startups at the UW’s Clean Energy Test Beds.
“We’re doing walks through factories. We’re meeting with the CEOs and the inventors, diving deep into their technology,” Shor said. But the series also has “the ‘Carl Sagan vibe,’” he added, explaining “how does this technology actually impact humanity, and why does it matter to the average person?”
Advertisement
Additional episodes will be shot in Portland and Vancouver, B.C. The plan is to air the series later this year at energy events in Oregon and Las Vegas, plus other area venues.
Future Pacific Northwest Battery Collaborative plans include a job fair and fundraising gala. Shor also envisions a convention where the entrepreneurs and innovators could set up booths to show off their technologies. The ideas keep coming.
“This is playing my little role in trying to tackle climate change, to try to advance the energy transition,” he said. “It helps with equity, it helps with economic opportunity …. It makes me happy.”
Clocks come in many styles and sizes, with perhaps the most visually pleasing ones involving marbles. Watching these little spheres obey gravity and form clearly readable numbers on a clock has strong mesmerizing qualities. If you’re not into really big marble clocks, or cannot quite find the space for a desk-sized clock, then the tiny marble clock by [Jens] may be an option.
While he totally loved the massive marble clock that [Ivan Miranda] built, it is a massive contraption that’s hard to justify as a permanent installation. His take on the concept thus makes it as small as possible, by using a pick-and-place style arm to place the marbles instead. Although the marbles don’t do a lot of rolling this way, it’s decidedly more quiet, and replace the rumbling and click-clacking of marbles with the smooth motion of a robotic arm.
Another benefit of this clock is that it’s cheap to make, with a price tag of less than $23. A big part of this is the use of cheap SG90 micro servos, and a permanent magnet along with a mechanism that pushes the marble off said magnet. Perhaps the biggest issue with this clock is that the arm somewhat obscures the time while it’s moving around, but it’s definitely another interesting addition to the gallery of marble clocks.
Advertisement
We have previously seen such clocks built out of wood and brass as well as 3D-printed using pendulum mechanisms, which can be made pretty compact as well, albeit with a more analog vibe.
Scenario Modeling and Array Design for Non-Terrestrial Networks (NTNs)
Non-terrestrial networks (NTNs) using low earth orbit (LEO) satellites present unique technical challenges, from managing large satellite constellations to ensuring reliable communication links. In this webinar, we’ll explore how to address these complexities using comprehensive modeling and simulation techniques. Discover how to model and analyze satellite orbits, onboard antennas and arrays, transmitter power amplifiers (PAs), signal propagation channels, and the RF and digital receiver segments—all within an integrated workflow. Learn the importance of including every link component to achieve accurate, reliable system performance.
Highlights include:
Modeling large satellite constellations
Analyzing and visualizing time-varying visibility and link closure
Using graphical apps for antenna analysis and RF component design
Modeling PAs and digital predistortion
Simulating interference effects in communication links
Republican lawmakers in Utah have long been on the cutting edge of shitty policymaking when it comes to regulating the internet. The latest chapter in that legacy is a proposed tax on porn and adult content purchased in the state’s digital space.
Originally proposed by a pair of Republican lawmakers in the Utah state legislature earlier this year, Senate Bill (SB) 73 would levy a so-called “material harmful to minors” tax at 2 percent on revenues generated by the sale of online porn (it was originally 7 percent). Having been amended and passed through the state Senate with considerable support, SB 73 is on track to clear the hurdles of the House of Representatives and be signed into law by Gov. Spencer Cox, a Republican and staunch anti-pornography activist like the bill’s sponsors.
This activism from Gov. Cox and the sponsors of porn tax bill—Republican state Sen. Calvin R. Musselman and state Rep. Steve Eliason—could presage a far more corrosive and expansive campaign against civil liberties and key freedom of expression protections that cover sexually-related speech.
First off, SB 73 would fund a variety of efforts for Utah’s state government. Such efforts benefiting from the funds under the proposal would include enforcement efforts for the state’s social media and pornography age verification laws.
Advertisement
But the bill goes further, especially after several rounds of being amended in the Senate and the House to include the mention of web traffic sourced from virtual private networks (VPNs) and other proxies. This bill would make it illegal to circumvent content blocks implemented by platforms due to local age verification laws, making it punishable by a bevy of civil penalties. Nonetheless, what goes well beyond extreme is that there is a provision in the bill that would also make it illegal for websites covered by age verification laws (e.g., a porn site) to offer Utah-based users information about using VPNs to get around any content blocks securely.
Consider the following language in the current form of Senate Bill 73 regarding VPN “facilitation”:
“A commercial entity that operates a website that contains a substantial portion of material harmful to minors may not facilitate or encourage the use of a virtual private network, proxy server, or other means to circumvent age verification requirements, including by providing: (a) instructions on how to use a virtual private network or proxy server to access the website; or (b) means for individuals in this state to circumvent geofencing or blocking.”
Utah’s bill doesn’t go that far on the concerns of records, but it certainly conjures up civil liberties concerns. Aside from the glaring privacy concerns related to age verification tech, Utah has no right to restrict the communications of a private company to its customers. This goes double for attempts to supersede interstate commerce on a category of products and services that are lawful. And don’t forget the dimensions of the porn tax. SB 73’s approach is expansive and blatantly violates the First Amendment rights of millions of people, not just those who live within the state boundaries of Utah.
Advertisement
The tax is a textbook “sin tax” a jurisdiction would levy on something like alcohol, tobacco, and gambling. But what is different between the purchase of a six-pack of beer versus wanking off alone in your home is that buying that beer from the liquor store isn’t necessarily considered expressive in its nature. Producing, selling, and consuming pornography are matters of protected sexual speech so long nothing illegal and criminal occur. Porn taxes like the one proposed in SB 73 explicitly outline “covered entities,” to include all entities that sell adult content through clip sales, subscriptions, and fan sites. And with total Utah sales, revenues are then taxed at the 2 percent levy and then paid to the state each year.
This might be an incidental bump in the road for many of the larger platforms, like Pornhub or OnlyFans, but this type of policymaking is a vindictive ploy to make operating a small and medium business in this space excruciatingly harder. I do see the Utah bill passing this legislative session, which would lead to a potential legal standoff in a federal courthouse. But I am not holding my breath for anything more beyond that.
Michael McGrady covers the tech and legal sides of theonlinepornbusiness.
Video Friday is your weekly selection of awesome robotics videos, collected by your friends at IEEE Spectrum robotics. We also post a weekly calendar of upcoming robotics events for the next few months. Please send us your events for inclusion.
The functional replication and actuation of complex structures inspired by nature is a longstanding goal for humanity. Creating such complex structures combining soft and rigid features and actuating them with artificial muscles would further our understanding of natural kinematic structures. We printed a biomimetic hand in a single print process comprised of a rigid skeleton, soft joint capsules, tendons, and printed touch sensors.
This is our latest work on the trajectory planning method for floating-based articulated robots, enabling the global path searching in complex and cluttered environments.
OmniPlanner is a unified solution for exploration and inspection path planning (as well as target reach) across aerial, ground, and underwater robots. It has been verified through extensive simulations and a multitude of field tests, including in underground mines, ballast water tanks, forests, university buildings, and submarine bunkers.
In the ARISE project, the FZI Research Center for Information Technology and its international partners ETH Zurich, University of Zurich, University of Bern, and University of Basel took a major step toward future lunar missions by testing cooperative autonomous multi-robot teams under outdoor conditions.
Gugusse and the Automaton’ is a 1897 French film by Georges Méliès featuring a humanoid robot in nearly as realistic of a way as some of the humanoid promo videos we’ve seen lately.
Anca Dragan is no stranger to Waymo. She worked with us for six years while also at UC Berkeley and now, GoogleDeepMind. Her focus on making AI safer helped Waymo as it launched commercially. In this final episode of our season, Anca describes how her work enables AI agents to work fluently with people, based on human goals and values.
This UPenn GRASP SFI Seminar is by Junyao Shi, on “Unlocking Generalist Robots with Human Data and Foundation Models.”
Advertisement
Building general-purpose robots remains fundamentally constrained by data scarcity and labor-intensive engineering. Unlike vision and language, robotics lacks large, diverse datasets spanning tasks, environments, and embodiments, limiting both scalability and generalization. This talk explores how human data and foundation models trained at scale can help overcome these bottlenecks.
If you’ve spent any time following gaming news in early 2026, you might think the end of Xbox is right around the corner. Between reports of a 32% year-over-year drop in hardware revenue, the sudden departure of longtime Xbox boss Phil Spencer, and wild speculation that Microsoft might pivot the entire gaming division toward AI, the internet has been flooded with dramatic takes about the “death of Xbox.”
Official Xbox Podcast
But the eulogies are premature. Despite the noise, Xbox still sits on one of the most powerful portfolios in gaming, including Halo, Forza, Gears of War, Call of Duty, Minecraft, and more. Microsoft also has the financial backing, infrastructure, and studio network to remain a major player for decades. The real issue isn’t survival, but identity.
You see, for several years, Xbox leadership pushed an ambitious idea that “every screen is an Xbox.” The strategy expanded the brand through cloud gaming, PC integration, and Game Pass across multiple platforms. While that approach broadened reach, it also created confusion about what Xbox actually is. Now, under the new leadership of Microsoft Gaming CEO Asha Sharma, the company appears to be acknowledging that confusion and attempting a course correction.
Sharma recently confirmed Project Helix, the codename for Xbox’s next-generation hardware, promising a device that will “lead in performance and play your Xbox and PC games.” That announcement alone signals a shift in direction. Xbox isn’t ending, but it is entering a critical rebuilding phase. And if the company wants to return to its former glory, experts and players alike largely agree that three major changes are essential.
1. Nail the execution of Project Helix
One of the biggest challenges Xbox faces today is simple: many players aren’t sure why they should buy an Xbox console anymore.
If the same games appear on PC, and sometimes even on rival platforms, what makes the Xbox console special? That’s where Project Helix could become the most important product Microsoft has released in years. Rumored for a 2027 launch, Helix is expected to be a hybrid system, essentially a powerful AMD-powered console running a “console-ized” version of Windows. The promise is compelling: the simplicity of a traditional console combined with the flexibility of a gaming PC.
Imagine a device that boots straight into a controller-friendly interface but also lets players access platforms like Steam or Epic from the living room. If done right, Helix could blur the line between PC and console in a way no competitor currently offers. But execution will determine everything. Helix must never feel like a desktop computer awkwardly connected to a TV. Instead, it needs to launch into a seamless controller-first experience, as the “Xbox Full Screen Experience” we saw on the ROG Xbox Ally, preserving the plug-and-play simplicity that console players expect.
If Microsoft can successfully merge the PC and console ecosystems without sacrificing ease of use, Helix won’t just save Xbox hardware, but it could redefine what a console is. Yes, it’s likely going to be expensive, with rumors suggesting a price tag that could cross the $1,000 mark. But Xbox could still justify that premium if it delivers on the other two pillars that matter just as much.
2. Let the studios deliver the games
The second major fix is both obvious and unavoidable: Xbox needs more great games, more consistently.
Advertisement
Over the past decade, Microsoft has spent nearly $100 billion acquiring studios, including Bethesda and Activision Blizzard. On paper, that gives Xbox one of the strongest first-party lineups in gaming history. Yet the results have been uneven. Franchises like Halo, Gears of War, and Forza, once the backbone of the platform, have seen long development gaps. Meanwhile, studio closures, layoffs, and shifting corporate priorities have created uncertainty inside Microsoft’s gaming division.
Halo
To further add to the injury, when Sharma took over, some players worried that her background in AI-driven tech companies might push Xbox toward algorithm-generated content. Thankfully, she has quickly pushed back on that idea, stating that Microsoft will not “chase short-term efficiency or flood our ecosystem with soulless AI slop.” Now the company needs to prove it.
Xbox
Microsoft now owns some of the most talented developers in the world. What they need most is stability. Fewer shifting mandates, fewer corporate interruptions, and enough time to create the kind of system-defining games that drive entire console generations. Because ultimately, subscriptions and hardware don’t sell themselves. Great games do. The upcoming Forza Horizon 6 is already generating plenty of buzz and appears well on track to be a major success. However, Microsoft will need a steady stream of titles, especially strong exclusives, if it hopes to match the kind of consistent first-party momentum Sony has built on the PlayStation side.
3. Rebuild the culture around Xbox
Finally, there’s one part of the Xbox experience that often gets overlooked: the community culture. For many fans, the Xbox 360 era still feels like the golden age of the platform. Profiles felt personal, avatars actually mattered, and the dashboard felt like a social space where gamers could hang out. It wasn’t just a storefront pushing subscriptions and ads.
Xbox 360
Over time, much of that personality has disappeared. Today, the Xbox dashboard is often criticized for feeling cluttered with Game Pass promotions and advertisements. Across communities like Reddit, ResetEra, and Xbox Insider forums, the message from players is clear: bring back the personality. Fans want things like dynamic themes, meaningful achievement rewards, deeper avatar integration, and more ways to personalize the UI so the console feels like their space again.
Billy Freeman / Unsplash
Players are also asking Xbox to double down on something it once did better than anyone else: game preservation. The Backward Compatibility program was hugely popular, and with Activision Blizzard now under Microsoft’s umbrella, fans want to see classic titles return. If Xbox can become the place where decades of gaming history remain playable on modern hardware, it could turn preservation into one of its biggest strengths.
The road back
Long story short, Xbox isn’t going anywhere anytime soon. The brand still holds enormous influence in the gaming industry, backed by Microsoft’s resources and a massive network of studios and services. However, the platform is at a turning point.
For Xbox to truly thrive again, the solution isn’t chasing every new trend. It’s about focusing on the basics: delivering great games consistently, launching a strong next-generation hardware platform, and reconnecting with the community that built the brand. If Microsoft gets these fundamentals right, the “Xbox is dying” narrative could quickly fade, and the next chapter of Xbox might end up being its most exciting yet.
MSI MEG Vision X AI 13.3-inch touchscreen doubles as a monitoring hub for creatives and professionals
GPU selection dictates performance for gaming, rendering, and professional workloads alike
Lobster-like chassis combines expandability with unconventional aesthetics
MSI has launched the MEG Vision X AI series, a barebones all-in-one PC which combines high-end gaming hardware with a strikingly unconventional design.
The system features a full-size tower measuring 299.3mm wide, 502.7mm deep, and 423.4mm tall, weighing approximately 18.3kg, and a PS3-esque appendage and protrusions that suggest both function and a distinctive aesthetic.
The device includes a 13.3-inch touchscreen intended for system monitoring, quick toggles, or dedicated status displays, allowing creatives to access software shortcuts, monitor rendering progress, or adjust project settings without switching focus from their primary display.
Interactive touchscreen enhances workflow and monitoring
The unique look of this device promoted TechRadar Pro editor Desire Athow to quip the casing resembled, “a lobster that hadn’t completely shed its hard exoskeleton to grow,” capturing the layered and almost organic appearance of the chassis, emphasizing the sense of a device that is both protective and expandable, housing high-end components while presenting a unique surface.
MSI appears to have embraced this aesthetic to showcase the interactive touchscreen while accommodating a full-size tower structure capable of housing top-tier components.
The device is larger than regular compact all-in-one PCs, suggesting the company prioritizes cooling, power delivery, and expandability over minimalism.
Advertisement
Performance is anchored by Intel’s Core Ultra 7 265K CPU on a Z890 platform, paired with 64GB of DDR5 memory.
Sign up to the TechRadar Pro newsletter to get all the top news, opinion, features and guidance your business needs to succeed!
GPU options split the series into two clear tiers, a GeForce RTX 5080X configuration at $4,640 and a GeForce RTX 5070 Ti model at $4,082.
MSI indicates that CPU and RAM are consistent across models, meaning buyers make performance choices largely through GPU selection.
Advertisement
This ensures that professional applications like 3D rendering, video editing, and simulation software benefit from dedicated GPU acceleration alongside gaming performance.
The MEG Vision X AI supports both wired and wireless connections, with Intel Killer E5000 5GbE for the former and Wi-Fi 7 or Bluetooth 5.4 for the latter.
It also includes two Thunderbolt 4 ports, which support fast external storage, docking, or display expansion.
This connectivity allows professionals to attach high-speed NVMe drives or multi-monitor setups, which can streamline workflows for designers, animators, and video editors.
Advertisement
Power is supplied by an 850W 80 PLUS Gold PSU, providing adequate headroom for sustained GPU loads.
Although the primary audience for the device is gamers, its hardware and expandability suggest it could also serve as a versatile platform for creators who require both raw performance and reliable workstation capabilities.
Enterprise AI applications that handle large documents or long-horizon tasks face a severe memory bottleneck. As the context grows longer, so does the KV cache, the area where the model’s working memory is stored.
A new technique developed by researchers at MIT addresses this challenge with a fast compression method for the KV cache. The technique, called Attention Matching, manages to compact the context by up to 50x with very little loss in quality.
While it is not the only memory compaction technique available, Attention Matching stands out for its execution speed and impressive information-preserving capabilities.
The memory bottleneck of the KV cache
Large language models generate their responses sequentially, one token at a time. To avoid recalculating the entire conversation history from scratch for every predicted word, the model stores a mathematical representation of every previous token it has processed, also known as the key and value pairs. This critical working memory is known as the KV cache.
Advertisement
The KV cache scales with conversation length because the model is forced to retain these keys and values for all previous tokens in a given interaction. This consumes expensive hardware resources. “In practice, KV cache memory is the biggest bottleneck to serving models at ultra-long context,” Adam Zweiger, co-author of the paper, told VentureBeat. “It caps concurrency, forces smaller batches, and/or requires more aggressive offloading.”
In modern enterprise use cases, such as analyzing massive legal contracts, maintaining multi-session customer dialogues, or running autonomous coding agents, the KV cache can balloon to many gigabytes of memory for a single user request.
To solve this massive bottleneck, the AI industry has tried several strategies, but these methods fall short when deployed in enterprise environments where extreme compression is necessary. A class of technical fixes includes optimizing the KV cache by either evicting tokens the model deems less important or merging similar tokens into a single representation. These techniques work for mild compression but “degrade rapidly at high reduction ratios,” according to the authors.
Real-world applications often rely on simpler techniques, with the most common approach being to simply drop the older context once the memory limit is reached. But this approach causes the model to lose older information as the context grows long. Another alternative is context summarization, where the system pauses, writes a short text summary of the older context, and replaces the original memory with that summary. While this is an industry standard, summarization is highly lossy and heavily damages downstream performance because it might remove pertinent information from the context.
Advertisement
Recent research has proven that it is technically possible to highly compress this memory using a method called Cartridges. However, this approach requires training latent KV cache models through slow, end-to-end mathematical optimization. This gradient-based training can take several hours on expensive GPUs just to compress a single context, making it completely unviable for real-time enterprise applications.
How attention matching compresses without the cost
Attention Matching achieves high-level compaction ratios and quality while being orders of magnitude faster than gradient-based optimization. It bypasses the slow training process through clever mathematical tricks.
The researchers realized that to perfectly mimic how an AI interacts with its memory, they need to preserve two mathematical properties when compressing the original key and value vectors into a smaller footprint. The first is the “attention output,” which is the actual information the AI extracts when it queries its memory. The second is the “attention mass,” which acts as the mathematical weight that a token has relative to everything else in the model’s working memory. If the compressed memory can match these two properties, it will behave exactly like the massive, original memory, even when new, unpredictable user prompts are added later.
“Attention Matching is, in some ways, the ‘correct’ objective for doing latent context compaction in that it directly targets preserving the behavior of each attention head after compaction,” Zweiger said. While token-dropping and related heuristics can work, explicitly matching attention behavior simply leads to better results.
Advertisement
Before compressing the memory, the system generates a small set of “reference queries” that act as a proxy for the types of internal searches the model is likely to perform when reasoning about the specific context. If the compressed memory can accurately answer these reference queries, it will very likely succeed at answering the user’s actual questions later. The authors suggest various methods for generating these reference queries, including appending a hidden prompt to the document telling the model to repeat the previous context, known as the “repeat-prefill” technique. They also suggest a “self-study” approach where the model is prompted to perform a few quick synthetic tasks on the document, such as aggregating all key facts or structuring dates and numbers into a JSON format.
With these queries in hand, the system picks a set of keys to preserve in the compacted KV cache based on signals like the highest attention value. It then uses the keys and reference queries to calculate the matching values along with a scalar bias term. This bias ensures that pertinent information is preserved, allowing each retained key to represent the mass of many removed keys.
This formulation makes it possible to fit the values with simple algebraic techniques, such as ordinary least squares and nonnegative least squares, entirely avoiding compute-heavy gradient-based optimization. This is what makes Attention Matching super fast in comparison to optimization-heavy compaction methods. The researchers also apply chunked compaction, processing contiguous chunks of the input independently and concatenating them, to further improve performance on long contexts.
Attention matching in action
To understand how this method performs in the real world, the researchers ran a series of stress tests using popular open-source models like Llama 3.1 and Qwen-3 on two distinct types of enterprise datasets. The first was QuALITY, a standard reading comprehension benchmark using 5,000 to 8,000-word documents. The second, representing a true enterprise challenge, was LongHealth, a highly dense, 60,000-token dataset containing the complex medical records of multiple patients.
The key finding was the ability of Attention Matching to compact the model’s KV cache by 50x without reducing the accuracy, while taking only seconds to process the documents. To achieve that same level of quality previously, Cartridges required hours of intensive GPU computation per context.
Advertisement
Attention Matching with Qwen-3 (source: arXiv)
When dealing with the dense medical records, standard industry workarounds completely collapsed. The researchers noted that when they tried to use standard text summarization on these patient records, the model’s accuracy dropped so low that it matched the “no-context” baseline, meaning the AI performed as if it had not read the document at all.
Attention Matching drastically outperforms summarization, but enterprise architects will need to dial down the compression ratio for dense tasks compared to simpler reading comprehension tests. As Zweiger explains, “The main practical tradeoff is that if you are trying to preserve nearly everything in-context on highly information-dense tasks, you generally need a milder compaction ratio to retain strong accuracy.”
The researchers also explored what happens in cases where absolute precision isn’t necessary but extreme memory savings are. They ran Attention Matching on top of a standard text summary. This combined approach achieved 200x compression. It successfully matched the accuracy of standard summarization alone, but with a very small memory footprint.
Advertisement
One of the interesting experiments for enterprise workflows was testing online compaction, though they note that this is a proof of concept and has not been tested rigorously in production environments. The researchers tested the model on the advanced AIME math reasoning test. They forced the AI to solve a problem with a strictly capped physical memory limit. Whenever the model’s memory filled up, the system paused, instantly compressed its working memory by 50 percent using Attention Matching, and let it continue thinking. Even after hitting the memory wall and having its KV cache shrunk up to six consecutive times mid-thought, the model successfully solved the math problems. Its performance matched a model that had been given massive, unlimited memory.
There are caveats to consider. At a 50x compression ratio, Attention Matching is the clear winner in balancing speed and quality. However, if an enterprise attempts to push compression to extreme 100x limits on highly complex data, the slower, gradient-based Cartridges method actually outperforms it.
The researchers have released the code for Attention Matching. However, they note that this is not currently a simple plug-and-play software update. “I think latent compaction is best considered a model-layer technique,” Zweiger notes. “While it can be applied on top of any existing model, it requires access to model weights.” This means enterprises relying entirely on closed APIs cannot implement this themselves; they need open-weight models.
The authors note that integrating this latent-space KV compaction into existing, highly optimized commercial inference engines still requires significant effort. Modern AI infrastructure uses complex tricks like prefix caching and variable-length memory packing to keep servers running efficiently, and seamlessly weaving this new compaction technique into those existing systems will take dedicated engineering work. However, there are immediate enterprise applications. “We believe compaction after ingestion is a promising use case, where large tool call outputs or long documents are compacted right after being processed,” Zweiger said.
Advertisement
Ultimately, the shift toward mechanical, latent-space compaction aligns with the future product roadmaps of major AI players, Zweiger argues. “We are seeing compaction to shift from something enterprises implement themselves into something model providers ship,” Zweiger said. “This is even more true for latent compaction, where access to model weights is needed. For example, OpenAI now exposes a black-box compaction endpoint that returns an opaque object rather than a plain-text summary.”
Citizens and law enforcement officials alike would probably be quick to tell you that speeding drivers rank among the most dangerous issues they face on the roadways every day. While the onus of obeying speed limits on the road ultimately rests on the person in the driver’s seat, authorities are expected to help control excessive speeding by catching those drivers in the act and issuing citations as punishment.
That job is particularly tricky, as the number of officers on patrol is typically outnumbered greatly by the number of citizens at the wheel of their own vehicles. Some municipalities have, however, sought to tilt the situation in their favor by setting up speed traps. Similarly, traffic light cameras have become regular fixtures in helping monitor and control traffic patterns. Some local forces are taking matters a step further by using so-called “Speed Jeeps,” which are stationary, unmanned cruisers equipped with cameras to catch and ticket speeding drivers.
Advertisement
Commerce City, Colorado, has started rolling out such vehicles in March, with authorities in the Denver suburb looking to use them to help enforce speed limits in school zones, residential areas, and work zones. It remains to be seen how effective the move will be, as speed cameras have sometimes caused controversy in their alleged overreach. Still, according to Denver 7 News, some Commerce City residents are fully behind the use of Speed Jeeps if they help make their streets safer.
Advertisement
Here’s how Speed Jeeps actually work
Speed cameras are, of course, not legal in every city and state in the U.S. However, areas such as Montgomery County, Maryland, have effectively used them to control speeding in areas of concern. Commerce City has now joined the list of municipalities hoping to use tech to increase community safety, with its Speed Jeeps rotating between locations and adding mobility to the mix.
The fact that Speed Jeeps are designed to look like real police cruisers may make them even more effective than just their cameras, as few things will get a speeder to tap the brakes faster than the sight of a cop. The unmanned vehicles are, obviously, not designed to chase after speeders as a normal officer might. Instead, their cameras are activated when a speeding vehicle enters the range of its on-board radar gun. Once activated, the camera snaps a shot of the vehicle’s front end and driver. A separate camera then takes a shot of the rear license plate once the speeding car passes.
From there, local law enforcement will collect additional information about the alleged infraction and then decide whether to issue a citation. If deemed necessary, the citation will be mailed to the vehicle’s registered address. Upon receipt, the recipient will have a chance to either pay the fine or challenge the ruling in court.
Garry Duffy says that entrepreneurship should be taught at an undergraduate level.
Ireland is doubling down on building a strong research-to-market pipeline in the hopes of creating innovative global companies with homegrown roots.
To do this, Research Ireland has tapped leading universities across the country to deliver what its CEO, Diarmuid O’Brien, calls “one of the most proactive, imaginative and potentially disruptive programmes” in its history.
Last year, the Government announced three hubs to act as a funding mechanism, support system and testing ground for researchers attempting to commercialise their ideas.
Advertisement
Academics need this kind of support, says Garry Duffy, the director of the ARC Hub for Healthtech at the University of Galway, which officially launched just last month.
“Commercialisation is generally new to people – particularly researchers. And it’s a new language and it’s a new acumen, and you have to try and build that. And that’s what we’re really trying to do with the ARC Hub,” Duffy says. ARC, quite aptly, stands for ‘Accelerating Research to Commercialisation’.
With a backing of €34.3m from the Irish Government and the EU, the ARC health-tech hub is co-run by Atlantic Technological University and RCSI University of Medicine and Health Sciences, with other major institutions also taking part.
The Government announced two other hubs last year as well, one for therapeutics and one for ICT, boasting a combined funding that exceeded €60m.
Advertisement
The idea behind the hubs is to create a nurturing environment for entrepreneurial scientists and engineers to carry out research that will lead to commercial impact.
Duffy cites Dublin start-up ProVerum as a success story he would like to replicate in the health-tech hub he leads.
The 2016-founded Trinity College Dublin spin-out is the creator behind ‘ProVee’, a minimally invasive solution for treating benign prostatic hyperplasia.
ProVerum raised $80m in a Series B round last August. The start-up’s co-founder Ríona Ní Ghriallais is on the ARC health-tech advisory board.
Advertisement
The ARC Hub for Healthtech launched last month, with 23 projects across major areas – including sensors, implantables and AI – already in the pipeline.
Researchers, with the help of industry professionals, are creating commercial solutions for health issues such as hypertension management, ovarian cancer and falls among the elderly, Duffy says. Some projects have already generated clinical evidence to support the future impact of the various technologies.
The health-tech hub is also inviting around 22 new projects in its second call, which would give a total of around 45 projects under its remit.
Peter Power, the head of the European Commission Representation in Ireland, called the ARC Hub for Healthtech an “operation of strategic importance”, while Minister for Further and Higher Education, Research, Innovation and Science James Lawless, TD said that he believes the hub “has the potential to deliver game changing acceleration of research commercialisation”.
Advertisement
Duffy believes entrepreneurship should be taught to students early on in their higher education. Hackathons and labs that nurture students to think commercially have had a positive impact, he notes.
“I feel like we’re evolving into a nice ecosystem in Ireland where it’s becoming a bit of a norm to think of a spin-out company as an outcome for university education.”
Duffy is a professor at the University of Galway, and head of the anatomy and regenerative medicine department at RCSI University of Medicine and Health Sciences.
Don’t miss out on the knowledge you need to succeed. Sign up for the Daily Brief, Silicon Republic’s digest of need-to-know sci-tech news.