Connect with us
DAPA Banner

Tech

GeekWire Awards: From the farm to space, Next Tech Titan finalists growing to meet big challenges

Published

on

(Company logos)

The path from successful startup to industry heavyweight is often marked by the ability to solve massive, complex problems at scale — whether those challenges are on a farm, battlefield or in low-Earth orbit.

This GeekWire Award, presented by Baird, takes notice of the next dominant force in Pacific Northwest tech. The Next Tech Titan finalists are: Overland AI, Carbon Robotics, Stoke Space, Chainguard and MotherDuck.

Now in its 18th year, the GeekWire Awards is the premier event recognizing the top leaders, companies and breakthroughs in Pacific Northwest tech, bringing together hundreds of people to celebrate innovation and the entrepreneurial spirit. It takes place May 7 at the Showbox SoDo in Seattle.

Last year’s Next Tech Titan winner was Truveta, a Bellevue, Wash.-based company that aims to aggregate medical records data from partner institutions to link treatments with outcomes and underlying health. Truveta raised $320 million in fresh funding in 2025 to push its valuation above $1 billion.

Continue reading for information on the 2026 Next Tech Titan finalists, who were chosen by a panel of independent judges from community nominations. You can help pick the winner: Cast your ballot here or in the embedded form at the bottom. Voting runs through April 10.

Overland AI develops autonomous vehicle software and hardware designed specifically for complex, off-road environments. The company’s platform allows robotic vehicles to navigate high-speed, unpredictable terrain where GPS and cellular signals are often unavailable. Overland is focused on operational integration with the U.S. Army and Marine Corps, and is a key player in the emerging defense-tech corridor of the Pacific Northwest.

Advertisement

GeekWire first covered Overland AI in 2022 when it was a small, stealthy group of researchers spinning out of the University of Washington’s Robot Learning Laboratory. The company, No. 12 on the GeekWire 200, has grown to more than 100 employees, raised more than $140 million, and opened a 22,000 square-foot production facility in Seattle since then.

Ag-tech startup Carbon Robotics builds AI-powered machinery designed to eliminate weeds without the use of chemical herbicides. Its flagship LaserWeeder uses computer vision to identify and zap weeds with lasers, a process powered by the company’s “Large Plant Model.” This AI model, trained on 150 million labeled plants, allows the machines to adapt to new crops and environments in minutes. The company is also expanding into autonomous farm equipment with its Carbon ATK platform and an unrevealed new AI robot.

Founded in 2018 by Isilon Systems co-founder Paul Mikesell, the Seattle-based company has raised $177 million to date and employs about 260 people. Its LaserWeeders are now active on hundreds of farms across 15 countries, helping growers significantly reduce labor and pesticide costs. Carbon is No. 10 on the GeekWire 200.

Stoke Space is developing Nova, a medium-lift rocket designed for 100% reusability and rapid turnaround between flights. Unlike competitors that focus on heavy-lift vehicles, the Kent, Wash.-based company is targeting the medium-lift market with a unique second-stage design featuring an actively cooled heatshield for atmospheric reentry. The goal is to provide a more flexible and cost-effective launch platform that can be reused as seamlessly as an aircraft.

Advertisement

Founded by former Blue Origin and SpaceX engineers, Stoke Space has raised $1.34 billion to date, including a massive $860 million Series D round concluded in early 2026. The company, No. 8 on the GeekWire 200, is currently preparing for its first orbital launch from Cape Canaveral later this year and has already been selected by the U.S. Space Force for national security launches.

Chainguard secures the “software supply chain” by protecting the open-source components and container images used in modern cloud applications. The company’s tools allow developers to use verified, vulnerability-free code, automating the process of keeping foundational software secure. By focusing on the root of software production, Chainguard helps engineering teams eliminate security risks without slowing down development cycles.

Founded in 2021 and based in Kirkland, Wash., the startup has raised $892 million to date, reaching a $3.5 billion valuation. In fiscal year 2025, the company grew its annual recurring revenue sevenfold to $40 million. Now employing more than 500 people and serving over 200 customers — including GitLab and Hewlett Packard Enterprise — Chainguard is No. 2 on the GeekWire 200.

MotherDuck provides a serverless analytics platform built on the open-source DuckDB database engine. Designed for “small data” that doesn’t reach petabyte scale, the technology allows users to run fast SQL queries locally in a browser or in the cloud without the complexity of distributed architectures. By merging local processing speed with cloud scalability, the platform aims to make data analysis more cost-effective and accessible.

Advertisement

Founded in 2022 by former Google BigQuery founding engineer Jordan Tigani, the Seattle startup has raised more than $100 million and is No. 25 on the GeekWire 200.

Astound Business Solutions is the presenting sponsor of the 2026 GeekWire Awards. Thanks also to gold sponsors Amazon Sustainability, BairdBECU, JLLFirst Tech and Wilson Sonsini, and silver sponsors Prime Team Partners.

The event will feature a VIP reception, sit-down dinner and fun entertainment mixed in. Tickets go fast. A limited number of half-table and full-table sponsorships available. Contact events@geekwire.com to reserve a spot for your team today.

(function(t,e,s,n){var o,a,c;t.SMCX=t.SMCX||[],e.getElementById(n)||(o=e.getElementsByTagName(s),a=o[o.length-1],c=e.createElement(s),c.type=”text/javascript”,c.async=!0,c.id=n,c.src=”https://widget.surveymonkey.com/collect/website/js/tRaiETqnLgj758hTBazgd5M58tggxeII7bOlSeQcq8A_2FgMSV6oauwlPEL4WBj_2Fnb.js”,a.parentNode.insertBefore(c,a))})(window,document,”script”,”smcx-sdk”); Create your own user feedback survey

Advertisement

Source link

Continue Reading
Click to comment

You must be logged in to post a comment Login

Leave a Reply

Tech

OpenAI Acquires Popular Tech-Industry Talk Show TBPN

Published

on

OpenAI is acquiring tech news podcast TBPN, a fast-growing daily show hosted by John Coogan and Jordi Hays. OpenAI says TBPN will keep its editorial independence, even though the acquisition is widely viewed as part of a broader effort to influence public discourse around AI. CNBC reports: In the announcement, OpenAI CEO of AGI Deployment Fidji Simo wrote that their mission of bringing artificial general intelligence comes with a responsibility to have a space for “constructive conversation about the changes AI creates.” Altman has appeared on TBPN multiple times and is a frequent presence across media and podcasts, even hitting NBC’s “Tonight Show Starring Jimmy Fallon” in December.

The announcement says TBPN will maintain editorial independence and continue to choose its own guests. “TBPN is my favorite tech show. We want them to keep that going and for them to do what they do so well,” Altman wrote in a post on X. “I don’t expect them to go any easier on us, am sure I’ll do my part to help enable that with occasional stupid decisions.” OpenAI did not disclose the terms of the deal but said TBPN will be housed within its strategy organization. “While we’ve been critical of the industry at times, after getting to know Sam and the OpenAI team, what stood out most was their openness to feedback and commitment to getting this right,” wrote Hays in a statement. “Moving from commentary to real impact in how this technology is distributed and understood globally is incredibly important to us.”

Source link

Continue Reading

Tech

Evolution of Ransomware: Multi-Extortion Ransomware Attacks

Published

on

Hand pointing at a cybersecurity symbol

Ransomware’s Real-World Impact Across Industries

In February 2026, the University of Mississippi Medical Center (UMMC) fell victim to a ransomware attack. The incident took the Epic electronic health record system offline across 35 clinics and more than 200 telehealth sites, forcing the cancellation of chemotherapy appointments and the postponement of non-emergency surgeries. Medical staff were required to revert to paper-based workflows, leaving countless patients to bear the consequences.

UMMC is far from an isolated case. According to recent data, 93% of U.S. healthcare organizations experienced at least one cyberattack in 2025, and 72% of respondents reported that at least one incident directly disrupted patient care.

The manufacturing and financial sectors are equally exposed. In February 2026, payment processing network BridgePay suffered a ransomware attack that took its APIs, virtual terminals, and payment pages completely offline. Across all industries, publicly disclosed ransomware attacks surged 49% year-over-year in 2025, reaching 1,174 confirmed incidents.

As hospitals halt treatments, financial institutions freeze transactions, and manufacturers shut down production lines, ransomware has firmly established itself as a direct business risk with tangible operational consequences.

Advertisement

The Evolution of Ransomware: Double Extortion

Early ransomware operated on a straightforward premise: infiltrate a system, encrypt files, and demand payment in exchange for the decryption key. As organizations began countering this tactic by restoring from backups rather than paying ransoms, threat actors responded by developing a more lucrative model — double extortion.

In a double extortion attack, adversaries first exfiltrate sensitive files — such as patient records and billing data — before encrypting the target system. Victims are then pressured on two fronts: pay to receive the decryption key, or face public exposure of the stolen data.

Backups alone are insufficient against this model. Since attackers already possess the data, refusing to pay the ransom can result in the public release of sensitive files, exposing organizations to significant business losses and regulatory consequences.

The threat landscape has continued to escalate, with triple extortion cases on the rise — a tactic in which attackers directly contact a victim organization’s customers or partners to apply additional pressure.

Advertisement

As of 2025, 124 active ransomware groups have been identified, 73 of which are newly emerged.

The proliferation of AI-powered tools has lowered the barrier to entry for cybercrime, making ransomware capabilities increasingly accessible to less sophisticated actors.

D.AMO makes stolen data unreadable.

See how D.AMO defends against every stage of a ransomware attack.

Advertisement

Learn More

A Defense Architecture for Multi-Extortion Threats

The rise of multi-extortion ransomware fundamentally changes the assumptions underlying traditional defense strategies. Perimeter-based prevention alone is no longer sufficient.

Organizations need a security posture that protects data from being weaponized after a breach — rendering exfiltrated data unreadable, blocking ransomware from accessing files in the first place, and enabling rapid recovery even when an attack succeeds.

Muti-extortion ransomware attack flow diagram

D.AMO: Blocking Every Stage of a Ransomware Attack

D.AMO, developed by Penta Security, is an encryption-based data protection platform designed to address every phase of a multi-extortion ransomware attack. It delivers integrated encryption, access control, and backup recovery across on-premises and cloud environments.

By applying file encryption and process-based access control technologies, D.AMO protects critical data stored on servers and PCs — safeguarding sensitive information against malicious programs through robust access enforcement. D.AMO’s key capabilities are as follows:

Advertisement

Folder-Level File Encryption

D.AMO KE encrypts all files within administrator-designated folders at the OS level. Deployable via an installer without source code modification, it operates using kernel-level encryption technology, enabling fast and secure encryption on existing systems with no disruption to the user experience.

Encryption policies are applied at the folder level, ensuring consistent protection with minimal operational overhead. Critically, even if an attacker exfiltrates sensitive data, the files remain encrypted — neutralizing the data exposure threat that is central to double extortion.

Access Control

D.AMO KE enforces strict access control over processes and OS users, permitting only explicitly authorized access. Ransomware and other malicious applications are automatically blocked from accessing encrypted folders, preventing unauthorized file manipulation.

Advertisement

All blocked activity is recorded through an audit log function and can be reviewed centrally via D.AMO Control Center.

Backup and Recovery

Even in the event of a successful attack, organizations can resume operations through an independently managed recovery system. With D.AMO in place, the ability to restore from backup significantly reduces dependence on decryption key negotiations with threat actors.

As multi-extortion tactics become the norm, neutralizing the data attackers seek to exploit has become a strategic priority. Organizations need the ability to render exfiltrated data unreadable, prevent ransomware from accessing files, and recover rapidly when incidents occur.

D.AMO addresses each stage of a ransomware attack within a single integrated platform — combining encryption, process-based access control, and backup recovery into a unified line of defense.

Advertisement

Want to learn more? Download the D.AMO Data Sheet.  

Sponsored and written by Penta Security.

Source link

Advertisement
Continue Reading

Tech

Samsung’s Color E-Paper Gives Retailers a Simple Way to Refresh Every Sign on the Spot

Published

on

Samsung Color E-Paper Retail
Retail locations have always employed posters and signage to capture the attention of customers, but changing them out on a daily basis is a time-consuming task for workers. Samsung has developed a solution: displays that appear like printed paper but feature digital flexibility. These are known as Color E-Paper and are available in a variety of sizes to meet your needs, including a 13-inch variant that is nearly identical to an A4 sheet, a 20-inch version that matches an A3, and a 32-inch choice if you have a large wall or window to fill.



Each of these display units is extremely thin (under 18mm) and light enough to hang almost anywhere. The 13-inch version may be hung on a door, slid onto a counter, or even fixed to a shelf without the need for additional hardware; simply use the brackets and stands that come with it. Inside each one are millions of very small receptacles, or tiny cups, filled with colored ink particles in red, yellow, white, and blue. Electric signals just nudge the appropriate ones to the top, forming the image or words you wish to view.

Sale


PhotoTag Smart Item Finder, Bluetooth Tracker Tag Works with Apple Find My & Google Find Hub (iOS…
  • 【Works with Apple Find My & Google Find Hub】Compatible with both iPhone and Android, track your essentials with confidence. Allowing you to locate…
  • 【Customizable Color E-Ink Display – Make It Truly Yours】Unlike ordinary tracker tags, PhotoTag features a customizable color E-Ink screen…
  • 【“If Found” Contact Card – Help Lost Items Return】Turn your tracker into a smart return tag. Display your name, phone number, email, or…

Samsung engineers have devised a technique to mix and match those four colors to create a stunning 2.5 million different shades, almost as lovely as the real thing, and much like a real printed page, once the image is up, it consumes no electricity. The only time it requires power is when you change the content, and a full charge on the battery will last for weeks or even months, depending on how frequently you wish to update the data.

Advertisement

Samsung Color E-Paper Retail
To get the fresh image on the screen, simply utilize the mobile app, which is accessible for most smartphones running the latest Android or iOS versions. Simply launch the app, select a layout, see how the colours will appear, and transmit it to the display. If you’re part of a large chain with multiple stores, Samsung has a cloud platform that can manage everything remotely and link with any other screens you already have in place. Bye bye ladder, heaps of paper to print, and no more waiting for the next delivery truck; it seems like a dream come true.

A handful of the models’ housings blend crushed recycled plastic with a special plant-based glue derived from the microscopic particles that floats on water, and the packaging is entirely made of paper rather than plastic. By selecting these solutions, retailers may reduce their environmental effect without sacrificing durability or the feel of real paper, which draws customers in. Retailers that have begun testing the displays have raved about the outcomes on shelves, café walls, and franchise shop gateways. You can swap out a menu board in a diner to go from breakfast to lunch in two seconds flat, highlight a flash sale in every single clothing store at the exact same time, and do the same thing for class schedules or seasonal notices in community areas without messing around.

Samsung Color E-Paper Retail
Samsung plans to release the 20-inch models in the second half of 2026, but the smaller and larger sizes are now available in a few locations, and the early versions appear to be working properly in real stores as we speak. In the future, you may see updates that allow you to choose from different colors or batteries that last longer. For those tired of the never-ending cycle of printing, laminating, and continuously replacing signs, these displays quietly get the job done one gorgeous, low power panel at a time.

Source link

Advertisement
Continue Reading

Tech

Apple's iPad is still showing the world how to do tablets, 16 years later

Published

on

The iPad was mocked at launch, has been threatened by rivals throughout, and yet still remains the best-selling tablet ever made, 16 years after it first shipped to customers on April 3, 2010.


It’s easy to name alternatives to the iPad, you could be here all day listing myriad Android tablets. But it’s impossible to name even one true iPad competitor.
For after all of these years since it launched, and after all of the rival devices that have launched after that moment, there isn’t any one tablet that sells enough on its own to compete with the iPad. Its competition is the mass of cheaper rivals, which is not to be ignored, yet none of them have come close to the success of the iPad.
The closest is surely the Microsoft Surface, but if that’s the best and the best-known rival, it doesn’t appear to be doing all that well.
Continue Reading on AppleInsider | Discuss on our Forums

Source link

Continue Reading

Tech

Python Blood Could Hold the Secret To Healthy Weight Loss

Published

on

Longtime Slashdot reader fahrbot-bot writes: CU Boulder researchers are reporting that they have discovered an appetite-suppressing compound in python blood that helps the snakes consume enormous meals and go months without eating yet remain metabolically healthy. The findings were published in the journal Natural Metabolism on March 19, 2026.

Pythons can grow as big as a telephone pole, swallow an antelope whole, and go months or even years without eating — all while maintaining a healthy heart and plenty of muscle mass. In the hours after they eat, research has shown, their heart expands 25% and their metabolism speeds up 4,000-fold to help them digest their meal. The team measured blood samples from ball pythons and Burmese pythons, fed once every 28 days, immediately after they ate a meal. In all, they found 208 metabolites that increased significantly after the pythons ate. One molecule, called para-tyramine-O-sulfate (pTOS) soared 1,000-fold.

Further studies, done with Baylor University researchers, showed that when they gave high doses of pTOS to obese or lean mice, it acted on the hypothalamus, the appetite center of the brain, prompting weight loss without causing gastrointestinal problems, muscle loss or declines in energy. The study found that pTOS, which is produced by the snake’s gut bacteria, is not present in mice naturally. It is present in human urine at low levels and does increase somewhat after a meal. But because most research is done in mice or rats, pTOS has been overlooked. “We’ve basically discovered an appetite suppressant that works in mice without some of the side-effects that GLP-1 drugs have,” said senior author Leslie Leinwand, a distinguished professor of Molecular, Cellular and Developmental Biology who has been studying pythons in her lab for two decades. Drugs like Ozempic and Wegovy act on the hormone glucagon-like petide-1 (GLP-1).

Source link

Advertisement
Continue Reading

Tech

The Best Samsung Galaxy S26 Cases (2026): S26, S26+, and S26 Ultra

Published

on

Other Cases to Consider

Image may contain Electronics Mobile Phone and Phone

Photograph: Louryn Strampe

Spigen Tough Armor and Nano Pop MagFit Cases: These affordable cases both look and perform well for the price. The Nano Pop case was just a little too slippery for me, and the Tough Armor case kickstand was flimsier than I’d have liked. But if none of our other recommendations tickle your fancy, these options aren’t the worst.

Dbrand Tank Case for $60: This case looks very tactical. If that’s the look you’re after, it’s worth considering. For me, I found the back textures to be a little overstimulating and unpleasant. I wasn’t ever able to forget about my phone case. The buttons are swappable, and there are camera covers to help ensure a cohesive aesthetic. The case is durable and sturdy, and it makes it easy to get a good grip on the phone. It just comes down to the kind of design you prefer.

Poetic Spartan, Revolution, and Guardian Cases: I thought all three of these cases were just fine. I liked the Revolution’s built-in camera privacy cover, which also helped to protect the large camera array from bumps and bruises. But I wasn’t a fan of the rest of the design—while the built-in kickstand is a neat feature, the entirety of the case was too bulky for my preferences. The Guardian was the thinnest, and I liked it for the most part although the black grippy edges were a little bulkier than I wanted them to be. I didn’t like the Spartan case’s built-in metal ring, tactical design, or rigid bumper corners. Overall, the Poetic cases I tried had appealing prices but their designs weren’t my favorite. All three of these cases come with screen protectors, which work just fine (though you’ll have to install them the old-fashioned way).

Advertisement

More Good Screen Protectors

Spigen AluminaCore Screen Protectors (2-pack) for $19: Installation was easy, with a foolproof frame and a peel-off sticker that leaves the protector exactly where you want it. I had some initial issues with bubbles (that I was able to remove with the included squeegee) which is why these aren’t my top pick. I do like that you get two in case of issues with installation, or as a replacement when you inevitably crack the first one.

Cases to Avoid

Samsung Slim Magnet Case for $70: If this case cost $20, then sure. But it’s $70 for an exceedingly thin plastic shell with a ring of magnets built in. The build feels flimsy, and the case feels slippery too. It’s almost easier to grip the phone with no protection rather than to hold it with this case on it. There are simply too many other options on the market for this one to be worthy of a recommendation.


Power up with unlimited access to WIRED. Get best-in-class reporting and exclusive subscriber content that’s too important to ignore. Subscribe Today.

Advertisement

Source link

Continue Reading

Tech

Fan fiction website AO3 is finally coming out of beta

Published

on

The famous fan fiction website Archive of Our Own or AO3 has finally exited open beta, 17 years after it launched way back in 2009. AO3 is a nonprofit created by the by the Organization for Transformative Works. In an announcement, the team reminisced about its early days and how volunteers had to manually send out invitations to prospective writers. Upon launching the website on open beta, it only had 347 accounts and hosted 6,598 works. Now, it has 10 million registered users and is hosting 17 million fan-created works.

The team has highlighted some of the most useful features it has added over the past 17 years, including its tagging system. It also mentioned a feature it calls “Orphaning,” which allows authors to leave their works online even after deleting their account. In addition, it released the ability to download fanworks in AZW3, EPUB, MOBI, PDF or HTML format for offline access.

Even though the website has only just exited open beta, it has been stable for a long time. Users will not see huge changes, but the team also promised that it will not stop improving the fan fiction portal. It says its contributors and volunteers will continue tweaking the website, and it also continues to welcome anybody who has coding knowledge to contribute their time.

Source link

Advertisement
Continue Reading

Tech

Let Us Text You the Very Best Deals Directly to Your Phone

Published

on

Even though the Amazon Big Spring Sale has come to a close, that’s no reason to pay full price for the devices and gadgets you want. There are still some exceptional savings to be had with deals popping up left and right. Whether you’re picking up a new laptop or shopping for kitchen essentials, retailers regularly run sales that slash the amount you have to pay. But we know that it can be a real chore finding the best prices out there, and finding the time to trawl the web for the right deal isn’t always in the cards.

Well, luckily for you, it’s our job to search through all the sales. CNET’s shopping experts focus on finding deals that genuinely save you money. We know how to avoid the ones padded by inflated list prices or clever wording. If the discount isn’t real or the product isn’t worth owning, it doesn’t make the cut.

The team and I continually track and handpick the best offers from your favorite retailers, including Amazon and Walmart, as well as others, for our CNET Deals text subscribers. I’ll text the best sales at no cost straight to your phone, so you can keep an eye on the hottest drops and jump on them before everyone else does. It is never a bad time to save money, but with recent holiday expenses behind us, finding affordable items in early 2026 is more welcome than ever.

Advertisement

Why go through the effort of sifting through sales on your own when we can do it for you? Signing up for the CNET Deals text group (just scroll down) takes less than a minute. It is safe and trusted, plus you can opt out anytime. And the best part: It is free. The service costs nothing, and you’ll save money on products you love.

More about our deals text curation

This is the good stuff, not just “discounts” on items that were artificially inflated last week. We vet every deal to ensure the price is accurate and that the product is in stock when we send the text.

We send out a major Deal of the Day most days. During big shopping events, we send two texts a day on standout sales. If we find multiple deals at an ultra-low price, we hook you up in a single text.

My team and I apply the same care that we do across all of CNET, just in a bite‑size format. With daily deals texting, you will receive the same level of deep research and the same confirmation that these discounts are legitimate. And there is no AI pulling the strings. Real people are behind these texts and the research that has gone into finding the best deals.

Advertisement

We are a passionate and dedicated group of bargain hunters. So when we uncover something interesting for an affordable price, usually less than $50 with a significant discount, you will hear about it. If we find a cool thing on sale, we share that discovery. It is as simple as that. Join us today.

Source link

Advertisement
Continue Reading

Tech

How Apple keeps redefining personal computing at 50

Published

on

For a 50-year-old company, Apple remains pretty hip and nimble. This week, Devindra and Senior Reporter Igor Bonifacic dive into Apple’s big birthday, the state of the company today and what the next 50 years could bring. It remains one of the few PC companies that’s still firmly committed to the idea of personal computing. Also, we celebrate the successful launch of NASA’s Artemis II mission, which will bring us back to the Moon (but just for a close look).

Subscribe!

Topic

  • Apple at 50: Why it’s still all about personal computing – 1:16

  • Artemis II is safely on its way to the moon, but they’re having problems with Outlook – 37:48

  • SpaceX files for the largest IPO ever, what’s driving their hopes for a 1.75 Trillion valuation? – 40:52

  • Another Starlink satellite broke up in orbit, the second in 6 months – 47:21

  • Anthropic accidentally leaked source code for Claude Code – 52:17

  • FCC issues ban on all foreign-made WiFi routers – 57:18

  • Around Engadget – 1:02:09

  • Pop culture picks – 1:08:20

Credits

Hosts : Devindra Hardawar and Igor Bonifacic
Producer: Ben Ellman
Music: Dale North and Terrence O’Brien

Source link

Advertisement
Continue Reading

Tech

Arcee’s new, open source Trinity-Large-Thinking is the rare, powerful U.S.-made AI model that enterprises can download and customize

Published

on

The baton of open source AI models has been passed on between several companies over the years since ChatGPT debuted in late 2022, from Meta with its Llama family to Chinese labs like Qwen and z.ai. But lately, Chinese companies have started pivoting back towards proprietary models even as some U.S. labs like Cursor and Nvidia release their own variants of the Chinese models, leaving a question mark about who will originate this branch of technology going forward.

One answer: Arcee, a San Francisco based lab, which this week released AI Trinity-Large-Thinking—a 399-billion parameter text-only reasoning model released under the uncompromisingly open Apache 2.0 license, allowing for full customizability and commercial usage by anyone from indie developers to large enterprises.

The release represents more than just a new set of weights on AI code sharing community Hugging Face; it is a strategic bet that “American Open Weights” can provide a sovereign alternative to the increasingly closed or restricted frontier models of 2025.

This move arrives precisely as enterprises express growing discomfort with relying on Chinese-based architectures for critical infrastructure, creating a demand for a domestic champion that Arcee intends to fill.

Advertisement

As Clément Delangue, co-founder and CEO of Hugging Face, told VentureBeat in a direct message on X: “The strength of the US has always been its startups so maybe they’re the ones we should count on to lead in open-source AI. Arcee shows that it’s possible!”

Genesis of a 30-person frontier lab

To understand the weight of the Trinity release, one must understand the lab that built it. Based in San Francisco, Arcee AI is a lean team of only 30 people.

While competitors like OpenAI and Google operate with thousands of engineers and multibillion-dollar compute budgets, Arcee has defined itself through what CTO Lucas Atkins calls “engineering through constraint”.

The company first made waves in 2024 after securing a $24 million Series A led by Emergence Capital, bringing its total capital to just under $50 million. In early 2026, the team took a massive risk: they committed $20 million—nearly half their total funding—to a single 33-day training run for Trinity Large.

Advertisement

Utilizing a cluster of 2048 NVIDIA B300 Blackwell GPUs, which provided twice the speed of the previous Hopper generation, Arcee bet the company’s future on the belief that developers needed a frontier model they could truly own.

This “back the company” bet was a masterclass in capital efficiency, proving that a small, focused team could stand up a full pipeline and stabilize training without endless reserves.

Engineering through extreme architectural constraint

Trinity-Large-Thinking is noteworthy for the extreme sparsity of its attention mechanism. While the model houses 400 billion total parameters, its Mixture-of-Experts architecture means that only 1.56%, or 13 billion parameters, are active for any given token.

This allows the model to possess the deep knowledge of a massive system while maintaining the inference speed and operational efficiency of a much smaller one—performing roughly 2 to 3 times faster than its peers on the same hardware. Training such a sparse model presented significant stability challenges.

Advertisement

To prevent a few experts from becoming “winners” while others remained untrained “dead weight,” Arcee developed SMEBU, or Soft-clamped Momentum Expert Bias Updates.

This mechanism ensures that experts are specialized and routed evenly across a general web corpus. The architecture also incorporates a hybrid approach, alternating local and global sliding window attention layers in a 3:1 ratio to maintain performance in long-context scenarios.

The data curriculum and synthetic reasoning

Arcee’s partnership with fellow startup DatologyAI provided a curriculum of over 10 trillion curated tokens. However, the training corpus for the full-scale model was expanded to 20 trillion tokens, split evenly between curated web data and high-quality synthetic data.

Unlike typical imitation-based synthetic data where a smaller model simply learns to mimic a larger one, DatologyAI utilized techniques to synthetically rewrite raw web text—such as Wikipedia articles or blogs—to condense the information.

Advertisement

This process helped the model learn to reason over concepts and information rather than merely memorizing exact token strings.

To ensure regulatory compliance, tremendous effort was invested in excluding copyrighted books and materials with unclear licensing, attracting enterprise customers who are wary of intellectual property risks associated with mainstream LLMs.

This data-first approach allowed the model to scale cleanly while significantly improving performance on complex tasks like mathematics and multi-step agent tool use.

The pivot from yappy chatbots to reasoning agents

The defining feature of this official release is the transition from a standard “instruct” model to a “reasoning” model.

Advertisement

By implementing a “thinking” phase prior to generating a response—similar to the internal loops found in the earlier Trinity-Mini—Arcee has addressed the primary criticism of its January “Preview” release.

Early users of the Preview model had noted that it sometimes struggled with multi-step instructions in complex environments and could be “underwhelming” for agentic tasks.

The “Thinking” update effectively bridges this gap, enabling what Arcee calls “long-horizon agents” that can maintain coherence across multi-turn tool calls without getting “sloppy”.

This reasoning process enables better context coherence and cleaner instruction following under constraint. This has direct implications for Maestro Reasoning, a 32B-parameter derivative of Trinity already being used in audit-focused industries to provide transparent “thought-to-answer” traces.

Advertisement

The goal was to move beyond “yappy” or inefficient chatbots toward reliable, cheap, high-quality agents that stay stable across long-running loops.

Geopolitics and the case for American open weights

The significance of Arcee’s Apache 2.0 commitment is amplified by the retreat of its primary competitors from the open-weight frontier.

Throughout 2025, Chinese research labs like Alibaba’s Qwen and z.ai (aka Zhupai) set the pace for high-efficiency MoE architectures.

However, as we enter 2026, those labs have begun to shift toward proprietary enterprise platforms and specialized subscriptions, signaling a move away from pure community growth.

Advertisement

The fragmentation of these once-prolific teams, such as the departure of key technical leads from Alibaba’s Qwen lab, has left a void at the high end of the open-weight market. In the United States, the movement has faced its own crisis.

Meta’s Llama division notably retreated from the frontier landscape following the mixed reception of Llama 4 in April 2025, which faced reports of quality issues and benchmark manipulation.

For developers who relied on the Llama 3 era of dominance, the lack of a current 400B+ open model created an urgent need for an alternative that Arcee has risen to fill.

Benchmarks and how Arcee’s Trinity-Large-Thinking stacks up to other U.S. frontier open source AI model offerings

Trinity-Large-Thinking’s performance on agent-specific evaluations establishes it as a legitimate frontier contender. On PinchBench, a critical metric for evaluating model capability on autonomous agentic tasks, Trinity achieved a score of 91.9, placing it just behind the proprietary market leader, Claude Opus 4.6 (93.3).

Advertisement
Arcee Trinity-Large-Thinking benchmark comparison chart

Arcee Trinity-Large-Thinking benchmark comparison chart. Credit: Arcee

This competitiveness is mirrored in IFBench, where Trinity’s score of 52.3 sits in a near-dead heat with Opus 4.6’s 53.1, indicating that the reasoning-first “Thinking” update has successfully addressed the instruction-following hurdles that challenged the model’s earlier preview phase.

The model’s broader technical reasoning capabilities also place it at the high end of the current open-source market. It recorded a 96.3 on AIME25, matching the high-tier Kimi-K2.5 and outstripping other major competitors like GLM-5 (93.3) and MiniMax-M2.7 (80.0).

While high-end coding benchmarks like SWE-bench Verified still show a lead for top-tier closed-source models—with Trinity scoring 63.2 against Opus 4.6’s 75.6—the massive delta in cost-per-token positions Trinity as the more viable sovereign infrastructure layer for enterprises looking to deploy these capabilities at production scale.

Advertisement

When it comes to other U.S. open source frontier model offerings, OpenAI’s gpt-oss tops out at 120 billion parameters, but there’s also Google with Gemma (Gemma 4 was just released this week) and IBM’s Granite family is also worth a mention, despite having lower benchmarks. Nvidia’s Nemotron family is also notable, but is fine-tuned and post-trained Qwen variants.

Benchmark

Arcee Trinity-Large

gpt-oss-120B (High)

Advertisement

IBM Granite 4.0

Google Gemma 4

GPQA-D

76.3%

Advertisement

80.1%

74.8%

84.3%

Tau2-Airline

Advertisement

88.0%

65.8%*

68.3%

76.9%

Advertisement

PinchBench

91.9%

69.0% (IFBench)

89.1%

Advertisement

93.3%

AIME25

96.3%

97.9%

Advertisement

88.5%

89.2%

MMLU-Pro

83.4%

Advertisement

90.0% (MMLU)

81.2%

85.2%

So how is an enterprise supposed to choose between all these?

Advertisement

Arcee Trinity-Large-Thinking is the premier choice for organizations building autonomous agents; its sparse 400B architecture excels at “thinking” through multi-step logic, complex math, and long-horizon tool use. By activating only a fraction of its parameters, it provides a high-speed reasoning engine for developers who need GPT-4o-level planning capabilities within a cost-effective, open-source framework.

Conversely, gpt-oss-120B serves as the optimal middle ground for enterprises that require high-reasoning performance but prioritize lower operational costs and deployment flexibility.

Because it activates only 5.1B parameters per forward pass, it is uniquely suited for technical workloads like competitive code generation and advanced mathematical modeling that must run on limited hardware, such as a single H100 GPU.

Its configurable reasoning effort—offering “Low,” “Medium,” and “High” modes—makes it the best fit for production environments where latency and accuracy must be balanced dynamically across different tasks.

Advertisement

For broader, high-throughput applications, Google Gemma 4 and IBM Granite 4.0 serve as the primary backbones. Gemma 4 offers the highest “intelligence density” for general knowledge and scientific accuracy, making it the most versatile option for R&D and high-speed chat interfaces.

Meanwhile, IBM Granite 4.0 is engineered for the “all-day” enterprise workload, utilizing a hybrid architecture that eliminates context bottlenecks for massive document processing. For businesses concerned with legal compliance and hardware efficiency, Granite remains the most reliable foundation for large-scale RAG and document analysis.

Ownership as a feature for regulated industries

In this climate, Arcee’s choice of the Apache 2.0 license is a deliberate act of differentiation. Unlike the restrictive community licenses used by some competitors, Apache 2.0 allows enterprises to truly own their intelligence stack without the “black box” biases of a general-purpose chat model.

“Developers and Enterprises need models they can inspect, post-train, host, distill, and own,” Lucas Atkins noted in the launch announcement.

Advertisement

This ownership is critical for the “bitter lesson” of training small models: you usually need to train a massive frontier model first to generate the high-quality synthetic data and logits required to build efficient student models.

Furthermore, Arcee has released Trinity-Large-TrueBase, a raw 10-trillion-token checkpoint. TrueBase offers a rare, “unspoiled” look at foundational intelligence before instruction tuning and reinforcement learning are applied. For researchers in highly regulated industries like finance and defense, TrueBase allows for authentic audits and custom alignments starting from a clean slate.

Community verdict and the future of distillation

The response from the developer community has been largely positive, reflecting the desire for more open weights, U.S.-made mdoels.

On X, researchers highlighted the disruption, noting that the “insanely cheap” prices for a model of this size would be a boon for the agentic community.

Advertisement

On open AI model inference website OpenRouter, Trinity-Large-Preview established itself as the #1 most used open model in the U.S., serving over 80.6 billion tokens on peak days like March 1, 2026.

The proximity of Trinity-Large-Thinking to Claude Opus 4.6 on PinchBench—at 91.9 versus 93.3—is particularly striking when compared to the cost. At $0.90 per million output tokens, Trinity is approximately 96% cheaper than Opus 4.6, which costs $25 per million output tokens.

Arcee’s strategy is now focused on bringing these pretraining and post-training lessons back down the stack. Much of the work that went into Trinity Large will now flow into the Mini and Nano models, refreshing the company’s compact line with the distillation of frontier-level reasoning.

As global labs pivot toward proprietary lock-in, Arcee has positioned Trinity as a sovereign infrastructure layer that developers can finally control and adapt for long-horizon agentic workflows.

Advertisement

Source link

Continue Reading

Trending

Copyright © 2025