Connect with us
DAPA Banner

Tech

Samsung stopping sales of the TriFold is actually a good thing

Published

on

Samsung’s decision to quietly pull the plug on its most futuristic foldable might sound like a step backwards, but I don’t think it is. 

The Galaxy Z TriFold has always been less a mainstream gadget and more a very expensive proof of concept – a flashy way for Samsung to show off what its foldable tech can do. That’s why the news that Samsung is reportedly winding down production, just months after launch, shouldn’t be seen purely as a loss. 

Yes, it’s frustrating for fans who’ve no doubt been refreshing product pages only to watch restocks vanish within minutes, but with skyrocketing component prices, supply chain chaos and a mobile division under pressure to justify every penny it spends, the TriFold was always going to be first on the chopping block.

In reality, shelving the TriFold now could be exactly what Samsung needs: a chance to refocus on devices that more people can actually buy, that developers will actually support, and that move foldables forward in ways that matter – like the rumoured ‘Wide Fold’ and the next-gen Z Fold. 

Advertisement

Losing the snazzy TriFold might sting right now, but it could make Samsung’s foldable future all the stronger for it.

Advertisement

Samsung is reportedly winding down production of the Galaxy Z TriFold

A new report from South Korean publication Donga claimed this week that Samsung is preparing to wind down sales of the Galaxy Z TriFold, just months after its initial launch – a surprising move considering how quickly it sells out whenever stock does drop.

Per the report, Samsung is expected to release a final batch of TriFolds in its home region of South Korea this week, after which, sales could come to an end. Now it’s worth noting that the report is explicitly about sales in South Korea, with no word on whether the same pattern will play out in regions like the US – but it seems increasingly likely.

Advertisement
Galaxy trifold open watching a videoGalaxy trifold open watching a video
Image Credit (Trusted Reviews)

After all, Samsung has always been transparent about the TriFold and how it wasn’t a mass-market product, rather a showcase of what the company’s foldable tech is truly capable of. 

That explains not only the incredibly high $2,899 price tag in the States, but also comments from those who have used it on build quality not quite matching that of the comparatively cheaper, mass-produced Samsung Galaxy Z Fold 7

While that news will be disappointing for those still waiting to get their hands on the tri-folding 10-inch smartphone, the decision to stop sales isn’t all that surprising if you’ve been keeping an eye on Samsung more broadly recently. 

Advertisement

Not a surprise given soaring costs

A separate report that also surfaced this week, this time from the South Korean outlet FNNews, claims that despite a record number of Galaxy S26 range pre-orders, the Device eXperience (DX) division is essentially in crisis mode. That includes not only Samsung’s smartphones, but also wearables, smart TVs and home appliances. 

Advertisement

Of course, just like every other tech issue surfacing in 2026, it all seems to be down to the rapidly rising costs of components. The all-important RAM needed to power most tech has surged in the past nine months or so, with some estimates putting the rise as high as 850%, as AI data centres hoover up as much RAM as possible. 

Galaxy trifold open screenGalaxy trifold open screen
Image Credit (Trusted Reviews)

There’s also the ongoing conflict in the Middle East, which has reportedly increased logistics costs. 

A Samsung spokesperson who spoke to FNNews explained that “with raw material costs under extreme pressure from rising semiconductor prices, and logistics costs increasing on top of that, we ultimately had no choice but to put the MX division under emergency management.”

With the DX division under such tight constraints, it makes sense for Samsung to stop production of the TriFold – it’s expensive and, no doubt, more difficult to manufacture than the flip- and book-style foldables the company has cranked out over the past few years – and double down on smartphones that it knows will sell well.

Advertisement

Advertisement

Samsung can now focus on more consumer-ready foldable tech – like the ‘Wide Fold’

It might sound like it’s all doom and gloom over at Samsung, but honestly, it’s probably for the best. Tighter purse strings should force Samsung to double down on its core products – both foldable and non-foldable – rather than putting increasingly limited resources into niche, hyper-expensive products like the TriFold that aren’t ready for prime time just yet. 

Samsung Galaxy Z Fold 7Samsung Galaxy Z Fold 7
Image Credit (Trusted Reviews)

That’s more important than usual right now, with Samsung rumoured to be working on not one but two book-style foldables for mass consumer release later this year. 

The most recent reports claim that, in addition to a successor to the Galaxy Z Fold 7 that’s expected to offer a similarly slimline design and a boxy inner aspect ratio, the company is working on a second foldable, unofficially dubbed the ‘Z Wide Fold’. 

The key difference, as the unofficial moniker suggests, is the shape of the foldable. While Samsung’s regular Z Fold offers a thin and narrow aspect ratio, even with big improvements on this front with the Fold 7, and an almost perfectly 1:1 inner display, the ‘Wide’ Fold is said to be more passport-shaped.

Advertisement

Advertisement
Google Pixel Fold unfoldedGoogle Pixel Fold unfolded
Image Credit (Trusted Reviews)

That’s not a new idea – it was used by the original Oppo Find N, as well as Google’s first-gen Pixel Fold – but it has fallen out of trend in recent times. 

Samsung looks to revive it, and the foldable experience will be all the better for it; I’ve long been a fan of the passport-shaped foldable, not only because of the shorter, wider outer panel, but also because the inner screen more closely resembles a regular 4:3 tablet-sized screen. 

That should hopefully put an end to apps that, even after all this time, still can’t quite handle the boxy aspect ratio of foldables. With an aspect ratio closer to a tablet or even a smartphone in horizontal orientation, it becomes almost trivial to support, with no major rejigging of the UI required.  

If there were a choice between Samsung putting effort into its extremely expensive, not-quite-polished TriFold or the Wide Fold, especially under the new budget constraints, I know which I’d go for. 

Advertisement

Source link

Continue Reading
Click to comment

You must be logged in to post a comment Login

Leave a Reply

Tech

Trump officials may be encouraging banks to test Anthropic’s Mythos model

Published

on

Treasury Secretary Scott Bessent and Federal Reserve Chair Jerome Powell summoned bank executives for a meeting this week where they encouraged the executives to use Anthropic’s new Mythos model to detect vulnerabilities, according to Bloomberg

Indeed, while JPMorgan Chase was the only bank listed as one of the initial partner organizations with access to the model, Goldman Sachs, Citigroup, Bank of America, and Morgan Stanley are reportedly testing Mythos as well.

Anthropic announced the model this week but said it would be limiting access for now, in part because Mythos — despite not being trained specifically for cybersecurity — is too good at finding security vulnerabilities. (Others suggested this was hype or simply a smart enterprise sales strategy.)

The report is particularly surprising since Anthropic is currently battling the Trump administration in court over the Department of Defense’s designation of Anthropic as a supply-chain risk; that designation came after negotiations fell apart over the company’s efforts to limit how its AI models can be used by the government.

Advertisement

Meanwhile, the Financial Times reports that U.K. financial regulators are also discussing the risk posed by Mythos.

Source link

Continue Reading

Tech

OpenAI’s new $100 ChatGPT Pro plan targets Claude Max with five times the Codex access

Published

on

In short: OpenAI launched a new $100 per month Pro plan for ChatGPT on 9 April 2026, inserting a new tier between the existing $20 Plus plan and the $200 Pro plan and directly targeting Anthropic’s Claude Max, which is also priced at $100 per month. The new plan offers five times more Codex usage than Plus, access to the same model suite as the $200 tier, and a launch promotion that temporarily doubles that advantage: through 31 May 2026, subscribers get ten times the Codex usage of Plus. The move follows Codex crossing three million weekly users on 8 April, a growth rate the company describes as a 5x increase in three months.

What the $100 plan includes, and where it sits in ChatGPT’s pricing structure

The new plan is the sixth pricing tier in ChatGPT’s current structure, which now runs from a free account with advertising, through a $8 per month Go plan, the $20 per month Plus plan, to two versions of Pro at $100 and $200 per month, a $25 per user per month Business plan, and custom-priced Enterprise contracts. The $100 Pro plan sits directly between Plus and the existing $200 Pro tier, offering five times the Codex usage of Plus and targeting what OpenAI describes as “longer, high-effort Codex sessions” that Plus subscribers hit the ceiling on. The $200 Pro plan, by comparison, provides 20 times the Codex usage of Plus, making it four times more Codex-intensive than the new $100 tier.

Despite the difference in usage limits, both Pro tiers give access to the same model suite: the exclusive GPT-5.4 Pro model, unlimited use of GPT-5.4 Instant and GPT-5.4 Thinking, and all other features available on the $200 plan. The differentiation between the two tiers is usage volume, not capability. As a launch promotion, subscribers to the new $100 plan will receive ten times the Codex usage of Plus through 31 May 2026; after that date, the standard five times limit applies. OpenAI also announced a rebalancing of the Plus plan’s Codex allocation alongside the new tier, shifting Plus towards steadier day-to-day usage rather than allowing the longer burst sessions that the $100 plan is intended to serve.

Codex demand: the numbers that prompted the new tier

On 8 April 2026, the day before the $100 plan was announced, Sam Altman posted on X that OpenAI was resetting Codex’s usage limits across all plans “to celebrate 3M weekly codex users,” and committed to repeating the reset for every additional million users until Codex reaches ten million weekly users. Thibault Sottiaux, who leads the Codex product, stated: “Three million people are now using Codex weekly, up from two million a little under a month ago.” OpenAI described the growth trajectory as a 5x increase in the preceding three months, with 70% month-over-month user growth.

Advertisement

The 💜 of EU tech

The latest rumblings from the EU tech scene, a story from our wise ol’ founder Boris, and some questionable AI art. It’s free, every week, in your inbox. Sign up now!

The scale of that growth reflects a shift in how developers are using AI coding tools. OpenAI rolled out a dedicated Codex app for macOS in February 2026, designed to move beyond line-by-line code generation into what the company called agentic, multi-task coding workflows: orchestrating multiple agents in parallel, running background jobs, and handling instructions that span hours rather than seconds. That architecture, with its longer-running sessions and heavier compute demands, is precisely the usage pattern that the $100 plan is priced to capture. A Plus subscriber who uses Codex for extended autonomous engineering tasks hits usage limits well before their billing cycle ends; the $100 plan is designed to be the next logical tier rather than a jump to $200.

The Claude Max comparison

OpenAI made no attempt to obscure the competitive framing. The new plan is priced identically to Anthropic’s Claude Max 5x tier, which also costs $100 per month and includes elevated limits for Claude Code, Anthropic’s terminal-based agentic coding product. Claude Code has become the fastest-growing part of Anthropic’s commercial portfolio, with an estimated $2.5 billion in annualised revenue by early 2026, and Anthropic has been constructing a developer ecosystem around it: Anthropic launched a marketplace for Claude-powered enterprise software in March 2026, with launch partners including Snowflake, Harvey, and Replit, connecting enterprise buyers with third-party applications built on Claude.

Advertisement

The competitive dynamic sharpened further in the week before OpenAI’s announcement. On 4 April 2026, Anthropic banned third-party agents from Claude Pro and Max subscriptions, preventing subscribers from routing their plan’s usage limits through external frameworks such as OpenClaw; users wanting to continue using those tools must now pay separately under a new per-session “extra usage” system. OpenAI’s announcement went in the opposite direction, increasing Codex availability at the $100 price point and doubling it temporarily to mark the launch. The contrast, at the identical price, was visible enough that most coverage described the new plan as a direct response to Anthropic’s developer subscriber base.

What OpenAI’s pricing move signals

The new tier arrives during a period of accelerating commercial momentum for OpenAI. OpenAI’s $122 billion raise at an $852 billion valuation, completed in March 2026, was led by SoftBank, NVIDIA, and Amazon, and included $3 billion from individual retail investors, a structure that many analysts read as groundwork for an IPO expected as early as the fourth quarter of 2026. The company is generating $2 billion in revenue per month and has more than 50 million paid subscribers across its plans. The $100 plan is part of a deliberate effort to fill the pricing gap between $20 and $200 that had, until now, left a large segment of heavy but not enterprise-grade users without a compelling upgrade path.

The model powering the Pro tiers, GPT-5.4, which launched in March 2026 and introduced native computer use directly into Codex and the API, is the clearest statement of where OpenAI sees the next phase of developer adoption going: not prompting, but autonomous agents operating software, navigating file systems, and running multi-step workflows across applications for hours at a time. The $100 plan is the pricing expression of that bet. Whether it moves enough developers at the $100 Claude Max price point to make a measurable difference in Anthropic’s subscriber base will be visible in both companies’ next quarterly metrics.

Advertisement

Source link

Continue Reading

Tech

‘Marshals’ Release Schedule: When Episode 7 Hits Paramount Plus

Published

on

Marshals, a new Yellowstone spinoff starring Luke Grimes as Kayce Dutton, is airing on CBS right now. You can also tune in with Paramount Plus. The Yellowstone sequel series sees Grimes’ former Navy SEAL join an elite unit of US Marshals to bring range justice to Montana, according to a synopsis from CBS.

The show includes Yellowstone actors Gil Birmingham as Thomas Rainwater, Mo Brings Plenty as Mo and Brecken Merrill as Tate. Spencer Hudnut is the showrunner of Marshals — formerly known as Y: Marshals — and Taylor Sheridan is an executive producer.

Advertisement

When to watch new Marshals episodes on Paramount Plus

Episode 7 of Marshals airs on CBS on Sunday, April 12. Viewing options for Paramount Plus customers vary by subscription tier. You can watch the episode live if you have Paramount Plus Premium, which includes your local CBS station. If you subscribe to Paramount Plus Essential, you can watch the installment on demand the following Monday, but not live on Sunday.

Here’s a release schedule for the next two episodes of Marshals.

  • Episode 7, Family Business: Premieres on CBS/Paramount Plus Premium on April 12 at 8 p.m. ET/8 p.m. PT/7 p.m. CT. Streams on Paramount Plus Essential on April 13.
  • Episode 8, Blowback: Premieres on CBS/Paramount Plus Premium on April 19 at 8 p.m. ET/8 p.m. PT/7 p.m. CT. Streams on Paramount Plus Essential on April 20.

You can also watch CBS and the seventh episode of Marshals without cable with a live TV streaming service such as YouTube TV, Hulu Plus Live TV or the DirecTV MyNews skinny bundle. In addition to offering a lower-cost option, Paramount Plus lets you watch the other two Yellowstone spinoffs: the prequels 1883 and 1923.

Advertisement

James Martin/CNET

After a price increase in early 2026, the ad-supported Essential version runs $9 per month or $90 per year. The ad-free Premium version runs $14 per month or $140 per year. Paying more for Premium gives you downloads, the ability to watch more Showtime programming than Essential and access to your live, local CBS station.

Source link

Advertisement
Continue Reading

Tech

Sunday Reboot: MacBook Neo upgrades, masses of Mac minis, and iPhone re-entry

Published

on

In this week’s “Sunday Reboot,” a storage upgrade for the MacBook Neo, an excuse to buy many Mac minis, and the iPhones come back to Earth with a late congratulatory message.

Silhouetted person in shadow peers through a narrow opening toward a brightly lit wall of stacked silver computer mini desktops with ports and indicator lights
Image credits: NASA/Overcast

Sunday Reboot is a weekly column covering some of the lighter stories within the Apple reality distortion field from the past seven days. All to get the next week underway with a good first step.
This week, researchers managed to get around Apple Intelligence security measures using prompt injection techniques, a repairability report panned Apple’s hardware again, and Apple’s lawsuit with Epic Games over the App Store continued to roll on. There was also a bug found to break Mac networking every 49 days, 17 hours, two minutes, and 47 seconds.
Continue Reading on AppleInsider | Discuss on our Forums

Source link

Continue Reading

Tech

Tesla Achieves European Breakthrough as Full Self-Driving Supervised Reaches Dutch Roads

Published

on

Tesla FSD Europe Netherlands Launch Test Drive
Tesla finally received approval on April 10 from the Dutch vehicle regulator, the RDW, for its Full Self-Driving Supervised system to be used on European roads. They were the first to receive approval for this advanced technology across Europe, marking a significant milestone for the company. This means that the program has been cleared to run on public roads in the Netherlands, and the distribution began the next day, April 11, for a limited number of early testers who had been patiently waiting.



Drivers with Hardware 4 (HW4) computers in their vehicles received an upgrade to version 2026.3.6, which included the European-tuned build of FSD 14.2.2.5. Before turning on the system, drivers must complete a fast tutorial followed by a mini test within the car interface. Once that’s done, they can take their hands off the wheel under appropriate situations, and cameras will watch their eyes to see if they’re paying attention. If they become distracted, the system will begin to display visual alarms, followed by sounds and vibrations if they do not return to it, and if all else fails, the car will slow down and come to a safe stop on its own.


LEGO Icons Ford Model T Building Set for Adults – Collectible Kit with Detailed Features for Bedroom…
  • COLLECTIBLE MODEL CAR KIT – Embark on a nostalgic journey with the LEGO Icons Ford Model T (11376) building set for adults ages 18 and up
  • VINTAGE CAR DISPLAY PIECE – Recreate the iconic 1910s automobile with gleaming black bodywork, golden accents, a foldback fabric roof and tall…
  • AUTHENTIC MODEL T FUNCTIONS – Fold back the model car’s roof, fold down the split windshield, lift the hood from both sides to see the engine, turn…

Overall, this was the product of 18 months of testing, which included more than 1.5 million kilometers of driving on European highways, as well as numerous controlled scenarios on closed tracks. Before approving the system, regulators reviewed almost 400 compliance points. The RDW pronounced it a beneficial addition to road safety, but stressed that drivers must remain in the driver’s seat and ready to take over at any time.


The European version of the software is substantially different from the one accessible in the United States. This is primarily due to the way regulators work here, which requires them to do pre-market checks, as opposed to their US counterparts’ self-certification strategy. As a result, the Dutch construction is more cautious and limits some of the more aggressive driving characteristics available elsewhere. Automatic turns at junctions and navigation-based lane changes are still accessible, but several parking-lot summoning capabilities found in the United States are not available in the Netherlands.

Advertisement


Subscribers will pay 99 euros per month to receive the system, or 49 euros per month if they already have Enhanced Autopilot. Alternatively, they can purchase it outright for 7,500 euros. Tesla claims that the system leverages billions of kilometers of real-world data collected worldwide, and Elon Musk has just stated that the RDW review process was particularly rigorous.

Tesla FSD Europe Netherlands Launch Test Drive
The Dutch approval is now a one-time occurrence, although it has a provisional validity period of at least 36 months. It means that other European states can adopt it on their own, and authorities in Germany, France, and Italy are expected to do so within the next 4 to 8 weeks. Tesla’s goal is to have the system more widely accepted across the EU by the summer, allowing millions of drivers to use it without having to repeat the testing procedure in each nation.
[Source]

Source link

Advertisement
Continue Reading

Tech

Apple reportedly testing out four different styles for its smart glasses that will rival Meta Ray-Bans

Published

on

Apple may be late to the smart glasses market, but it could be covering all its bases with up to four potential styles for its upcoming product. According to Bloomberg‘s Mark Gurman, Apple could launch some or all of the four styles it’s currently testing for its smart glasses.

Gurman reported Apple is testing out a large rectangular frame that’s comparable to Ray-Ban Wayfarers, a slimmer rectangular design like the glasses that Apple CEO Tim Cook wears, a larger oval or circular frame and a smaller oval or circle option. Apple is also working on a range of colors, including black, ocean blue and light brown, according to Bloomberg.

Internally code-named N50 for now, Apple’s upcoming smart glasses will compete directly with the second-gen Ray-Ban Meta model. While similar, Apple might be differentiating its design with “vertically oriented oval lenses with surrounding lights,” according to the report. Like Meta’s smart glasses, Apple’s upcoming product will capture photos and videos, but is meant to better sync with an iPhone, allowing users to take advantage of Apple’s ecosystem for editing, sharing, phone calls, notifications, music and even its voice assistant, according to Gurman. The release of Apple’s smart glasses could even coincide with the upcoming improved Siri that should arrive with iOS 27.

Gurman reported that Apple could reveal its smart glasses as soon as the end of 2026 or early 2027, followed by an official release sometime in 2027. As for the competition, Meta released its latest model that’s better suited for prescription lenses and offers a more customizable fit.

Advertisement

Source link

Continue Reading

Tech

The MacBook Neo is moonlighting as a Windows gaming machine, and it’s doing it well

Published

on

Apple didn’t position its most affordable MacBook as a gaming machine. The MacBook Neo, a budget-leaning laptop that runs on Apple’s A18 Pro chip, the same chip that powers the iPhone 16 Pro models, has been put through a Windows 11 gaming test for YouTuber ETA Prime. 

Turns out, the results are genuinely surprising. Using Parallels Desktop, a virtualization app (paid) with 3D hardware acceleration, the channel ran Windows 11 ARM directly on the Neo’s 8GB RAM (allocating 5GB to the virtual environment), and it did better than most people would think it would. 

What games actually ran well?

Dirt 3 held 75 fps at 1200p on high settings, while Portal 2 cleared 100 fps on medium settings. Skyrim, on the other hand, maintained roughly 60 fps at 1200p resolution on medium graphics settings, while Marvel Cosmic Invasion averaged around 60 fps at the maximum resolution.

Advertisement

What helped performance was games running as native Windows-on-ARM applications. However, GTA V was among the notable stumbles, as the frame rates through the Parelles weren’t playable at all. However, according to Notebookcheck, the game runs acceptably via Crossover. 

Why does this matter for everyday MacBook Neo users?

For users who work on their Mac but occasionally enjoy playing Windows-only games, MacBook Neo’s ability to run native titles via the Parallels app comes as good news. The cost? Parallels Desktop’s Standard tier costs $99.99 per year, which could add to your weekend leisure sessions. 

Anyways, the bigger takeaway is that the MacBook Neo, even with 8GB of RAM (highlighted as a constraint in the video), can run low-to-mid-range Windows games. It also changes the notion around budget Apple hardware being primarily for productivity-based tasks. 

As virtualization tech continues to improve and Apple provides more RAM in future generations of the MacBook Neo, it could redefine what “budget” actually means for Apple buyers, bridging the gap between MacBook and Windows laptops even further. 

Advertisement

Source link

Continue Reading

Tech

How the Budget-Friendly BougeRV 23-Quart 12V Fridge Keeps Food Fresh Through Every Drive

Published

on

BougeRV 23 Quart 12V Portable Fridge Car
Summer heat makes any travel difficult, especially if you’re transporting groceries and / or cold drinks. Drivers are frequently forced to rely on old, simple coolers with ice that melts faster than a popsicle on a hot day, leaving everything wet by the time they reach. That’s where the BougeRV 23-quart unit, priced at $159.97 (was $189.99), comes in, a more practical solution that plugs directly into your car’s normal 12V socket and keeps items perfectly chilled without any of the fuss.



The unit is 22 inches long and weighs just more than 21 pounds, so it can fit into even the smallest trunks or backseats. It also has a built-in handle, making it simple to pull out at a rest break or transport home after a long shopping excursion. Inside, there’s enough space for a couple days’ worth of food or a full load of drinks and snacks for a family road trip.

Sale


BougeRV 12 Volt Refrigerator 12V Car Fridge 23 Quart Portable Freezer Compressor Cooler 12/24V DC…
  • What You Get: The CR22 12V refrigerator comes with a 2-year Tech Support. If you have any questions about the product, please REACH OUT TO BougeRV, as…
  • Fast Cooling Down to 32℉: With Compressor refrigeration technology, this 12v car refrigerator could achieve 15 min fast cooling from 77℉ to…
  • 45W Low Power Consumption: With ECO energy saving mode, this 23 Qt portable refrigerator’s operating power is less than 36W. Even running on MAX…


It’s powered by a 12-volt socket, which is found in practically every modern automobile, and there are alternatives for residential outlets or even solar power if you are parked for an extended period of time. The compressor system kicks in quickly, about 15 minutes, and maintains a consistent temperature between 8 degrees below zero and 50 degrees Fahrenheit, allowing you to choose between fridge and freezer mode as needed.

Advertisement

BougeRV 23 Quart 12V Portable Fridge Car
The portable fridge uses very little energy (around 36 watts in environment mode), and the smart cycling keeps your daily power consumption under one kilowatt-hour even on the warmest days. To be on the safe side, there’s a built-in battery monitor that will turn it off before it consumes your vehicle’s battery, so you don’t have to worry about that.

BougeRV 23 Quart 12V Portable Fridge Car
People who have used it on road trips note that it works effectively, keeping perishables from spoiling without having to constantly add ice, and it absorbs bumps in the road well, even while traveling at a 30-degree angle. If you’re only running to the store for a quick shopping trip, the fridge will keep running until you return home, even if you get stopped in traffic.

Source link

Continue Reading

Tech

New FCC router rules could trap millions using outdated ISP hardware as supply chain limits stall upgrades and complicate security fixes

Published

on


  • FCC rules block new foreign routers while old, vulnerable ones stay in homes longer
  • ISP customers cannot upgrade routers even when security risks become widely known
  • Router approvals now depend on waivers that may slow down nationwide replacements

The Federal Communications Commission (FCC) has issued new rules intended to address security risks posed by routers produced outside the United States.

A number of recent incidents have shown foreign routers are vulnerable to cyberattacks, with campaigns like Flax, Volt, and Salt Typhoon making headlines across the world.

Source link

Continue Reading

Tech

Your developers are already running AI locally: Why on-device inference is the CISO’s new blind spot

Published

on

For the last 18 months, the CISO playbook for generative AI has been relatively simple: Control the browser.

Security teams tightened cloud access security broker (CASB) policies, blocked or monitored traffic to well-known AI endpoints, and routed usage through sanctioned gateways. The operating model was clear: If sensitive data leaves the network for an external API call, we can observe it, log it, and stop it. But that model is starting to break.

A quiet hardware shift is pushing large language model (LLM) usage off the network and onto the endpoint. Call it Shadow AI 2.0, or the “bring your own model” (BYOM) era: Employees running capable models locally on laptops, offline, with no API calls and no obvious network signature. The governance conversation is still framed as “data exfiltration to the cloud,” but the more immediate enterprise risk is increasingly “unvetted inference inside the device.”

When inference happens locally, traditional data loss prevention (DLP) doesn’t see the interaction. And when security can’t see it, it can’t manage it.

Advertisement

Why local inference is suddenly practical

Two years ago, running a useful LLM on a work laptop was a niche stunt. Today, it’s routine for technical teams.

Three things converged:

  • Consumer-grade accelerators got serious: A MacBook Pro with 64GB unified memory can often run quantized 70B-class models at usable speeds (with practical limits on context length). What once required multi-GPU servers is now feasible on a high-end laptop for many real workflows.

  • Quantization went mainstream: It’s now easy to compress models into smaller, faster formats that fit within laptop memory often with acceptable quality tradeoffs for many tasks.

  • Distribution is frictionless: Open-weight models are a single command away, and the tooling ecosystem makes “download → run → chat” trivial.

The result: An engineer can pull down a multi‑GB model artifact, turn off Wi‑Fi, and run sensitive workflows locally, source code review, document summarization, drafting customer communications, even exploratory analysis over regulated datasets. No outbound packets, no proxy logs, no cloud audit trail.

From a network-security perspective, that activity can look indistinguishable from “nothing happened”.

Advertisement

The risk isn’t only data leaving the company anymore

If the data isn’t leaving the laptop, why should a CISO care?

Because the dominant risks shift from exfiltration to integrity, provenance, and compliance. In practice, local inference creates three classes of blind spots that most enterprises have not operationalized.

1. Code and decision contamination (integrity risk)

Local models are often adopted because they’re fast, private, and “no approval required.” The downside is that they’re frequently unvetted for the enterprise environment.

A common scenario: A senior developer downloads a community-tuned coding model because it benchmarks well. They paste in internal auth logic, payment flows, or infrastructure scripts to “clean it up.” The model returns output that looks competent, compiles, and passes unit tests, but subtly degrades security posture (weak input validation, unsafe defaults, brittle concurrency changes, dependency choices that aren’t allowed internally). The engineer commits the change.

Advertisement

If that interaction happened offline, you may have no record that AI influenced the code path at all. And when you later do incident response, you’ll be investigating the symptom (a vulnerability) without visibility into a key cause (uncontrolled model usage).

2. Licensing and IP exposure (compliance risk)

Many high-performing models ship with licenses that include restrictions on commercial use, attribution requirements, field-of-use limits, or obligations that can be incompatible with proprietary product development. When employees run models locally, that usage can bypass the organization’s normal procurement and legal review process.

If a team uses a non-commercial model to generate production code, documentation, or product behavior, the company can inherit risk that shows up later during M&A diligence, customer security reviews, or litigation. The hard part is not just the license terms, it’s the lack of inventory and traceability. Without a governed model hub or usage record, you may not be able to prove what was used where.

3. Model supply chain exposure (provenance risk)

Local inference also changes the software supply chain problem. Endpoints begin accumulating large model artifacts and the toolchains around them: ownloaders, converters, runtimes, plugins, UI shells, and Python packages.

Advertisement

There is a critical technical nuance here: The file format matters. While newer formats like Safetensors are designed to prevent arbitrary code execution, older Pickle-based PyTorch files can execute malicious payloads simply when loaded. If your developers are grabbing unvetted checkpoints from Hugging Face or other repositories, they aren’t just downloading data — they could be downloading an exploit.

Security teams have spent decades learning to treat unknown executables as hostile. BYOM requires extending that mindset to model artifacts and the surrounding runtime stack. The biggest organizational gap today is that most companies have no equivalent of a software bill of materials for models: Provenance, hashes, allowed sources, scanning, and lifecycle management.

Mitigating BYOM: treat model weights like software artifacts

You can’t solve local inference by blocking URLs. You need endpoint-aware controls and a developer experience that makes the safe path the easy path.

Here are three practical ways:

Advertisement

1. Move governance down to the endpoint

Network DLP and CASB still matter for cloud usage, but they’re not sufficient for BYOM. Start treating local model usage as an endpoint governance problem by looking for specific signals:

  • Inventory and detection: Scan for high-fidelity indicators like .gguf files larger than 2GB, processes like llama.cpp or Ollama, and local listeners on common default port 11434.

  • Process and runtime awareness: Monitor for repeated high GPU/NPU (neural processing unit) utilization from unapproved runtimes or unknown local inference servers.

  • Device policy: Use mobile device management (MDM) and endpoint detection and response (EDR) policies to control installation of unapproved runtimes and enforce baseline hardening on engineering devices. The point isn’t to punish experimentation. It’s to regain visibility.

2. Provide a paved road: An internal, curated model hub

Shadow AI is often an outcome of friction. Approved tools are too restrictive, too generic, or too slow to approve. A better approach is to offer a curated internal catalog that includes:

Advertisement
  • Approved models for common tasks (coding, summarization, classification)

  • Verified licenses and usage guidance

  • Pinned versions with hashes (prioritizing safer formats like Safetensors)

  • Clear documentation for safe local usage, including where sensitive data is and isn’t allowed. If you want developers to stop scavenging, give them something better.

3. Update policy language: “Cloud services” isn’t enough anymore

Most acceptable use policies talk about SaaS and cloud tools. BYOM requires policy that explicitly covers:

  • Downloading and running model artifacts on corporate endpoints

  • Acceptable sources

  • License compliance requirements

  • Rules for using models with sensitive data

  • Retention and logging expectations for local inference tools This doesn’t need to be heavy-handed. It needs to be unambiguous.

The perimeter is shifting back to the device

For a decade we moved security controls “up” into the cloud. Local inference is pulling a meaningful slice of AI activity back “down” to the endpoint.

5 signals shadow AI has moved to endpoints:

Advertisement
  • Large model artifacts: Unexplained storage consumption by .gguf or .pt files.

  • Local inference servers: Processes listening on ports like 11434 (Ollama).

  • GPU utilization patterns: Spikes in GPU usage while offline or disconnected from VPN.

  • Lack of model inventory: Inability to map code outputs to specific model versions.

  • License ambiguity: Presence of “non-commercial” model weights in production builds.

Shadow AI 2.0 isn’t a hypothetical future, it’s a predictable consequence of fast hardware, easy distribution, and developer demand. CISOs who focus only on network controls will miss what’s happening on the silicon sitting right on employees’ desks.

The next phase of AI governance is less about blocking websites and more about controlling artifacts, provenance, and policy at the endpoint, without killing productivity.

Jayachander Reddy Kandakatla is a senior MLOps engineer.

Welcome to the VentureBeat community!

Advertisement

Our guest posting program is where technical experts share insights and provide neutral, non-vested deep dives on AI, data infrastructure, cybersecurity and other cutting-edge technologies shaping the future of enterprise.

Read more from our guest post program — and check out our guidelines if you’re interested in contributing an article of your own!

Source link

Advertisement
Continue Reading

Trending

Copyright © 2025