Connect with us
DAPA Banner

Tech

Your developers are already running AI locally: Why on-device inference is the CISO’s new blind spot

Published

on

For the last 18 months, the CISO playbook for generative AI has been relatively simple: Control the browser.

Security teams tightened cloud access security broker (CASB) policies, blocked or monitored traffic to well-known AI endpoints, and routed usage through sanctioned gateways. The operating model was clear: If sensitive data leaves the network for an external API call, we can observe it, log it, and stop it. But that model is starting to break.

A quiet hardware shift is pushing large language model (LLM) usage off the network and onto the endpoint. Call it Shadow AI 2.0, or the “bring your own model” (BYOM) era: Employees running capable models locally on laptops, offline, with no API calls and no obvious network signature. The governance conversation is still framed as “data exfiltration to the cloud,” but the more immediate enterprise risk is increasingly “unvetted inference inside the device.”

When inference happens locally, traditional data loss prevention (DLP) doesn’t see the interaction. And when security can’t see it, it can’t manage it.

Advertisement

Why local inference is suddenly practical

Two years ago, running a useful LLM on a work laptop was a niche stunt. Today, it’s routine for technical teams.

Three things converged:

  • Consumer-grade accelerators got serious: A MacBook Pro with 64GB unified memory can often run quantized 70B-class models at usable speeds (with practical limits on context length). What once required multi-GPU servers is now feasible on a high-end laptop for many real workflows.

  • Quantization went mainstream: It’s now easy to compress models into smaller, faster formats that fit within laptop memory often with acceptable quality tradeoffs for many tasks.

  • Distribution is frictionless: Open-weight models are a single command away, and the tooling ecosystem makes “download → run → chat” trivial.

The result: An engineer can pull down a multi‑GB model artifact, turn off Wi‑Fi, and run sensitive workflows locally, source code review, document summarization, drafting customer communications, even exploratory analysis over regulated datasets. No outbound packets, no proxy logs, no cloud audit trail.

From a network-security perspective, that activity can look indistinguishable from “nothing happened”.

Advertisement

The risk isn’t only data leaving the company anymore

If the data isn’t leaving the laptop, why should a CISO care?

Because the dominant risks shift from exfiltration to integrity, provenance, and compliance. In practice, local inference creates three classes of blind spots that most enterprises have not operationalized.

1. Code and decision contamination (integrity risk)

Local models are often adopted because they’re fast, private, and “no approval required.” The downside is that they’re frequently unvetted for the enterprise environment.

A common scenario: A senior developer downloads a community-tuned coding model because it benchmarks well. They paste in internal auth logic, payment flows, or infrastructure scripts to “clean it up.” The model returns output that looks competent, compiles, and passes unit tests, but subtly degrades security posture (weak input validation, unsafe defaults, brittle concurrency changes, dependency choices that aren’t allowed internally). The engineer commits the change.

Advertisement

If that interaction happened offline, you may have no record that AI influenced the code path at all. And when you later do incident response, you’ll be investigating the symptom (a vulnerability) without visibility into a key cause (uncontrolled model usage).

2. Licensing and IP exposure (compliance risk)

Many high-performing models ship with licenses that include restrictions on commercial use, attribution requirements, field-of-use limits, or obligations that can be incompatible with proprietary product development. When employees run models locally, that usage can bypass the organization’s normal procurement and legal review process.

If a team uses a non-commercial model to generate production code, documentation, or product behavior, the company can inherit risk that shows up later during M&A diligence, customer security reviews, or litigation. The hard part is not just the license terms, it’s the lack of inventory and traceability. Without a governed model hub or usage record, you may not be able to prove what was used where.

3. Model supply chain exposure (provenance risk)

Local inference also changes the software supply chain problem. Endpoints begin accumulating large model artifacts and the toolchains around them: ownloaders, converters, runtimes, plugins, UI shells, and Python packages.

Advertisement

There is a critical technical nuance here: The file format matters. While newer formats like Safetensors are designed to prevent arbitrary code execution, older Pickle-based PyTorch files can execute malicious payloads simply when loaded. If your developers are grabbing unvetted checkpoints from Hugging Face or other repositories, they aren’t just downloading data — they could be downloading an exploit.

Security teams have spent decades learning to treat unknown executables as hostile. BYOM requires extending that mindset to model artifacts and the surrounding runtime stack. The biggest organizational gap today is that most companies have no equivalent of a software bill of materials for models: Provenance, hashes, allowed sources, scanning, and lifecycle management.

Mitigating BYOM: treat model weights like software artifacts

You can’t solve local inference by blocking URLs. You need endpoint-aware controls and a developer experience that makes the safe path the easy path.

Here are three practical ways:

Advertisement

1. Move governance down to the endpoint

Network DLP and CASB still matter for cloud usage, but they’re not sufficient for BYOM. Start treating local model usage as an endpoint governance problem by looking for specific signals:

  • Inventory and detection: Scan for high-fidelity indicators like .gguf files larger than 2GB, processes like llama.cpp or Ollama, and local listeners on common default port 11434.

  • Process and runtime awareness: Monitor for repeated high GPU/NPU (neural processing unit) utilization from unapproved runtimes or unknown local inference servers.

  • Device policy: Use mobile device management (MDM) and endpoint detection and response (EDR) policies to control installation of unapproved runtimes and enforce baseline hardening on engineering devices. The point isn’t to punish experimentation. It’s to regain visibility.

2. Provide a paved road: An internal, curated model hub

Shadow AI is often an outcome of friction. Approved tools are too restrictive, too generic, or too slow to approve. A better approach is to offer a curated internal catalog that includes:

Advertisement
  • Approved models for common tasks (coding, summarization, classification)

  • Verified licenses and usage guidance

  • Pinned versions with hashes (prioritizing safer formats like Safetensors)

  • Clear documentation for safe local usage, including where sensitive data is and isn’t allowed. If you want developers to stop scavenging, give them something better.

3. Update policy language: “Cloud services” isn’t enough anymore

Most acceptable use policies talk about SaaS and cloud tools. BYOM requires policy that explicitly covers:

  • Downloading and running model artifacts on corporate endpoints

  • Acceptable sources

  • License compliance requirements

  • Rules for using models with sensitive data

  • Retention and logging expectations for local inference tools This doesn’t need to be heavy-handed. It needs to be unambiguous.

The perimeter is shifting back to the device

For a decade we moved security controls “up” into the cloud. Local inference is pulling a meaningful slice of AI activity back “down” to the endpoint.

5 signals shadow AI has moved to endpoints:

Advertisement
  • Large model artifacts: Unexplained storage consumption by .gguf or .pt files.

  • Local inference servers: Processes listening on ports like 11434 (Ollama).

  • GPU utilization patterns: Spikes in GPU usage while offline or disconnected from VPN.

  • Lack of model inventory: Inability to map code outputs to specific model versions.

  • License ambiguity: Presence of “non-commercial” model weights in production builds.

Shadow AI 2.0 isn’t a hypothetical future, it’s a predictable consequence of fast hardware, easy distribution, and developer demand. CISOs who focus only on network controls will miss what’s happening on the silicon sitting right on employees’ desks.

The next phase of AI governance is less about blocking websites and more about controlling artifacts, provenance, and policy at the endpoint, without killing productivity.

Jayachander Reddy Kandakatla is a senior MLOps engineer.

Welcome to the VentureBeat community!

Advertisement

Our guest posting program is where technical experts share insights and provide neutral, non-vested deep dives on AI, data infrastructure, cybersecurity and other cutting-edge technologies shaping the future of enterprise.

Read more from our guest post program — and check out our guidelines if you’re interested in contributing an article of your own!

Source link

Advertisement
Continue Reading
Click to comment

You must be logged in to post a comment Login

Leave a Reply

Tech

Today’s NYT Connections Hints, Answers for April 13 #1037

Published

on

Looking for the most recent Connections answers? Click here for today’s Connections hints, as well as our daily answers and hints for The New York Times Mini Crossword, Wordle, Connections: Sports Edition and Strands puzzles.


Today’s NYT Connections puzzle is very tricky, especially the purple category. Read on for clues and today’s Connections answers.

The Times has a Connections Bot, like the one for Wordle. Go there after you play to receive a numeric score and to have the program analyze your answers. Players who are registered with the Times Games section can now nerd out by following their progress, including the number of puzzles completed, win rate, number of times they nabbed a perfect score and their win streak.

Advertisement

Read more: Hints, Tips and Strategies to Help You Win at NYT Connections Every Time

Hints for today’s Connections groups

Here are four hints for the groupings in today’s Connections puzzle, ranked from the easiest yellow group to the tough (and sometimes bizarre) purple group.

Yellow group hint: Let’s all go to the movies.

Advertisement

Green group hint: Abracadabra!

Blue group hint: Television people.

Purple group hint: Think hats.

Answers for today’s Connections groups

Yellow group: Seen outside a theater.

Advertisement

Green group: Accessories for a magician.

Blue group: TV show title surnames.

Purple group: They have caps.

Read more: Wordle Cheat Sheet: Here Are the Most Popular Letters Used in English Words

Advertisement

What are today’s Connections answers?

completed NYT Connections puzzle for April 13, 2026

The completed NYT Connections puzzle for April 13, 2026.

NYT/Screenshot by CNET

The yellow words in today’s Connections

The theme is seen outside a theater. The four answers are box office, marquee, ticket line and velvet rope.

The green words in today’s Connections

The theme is accessories for a magician. The four answers are cape, handkerchief, magic wand and rabbit.

Advertisement

The blue words in today’s Connections

The theme is TV show title surnames. The four answers are House, Lasso, Montana and Soprano.

The purple words in today’s Connections

The theme is they have caps. The four answers are baseball player, camera lens, mushroom and pen.

Source link

Advertisement
Continue Reading

Tech

Three Apple Stores closing in June, one was unionized

Published

on

Three Apple Store locations in struggling malls are set to close permanently as summer kicks off, with one of them the controversial unionized store in Towson, Maryland.

Bright modern Apple retail store with glass front, wooden tables displaying iPhones, iPads, laptops and accessories, illuminated product posters on both walls, and large glowing Apple logo overhead
Apple Trumbull | Image Credit: Apple

The stores in question are Apple North County, in Escondido, California, Apple Trumbull in Trumbull, Connecticut, and Apple Towson Town Center in Towson, Maryland. Notably, Apple Towson was Apple’s first unionized store.
Most employees will be shifted to nearby locations with no further action required by the employee, provided they agree to stay with the company. The unionized Towson employees will be eligible to apply for open roles at Apple, as per the existing bargaining agreement.
Continue Reading on AppleInsider | Discuss on our Forums

Source link

Continue Reading

Tech

Today’s NYT Mini Crossword Answers for April 13

Published

on

Looking for the most recent Mini Crossword answer? Click here for today’s Mini Crossword hints, as well as our daily answers and hints for The New York Times Wordle, Strands, Connections and Connections: Sports Edition puzzles.


Need some help with today’s Mini Crossword? Read on for all the answers. And if you could use some hints and guidance for daily solving, check out our Mini Crossword tips.

If you’re looking for today’s Wordle, Connections, Connections: Sports Edition and Strands answers, you can visit CNET’s NYT puzzle hints page.

Advertisement

Read more: Tips and Tricks for Solving The New York Times Mini Crossword

Let’s get to those Mini Crossword clues and answers.

completed-nyt-mini-crossword-puzzle-for-april-13-2026.png

The completed NYT Mini Crossword puzzle for April 13, 2026.

Advertisement

NYT/Screenshot by CNET

Mini across clues and answers

1A clue: Symbol in the middle of Captain America’s shield
Answer: STAR

5A clue: Chanel’s interlocking C’s, e.g.
Answer: LOGO

6A clue: Content creator’s output
Answer: VIDEO

Advertisement

7A clue: Two-word denial
Answer: IDONT

8A clue: Fits one inside of the other
Answer: NESTS

Mini down clues and answers

1D clue: Part of an office presentation
Answer: SLIDE

2D clue: List that might have check boxes
Answer: TODOS

Advertisement

3D clue: Contract negotiator
Answer: AGENT

4D clue: Jimmy Fallon’s house band, with “The”
Answer: ROOTS

6D clue: French for “wine”
Answer: VIN

Advertisement

Source link

Continue Reading

Tech

Why Google’s Nano Banana Pro Image Model Has Such A Weird Name

Published

on





Naming AI products is a bit hit-or-miss. Some names sound as if they were polished in a branding lab for six months, while others feel as though they were just pulled from a hat. Claude has a certain elegance. Gemini is fine. ChatGPT, on the other hand, is a rubbish name and only became familiar through brute force when it was suddenly absolutely everywhere

Nano Banana, Google Gemini’s AI image generator that enables anyone to create realistic-looking pictures, is called Gemini 3 Pro Image Preview in Google’s technical documentation. However, the name “Nano Banana” is both more official and less official than you might think. Google openly calls it Nano Banana Pro — and even Nano Banana 2, now — but that wasn’t the original plan. 

Nano Banana Pro has such a weird name because that moniker was never intended to be taken seriously. The team needed a temporary name for Arena.ai (then called LMArena), the crowdsourced model-testing platform where systems are compared anonymously. The codename wasn’t chosen until the last minute. Product Manager Naina Raisinghani was pushed to come up with something on the spot and suggested Nano Banana. It was a combination of two of her nicknames. “Some of my friends call me Naina Banana, and others call me Nano because I’m short and I like computers. So I just smushed my two nicknames together,” Naina revealed on Google’s blog, The Keyword.

Advertisement

Nano Banana quickly caught on

Despite Google’s attempts to keep its identity secret on Arena.ai, some people were quick to speculate that the highly rated new image generation and editing tool was a Google product. It was initially uploaded to Arena.ai on August 12, 2025. Within days, users were sharing their AI-generated creations on social media. After a week of speculation, a couple of X posts fueled users’ suspicions. Product Lead for Google AI Studio, Logan Kilpatrick, posted a banana emoji, and Naina Raisinghani, the developer behind the name, shared a picture of a banana gaffer-taped to a wall. Nano Banana was officially launched on August 26, 2025, upstaging ChatGPT as the most popular AI image generator.

It’s not the first tech product with “banana” in its name. We might be more familiar with Apple, Blackberry, and Raspberry Pi, but you can also purchase a bananaphone — a banana-shaped Bluetooth headset to pair with your smartphone. There’s also a 2019 research paper with a BANANAS algorithm, which stands for Bayesian Optimization with Neural Architectures for Neural Architecture Search. (You have to respect the contrivance even if it doesn’t quite work.) Tech companies are still naming things after fruit. OpenAI internally used “Strawberry” for the project that became o1, and Meta is currently working on an AI model nicknamed “Avocado.”

Nano Banana may not have been meant as the official name, but it stuck because people liked it. Companies spend fortunes chasing that kind of stickiness, and Google stumbled into it. The model got noticed, the odd codename was memorable, and Google was smart enough not to crush the joke with a committee-approved replacement.

Advertisement



Source link

Advertisement
Continue Reading

Tech

Today’s NYT Connections: Sports Edition Hints, Answers for April 13 #567

Published

on

Looking for the most recent regular Connections answers? Click here for today’s Connections hints, as well as our daily answers and hints for The New York Times Mini Crossword, Wordle and Strands puzzles.


Today’s Connections: Sports Edition is a tough one. If you’re struggling with it but still want to solve it, read on for hints and the answers.

Connections: Sports Edition is published by The Athletic, the subscription-based sports journalism site owned by The Times. It doesn’t appear in the NYT Games app, but it does in The Athletic’s own app. Or you can play it for free online.

Advertisement

Read more: NYT Connections: Sports Edition Puzzle Comes Out of Beta

Hints for today’s Connections: Sports Edition groups

Here are four hints for the groupings in today’s Connections: Sports Edition puzzle, ranked from the easiest yellow group to the tough (and sometimes bizarre) purple group.

Yellow group hint: Get your glove ready!

Advertisement

Green group hint: Sweat equity.

Blue group hint: There used to be a ballpark.

Purple group hint: Not night.

Answers for today’s Connections: Sports Edition groups

Yellow group: Field a baseball.

Advertisement

Green group: Work hard.

Blue group: Former MLB stadiums.

Purple group: ____ day.

Read more: Wordle Cheat Sheet: Here Are the Most Popular Letters Used in English Words

Advertisement

What are today’s Connections: Sports Edition answers?

completed NYT Connections: Sports Edition puzzle for April 13, 2026

The completed NYT Connections: Sports Edition puzzle for April 13, 2026.

NYT/Screenshot by CNET

The yellow words in today’s Connections

The theme is field a baseball. The four answers are catch, field, pick and scoop.

The green words in today’s Connections

The theme is work hard. The four answers are grind, labor, strain and toil.

Advertisement

The blue words in today’s Connections

The theme is former MLB stadiums. The four answers are Polo, Shea, Turner and Veterans.

The purple words in today’s Connections

The theme is ____ day. The four answers are draft, game, opening and Ryan.

Source link

Advertisement
Continue Reading

Tech

Intel and SambaNova just built a three-chip AI machine that splits work between GPUs, RDUs, and Xeon

Published

on


  • GPUs handle prefill operations by converting prompts into key-value caches
  • SambaNova RDUs generate tokens at high throughput and low latency
  • Intel Xeon 6 processors manage workload distribution and execute compiled code

Intel and SambaNova Systems have introduced a joint hardware blueprint combining GPUs, SambaNova RDUs, and Intel Xeon 6 processors for large-scale inference workloads.

The system assigns GPUs to prefill operations, RDUs to decoding, and Xeon CPUs to execution and orchestration tasks across agent-driven environments.

Source link

Advertisement
Continue Reading

Tech

Amazon laid off 30,000 workers while CEO Andy Jassy got a 30% pay bump

Published

on


Amazon published its annual proxy statement yesterday, revealing Jassy’s compensation for last year.
Read Entire Article
Source link

Continue Reading

Tech

Green Powered Challenge: Solar Powered Pi Hosts Websites In RAM

Published

on

If you started with computers early enough, you’ll remember the importance of the RAMdisk concept: without a hard drive and with floppies slow and swapping constantly, everything had to live in RAM. That’s not done much these days, but [Quackieduckie]’s solar powered Pi Zero W web server has gone back to it to save its SD card.

Sustainability and low power is the name of the game. Starting with a Pi Zero W means low power is the default; a an SLS-printed aluminum case that doubles as the heat sink– while looking quite snazzy–saves power that would otherwise be used for cooling. The STLs are available through the project page if you like the look and have a hankering for passively cooled Pi. Even under load [Quackieduckie] reports temperatures of just 29.9°C,  less than a degree over idle.

The software stack is of course key to a server, and here he’s using Alpine Linux running in “diskless mode”– that’s the equivalent of what us oldsters would think of as the RAMdisk. That’s not that unusual for servers, but we don’t see it much on these pages. It’s a minimal setup to save processing, and thus electrical power, with only a handful of services kept running: lighttpd, a lightweight webserver, and duckiebox, a python-based file server, along with SSHD and dchron; together they consume 27 MB of RAM, leaving the rest of the 512 MB DDR2 the Pi comes with to quickly serve up websites without the overhead of SD card access.

As a webserver, [Quackieduckie] tested it with 50 simultaneous connections, which would be rather a lot for most small, personal web sites, and while it did slow down to an average 1.3s per response that’s perfectly usable and faster than we’d have expected from this hardware. While the actual power consumption figures aren’t given, we know from experience it’s not going to be drawing more than a watt or so. With a reasonably sized battery and solar cell– [Quackieduckie] suggests 20W–it should run until the cows come home.

Advertisement

This isn’t the first solar-powered web server we’ve seen, but this one was submitted for the 2026 Green Powered Challenge, which runs until April 24th.

Source link

Advertisement
Continue Reading

Tech

The Complex Transformations Underlying MC Escher’s Works

Published

on

Self-similar images are rather common, which are images in which the same image is repeated on a smaller scale somewhere within the image that one is looking at, something which is also referred to as the Droste effect. Yet in [MC Escher]’s 1956 Prentententoonstelling (‘picture gallery’) drawing, this self-similar image is somehow also the foreground image, from where it just keeps looping around in an endless dance. How this effect is accomplished and what the mathematical transformations behind it are and how they work is explained in a recent video by [3Blue1Brown].

The video uses previous work by [B. de Smit] and [H. W. Lenstra Jr] whose 2003 paper detailed the underlying transformations, as well as the mystery of the center of the work.

Although [MC Escher] created a transformation grid with square rectangles into which a non-transformed image could be copied verbatim, he left the center as a void with just his signature in it, leaving many to guess how one might be able to fill in this area with something that made sense. In the work by [Smit] et al. it was postulated that by treating the work as having been drawn on an elliptic curve over a field of complex numbers this might be possible.

While the transformation is simple enough at first, with just four rectangles at different zoom levels to make up the corners, the trick is to connect these rectangles. Using the demonstrated complex method this can be automated, with the central void now filled in and creating its own Droste effect. This once again demonstrates the beautifully complex mathematics in [Escher]’s works, despite him never having had any formal mathematical education.

Advertisement

Source link

Advertisement
Continue Reading

Tech

Best 2-in-1 Laptops (2026): Microsoft, Lenovo, and the iPad

Published

on

There will always be a use case for owning both a laptop and a tablet as stand-alone products. But the 2-in-1 laptop is the utopian dream of combining these two into a single device.

Of all the models I’ve tested, no 2-in-1 laptop is equally good at being both a tablet and a laptop. They always lean toward one or the other. But that doesn’t mean you shouldn’t buy one, especially since the convenience of having both in one device makes it an easier pill to swallow, price-wise.

The products below should meet most people’s needs. But if none are a fit for you, check out our other computer buying guides, including the Best Cheap Laptops, the Best Tablets, and the Best iPad.

Table of Contents

Advertisement

Detachable Tablets

Microsoft

Surface Pro 13-inch (11th Edition, 2024)

If you want a 2-in-1, think first about a detachable tablet. These are basically tablets that attach to a keyboard. This form factor emphasizes being able to switch between tablet and laptop modes. It’s just as functional as a tablet as it is as a laptop. The Surface Pro is the epitome of this design, pioneering the idea of a tablet with a built-in kickstand that runs a full version of Windows.

Advertisement

Microsoft has refined the hardware over the years, but it wasn’t until the 2024 model that it came into its own. That’s largely thanks to the Qualcomm Snapdragon X Elite (and Plus) processor, which finally gave the device an appropriate amount of performance and battery life. While it’s not cheap (especially once you include the Type Cover), I love that you can now use the keyboard while detached from the screen, making it even more adaptable in scenarios away from a desk. To compete with the iPad Pro, there’s even an OLED model (with 120-Hz refresh rate) available, which really brings visuals on the display to life.

Last year, Microsoft came out with a smaller and more affordable model, the Surface Pro 12. This is the most successful small tablet Microsoft has ever made, and a big reason is because it doesn’t cheap out on quality or shrink down the size too much. With a 12-inch screen, it still allows the keyboard to be large enough to be comfortable typing on. It doesn’t have the option for an OLED screen, but this is still a surprisingly premium-feeling device that is even more portable than its older sibling.

Not only is the Surface Pro 12 cheaper overall, it’s also the only 256-GB storage model on offer. Because Surface devices run a full version of Windows, they are the best 2-in-1 devices to use as full laptop replacements. While the hardware is there to make for a good tablet, Windows isn’t so friendly with touch and doesn’t have a touch-first app ecosystem to support it. That’s where iPads come into play.

The iPad Air and iPad Pro are the best tablets you can buy, largely thanks to the breadth of touch-first apps available in the App Store. In many ways, that’s what makes an iPad such an ideal 2-in-1 laptop, especially if you actually want to use it as a tablet. They are also easier to hold in one hand, as they are lighter than the Surface devices. These days, these iPads are increasingly legitimate laptop replacements too. With the Magic Keyboard attachment, you can add an additional USB-C port and a full-size keyboard and trackpad. I like that this design doesn’t rely on a kickstand either, which makes it easier to use on your lap than the Surface.

Advertisement

iPadOS still isn’t perfect, but with the introduction of windowing and better cursor support, they work as laptops better than ever. The latest model I tested, the M4 iPad Air, is immensely powerful, and with the Magic Keyboard attached, it’s a really solid 2-in-1 laptop that comes in cheaper than the Surface Pro with the keyboard included. It’s plenty of performance for just about anything you’d want to do with an iPad, especially if you opt for the larger 13-inch model. My only real complaint is that the palm rests on the Magic Keyboard are quite small.

Source link

Continue Reading

Trending

Copyright © 2025