Connect with us
DAPA Banner

Tech

Nvidia introduces Vera Rubin, a seven-chip AI platform with OpenAI, Anthropic and Meta on board

Published

on

Nvidia on Monday took the wraps off Vera Rubin, a sweeping new computing platform built from seven chips now in full production — and backed by an extraordinary lineup of customers that includes Anthropic, OpenAI, Meta and Mistral AI, along with every major cloud provider.

The message to the AI industry, and to investors, was unmistakable: Nvidia is not slowing down. The Vera Rubin platform claims up to 10x more inference throughput per watt and one-tenth the cost per token compared with the Blackwell systems that only recently began shipping. CEO Jensen Huang, speaking at the company’s annual GTC conference, called it “a generational leap” that would kick off “the greatest infrastructure buildout in history.” Amazon Web Services, Google Cloud, Microsoft Azure and Oracle Cloud Infrastructure will all offer the platform, and more than 80 manufacturing partners are building systems around it.

“Vera Rubin is a generational leap — seven breakthrough chips, five racks, one giant supercomputer — built to power every phase of AI,” Huang declared. “The agentic AI inflection point has arrived with Vera Rubin kicking off the greatest infrastructure buildout in history.”

In any other industry, such rhetoric might be dismissed as keynote theater. But Nvidia occupies a singular position in the global economy — a company whose products have become so essential to the AI boom that its market capitalization now rivals the GDP of mid-sized nations. When Huang says the infrastructure buildout is historic, the CEOs of the companies actually writing the checks are standing behind him, nodding.

Advertisement

Dario Amodei, the chief executive of Anthropic, said Nvidia’s platform “gives us the compute, networking and system design to keep delivering while advancing the safety and reliability our customers depend on.” Sam Altman, the chief executive of OpenAI, said that “with Nvidia Vera Rubin, we’ll run more powerful models and agents at massive scale and deliver faster, more reliable systems to hundreds of millions of people.”

Inside the seven-chip architecture designed to power the age of AI agents

The Vera Rubin platform brings together the Nvidia Vera CPU, Rubin GPU, NVLink 6 Switch, ConnectX-9 SuperNIC, BlueField-4 DPU, Spectrum-6 Ethernet switch and the newly integrated Groq 3 LPU — a purpose-built inference accelerator. Nvidia organized these into five interlocking rack-scale systems that function as a unified supercomputer.

The flagship NVL72 rack integrates 72 Rubin GPUs and 36 Vera CPUs connected by NVLink 6. Nvidia says it can train large mixture-of-experts models using one-quarter the GPUs required on Blackwell, a claim that, if validated in production, would fundamentally alter the economics of building frontier AI systems.

The Vera CPU rack packs 256 liquid-cooled processors into a single rack, sustaining more than 22,500 concurrent CPU environments — the sandboxes where AI agents execute code, validate results and iterate. Nvidia describes the Vera CPU as the first processor purpose-built for agentic AI and reinforcement learning, featuring 88 custom-designed Olympus cores and LPDDR5X memory delivering 1.2 terabytes per second of bandwidth at half the power of conventional server CPUs.

Advertisement

The Groq 3 LPX rack, housing 256 inference processors with 128 gigabytes of on-chip SRAM, targets the low-latency demands of trillion-parameter models with million-token contexts. The BlueField-4 STX storage rack provides what Nvidia calls “context memory” — high-speed storage for the massive key-value caches that agentic systems generate as they reason across long, multi-step tasks. And the Spectrum-6 SPX Ethernet rack ties it all together with co-packaged optics delivering 5x greater optical power efficiency than traditional transceivers.

Why Nvidia is betting the future on autonomous AI agents — and rebuilding its stack around them

The strategic logic binding every announcement Monday into a single narrative is Nvidia’s conviction that the AI industry is crossing a threshold. The era of chatbots — AI that responds to a prompt and stops — is giving way to what Huang calls “agentic AI“: systems that reason autonomously for hours or days, write and execute software, call external tools, and continuously improve.

This isn’t just a branding exercise. It represents a genuine architectural shift in how computing infrastructure must be designed. A chatbot query might consume milliseconds of GPU time. An agentic system orchestrating a drug discovery pipeline or debugging a complex codebase might run continuously, consuming CPU cycles to execute code, GPU cycles to reason, and massive storage to maintain context across thousands of intermediate steps. That demands not just faster chips, but a fundamentally different balance of compute, memory, storage and networking.

Advertisement

Nvidia addressed this with the launch of its Agent Toolkit, which includes OpenShell, a new open-source runtime that enforces security and privacy guardrails for autonomous agents. The enterprise adoption list is remarkable: Adobe, Atlassian, Box, Cadence, Cisco, CrowdStrike, Dassault Systèmes, IQVIA, Red Hat, Salesforce, SAP, ServiceNow, Siemens and Synopsys are all integrating the toolkit into their platforms. Nvidia also launched NemoClaw, an open-source stack that lets users install its Nemotron models and OpenShell runtime in a single command to run secure, always-on AI assistants on everything from RTX laptops to DGX Station supercomputers.

The company separately announced Dynamo 1.0, open-source software it describes as the first “operating system” for AI inference at factory scale. Dynamo orchestrates GPU and memory resources across clusters and has already been adopted by AWS, Azure, Google Cloud, Oracle, Cursor, Perplexity, PayPal and Pinterest. Nvidia says it boosted Blackwell inference performance by up to 7x in recent benchmarks.

The Nemotron coalition and Nvidia’s play to shape the open-source AI landscape

If Vera Rubin represents Nvidia’s hardware ambition, the Nemotron Coalition represents its software ambition. Announced Monday, the coalition is a global collaboration of AI labs that will jointly develop open frontier models trained on Nvidia’s DGX Cloud. The inaugural members — Black Forest Labs, Cursor, LangChain, Mistral AI, Perplexity, Reflection AI, Sarvam and Thinking Machines Lab, the startup led by former OpenAI executive Mira Murati — will contribute data, evaluation frameworks and domain expertise.

Advertisement

The first model will be co-developed by Mistral AI and Nvidia and will underpin the upcoming Nemotron 4 family. “Open models are the lifeblood of innovation and the engine of global participation in the AI revolution,” Huang said.

Nvidia also expanded its own open model portfolio significantly. Nemotron 3 Ultra delivers what the company calls frontier-level intelligence with 5x throughput efficiency on Blackwell. Nemotron 3 Omni integrates audio, vision and language understanding. Nemotron 3 VoiceChat supports real-time, simultaneous conversations. And the company previewed GR00T N2, a next-generation robot foundation model that it says helps robots succeed at new tasks in new environments more than twice as often as leading alternatives, currently ranking first on the MolmoSpaces and RoboArena benchmarks.

The open-model push serves a dual purpose. It cultivates the developer ecosystem that drives demand for Nvidia hardware, and it positions Nvidia as a neutral platform provider rather than a competitor to the AI labs building on its chips — a delicate balancing act that grows more complex as Nvidia’s own models grow more capable.

From operating rooms to orbit: how Vera Rubin’s reach extends far beyond the data center

Advertisement

The vertical breadth of Monday’s announcements was almost disorienting. Roche revealed it is deploying more than 3,500 Blackwell GPUs across hybrid cloud and on-premises environments in the U.S. and Europe — the largest announced GPU footprint in the pharmaceutical industry. The company is using the infrastructure for biological foundation models, drug discovery and digital twins of manufacturing facilities, including its new GLP-1 facility in North Carolina. Nearly 90 percent of Genentech’s eligible small-molecule programs now integrate AI, Roche said, with one oncology molecule designed 25 percent faster and a backup candidate delivered in seven months instead of more than two years.

In autonomous vehicles, BYD, Geely, Isuzu and Nissan are building Level 4-ready vehicles on Nvidia’s Drive Hyperion platform. Nvidia and Uber expanded their partnership to launch autonomous vehicles across 28 cities on four continents by 2028, starting with Los Angeles and San Francisco in the first half of 2027. The company introduced Alpamayo 1.5, a reasoning model for autonomous driving already downloaded by more than 100,000 automotive developers, and Nvidia Halos OS, a safety architecture built on ASIL D-certified foundations for production-grade autonomy.

Nvidia also released the first domain-specific physical AI platform for healthcare robotics, anchored by Open-H — the world’s largest healthcare robotics dataset, with over 700 hours of surgical video. CMR Surgical, Johnson & Johnson MedTech and Medtronic are among the adopters.

And then there was space. The Vera Rubin Space Module delivers up to 25x more AI compute for orbital inferencing compared with the H100 GPU. Aetherflux, Axiom Space, Kepler Communications, Planet Labs and Starcloud are building on it. “Space computing, the final frontier, has arrived,” Huang said, deploying the kind of line that, from another executive, might draw eye-rolls — but from the CEO of a company whose chips already power the majority of the world’s AI workloads, lands differently.

Advertisement

The deskside supercomputer and Nvidia’s quiet push into enterprise hardware

Amid the spectacle of trillion-parameter models and orbital data centers, Nvidia made a quieter but potentially consequential move: it launched the DGX Station, a deskside system powered by the GB300 Grace Blackwell Ultra Desktop Superchip that delivers 748 gigabytes of coherent memory and up to 20 petaflops of AI compute performance. The system can run open models of up to one trillion parameters from a desk.

Snowflake, Microsoft Research, Cornell, EPRI and Sungkyunkwan University are among the early users. DGX Station supports air-gapped configurations for regulated industries, and applications built on it move seamlessly to Nvidia’s data center systems without rearchitecting — a design choice that creates a natural on-ramp from local experimentation to large-scale deployment.

Nvidia also updated DGX Spark, its more compact system, with support for clustering up to four units into a “desktop data center” with linear performance scaling. Both systems ship preconfigured with NemoClaw and the Nvidia AI software stack, and support models including Nemotron 3, Google Gemma 3, Qwen3, DeepSeek V3.2, Mistral Large 3 and others.

Advertisement

Adobe and Nvidia separately announced a strategic partnership to develop the next generation of Firefly models using Nvidia’s computing technology and libraries. Adobe will also build a cloud-native 3D digital twin solution for marketing on Nvidia Omniverse and integrate Nemotron capabilities into Adobe Acrobat. The partnership spans creative tools including Photoshop, Premiere Pro, Frame.io and Adobe Experience Platform.

Building the factories that build intelligence: Nvidia’s AI infrastructure blueprint

Perhaps the most telling indicator of where Nvidia sees the industry heading is the Vera Rubin DSX AI Factory reference design — essentially a blueprint for constructing entire buildings optimized to produce AI. The reference design outlines how to integrate compute, networking, storage, power and cooling into a system that maximizes what Nvidia calls “tokens per watt,” along with an Omniverse DSX Blueprint for creating digital twins of these facilities before they are built.

The software stack includes DSX Max-Q for dynamic power provisioning — which Nvidia says enables 30 percent more AI infrastructure within a fixed-power data center — and DSX Flex, which connects AI factories to power-grid services to unlock what the company estimates is 100 gigawatts of stranded grid capacity. Energy leaders Emerald AI, GE Vernova, Hitachi and Siemens Energy are using the architecture. Nscale and Caterpillar are building one of the world’s largest AI factories in West Virginia using the Vera Rubin reference design.

Advertisement

Industry partners Cadence, Dassault Systèmes, Eaton, Jacobs, Schneider Electric, Siemens, PTC, Switch, Trane Technologies and Vertiv are contributing simulation-ready assets and integrating their platforms. CoreWeave is using Nvidia’s DSX Air to run operational rehearsals of AI factories in the cloud before physical delivery.

“In the age of AI, intelligence tokens are the new currency, and AI factories are the infrastructure that generates them,” Huang said. It is the kind of formulation — tokens as currency, factories as mints — that reveals how Nvidia thinks about its place in the emerging economic order.

What Nvidia’s grand vision gets right — and what remains unproven

The scale and coherence of Monday’s announcements are genuinely impressive. No other company in the semiconductor industry — and arguably no other technology company, period — can present an integrated stack spanning custom silicon, systems architecture, networking, storage, inference software, open models, agent frameworks, safety runtimes, simulation platforms, digital twin infrastructure and vertical applications from drug discovery to autonomous driving to orbital computing.

Advertisement

But scale and coherence are not the same as inevitability. The performance claims for Vera Rubin, while dramatic, remain largely unverified by independent benchmarks. The agentic AI thesis that underpins the entire platform — the idea that autonomous, long-running AI agents will become the dominant computing workload — is a bet on a future that has not yet fully materialized. And Nvidia’s expanding role as a provider of models, software, and reference architectures raises questions about how long its hardware customers will remain comfortable depending so heavily on a single supplier for so many layers of their stack.

Competitors are not standing still. AMD continues to close the gap on data center GPU performance. Google’s TPUs power some of the world’s largest AI training runs. Amazon’s Trainium chips are gaining traction inside AWS. And a growing cohort of startups is attacking various pieces of the AI infrastructure puzzle.

Yet none of them showed up at GTC on Monday with endorsements from the CEOs of Anthropic and OpenAI. None of them announced seven new chips in full production simultaneously. And none of them presented a vision this comprehensive for what comes next.

There is a scene that repeats at every GTC: Huang, in his trademark leather jacket, holds up a chip the way a jeweler holds up a diamond, rotating it slowly under the stage lights. It is part showmanship, part sermon. But the congregation keeps growing, the chips keep getting faster, and the checks keep getting larger. Whether Nvidia is building the greatest infrastructure in history or simply the most profitable one may, in the end, be a distinction without a difference.

Advertisement

Source link

Continue Reading
Click to comment

You must be logged in to post a comment Login

Leave a Reply

Tech

Restaurants are forcing us to put phones away, and I’m not complaining

Published

on

A growing number of bars and restaurants across the United States are embracing a phone-free experience, reflecting a broader cultural shift toward reducing screen time and encouraging real-world connection. From upscale supper clubs to neighborhood cocktail bars, establishments are introducing policies that either restrict phone usage or actively incentivize guests to put their devices away.

At the heart of this trend is a rising awareness of the negative effects smartphones and social media can have on attention, memory, and interpersonal relationships. Studies continue to highlight how constant digital engagement impacts learning, socialization, and even self-esteem. With Americans reportedly checking their phones around 144 times a day and spending nearly 4.5 hours on their devices, the pushback against screen dependency is gaining traction.

Younger generations, particularly Gen Z, are leading this shift

Surveys indicate that a significant portion of them intentionally disconnect from their devices, followed by millennials and older age groups. This growing appetite for “analog” experiences is now influencing the hospitality industry in noticeable ways.

Restaurants and bars in at least 11 U.S. states have already introduced some form of phone restriction. Washington, D.C., currently leads with the highest number of such venues. Some establishments take a strict approach, such as locking phones away in secure pouches for the duration of a visit, while others offer softer incentives like free desserts for diners who keep their devices off the table.

The reasoning behind these policies is simple: removing phones enhances human interaction. Business owners and industry experts argue that without digital distractions, guests are more engaged with their company, their surroundings, and even their food. Chefs have also noted that phones can detract from the dining experience, making meals feel less memorable.

Advertisement

For customers, the impact can be surprisingly profound

Many report feeling more present and emotionally connected during phone-free outings. Experiences that might otherwise be fragmented by notifications become more immersive and meaningful.

Looking ahead, the trend is expected to expand beyond independent venues. As digital fatigue continues to grow and awareness of screen-time effects increases, more mainstream chains and public spaces may experiment with similar policies. While not everyone may be ready to give up their phones during a night out, the rise of phone-free dining suggests a clear shift: people are beginning to value presence over perpetual connectivity.

Restaurants are finally pushing back against the constant glow of screens at the table, and honestly, it feels long overdue. Dining out was never meant to compete with notifications and endless scrolling. By nudging people to put their phones away, these places are restoring something we’ve quietly lost – real conversation, attention, and presence. It may feel restrictive at first, but the payoff is a far more meaningful experience.

Source link

Advertisement
Continue Reading

Tech

Rec Room shutting down: Once valued at $3.5B, social gaming platform finds profits elusive

Published

on

Rec Room, the Seattle-based social gaming company once valued at $3.5 billion, is shutting down its platform on June 1, leaving the future of the company and its employees unclear.

The company made the announcement Monday afternoon, saying it couldn’t find a path to profitability even after serving more than 150 million players over the past decade.

“Despite this popularity, we never quite figured out how to make Rec Room a sustainably profitable business,” the company said in its post announcing the news. “Our costs always ended up overwhelming the revenue we brought in.”

FOLLOW-UP: Snap acquires assets from Rec Room amid shutdown

Advertisement

The platform will go dark at noon Pacific on June 1. Starting immediately, Rec Room is blocking new account creation, new friend requests, and new subscriptions to its Rec Room Plus membership. Creators can no longer publish new monetized content. Token purchases end May 1, creator earnings stop May 18, and a final creator payout will be processed on June 1.

Rec Room users, posting in the community Discord server, expressed shock and surprise, with some holding out hope that the announcement was an early April Fool’s joke.

Alas, it appears not.

“We spent a long time trying to find a way to make the numbers work,” the post said. “But with the recent shift in the VR market, along with broader headwinds in gaming, the path to profitability has gotten tough enough that we’ve made the difficult decision to shut things down.”

Advertisement

The company said it was making the decision now “while we still have the ability to wind things down thoughtfully and do right by the people who built this with us.”

Nick Fajt
Nick Fajt, co-founder and CEO of Rec Room, in Seattle in 2017. (GeekWire File Photo / Kurt Schlosser)

Rec Room was founded in 2016 by Nick Fajt, Cameron Brown and a handful of other co-founders under the name Against Gravity. The Seattle startup built a cross-platform social gaming app that lets players create and share games, virtual goods and experiences across phones, consoles, PCs and VR headsets.

The company attracted backing from Sequoia Capital, Index Ventures, Madrona Venture Group, Coatue Management and others, raising $294 million across six rounds. Its December 2021 Series F valued the company at $3.5 billion, making it one of Seattle’s most prominent unicorns.

Rec Room’s popularity surged during the pandemic as players flocked to virtual hangouts, and the company said it surpassed 100 million lifetime users. But growth in the broader gaming market slowed in the years that followed, and Rec Room’s ambitions outpaced its revenue.

Rec Room laid off 16% of its staff in March 2025 and then cut roughly half its remaining workforce five months later, eliminating 141 positions and shrinking from about 310 employees to just over 100 people at the time.

Advertisement

Fajt said back then that the company needed to become self-sustaining and could no longer count on raising more money, but noted that Rec Room had enough runway to operate into 2029.

“If we had just kept going, we would have run out of money in the next couple of years,” he wrote at the time. “And with no money left, we would have had to lay everyone off.”

The company bet heavily on a vision of letting anyone create games on any device. It rolled out AI features including Maker AI for game creation and an artificial intelligence companion called Roomie, though the per-user costs of AI exceeded subscription revenue.

As of last September, revenue from user-generated content was growing about 70% year over year, and creators earned more than $1 million in a single quarter for the first time.

Advertisement

However, as noted by Fajt in public posts, the margins on user generated content were thin: Rec Room keeps only about 30 cents of every dollar of sales of user-generated content, after paying platforms and creators, compared with 70 cents on sales of first-party content.

Source link

Continue Reading

Tech

Today’s NYT Connections: Sports Edition Hints, Answers for April 6 #560

Published

on

Looking for the most recent regular Connections answers? Click here for today’s Connections hints, as well as our daily answers and hints for The New York Times Mini Crossword, Wordle and Strands puzzles.


Today’s Connections: Sports Edition is a tough one. If you’re struggling with it but still want to solve it, read on for hints and the answers.

Connections: Sports Edition is published by The Athletic, the subscription-based sports journalism site owned by The Times. It doesn’t appear in the NYT Games app, but it does in The Athletic’s own app. Or you can play it for free online.

Advertisement

Read more: NYT Connections: Sports Edition Puzzle Comes Out of Beta

Hints for today’s Connections: Sports Edition groups

Here are four hints for the groupings in today’s Connections: Sports Edition puzzle, ranked from the easiest yellow group to the tough (and sometimes bizarre) purple group.

Yellow group hint: City of Angels.

Advertisement

Green group hint: Winter football.

Blue group hint: Like Hemsworth, but in hoops.

Purple group hint: Cinderellas.

Answers for today’s Connections: Sports Edition groups

Yellow group: A Los Angeles athlete.

Advertisement

Green group: College football bowl games.

Blue group: Basketball Chrises.

Purple group: Men’s NCAA tournament 16-seeds.

Read more: Wordle Cheat Sheet: Here Are the Most Popular Letters Used in English Words

Advertisement

What are today’s Connections: Sports Edition answers?

completed NYT Connections: Sports Edition puzzle for April 6, 2026.

The completed NYT Connections: Sports Edition puzzle for April 6, 2026.

NYT/Screenshot by CNET

The yellow words in today’s Connections

The theme is a Los Angeles athlete. The four answers are Clipper, King, Ram and Spark.

The green words in today’s Connections

The theme is college football bowl games. The four answers are Fiesta, Orange, Rose and Sugar.

Advertisement

The blue words in today’s Connections

The theme is basketball Chrises. The four answers are Bosh, Mullin, Paul and Webber.

The purple words in today’s Connections

The theme is men’s NCAA tournament 16-seeds. The four answers are Howard, Long Island, Prairie View A&M and Siena.

Source link

Advertisement
Continue Reading

Tech

Samsung’s next big audio bet might skip your ears entirely

Published

on

Samsung could be preparing to shake up its audio lineup with a radically different kind of earbuds – ones that don’t even rely on your ear canal. According to recent leaks, the company is working on a new product, possibly called “Galaxy Buds Able,” and early signs suggest these could use bone conduction technology instead of traditional speaker drivers.

Multiple leaks and certifications, including a recent appearance on India’s BIS database, indicate that the product is actively in development. While details remain limited, the unusual model numbering and repeated references across sources hint that this isn’t just another incremental Galaxy Buds refresh, but potentially an entirely new category.

Bone conduction audio works very differently from conventional earbuds

Instead of pushing sound waves through your ear canal, it sends vibrations through your skull directly to the inner ear, effectively bypassing the eardrum. This allows for an open-ear design, meaning users can still hear their surroundings while listening to audio—something traditional in-ear or noise-canceling earbuds often block out.

That shift matters more than it might seem. As wearable tech evolves, companies are increasingly looking at ways to blend digital experiences with real-world awareness. Bone conduction could make earbuds safer for outdoor use, more comfortable for long sessions, and even more accessible for users who struggle with in-ear designs. It also opens doors for new health and assistive applications, especially when combined with Samsung’s growing interest in wellness-focused audio features.

Advertisement

For users, the appeal is straightforward. Imagine listening to music, taking calls, or interacting with voice assistants without isolating yourself from your environment. Whether you’re commuting, working out, or just walking through a busy street, this kind of tech promises a more natural and less intrusive experience.

Looking ahead, timing could be key

Reports suggest Samsung may be positioning these earbuds for a major launch alongside its next-generation foldables, such as the Galaxy Z Fold 8 and Flip 8. If that happens, the “Buds Able” could represent the company’s push into more experimental, next-gen hardware – going beyond iterative upgrades and into entirely new user experiences.

While nothing is official yet, the direction is clear: Samsung isn’t just refining earbuds anymore – it may be redefining how we hear them.

Advertisement

Source link

Continue Reading

Tech

Portal Space’s ‘Mini-Nova’ payload goes into orbit to test technologies for maneuverable space vehicles

Published

on

A SpaceX Falcon 9 rocket sends Portal Space Systems’ “Mini-Nova” technology demonstration payload and more than 100 other payloads into orbit from Vandenberg Space Force Base in California. (SpaceX Photo)

Bothell, Wash.-based Portal Space Systems has made its first foray into Earth orbit, in the form of a piggyback payload that will test technologies for highly maneuverable space vehicles.

The instrument package, which is about the size of a tissue box, was one of 119 payloads sent into orbit at 4:02 a.m. PT today from Vandenberg Space Force Base in California for SpaceX’s Transporter-16 satellite rideshare mission. Portal’s “Mini-Nova” payload was attached to Momentus’ Vigoride-7 orbital service vehicle for the ride on a SpaceX Falcon 9 rocket.

Minutes after launch, the Falcon 9’s first-stage booster landed autonomously on a drone ship that was stationed in the Pacific. Meanwhile, the second stage proceeded to orbit and deployed Vigoride-7 and other spacecraft.

“I’ve said for a long time that a company only really becomes a space company once it gets to space, and with last night’s launch out of Vandenberg, that’s now true for Portal,” the company’s co-founder and CEO, Jeff Thornburg, said in a LinkedIn post.

Model of Mini-Nova payload, held by Portal Space Systems CEO Jeff Thornburg
The Mini-Nova technology demonstration payload is about the size of a tissue box. (GeekWire Photo / Alan Boyle)

“We know that Mini-Nova is healthy, but it will be a few days before we get to download telemetry,” Thornburg told GeekWire in an email.

Mini-Nova will remain attached to Vigoride-7 for its demonstration mission. Over the next six months, Portal will use the payload to test the “brains and critical power systems for our upcoming Starburst and Supernova vehicles,” Thornburg said.

Advertisement

Both of those vehicles will be capable of maneuvering rapidly in orbit to rendezvous with other objects in space for a variety of purposes — including surveillance and space domain awareness, in-space servicing and space-junk disposal. Supernova will make use of an innovative solar thermal propulsion system that could cut the time required for orbital maneuvers from weeks to hours.

Thornburg said the first Starburst vehicle is due for launch as early as October on SpaceX’s Transporter-18 mission. The first Supernova vehicle is expected to be ready for flight in 2027.

Portal was founded in 2021 and has received millions of dollars in support from the U.S. Space Force and the Department of Defense. Last year, the startup raised $17.5 million in an oversubscribed seed funding round.

Source link

Advertisement
Continue Reading

Tech

Today’s NYT Mini Crossword Answers for April 6

Published

on

Looking for the most recent Mini Crossword answer? Click here for today’s Mini Crossword hints, as well as our daily answers and hints for The New York Times Wordle, Strands, Connections and Connections: Sports Edition puzzles.


Need some help with today’s Mini Crossword? I must say, 6-Across really stumped me, but I get it now. Read on for all the answers. And if you could use some hints and guidance for daily solving, check out our Mini Crossword tips.

If you’re looking for today’s Wordle, Connections, Connections: Sports Edition and Strands answers, you can visit CNET’s NYT puzzle hints page.

Advertisement

Read more: Tips and Tricks for Solving The New York Times Mini Crossword

Let’s get to those Mini Crossword clues and answers.

completed-nyt-mini-crossword-puzzle-for-april-6-2026.png

The completed NYT MIni Crossword puzzle for April 6, 2026.

Advertisement

NYT/Screenshot by CNET

Mini across clues and answers

1A clue: Transfusion cocktail = ___, ginger ale, grape juice and lime
Answer: VODKA

6A clue: Body guard?
Answer: APRON

7A clue: Temporary Instagram update
Answer: STORY

Advertisement

8A clue: Big name in hiking sandals
Answer: TEVA

9A clue: TV room
Answer: DEN

Mini down clues and answers

1D clue: Reaching far and wide
Answer: VAST

2D clue: Chose
Answer: OPTED

Advertisement

3D clue: Went by car
Answer: DROVE

4D clue: Book in a mosque, using a non-standard spelling
Answer: KORAN

5D clue: “Got ___ bright ideas?”
Answer: ANY

Advertisement

Source link

Continue Reading

Tech

Light up your life with the Philips Hue Omniglow, the best Hue lightstrip yet

Published

on

Why you can trust TechRadar


We spend hours testing every product or service we review, so you can be sure you’re buying the best. Find out more about how we test.

Philips Hue Omniglow: one-minute review

Hue Omniglow lightstrip held in a hand

(Image credit: Future)

Specifications

Length: 3m (also 5m and 10m in some markets)

Advertisement

Brightness: up to 2,700 lumens at 6,500K (3m)

Colors: white, warm white, and multicolor

The Philips Hue Omniglow is the best Hue lightstrip yet. It’s a classier kind of LED strip: where other models have visible LEDs, the Omniglow delivers seamless color gradients and smoothly moving light effects. The results are very impressive and the Hue app makes it easy to select, edit or create scenes either solo or as part of a wider Hue setup. If you’ve already got a Hue system you can add it in seconds and then include it in your scenes and automations. As with other Hue lights you’ll need a Philips Hue Bridge or Bridge Pro to access advanced features such as custom scenes and smart home integration.

The Omniglow is easy to install and set up, although if you’re mounting it up high you might curse the short power cable. The only real downside is the length: you can shorten the Omniglow but not extend it, and longer versions are not widely available in the UK or US. While European customers can choose between 3m, 5m and 10m models, the US and UK are currently limited to the 3m model only.

Advertisement

Source link

Continue Reading

Tech

iCloud email goes down for some users in an Easter Sunday outage

Published

on

Apple users encountered issues accessing iCloud, in what was a rare Sunday outage for the company’s email, cloud storage, and associated services.

Green circle, yellow rotated square, and red triangle arranged horizontally on a dark gray background
Apple service outage icons

Users of iCloud, Apple’s online services, ware reporting issues in being able to access files. Sites including DownDetector and StatusGator showed a sudden surge of reports from thousands of users, encountering problems since 10 A.M. Eastern.
The reported issues, for the most part, raised iCloud as being the problem. The range of issues was wide, including claims of iCloud Mail being unavailable, Find My devices disappearing in the app, and an inability to access files stored on the service.
Continue Reading on AppleInsider | Discuss on our Forums

Source link

Continue Reading

Tech

LinkedIn secretly scans 6,000+ browser extensions and fingerprints your device

Published

on

In short: Every time you visit LinkedIn in a Chrome-based browser, a hidden JavaScript routine silently probes your browser for more than 6,000 installed extensions, collects 48 hardware and software characteristics about your device, encrypts the resulting fingerprint, and attaches it to every API request you make during your session. The practice, labelled “BrowserGate” by researchers, is not disclosed in LinkedIn’s privacy policy. LinkedIn says it is a security measure; critics say it is covert surveillance of a billion users’ browsing behaviour at industrial scale.

There is a routine that runs on your computer every time you open LinkedIn. You cannot see it, you were not told about it, and it is not described in the company’s privacy policy. According to an investigation published in early April 2026 by Fairlinked e.V., a European association of commercial LinkedIn users, the platform injects a 2.7-megabyte JavaScript bundle into its website that silently scans visitors’ browsers for the presence of more than 6,000 specific Chrome extensions, assembles a detailed fingerprint of their hardware, encrypts it, and transmits the result to LinkedIn’s servers, where it is attached to every subsequent action taken during the session.

The investigation, independently confirmed by BleepingComputer, which verified the scanning behaviour through its own testing, has been dubbed “BrowserGate.” LinkedIn disputes many of the report’s characterisations. The technical facts are not in dispute.

What the script does

LinkedIn calls its scanning system “Spectroscopy.” When a user loads the LinkedIn website, the script fires off up to 6,222 simultaneous requests, each one probing for a specific browser extension by attempting to access files associated with that extension’s ID. The presence or absence of a file in the response indicates whether the extension is installed. The entire operation runs silently in the background, without a visible prompt or notification of any kind.

Advertisement

The 💜 of EU tech

The latest rumblings from the EU tech scene, a story from our wise ol’ founder Boris, and some questionable AI art. It’s free, every week, in your inbox. Sign up now!

Beyond extensions, the script collects 48 distinct characteristics of the user’s device: CPU core count, available memory, screen resolution, timezone, language settings, battery status, audio hardware information, and storage capacity, among others. Individually, these attributes are unremarkable. Combined, they form a device fingerprint specific enough to identify a user even after cookies are cleared.

Once compiled, the data is serialised to JSON and encrypted using an RSA public key, LinkedIn’s internal identifier for the key is “apfcDfPK”,  before being transmitted to telemetry endpoints including li/track and /platform-telemetry/li/apfcDf. The fingerprint is then permanently injected as an HTTP header into every API request made during the session, meaning LinkedIn receives it with every search, every profile view, every message sent.

Advertisement

What it is looking for

The question of which extensions LinkedIn is scanning for makes the surveillance more sensitive than simple fraud detection would require. According to the BrowserGate report, LinkedIn’s list includes more than 200 products that compete directly with its own sales tools, among them Apollo, Lusha, and ZoomInfo. Because LinkedIn knows the employer of each registered user, systematically scanning for the presence of a competitor’s tool gives the platform visibility into which companies are evaluating or deploying rival products.

The list also reportedly includes tools associated with neurodivergent conditions, religious practice, political interests, and job-hunting activity, categories that, in the European Union, qualify as sensitive personal data subject to heightened protection under the General Data Protection Regulation. Knowing that a user is running a job-search extension, for instance, is a meaningful inference about their employment intentions, drawn without consent.

The scale of the operation has grown substantially over time. LinkedIn began scanning for 38 specific extensions in 2017. By 2024, that number had grown to 461. By February 2026, the list had reached 6,167, a 1,252% increase in two years. BleepingComputer’s testing confirmed the scanning was active as of early April 2026.

LinkedIn’s defence and the source of the report

LinkedIn’s response to BleepingComputer was pointed. “The claims made on the website linked here are plain wrong,” a spokesperson said. “The person behind them is subject to an account restriction for scraping and other violations of LinkedIn’s Terms of Service. To protect the privacy of our members, their data, and to ensure site stability, we do look for extensions that scrape data without members’ consent or otherwise violate LinkedIn’s Terms of Service.” The company added that it does not use the data to “infer sensitive information about members.

Advertisement

The platform’s characterisation of the source matters. Fairlinked e.V. is connected to Teamfluence Signal Systems OÜ, an Estonian company whose managing directors include Steven Morell and Jan Liebling. Teamfluence makes a Chrome extension, also called Teamfluence, that LinkedIn restricted for alleged terms of service violations. The company subsequently filed a preliminary injunction against LinkedIn Ireland Unlimited Company and LinkedIn Germany GmbH at the Regional Court of Munich, alleging violations of the Digital Markets Act, EU competition law, and German data protection rules. In January 2026, the Munich court denied the injunction, finding that LinkedIn’s actions did not constitute unlawful obstruction or discrimination.

The financial dispute between the parties does not change the technical findings, which were verified independently. It does mean the framing of those findings is contested, and readers should weigh both the substance of the claim and its provenance.

The regulatory backdrop

This is not LinkedIn’s first serious encounter with European data protection enforcement. In October 2024, the Irish Data Protection Commission, which regulates LinkedIn in the EU through its Irish subsidiary, fined the company €310 million, approximately $334 million , for processing users’ personal data for targeted advertising without a valid legal basis. The decision found that LinkedIn’s consent mechanisms did not meet GDPR’s requirement that consent be “freely given.” LinkedIn was ordered to bring its data processing into compliance.

The BrowserGate investigation drops into that context. The legal question of whether scanning for 6,000 browser extensions constitutes processing of special-category personal data, and whether users’ lack of awareness of the practice renders any implied consent invalid,  is exactly the kind of question the Irish Data Protection Commission has already shown it is willing to adjoin in court. Europe’s evolving digital regulation framework has been moving steadily toward requiring explicit disclosure of all significant data collection, and a scanning operation of this scale, conducted without any mention in a privacy policy, appears difficult to square with that direction of travel.

Advertisement

LinkedIn is a Microsoft subsidiary, acquired in 2016 for $26.2 billion. Microsoft has been aggressively expanding its AI capabilities in 2026, with LinkedIn’s vast dataset of professional identity and employment history forming a significant part of the data infrastructure on which those capabilities rest. The relationship between LinkedIn’s data collection practices and Microsoft’s broader AI ambitions is not addressed in LinkedIn’s privacy policy either.

What this means for users

LinkedIn has more than one billion registered users. The majority access the platform through Chrome-based browsers, meaning the Spectroscopy scan runs routinely on the devices of a significant fraction of the global professional workforce, collecting a fingerprint that is precise enough to persist across cookie resets and potentially across devices.

Short of using a non-Chromium browser such as Firefox, which would limit but not necessarily eliminate LinkedIn’s fingerprinting capabilities, there is no user-facing setting that prevents the scanning. The platform does not offer an opt-out, because it does not disclose the practice in the first place. The 2026 push for governed and transparent AI and data practices is built on precisely the premise that invisible data collection of this kind should not be the default.

Whether regulators move quickly enough to change that default at LinkedIn’s scale remains to be seen. Security firms increasingly built to detect exactly this kind of covert data harvesting are becoming a growth sector in their own right, a market indicator that the gap between what platforms collect and what users understand is still very wide. The year 2025 normalised AI-powered data collection at a pace that regulation has yet to match. BrowserGate is a case study in what that lag looks like from the inside of a browser.

Advertisement

Source link

Continue Reading

Tech

Vibe coding drove an 84% jump in App Store submissions. Apple is cracking down.

Published

on

In short: AI-powered “vibe coding” tools have driven an 84% jump in new app submissions to Apple’s App Store in a single quarter, according to reporting by The Information, the largest surge in a decade. The flood is straining Apple’s review infrastructure, with approval times ballooning from 24 hours to as many as 30 days. Apple has responded by pulling apps that violate its self-containment rules, triggering a standoff with the platforms fuelling the boom.

Apple’s App Store is receiving more new apps than at any point in the past ten years. The cause is not a wave of professional developers: it is a term that Collins English Dictionary named its word of the year for 2025, coined by Andrej Karpathy, a co-founder of OpenAI and former AI lead at Tesla, in a single social media post in February of that year. Vibe coding ,the practice of building software by describing what you want in plain language and letting a large language model write the code, has lowered the barrier to app development so dramatically that it is now overwhelming the infrastructure Apple built to gatekeep its platform.

According to reporting by The Information, the number of new apps submitted to the App Store rose 84% in a single quarter as vibe coding went mainstream. The figure corroborates broader data from Sensor Tower, which tracked a 56% year-on-year spike in iOS app launches in December 2025 and a 54.8% rise in January 2026, the highest growth rates in four years. Apple’s full-year 2025 total reached 557,000 new app submissions, the largest annual wave since 2016.

The tools behind the flood

The surge is attributable to a small cluster of platforms that have turned natural language into deployable software. Cursor, made by Anysphere and used by seven million developers, surpassed $2 billion in annualised revenue in March 2026 and was valued at $29.3 billion after a $2.3 billion funding round co-led by Accel and Coatue in November 2025. Lovable, which targets non-technical builders, reached $200 million in annualised revenue in late 2025, a fiftyfold increase in a single year, and raised $330 million in a Series B at a $6.6 billion valuation in December 2025. Replit generated $240 million in revenue during 2025, serves more than 150,000 paying customers, and is targeting $1 billion in revenue for 2026. Bolt.new has become a popular entry point for rapid idea-to-prototype work.

Advertisement

The commercial argument for these platforms is straightforward: anyone with an idea and an internet connection can now build and submit an app. The problem for Apple is that the same dynamic that makes vibe coding commercially compelling is structurally incompatible with how the App Store review process works.

Advertisement

Why Apple has a structural problem

Vibe coding’s power lies in generating and executing new code on demand,  in response to user prompts, in real time, without a fixed codebase. Apple’s App Store review process was designed for a different model: a developer submits a static build, Apple reviews it, and the approved build is what users receive. Guideline 2.5.2 of Apple’s App Review Guidelines states explicitly that apps “may not download, install, or execute code which introduces or changes features or functionality of the app.” Vibe coding apps, almost by definition, do exactly that.

The volume consequences are already visible in Apple’s infrastructure. Developers submitting to the App Store in March 2026 reported review delays of seven to 30 or more days, against a historical baseline of 24 to 48 hours,  with the majority of delay time spent in the “Waiting for Review” queue before a reviewer picks up the submission. The flood of AI-generated apps is straining a system designed for a world in which building an app took months, not minutes.

The crackdown begins

Apple’s enforcement response has been progressive and, at times, opaque. In mid-March 2026, reports emerged that Apple had quietly blocked updates for a set of vibe coding apps, including Replit and Vibecode,  without public explanation. Developers described receiving rejections citing Guideline 2.5.2 but receiving no advance warning that enforcement was intensifying.

The most prominent casualty was Anything, an app that let users build small tools and automations through natural language prompts. Its co-founder, Dhruv Amin, said Apple had been preventing updates since December 2025 before pulling the app entirely on 30 March 2026. Amin attempted to reach a compromise by modifying the app so that vibe-coded outputs would be previewed in a web browser rather than executed inside the app itself; Apple blocked that update and removed the app regardless.

Advertisement

An Apple spokesperson told The Information that the company was not targeting vibe coding as a category but rather enforcing guidelines that prevent apps from changing their behaviour after review. The distinction, in practice, is narrow: the defining capability of a vibe coding app is its ability to generate and run new functionality on demand, which is precisely what Guideline 2.5.2 prohibits.

The counterargument

Critics of Apple’s position have been pointed. A CNBC column published at the end of March 2026 argued that Apple’s crackdown “puts it on the wrong side of history,” contending that the review-based model was conceived for a world that no longer exists and that blocking vibe coding apps disadvantages the platform against Android, which applies fewer constraints on dynamic code execution.

The deeper tension is one of gatekeeping economics. Apple’s App Store review process is not only a safety mechanism: it is the basis of the 15–30% commission the company collects on in-app purchases and subscriptions. A wave of vibe-coded apps that bypass review , by generating code outside the approved bundle, is also, in structural terms, a challenge to the business logic of the store itself. Regulators in Europe have been scrutinising Apple’s App Store gatekeeping under the Digital Markets Act, and the vibe coding dispute adds another dimension to that ongoing examination.

A platform reckoning

What vibe coding has exposed is a mismatch between the speed at which AI can generate software and the speed at which existing review infrastructure can evaluate it. Apple reviewed roughly 200,000 weekly app submissions at the height of its 2025 volume, and the surge has outpaced that capacity. The platform now faces a choice between expanding its review capacity significantly, updating its guidelines to accommodate dynamic code execution in controlled ways, or continuing to enforce existing rules and accepting the friction that creates with a rapidly growing class of developers.

Advertisement

The capital being deployed into AI infrastructure in 2026 makes it unlikely that the volume of vibe-coded apps will slow on its own. The tools are becoming faster and cheaper; the category is producing some of the highest-growth companies in the technology industry. As AI moves from novelty to commercial infrastructure, the question of who controls the distribution layer,  and on what terms, is becoming the central battleground of the platform era. Apple built the App Store as an answer to that question. Vibe coding is making it ask the question again from the beginning. The AI acceleration of 2025 has arrived at the gate. Apple is deciding whether to open it.

Source link

Advertisement
Continue Reading

Trending

Copyright © 2025