Connect with us
DAPA Banner

Tech

A Kid With a Fake Mustache Tricked an Online Age-Verification Tool

Published

on

Meta is beefing up its age-verification mechanisms with an AI system that analyzes images and videos on Instagram and Facebook for “visual cues,” such as height and bone structure, to identify and delete accounts of users under the age of 13. The company announced the move amid a wave of cases in which hundreds of children have managed to evade social network access restrictions, even through simple tricks such as drawing on a mustache.

The new approach is part of a series of measures Meta adopted as part of an AI-based security strategy designed to correct the limitations of traditional methods, which rely heavily on self-reported age. With this change, the company seeks to reduce the ease with which minors access platforms that, in theory, are restricted to them.

In a press release, Meta explained that it is implementing several tools to identify contextual indicators that allow estimating a person’s age. This process includes the analysis of posts, comments, bios, and descriptions, with special attention to references related to school years or birthday celebrations—elements that can offer clues about the real age of the person who manages the account.

These tools are in addition to automated analysis techniques aimed at detecting physical traits from imagery shared to Meta’s social platforms. These include characteristics such as height and bone structure. Meta is careful to stipulate that this system is not face recognition, as it does not seek to identify specific individuals in images or videos. Instead, the company notes that, “by combining these visual insights with our analysis of text and interactions, we can significantly increase the number of underage accounts we identify and remove.”

Advertisement

If, based on these elements, Meta suspects that an account is managed by a child under 13, it will be suspended. The user will have to revalidate their age using the procedures established by the company to regain access; otherwise, the profile will be permanently deleted.

Meta also announced that it will expand the scope of its technology to detect users between the ages of 13 and 15 and automatically assign them teen accounts. This type of profile incorporates content restrictions and parental controls enabled by default, with the aim of providing a safer environment for this age group.

Meta began implementing age-verification tech in 2024 for Instagram users in the United States, Australia, Canada, and the United Kingdom. Now, the mechanism will be extended to Instagram accounts in Brazil and 27 European Union countries. In addition, these practices will be applied for the first time to Facebook users in the US, with plans to expand to the EU and UK next month.

Looking All Grown-Up

The new measures have been interpreted as a response to a preliminary ruling recently issued by the European Commission, which concluded that the company led by Mark Zuckerberg is in breach of the Digital Services Act for allegedly failing to effectively prevent children under 13 from using its platforms. The EU body found that the company lacks sufficiently effective mechanisms to block such access and that its current systems for identifying and suspending accounts below the age threshold are insufficient.

Advertisement

These criticisms are supported by the results of a survey conducted by the nonprofit Internet Matters. After surveying nearly 1,300 children and their parents in the UK, the study revealed that approximately one-third of children have successfully evaded government-imposed restrictions on access to social networking sites. In some cases, the methods employed are particularly striking.

The report, titled “The Online Safety Act: Are Children Safe Online?” showed that 46 percent of 9- to 16-year-olds believe that circumventing age controls is very easy. In total, however, only 32 percent admitted to breaking the rules.

Source link

Advertisement
Continue Reading
Click to comment

You must be logged in to post a comment Login

Leave a Reply

Tech

YG Acoustics Titan Loudspeakers Are Really $1 Million Per Pair

Published

on

The ultra high-end loudspeaker category has been rather busy in 2026, which is either a sign that the market still has serious money to spend or that nobody in this business knows how to tap the brakes. Wilson Audio has already made a very large statement with the $788,000 per pair Autobiography, while Børresen’s M8 Gold Signature pushes the conversation into seven figure territory. Now YG Acoustics has entered the same rarefied air with Titan, the first model in its new flagship Ultimate Range and arguably the most talked about new loudspeaker at AXPONA 2026.

This is not a speaker aimed at the casually curious. Titan is YG’s attempt to plant a flag at the top of the mountain, where the air is thin, the rooms are large, and your bank account needs to barely flinch when the invoice is opened.

Attack of the Titans?

Five years of research and three years of product development later, YG Acoustics has arrived at Titan, the most ambitious loudspeaker the company has released to date. It is the first model in YG’s new Ultimate Range and is designed to showcase the brand’s latest engineering work while pushing further into the ultra high end loudspeaker category.

Titan is not trying to disappear visually. It stands approximately 7 feet tall, or 84.5 inches, and weighs about 1,000 pounds. That is half a ton before the crates, the room treatment, and the quiet conversation with your financial advisor. Inside the cabinet, YG uses a seven driver symmetrical array supported by a custom designed sub bass driver.

Advertisement
yg-acoustics-titan-loudspeaker-front

YG Acoustics Titan Versions & Pricing

YG Acoustics offers the Titan in three configurations, giving ultra high-end buyers some room to choose between passive operation, active sub bass support, and a fully active system. “Some room” being relative when the starting point is $880,000 per pair.

Titan Passive: $880,000 per pair

The Titan Passive is a fully passive five-way loudspeaker that uses an external crossover with YG Acoustics’ Ultracoherent circuits. Crossover points are specified at 35 Hz, 90 Hz, 360 Hz, and 1.85 kHz.

Titan with Active Sub: $910,000 per pair

The Titan with Active Sub uses a passive four-way system with an external crossover and Ultracoherent circuits at 90 Hz, 360 Hz, and 1.85 kHz. Sub-bass duties are handled by a dedicated external 1,000-watt amplifier with DSP tuned crossover control for the 12.5-inch (32 cm) driver, including room correction support. The active sub-bass section is time aligned to integrate with the passive drivers.

YG Titan sub-Bass Driver and 1KW Amp (Active Sub Option)

Titan Live: $1,000,000 per pair

The Titan Live is the fully integrated active version of YG Acoustics Titan. It includes dedicated amplification and DSP for each channel, covering the high, midrange, low, and sub bass sections. Each tower is partnered with an external amplifier using 8 x 700-watts with an optimized DSP crossover.

The system connects to the Live Controller using glass fiber optic cables, with support for Roon Ready streaming, analog and digital inputs, and a high quality phono stage. In other words, this is the version for buyers who want the full YG ecosystem, not just a pair of loudspeakers and a nervous conversation with their amplifier dealer.

Advertisement

The Drivers

High frequencies are handled by YG’s lattice tweeter set inside a unique oval waveguide flanked by proprietary YG aluminum cone drivers arranged symmetrically above and below it: a pair of 15 cm (6-inch) mid-range drivers, a pair of 18 cm (7.25-inch) mid-bass drivers, and a pair of 26 cm (10.25-inch) bass drivers.

yg-acoustics-titan-loudspeaker-top-half

The tweeter, midrange, mid-bass, and bass array is supported by a custom 32 cm (12.5 inch) YG aluminum cone sub bass driver with an ultra high field magnet structure. YG says each driver is phase aligned across a wide frequency range to support a point source presentation.

Advertisement. Scroll to continue reading.

The goal is improved clarity, more precise imaging, and a wider listening area. That matters with a loudspeaker this large, because nobody spending this kind of money wants to sit with their head locked in one exact spot like they are being scanned for replicant behavior.

More details are included in the specifications chart later in the article.

Advertisement

Driver array includes:

  • One 26 mm (1”) Tweeter: Proprietary YG Lattice hybrid construction
  • Dual 15 cm (6 inch) midrange drivers: Proprietary YG aluminum cone drivers with neodymium magnets and close pair matching.
  • Dual 18.5 cm (7.25 inch) midbass drivers: Proprietary YG aluminum cone drivers with neodymium magnets and close pair matching.
  • Dual 26 cm (10.25 inch) bass drivers: Proprietary YG aluminum cone drivers with ultra high field magnets and close pair matching.
  • 32 cm (12.5 inch) sub bass driver: Proprietary YG aluminum cone driver with an ultra high field magnet and close matching.

Cabinet Construction

The Titan cabinet uses a five layer construction, another example of YG Acoustics’ focus on structural rigidity, resonance control, and fit and finish.

yg-acoustics-titan-loudspeaker-angle

The side panels use three layers of aerospace grade aluminum alloy, with precision engineered damping materials between them. That creates the five layer structure and gives the cabinet additional stiffness without relying on brute mass alone.

The front faceplates are machined from solid aerospace grade aluminum measuring 75 mm, or 3 inches, thick. YG says the aluminum is heat treated to optimize its crystalline structure before being milled to extremely tight tolerances.

Inside the cabinet, Titan uses advanced bracing, composite resin damping, and computationally optimized diffuser and absorber structures to reduce internal resonances and help the drivers operate with fewer cabinet related distortions.

In addition, the subwoofer section at the bottom maximizes cabinet volume by tapping into a channel that runs up the entire back of the Titan.

The Passive Crossover

The passive crossover is the result of the art and science of speaker engineering. The crossover is built on bespoke multi-layer PCB material with advanced dielectric properties and internal resonance damping.

Advertisement
yg-acoustics-titan-external-crossover

Each crossover circuit is precisely milled in-house from extra-thick, high-purity copper, with traces optimized through simulation to eliminate interference and distortion. The crossover is housed externally in an enclosure crafted entirely from a specially selected polymer material. The external placement eliminates any field interactions with the crossover signal.

YG Acoustics Titan Specifications

YG Acoustics Model  Titan
Speaker Design 5-Way (Available in passive, Active Sub, and Live versions)
Price per pair $880,000 (passive)
$910,000 (passive with active sub)
$1,000,000 (fully active)
Cabinet  Construction Side Panels: 3 aerospace aluminum layers with two damping layers in between for a total of five.

Front and Back: 3″ thick monolithic slabs of aerospace aluminum with specially engineered damping chambers

Speaker Type Floorstanding loudspeaker
Tweeter One 26 mm (1”) proprietary YG Lattice hybrid tweeter 
Midrange Dual 15 cm (6”) proprietary YG aluminum cone midrange drivers, neodymium magnets, and exceptional matching 
Advertisement

Dual 18.5 cm (7.25”) proprietary YG aluminum cone mid bass drivers, neodymium magnets, and exceptional matching 

Woofer Dual 26 cm (10.25”) proprietary YG aluminum cone bass drivers, ultra-high field magnets, and exceptional matching 

32 cm (12.5”) proprietary YG aluminum cone sub-bass driver, ultra-high field magnets, exceptional matching 

Crossover Passive: External crossover with Ultracoherent circuits at 35 Hz, 90 Hz, 360 Hz, and 1.85 kHz
Frequency Response Usable output extends from below 20 Hz to 40 kHz
Sensitivity 88 dB
Impedance 4 Ohms average
2.2 Ohms minimum
Dimensions (HxWxD) 215 x 54 x 108 cm
(84.5 x 21.5 x 42.5 inches)
Weight  455 kg (1,000 lbs) per tower unpacked
yg-acoustics-titan-loudspeaker-left
YG Acoustics Titan Loudspeaker with external crossover behind speaker.

The Bottom Line 

The YG Acoustics Titan is not just the first model in the company’s new Ultimate Range. It is YG’s statement that it wants a seat at the same very expensive table as the Wilson Audio Autobiography, Børresen M8 Gold Signature, and Sonus faber Suprema. That table does not come with a kids’ menu.

What makes Titan interesting is not one single trick feature, but the complete system approach: a massive seven driver symmetrical array, dedicated 12.5-inch sub bass driver, five layer cabinet construction, phase aligned driver integration, and three available configurations: passive, active sub bass, and fully active Titan Live. That flexibility is unusual for a loudspeaker at this level, especially when buyers can choose between traditional external amplification, active low frequency support, or a complete YG controlled ecosystem with amplification, DSP, streaming, digital inputs, analog inputs, and a phono stage.

Advertisement

Prospective buyers need to think beyond the loudspeaker price. The passive and active sub versions still require serious amplification, sources, cabling, setup expertise, and a room large enough to let a 7 foot tall, 1,000 pound loudspeaker breathe. With output extending below 20 Hz and up to 40 kHz, a 4 ohm average impedance, and a 2.4 ohm minimum load, Titan is not something you casually drop into a living room next to the sectional and hope for the best. It needs space, power, proper setup, and a system built around it.

Advertisement. Scroll to continue reading.

Titan is aimed at listeners who already understand the cost of entry in the ultra high-end category and want YG’s most ambitious expression of scale, control, and engineering. Everyone else can admire it from a safe distance, preferably outside the freight elevator.

yg-acoustics-titan-loudspeaker-gold
YG Acoustics Titan in special gold finish that could double the price.

Price & Availability

YG Acoustics Titan (aluminum version) is available now through Authorized Dealers in three configurations:

  • Titan Passive – $880,000 per pair
  • Titan w/Active Sub – $910,000 per pair
  • Titan Live – $1,000,000 per pair

A gold version is also available for roughly $2 million per pair, depending on market pricing and thickness of gold plating.

Sound Clip from AXPONA 2026

For more information: yg-acoustics.com

Advertisement

Source link

Continue Reading

Tech

Google’s AI Search Results Will Now Turn To Reddit For ‘Expert Advice’

Published

on

Google is updating AI Overviews and AI Mode to more prominently surface “Expert Advice” from public discussions, social platforms, forums, blogs, and Reddit. Engadget reports: Via a new “Expert Advice” section that can appear in AI responses, Google will display “a preview of perspectives from public online discussions, social media and other firsthand sources.” In the sample screenshot the company provided, quotes from forums, WordPress blogs and Reddit were arranged above links to their respective sources. Google plans to add more context to these links, too, showing “a creator’s name, handle or community name,” so you can judge what you might want to click through and read from a glance.

Google will also start recommending in-depth articles at the end of AI responses for further exploration of a given topic, and link to more sources directly in its generated answers rather than just at the end. If you subscribe to any publications, AI responses will also highlight sources from the subscriptions you link to your Google account.

Source link

Continue Reading

Tech

Inside the 15,500 malicious domains secretly using ad trackers to push AI investment scams across the web

Published

on


  • 15,500 domains were actively used to deliver cloaked AI investment scams
  • Cloaking ensures harmful content is shown only to targeted victims
  • Commercial tracking software allows cybercriminals to scale operations without building infrastructure

Cloaking has shifted from a supporting tactic into a central layer of cybercriminal infrastructure, and commercial tools are now widely embedded in cybercrime operations at scale.

A four-month analysis of malicious activity by Infoblox and Confiant identified roughly 15,500 domains linked to malicious tracker deployments.

Source link

Advertisement
Continue Reading

Tech

Barry Diller trusts Sam Altman. But ‘trust is irrelevant’ as AGI nears, he says.

Published

on

Billionaire media mogul Barry Diller doesn’t think OpenAI CEO Sam Altman is untrustworthy, despite recent reporting to the contrary. Onstage at The Wall Street Journal’s “Future of Everything” conference this week, Diller vouched for the AI exec, who has been accused by some former colleagues and board members of being manipulative and deceptive at times.

Diller, who is friendly with Altman, was responding to a question about whether or not people should put their faith in Altman to ensure that artificial intelligence benefits humanity.

In particular, he was asked about the theoretical form of AI known as artificial general intelligence, or AGI, which could one day outperform humans on any task.

The media exec, a co-founder of Fox Broadcasting and chairman of IAC and Expedia Group, said that while he believes Altman is sincere in his pursuits, that’s not really the area of concern people should be focused on. Rather, it’s the unknown consequences that will result from AI.

Advertisement

“One of the big issues with AI is it goes way beyond trust,” Diller said. “It may be that trust is irrelevant because the things that are happening are a surprise to the people who are making those things happen. And I’ve spent a lot of time with various people who’ve been in the creation mode of AI, and they have a sense of wonder themselves. So…it’s the great unknown. We don’t know. They don’t know,” he explained.

“We have embarked on something that is going to change almost everything. It is not under-reported. Now, whether these huge investments are going to come through — I couldn’t care less. I’m not invested in it, but progress is going to be made,” Diller added.

Still, the media mogul said he believes that most of the people leading the charge are good stewards, saying he believes that Altman is sincere and “a decent person with good values.” (Diller wouldn’t say which of the AI leaders he thinks is insincere, we should note.)

Techcrunch event

Advertisement

San Francisco, CA
|
October 13-15, 2026

“But the issue is not their stewardship. The issue is … it’s dealing truly with the unknown. They don’t know what can happen once you get AGI, and we’re close to it. We’re not there yet, but we’re getting closer and closer, quicker and quicker. And we must think about guardrails,” Diller noted.

Advertisement

Plus, he warned, if humans don’t think about guardrails, then the alternative is that “another force, an AGI force, will do it themselves. And once that happens, once you unleash that, there’s no going back,” Diller said.

When you purchase through links in our articles, we may earn a small commission. This doesn’t affect our editorial independence.

Source link

Advertisement
Continue Reading

Tech

Snap’s $400 Million Deal With Perplexity Is Dead

Published

on





Snap’s $400 million deal with Perplexity to put the AI search engine directly in Snapchat is dead. The two companies “amicably ended the relationship” earlier this year, Snap disclosed in its latest earnings report.

The two companies first announced the partnership last November, with plans to make Perplexity’s AI search technology a prominent part of the Snapchat app. Snap said at the time that it expected to start seeing revenue from the partnership in “early 2026.” The feature began testing in Snapchat, but was never fully rolled out, according to a help page. The deal has been on shaky ground for some time, with the Snapchat maker saying earlier this year that the two sides had “yet to mutually agree on a path to a broader roll out.”

In a statement, a spokesperson for Perplexity said that the planned feature was “not the right fit” for either company. “After working together, Snap and Perplexity determined that the original implementation was not the right fit for each company’s product goals and have resolved the matter amicably on confidential terms,” the spokesperson said. “Perplexity continues to value Snapchat as a platform for reaching key audiences, remains active on the Snap platform, and expects to continue using Snap’s advertising products.”

Advertisement

Elsewhere, Snap has explored other ways of bringing revenue-generating AI features to Snapchat’s chat feature.The company recently rolled out “AI Sponsored Snaps,” which allows brands to surface AI agents in users’ conversations. During a call with analysts, CEO Evan Spiegel said that the feature was proof “that chat can be monetized in a way that’s really native to Snapchat.”

Spiegel also spoke about the company’s upcoming plans to show off its new AR glasses, which will be the first consumer-ready version of Specs. “The way that people are using their computers is changing really dramatically, and I think that that’s going to be evident in the adoption of wearables and the adoption of Specs over time,” he said. “Because people are going to spend less time hunched over their computers or their phones typing away on keyboards, and spend more time supervising agents who are doing that work on their behalf.” The company is expected to share more about Specs next month at the Augmented World Expo (AWE) event in Long Beach, California.

Advertisement



Source link

Continue Reading

Tech

Energizer Releases Coin Lithium Batteries That Won’t Cause Burning If Accidentally Swallowed

Published

on





Coin batteries don’t make headlines all that often, but today shifts that trend with a smart new development. Energizer has released a new Ultimate Child Shield version of its 20mm coin lithium batteries. As the name implies, the tech is aimed at reducing the number of health risks if a child accidentally swallows the battery. These coin batteries don’t cause burning in the esophagus if ingested, and they also contain a dye that will color a child’s mouth blue if it comes in contact with saliva.

According to Energizer’s press release, more than 3,500 coin lithium battery ingestion incidents are reported in the US each year. Traditional coin batteries can cause esophageal burning within 15 minutes of ingestion, and the addition of the dye can help a caregiver notice quickly if an accident has occurred and get help for the child. Ultimate Child Shield tech is available on the company’s size 2032, 2025 and 2016 batteries, which power a variety of devices including watches and AirTags.

Advertisement



Source link

Advertisement
Continue Reading

Tech

Zillow Group leans into AI as revenue climbs 18% in flat housing market

Published

on

Bigstock Photo

Zillow Group says its engineers are shipping 40% more code each, on average, thanks to internal use of AI tools, allowing them to move features faster from concept to launch.

That’s one of the AI claims the company made in its first-quarter shareholder letter Wednesday, which read at times less like a financial recap and more like a tech strategy blueprint.

“We’re embedding AI throughout the real estate experience in ways that make Zillow increasingly indispensable, and we’re innovating with speed and intention,” CEO Jeremy Wacksman said in the letter, as the Seattle-based company reported revenue up 18% to $708 million in a housing market that was essentially flat. 

Other examples of AI implementation from the letter include:

  • Consumer AI search: Zillow has begun rolling out an AI-powered search mode to about 5% of its audience, or millions of users. The company said early signals show deeper conversations and more actionable engagement compared to traditional search.
  • Agent tools: Follow Up Boss, Zillow’s CRM tool for real estate teams, is becoming an “AI-powered workflow engine” for coordination, prioritization, and outreach. Monthly active users are up more than 70% since Zillow acquired the product at the end of 2023
  • Rental leasing: AI Assist, a leasing assistant embedded in multifamily listings, handles lead management, applicant screening, and lease coordination for property managers.

Wacksman also addressed competition from general-purpose AI platforms, saying Zillow’s proprietary data, deep consumer engagement, and end-to-end transaction tools give it advantages that are difficult to replicate. Zillow launched a partnership with ChatGPT last October, feeding its listings, photos, and pricing into OpenAI’s platform and funneling users back to Zillow for tours and financing. 

In addition to its flagship Zillow homes portal, Zillow Group includes real estate brands such as Trulia, StreetEasy, HotPads, Follow Up Boss, ShowingTime, dotloop, and Zillow Home Loans.

Advertisement

The company has cut jobs twice in the past 18 months, including about 200 positions in January, which it attributed to performance rather than AI-driven reductions.

Financial highlights: 

  • Net income rose to $46 million from $8 million a year ago. 
  • Purchase loan origination volume through Zillow Home Loans nearly doubled, rising 96% to $1.5 billion, making it a top-25 purchase lender nationally. 
  • Rentals revenue jumped 42% to $183 million, driven by 57% growth in multifamily revenue. 
  • The company repurchased 13.5 million shares for $626 million during the quarter, consuming nearly half its cash reserves.

Competitive moves: The day before earnings, Zillow announced a partnership with Realtor.com to extend its Zillow Preview pre-market listings across both platforms. Zillow Preview, launched seven weeks ago, now has more than 60 brokerage partners. 

The move is part of a broader industry fight over listing transparency that put Zillow at odds with Compass, which sued over Zillow’s ban on private listings before settling in March.

Litigation costs: The company flagged $11 million in incremental legal expenses in Q1, expected to rise to about $20 million in Q2 as the FTC trial over Zillow’s rental syndication agreement with Redfin approaches. A CoStar copyright case also remain active.

Advertisement

Market reaction: Zillow shares dropped about 6% in after-hours trading, driven less by the Q1 results (which beat analyst estimates) than by Q2 revenue guidance of $750 million to $765 million, with the midpoint slightly below Wall Street expectations. 

The company is projecting revenue growth in the mid teens for the full year and planning for the housing market to remain at the bottom of the cycle.

Correction: References to the status of the Compass lawsuit were corrected after publication.

Source link

Advertisement
Continue Reading

Tech

Elon Musk’s Last-Ditch Effort to Control OpenAI: Recruit Sam Altman to Tesla

Published

on

A few months before Elon Musk left OpenAI’s board of directors in February 2018, he tried to recruit Sam Altman to join a “world-class AI lab” within Tesla. Musk went as far as offering the OpenAI CEO a Tesla board seat, according to emails and testimony presented in federal court on Wednesday during the Musk v. Altman trial. The emails were shown to a jury during the cross examination of Shivon Zilis, a former OpenAI adviser and board member who is also the mother of four of Musk’s children.

Musk’s core claim in this lawsuit is that Altman and OpenAI president Greg Brockman effectively stole a nonprofit, using the $38 million Musk invested to create a private company worth more than $800 billion today. On Wednesday, lawyers for Musk showed video depositions of former OpenAI CTO Mira Murati and former OpenAI board member Helen Toner, to raise concerns over Altman’s alleged history of deceit.

OpenAI’s legal team has responded to Musk’s claims by questioning his true motives, arguing that the Tesla CEO has had “sour grapes” ever since he failed to assume control of OpenAI in 2017. He has since started a rival, for-profit AI lab. OpenAI’s lawyers used Zilis’ cross-examination on Wednesday to bring up evidence about Musk’s alleged plans to subvert OpenAI, and tried to suggest Zilis was privy to those plans. As it pertains to this case, one of Zilis’ most important roles at OpenAI was acting as a conduit between Musk and Altman.

In a text from February 2018 presented as evidence, Zilis—then an OpenAI adviser, as well as a Neuralink and Tesla executive—asked Altman, “Did you think through a B Corp subsidiary of Tesla?”

Advertisement

“There was documentary evidence that, at several points, Mr. Musk had contemplated seeking to join Sam Altman to the board and offered that option,” said OpenAI lawyer William Savitt outside the courthouse on Wednesday. “It was part of Mr. Musk’s effort to corrupt OpenAI and absorb it into Tesla … he was trying to get Altman to abandon the mission and be part of Tesla.”

In an email to Tesla’s VP of communications, Sarah O’Brien, from November 2017, Zilis shared a draft of an FAQ page about an event Tesla was planning to hold at the NeurIPS AI conference. “The purpose of this event is to share that Tesla is building a world leading AI lab(?) which will rival the likes of Google / DeepMind and Facebook AI Research,” the drafted FAQ read. The document continues, “One major issue for Tesla is when people think of Elon and AI, they think of OpenAI.”

Another part of the FAQ labeled “Who?” lists several Tesla executives who were planned to lead the unit, including Musk and Andrej Karpathy, a former OpenAI researcher. Altman’s name is listed next to Musk’s with two question marks beside it.

The FAQ is marked up with notes including that Altman could be a moderator for the NeurIPS event, which “could be a forcing function for Sam to commit to TeslaAI.” Another note reads that Tesla AI’s “strategy had yet to be defined and some of it may be deeply proprietary.”

Advertisement

Zilis testified on Wednesday that Altman never ended up joining Tesla, and the AI lab and the NeurIPS launch event never came to fruition. She also testified that Musk reached out to Karpathy about recruiting him to Tesla. Savitt told reporters that Zilis’ testimony on Karpathy is “directly contrary to what Mr. Musk told the jury just a few days ago.” Earlier in this trial, Musk testified that Karpathy left OpenAI of his own volition.

Source link

Continue Reading

Tech

AI vision agents use 45x more tokens than APIs in benchmark

Published

on

AI

For AI agents, seeing is expensive

Businesses deploying AI agents to automate computer usage may be spending far more money than necessary if those agents try to emulate human visual interaction.

Reflex, an enterprise application platform, recently set out to compare vision agents with API agents.

Advertisement

A vision agent in this context refers to an AI agent that mimics human interaction by relying on image processing and optical character recognition to operate an application. In this instance, that’s Claude Sonnet navigating a web app user interface via browser-use 0.12, a tool for automated web browser operation.

An API agent here refers to Claude Sonnet interacting with a web app via tools and APIs. The agent calls the same handling mechanisms that the UI calls and receives structured data in response, rather than a web page screenshot that must be analyzed.

“Two agents target the same running app: one drives the UI via screenshots and clicks, the other calls the app’s HTTP endpoints directly,” explained Palash Awasthi, head of growth at Reflex, in a blog post. “Same Claude Sonnet, same pinned dataset, same task. The interface is the only variable.”

The following task was presented to each agent: “A customer named Smith has complained about a recent order. Find the Smith with the most orders, accept all their pending reviews, and mark their most-recent ordered order as delivered.”

Advertisement

According to Awasthi, the API agent completed the task in just eight calls. It listed pending customer reviews, accepted them, and marked the order delivered. 

The vision agent, however, found only one of four pending reviews because it failed to scroll the page where it would have seen the three other reviews hidden off-screen.

Analyzing and interpreting a web page visually is fundamentally more challenging for an AI model than interacting with API calls and tools.

When the prompt was revised to help the vision model perform better, the vision agent still took ~17 minutes, significantly longer than the API agent at ~20 seconds. The vision agent also consumed a lot more tokens – ~45x more.

Advertisement

The company made its test available as a benchmark for those interested in trying to reproduce the results.

Awasthi said that the cost difference between the two approaches reflects the architecture – vision agents need to see and seeing is costly – each screenshot demands thousands of input tokens to process.

Anthropic estimates that processing a 1000×1000-pixel image with Claude Sonnet 4.6 uses about 1,334 tokens

The vision agent expended around 500,000 input tokens and around 38,000 output tokens to complete its task. The API agent used around 12,150 input tokens and around 934 output tokens.

Advertisement

For Awasthi, the lesson is that while vision agents may be necessary for interacting with apps you don’t control, inwardly focused agents should target APIs. ®

Source link

Continue Reading

Tech

OpenAI could launch its AI phone sooner than expected

Published

on

OpenAI’s rumoured smartphone might be arriving earlier than anyone expected.

According to analyst Ming-Chi Kuo, the company is now aiming to mass-produce its first “AI agent phone” in the first half of 2027. This is a significant shift from earlier timelines, with earlier estimates pointing closer to 2028.

The accelerated push suggests OpenAI is getting more serious about hardware, and quickly. Kuo says the move is partly tied to a potential IPO. A flagship device could help strengthen its pitch to investors. Additionally, there is growing competition in the emerging “AI phone” space.

Details are still thin, but the early picture is interesting. The device will run on a custom chip based on MediaTek’s Dimensity 9600, potentially built on TSMC’s next-gen process. It could also feature dual AI processors. These processors would handle things like vision and language tasks simultaneously. This is a hint at how OpenAI wants this to function less like a traditional smartphone and more like an always-on assistant.

Advertisement

One of the headline upgrades is the image signal processor. This will improve how the AI “sees” the world through the camera with enhanced HDR and real-time sensing. Combined with fast memory and storage, the device will have tighter security isolation. Therefore, the aim seems to be a device that can process context continuously, rather than waiting for users to open apps.

Advertisement

That’s the bigger shift here. Kuo believes AI agent phones will move away from the usual app-based experience. Instead, they will let users complete tasks through a context-aware interface. This is something OpenAI is uniquely exploring. This is especially true if it controls both hardware and software.

There’s still some overlap with OpenAI’s other hardware plans. The company is also reportedly working on a screen-less AI device with Jony Ive, now expected around 2027. This comes alongside longer-term ideas like smart glasses and earbuds. If all of this lands, OpenAI could quickly find itself competing with the likes of Apple across multiple product categories.

Advertisement

For now, the AI phone remains unofficial. But if this timeline holds, OpenAI’s biggest hardware play might be closer than expected.

Source link

Advertisement
Continue Reading

Trending

Copyright © 2025