Connect with us

Technology

The brightest bling of TechCrunch Disrupt 2024

Published

on

matt-mullenweg-bling-watch

At TechCrunch Disrupt, our team sits front and center, furiously typing away at our laptops to publish real-time news from impressive speakers like NFL quarterback-turned-founder Colin Kaepernick, Perplexity AI founder Aravind Srinivas, and Ashton Kutcher. We are a well-oiled machine. Some writers have been covering Disrupt since the old days, before Meta was Meta, and the show “Silicon Valley” hadn’t given us the free marketing of a lifetime (thank you, HBO). And yet, this year, even the longest-tenured TechCrunch editors were a bit distracted by the stunning jewelry gracing the stage. So, for those of you who were more focused on the actual substance of these conversations with influential tech players, fret not. Behold: our shimmering roundup of the best bling at TechCrunch Disrupt.

Matt Mullenweg, Founder and CEO of Automattic

matt-mullenweg-bling-watch
Image Credits:Kimberly White / Getty Images

Google, if you’re listening, I have a pitch for the next Google Lens commercial. So, imagine, a group of TechCrunch writers are watching their editor-in-chief host a thoughtful, informative conversation with Matt Mullenweg, the WordPress founder who’s been in the news for his battles with WP Engine. But almost as interesting as Mullenweg’s comments on stage is his watch, a $240,000 fashion statement. The watch’s sapphire detail alone takes over 185 hours to make, according to the website of designer Maximilian Büsser. How did we figure out exactly what bling Mullenweg was wearing? That’s right, Google. We used Circle to Search.

Why is this watch, which is inspired by the design of the KittyHawk aircraft, so expensive? The designer’s website states: “Every component and form has a technical purpose; nothing is superfluous and every line and curve is in poetic harmony. Articulated lugs ensure supreme comfort. Highly legible time is a fringe benefit.”

Yes, it’s ridiculous to buy a bracelet for the cost of a house. But hey, we all spend impulsively sometimes — for example, I spent an extra $2 at the grocery store this weekend to buy pre-sliced mangoes. This is definitely the same kind of frivolity.

Mary Barra, Chair & CEO of General Motors

Image Credits:Kimberly White/Getty Images for TechCrunch) / Getty Images
TechCrunch’s Matt Rosoff and GM’s Mary Barra,Image Credits:Kimberly White/Getty Images for TechCrunch / Getty Images

What’s fascinating about Mary Barra’s bling is its utility. She’s got one diamond ring on each hand, yet beneath the sleeve of her high-end houndstooth blazer (is this the one?), we see an Apple Watch peeking out. Barra’s choice of jewelry displays the duality of the role of the modern CEO: you need to act polished and elegant on the exterior, but when no one’s watching, you still need to be efficient and reliable. And, like my own Apple Watch SE and its disappointing battery life, you can’t survive unless you recharge your batteries each night.

Denise Dresser, Chief Executive Officer, Slack, from Salesforce 

Image Credits:Kimberly White / Getty Images
Techcrunch’s Brian Heater and Slack’s Denise Dresser

Slack CEO Denise Dresser gets some style points: her rings aren’t just bling for the sake of bling, like a massive wedding ring that could rival Mullenweg’s aircraft watch in price. Dresser’s bling has a bit of spunk. One ring is a large gold square, which seems like it might have a maroon-ish band, or it could be the reflection of light. We may never know. The other ring… I want to say it looks like an octopus signet, but I can’t tell.

Tony Fadell, Nest founder

Image Credits:Kimberly White / Getty Images

Nest founder Tony Fadell’s interview at Disrupt was explosive — he took shots at Sam Altman, but also, we couldn’t look away from his large ruby ring. Unfortunately, Google Lens is coming up empty here, even though our source image is pretty clear. This ring could be a $8,400 designer ring — it could also be $10 costume jewelry from Amazon. The truth is probably somewhere closer to the middle.

Erin Foster, co-founder of Favorite Daughter

Image Credits:Kimberly White / Getty Images

Erin Foster was already well-known for her various pursuits with her sister Sara, including a hit podcast, clothing line, and venture firm. She and her sister were even co-heads of creative at Bumble. But now, Erin Foster has reached a new level of fame: she’s the creator of the Netflix show “Nobody Wants This,” which has remained on the top of the Netflix charts since its premiere in September.

What does Foster’s TV writing success have to do with her bling? That’s not just a ring: it’s a wedding ring. The central romance of “Nobody Wants This” — a sex podcaster and a rabbi — was inspired in part by her own experience falling in love with a Jewish man and navigating their cultural differences. If there’s a wedding on “Nobody Wants This,” we can only hope that Kristen Bell’s character’s ring is as spectacular.

Advertisement

Jingna Zhang, founder and CEO of Cara

Image Credits:Katelyn Tucker/ Slava Blazer Photography) / Flickr (opens in a new window)

Jingna Zhang isn’t just the founder of a social platform, Cara. She’s an artist. And as an artist, she knows a thing or two about style — after all, her photographs have appeared in global fashion magazines like Vogue, Elle, and Harper’s Bazaar. At Disrupt, she goes for a look that highlights her jewelry: all black, head to toe, save for a big shiny jewel on her left hand.

Brandie Nonnecke, Director, CITRIS Policy Lab

Image Credits:Katelyn Tucker/ Slava Blazer Photography / Flickr (opens in a new window)

Brandie Nonnecke may be an academic, but her bling is on point — it’s giving tenure. Her gold, crinkle-effect earrings could be from anywhere — the style is quite popular, ranging from cheap Shein dupes, to mid-range gold-plated studs, to more expensive looks from a designer that I wish I hadn’t just discovered, because now I wish I had a spare $400 to buy these earrings.

Nonnecke studies and teaches AI policy, and if you’re reading this, Brandie, I have good news: Google Lens has no idea where your ring is from, and with the exception of Mullenweg’s watch and Barra’s jacket, it has proven very unhelpful at writing this post. Maybe our robot overlords aren’t closing in on us as quickly as we imagined.

Source link

Advertisement
Continue Reading
Advertisement
Click to comment

You must be logged in to post a comment Login

Leave a Reply

Technology

NYT Strands today: hints, spangram and answers for Tuesday, November 5

Published

on

NYT Strands today: hints, spangram and answers for Saturday, September 21

Strands is a brand new daily puzzle from the New York Times. A trickier take on the classic word search, you’ll need a keen eye to solve this puzzle.

Like Wordle, Connections, and the Mini Crossword, Strands can be a bit difficult to solve some days. There’s no shame in needing a little help from time to time. If you’re stuck and need to know the answers to today’s Strands puzzle, check out the solved puzzle below.

How to play Strands

You start every Strands puzzle with the goal of finding the “theme words” hidden in the grid of letters. Manipulate letters by dragging or tapping to craft words; double-tap the final letter to confirm. If you find the correct word, the letters will be highlighted blue and will no longer be selectable.

If you find a word that isn’t a theme word, it still helps! For every three non-theme words you find that are at least four letters long, you’ll get a hint — the letters of one of the theme words will be revealed and you’ll just have to unscramble it.

Advertisement

Every single letter on the grid is used to spell out the theme words and there is no overlap. Every letter will be used once, and only once.

Each puzzle contains one “spangram,” a special theme word (or words) that describe the puzzle’s theme and touches two opposite sides of the board. When you find the spangram, it will be highlighted yellow.

The goal should be to complete the puzzle quickly without using too many hints.

Hint for today’s Strands puzzle

Today’s theme is “More than just sports”

Advertisement

Here’s a hint that might help you: clubs you might join.

Today’s Strand answers

NYT Strands logo.
NYT

Today’s spanagram

We’ll start by giving you the spangram, which might help you figure out the theme and solve the rest of the puzzle on your own:

Today’s Strands answers

  • BAND
  • CHOIR
  • ORCHESTRA
  • DRAMA
  • DEBATE
  • YEARBOOK






Source link

Advertisement
Continue Reading

Technology

Meta’s AI adult classifier will detect age falsification attempts

Published

on

Featured image for Meta takes on OpenAI with its AI video generator

This year, social media companies have been in the spotlight of the authorities. Lawsuits have hit big names like Meta and TikTok for their failure to adequately protect underage users. Under all the pressure, some, like Instagram, have been implementing harsh privacy measures on teen accounts. Now, Meta has offered insight into its new AI-powered adult classifier.

For months now, underage accounts (users under 16) on Instagram have received the “teen account” label. Profiles labeled as such have the most restrictive privacy restrictions by default. This should prevent children or teens from directly contacting potential bad actors or predators. Because these restrictions may limit features, some teens may try to bypass them.

Meta offers more details about the AI-powered adult classifier that Instagram will get

One way that minors might try to get around teen account restrictions is to create a new profile with a fake birth date. With that in mind, Meta announced in September that it will launch an AI-powered adult classifier tool to automatically detect such cases. Now Allison Hartnett, Meta’s director of product management for youth and social impact, has revealed more details about how it will work.

According to Hartnett, the tool will analyze multiple parameters to make a decision. These include the accounts a user follows in particular and the type of content they tend to interact with. Meta’s systems will also be on the lookout for potentially suspicious behavior when creating a new Instagram account, for example, using an email associated with an existing profile or even obtaining the device ID. This way, they can make a more accurate decision about who is creating a new profile.

Advertisement

Meta’s AI adult classifier will be able to label accounts suspected of belonging to minors as “teens,” automatically applying all restrictions. Accounts with those restrictions cannot have them removed without prior authorization from a parent. The company promises to provide an appeal tool if it incorrectly labels an account as “teen.” However, there is no date yet for the appeal tool’s availability.

Instagram will ask for valid IDs or AI-powered facial analysis when trying to change age

There may also be cases of teenagers trying to remove restrictions by changing their date of birth. Here, Instagram will ask for a valid government-issued ID. Users will also have the option to upload a selfie video through Yoti’s technology. The latter offers advanced AI-powered recognition services that can even determine a person’s age. Meta has already turned to Yoti to verify the age of users of Facebook’s dating option.

Source link

Advertisement
Continue Reading

Technology

Meta opens its Llama AI models to government agencies for national security

Published

on

Meta opens its Llama AI models to government agencies for national security

Meta is opening up its Llama AI models to government agencies and contractors working on national security, the company said in . The group includes more than a dozen private sector companies that partner with the US government, including Amazon Web Services, Oracle and Microsoft, as well as defense contractors like Palantir and Lockheed Martin.

Mark Zuckerberg hinted at the move last week during Meta’s earnings call, when the company was “working with the public sector to adopt Llama across the US government.” Now, Meta is offering more details about the extent of that work.

Oracle, for example, is “building on Llama to synthesize aircraft maintenance documents so technicians can more quickly and accurately diagnose problems, speeding up repair time and getting critical aircraft back in service.” Amazon Web Services and Microsoft, according to Meta, are “using Llama to support governments by hosting our models on their secure cloud solutions for sensitive data.”

Meta is also providing similar access to Llama to governments and contractors in the UK, Canada, Australia and New Zealand, Bloomberg . In a blog post, Meta’s President of Global Affairs, Nick Clegg, suggested the partnerships will help the US compete with China in the global arms race over artificial intelligence. “We believe it is in both America and the wider democratic world’s interest for American open source models to excel and succeed over models from China and elsewhere,” he wrote. “As an American company, and one that owes its success in no small part to the entrepreneurial spirit and democratic values the United States upholds, Meta wants to play its part to support the safety, security and economic prosperity of America – and of its closest allies too.”

Advertisement

Source link

Continue Reading

Technology

UC San Diego, Tsinghua University researchers just made AI way better at knowing when to ask for help

Published

on

UC San Diego, Tsinghua University researchers just made AI way better at knowing when to ask for help

Join our daily and weekly newsletters for the latest updates and exclusive content on industry-leading AI coverage. Learn More


A team of computer scientists has developed a method that helps artificial intelligence understand when to use tools versus relying on built-in knowledge, mimicking how human experts solve complex problems.

The research from the University of California San Diego and Tsinghua University demonstrates a 28% improvement in accuracy when AI systems learn to balance internal knowledge with external tools — a critical capability for deploying AI in scientific work.

How scientists taught AI to make better decisions

“While integrating LLMs with tools can increase reliability, this approach typically results in over-reliance on tools, diminishing the model’s ability to solve simple problems through basic reasoning,” the researchers write in their paper. “In contrast, human experts first assess problem complexity using domain knowledge before choosing an appropriate solution approach.”

Advertisement

The new method, called “Adapting While Learning,” uses a two-step process to train AI systems. First, the model learns directly from solutions generated using external tools, helping it internalize domain knowledge. Then, it learns to categorize problems as either “easy” or “hard” and decides whether to use tools accordingly.

The two-step process researchers developed to teach AI systems when to use tools versus rely on internal knowledge, mirroring how human experts approach problem-solving. (Credit: UC San Diego / Tsinghua University)

Small AI model outperforms larger systems on complex tasks

What makes this development significant is its efficiency-first approach. Using a language model with just 8 billion parameters — far smaller than industry giants like GPT-4 — the researchers achieved a 28.18% improvement in answer accuracy and a 13.89% increase in tool usage precision across their test datasets. The model demonstrated particular strength in specialized scientific tasks, outperforming larger models in specific domains.

This success challenges a fundamental assumption in AI development: that bigger models necessarily yield better results. Instead, the research suggests that teaching AI when to use tools versus rely on internal knowledge — much like training a junior scientist to know when to trust their calculations versus consult specialized equipment — may be more important than raw computational power.

Examples of how the AI system handles different types of climate science problems: a simple temperature calculation (top) and a complex maritime routing challenge (bottom). (Credit: UC San Diego / Tsinghua University)

The rise of smaller, smarter AI models

This research aligns with a broader industry shift toward more efficient AI models in 2024. Major players including Hugging Face, Nvidia, OpenAI, Meta, Anthropic, and H2O.ai have all released smaller but highly capable models this year.

Hugging Face’s SmolLM2, with versions as small as 135 million parameters, can run directly on smartphones. H2O.ai’s compact document analysis models have outperformed tech giants’ larger systems on specialized tasks. Even OpenAI entered the small model arena with GPT-4o Mini, offering similar capabilities at a fraction of the cost.

This trend toward “AI downsizing” reflects growing recognition that bigger isn’t always better — specialized, efficient models can often match or exceed the performance of their larger counterparts while using far fewer computational resources.

Advertisement

The technical approach involves two distinct learning phases. During training, the model first undergoes what the researchers call “World Knowledge Distillation” (WKD), where it learns from solutions generated using external tools. This helps it build up internal expertise.

The second phase, “Tool Usage Adaptation” (TUA), teaches the system to classify problems based on its own confidence and accuracy in solving them directly. For simpler problems, it maintains the same approach as in WKD. But for more challenging problems, it learns to switch to using external tools.

Business impact: More efficient AI systems for complex scientific work

For enterprises deploying AI systems, this research addresses a fundamental challenge that has long plagued the industry. Current AI systems represent two extremes: they either constantly reach for external tools — driving up computational costs and slowing down simple operations — or dangerously attempt to solve everything internally, leading to potential errors on complex problems that require specialized tools.

This inefficiency isn’t just a technical issue — it’s a significant business problem. Companies implementing AI solutions often find themselves paying premium prices for cloud computing resources to run external tools, even for basic tasks their AI should handle internally. On the flip side, organizations that opt for standalone AI systems risk costly mistakes when these systems attempt complex calculations without proper verification tools.

Advertisement

The researchers’ approach offers a promising middle ground. By teaching AI to make human-like decisions about when to use tools, organizations could potentially reduce their computational costs while maintaining or even improving accuracy. This is particularly valuable in fields like scientific research, financial modeling, or medical diagnosis, where both efficiency and precision are crucial.

Moreover, this development suggests a future where AI systems could be more cost-effective and reliable partners in scientific work, capable of making nuanced decisions about when to leverage external resources — much like a seasoned professional who knows exactly when to consult specialized tools versus rely on their expertise.

The power of knowing when to ask for help

Beyond the immediate technical achievements, this research challenges the bigger-is-better paradigm that has dominated AI development. In demonstrating that a relatively small model can outperform its larger cousins by making smarter decisions about tool use, the team points toward a more sustainable and practical future for AI.

The implications extend far beyond academic research. As AI increasingly enters domains where mistakes carry real consequences – from medical diagnosis to climate modeling – the ability to know when to seek help becomes crucial. This work suggests a future where AI systems won’t just be powerful, but prudent – knowing their limitations just as skilled professionals do.

Advertisement

In essence, the researchers have taught AI something fundamentally human: sometimes the smartest decision is knowing when to ask for help.


Source link
Continue Reading

Technology

GM says it has become the No. 2 seller of EVs in the US

Published

on

GM says it has become the No. 2 seller of EVs in the US

GM is claiming the number two spot in EV sales in the US for the third quarter of this year, selling 32,000 electric vehicles. The automaker produces EVs across multiple brands running on the same platform, like Chevy’s Silverado, Blazer, and Equinox EVs, as well as the GMC Hummer EV and the Cadillac Lyriq.

GM says it has sold a total of 370,000 EVs in North America since 2016, including 300,000 in the US specifically. Tesla is still the undisputed leader, with more than 5 million vehicles sold since 2008.

In an email with The Verge, GM’s executive director of finance and sales communications James Cain wrote that sales have accelerated since the company built a dedicated EV platform (formerly known as Ultium) and began producing battery cells through its joint ventures with LG and Samsung SDI. GM’s third-quarter EV sales beat out rival Ford by about 8,600 units, according to Kelley Blue Book, as reported by The New York Times.

Meanwhile, Ford spokesperson Dan Barbossa claims the Blue Oval remains “America’s No. 2 best-selling EV brand behind Tesla.” In an email with The Verge, Barbossa wrote:

Advertisement

We remain the No. 2 brand. GM is adding every brand EV (Chevy, GMC, Cadillac, etc) they sell and making a different claim.

Still, GM has a ways to go before it achieves the goal of producing 1 million EVs, which it previously projected it would accomplish by 2025. The company later distanced itself from that target when it became clear that production troubles, charging difficulties, and high interest rates would slow down the rate of growth in EV sales in the US.

Ford had a strong early start with solid sales of its all-electric Mustang Mach-E, launched in 2019, and the F-150 Lightning electric truck in 2022. During that timeframe, GM only had the Chevy Bolt, built on an older battery platform. The Hummer EV truck launched in 2020, but overall EV sales were slow amid production troubles.

Ford also hit some snags along the way, including parts shortages. The company has lost billions of dollars in its Model e division, where revenues have not kept up with spending. Ford recently canceled a planned three-row SUV and has paused production of the F-150 Lightning until next year. Ford is placing a lot of its hopes on its skunkworks team in Silicon Valley, developing its next-gen platform for cheaper EVs.

Source link

Advertisement

Continue Reading

Technology

NYT Strands today — hints, answers and spangram for Tuesday, November 5 (game #247)

Published

on

NYT Strands homescreen on a mobile phone screen, on a light blue background

Strands is the NYT’s latest word game after the likes of Wordle, Spelling Bee and Connections – and it’s great fun. It can be difficult, though, so read on for my Strands hints.

Want more word-based fun? Then check out my Wordle today, NYT Connections today and Quordle today pages for hints and answers for those games.

Source link

Continue Reading

Trending

Copyright © 2024 WordupNews.com