Connect with us

Technology

Meta unveils AI tools to give robots a human touch in physical world

Published

on

Meta unveils AI tools to give robots a human touch in physical world

Join our daily and weekly newsletters for the latest updates and exclusive content on industry-leading AI coverage. Learn More


Meta made several major announcements for robotics and embodied AI systems this week. This includes releasing benchmarks and artifacts for better understanding and interacting with the physical world. Sparsh, Digit 360 and Digit Plexus, the three research artifacts released by Meta, focus on touch perception, robot dexterity and human-robot interaction. Meta is also releasing PARTNR a new benchmark for evaluating planning and reasoning in human-robot collaboration.

The release comes as advances in foundational models have renewed interest in robotics, and AI companies are gradually expanding their race from the digital realm to the physical world.

There is renewed hope in the industry that with the help of foundation models such as large language models (LLMs) and vision-language models (VLMs), robots can accomplish more complex tasks that require reasoning and planning.

Advertisement

Tactile perception

Sparsh, which was created in collaboration with the University of Washington and Carnegie Mellon University, is a family of encoder models for vision-based tactile sensing. It is meant to provide robots with touch perception capabilities. Touch perception is crucial for robotics tasks, such as determining how much pressure can be applied to a certain object to avoid damaging it. 

The classic approach to incorporating vision-based tactile sensors in robot tasks is to use labeled data to train custom models that can predict useful states. This approach does not generalize across different sensors and tasks.

Meta Sparsh architecture Credit: Meta

Meta describes Sparsh as a general-purpose model that can be applied to different types of vision-based tactile sensors and various tasks. To overcome the challenges faced by previous generations of touch perception models, the researchers trained Sparsh models through self-supervised learning (SSL), which obviates the need for labeled data. The model has been trained on more than 460,000 tactile images, consolidated from different datasets. According to the researchers’ experiments, Sparsh gains an average 95.1% improvement over task- and sensor-specific end-to-end models under a limited labeled data budget. The researchers have created different versions of Sparsh based on various architectures, including Meta’s I-JEPA and DINO models.

Touch sensors

In addition to leveraging existing data, Meta is also releasing hardware to collect rich tactile information from the physical. Digit 360 is an artificial finger-shaped tactile sensor with more than 18 sensing features. The sensor has over 8 million taxels for capturing omnidirectional and granular deformations on the fingertip surface. Digit 360 captures various sensing modalities to provide a richer understanding of the environment and object interactions. 

Digit 360 also has on-device AI models to reduce reliance on cloud-based servers. This enables it to process information locally and respond to touch with minimal latency, similar to the reflex arc in humans and animals.

Advertisement
Meta Digit 360 Credit: Meta

“Beyond advancing robot dexterity, this breakthrough sensor has significant potential applications from medicine and prosthetics to virtual reality and telepresence,” Meta researchers write.

Meta is publicly releasing the code and designs for Digit 360 to stimulate community-driven research and innovation in touch perception. But as in the release of open-source models, it has much to gain from the potential adoption of its hardware and models. The researchers believe that the information captured by Digit 360 can help in the development of more realistic virtual environments, which can be big for Meta’s metaverse projects in the future.

Meta is also releasing Digit Plexus, a hardware-software platform that aims to facilitate the development of robotic applications. Digit Plexus can integrate various fingertip and skin tactile sensors onto a single robot hand, encode the tactile data collected from the sensors, and transmit them to a host computer through a single cable. Meta is releasing the code and design of Digit Plexus to enable researchers to build on the platform and advance robot dexterity research.

Meta will be manufacturing Digit 360 in partnership with tactile sensor manufacturer GelSight Inc. They will also partner with South Korean robotics company Wonik Robotics to develop a fully integrated robotic hand with tactile sensors on the Digit Plexus platform.

Evaluating human-robot collaboration

Meta is also releasing Planning And Reasoning Tasks in humaN-Robot collaboration (PARTNR), a benchmark for evaluating the effectiveness of AI models when collaborating with humans on household tasks. 

Advertisement

PARTNR is built on top of Habitat, Meta’s simulated environment. It includes 100,000 natural language tasks in 60 houses and involves more than 5,800 unique objects. The benchmark is designed to evaluate the performance of LLMs and VLMs in following instructions from humans. 

Meta’s new benchmark joins a growing number of projects that are exploring the use of LLMs and VLMs in robotics and embodied AI settings. In the past year, these models have shown great promise to serve as planning and reasoning modules for robots in complex tasks. Startups such as Figure and Covariant have developed prototypes that use foundation models for planning. At the same time, AI labs are working on creating better foundation models for robotics. An example is Google DeepMind’s RT-X project, which brings together datasets from various robots to train a vision-language-action (VLA) model that generalizes to various robotics morphologies and tasks.


Source link
Continue Reading
Advertisement
Click to comment

You must be logged in to post a comment Login

Leave a Reply

Technology

Hyundai’s cutesy Inster EV doesn’t need to be quick

Published

on

Hyundai’s cutesy Inster EV doesn’t need to be quick

The reviews for Hyundai’s little electric SUV that could are trickling in, and it’s clear that the Inster is a delightful way to move about town — regardless of its lack of quickness compared to other similarly-sized EVs. The Inster’s top speed for the long-range version is about 93 miles per hour (or 150 km/h), and it has a zero to 62 mph (100km/h) acceleration in 10.6 seconds, according to the specs Hyundai published today.

Hyundai also revealed more details about the Inster’s price, with European reviewers saying it’s expensive compared to similar competition at £23,495 (about $25,477). In the US, however, that’s a price we can only dream about since our most affordable options include the $35,000 Chevy Equinox EV or the hope Tesla will deliver a cheaper car for around $25,000.

Hyundai uses the Casper name in Korea only.
Image: Hyundai

One newer compact EV that has made it to the US is the Fiat 500e. At 143 inches in length, it’s only about 7 inches shorter than the Inster at 150.59 inches. However, the Inster has more internal storage since it’s SUV-shaped, plus it has a range of about 230 miles WTLP on the long range 49 kWh battery compared to the 500e’s, which is under 200 miles.

Advertisement

Source link

Continue Reading

Technology

NYT Connections today — hints and answers for Saturday, November 2 (game #510)

Published

on

NYT Connections homescreen on a phone, on a purple background

Good morning! Let’s play Connections, the NYT’s clever word game that challenges you to group answers in various categories. It can be tough, so read on if you need clues.

What should you do once you’ve finished? Why, play some more word games of course. I’ve also got daily Wordle hints and answers, Strands hints and answers and Quordle hints and answers articles if you need help for those too.

Source link

Continue Reading

Technology

NYT Strands today: hints, spangram and answers for Saturday, November 2

Published

on

NYT Strands today: hints, spangram and answers for Saturday, September 21

Strands is a brand new daily puzzle from the New York Times. A trickier take on the classic word search, you’ll need a keen eye to solve this puzzle.

Like Wordle, Connections, and the Mini Crossword, Strands can be a bit difficult to solve some days. There’s no shame in needing a little help from time to time. If you’re stuck and need to know the answers to today’s Strands puzzle, check out the solved puzzle below.

How to play Strands

You start every Strands puzzle with the goal of finding the “theme words” hidden in the grid of letters. Manipulate letters by dragging or tapping to craft words; double-tap the final letter to confirm. If you find the correct word, the letters will be highlighted blue and will no longer be selectable.

If you find a word that isn’t a theme word, it still helps! For every three non-theme words you find that are at least four letters long, you’ll get a hint — the letters of one of the theme words will be revealed and you’ll just have to unscramble it.

Advertisement

Every single letter on the grid is used to spell out the theme words and there is no overlap. Every letter will be used once, and only once.

Each puzzle contains one “spangram,” a special theme word (or words) that describe the puzzle’s theme and touches two opposite sides of the board. When you find the spangram, it will be highlighted yellow.

The goal should be to complete the puzzle quickly without using too many hints.

Hint for today’s Strands puzzle

Today’s theme is “Good on paper”

Advertisement

Here’s a hint that might help you: typical tools found at work.

Today’s Strand answers

NYT Strands logo.
NYT

Today’s spanagram

We’ll start by giving you the spangram, which might help you figure out the theme and solve the rest of the puzzle on your own:

Today’s Strands answers

  • RULER
  • PENCILS
  • SCISSORS
  • PRINTER
  • STAPLER






Source link

Continue Reading

Technology

Google TV Freeplay free channel app disabled due to crashes

Published

on

Google TV Streamer unboxed before international availability

In the US, Google TV products have access to over 150 free (ad-supported) channels, which is great. Recently, Google improved the experience of enjoying these channels with a new Google TV Freeplay, which includes a useful live guide. However, the Google TV Freeplay app has been disabled due to unexpected technical issues.

Google TV Freeplay app’s new features were causing crashes, so Google disabled it

Google began the rollout of the revamped Freeplay app in early September. The update made it much easier to use than before, where the experience of searching for channels and shows was quite cumbersome. The mere presence of a programming guide totally changed the reality of the app. Sadly, it seems that the guide was causing crashes on some devices.

As spotted by 9to5Google, the Google TV Freeplay app disappeared from all Google TV devices this week. The only product where it is still active and working without problems is the latest Google TV Streamer. The reason for the app’s disappearance was unknown, as not everyone was aware of the crash issues. However, an official statement from Google shed more light on what’s going on.

“While rolling out the new Google TV Freeplay guide, we discovered an issue that can increase crashes for some users. We have disabled access to the new guide on affected devices until the fix is ​​in place. The fix will begin rolling out soon, and the updated guide will be available for all Google TV devices in the coming months,” a company spokesperson said.

Advertisement

Freeplay could be back soon without the live guide

So, given the facts, by “affected devices” Google is referring to all Google TV products except the latest Streamer. While the company claims that the fix will be available “soon,” they also mention that the guide will be available “in the coming months.” So, possibly the Google TV Freeplay app will return in a few days, but without the latest improvements to the user experience. Let’s hope that a fixed version of the app with all the new features won’t take long to arrive.

Source link

Continue Reading

Technology

Instagram reorganizes message requests for creators

Published

on

Instagram reorganizes message requests for creators

A new update for Instagram posted earlier today could fix one of the most frustrating problems for creators. Adam Mosseri, the head of Instagram, announced a new filtering update on Instagram for creators’ inboxes.

Instagram users with a creator-designated account can now filter message requests in their inbox based on its sender in a similar way to Gmail’s labels. Creators can still sort their messages by the most “recent” received and by the “number of followers” but they can now filter out certain messages. The new filters include requests and messages from “verified accounts,” “businesses,” “creators” and “subscribers.”

The update also includes a way to sort all of your story replies on Instagram. If you go to the top of your inbox, you can also sort and filter your story replies “in case you just wanna get to these requests really quickly and easily,” Mosseri says.

“Now there’s a lot more to do to improve the inbox for creators and requests but hopefully this is one step in the right direction,” Mosseri adds in his video. He also said this feature was one a lot of creators were asking for, so hopefully Instagram will be adding more inbox tools in the near future to make that part of the app a bit cleaner.

Advertisement

Instagram has been toying with new ways to update its platform for higher profile users and creators for a long time now. The company started testing its creator account concept in 2018 that allowed celebrities and more famous social media stars to filter their direct messages and track stats of their followers.

Source link

Continue Reading

Technology

AI on your smartphone? Hugging Face’s SmolLM2 brings powerful models to the palm of your hand

Published

on

AI on your smartphone? Hugging Face’s SmolLM2 brings powerful models to the palm of your hand

Join our daily and weekly newsletters for the latest updates and exclusive content on industry-leading AI coverage. Learn More


Hugging Face today has released SmolLM2, a new family of compact language models that achieve impressive performance while requiring far fewer computational resources than their larger counterparts.

The new models, released under the Apache 2.0 license, come in three sizes — 135M, 360M and 1.7B parameters — making them suitable for deployment on smartphones and other edge devices where processing power and memory are limited. Most notably, the 1.7B parameter version outperforms Meta’s Llama 1B model on several key benchmarks.

Performance comparison shows SmolLM2-1B outperforming larger rival models on most cognitive benchmarks, with particularly strong results in science reasoning and commonsense tasks. Credit: Hugging Face

Small models pack a powerful punch in AI performance tests

“SmolLM2 demonstrates significant advances over its predecessor, particularly in instruction following, knowledge, reasoning and mathematics,” according to Hugging Face’s model documentation. The largest variant was trained on 11 trillion tokens using a diverse dataset combination including FineWeb-Edu and specialized mathematics and coding datasets.

This development comes at a crucial time when the AI industry is grappling with the computational demands of running large language models (LLMs). While companies like OpenAI and Anthropic push the boundaries with increasingly massive models, there’s growing recognition of the need for efficient, lightweight AI that can run locally on devices.

Advertisement

The push for bigger AI models has left many potential users behind. Running these models requires expensive cloud computing services, which come with their own problems: slow response times, data privacy risks and high costs that small companies and independent developers simply can’t afford. SmolLM2 offers a different approach by bringing powerful AI capabilities directly to personal devices, pointing toward a future where advanced AI tools are within reach of more users and companies, not just tech giants with massive data centers.

A comparison of AI language models shows SmolLM2’s superior efficiency, achieving higher performance scores with fewer parameters than larger rivals like Llama3.2 and Gemma, where the horizontal axis represents the model size and the vertical axis shows accuracy on benchmark tests. Credit: Hugging Face

Edge computing gets a boost as AI moves to mobile devices

SmolLM2’s performance is particularly noteworthy given its size. On the MT-Bench evaluation, which measures chat capabilities, the 1.7B model achieves a score of 6.13, competitive with much larger models. It also shows strong performance on mathematical reasoning tasks, scoring 48.2 on the GSM8K benchmark. These results challenge the conventional wisdom that bigger models are always better, suggesting that careful architecture design and training data curation may be more important than raw parameter count.

The models support a range of applications including text rewriting, summarization and function calling. Their compact size enables deployment in scenarios where privacy, latency or connectivity constraints make cloud-based AI solutions impractical. This could prove particularly valuable in healthcare, financial services and other industries where data privacy is non-negotiable.

Industry experts see this as part of a broader trend toward more efficient AI models. The ability to run sophisticated language models locally on devices could enable new applications in areas like mobile app development, IoT devices, and enterprise solutions where data privacy is paramount.

The race for efficient AI: Smaller models challenge industry giants

However, these smaller models still have limitations. According to Hugging Face’s documentation, they “primarily understand and generate content in English” and may not always produce factually accurate or logically consistent output.

Advertisement

The release of SmolLM2 suggests that the future of AI may not solely belong to increasingly large models, but rather to more efficient architectures that can deliver strong performance with fewer resources. This could have significant implications for democratizing AI access and reducing the environmental impact of AI deployment.

The models are available immediately through Hugging Face’s model hub, with both base and instruction-tuned versions offered for each size variant.


Source link
Continue Reading

Trending

Copyright © 2024 WordupNews.com