Trillium offers 4x training boost, 3x inference improvement over TPU v5e
Enhanced HBM and ICI bandwidth for LLM support
Scales up to 256 chips per pod, ideal for extensive AI tasks
Google Cloud has unleashed its latest TPU, Trillium, the sixth-generation model in its custom AI chip lineup, designed to power advanced AI workloads.
First announced back in May 2024, Trillium is engineered to handle large-scale training, tuning, and inferencing with improved performance and cost efficiency.
The release forms part of Google Cloud’s AI Hypercomputer infrastructure, which integrates TPUs, GPUs, and CPUs alongside open software to meet the increasing demands of generative AI.
A3 Ultra VMs arriving soon
Trillium promises significant improvements over its predecessor, TPU v5e, with over a 4x boost in training performance and up to a 3x increase in inference throughput. Trillium delivers twice the HBM capacity and doubled Interchip Interconnect (ICI) bandwidth, making it particularly suited to large language models like Gemma 2 and Llama, as well as compute-heavy inference applications, including diffusion models such as Stable Diffusion XL.
Google is keen to stress Trillium’s focus on energy efficiency as well, with a claimed 67% increase compared to previous generations.
Advertisement
Google says its new TPU has demonstrated substantially improved performance in benchmark testing, delivering a 4x increase in training speeds for models such as Gemma 2-27b and Llama2-70B. For inference tasks, Trillium achieved 3x greater throughput than TPU v5e, particularly excelling in models that demand extensive computational resources.
Scaling is another strength of Trillium, according to Google. The TPU can link up to 256 chips in a single, high-bandwidth pod, expandable to thousands of chips within Google’s Jupiter data center network, providing near-linear scaling for extensive AI training tasks. With Multislice software, Trillium maintains consistent performance across hundreds of pods.
Sign up to the TechRadar Pro newsletter to get all the top news, opinion, features and guidance your business needs to succeed!
Tied in with the arrival of Trillium, Google also announced the A3 Ultra VMs featuring Nvidia H200 Tensor Core GPUs. Scheduled for preview this month they will offer Google Cloud customers a high-performance GPU option within the tech giant’s AI infrastructure.
Advertisement
Trillium TPU, built to power the future of AI – YouTube
Gemini Live is now out for the general public to use, and it’s one of Google’s most advanced AI tools. However, it’s not much more than a friendly voice to answer your general questions. This could change, according to a new report. Gemini Live might talk to you about your files sometime soon.
In case you don’t know what Gemini Live is, here’s a refresher. Google recently released this feature for free Gemini users. It’s a much more conversational version of Gemini. It will give you answers and talk to you in a very human manner. It strives to break the barrier between human beings and AI voices, however unsettling that may be.
To use Gemini Live, make sure that your Gemini app is fully updated. After that, either open the app or summon Gemini using the hot word. You’ll see a little waveform icon on the bottom right of the screen. Once you tap that option, Gemini Live will activate and let you choose your voice.
Gemini Live could eventually talk about your files
You can upload files to Gemini to read through, but you could only interact with them in the main Gemini app. When it comes to Gemini Live, you’re not really able to do much outside of having a conversation. However, thanks to a new APK deep-dive, it looks like Google wants Gemini to look at your files.
Advertisement
Since this news is from an APK deep dive, you’ll want to take it cautiously. Google placed code in the current version of the Google app that hints at new features, but it doesn’t guarantee that it will implement them. There’s always the chance that the company will shelf them before the official launch.
The deep dive, performed by AssembleDebug (in collaboration with Android Authority), shows some strings hidden within the 15.45.33.ve.arm64 beta of the Google app.
<string name=”assistant_zero_state_suggestions_open_live_snippet_highlight”>Open Live</string> <string name=”assistant_zero_state_suggestions_open_live_snippet_simplified”>Talk about attachment</string> <string name=”assistant_zero_state_suggestions_open_live_text”>Open Live with attachment</string>
These strings point to Google letting Gemini Live look through the files you uploaded. With this feature enabled, you’ll be able to have a conversation about the media you uploaded. This could have some major implications.
Advertisement
Imagine uploading a PDF of a textbook and having a 1-on-1 conversation about the media in it with Gemini Live. It’ll be like having your own tutor. That’s only one possibility. Who knows how this could help make Gemini Live more useful?
Cadillac is adding to its fleet of EVs with a new luxury SUV. is a three-row, all-electric SUV that will hit showrooms and dealerships sometime next summer with a starting price of $78,790.
The Vistiq’s dual-motor, all-wheel drive system runs on a 102 kWh battery pack with a range of 300 miles that produces 615 horsepower and 650 pound-feet of torque. The Vistiq also supports vehicle-to-home (V2H) bidirectional charging capabilities: it can charge at home, and also deliver electricity to your house during a power outage. Using the features requires buying the GM Energy V2H bundle though.
The SUV’s design borrows aesthetically from other Cadillac EVs. Like the , it has flush door handles, and features similar looking lights and side panels. It also matches the Lyriq’s 300 mile range. The “swept-back windshield” and “Black Crystal Shield grill” evoke the Escalade IQ.
Of course, the Vistiq’s power and price are different from its Cadillac EV’s. The new Cadillac EV SUV is less expensive than an Escalade IQ ($129,990) but more than a Lyriq ($58,595), and the has a higher peak battery range at 450 miles.
Advertisement
The Vistiq comes with a 23-speaker AKG7 Studio Audio system with Dolby Atmos. The Android-powered infotainment system is baked into a 33-inch high resolution LED display. The Verge also that the new EV’s navigation system uses Google Maps and can run other apps from the Google Play Store.
Apple CarPlay and Android Auto won’t be available in Cadillac’s newest EV. is phasing out Apple CarPlay and Android Auto from its EVs and plans to go with Android Automotive. GM’s Executive Director of Digital Cockpit Experience Edward Kummer said in that the carmaker didn’t want any features in its EVs “that are dependent on a person having a cellphone.”
Covering your surfboard in bright lights sounds like an open invitation to great white sharks, but research released Tuesday by Australian scientists found it might actually stave off attacks.
Biologist Laura Ryan said the predator often attacked its prey from underneath, occasionally mistaking a surfer’s silhouette for the outline of a seal.
Ryan and her fellow researchers showed that seal-shaped boards decked with bright horizontal lights were less likely to be attacked by great white sharks.
Advertisement
This appeared to be because the lights distorted the silhouette on the ocean’s surface, making it appear less appetizing.
“There is this longstanding fear of white sharks and part of that fear is that we don’t understand them that well,” said Ryan, from Australia’s Macquarie University.
Seal-shaped decoys were strung with different configurations of LED lights and towed behind a boat to see which attracted the most attention.
Advertisement
Brighter lights were better at deterring sharks, the research found, while vertical lights were less effective than horizontal.
Macquarie University Professor Nathan Hart, one of the study’s authors, said the lights caused a “complex interaction” with the shark’s behavior.
“It’s like an invisibility cloak but with the exception that we are splitting the object, the visual silhouette, into smaller bits,” Hart said.
The study’s authors released a video showing some of the research in action.
Advertisement
Ryan said the results were better than expected and is now in the process of building prototypes for use on the underside of kayaks and surfboards.
Australia has some of the world’s most comprehensive shark management measures, including monitoring drones, shark nets and a tagging system that alerts authorities when a shark is near a crowded beach.
Ryan said her research could allow less invasive mitigation methods to be used.
Advertisement
More research was needed to see if bull and tiger sharks — which have different predatory behavior — responded to the lights in a similar way, the authors said.
There have been more than 1,200 shark incidents in Australia since 1791, of which 255 resulted in death, official data shows.
Advertisement
Great white sharks were responsible for 94 of those deaths.
The overall number of deadly shark attacks worldwide in 2023 remained relatively low, but it was still twice the previous year’s total, according to the latest iteration of the International Shark Attack File — a database of global shark attacks run by the University of Florida.
The report noted that a “disproportionate” amount of people died from shark bites in Australia last year compared with other countries around the world.
Join our daily and weekly newsletters for the latest updates and exclusive content on industry-leading AI coverage. Learn More
Today, data ecosystem giant Snowflake kicked off its BUILD developer conference with the announcement of a special new offering: Snowflake Intelligence.
Set to launch in private preview soon, Snowflake Intelligence is a platform that will help enterprise users set up and deploy dedicated ‘data agents’ to extract relevant business insights from their data, hosted within their data cloud instance and beyond, and then use those insights to take actions across different tools and applications, like Google Workspace and Salesforce.
The move comes as the rise of AI agents continues to be a prominent theme in the enterprise technology landscape, with both nimble startups and large-scale enterprises (like Salesforce) adopting them. It will further strengthen Snowflake’s position in the data domain, leaving the ball in rival Databricks’ court to come back with something bigger.
Advertisement
However, it is important to note that Snowflake isn’t the very first company to toy with the idea of AI agents for improved data operations.
Other startups including Redbird, Altimate AI and Connecty AI, are also exploring with the idea of agents to help users better manage and extract value (in the form of AI and analytical applications) from their datasets. One key benefit of Snowflake’s is that the agent creation and deployment platform will live within the same cloud data warehouse or lakehouse provider, eliminating the need for another tool.
What to expect from Snowflake’s data agents?
Ever since Neeva AI CEO Sridhar Ramaswamy took over as CEO, Snowflake has been integrating AI capabilities on top of its core data platform to help customers take advantage of all their datasets, without running into technical complexities.
From the Document AI feature launched last year to help teams extract data from their unstructured documents and to fully-managed open LLM solution Cortex AI to Snowflake Copilot, an assistant built with Cortex to write SQL queries in natural language and extract insights from data, Snowflake has been busy adding such AI features.
Advertisement
However, until now, the AI smarts were only limited to working with the data hosted within users’ respective Snowflake instances, not other sources.
How Snowflake Intelligence data agents work
With the launch of Snowflake Intelligence, the company is expanding these capabilities, giving teams the option to set up enterprise-grade data agents that could tap not only business intelligence data stored in their Snowflake instance, but also structured and unstructured data across siloed third-party tools — such as sales transactions in a database, documents in knowledge bases like SharePoint, information in tools like Slack, Salesforce, and Google Workspace.
According to the company, the platform, underpinned by Cortex AI’s capabilities, integrates different data systems with a single governance layer and then uses Cortex Analyst and Cortex Search (part of Cortex AI architecture) to deploy agents that accurately retrieve and process specific data assets from both unstructured and structured data sources to provide relevant insights.
The users interact with the agents in natural language, asking business-related questions covering different subjects, while the agents identify the relevant internal and external data sources, covering data types like PDFs, tables, etc., for those subjects and run analysis and summarization jobs to provide answers.
Advertisement
But that’s not all. Once the relevant data is surfaced, the user can ask the data agents to go a step further and take specific actions around the generated insights.
For instance, a user can ask their data agent to enter the surfaced insights into an editable form and upload the file to their Google Drive. The agent would immediately analyze the query, plan and make required API function calls to connect to the relevant tools and execute the task. It can even be used for writing to Snowflake tables and making data modifications.
We’ve reached out to Snowflake with specific questions about these data agents, including the breadth of data sources they can cover and tasks they can (or cannot) execute, but have not heard from the company at the time of writing.
It also remains to be seen how quickly and easily users can create and set up these data agents. For now, the company has only said it only takes a “few steps” to deploy them.
Baris Gultekin, the head of AI at Snowflake says the unified platform “represents the next step in Snowflake’s AI journey, further enabling teams to easily, and safely, advance their businesses with data-driven insights they can act on to deliver measurable impact.”
Advertisement
No word on widespread availability
While the idea of having agents that could answer questions about business data and then take specific actions with the generated insights to do organizational work sounds very tempting, it is pertinent to note that the capability has just been announced yet.
Snowflake has not given a timeline on its availability. It only says that the unified platform will go into private preview very soon.
However, the competition is intensifying fast, including from AI model provider startups such as Anthropic with its new Computer Use mode, giving users more options to choose from when it comes to turning autonomous agents loose on business data, and completing tasks from a user’s text prompt instructions.
The company also notes that Snowflake Intelligence will be natively integrated with the company’s Horizon Catalog at the foundation level, allowing users to run agents for insights right where they discover, manage and govern their data assets. It will be compatible with both Apache Iceberg and Polaris, the company added.
A jet of liquid can bounce off of a hot plate without ever touching it. This extension of the Leidenfrost effect – the phenomenon that allows beads of water to skitter across a scorching pan – could help improve cooling processes, from nuclear reactors to firefighting.
Though first described nearly 300 years ago, the Leidenfrost effect has only been tested with fluid droplets, not squirts of liquid. Until now.
The media industry today may not have a very favorable view of AI — a technology that’s already been used to replace reporters with AI-written copy, while other AI companies have scooped up journalists’ work to feed their chatbots’ data demands, but without returning traffic to the publisher as search engines once did. However, one startup, an AI newsreader called Particle from former Twitter engineers, believes that AI could serve a valuable role in the media industry by helping consumers make sense of the news and dig deeper into stories, while still finding a way to support the publishers’ businesses.
Backed by $4.4 million in seed funding and a $10.9 million Series A led by Lightspeed, Particle was founded last year by the former senior director of Product Management at Twitter, Sara Beykpour, who worked on products like Twitter Blue, Twitter Video, and conversations, and who spearheaded the experimental app, twttr. Her co-founder is a former senior engineer at Twitter and Tesla, Marcel Molina.
From the consumers’ perspective, the core idea behind Particle is to help readers better understand the news with the help of AI technology. More than just summarizing stories into key bullet points for quick catch-ups, Particle offers a variety of clever features that let you approach the news in different ways.
But instead of simply sucking up publishers’ work for its own use, Particle aims to compensate publishers or even drive traffic back to news sites by prominently showcasing and linking to sources directly underneath its AI summaries.
To start, Particle has partnered with specific publishers to host some of their content in the app via their APIs, including outlets like Reuters, AFP, and Fortune. These partners receive better positioning and their links are highlighted in gold above others.
Advertisement
Already, beta tests indicate that readers are clicking through to publishers’ sites because of the app’s design and user interface, though that could shift now that the app is launching beyond news junkies to the general public. In time, the company intends to introduce other ways to work with the media, too, in addition to sending them referral traffic. The team is also having discussions with publishers about providing its users access to paywalled content in a way that makes sense for all parties.
“Having deep partnerships and collaboration is one of the things that we’re really interested in,” notes Beykpour.
To help with its traffic referral efforts, the app’s article section includes big tap targets, making it easy for readers to click through to the publisher’s site. Plus, Particle includes the faces of the journalists on their bylines, and readers can follow through links to publisher profiles to read more of their content or follow them.
Using the app’s built-in AI tools, news consumers can switch between different modes like “Explain Like I’m 5,” to get a simplified version of a complicated story or those that summarize “just the facts,” (or the 5W’s — who, what, when, where, and why). You can have the news summarized in another language besides English, or listen to an audio summary of a story or a personalized selection of stories while on the go. Particle can also pull out important quotes from a story and other links of reference.
But two of the more interesting features involve how Particle leverages AI to help present the news from different angles and allows you to further engage with the story at hand by asking questions.
In Particle, one tool called “Opposite Sides” aims to break users’ filter bubbles by presenting different viewpoints from the same story. This model has been tried before by other news apps, including the startup Brief and SmartNews. Unlike earlier efforts, Particle includes a story spectrum that shows how news is being reported across both “red” and “blue”-leaning sites, with bubbles placed to indicate how far to the left or right the news’ positioning is, and how outsized the coverage may be from one side or the other. The AI will also summarize both sides’ positions, allowing news consumers to reach their own opinions about the matter.
Advertisement
However, the app’s killer feature is an AI chatbot that lets you ask questions and get instant answers about a story. The app will include suggested questions and those asked by others. For example, if you’re reading about Trump’s immigration policy plans, you could ask the chatbot things like “What are the potential legal challenges to Trump’s deportation plans?” or “What are the potential costs of mass deportation?” among other things. Particle will then use its AI technology to find those answers and fact-check them for accuracy.
“The chat function uses OpenAI as well as…our own pre-processing and post-processing,” explains Beykpour, in an interview with TechCrunch. “It uses the content, searches the web a little bit — if it wants to find extra information on the web — to generate those answers.” She says that after the answer is generated, Particle includes an extra step where the AI has to go find the supporting material that matches those answers.
Overall, the app encompasses tech like OpenAI’s GPT-4o and GPT-4o mini, Anthropic, Cohere, and others, including more traditional AI technologies, which are not LLM-based, from Google.
“We have a processing pipeline that takes related content and summarizes it into bullet points, into a headline, sub-headline, and does all the extractions,” she continues. “Then…we pull out quotes and links and all sorts of relevant information about [the story]. And we have our own algorithms to rank, so that the most important or relevant link is the one that you see first — or what we think is the most important or relevant quote is the one that you see first.”
The company claims that its technology reduces AI accuracy problems that would otherwise occur one out of 100 times, and reduces their likelihood to one out of 10,000 times.
Advertisement
Particle will also use human editors as it grows to help better manage the AI content and curate its homepage, she notes.
You must be logged in to post a comment Login