Connect with us

Technology

From cost center to competitive edge: The strategic value of custom AI Infrastructure

Published

on

From cost center to competitive edge: The strategic value of custom AI Infrastructure

Join our daily and weekly newsletters for the latest updates and exclusive content on industry-leading AI coverage. Learn More


This article is part of a VB Special Issue called “Fit for Purpose: Tailoring AI Infrastructure.” Catch all the other stories here.

AI is no longer just a buzzword — it’s a business imperative. As enterprises across industries continue to adopt AI, the conversation around AI infrastructure has evolved dramatically. Once viewed as a necessary but costly investment, custom AI infrastructure is now seen as a strategic asset that can provide a critical competitive edge.

Mike Gualtieri, vice president and principal analyst at Forrester, emphasizes the strategic importance of AI infrastructure. “Enterprises must invest in an enterprise AI/ML platform from a vendor that at least keeps pace with, and ideally pushes the envelope of, enterprise AI technology,” Gualtieri said. “The technology must also serve a reimagined enterprise operating in a world of abundant intelligence.” This perspective underscores the shift from viewing AI as a peripheral experiment to recognizing it as a core component of future business strategy.

Advertisement

The infrastructure revolution

The AI revolution has been fueled by breakthroughs in AI models and applications, but those innovations have also created new challenges. Today’s AI workloads, especially around training and inference for large language models (LLMs), require unprecedented levels of computing power. This is where custom AI infrastructure comes into play.

>>Don’t miss our special issue: Fit for Purpose: Tailoring AI Infrastructure.<<

“AI infrastructure is not one-size-fits-all,” says Gualtieri. “There are three key workloads: data preparation, model training and inference.” Each of these tasks has different infrastructure requirements, and getting it wrong can be costly, according to Gualtieri. For example, while data preparation often relies on traditional computing resources, training massive AI models like GPT-4o or LLaMA 3.1 necessitates specialized chips such as Nvidia’s GPUs, Amazon’s Trainium or Google’s TPUs.

Nvidia, in particular, has taken the lead in AI infrastructure, thanks to its GPU dominance. “Nvidia’s success wasn’t planned, but it was well-earned,” Gualtieri explains. “They were in the right place at the right time, and once they saw the potential of GPUs for AI, they doubled down.” However, Gualtieri believes that competition is on the horizon, with companies like Intel and AMD looking to close the gap.

Advertisement

The cost of the cloud

Cloud computing has been a key enabler of AI, but as workloads scale, the costs associated with cloud services have become a point of concern for enterprises. According to Gualtieri, cloud services are ideal for “bursting workloads” — short-term, high-intensity tasks. However, for enterprises running AI models 24/7, the pay-as-you-go cloud model can become prohibitively expensive.

“Some enterprises are realizing they need a hybrid approach,” Gualtieri said. “They might use the cloud for certain tasks but invest in on-premises infrastructure for others. It’s about balancing flexibility and cost-efficiency.”

This sentiment was echoed by Ankur Mehrotra, general manager of Amazon SageMaker at AWS. In a recent interview, Mehrotra noted that AWS customers are increasingly looking for solutions that combine the flexibility of the cloud with the control and cost-efficiency of on-premise infrastructure. “What we’re hearing from our customers is that they want purpose-built capabilities for AI at scale,” Mehrotra explains. “Price performance is critical, and you can’t optimize for it with generic solutions.”

To meet these demands, AWS has been enhancing its SageMaker service, which offers managed AI infrastructure and integration with popular open-source tools like Kubernetes and PyTorch. “We want to give customers the best of both worlds,” says Mehrotra. “They get the flexibility and scalability of Kubernetes, but with the performance and resilience of our managed infrastructure.”

Advertisement

The role of open source

Open-source tools like PyTorch and TensorFlow have become foundational to AI development, and their role in building custom AI infrastructure cannot be overlooked. Mehrotra underscores the importance of supporting these frameworks while providing the underlying infrastructure needed to scale. “Open-source tools are table stakes,” he says. “But if you just give customers the framework without managing the infrastructure, it leads to a lot of undifferentiated heavy lifting.”

AWS’s strategy is to provide a customizable infrastructure that works seamlessly with open-source frameworks while minimizing the operational burden on customers. “We don’t want our customers spending time on managing infrastructure. We want them focused on building models,” says Mehrotra.

Gualtieri agrees, adding that while open-source frameworks are critical, they must be backed by robust infrastructure. “The open-source community has done amazing things for AI, but at the end of the day, you need hardware that can handle the scale and complexity of modern AI workloads,” he says.

The future of AI infrastructure

As enterprises continue to navigate the AI landscape, the demand for scalable, efficient and custom AI infrastructure will only grow. This is especially true as artificial general intelligence (AGI) — or agentic AI — becomes a reality. “AGI will fundamentally change the game,” Gualtieri said. “It’s not just about training models and making predictions anymore. Agentic AI will control entire processes, and that will require a lot more infrastructure.”

Advertisement

Mehrotra also sees the future of AI infrastructure evolving rapidly. “The pace of innovation in AI is staggering,” he says. “We’re seeing the emergence of industry-specific models, like BloombergGPT for financial services. As these niche models become more common, the need for custom infrastructure will grow.”

AWS, Nvidia and other major players are racing to meet this demand by offering more customizable solutions. But as Gualtieri points out, it’s not just about the technology. “It’s also about partnerships,” he says. “Enterprises can’t do this alone. They need to work closely with vendors to ensure their infrastructure is optimized for their specific needs.”

Custom AI infrastructure is no longer just a cost center — it’s a strategic investment that can provide a significant competitive edge. As enterprises scale their AI ambitions, they must carefully consider their infrastructure choices to ensure they are not only meeting today’s demands but also preparing for the future. Whether through cloud, on-premises, or hybrid solutions, the right infrastructure can make all the difference in turning AI from an experiment into a business driver


Source link
Continue Reading
Advertisement
Click to comment

You must be logged in to post a comment Login

Leave a Reply

Technology

Bret Taylor of Sierra joins Disrupt 2024

Published

on

Bret Taylor of Sierra joins TechCrunch Disrupt 2024

We’re thrilled to welcome Bret Taylor to TechCrunch Disrupt 2024. As the former co-CEO of Salesforce, founder of Quip, former CTO of Facebook, the co-creator of Google Maps, and current chairman of the board at OpenAI, Taylor needs very little introduction. 

Bret Taylor is one of the most influential players in the world of enterprise and SaaS, so we invited him to join us for a fireside chat on our SaaS Stage at Disrupt 2024 on October 29. 

In our conversation with Bret, we’ll explore his innovative AI startup, Sierra, recently launched in partnership with former Google executive Clay Bavor. Sierra envisions AI agents as the next major technological breakthrough, much like the rise of websites and mobile applications. These agents could become crucial digital resources for businesses, ultimately enhancing the digital customer experience.

Join us at Disrupt 2024, taking place at Moscone West in San Francisco from October 28-30, as we delve into this topic with Bret, along with a host of other insights. Register your passes today and enjoy savings of up to $600. This offer ends tonight at 11:59 p.m. PT.

Advertisement

Source link

Continue Reading

Servers computers

How to install Server Rack 9U Full Complete Video

Published

on

How to install Server Rack 9U Full Complete Video



Server Rack install Full Video 9U And any question ask me
thank you all
Whatsapp : +8801710532227

source

Continue Reading

Technology

Steam will let you sue Valve now

Published

on

Steam will let you sue Valve now

Steam just removed its forced arbitration policy, opening the door for lawsuits against its parent company, Valve. In an update on Thursday, Steam says its subscriber agreement “now provides that any disputes are to go forward in court instead of arbitration.”

Many companies include a forced arbitration clause in their user agreement, waiving a person’s right to a trial in court. Arbitration involves settling a dispute outside a legal system before an impartial third party. This method is often faster but may not get the best results for consumers, as arbitrators don’t need to consider the law when issuing a decision.

Previously, Steam’s user agreement said, “you and Valve agree to resolve all disputes and claims between us in individual binding arbitration” for all disputes related to Steam, your account, hardware, or the company’s content and services. The new agreement eliminates any mention of a binding arbitration policy.

Steam doesn’t say why it decided to suddenly remove the forced arbitration clause. As pointed out by 404 Media, a group of plaintiffs recently challenged Valve’s forced arbitration policy and were able to file a class action lawsuit over Steam’s dominance.

Advertisement

Source link

Continue Reading

Servers computers

How to make a fire protection server cabinet? Lehmann F90 Server Rack – Made in Germany

Published

on

How to make a fire protection server cabinet? Lehmann F90 Server Rack - Made in Germany



19-inch server cabinets with integrated fire protection from Lehmann IT

19-inch server cabinets with integrated fire protection can basically be placed in all locations where there is an increased risk of fire in order to optimally protect the devices located in the server cabinet as well as all relevant data. For this purpose, we at Lehmann IT have developed and designed various solutions that can be used for numerous applications and requirements.

When should a server cabinet with fire protection be considered?
Basically, a 19-inch server cabinet with an integrated fire protection function can never hurt, as fires can occur anytime and anywhere, often caused by electricity or human error. However, these models are especially interesting for companies that want to place the cabinets in locations with increased fire risk.

As a rule, however, fire-protected server cabinets are used in the industrial sector. In times of Industry 4.0 and the advancing digitalization of industrial processes, reliable IT infrastructures are increasingly needed in production halls. These are often subject to high temperatures because the machines used there radiate heat during operation or the production processes themselves require these temperatures. In addition, the amount of dust is very high, which is also harmful to unprotected IT equipment. This is because this dust accumulates on and in the equipment and can sometimes clog air slots and fans, which has a strong negative impact on air circulation in the cabinet, resulting in so-called hot spots. At these points, there is an extremely high temperature locally, which can cause considerable damage to the equipment.

Especially in industrial halls, IT equipment should therefore be well protected, which is why the purchase of a server cabinet is inevitable. Since the risk of fire in these industrial halls is very high due to the high temperature or the use of highly flammable substances, it is recommended to immediately choose a model with integrated fire protection. In the event of a fire, this will not only protect the equipment but also all the data stored on the servers.

What consequences can fire have for IT equipment?
Even small and quickly contained fires can cause damage with high costs. Even server racks without integrated fire protection can be severely affected in a very short time. If a fire is not quickly detected and appropriately fought, IT equipment will be very badly damaged as a result of sooting and corrosion. Smoke development also has severe consequences for the servers, routers, and switches in the cabinet. The technology is no longer salvageable as a result of a fire, and the data on the devices can also be irretrievably lost. A server cabinet with fire protection not only protects the devices themselves but also gives you enough time to save the data.

How does fire protection in the server cabinet work?
The fire protection server cabinets are certified according to the relevant DIN and EN guidelines and offer fire resistance of approximately 90 minutes. In addition, the models comply with the IP54 protection class, which is an indicator of protection against dust and water, however, both of which can also be potential hazards for server cabinets in industrial halls. The surface of the server cabinets is highly resistant to chemicals and moisture. The cabinets have sensors that measure the exhaust air temperature and smoke development in the cabinet. If this is too high, all ventilation openings are immediately closed.

According to the European classification EN 13501-1, our server cabinets have the marking As1d0. The “A” indicates that the cabinet itself does not contribute to the fire, which is the highest rating for this fire safety aspect. The “s1” (for low smoke emission) and “d0” (no burning drip) markings are also the highest indexes for the aspects considered, according to the EN directive.

The classification EN 13501-2 is also important for the classification of fire protection for various devices. Our cabinets have the specification EI 90, which means that the cabinet is fire resistant for 90 minutes. The inner and outer walls of the server cabinet are non-load-bearing, as no loads are applied to them. This is expressed by the two letters “EI”.

Lehmann IT – Your partner for server cabinets
If you have questions about our server and network cabinets or need help choosing a suitable model, we are the right contact for you. We are looking forward to your inquiry.

https://www.lehmann-it.de/en/it-fire-protection.html .

source

Continue Reading

Technology

Nvidia Container Toolkit found to have worrying security flaws

Published

on

Nvidia Container Toolkit found to have worrying security flaws

NVIDIA Container Toolkit and GPU Operator were carrying a critical vulnerability that allowed threat actors access to the underlying host’s file system, experts have warned.

Cybersecurity researchers at Wiz discovered and reported the flaw, tracked as CVE-2024-0132, and carries a vulnerability score of 9.0/10 – critical, to Nvidia on September 1, 2024.

Source link

Continue Reading

Technology

Meta fined €91m after password storage investigation

Published

on

Meta fined €91m after password storage investigation

Facebook parent company Meta has been fined €91m (£75m) by the Irish Data Protection Commission (DPC) following an investigation into the storage of passwords.

An inquiry was launched in April 2019 after Meta notified the DPC that it had inadvertently stored certain passwords of social media users on its internal systems without encryption.

The DPC submitted a draft decision to other European data watchdogs in June 2024.

No objections were raised by the other authorities.

Advertisement

Meta has been found to have four breaches of General Data Protection Regulation (GDPR).

DPC deputy commissioner Graham Doyle said: “It is widely accepted that user passwords should not be stored in ‘plaintext’ considering the risks of abuse that arise from persons accessing such data.

“It must be borne in mind, that the passwords the subject of consideration in this case are particularly sensitive, as they would enable access to users’ social media accounts.” he added.

The decision, which was made by the commissioners for data protection, Dr Des Hogan and Dale Sunderland, and notified to Meta on 26 September, includes a reprimand and a fine.

Advertisement

In May 2023, Meta was fined €1.2bn (£1bn) for mishandling data when transferring it between Europe and the United States.

That fine was also issued by Ireland’s DPC; the largest fine imposed under the EU’s GDPR privacy law.

In 2022, Meta was fined €265m (£220m) after data from 533m people in 106 countries was published on a hacking forum having been “scraped” from Facebook years earlier.

Source link

Advertisement

Continue Reading

Trending

Copyright © 2024 WordupNews.com