Connect with us
DAPA Banner

Tech

How Long After A Tire Rotation Should You Re-Torque Your Wheels?

Published

on





A good set of quality tires can typically withstand everything from rough roads to bad weather, and a lot more. Of course, getting the most out of your tires means doing preventative maintenance as well, and that’s where regular tire rotations come into play. But it’s also a good idea to re-torque your wheels about 30 miles after your rotation. It’s a practice that can potentially save you from some problems later on.

“Re-torque” simply means to re-tighten, as your lug nuts can loosen over time. This can sometimes be caused by heat, but motion can be a big contributor as well. Even the weight of your vehicle can add to the problem. Despite how well the wheel was secured during your rotation, exterior forces can impact your tires. Because of this, you may end up with uneven tread wear, or possibly a loose wheel, which could make for a dangerous situation.

It’s important to note that your lug nuts may not move that much, if at all, when you re-torque them. After all, wheel and tire assemblies can vary, and if you don’t drive that often, you might be just fine. In fact, you may be able to go from one tire rotation to the next without an issue. But it’s better to be safe than sorry. When in doubt, stop by your local garage and have a technician take a look. It might cost you a little time, but it could save you some grief in the long run.

Advertisement

The proper technique for re-torquing your wheels

You might not rotate your own tires at home, but you can re-torque your own wheels. Before you begin, consult your vehicle’s owner’s manual. You should be able to find some useful information about the correct torque specifications for your make and model. This is important, because every vehicle is different in terms of how much force it takes to properly secure your wheels. Too little torque and your wheels could come loose. Too much, and you’re risking possible damage to the wheel.

If you’re re-tightening the lug nuts while your vehicle is on the ground, the weight of your car should keep the tires stationary. Be sure to park on a flat surface and put on your parking brake. Next, use a torque wrench to tighten each lug nut in a crisscross/star pattern to the proper specification according to your owner’s manual. But beware that if you use a tire iron, you won’t be able to achieve the exact torque as specified in your manual.

Advertisement

If you do have access to a lift and want to tighten your lug nuts that way, the biggest difference is that your vehicle will be off the ground. This is where an actual torque wrench will come in handy, as the wheel would be less likely to move as much during the tightening process. Just follow the same crisscross pattern, tighten the bolts to the proper specifications via your owner’s manual, and you’re all set.



Advertisement

Source link

Continue Reading
Click to comment

You must be logged in to post a comment Login

Leave a Reply

Tech

Anthropic loses bid to pause Pentagon blacklisting as AI legal battle escalates

Published

on


The ruling keeps Anthropic locked out of DoD contracts for now, even though a separate federal court in California recently barred the Trump administration from enforcing a broader ban on the use of Claude.
Read Entire Article
Source link

Continue Reading

Tech

Some Windows 3.1 apps were simply "too evil" for Windows 95 to support, says Microsoft veteran

Published

on


Microsoft veteran Raymond Chen is once again spilling the beans on how Windows 95 became one of the most influential operating systems ever. Back in the Nineties, Microsoft developers were busy working on many custom solutions to make the new OS compatible with previous software products. However, a few programs…
Read Entire Article
Source link

Continue Reading

Tech

Kia cuts EV target, confirms electric pickup, and plans to put Atlas robots in its Georgia factories

Published

on

In short: On the day that 25% US tariffs on South Korean imports took effect, Kia held its 2026 CEO Investor Day in Seoul and presented a plan built for a changed world: a quietly reduced EV sales target for 2030, a major expansion of its hybrid range, the first confirmation of a North American electric pickup truck, and a commitment to deploy Boston Dynamics’ Atlas humanoid robots in its Georgian factories from 2028. The five-year investment plan reaches KRW 49 trillion, and the company is targeting KRW 170 trillion in revenue by 2030.

Kia President and CEO Ho-sung Song opened the event with a statement of direction: “EVs, HEVs, autonomous driving, and robotics will serve as key drivers for Kia’s fastest growth to date.” The framing is deliberately broad,  a recognition that the path to Kia’s 2030 ambitions no longer runs through battery-electric vehicles alone, and that the company must build revenue across multiple technology bets simultaneously.

A lower EV target, a bigger hybrid push

The most numerically significant announcement at this year’s event is one Kia did not frame as a retreat. The company’s 2030 EV sales target now stands at 1 million units annually, across a lineup that will expand to 14 models. That figure represents a reduction of roughly 20% from the approximately 1.26 million target set at last year’s investor day, and a sharper fall from the 1.6 million target Kia set at its 2023 event. The causes are well understood: the elimination of US EV subsidies, the slowdown in US battery-electric sales, and the weight of import tariffs that cost the group KRW 3.3 trillion (approximately $2.3 billion) in 2025 alone.

In place of the lost EV volume, Kia is expanding its hybrid offer substantially. Annual HEV sales are now targeted at 1.1 million units by 2030, supported by a lineup growing to 13 models. Combined with the EV target, Kia plans to sell 2.1 million electrified passenger vehicles per year by the end of the decade, out of a total of 4.13 million units and a targeted global market share of 4.5%. Its purpose-built vehicle (PBV) range, comprising the PV5, PV7, and PV9 commercial models, adds a further 232,000 unit target by 2030. Regionally, Kia is targeting 1.02 million units in the US, 746,000 in Europe, and 1.48 million in emerging markets.

Advertisement

The immediate financial picture is more pressing than the 2030 targets. For 2026, Kia is projecting KRW 122.3 trillion in sales and KRW 10.2 trillion in operating profit — a recovery from the tariff-hit prior year, premised on the 15% tariff rate established under the Korea-US agreement in late 2025, which replaced the previous 25% rate. Whether that rate holds under continued trade policy pressure remains an open question.

The 💜 of EU tech

The latest rumblings from the EU tech scene, a story from our wise ol’ founder Boris, and some questionable AI art. It’s free, every week, in your inbox. Sign up now!

A pickup truck for America, and for the tariff era

The announcement that received the most immediate attention is Kia’s confirmation that it will build a mid-size electric pickup truck aimed specifically at North America. The model will be built on a next-generation EV platform, and the company is targeting a 7% share of the North American pickup truck market, implying annual sales of approximately 90,000 units in the medium to long term.

Advertisement

Kia did not confirm where the vehicle will be manufactured, but the strategic logic is clear. Both of the group’s US facilities — Hyundai Motor Group Metaplant America in Georgia and Kia’s own manufacturing plant in West Point, Georgia — are positioned to produce vehicles that avoid import tariffs, including both the longstanding “Chicken Tax” applied to light trucks and the newer EV import levies. The timing of the announcement, made on the same day that 25% reciprocal tariffs on South Korean imports came into force, underlines the degree to which Kia is reconfiguring its product strategy around US production.

Atlas on the factory floor

Kia also used the investor day to advance its timeline for deploying Boston Dynamics’ Atlas humanoid robots in its manufacturing operations. Atlas robots — trained at Hyundai Motor Group’s Robotics Metaplant Application Centre — are scheduled to begin sequencing tasks at HMGMA in 2028, with more complex assembly operations beginning by 2030. The programme will then expand to Kia AutoLand Georgia in the second half of 2029.

The contest to deploy humanoid robots in production environments at scale has been building for several years, with automakers positioned as early adopters given the structured and predictable nature of assembly line work. Boston Dynamics unveiled a production-ready version of Atlas at CES 2026 and said all 2026 deployments were already committed. As humanoid robots move from demonstration to production-line deployment, manufacturers are working out what tasks the technology can handle reliably and which require further development before genuine integration into complex assembly. Kia’s roadmap, sequencing tasks first, assembly later — reflects that staged approach.

Boston Dynamics is a subsidiary of Hyundai Motor Group, which gives Kia preferential access to Atlas deployments. Alongside the factory programme, Kia is exploring last-mile logistics applications that combine its PBV range with Boston Dynamics’ Stretch logistics robot for warehouse operations and its Spot quadruped for on-site delivery.

Advertisement

Software-defined vehicles, autonomy, and the financial plan

Kia’s technology roadmap beyond hardware commits the company to completing its first software-defined vehicle model, equipped with highway-level 2+ autonomous driving capability, by the end of 2027. Urban autonomous driving at Level 2++ is targeted for rollout from early 2029. The competitive context for higher-level autonomy is shifting quickly, with robotaxi operators expanding their geographic footprints and the gap between technology leaders and production vehicle manufacturers becoming harder to ignore. Kia’s AV programme, while more conservative than pure-play autonomous operators, is designed to bring meaningful driver assistance into high-volume production vehicles rather than limited commercial fleets.

The financial scaffolding for all of this is KRW 49 trillion in investment over the five-year period from 2026 to 2030, of which KRW 21 trillion is earmarked for future business areas including robotics, SDVs, and autonomous driving. The year 2025 crystallised how unevenly the AI and technology dividend was being distributed across industries, and Kia’s investment plan reflects an explicit attempt to ensure that the automotive business captures value from the automation and software transitions rather than ceding it to technology companies entering the mobility space. By 2030, Kia is targeting KRW 170 trillion in annual revenue and a 10% operating profit margin, implying KRW 17 trillion in operating profit. Whether that margin is achievable depends heavily on how trade policy, EV demand, and the pace of hybrid uptake develop over the next four years. The convergence of automotive hardware and AI-driven mobility software is accelerating, and Kia’s investor day is, in aggregate, a bet that traditional automakers can compete in both domains if they act now.

Source link

Advertisement
Continue Reading

Tech

How AI-Powered Identity Verification is Redefining Business Security

Published

on

Passwords have been the standard of online security. Next was the two-factor authentication. Then security questions, CAPTCHA, and fingerprinting of devices. Every layer introduced with a new threat. Both were ultimately defeated by more advanced scams.

The trend is obvious: any security system relying on what one knows or possesses will be susceptible to theft, copying, or social engineering. The one verification level that is truly hard to counterfeit is who someone is – and that is exactly where artificial intelligence has transformed all that.

Identity verification using AI is no longer a niche technology that is implemented only by banks and governmental agencies. It is also going to be the minimum security requirement of any business onboarding clients digitally, transacting high-value deals, or working within a regulated sector in 2026. The knowledge of how it works, why it is important, and how to apply it is now a business competency rather than an IT issue.

The Issue Classic Security Cannot Address

It is only prudent to know what AI-driven identity verification is meant to address before delving into how it works, since the threat landscape has changed drastically.

Advertisement

Credential breach has rendered credentials a worthless security signal. The Cost of a Data Breach Report by IBM indicated that in 2024, the mean data breach involved more than 25,000 records. Out of the thousands of breaches that have taken place worldwide in the last ten years, billions of usernames and passwords, social security numbers, dates of birth, and answers to security questions are now being sold on the dark web. With access to such databases, a fraudster can easily pass through traditional credential checks since the credentials are authentic, only that they are owned by a different person.

The synthetic identity fraud has generated a new breed of criminal. More than stealing an existing identity, advanced fraudsters are building identities, assembling a real Social Security number (usually that of a child or an aged individual with no credit history) with invented names, addresses, and biographical information. These artificial identities can withstand a simple verification check since some of the information is authentic. They are mostly unnoticed by traditional rules-based fraud detection systems.

Deepfakes created by AI have defeated selfie-based authentication. The fast development of generative AI has brought about tools that are capable of generating photorealistic fake images, videos, and even real-time video feeds of non-existent individuals within minutes. The days of systems utilizing a mere selfie photo to verify identity are long gone, with fraudsters capable of uploading a deepfake image that, visually, resembles a real photo.

Credential theft, synthetic identity fraud, and AI-generated deepfakes are the three converging threats that next-generation AI-powered identity verification is designed to deal with.

Advertisement

The reality of what AI-Powered Identity Verification does

Identity verification is not just an AI-based technology. It is a multi-tiered system of a series of AI models operating together to determine with high probability that an individual is who they claim they are.

Document Authentication

The initial layer is document checking. A user enters a government-issued identity document, passport, driver’s license, national ID card, and an AI model compares it with thousands of known document templates that exist in the world.

The level of the analysis is much higher than determining whether the document is real. Machine learning algorithms trained on millions of real and fake documents analyze the quality of microprints, the presence of UV patterns, holographic elements, font authenticity, MRZ (Machine Readable Zone) information integrity, and pixel-level anomalies (which signify editing and manipulation). Digitally manipulated documents (even in subtle ways) are detected within seconds.

Advertisement

The system of document verification is available in modern document verification systems that can verify more than 14,000 types of documents representing more than 190 countries, which would not otherwise be feasible to verify manually.

Biometric Face Matching

When the document has been verified, the system will compare the face on the document to a live selfie or a video submission by the individual purporting to be the document holder. In AI facial recognition models, the geometric distance between facial features, such as the distance between eyes, nose shape, jaw angle, and a confidence score of the match, is calculated.

It is a quick, precise, and much more dependable method than a visual inspection by people. Research by the National Institute of Standards and Technology (NIST) consistently reported that the best facial recognition algorithms perform better than human examiners in face matching tasks, especially when there are changes in lighting, angle, and age.

Advertisement

Liveness Detection

It is the layer that deals with deepfake fraud in particular, and it is in this area that AI has achieved the most critical progress.

Liveness detection identifies when the face presented is that of a real, physically present human being, or whether it is a photograph, printed mask, video recording, or deepfake generated by a computer AI. Passive liveness detection examines a single image of slight signs of non-liveness: texture anomalies, unnatural light reflection, absence of micro-movements, or compression artifacts suggesting a screen capture. Active liveness detection requires the user to do randomized behaviors: blink, move their head, smile, which are virtually impossible to impersonate by a still image and computationally infeasible to spoof by a live deepfake.

Passive and active liveness detection combined has increased the threshold to deepfake fraud attacks to the extent that the cost of a successful attack is usually more economical than the fraudulent value, and AI-generated identity fraud attacks are thus not economical in most criminal activities.

Advertisement

Cross-Referencing of Data and AML Screening

Outside the biometric layer, identity verification systems built with AI will cross-verify the verified identity data against external databases in real-time. This encompasses global sanctions lists, Politically Exposed Persons (PEP) databases, adverse media sources, and watchlists that are managed by regulatory agencies such as the OFAC, the UN, and the EU.

It is this AML screening layer that makes identity verification a compliance tool, as well as a security tool, such that businesses can fulfill their Know Your Customer (KYC) and Anti-Money Laundering (AML) requirements alongside the verification check, instead of as a downstream operation.

The Importance of Thematically Integrated Security to Business Security – Not Just Compliance

The argument of AI-based identity verification as compliance is well-established. In practically every jurisdiction, financial services companies, fintech, and other regulated businesses are required to perform KYC and AML processes on a compulsory basis. Failing to meet them carries substantial financial penalties and reputational risk.

Advertisement

However, the business security case is far bigger than regulatory compliance – and to most businesses, the non-compliance risks pale in comparison with the direct losses of fraud that can be easily facilitated by poor identity verification.

Businesses are directly affected by account takeover fraud. Once a fraudster manages to create a successful impersonation of an authentic customer in the process of onboarding or recovering an account, they access the available accounts, payment methods, and stored credit. The ensuing chargebacks, frauds, and dispute settlements are more on the business side than the card network. Account takeover fraud is a major and increasing direct operating expense to e-commerce companies and financial technology applications.

New account fraud generates unpayable debts. Synthetic identity fraud generally leads to the so-called bust-out schemes in which a fraudster accumulates credit exposure on a variety of products, and then defaults on all of them at once. To lenders, credit providers, and buy-now-pay-later sites, the damages of a single synthetic identity that has been nurtured over months can go into tens of thousands of dollars.

Financial loss is compounded by reputational loss because of instances of fraud. In cases where clients of a business fall victim to fraud by a security breach on a platform, the reputational loss is more than just the direct financial loss. The loss of customers, media attention, and regulatory investigations after a fraud incident can be even more expensive than the actual losses incurred in the fraud itself – especially to a business in which trust is the product.

Advertisement

At the onboarding stage, AI-based identity checks prevent the vast majority of such attack vectors, prior to the creation of a fraudulent account. Compliance cost avoidance is only part of the payback; it is the avoidance of downstream fraud losses that grow with business expansion.

Real-life Application: What Companies Should know

The practical considerations of AI-powered identity verification extend beyond the technology when business leaders consider this technology.

Should Be API-First Deployment

Contemporary identity verification systems are implemented through API integration – linking your onboarding process with the verification service without having the customer leave your site. This retains the customer experience and facilitates instant verification decisions. Find options that enable integration of SDKs in mobile applications and provide a webhook-based delivery of decisions to reduce the onboarding latency.

Advertisement

Risk Level should be configurable to determine Verification Decisions

Customers do not pose the same fraud risk, and not all transactions need the same level of verification. An effective AI-based solution enables companies to set up verification processes according to risk indicators – introduce lightweight document verification to transactions with low risks and complete biometric verification with liveness detection to high-value or high-risk onboarding situations. This risk-based model maintains conversion rates among legitimate customers and focuses verification resources where the fraud risk is the greatest.

Audit Trails are Not Negotiable

Each verification decision, be it approval, rejection, or flagged to undergo manual review, should be recorded with a time stamp, the particular methods used to verify, the confidence levels delivered by the methods, and the documentation. Such an audit trail is necessary in regulatory audits, chargeback audits, and internal fraud audits. Firms that are subject to FINTRAC, GDPR, or other regulations must generate such records when they are requested, usually in 30 days or fewer.

Advertisement

Should Be Constructed Human Review Escalation

AI verification systems are extremely precise, yet no computerized system can be 100 percent confident in all cases. Good implementations involve a queue of cases with AI confidence less than a set-point – often around 5-10% of all verifications. The edge cases that are not detected by the automated systems are picked by human reviewers looking at the flagged cases, and their verdicts are used to inform further improvement of the model.

Select a Partner that has Worldwide Document Covers

When you have customers in a variety of countries, your identity verification provider should accept document types in those countries. An optimized system for North American documents will result in an unacceptable high rate of false rejection of customers with a Southeast Asian, Middle Eastern, or African identity document. Such solutions as the document verification offered by Shufti Pro can work with documents issued in 190+ countries – an essential feature that businesses with international clientele can use.

Advertisement

The Competitive Advantage of this Right

The divide between companies that have invested in solid identity verification infrastructure and those that have not is widening, and the difference has repercussions beyond losses in fraud.

The relations between payment processors are based on fraud indicators. The card networks and payment processors keep a close eye on the chargeback rates and the fraud rates. Companies with low fraud traces due to proper identity checking receive superior processing rates, increased transaction limits, and preference of merchants. Companies that have higher fraud rates will be charged higher fees, delays in processing, and, in the worst case, the merchant account will be shut down.

Security posture is also necessary to acquire enterprise clients. Enterprise customers: Large enterprises (especially in the financial services, medical, and government contracting) perform vendor security testing before contracting. Documented, auditable identity verification and fraud prevention program is becoming a condition to winning enterprise business, and not a differentiator.

Fraud infrastructure is studied in investor due diligence. In the case of growth-stage businesses that are in need of investment, fraud prevention infrastructure is part of due diligence. Fintech, e-commerce, and SaaS investors prefer to observe that the business has developed security basics that can scale up since fraud losses that can be controlled at the early stage become existential at the growth stage when the infrastructure is lacking.

Advertisement

The Future: The Future of AI Identity Verification

The technology does not stand still. Several trends are underway transforming AI-driven identity verification in 2026 and beyond.

Continuous authentication has passed onboarding. Instead of authenticating identity when creating an account, AI systems are starting to track behavioral indicators, such as typing patterns, mouse motions, transaction activities, etc., in real time, and used in the course of a user session, which indicates anomalies that may indicate account takeover.

There is an increasing regulatory trend toward decentralized identity frameworks, in which verified credentials are stored by the user, but not by individual businesses, both in the EU and Canada. These frameworks minimise the data liability that businesses already bear when it comes to storing identity documents and biometric data.

There are ever-growing regulatory requirements across the world. Fintrac of Canada, the AML package of the EU, and other systems in Asia-Pacific are increasing standards of identity verification – that is, what is best practice now will become legal minimum tomorrow.

Advertisement

Concluision

The paradigm change that AI-enabled identity verification will be a transition to proactive security, rather than reactive security. Conventional methods identified fraud only once it occurred, by way of chargeback, account audits, and forensic audits. Verification, which is AI-based, detects fraud when it is attempted – before creation of a fraudulent account, before a stolen identity being impersonated, before a deep fake passing through an onboarding test.

In the case of businesses that are scaling, that change does not qualify as a security upgrade. It is a foundation. Survivable losses of fraud at a small scale are devastating at the growth stage. It is the businesses that develop strong identity verification infrastructure early that develop without the compounding drag of costs associated with fraud, compliance failures, and reputational incidents slowing them down.

With the cost of impersonation in a digital economy falling to almost zero, the cost of not authenticating identity is increasing year after year. 

Advertisement

Source link

Continue Reading

Tech

Decentralized AI Training Turns Homes Into Data Hubs

Published

on

Artificial intelligence harbors an enormous energy appetite. Such constant cravings are evident in the hefty carbon footprint of the data centers behind the AI boom and the steady increase over time of carbon emissions from training frontier AI models.

No wonder big tech companies are warming up to nuclear energy, envisioning a future fueled by reliable, carbon-free sources. But while nuclear-powered data centers might still be years away, some in the research and industry spheres are taking action right now to curb AI’s growing energy demands. They’re tackling training as one of the most energy-intensive phases in a model’s life cycle, focusing their efforts on decentralization.

Decentralization allocates model training across a network of independent nodes rather than relying on one platform or provider. It allows compute to go where the energy is—be it a dormant server sitting in a research lab or a computer in a solar-powered home. Instead of constructing more data centers that require electric grids to scale up their infrastructure and capacity, decentralization harnesses energy from existing sources, avoiding adding more power into the mix.

Hardware in harmony

Training AI models is a huge data center sport, synchronized across clusters of closely connected GPUs. But as hardware improvements struggle to keep up with the swift rise in size of large language models, even massive single data centers are no longer cutting it.

Advertisement

Tech firms are turning to the pooled power of multiple data centers—no matter their location. Nvidia, for instance, launched the Spectrum-XGS Ethernet for scale-across networking, which “can deliver the performance needed for large-scale single job AI training and inference across geographically separated data centers.” Similarly, Cisco introduced its 8223 router designed to “connect geographically dispersed AI clusters.”

Other companies are harvesting idle compute in servers, sparking the emergence of a GPU-as-a-Service business model. Take Akash Network, a peer-to-peer cloud computing marketplace that bills itself as the “Airbnb for data centers.” Those with unused or underused GPUs in offices and smaller data centers register as providers, while those in need of computing power are considered as tenants who can choose among providers and rent their GPUs.

“If you look at [AI] training today, it’s very dependent on the latest and greatest GPUs,” says Akash cofounder and CEO Greg Osuri. “The world is transitioning, fortunately, from only relying on large, high-density GPUs to now considering smaller GPUs.”

Software in sync

In addition to orchestrating the hardware, decentralized AI training also requires algorithmic changes on the software side. This is where federated learning, a form of distributed machine learning, comes in.

Advertisement

It starts with an initial version of a global AI model housed in a trusted entity such as a central server. The server distributes the model to participating organizations, which train it locally on their data and share only the model weights with the trusted entity, explains Lalana Kagal, a principal research scientist at MIT’s Computer Science and Artificial Intelligence Laboratory (CSAIL) who leads the Decentralized Information Group. The trusted entity then aggregates the weights, often by averaging them, integrates them into the global model, and sends the updated model back to the participants. This collaborative training cycle repeats until the model is considered fully trained.

But there are drawbacks to distributing both data and computation. The constant back and forth exchanges of model weights, for instance, result in high communication costs. Fault tolerance is another issue.

“A big thing about AI is that every training step is not fault-tolerant,” Osuri says. “That means if one node goes down, you have to restore the whole batch again.”

To overcome these hurdles, researchers at Google DeepMind developed DiLoCo, a distributed low-communication optimization algorithm. DiLoCo forms what Google DeepMind research scientist Arthur Douillard calls “islands of compute,” where each island consists of a group of chips. Every island holds a different chip type, but chips within an island must be of the same type. Islands are decoupled from each other, and synchronizing knowledge between them happens once in a while. This decoupling means islands can perform training steps independently without communicating as often, and chips can fail without having to interrupt the remaining healthy chips. However, the team’s experiments found diminishing performance after eight islands.

Advertisement

An improved version dubbed Streaming DiLoCo further reduces the bandwidth requirement by synchronizing knowledge “in a streaming fashion across several steps and without stopping for communicating,” says Douillard. The mechanism is akin to watching a video even if it hasn’t been fully downloaded yet. “In Streaming DiLoCo, as you do computational work, the knowledge is being synchronized gradually in the background,” he adds.

AI development platform Prime Intellect implemented a variant of the DiLoCo algorithm as a vital component of its 10-billion-parameter INTELLECT-1 model trained across five countries spanning three continents. Upping the ante, 0G Labs, makers of a decentralized AI operating system, adapted DiLoCo to train a 107-billion-parameter foundation model under a network of segregated clusters with limited bandwidth. Meanwhile, popular open-source deep learning framework PyTorch included DiLoCo in its repository of fault tolerance techniques.

“A lot of engineering has been done by the community to take our DiLoCo paper and integrate it in a system learning over consumer-grade internet,” Douillard says. “I’m very excited to see my research being useful.”

A more energy-efficient way to train AI

With hardware and software enhancements in place, decentralized AI training is primed to help solve AI’s energy problem. This approach offers the option of training models “in a cheaper, more resource-efficient, more energy-efficient way,” says MIT CSAIL’s Kagal.

Advertisement

And while Douillard admits that “training methods like DiLoCo are arguably more complex, they provide an interesting tradeoff of system efficiency.” For instance, you can now use data centers across far apart locations without needing to build ultrafast bandwidth in between. Douillard adds that fault tolerance is baked in because “the blast radius of a chip failing is limited to its island of compute.”

Even better, companies can take advantage of existing underutilized processing capacity rather than continuously building new energy-hungry data centers. Betting big on such an opportunity, Akash created its Starcluster program. One of the program’s aims involves tapping into solar-powered homes and employing the desktops and laptops within them to train AI models. “We want to convert your home into a fully functional data center,” Osuri says.

Osuri acknowledges that participating in Starcluster will not be trivial. Beyond solar panels and devices equipped with consumer-grade GPUs, participants would also need to invest in batteries for backup power and redundant internet to prevent downtime. The Starcluster program is figuring out ways to package all these aspects together and make it easier for homeowners, including collaborating with industry partners to subsidize battery costs.

Backend work is already underway to enable homes to participate as providers in the Akash Network, and the team hopes to reach its target by 2027. The Starcluster program also envisions expanding into other solar-powered locations, such as schools and local community sites.

Advertisement

Decentralized AI training holds much promise to steer AI toward a more environmentally sustainable future. For Osuri, such potential lies in moving AI “to where the energy is instead of moving the energy to where AI is.”

From Your Site Articles

Related Articles Around the Web

Source link

Advertisement
Continue Reading

Tech

Avalanche Energy lands share of $5.2M DOD award to develop long-lasting ‘nuclear batteries’

Published

on

An early prototype of Avalanche Energy’s radiovoltaic converter for the DARPA Rads to Watts program is exposed to high-energy ion-beam irradiation. (Avalanche Photo)

Seattle fusion startup Avalanche Energy was awarded a share of a $5.2 million contract announced Wednesday from the U.S. Department of Defense to develop compact nuclear batteries.

The award comes from the DARPA Rads to Watts program, which is focused on building long-lasting batteries for defense and space applications where chemical batteries, solar power and refueling are not possible.

Avalanche is focused on engineering micro-fabricated energy cells that turn alpha particles emitted by radioactive material into electricity. The process, the team said, is analogous to solar cells converting photons into electricity.

“The goals are to produce a device that has a long lifetime, and that can produce orders of magnitude more power than current technologies,” said Daniel Velázquez, Avalanche’s physicist and materials science lead. The target is a battery that could continuously power a laptop computer, for example, for many months but weighs roughly 10 pounds.

And the timeline is tight. By the end of the 30-month program, the objective is to validate the physics involved and develop a power-producing prototype.

Advertisement

“It’s very ambitious,” Velázquez said.

Avalanche is leading the team tackling DARPA’s nuclear battery challenge, which includes the University of Utah, Caltech, Los Alamos National Laboratory and McQuaide Microsystems.

Others are also working on nuclear batteries, including Seattle’s Zeno Power. The startup plans to demonstrate its first full-scale radioisotope power system this year and commercially produce nuclear batteries by 2027.

While Avalanche is ultimately working to develop a compact device that creates energy from fusion — the reactions that power the sun — the DARPA project feeds directly into that longer-term goal, Velázquez said. There are direct parallels to capturing energy from a nuclear battery and from fusion reactions.

Advertisement

That should help the company compete in the global race to commercialize fusion power, which could provide nearly limitless clean energy. To support domestic enterprises, the Department of Energy is slated to commit a record-setting $135 million over 18 months to accelerate fusion research, Axios reported today.

Demand for new power is spiking with the expansion of data centers and the shift from fossil fuels to electrification.

Since launching in 2018, Avalanche has pursued multiple lines of revenue. Last month, the company announced it’s part of a team receiving $1.25 million from AFWERX, the innovation arm of the Department of the Air Force, to develop advanced materials for extreme environments.

Other efforts include using its fusion machine to produce neutrons for commercial customers; a Pentagon contract to develop technology for space propulsion; and a state grant to launch FusionWERX, a commercial-scale testing facility for fusion technologies in Eastern Washington.

Advertisement

In February, Avalanche announced $29 million in new funding from investors, bringing its total to more than $105 million across venture capital and government grants — a war chest the company is deploying across fusion, propulsion and now compact nuclear batteries.

Source link

Continue Reading

Tech

Fi Mini for Cats Review: Track Your Pets and Monitor Their Activity

Published

on

Within the app, you can add safe zones, more pets with Fi trackers, and other users who can also track and monitor the pet. There’s a Health tab where you can add and store things like vet records, receipts, and insurance information, and add vets to easily share your pet’s documents and get appointment reminders. You can also set up the Fi app on your Apple Watch to have even quicker access to monitor your pet’s location, activity, and safety (including Lost Mode) without needing a phone.

When you open the app, you’ll see a map with live tracking showing where your pet is currently, as well as a notification of the last time they were outside and where they were. With the latter, you can pull up stats like location timeline, showing where they were and when. If you dive into any day when the tracker left the home, it will recreate the route, following the path and calculating the distance the pet traveled.

There’s also health-monitoring data from activity and sleep tracking, which is most useful for an indoor-only pet like mine. Like other health-tracking collars, stats for sleep and activity aren’t 100 percent accurate, as the app uses GPS to track movement, categorizing “activity” when the animal is moving and “sleep” when the pet is still for a prolonged period. This means that if Basil was awake but stationary, the app may inaccurately categorize this as sleep.

Image may contain Text

Fi Mini App source Molly Higgins

In the Rest tab, you can see sleep metrics, including a daily summary of deep sleep, naps, and interruptions during nightly sleep. You can compare this over time, and the app notes how much more or less Basil slept than the night before. It also compares stats historically, by week, month, and year, so you can track trends and better understand your pet’s normal sleep schedule.

Advertisement

The Activity tab is similar, tracking activity by day, week, and month, noting in the day’s timeline when the pet was most active and for how long. This also compares activity to the day before. I liked looking at the weekly report, comparing days during the week to see which he was most active during and if any patterns in activity popped up.

For example, I noticed that his sleep-versus-activity schedule was similar to mine, except he was active between 4:45 and 6:30 am (while I was still asleep), because that’s when his automatic feeder goes off for breakfast and my roommate is getting ready to leave for work. He was most active in the evenings, when I feed him dinner, have dedicated playtime, and my roommates are home, so there’s more activity to keep him awake. Historical comparison is also a super helpful way to track whether your pet is sleeping more or becoming more lethargic—an early warning sign of a bigger health problem.

Not Without Its Quirks

Since my cat is indoor-only, I ran some experiments to track location, using GPS on both the Fi Mini tracker and my phone. I also had a friend take the tracker out without my phone nearby to see whether I’d get pinged that “Basil” had left the safe zone.

Although it is better than not being alerted at all, the Fi’s GPS has limitations (as did the Tractive tracker I tested). It needs a strong signal to communicate with cell towers for accurate location. If your phone is close to the smart collar (via Bluetooth), it uses that instead of the Fi’s GPS, making it more accurate and alerting quicker. If the pet gets loose and is out of range of your phone, it uses the collar’s cellular antenna (in this case, Verizon cell towers). But because the Fi’s antenna isn’t as strong as a phone’s, location accuracy is lower, and the connection can be very spotty, especially if your pet is in the country or on acreage where cell towers are farther away.

Advertisement

Source link

Continue Reading

Tech

Eurail says December data breach impacts 300,000 individuals

Published

on

Eurail

Eurail B.V., a European travel operator that provides digital passes covering 33 national railways, says attackers stole the personal information of over 300,000 individuals in a December 2025 data breach.

Eurail is a Netherlands-based company that sells Interrail and Eurail passes for multi-country train travel across Europe, passes that are also available to young Europeans through the EU’s DiscoverEU program.

When it disclosed the incident in February, the company said the attackers gained access to travelers’ sensitive information, including full names, passport details, ID numbers, bank account IBANs, health information, and contact details (email addresses, phone numbers), after breaching its customer database.

Wiz

Eurail also warned at the time that the threat actors had published a sample of the stolen data on Telegram and were attempting to sell it on the dark web.

“The evidence showed that an unauthorized actor transferred files from our network on December 26, 2025,” the European train travel company said in breach notification letters sent to affected individuals on March 27.

Advertisement

“We reviewed the files involved and, on February 25, 2026, determined that they contained some of your information. The information included your name and passport number.”

The same day, Eurail revealed in a filing with the Office of Oregon’s Attorney General that the resulting data breach impacted 308,777 individuals.

Eurail data breach filing with Oregon's OAG
Eurail data breach filing with Oregon’s OAG (BleepingComputer)

​While Eurail said that it didn’t store financial information or passport photocopies on the compromised systems, the European Commission warned in a separate alert that this type of data (as well as health information) may have been exposed for young travelers who received a Pass through the DiscoverEU program.

Eurail told customers whose information was exposed in the breach to remain vigilant against potential phishing attacks and scams, and advised them to update their Rail Planner app account passwords and reset them on any other platform where they’re also used.

The company added that customers should monitor their bank account activity and report any suspicious transactions to their bank as soon as possible.

Advertisement

Last month, the European Commission also confirmed a data breach after the Europa.eu web platform was hacked in a cyberattack claimed by the ShinyHunters extortion gang.

Automated pentesting proves the path exists. BAS proves whether your controls stop it. Most teams run one without the other.

This whitepaper maps six validation surfaces, shows where coverage ends, and provides practitioners with three diagnostic questions for any tool evaluation.

Source link

Advertisement
Continue Reading

Tech

After Wi-Fi 7's Speed Push, Wi-Fi 8 Is Turning to Reliability

Published

on

Wi-Fi 8 is already taking shape, and while it won’t raise peak speeds beyond Wi-Fi 7, it promises something just as important: more reliable, lower-latency wireless performance where it actually matters.

Read Entire Article
Source link

Continue Reading

Tech

AMD finally puts a price tag on the Ryzen 9 9950X3D2 and it’s hefty

Published

on

AMD has locked in the price for its Ryzen 9 9950X3D2, and it lands high at $899. This is a flagship desktop chip aimed at people who rely on fast systems every day and don’t want to rebuild everything just to get there.

The processor introduces a dual 3D V-Cache setup, which AMD is using to push both gaming and heavy workloads forward at the same time. It also fits into the current AM5 ecosystem, so users with compatible boards and memory can upgrade without replacing the core of their system.

It goes on sale April 22, though there’s no detail yet on how widely it will be available at launch.

Who this $899 chip is for

At $899, this chip sits well outside the mainstream. AMD is going after creators, developers, and power users who notice slowdowns immediately and are willing to pay to avoid them.

Advertisement

The world’s first dual 3D V-Cache™ technology desktop processor.

AMD Ryzen™ 9 9950X3D2 Dual Edition processor

Available April 22 | $899

Workstation-class performance meets the AM5 platform, no new motherboard or memory required.

Built for developers and content creators… pic.twitter.com/rN4ysy45X6

Advertisement

— David McAfee (@McAfeeDavid_AMD) April 8, 2026

The pricing also signals where this part sits in the stack. It’s a top-tier option in the Ryzen 9000 lineup, built to prioritize sustained performance in demanding scenarios rather than broad affordability.

There’s a clear tradeoff here, where you’re paying more upfront to reduce waiting time during real work.

How dual V-Cache changes things

The dual 3D V-Cache design builds on AMD’s earlier work with stacked cache, but pushes it further. Instead of leaning mostly toward gaming gains, this version is meant to handle a wider mix of tasks without compromise.

Advertisement

That shift is important because earlier X3D chips often felt specialized. This one is positioned as more balanced, giving users who split time between games and production workloads a stronger reason to consider it.

Still, AMD hasn’t shared detailed performance figures in this material, so it’s not yet clear how much improvement shows up across different types of work.

Should you upgrade now

Compatibility is one of the more practical advantages here. The Ryzen 9 9950X3D2 works with existing AM5 motherboards and memory, which makes it a simpler upgrade for current users.

That helps take some pressure off the high asking price, especially if you’re already invested in the platform. Instead of planning a full rebuild, you can focus on swapping the processor and moving on.

With the April 22 release approaching, the decision comes down to whether you need the extra headroom now or can wait for more data.

Advertisement

Source link

Continue Reading

Trending

Copyright © 2025