Connect with us

Tech

Pete Hegseth Is Pushing Defense Employees to Volunteer With DHS

Published

on

The Department of Defense is putting more pressure on employees to volunteer to support the Department of Homeland Security’s immigration crackdown.

In a February 19 memo sent to civilians across the DOD, secretary of defense Pete Hegseth wrote that he expects “every supervisor to encourage their civilian employees to volunteer. Leadership must continue to promote this detail program and educate their civilian employees on its importance.” The memo, which was titled “Department of War Guidance to Encourage Support to the Department of Homeland Security Southern Border and Internal Immigration Enforcement Missions,” was sent to thousands of civilian DOD employees. The memo was first reported by GovExec and was also viewed by WIRED.

The instructions follow a June 2025 memo in which Hegseth authorized civilian employees to be detailed to DHS. But an Army civilian employee who spoke to WIRED on the condition of anonymity for fear of retaliation says that there is “definitely more pressure” now, “at least on the supervisory chain.”

The DOD and DHS did not respond to a request for comment.

Advertisement

“I received the obligatory announcement email with the first memo when it came out, and no one has talked about it at all, so much so that I had forgotten about it entirely,” says the Army civilian employee. “I don’t know anyone who has taken the job.” In a statement from August 2025, the DOD claims that “nearly 500 DoD civilians have signed up to participate and bring their skill sets to the border security and immigration enforcement mission at the participating DHS agencies.”

“While details and other short-term professional development opportunities are common for Army civilians, I have never heard of supervisors being REQUIRED to approve such details,” they say.

The employee noted that, as part of the Trump administration’s efforts to cut back on government jobs in the name of “efficiency,” Hegseth has sought to cut the department’s workforce. “I have taken up the duties of three departed colleagues on top of the job I was hired for as a result,” they say. This means it would be difficult for the department to lose anymore staff or for workers to step away from existing projects. The employee described this kind of request to volunteer for another federal agency as “very not common.” It’s not like the Defense Department has any spare time at the moment, either: Hegseth and DOD leadership are currently engaged in directing the US’s role in conflict with Iran.

DOD employees who want to volunteer to be detailed to DHS need to apply through USAJobs. According to the job posting, the Federal Emergency Management Agency (FEMA), which is part of DHS, will be reviewing applications. Volunteers will not only be sent to the southern border, but to “several ICE and CBP facilities throughout the interior of the United States.”

Advertisement

While some volunteer roles appear to be mundane tasks like “data entry,” others appear to be in the thick of immigration enforcement operations. These include assisting ICE and CBP in “developing concepts of operation and campaign plans to execute internal arrests and raids as well as patrols along the Southwest Border”; assisting ICE and CBP in “managing the physical flow of detained illegal aliens from arrest to deportation, as well as manage associated data”; and “managing the logistical planning to move law enforcement personnel, operational capabilities, and support equipment across the United States.”

The memo is just the latest in a series of changes across the federal government meant to enforce president Donald Trump’s immigration agenda. At the Department of Housing and Urban Development, a new rule would bar families with immigrant members from receiving certain forms of support from the agency, and at the General Services Administration, staff have been asked to assist ICE in procuring new physical spaces across the country.

Source link

Advertisement
Continue Reading
Click to comment

Leave a Reply

Your email address will not be published. Required fields are marked *

Tech

Daily Deal: The 2026 Complete Firewall Admin Bundle

Published

on

from the good-deals-on-cool-stuff dept

Transform your future in cybersecurity with 7 courses on next‑level packet control, secure architecture, and cloud‑ready defenses inside the 2026 Complete Firewall Admin Bundle. Courses cover IT fundamentals, topics to help you prepare for the CompTIA Server+ and CCNA exams, and more. It’s on sale for $25.

Note: The Techdirt Deals Store is powered and curated by StackCommerce. A portion of all sales from Techdirt Deals helps support Techdirt. The products featured do not reflect endorsements by our editorial team.

Filed Under: daily deal

Source link

Advertisement
Continue Reading

Tech

'X-Plane 12' flight simulator to take advantage of Nvidia CloudXR 6 in visionOS 26.4

Published

on

Apple Vision Pro owners will be able to take to the skies in the “X-Plane 12” flight simulator thanks to visionOS 26.4, which includes support for CloudXR.

Apple Vision Pro with two PSVR2 controllers arranged below it on a dark surface
Apple Vision Pro to get more gaming support via CloudXR

One of the main complaints surrounding Apple Vision Pro is its lack of a broad app ecosystem or VR game library. The latter could be addressed thanks to visionOS 26.4, and developers are already taking notice.
A post in the X-Plane blog shares that X-Plane 12 will launch on Apple Vision Pro “later this spring.” It seems to insinuate that it will launch alongside the release of visionOS 26.4, which includes support for the needed Nvidia CloudXR 6.0.
Continue Reading on AppleInsider | Discuss on our Forums

Source link

Continue Reading

Tech

Office.eu and the hope for a digitally sovereign Europe

Published

on

‘For many years, Europe has relied on American software and therefore created a certain risk of dependency’, said Office.eu’s CEO.

Much has happened over the past few years for Europe to decide to tighten its reigns on data in the region.

Examples include numerous comments from US president Donald Trump threatening action against the EU for its supposed discriminatory actions against US Big Tech, the possibility of the US government compelling US tech companies to share EU data, and not to mention the very same Big Tech companies dominating the region’s market while also violating its laws.

Microsoft 365 was found to have been tracking EU school students’ data illegally, Apple was found to be preventing EU users from exercising choice, and Meta was forcing users to pay or consent to its data tracking.

Advertisement

Last November, around 900 policymakers and other stakeholders gathered in Berlin to discuss ways to make Europe more technologically resilient, and less dependent on the US.

France and Germany took centre stage, announcing a joint taskforce on digital sovereignty, aiming for a more competitive and sovereign Europe.

“The Digital Sovereignty Summit sends a clear signal – Europe has what it takes to lead the digital age,” commented French president Emmanuel Macron at the time. “Europe is stepping up to accelerate the development of European innovation, to uphold strong data protection and to call for fair market conditions.”

The idea of digital sovereignty has finally taken centre stage in Europe, after years of the region conducting its business and administration on infrastructure made elsewhere.

Advertisement

It’s unsurprising then that a wholly European-owned alternative to Microsoft Office and Google Workspace, called Office.eu, made its debut last week, promising to enable organisations to regain control over their data and digital operations.

“We have seen more and more how essential it is to become cloud-independent and to rely on software that is built around European values,” said Maarten Roelfs, the CEO of Office.eu. The Hague-headquartered company was founded in 2024 and started operating early this year.

“For many years, Europe has relied on American software and therefore created a certain risk of dependency, but we have also given away the control over our own data. Office.eu proves that we now have a strong European alternative, with sovereignty, privacy and transparency at its core.”

Office.eu runs entirely on European data centres. Built on an open-source foundation, the platform offers a cloud drive, tools such as emails, spreadsheets, presentations and calendar, and video conference services.

Advertisement

Early access already has nearly 15,000 applicants, Office.eu told SiliconRepublic.com. A phased European rollout is planned for the second quarter.

Office.eu’s Office Suite is not the first of its kind, however. Infomaniak, which describes itself as an “ethical cloud with no bullshit”, launched ‘kSuite’ back in 2022, with a Microsoft Teams alternative called ‘kChat’, and the option to use Microsoft Office with documents hosted in Switzerland.

OpenDesk, created by the German Centre for Digital Sovereignty, offers similar tools for use by the German administration. There are others as well, such as NextCloud, which offers self-hosted file storage and a chat function.

“The Rubicon has been crossed. American tech firms can no longer offer assurances to European companies that their data sovereignty will be protected,” Roelfs told Stories of Purpose.

Advertisement

Don’t miss out on the knowledge you need to succeed. Sign up for the Daily Brief, Silicon Republic’s digest of need-to-know sci-tech news.

Source link

Advertisement
Continue Reading

Tech

Tony Hoare, Turing Award-Winning Computer Scientist Behind QuickSort, Dies At 92

Published

on

Tony Hoare, the Turing Award-winning pioneer who created the Quicksort algorithm, developed Hoare logic, and advanced theories of concurrency and structured programming, has died at age 92.

News of his passing was shared today in a blog post. The site I Programmer also commemorated Hoare in a post highlighting his contributions to computer science and the lasting impact of his work. Personal accounts have been shared on Hacker News and Reddit.

Many Slashdotters may know Hoare for his aphorism regarding software design: “There are two ways of constructing a software design: One way is to make it so simple that there are obviously no deficiencies, and the other way is to make it so complicated that there are no obvious deficiencies. The first method is far more difficult.”

Source link

Advertisement
Continue Reading

Tech

Nvidia and Thinking Machines Lab draw multi-year chip deal

Published

on

Nvidia has also made a ‘significant’ investment in Thinking Machines to support its long-term growth.

In a new multi-year partnership announced today (10 March), Thinking Machines Lab will be using Nvidia’s systems to train its AI frontier models. Deployment on Nvidia’s Vera Rubin platform is set to begin early next year.

Mira Murati launched her AI start-up Thinking Machines in early 2025, less than six months after quitting as OpenAI’s chief technology officer. By July 2025, the start-up had already hit $12bn in value after a $2bn seed round, all before it released its first product.

The start-up’s flagship product ‘Tinker’ is a training API for fine-tuning open-source models, designed for researchers and developers. Tinker can be used for a broad range of models, including Llama-3.2B, Qwen3.5, DeepSeek-V3.1 and Kimi-K2.

Advertisement

“Thinking Machines has brought together a world-class team to advance the frontier of AI,” said Nvidia CEO and founder Jensen Huang.

Huang has previously said that 1GW of AI data centre capacity costs up to $50bn, which would place the latest deal’s value to several billion dollars. In addition, the company has also made a “significant” investment in Thinking Machines to support its long-term growth.

“Nvidia’s technology is the foundation on which the entire field is built,” said Murati, Thinking Machines’ CEO. “This partnership accelerates our capacity to build AI that people can shape and make their own, as it shapes human potential in turn.”

This is the latest in a long range of supports Nvidia has been providing to fuel the AI race. The $4.4trn chipmaking giant participated in a freshly announced $1.03bn seed round into Advanced Machine Intelligence, supported Nscale in its $2bn Series C raise, and pumped $30bn into OpenAI (possibly its last time before OpenAI goes public).

Advertisement

Last month, Meta promised Nvidia billions of dollars on a multi-year partnership to support its data centre build-out. Meanwhile, Nvidia has been looking ahead to ensure AI-ready 6G infrastructure with a new partnership with telecommunication giants from around the globe.

Don’t miss out on the knowledge you need to succeed. Sign up for the Daily Brief, Silicon Republic’s digest of need-to-know sci-tech news.

Source link

Advertisement
Continue Reading

Tech

Trump Administration Won’t Rule Out Further Action Against Anthropic

Published

on

At Anthropic’s first court hearing challenging sanctions imposed by the Trump administration, the AI tech startup asked the government to commit that it wouldn’t levy additional penalties on the company. That didn’t happen.

“I am not prepared to offer any commitments on that issue,” James Harlow, a Justice Department attorney, told US district judge Rita Lin over video conference on Tuesday.

In fact, the government is gearing up to take another step designed to sideline the company from doing business with federal agencies. President Trump is currently finalizing an executive order that would formally ban usage of Anthropic tools across the government, according to a person at the White House familiar with the matter but not authorized to discuss it. Axios first reported on the plan.

Tuesday’s hearing stemmed from one of the two federal lawsuits Anthropic filed against the Trump administration on Monday, alleging that the government unconstitutionally designated it a supply-chain risk and turned it into a tech industry pariah. Billions of dollars in revenue for Anthropic is now at risk, with current customers and prospective ones dropping out of deals and demanding new terms, according to the company.

Advertisement

Anthropic seeking a preliminary court order suspending the risk designation and barring the administration from taking further punitive measures against the company.

The court appearance on Tuesday was to decide on the schedule for a preliminary hearing, and Anthropic is eager for it to happen soon to prevent further harm to its business. Michael Mongan, an attorney for Anthropic at WilmerHale, told Lin he was less concerned about delaying it until April if the Trump administration could commit to not taking additional action. “The actions of defendants are causing irreparable injuries, and those injuries are mounting day by day,” Mongan said.

After Harlow declined, Lin moved up the date of the hearing to March 24 in San Francisco, though that timeline was still later than Anthropic wanted. “The case is quite consequential from both sides, and I want to make sure I’m deciding on an expedited record but also a full record,” the judge said.

Scheduling in the other case, which is in Washington, DC, is on hold while Anthropic pursues an administrative appeal to the Department of Defense, which is expected to fail on Wednesday.

Advertisement

The months-long dispute between the Pentagon and Anthropic began when the AI startup refused to sign off on its current technologies being used by the military for any lawful purpose, which it fears could include broad surveillance of Americans and the launch of missiles without human supervision. The Defense Department contends usage decisions are its prerogative.

Several attorneys with expertise in government contracts and the US Constitution believe the administration’s action against Anthropic continues a pattern of abusing the law to punish perceived political enemies, including universities, media companies, and law firms (such as WilmerHale, the firm representing Anthropic). The experts believe Anthropic should prevail, but the challenge will be overcoming the deference that courts often give to national security arguments from the government, especially during times of war.

“If this is a one-off, you might give the president some deference,” says Harold Hongju Koh, a Yale Law School professor who worked in the Barack Obama presidential administration and has written about the Anthropic case. “But now, it’s just unmistakable that this is just the latest in a chain of events related to a punitive presidency.”

David Super, a Georgetown University Law Center professor who studies the constitution, says the provisions the Defense Department used to sanction Anthropic were designed to protect the country from potential sabotage by its enemies.

Advertisement

Source link

Continue Reading

Tech

Human Problems: It’s Not Always The Technology’s Fault

Published

on

from the it-prevents-us-from-fixing-societal-issues dept

We have met the enemy and he is us.

When a teenage boy in Orlando started texting Character.AI’s chatbot, it started as an innocent use of a new tool. Sewell Setzer III customized the chatbot to have the Game of Thrones-inspired persona of Daenerys Targaryen, the series’ prominent dragon-riding queen. In the months that followed, the boy developed a romantic connection with the chatbot. One night, he messaged the bot: “What if I told you I could come home right now?” The bot sent back, “[P]lease do, my sweet king.” Setzer was only fourteen years old when he died by suicide later that evening.

Setzer’s death is a tragedy. Like many parents in the wake of suicide, Seltzer’s mother is left searching for answers and accountability. Suicide often leaves behind a painful void, filled with questions that rarely yield satisfying explanations. 

In her search, Setzer’s mother sued the chatbot’s developer, Character Technologies, alleging that its chatbot caused her son’s death. The complaint describes the bot as a “defective” and “inherently dangerous” technology, and accuses the company of having “engineered Setzer’s harmful dependency on their products.” She is not alone. Three other families have brought similar suits against Character Technologies, and another has sued OpenAI, alleging the chatbots harmed their children.

Advertisement

Framing suicide and other harms as technology problems—as much of the current discourse around chatbots suggests—obscures underlying societal conditions and can undermine effective interventions. In effect, what are often described as “tech problems” are, more accurately, the result of human decisions, norms, and policies. They are, at their core, human problems. 

Historical Framing of Tech and Media in Creating and Sustaining Societal Problems

This is just the latest vintage whine, rebottled yet another time. Humanity has long sought to condemn new technologies and media for problems of the day. When the printing press made literature available to the masses, church and state condemned publications for causing immorality. Rock ‘n’ Roll and comic books were blamed for juvenile delinquency. Later, it was heavy metal and role-playing games. The advent of video games supposedly led to increased violence by adolescent boys.

The desire to hold technology companies responsible for human harms, however, has its immediate antecedent in social media. Over the past decade, users have sued social media platforms for offline violence committed by people they met online, failing to prevent cyberbullying, and hosting user-generated content that allegedly radicalized extremists. 

Advertisement

Like in Setzer’s case, parents have also sued social media companies after the deaths of their children, arguing that design choices, engagement mechanics, and algorithmic targeting played a role. Indeed, this is the central question at the heart of the current wave of “social media addiction” litigation that is currently being tried.

AI is just the latest technological scapegoat to which we seek to ascribe fault. It’s easier to hold technology responsible for our problems, especially when the technology is as uncanny as generative AI. We’re afraid of robots, perhaps not because of any harm they cause us, but because they show us how much we, as humanity, can harm ourselves. We would rather fault the technology du jour than confront the harder truths underneath. 

Death by Suicide as a Case Study

To put this into context, consider the allegations about the Character.AI chatbot and Setzer’s suicide. Suicide is a complex, deeply human problem. Among youth and young adults, it stands as the second leading cause of death. Suicide has no single cause. Public health experts have long recognized that risk emerges from a convergence of individual, relational, communal, and societal factors. These can include long-term effects of childhood trauma, substance abuse, social isolation, relationship loss, economic instability, and discrimination. On the surface, these may look like personal struggles, but they’re really the fallout of systemic failure. 

Advertisement

Access to lethal means compounds the risk of self-harm and suicide. In particular, the presence of firearms in the home has remained strongly associated with higher youth suicide rates. 

These systemic failures tend to hit teens the hardest. Studies consistently show that young people are facing rising rates of mental health challenges, especially due to and following the COVID-19 pandemic. This is compounded by chronically underfunded school counseling programs, inaccessible mental health care, and inconsistent support for youth in crisis. LGBTQ+ youth, in particular, bear the brunt, facing higher rates of bullying, depression, and suicidal ideation, all while increasingly being targeted by state policies that strip away protections and deny their identities. 

We don’t and can’t know for sure why Setzer or anyone else died by suicide. Tragically, teenage suicide is common. Indeed, it’s the subject of many songs. There’s no mechanism to definitively determine how Setzer and other victims felt when they started using Character.AI. However, as we likely all remember from our own lives, teenage years can be trying. As we mature physically and mentally, it can be difficult to express and accept ourselves. Other children can be cruel. Hormones can lead us to lash out in anger and withdraw into ourselves. 

In Setzer’s case, the complaint and public reporting indicate that he exhibited other signs and conditions commonly associated with elevated suicide risk, including anxiety and depression, withdrawal from teachers and peers, chronic lateness, significant sleep deprivation, and access to a firearm in the home. His interactions with fictional characters on the Character.AI service may suggest unmet emotional needs or a search for understanding and connection. At different points, he described a character as resembling a father figure and spoke about feelings of loneliness and a lack of romantic connection—experiences that are not uncommon for adolescents, particularly during periods of heightened vulnerability. According to the complaint, Setzer also raised the topic of suicide in earlier conversations with the chatbot, and those exchanges were promptly halted by the system. 

Advertisement

The uncomfortable truth about suicide is that it has existed as long as there have been people–sometimes for reasons we can understand, and often for reasons we never will. We are terrified that people die by suicide, not only because it is difficult to comprehend, but because the forces that drive someone there can feel disturbingly familiar.

Parents like Setzer’s can’t fix systemic governmental and societal failures. What feels more immediate and actionable is holding the technology companies accountable when their services appear to enable or amplify harm. It is far easier to fixate on the medium through which people express suicidal thoughts rather than ask where those thoughts came from or why they felt like the only option.

Legal Analysis of Faulting Tech

Legal doctrine appears to recognize that holding the technology responsible for these systemic failures is not viable. For example, because suicide is shaped by so many overlapping factors, tort claims against AI companies for causing a teen’s death—while understandable in their urgency—are, doctrinally speaking, a stretch. 

Advertisement

Under traditional tort principles, providers of generative AI systems and social media services are unlikely to bear legal responsibility in these cases. Claims based on intentional torts, such as battery, generally fail because providers of online services do not act with the intent to cause—or even to contribute to—physical harm. Therefore, Plaintiffs more commonly turn to negligence theories.

Negligence, however, requires more than just harm in fact. It demands both factual causation and proximate (i.e., legal) causation. In some situations, an online service or generative AI model might satisfy a but-for test because the harm would not have occurred without the service. But that is not sufficient. 

Proximate cause—what the law treats as a legally meaningful connection between conduct and injury—is where most of these claims falter. In many cases, particularly those involving such numerous and complex factors as suicide, the link between a provider’s conduct and the ultimate injury is typically too attenuated to meet this standard. 

Services such as social media and AI chatbots are typically designed as broad, general-purpose tools. The potentially implicatable content comes from other users’ behaviors, personalized interactions, or the user’s own actions. Even where excessive technology use—including social media—has been associated with elevated rates of suicidal ideation among youth and young adults, research has not established a direct causal link. As a result, courts are generally reluctant to find the technology service to be the legal cause of death. 

Advertisement

The Broader Ramifications of a Myopic Focus on Tech

Beyond legal error, focusing solely on technology obscures the path to real solutions. When we frame fundamentally human problems as technological ones, we deflect attention from the underlying conditions that lead to these tragedies and make it more likely they will recur. 

This framing guides policymakers and advocates toward seemingly easy, surface-level technological fixes such as imposing age-verification requirements, mandating disclosures about content moderation, or curbing algorithmic feeds. True, technology companies can—and should—consider how to help mitigate real-world harms. Yet these proposed interventions rest on the assumption that technology is the primary culprit, even though research increasingly shows that, in the right contexts, technology can actually help those in crisis. 

The appeal of reducing complex social issues to matters of redesigning or banning technology is understandable. Technology problems can feel tractable. They suggest clear targets and concrete fixes. 

Advertisement

What this logic ignores, however, is that the pre-technology status quo for many public health crises has long been dismal. The better question, then, isn’t whether technology causes harm, but whether it deepens an already broken baseline—or simply reflects it.

Technology, including generative AI, often acts less as a cause than a mirror. Our digital spaces often reflect the offline world, including its ills. 

Today, children face more pressure to excel at school and attend the best universities, even while job prospects stagnate and inflation soars. They have lost access to the kinds of public and community spaces that once offered structure, connection, and care. Libraries operate with reduced hours. Budget cuts have decimated after-school programs. Parks are monitored and restricted for loitering. Community centers that shuttered during the pandemic have never reopened. In many ways, technology—and social media in particular—has stepped in as a makeshift third space for teens. Yet rather than address the erosion of offline support, policymakers are now working to dismantle these digital communities too.

If human distress reflects deteriorating real-world social infrastructure, then optimizing digital services cannot restore stasis. Technological interventions address a symptom while the deeper human cancer persists.

Advertisement

A Pragmatic Path Forward

The path forward requires resisting the impulse to treat fundamentally human problems as technological ones. When new technologies appear alongside harm, the harder and more necessary questions are not simply how to regulate the tool, but what human choices produced the conditions in which harm emerged, which institutions failed or fell short, and what values should guide our response. These questions are more difficult—and often more uncomfortable—because they turn our attention inward, toward ourselves, rather than external and more convenient actors.

Instead of focusing our energies on systematically regulating platforms, we should direct our efforts toward these human problems. For suicide, public health experts point to a wide range of evidence-based strategies for preventing and mitigating risk factors. These include strengthening economic supports such as household financial stability and housing security; creating safer environments by reducing at-risk individuals’ access to lethal means; fostering healthy organizational policies and cultures; and improving access to healthcare by expanding insurance coverage for mental health services and increasing provider availability and remote access in underserved areas. Experts also emphasize the importance of promoting social connection and teaching problem-solving skills that can help individuals navigate periods of acute distress.

These and other socioeconomic reforms are not easy solutions. They aren’t just a matter of adjusting algorithms or restricting platform features. They demand uncomfortable conversations about how we structure work, education, and community life. They require sustained political commitment and resource allocation. Yet if we can achieve these results, we will create a better world than one derived from mere technological fixes.

Advertisement

In short, technology doesn’t cause suicide. It doesn’t cause a host of human problems for which it is often accused. Sadly, they have always been with us. 

But technology, used wisely, could help us mitigate these problems. For example, through processing massive amounts of data, AI can detect patterns that elude us humans. This alone could help reveal early warning signs or surface new protective factors. AI chatbots, for example, could help us identify teens who are at risk and create opportunities to intervene. 

But that kind of progress demands that we take responsibility for these problems. We must acknowledge that our governments, societies, communities, and even ourselves may have normalized and contributed to these harmful conditions. We may discover there’s no rhyme or reason to why teenagers commit suicide. But we may uncover that teen suicide isn’t random at all. It may stem from something we’ve unwittingly ignored, or perhaps built into the world. 

That possibility is far more unsettling than the idea of dangerous technology. It’s the idea that the danger might be us.

Advertisement

Kevin Frazier directs the AI Innovation and Law Program at the University of Texas School of Law and is a Senior Fellow at the Abundance Institute. Brian L. Frye is a Spears-Gilbert Professor of Law at the University of Kentucky J. David Rosenberg College of Law. Michael P. Goodyear is an Associate Professor at New York Law School. Jess Miers is an Assistant Professor of Law at The University of Akron School of Law

Filed Under: ai, blame, moral panic, negligence, proximate cause, suicide, tech, tort law

Source link

Advertisement
Continue Reading

Tech

Stephen Thaler’s Legendary AI Copyright Losing Streak Ends With Nowhere Left To Appeal

Published

on

from the public-domain-ftw dept

We’ve been covering Stephen Thaler’s quixotic quest to get copyright (and patent) protection for works generated entirely by his AI system “DABUS” for years now. If there’s one thing Thaler has proved beyond all reasonable doubt, it’s that you can be comprehensively, thoroughly, and repeatedly wrong at every level of the American legal system and still keep going. He loses everywhere, every time, at every level. The Copyright Office rejected him. A federal district court rejected him. The DC Circuit rejected him. The Patent Office rejected him. Courts rejected his parallel patent claims. Even the Trump administration—not exactly known for its nuanced intellectual property positions—told the Supreme Court not to bother hearing his appeal.

And now, the Supreme Court has declined to take up the case, putting the final period on what has been one of the most impressive losing streaks in recent IP law history.

Plaintiff Stephen Thaler had appealed to the justices after lower courts upheld a U.S. Copyright Office decision that the AI-crafted visual art at issue in the case was ineligible for copyright protection because it did not have a human creator.

That was always the fatal flaw with his argument. He wasn’t making the more nuanced claim that a human who uses AI as a tool should get some copyright protection. He was making the maximalist claim: the AI did it all by itself, and it (or rather, he, as the AI’s owner) should get the copyright anyway.

The image in question—”A Recent Entrance to Paradise,” of train tracks entering a portal surrounded by green and purple plant-like imagery—was, according to Thaler, created entirely by DABUS with no human creative input. Every single institution that looked at this said no.

Advertisement

A federal judge in Washington upheld the office’s decision in Thaler’s case in 2023, writing that human authorship is a “bedrock requirement of copyright.” The U.S. Court of Appeals for the District of Columbia Circuit affirmed the ruling in 2025.

Thaler’s lawyers, for their part, tried to argue that the stakes were too high for the Court to sit this one out:

With a refusal by the court to hear the appeal, Thaler’s lawyers said, “even if it later overturns the Copyright Office’s test in another case, it will be too late. The Copyright Office will have irreversibly and negatively impacted AI development and use in the creative industry during critically important years.”

That’s rich. The Copyright Office is already working through the genuinely harder questions in cases involving tools like Midjourney—cases where humans actually did have meaningful creative input. Those cases are moving through the system right now. The problem for Thaler is that he chose the worst possible vehicle to force a Supreme Court showdown: a case so maximalist in its claims (the AI did everything, humans did nothing, give us the copyright anyway) that courts could rule against him on the narrowest possible grounds without ever having to engage with the nuanced questions at all. His all-or-nothing bet made this an easy case.

Still, the question of what happens when a human uses AI as a creative tool—rather than letting the machine do everything—isn’t actually as novel or unsettled as many people seem to think.

Copyright law has required human creative choices since at least Burrow-Giles Lithographic Co. v. Sarony all the way back in 1884 in a case about whether or not photographs get covered by copyright. And the wonderful Feist Publications v. Rural Telephone Service from 1991 (a case we cite often) hammered the point home by establishing that copyright demands original creative expression. Consider how this already works with photography. A photographer who frames a shot of a landscape gets copyright protection in the creative choices they made—the composition, the angle, the timing, the lighting. But the landscape itself? No human created that. It gets no copyright. The camera mechanically captured what was in front of it, but the human’s original creative decisions (and only those original creative decisions) are what copyright protects.

Advertisement

AI-generated works should work roughly the same way. If a human’s creative input—through a sufficiently specific and expressive prompt, through selection and arrangement, through iterative creative choices—meaningfully shapes the output, that human contribution can be protected. But the parts that the AI generated autonomously, without human creative direction? Those are “the landscape.” They’re the thing no human authored.

There will certainly be disputes at the margins about exactly how much human input is enough, and where the line sits between “I told the AI to make something cool” and genuine creative direction. But the fundamental framework for handling this already exists. We’ve been here before with every new creative tool, from cameras to Photoshop. The principle has always been the same: copyright protects human creativity, regardless of the tool used to express it.

Thaler chose to fight for the one position that had no support in law (or in common sense). His losing streak is now complete, and there’s nowhere left to appeal. But the legacy of his many, many losses is actually kind of useful: he has, through sheer persistence, generated an incredibly clear and consistent body of authority establishing that purely AI-generated works, with no human creative input, do not get copyright protection.

So, thanks for that, I guess. Oh, and I guess we can confidently post that “Recent Entrance to Paradise” image as it, like the monkey selfie before it, is officially in the public domain.

Advertisement

Filed Under: ai, copyright, copyright office, copyrightable subject matter, dabus, human creativity, stephen thaler, supreme court

Source link

Continue Reading

Tech

HiFiMAN Arya WiFi and HE1000 WiFi First Impressions at CanJam NYC 2026: Are Wireless Planar Headphones Finally Ready for Audiophiles?

Published

on

For years, the biggest challenge in the wireless headphone space hasn’t been convenience or features. It’s been performance. High-end brands have spent the past decade refining flagship wired models that rely on dedicated headphone amplifiers and carefully tuned drivers, and translating that level of performance into a self contained wireless design is not a trivial exercise. That’s the context for the HiFiMAN Arya WiFi and HiFiMAN HE1000 WiFi, two open back wireless planar magnetic headphones introduced at CanJam NYC 2026.

HiFiMAN took its time with these for a reason. Integrating the necessary wireless electronics, amplification, DAC architecture, and battery systems into both earcups and the headband, while still preserving the acoustic character of the award winning wired Arya and HE1000 was not going to happen overnight.

HiFiMAN has experimented with wireless concepts before, but the Arya WiFi and HE1000 WiFi represent its most serious attempt yet to bring high bandwidth wireless audio to planar magnetic headphones.

Both models still support Bluetooth with codecs like aptX HD and LDAC, and they can function as USB Audio devices when connected directly. But those options are really just the supporting cast.

Advertisement

The real story here is WiFi. That’s where the bandwidth lives, and where HiFiMAN believes wireless listening can finally approach the kind of fidelity that planar magnetic headphones are known for.

hifiman-he1000-wifi-headphones-lifestyle
HiFiMAN HE1000 WiFi

Hymalaya R2R DAC: A Ladder Inside the Headphones

Both the HiFiMAN Arya WiFi and HiFiMAN HE1000 WiFi incorporate HiFiMAN’s proprietary Hymalaya R2R DAC, along with integrated amplification inside the earcups and planar magnetic drivers derived from the company’s established designs. That combination turns each headphone into something closer to a complete playback system rather than a passive transducer waiting for a DAC and amplifier to do the heavy lifting.

The Hymalaya DAC is built around a classic R2R ladder architecture, a design long favored in high end audio for its precise timing and direct signal conversion. Instead of relying on the delta sigma DAC chips found in most wireless headphones, an R2R DAC converts digital audio through a network of precision resistors that translate binary data directly into analog voltage. The upside can be excellent transient response and tonal accuracy. The downside is that traditional ladder DACs tend to be large, complex, and hungry for power.

HiFiMAN’s solution was to rethink the architecture from the ground up. The Hymalaya platform combines the R2R ladder with FPGA control and extremely low power consumption, allowing the company to shrink the circuit dramatically while maintaining support for high resolution formats. What makes this particularly impressive is the scale. Each earcup houses a compact DAC stage containing hundreds of precision resistors, carefully matched and integrated alongside the internal amplification and wireless electronics.

Advertisement
chris-boylan-hifiman-he1000-wifi-headphones-canjam-nyc-2026
Editor-at-Large, Chris Boylan wearing HiFiMAN HE1000 WiFi Headphones at CanJam NYC 2026

Packing that level of circuitry into both sides of a headphone while keeping noise low, power consumption manageable, and heat under control is not trivial. From an engineering standpoint alone, this is one of the more ambitious wireless headphone designs we encountered at CanJam NYC 2026.

At a private post-show gathering on Saturday evening, HiFiMAN explained that developing a product like the HiFiMAN HE1000 WiFi or HiFiMAN Arya WiFi can take 12 to 18 months, even when starting with established passive designs.

The company chose the HE1000 and Arya platforms for two key reasons. First, both models are built on proven driver technologies, which reduces the number of variables when integrating the DAC, amplification, wireless circuitry, and battery systems directly into the headphone. Second, these sit at the higher end of HiFiMAN’s lineup, where production volumes are naturally lower. That allows the company to maintain tighter control over quality and consistency while manufacturing the smaller batches required for a design this complex.

Advertisement. Scroll to continue reading.

HiFiMAN also made it clear that this technology will not immediately trickle down to more affordable models. A Sundara-based WiFi version is not on the roadmap anytime soon, and neither is a version with active noise cancellation. The company’s view is that both would need to meet the same performance standards relative to their price points before they would consider bringing them to market with so much established competition.

Advertisement

You May Want to Sit Down for the Price

The HiFiMAN HE1000 WiFi isn’t cheap. Not even close.

The flagship of HiFiMAN’s new WiFi headphone lineup carries an expected retail price of $2,699, placing it squarely in the same territory as many high-end wired reference headphones.

That number shouldn’t come as a total shock when you consider what’s inside. The HE1000 WiFi is essentially a self contained planar magnetic system with its own DAC, amplification, wireless receiver, battery system, and internal signal chain built directly into the headphone. In other words, it’s doing the job that normally requires a DAC, headphone amplifier, and source component sitting on your desk.

Still, there’s no getting around the sticker shock. Spending nearly three grand on wireless headphones will raise eyebrows even in the audiophile community. But compared to what it would cost to assemble a comparable desktop system for the wired HiFiMAN HE1000 lineage, the math starts to look a little less insane.

Advertisement

And if that price feels steep, the HiFiMAN Arya WiFi exists as the slightly less punishing option at around $1,449, though “budget” isn’t exactly the word most people would use here either. Focal and DALI have options around the same price that we’ve already reviewed for those interested in closed-back options.

hifiman-arya-wifi-headphones-canjam-nyc-2026
HiFiMAN Arya WiFi Headphones at CanJam NYC 2026

Two Ways to Go Wireless: HiFiMAN HE1000 WiFi and Arya WiFi

The HE1000 WiFi and Arya WiFi are built on the same basic premise, but they are not the same headphone with different price tags slapped on the box.

Both are open back wireless planar magnetic designs with WiFi as the primary connection, backed up by Bluetooth with LDAC and aptX HD and USB Audio support. Both also integrate HiFiMAN’s Hymalaya R2R DAC, internal amplification, and battery powered electronics directly into the headphone. That alone is unusual. Most wireless headphones still lean on Bluetooth and off the shelf chipsets. HiFiMAN decided to build something far more ambitious and a lot more complicated.

Where they begin to separate is with the driver platform.

The HE1000 WiFi is based on the company’s more upscale Nano Diaphragm driver architecture, paired with Stealth Magnet technology to reduce wave diffraction and preserve a cleaner path for the soundwave. It feels like the more luxurious and visually striking of the two, and frankly, at this price it had better. Chris Boylan and I both found it very manageable on the head. Not super lightweight, but certainly not some neck killing science project either. Clamping force felt right, and the changes to the suspension headband, yoke, and internal cable routing were very well executed.

Advertisement

The Arya WiFi uses a similar overall layout, but swaps in HiFiMAN’s Super Nano diaphragm driver, a thinner planar diaphragm intended to improve transient response and efficiency. It still uses Stealth Magnet technology, integrated amplification, and the Hymalaya DAC, but it is positioned as the more accessible entry into this new WiFi based range. Same core concept. Different driver implementation. Lower price of admission. Slightly less “you may need to explain this purchase to your family” energy.

Both also offer something you won’t find on the HE1000 Unveiled or Arya Unveiled: protective grilles in front of the drivers. HiFiMAN explained that because these headphones may be used in a wider range of real world environments, they needed to protect the drivers in a way the more purist home oriented wired models do not. That makes sense. Open back headphones this expensive do not need extra ways to get murdered.

Advertisement. Scroll to continue reading.

What also deserves real credit is the comfort and mechanical design. Because both earcups contain the DAC, amplification stage, wireless electronics, and battery systems, HiFiMAN had to rethink how the internal cabling runs through the headband and how the added hardware would affect overall balance and long term wear.

Advertisement

That could have gone badly. It didn’t.

The revised suspension headband, yoke structure, and internal cable routing appear to have been carefully engineered so the additional electronics don’t create pressure points or imbalance across the headband. Both Chris Boylan and I tried them and while neither headphone is super lightweight, they felt very manageable. Clamping force is well judged and the weight distribution works far better than one might expect from a headphone with this much hardware built into both earcups.

Visually, the HE1000 WiFi in particular is rather gorgeous. At nearly $2,700, it probably needs to be.

Setup is also a little more involved than typical Bluetooth pairing. We were given a quick walkthrough at the show, and while there are a few extra steps, the payoff is immediately noticeable. The WiFi connection offers significantly higher sound quality than Bluetooth, which is the entire point of the design.

Advertisement
HiFiMAN Arya WiFI and HE1000 WiFi Headphones
HiFiMAN Arya WiFi (left) and HE1000 WiFi (right) Headphones

So How Do they Actually Sound?

If you’re familiar with the HiFiMAN HE1000 Unveiled or HiFiMAN Arya Unveiled, their wireless siblings get you a sizable percentage of the way there. That’s impressive on its own, but it also needs a bit of context.

The Wi-Fi environment at the hotel in Times Square during CanJam NYC 2026 was far from ideal. Thousands of attendees were hammering the network all day streaming music and uploading content, and even 45 floors above the city the connection on our smartphones and tablets was inconsistent at best.

In other words, what we heard was likely only a glimpse of what these headphones are capable of under more controlled conditions.

My personal experience with the HE1000 Unveiled is that it rewards better electronics. Feed it a strong amplifier and a good DAC and it delivers a well balanced, near neutral presentation. There’s enough life and energy to keep things engaging without tipping into the sterile or analytical. Sub bass isn’t exactly seismic, but unless you live exclusively in EDM territory, it feels honest and realistic for a planar magnetic design.

Advertisement

The HE1000 WiFi gets surprisingly close to that presentation. The sub bass won’t rattle your fillings, but the pacing, detail, and overall resolution are very solid. Vocals come across clean and natural, and the soundstage is large and well organized with precise imaging.

Compared to something like the Focal Bathys MG, the character is very different. The Bathys MG is punchier and more forward, designed to sound impressive right out of the gate. The HiFiMAN approach leans more toward spaciousness, neutrality, and detail retrieval, which will feel far more familiar to anyone who spends time with open back planar headphones.

In other words, these still sound like HiFiMAN headphones. The surprise is how much of that DNA survived the jump to WiFi.

Advertisement. Scroll to continue reading.
Advertisement

Reviews of both coming in April.

Source link

Advertisement
Continue Reading

Tech

In a vote of confidence for Meta’s Threads, Kalshi adds sharing feature

Published

on

Prediction market Kalshi is making it easier for its users to have conversations on Meta’s social network Threads. Kalshi now offers a share option that will automatically embed the relevant prediction market chart into a Threads post.

Whether people want to discuss who’s going to win Best Picture or which reality TV contestant is going to go home (and possibly bet on the outcome on Kalshi), “with this integration, people can share their opinions alongside the forecasts they’re seeing on Kalshi,” the company said in a blog post.

It’s a move that echoes a successful social media strategy for both Kalshi and its biggest rival, Polymarket, on X. However, things have gotten complicated for Kalshi on X recently. In June, X named Polymarket as its “official” prediction market partner.

Last month, Kalshi removed its affiliate badges from X accounts run by its sponsored traders. This came after X enacted a policy that prohibits sponsored accounts from posting about sports betting. That policy was adopted after the prediction markets were reportedly busted for partnering with fake sports insiders who spread misinformation.

Advertisement

While the share button is not as significant as what the prediction markets have with X, it is a vote of confidence in the X rival just a couple of months after user data appeared to show Threads growing faster than X.

Source link

Continue Reading

Trending

Copyright © 2025