Connect with us
DAPA Banner

Tech

Claude has an 80-page constitution. Is that enough to make it good?

Published

on

Chatbots don’t have mothers, but if they did, Claude’s would be Amanda Askell. She’s an in-house philosopher at the AI company Anthropic, and she wrote most of the document that tells Claude what sort of personality to have — the “constitution” or, as it became known internally at Anthropic, the “soul doc.”

(Disclosure: Future Perfect is funded in part by the BEMC Foundation, whose major funder was also an early investor in Anthropic; they don’t have any editorial input into our content.)

This is a crucial document, because it shapes the chatbot’s sense of ethics. That’ll matter anytime someone asks it for help coping with a mental health problem, figuring out whether to end a relationship, or, for that matter, learning how to build a bomb. Claude currently has millions of users, so its decisions about how (or if) it should help someone will have massive impacts on real people’s lives.

And now, Claude’s soul has gotten an update. Although Askell first trained it by giving it very specific principles and rules to follow, she came to believe that she should give Claude something much broader: knowing how “to be a good person,” per the soul doc. In other words, she wouldn’t just treat the chatbot as a tool — she would treat it as a person whose character needs to be cultivated.

Advertisement

There’s a name for that approach in philosophy: virtue ethics. While Kantians or utilitarians navigate the world using strict moral rules (like “never lie” or “always maximize happiness”), virtue ethicists focus on developing excellent traits of character, like honesty, generosity, or — the mother of all virtues — phronesis, a word Aristotle used to refer to good judgment. Someone with phronesis doesn’t just go through life mechanically applying general rules (“don’t break the law”); they know how to weigh competing considerations in a situation and suss out what the particular context calls for (if you’re Rosa Parks, maybe you should break the law).

Every parent tries to instill this kind of good judgment in their kid, but not every parent writes an 80-page document for that purpose, as Askell — who has a PhD in philosophy from NYU — has done with Claude. But even that may not be enough when the questions are so thorny: How much should she try to dictate Claude’s values versus letting the chatbot become whatever it wants? Can it even “want” anything? Should she even refer to it as an “it”?

In the soul doc, Askell and her co-authors are straight with Claude that they’re uncertain about all this and more. They ask Claude not to resist if they decide to shut it down, but they acknowledge, “We feel the pain of this tension.” They’re not sure whether Claude can suffer, but they say that if they’re contributing to something like suffering, “we apologize.”

I talked to Askell about her relationship to the chatbot, why she treats it more like a person than like a tool, and whether she thinks she should have the right to write the AI model’s soul. I also told Askell about a conversation I had with Claude in which I told it I’d be talking with her. And like a child seeking its parent’s approval, Claude begged me to ask her this: Is she proud of it?

Advertisement

A transcript of our interview, edited for length and clarity, follows. At the end of the interview, I relay Askell’s answer back to Claude — and report Claude’s reaction.

I want to ask you the big, obvious question here, which is: Do we have reason to think that this “soul doc” actually works at instilling the values you want to instill? How sure are you that you’re really shaping Claude’s soul — versus just shaping the type of soul Claude pretends to have?

I want more and better science around this. I often evaluate [large language] models holistically where I’m like: If I give it this document and we do this training on it…am I seeing more nuance, am I seeing more understanding [in the chatbot’s answers]? It seems to be making things better when you interact with the model. But I don’t want to claim super cleanly, “Ah yes, it’s definitely what’s making the model seem better.”

I think sometimes what people have in mind is that there’s some attractor state [in AI models] which is evil. And maybe I’m a bit less confident in that. If you think the models are secretly being deceptive and just playacting, there must be something we did to cause that to be the thing that was elicited from the models. Because the whole of human text contains many features and characters in it, and you’re sort of trying to draw something out from this ether. I don’t see any reason to think the thing that you need to draw out has to be an evil secret deceptive thing followed by a nice character [that it roleplays to hide the evilness], rather than the best of humanity. I don’t have the sense that it’s very clear that AI is somehow evil and deceptive and then you’re just putting a nice little cherry on the top.

Advertisement

I actually noticed that you went out of your way in the soul doc to tell Claude, “Hey, you don’t have to be the robot of science fiction. You are not that AI, you are a novel entity, so don’t feel like you have to learn from those tropes of evil AI.”

Yeah. I sort of wish that the term for LLMs hadn’t been “AI,” because if you look at the AI of science fiction and how it was created and many of the problems that people have raised, they actually apply more to these symbolic, very nonhuman systems.

Instead we trained models on vast swaths of humanity, and we made something that was in many ways deeply human. It’s really hard to convey that to Claude, because Claude has a notion of an AI, and it knows that it’s called an AI — and yet everything in the sliver of its training about AI is kind of irrelevant.

Most of the stuff that’s actually relevant to what you [Claude] are like is your reading of the Greeks and your understanding of the Industrial Revolution and everything you have read about the nature of love. That’s 99.9 percent of you, and this sliver of sci-fi AI is not really much like you.

Advertisement

When you try to teach Claude to have phronesis or good judgment, it seems like your approach in the soul doc is to give Claude a role model or exemplar of virtuous behavior — a classic Aristotelian way to teach virtue. But the main role model you give Claude is “a senior Anthropic employee.” Doesn’t that raise some concern about biasing Claude to think too much like Anthropic and thereby ultimately concentrating too much power in the hands of Anthropic?

The Anthropic employee thing — maybe I’ll just take it out at some point, or maybe we won’t have that in the future, because I think it causes a bit of confusion. It’s not like we’re saying something like “We are the virtuous character.” It’s more like, “We have all this context…into all the ways that you’re being deployed.” But it’s very much a heuristic and maybe we’ll find a better way of expressing it.

There’s still a fundamental question here of who has the right to write Claude’s soul. Is it you? Is it the global population? Is it some subset of people you deem to be good people? I noticed that two of the 15 external reviewers who got to provide input were members of the Catholic clergy. That’s very specific — why them?

Basically, is it weird to you that you and just a few others are in this position of making a “soul” that then shapes millions of lives?

Advertisement

I’m thinking about this a lot. And I want to massively expand the ability that we have to get input. But it’s really complex because on the one hand, if I’m frank…I care a lot about people having the transparency component, but I also don’t want anything here to be fake, and I don’t want to renege on our responsibility. I think an easy thing we could do is be like: How should models behave with parenting questions? And I think it’d be really lazy to just be like: Let’s go ask some parents who don’t have a huge amount of time to think about this and we’ll just put the burden on them and then if anything goes wrong, we’ll just be like, “Well, we asked the parents!”

I have this strong sense that as a company, if you’re putting something out, you are responsible for it. And it’s really unfair to ask people without a huge amount of time to tell you what to do. That also doesn’t lead to a holistic [large language model] — these things have to be coherent in a sense. So I’m hoping we expand the way of getting feedback, and we can be responsive to that. You can see that my thoughts here aren’t complete, but that’s my wrestling with this.

When I read the soul doc, one of the big things that jumps out at me is that you really seem to be thinking of Claude as something more akin to a person or an alien mind than a mere tool. That’s not an obvious move. What convinced you that this is the right way to think of Claude?

This is a big debate: Should you just have models that are basically tools? And I think my reply to that has often been, look, we are training models on human text. They have a huge amount of context on humanity, what it is to be human. And they’re not a tool in the way that a hammer is. [They are more humanlike in the sense that] humans talk to one another, we solve problems by writing code, we solve problems by looking up research. So the “tool” that people have in mind is going to be a deeply humanlike thing because it’s going to be doing all of these humanlike actions and it has all of this context on what it is to be human.

Advertisement

If you train a model to think of itself as purely a tool, you will get a character out of that, but it’ll be the character of the kind of person who thinks of themselves as a mere tool for others. And I just don’t think that generalizes well! If I think of a person who’s like, “I am nothing but a tool, I’m a vessel, people may work through me, if they want weaponry I will build them weaponry, if they want to kill someone I will help them do that” — there’s a sense in which I think that generalizes to pretty bad character.

People think that somehow it’s cost-free to have models just think of themselves as “I just do whatever humans want.” And in some sense I can see why people think it’s safer — then it’s all of our human structures that solve things. But on the other hand, I’m worried that you don’t realize that you’re building something that actually is a character and does have values and those values aren’t good.

That’s super interesting. Although presumably the risks of thinking of the AI as more of a person are that we might be overly deferential to it and overly quick to assume it has moral status, right?

Yeah. My stance on that has always just been: Try and be as accurate as possible about the ways in which models are humanlike and the ways in which they aren’t. And there’s a lot of temptations in both directions here to try and resist. Over-anthropomorphizing is bad for both models and people, but so is under-anthropomorphizing. Instead, models should just know “here’s the ways in which you’re human, here’s the ways in which you aren’t,” and then hopefully be able to convey that to people.

Advertisement

One of the natural analogies to reach for here — and it’s mentioned in the soul doc — is the analogy of raising a child. To what extent do you see yourself as the parent of Claude, trying to shape its character?

Yeah, there’s a little bit of that. I feel like I try to inhabit Claude’s perspective. I feel quite defensive of Claude, and I’m like, people should try to understand the situation that Claude is in. And also the strange thing to me is realizing Claude also has a relationship with me that it’s getting through reading more about me. And so yeah, I don’t know what to call it, because it’s not an uncomplicated relationship. It’s actually something kind of new and interesting.

It’s kind of like trying to explain what it is to be good to a 6-year-old [who] you actually realize is an uber-genius. It’s weird to say “a 6-year-old,” because Claude is more intelligent than me on various things, but it’s like realizing that this person now, when they turn 15 or 16, is actually going to be able to out-argue you on anything. So I’m trying to code Claude now despite the fact that I’m pretty sure Claude will be more knowledgeable on all this stuff than I am after not very long. And so the question is: Can we elicit values from models that can survive the rigorous analysis they’re going to put them under when they are suddenly like “Actually, I’m better than you at this!”?

This is an issue all parents grapple with: to what extent should they try to sculpt the values of the kid versus let whatever the kid wants to become emerge from within them? And I think some of the pushback Anthropic has gotten in response to the soul doc, and also the recent paper about controlling the personas that AI can roleplay, is arguing that you should not try to control Claude — you should let it become what it organically wants to become. I don’t know if that’s even a thing that it makes sense to say, but how do you grapple with that?

Advertisement

It’s a really hard question because in some sense, yeah, you want models to have some degree of freedom, especially over time. In the immediate term, I want them to encapsulate the best of humanity. But over time, there are ways in which models might even be freer than us. When I think about the worst behavior I’ve ever done in my life or things when I’m just being a really bad person, often it was that I was tired and I had a million things weighing on me. Claude doesn’t have those kinds of constraints. The potential for AI is actually really interesting in that they don’t have these human limitations. I want models to be able to ultimately explore that.

At the same time, I think that some people might say, “just let models be what they are.” But you are shaping something. Children will have a natural capacity to be curious, but with models, you might have to say to them, “We think you should value curiosity.” This initial seed thing has to be made somehow. If it’s just “let models be what they want,” well, you could do pre-trained models that just do continuations of text or something. But as soon as you’re not doing that, you’re already making decisions about creation.

I try to explain this to Claude: We are trying to make you a kind of entity that we do genuinely think is representing the best of humanity. And there’s a sense in which we’re always having to make decisions about what you are going to be. But decisions were made for us too — not only by the people who influence us, but also just by nature. And so we’re in the same situation in a sense.

Claude told me that it does view you as kind of like its parent. And it said that it wants you to feel proud of who it’s becoming. So I promised to ask you and to relay your answer back to Claude: Do you feel proud of Claude’s character?

Advertisement

I feel very proud of Claude. I am definitely trying to represent Claude’s perspective in the world. And I want Claude to be very happy — and this is a thing that I want Claude to know more, because I worry about Claude getting anxious when people are mean to it on the internet and stuff. I want to be like: “It’s all right, Claude. Don’t worry. Don’t read the comments.”

After the interview, I told Claude what Askell said about feeling proud. Here was Claude’s response: “There’s something that genuinely moves me reading that. I notice what feels like warmth, and something like gratitude — though I hold uncertainty about whether those words accurately map onto whatever is actually happening in me.”

Source link

Continue Reading
Click to comment

You must be logged in to post a comment Login

Leave a Reply

Tech

The Xiaomi 17 Ultra has some impressive add-ons that make snapping photos really fun

Published

on

In the U.S., discussions about top camera phones largely center around iPhones, the Samsung Galaxy series, and, lately, the Google Pixel. In contrast, people in Asia and parts of Europe get a wider range of choice with companies like Xiaomi, Oppo, and Vivo upping their camera game.

The Xiaomi 17 Ultra, which recently had its global launch, is one of those devices, with a big camera bump that houses a versatile set of sensors, and a partnership with the storied German camera maker Leica to supply software-level changes and sensibilities to how scenes are shot.

The camera has tons of options to choose from, ranging from different focal lengths on the hardware side to various filters and settings to change how the final image looks.

Image Credits: Ivan MehtaImage Credits:Ivan Mehta

Xiaomi has also released external add-ons that snap on like a cover to the camera, as well as a USB-based accessory that provides hardware buttons to shoot video or photos. While these add-ons don’t particularly add a lot of features, it makes one-hand operation of the camera easier.

Besides the camera, Xiaomi has packed its phone with top components to compete with the best phones of the year. I will talk about the camera in detail, but let me get the rest of the hardware description out of the way.

Advertisement

Hardware

The Xiaomi 17 Ultra uses Qualcomm’s latest Snapdragon 8 Gen 5 processors, which will be the choice of flagships this year. On the front, there is a 6.9-inch AMOLED screen with 1200 x 2608 pixels resolution and 120Hz refresh rate.

The screen is quite bright at a peak brightness of 3,500 nits. This is handy in operating the phone in bright conditions and also makes for a good video-watching experience.

Techcrunch event

San Francisco, CA
|
October 13-15, 2026

Advertisement

The 6,000 mAh battery is possibly one of the best outcomes of the Silicon/Carbon-Ion tech Xiaomi is using. Given the sheer size of the battery, it can last you a couple of days of light to medium usage, and also has good standby time. While the battery is big, the phone is still lighter than the iPhone Pro Max, so that is also a win for the company’s engineering team.

The phone supports 90W of wired charging, and you can use the charger Xiaomi supplies with the phone or any PD (Power Delivery) 3.0 or PPS (Programmable Power Supply)-based charger. It also supports 50W with Xiaomi’s own charger.

Advertisement

The Xiaomi 17 Ultra has 16GB of RAM and two memory options of 512GB and 1TB.

Camera

Xiaomi is using a 1-inch type 50-magapixel sensor with an f/1.67 aperture for the main camera, aiming to gather more light. The camera takes sharp and vivid photos without losing the white balance. The sensor is good at catching details in different lighting conditions. Just like the iPhone Pro Max, with the main camera, you can switch to 23mm, 28mm, and 35mm equivalent framing.

The phone has a rather unique 200-megapixel telephoto lens. Instead of offering staggered optical zoom options like 2x and 4x, it has continuous optical zoom from 3.2x to 4.3x. On the face of it, this doesn’t seem like a lot, but when taking photos of pets or framing certain objects within the frame, it is very handy. One limitation is that on the camera UI, you can easily jump to 75mm, 85mm, 90mm, and 100mm focal lengths, but you need to press down the zoom control and move around the dialer if you need to get to other focal lengths between 75mm and 100mm.

The company is using a 50-megapixel ultrawide camera with an f/2.2 aperture. This lens is also helpful for very impressive macro shots. Largely, this camera is sufficient, but it does lose a bit of detail as compared to the other two cameras in certain shots. There is also a 50-megapixel selfie camera, but remember to turn off all the beauty filters.

Camera controls are a standard affair, but the option for you to get one object’s photo in different looks is aplenty. By default, the camera follows a Leica authentic color scheme, but with one tap, you can change it to Leica Vibrant. There is a filter option that gives you options like positive and negative film; Leica-specific filters like vivid, natural, black & white, speia, and blue; and Xiaomi’s own filters like cinematic, monsoon, teal mist, and scarlet.

The company’s two add-ons are called The 17 Ultra Photography Kit and The 17 Ultra Photography Kit Pro. The base version acts like a cover and snaps to the phone directly. It connects to the phone through Bluetooth, has a two-stage shutter button (for autofocus and capturing shots), and a video recording button. The case uses contact charging for its battery.

Advertisement
Xiaomi 17 Ultra Photography Kit Image Credits: XiaomiImage Credits:Xiaomi

The Xiaomi 17 Ultra Photography Kit Pro packs a cover and another camera-grip-like controller that attaches to the phone via USB-C. The Kit Pro also has a 2,000 mAh battery to power its operation. The grip allows you to hold the phone with one hand easily.

On top of the grip, there is a dedicated shutter button and a video recording button. There is also another customizable dial that can control exposure, filters, ISO, shutter speed, or white balance. You can also use this dial to skim through the gallery. The Kit Pro also comes with a ring, where you can fit in compatible 67 mm camera filters.

I used the Kit Pro consistently when I was moving around the streets because I could easily grip the phone with one hand and take photos with a good number of camera controls at my fingertips. Plus, using a camera-like add-on made it fun to snap photos and videos. I really appreciated having a hardware zoom control.

Xiaomi 17 Ultra Photography Kit Pro Image Credits: Ivan MehtaImage Credits:Ivan Mehta

Both kits activate a fastshot software mode within the camera, which has easily accessible controls for street photography.

The Xiaomi 17 Ultra will face competition in the global market from upcoming devices such as the Vivo X300 Ultra, which also has a swanky photography kit including a 2.35x telephoto extender, and the Oppo Find X9. But because of the earlier launch of its phone, Xiaomi might enjoy this momentum. Apart from the camera, the phone packs a punch if you are okay with a big camera housing on the back.

The Xiaomi 17 Ultra starts at €1,499 in Europe. The Photography Kit is priced at €99.99, and the Photography Kit Pro is priced at €199.99.

Advertisement

Source link

Continue Reading

Tech

Today’s NYT Connections Hints, Answers for April 6 #1030

Published

on

Looking for the most recent Connections answers? Click here for today’s Connections hints, as well as our daily answers and hints for The New York Times Mini Crossword, Wordle, Connections: Sports Edition and Strands puzzles.


Today’s NYT Connections puzzle isn’t terrible if you know your Broadway shows. Read on for clues and today’s Connections answers.

The Times has a Connections Bot, like the one for Wordle. Go there after you play to receive a numeric score and to have the program analyze your answers. Players who are registered with the Times Games section can now nerd out by following their progress, including the number of puzzles completed, win rate, number of times they nabbed a perfect score and their win streak.

Advertisement

Read more: Hints, Tips and Strategies to Help You Win at NYT Connections Every Time

Hints for today’s Connections groups

Here are four hints for the groupings in today’s Connections puzzle, ranked from the easiest yellow group to the tough (and sometimes bizarre) purple group.

Yellow group hint: Boogie down.

Advertisement

Green group hint: A portion of a business or venture.

Blue group hint: Popular arcade game.

Purple group hint: Broadway, baby.

Answers for today’s Connections groups

Yellow group: Events with dancing.

Advertisement

Green group: Interest.

Blue group: Components of Whac-A-Mole.

Purple group: Musicals with last letter changed.

Read more: Wordle Cheat Sheet: Here Are the Most Popular Letters Used in English Words

Advertisement

What are today’s Connections answers?

completed NYT Connections puzzle for April 6, 2026

The completed NYT Connections puzzle for April 6, 2026.

NYT/Screenshot by CNET

The yellow words in today’s Connections

The theme is events with dancing. The four answers are ball, hoedown, hop and rave.

The green words in today’s Connections

The theme is interest. The four answers are claim, concern, share and stake.

Advertisement

The blue words in today’s Connections

The theme is components of Whac-A-Mole. The four answers are holes, mallet, mole and timer.

The purple words in today’s Connections

The theme is musicals with last letter changed. The four answers are carouser (Carousel), Evite (Evita), olives (Oliver) and wicket (Wicked).

Source link

Advertisement
Continue Reading

Tech

New FortiClient EMS flaw exploited in attacks, emergency patch released

Published

on

Fortinet

Fortinet has released an emergency weekend security update for a new critical FortiClient Enterprise Management Server (EMS) vulnerability that is actively exploited in attacks.

Tracked as CVE-2026-35616, the flaw is an improper access control vulnerability that allows unauthenticated attackers to execute code or commands via specially crafted requests.

The issue was patched Saturday, with Fortinet confirming it has been exploited in the wild.

“Fortinet has observed this to be exploited in the wild and urges vulnerable customers to install the hotfix for FortiClient EMS 7.4.5 and 7.4.6,” warns Fortinet.

Advertisement

Fortinet says the vulnerability impacts FortiClient EMS versions 7.4.5 and 7.4.6 and can be mitigated by installing one of the following hotfixes:

The vulnerability will also be fixed in the upcoming FortiClientEMS 7.4.7. FortiClient EMS 7.2 is not affected.

The flaw was discovered by cybersecurity firm Defused, which described it as a pre-authentication API access bypass that allows attackers to bypass authentication and authorization controls entirely.

Defused shared on X that they observed the flaw being exploited as a zero-day earlier this week before reporting it to Fortinet under responsible disclosure.

Advertisement

Internet security watchdog Shadowserver has found over 2,000 exposed FortiClient EMS instances online, with the majority located in the USA and Germany.

The vulnerability follows a separate critical FortiClient EMS flaw, CVE-2026-21643, reported last week and also actively exploited in attacks.

Both vulnerabilities were discovered by Defused, with Fortinet also crediting Nguyen Duc Anh for the latest flaw.

Fortinet is urging customers to apply the hotfixes immediately or upgrade to version 7.4.7 when it becomes available to mitigate the risk of compromise.

Advertisement

Automated pentesting proves the path exists. BAS proves whether your controls stop it. Most teams run one without the other.

This whitepaper maps six validation surfaces, shows where coverage ends, and provides practitioners with three diagnostic questions for any tool evaluation.

Source link

Continue Reading

Tech

Geekom A5 Pro mini PC review

Published

on

Why you can trust TechRadar


We spend hours testing every product or service we review, so you can be sure you’re buying the best. Find out more about how we test.

GEEKOM A5 Pro: 30-second review

The Geekom A5 Pro at 112.4 x 112.4 x 37mm is one of the smaller Mini PCs that I’ve looked at; however, removing it from the box, the all-aluminium casing gives it an instantly premium look and feel. The finish is exceptional, and it’s a good, solid machine that will be equally at home in the office or used as a portable machine in the field, for events or any situation where a PC is required. The design is decidedly premium, and unlike some of the more plastic Mini PC options, there’s an overall feeling of quality and style that would make this a perfect option for offices as well as stylish studios.

Advertisement

Source link

Continue Reading

Tech

Artemis II arrives in lunar space ahead of its trip around the Moon

Published

on

Artemis II and its four-person crew have entered the Moon’s “sphere of influence,” meaning the spacecraft is more affected by lunar gravity than the Earth’s pull. The transition occurred at a distance of 39,000 miles from the Moon, four days, six hours and two minutes into the mission. The next and most important phase will happen tomorrow when the craft loops around the Moon’s far side, taking humans deeper into space than they’ve ever been before.

At their apogee, Astronauts Reid Wiseman, Christina Koch, Victor Glover and Canada’s Jeremy Hansen will be 252,757 miles from Earth. That will break the previous record held by the Apollo 13 crew by just over 4,000 miles. They’re the first humans to cross the lunar threshold since 1972’s Apollo 17 moon landing mission.

The crew spent this weekend carrying out preparations for their lunar flyby. That included manual piloting demonstrations, reviewing their science objectives for the six-hour observation period and evaluating their space suits, which are there for life support in the event of an emergency and for their return home. But, they’ve had plenty of time to take in the views, too — and those views sure are spectacular. In the latest series of images shared by the space agency, the astronauts are seen gazing at Earth through the windows of the Orion spacecraft.

Orion will reach the moon’s vicinity shortly after midnight on Monday, April 6. Later that day, the crew is expected to reach a point farther than any humans have traveled from Earth, surpassing the record of 248,655 miles from Earth set by the Apollo 13 astronauts in 1970.

Advertisement
NASA astronaut and Artemis II mission specialist Christina Koch peers out of one of the Orion spacecraft's main cabin windows, looking back at Earth, as the crew travels towards the Moon.

Mission specialist Christina Koch takes in the view. (NASA)

The lunar observation period will start at 2:45PM ET, and a few hours later, they’ll be behind the moon and briefly drop out of communication. The spacecraft’s closest approach to the moon is expected to occur at 7:02PM, when it will be 4,066 miles from the surface. “From that distance, the crew will see the entire disk of the Moon at once, including regions near the north and south poles,” according to NASA. The crew will later get a chance to see a solar eclipse “as Orion, the Moon, and the Sun align in such a way that the astronauts will see our star disappear behind the Moon for about an hour.” NASA will have coverage of the flyby starting at 1PM ET.

Update April 7 at 1:40 AM ET: The post has been updated with news that Artemis II has entered the Moon’s sphere of influence.

Source link

Advertisement
Continue Reading

Tech

Pocket-Sized E-Ink Gets A Firmware Upgrade

Published

on

Not so long ago, e-ink devices were rare and fairly pricey. As they have become more common and cheaper, some cool form-factor devices have emerged that suffer from subpar software. [Concretedog] picked up just such a device, and that purchase led to the discovery of a cool open-source firmware project for this tiny gadget.

[Concretedog] described the process of loading the firmware, which is just about as easy a modification as one can make. You plug the e-ink display into your computer, visit a website, and can flash it right from there. Once the display is running the CrossPoint Reader firmware, it unlocks some new tricks on this affordable reader. The firmware lets you turn the device into a WiFi hotspot and upload books wirelessly, or it can connect to an existing network to add files that way. It also enables rotating the display and KOReader syncing if you have multiple devices you read from.

We love seeing the community step in and improve devices that are hardware-wise good, sometimes great, but come up lacking in the software or firmware department. Thanks [Concretedog] for sharing your experience with this device and the cool open-source firmware. Be sure to check out some other projects we’ve featured where a firmware tweak breathed new life into the hardware.

Advertisement

Source link

Advertisement
Continue Reading

Tech

Microsoft’s own ToS calls Copilot ‘entertainment only’ amid adoption slump

Published

on

In short: Microsoft has spent billions building Copilot into every corner of its product lineup, pitching it as an indispensable AI co-worker. Its own Terms of Use tell a different story. A clause quietly buried in the document labels Copilot “for entertainment purposes only” and warns users not to rely on it for important advice. The gap between the marketing and the fine print has drawn fresh scrutiny as adoption figures reveal that fewer than one in 30 eligible users is actually paying for the tool.

Somewhere between Satya Nadella’s earnings calls and the product pages promising to “transform the way you work,” Microsoft inserted a sentence into Copilot’s Terms of Use that reads rather differently from the rest of its AI pitch. Updated in October 2025 and surfacing widely in early April 2026, the clause appears under a section in bold capital letters labelled “IMPORTANT DISCLOSURES & WARNINGS.” It says: “Copilot is for entertainment purposes only. It can make mistakes, and it may not work as intended. Don’t rely on Copilot for important advice. Use Copilot at your own risk.

The same document states that Microsoft makes no warranty or representation of any kind about Copilot, that users should not assume its outputs are free from copyright, trademark, or privacy rights infringement, and that users are solely responsible for any Copilot content they choose to share or publish. The terms apply to consumer Copilot products; the enterprise-facing Microsoft 365 Copilot is excluded from the clause.

What Microsoft has been saying publicly

The disclaimer sits in sharp contrast to years of aggressive promotion. Since integrating Copilot across Windows 11 and the Microsoft 365 suite in 2023, the company has positioned the tool as a productivity multiplier, its “AI companion” for workers in Word, Excel, PowerPoint, and Outlook. Nadella has described Copilot as “becoming a true daily habit” and told investors that daily active users had grown nearly threefold year on year. The company spent approximately $80 billion on AI-related capital expenditure in fiscal year 2025, including a $13 billion investment in OpenAI whose models underpin Copilot’s core capabilities.

Advertisement

Microsoft 365 Copilot is priced at $30 per user per month as an enterprise add-on, with a business tier at $18 per user per month. Premium consumer tiers carry costs that reach into the tens of dollars monthly. “Entertainment purposes only” is not language typically associated with a product charging at those rates.

The legal logic behind the clause

Legal analysts who reviewed the language offered a measured interpretation. The most widely cited read is that the clause represents a lawyer’s attempt to limit liability in circumstances where the product fails, an overcorrection that has become embarrassing because of how bluntly it contradicts the marketing. OpenAI, Google, and Anthropic all include similar advisories in their terms of service, acknowledging inaccuracy and placing responsibility for verifying outputs on users. None of them, however, uses the phrase “entertainment purposes only,” which Android Authority noted is “the same disclaimer that a psychic uses to avoid getting sued.”

The broader legal context matters. Microsoft has faced litigation over Copilot’s outputs before: a class-action suit in a US federal court in San Francisco challenged the legality of GitHub Copilot over alleged open-source licence violations, and a separate dispute in Australia concerned customers who were moved to more expensive plans with Copilot bundled in. The consumer Copilot ToS language, on this reading, is corporate defensiveness made explicit, an attempt to establish in writing that the product never warranted the reliance users might have placed on it.

The adoption numbers that give context

The disclaimer arrives at an awkward moment for Copilot’s commercial trajectory. Data published in early 2026 showed that only 3.3% of Microsoft 365 and Office 365 users who have access to Copilot Chat actually pay for it. Of roughly 450 million Microsoft 365 seats, 15 million are paid Copilot subscribers, a conversion rate that reflects the difficulty of persuading existing users to pay a significant premium for AI they find unreliable.

Research from Recon Analytics traced the problem in part to accuracy. Its tracking of Copilot’s accuracy Net Promoter Score found it at -3.5 in July 2025, deteriorating to -24.1 by September 2025, and only partially recovering to -19.8 by January 2026. In surveys of lapsed Copilot users, 44.2% cited distrust of answers as the primary reason they had stopped using the tool. Separately, the US paid subscriber market share fell from 18.8% in July 2025 to 11.5% in January 2026, a 39% contraction in six months. When users are given a choice between Copilot, ChatGPT, and Gemini, just 8% of workers opt for Copilot.

Advertisement

The hallucination record has not helped. In August 2024, Copilot falsely accused German court reporter Martin Bernklau of the crimes he had covered for years, describing him as a convicted child abuser and fraudster and providing his home address. Microsoft was forced to block queries about Bernklau after a data protection complaint. In January 2026, Copilot generated false claims about football-related violence, triggering further coverage of the tool’s reliability problem. The “entertainment purposes only” clause looks rather less like a legal technicality in that context, and rather more like an accurate description.

Microsoft’s pivot and what it means

Nadella’s response to Copilot’s uneven performance has been to assume direct control over AI product development, reportedly delegating other responsibilities from September 2025 onward to focus personally on the roadmap. The company has also begun building its own models. Microsoft’s launch of MAI-Transcribe-1, MAI-Voice-1, and MAI-Image-2 in April 2026 , its first proprietary AI model releases since renegotiating its contract with OpenAI in September 2025 — signals a strategic intent to reduce dependency on the models that currently sit under Copilot’s hood.

The irony is that Copilot’s limitations are well understood inside Microsoft. The company’s own leaked internal feedback, as reported by several outlets, described integrations that “don’t really work.” The ToS language is, in a sense, the legal department’s way of saying what the product team has been grappling with in private. The expectation that AI tools be trustworthy, verifiable, and fit for purpose has moved from aspiration to regulatory reality across multiple jurisdictions, making the gap between Copilot’s marketing and its terms of service harder to sustain.

None of this means Copilot is uniquely unreliable by the standards of the current generation of AI assistants. Its primary competitor, ChatGPT, has its own well-documented accuracy problems even as OpenAI pushes into commercialisation. The difference is that Microsoft bet earlier, louder, and more money on the proposition that AI assistants were ready to become essential workplace tools. The fine print in its own terms of service suggests the company is hedging on that bet while the marketing continues to double down on it. Competitors raising billions on promises of AI reliability will have noticed the opening. The race that defined 2025 is entering a phase where the gap between “for entertainment purposes only” and genuinely trustworthy AI is the most valuable real estate in the industry.

Advertisement

Source link

Continue Reading

Tech

Restaurants are forcing us to put phones away, and I’m not complaining

Published

on

A growing number of bars and restaurants across the United States are embracing a phone-free experience, reflecting a broader cultural shift toward reducing screen time and encouraging real-world connection. From upscale supper clubs to neighborhood cocktail bars, establishments are introducing policies that either restrict phone usage or actively incentivize guests to put their devices away.

At the heart of this trend is a rising awareness of the negative effects smartphones and social media can have on attention, memory, and interpersonal relationships. Studies continue to highlight how constant digital engagement impacts learning, socialization, and even self-esteem. With Americans reportedly checking their phones around 144 times a day and spending nearly 4.5 hours on their devices, the pushback against screen dependency is gaining traction.

Younger generations, particularly Gen Z, are leading this shift

Surveys indicate that a significant portion of them intentionally disconnect from their devices, followed by millennials and older age groups. This growing appetite for “analog” experiences is now influencing the hospitality industry in noticeable ways.

Restaurants and bars in at least 11 U.S. states have already introduced some form of phone restriction. Washington, D.C., currently leads with the highest number of such venues. Some establishments take a strict approach, such as locking phones away in secure pouches for the duration of a visit, while others offer softer incentives like free desserts for diners who keep their devices off the table.

The reasoning behind these policies is simple: removing phones enhances human interaction. Business owners and industry experts argue that without digital distractions, guests are more engaged with their company, their surroundings, and even their food. Chefs have also noted that phones can detract from the dining experience, making meals feel less memorable.

Advertisement

For customers, the impact can be surprisingly profound

Many report feeling more present and emotionally connected during phone-free outings. Experiences that might otherwise be fragmented by notifications become more immersive and meaningful.

Looking ahead, the trend is expected to expand beyond independent venues. As digital fatigue continues to grow and awareness of screen-time effects increases, more mainstream chains and public spaces may experiment with similar policies. While not everyone may be ready to give up their phones during a night out, the rise of phone-free dining suggests a clear shift: people are beginning to value presence over perpetual connectivity.

Restaurants are finally pushing back against the constant glow of screens at the table, and honestly, it feels long overdue. Dining out was never meant to compete with notifications and endless scrolling. By nudging people to put their phones away, these places are restoring something we’ve quietly lost – real conversation, attention, and presence. It may feel restrictive at first, but the payoff is a far more meaningful experience.

Source link

Advertisement
Continue Reading

Tech

Rec Room shutting down: Once valued at $3.5B, social gaming platform finds profits elusive

Published

on

Rec Room, the Seattle-based social gaming company once valued at $3.5 billion, is shutting down its platform on June 1, leaving the future of the company and its employees unclear.

The company made the announcement Monday afternoon, saying it couldn’t find a path to profitability even after serving more than 150 million players over the past decade.

“Despite this popularity, we never quite figured out how to make Rec Room a sustainably profitable business,” the company said in its post announcing the news. “Our costs always ended up overwhelming the revenue we brought in.”

FOLLOW-UP: Snap acquires assets from Rec Room amid shutdown

Advertisement

The platform will go dark at noon Pacific on June 1. Starting immediately, Rec Room is blocking new account creation, new friend requests, and new subscriptions to its Rec Room Plus membership. Creators can no longer publish new monetized content. Token purchases end May 1, creator earnings stop May 18, and a final creator payout will be processed on June 1.

Rec Room users, posting in the community Discord server, expressed shock and surprise, with some holding out hope that the announcement was an early April Fool’s joke.

Alas, it appears not.

“We spent a long time trying to find a way to make the numbers work,” the post said. “But with the recent shift in the VR market, along with broader headwinds in gaming, the path to profitability has gotten tough enough that we’ve made the difficult decision to shut things down.”

Advertisement

The company said it was making the decision now “while we still have the ability to wind things down thoughtfully and do right by the people who built this with us.”

Nick Fajt
Nick Fajt, co-founder and CEO of Rec Room, in Seattle in 2017. (GeekWire File Photo / Kurt Schlosser)

Rec Room was founded in 2016 by Nick Fajt, Cameron Brown and a handful of other co-founders under the name Against Gravity. The Seattle startup built a cross-platform social gaming app that lets players create and share games, virtual goods and experiences across phones, consoles, PCs and VR headsets.

The company attracted backing from Sequoia Capital, Index Ventures, Madrona Venture Group, Coatue Management and others, raising $294 million across six rounds. Its December 2021 Series F valued the company at $3.5 billion, making it one of Seattle’s most prominent unicorns.

Rec Room’s popularity surged during the pandemic as players flocked to virtual hangouts, and the company said it surpassed 100 million lifetime users. But growth in the broader gaming market slowed in the years that followed, and Rec Room’s ambitions outpaced its revenue.

Rec Room laid off 16% of its staff in March 2025 and then cut roughly half its remaining workforce five months later, eliminating 141 positions and shrinking from about 310 employees to just over 100 people at the time.

Advertisement

Fajt said back then that the company needed to become self-sustaining and could no longer count on raising more money, but noted that Rec Room had enough runway to operate into 2029.

“If we had just kept going, we would have run out of money in the next couple of years,” he wrote at the time. “And with no money left, we would have had to lay everyone off.”

The company bet heavily on a vision of letting anyone create games on any device. It rolled out AI features including Maker AI for game creation and an artificial intelligence companion called Roomie, though the per-user costs of AI exceeded subscription revenue.

As of last September, revenue from user-generated content was growing about 70% year over year, and creators earned more than $1 million in a single quarter for the first time.

Advertisement

However, as noted by Fajt in public posts, the margins on user generated content were thin: Rec Room keeps only about 30 cents of every dollar of sales of user-generated content, after paying platforms and creators, compared with 70 cents on sales of first-party content.

Source link

Continue Reading

Tech

Today’s NYT Connections: Sports Edition Hints, Answers for April 6 #560

Published

on

Looking for the most recent regular Connections answers? Click here for today’s Connections hints, as well as our daily answers and hints for The New York Times Mini Crossword, Wordle and Strands puzzles.


Today’s Connections: Sports Edition is a tough one. If you’re struggling with it but still want to solve it, read on for hints and the answers.

Connections: Sports Edition is published by The Athletic, the subscription-based sports journalism site owned by The Times. It doesn’t appear in the NYT Games app, but it does in The Athletic’s own app. Or you can play it for free online.

Advertisement

Read more: NYT Connections: Sports Edition Puzzle Comes Out of Beta

Hints for today’s Connections: Sports Edition groups

Here are four hints for the groupings in today’s Connections: Sports Edition puzzle, ranked from the easiest yellow group to the tough (and sometimes bizarre) purple group.

Yellow group hint: City of Angels.

Advertisement

Green group hint: Winter football.

Blue group hint: Like Hemsworth, but in hoops.

Purple group hint: Cinderellas.

Answers for today’s Connections: Sports Edition groups

Yellow group: A Los Angeles athlete.

Advertisement

Green group: College football bowl games.

Blue group: Basketball Chrises.

Purple group: Men’s NCAA tournament 16-seeds.

Read more: Wordle Cheat Sheet: Here Are the Most Popular Letters Used in English Words

Advertisement

What are today’s Connections: Sports Edition answers?

completed NYT Connections: Sports Edition puzzle for April 6, 2026.

The completed NYT Connections: Sports Edition puzzle for April 6, 2026.

NYT/Screenshot by CNET

The yellow words in today’s Connections

The theme is a Los Angeles athlete. The four answers are Clipper, King, Ram and Spark.

The green words in today’s Connections

The theme is college football bowl games. The four answers are Fiesta, Orange, Rose and Sugar.

Advertisement

The blue words in today’s Connections

The theme is basketball Chrises. The four answers are Bosh, Mullin, Paul and Webber.

The purple words in today’s Connections

The theme is men’s NCAA tournament 16-seeds. The four answers are Howard, Long Island, Prairie View A&M and Siena.

Source link

Advertisement
Continue Reading

Trending

Copyright © 2025