If you’re excited, or even just a little curious, about the future of augmented reality, Meta’s Orion prototype makes the most compelling case yet for the technology.
For Meta, Orion is about more than finally making AR glasses a reality. It’s also the company’s best shot at becoming less dependent on Apple and Google’s app stores, and the rules that come with them. If Orion succeeds, then maybe we won’t need smartphones for much at all. Glasses, Zuckerberg , might eventually become “the main way we do computing.”
At the moment, it’s still way too early to know if Zuckerberg’s bet will actually pay off. Orion is, for now, still a prototype. Meta hasn’t said when it might become widely available or how much it might cost. That’s partly because the company, which has already poured tens of billions of dollars into AR and VR research, still needs to figure out how to make Orion significantly more affordable than the $10,000 it costs to make the current version. It also needs to refine Orion’s hardware and software. And, perhaps most importantly, the company will eventually need to persuade its vast user base that AI-infused, eye-tracking glasses offer a better way to navigate the world.
Still, Meta has been eager to show off Orion since at Connect. And, after recently getting a chance to try out Orion for myself, it’s easy to see why: Orion is the most impressive AR hardware I’ve seen.
Advertisement
Meta has clearly gone to great lengths to make its AR glasses look, well, normal. While Snap has been mocked for its oversized Spectacles, Orion’s shape and size is closer to a traditional pair of frames.
Even so, they’re still noticeably wide and chunky. The thick black frames, which house an array of cameras, sensors and custom silicon, may work on some face shapes, but I don’t think they are particularly flattering. And while they look less cartoonish than Snap’s AR Spectacles, I’m pretty sure I’d still get some funny looks if I walked around with them in public. At 98 grams, the glasses were noticeably bulkier than my typical prescription lenses, but never felt heavy.
1 / 4
Meta Orion glasses
Advertisement
Meta’s Orion glasses are still quite bulky.
In addition to the actual glasses, Orion relies on two other pieces of kit: a 182-gram “wireless compute puck, which needs to stay near the glasses, and an electromyography (EMG) wristband that allows you to control the AR interface with a series of hand gestures. The puck I saw was equipped with its own cameras and sensors, but Meta told me they’ve since simplified the remote control-shaped device so that it’s mainly used for connectivity and processing.
When I first saw the three-piece Orion setup at Connect, my first thought was that it was an interesting compromise in order to keep the glasses smaller. But after trying it all together, it really doesn’t feel like a compromise at all.
You control Orion’s interface through a combination of eye tracking and gestures. After a quick calibration the first time you put the glasses on, you can navigate the AR apps and menus by glancing around the interface and tapping your thumb and index finger together. Meta has been experimenting with wrist-based neural interfaces for years, and Orion’s EMG wristband is the result of that work. The band, which feels like little more than a fabric watch band, uses sensors to detect the electrical signals that occur with even subtle movements of your wrist and fingers. Meta then uses machine learning to decode those signals and send them to the glasses.
Advertisement
That may sound complicated, but I was surprised by how intuitive the navigation felt. The combination of quick gestures and eye tracking felt much more precise than hand tracking controls I’ve used in VR. And while Orion also has hand-tracking abilities, it feels much more natural to quickly tap your fingers together than to extend your hands out in front of your face.
What it’s like to use Orion
Meta walked me through a number of demos meant to show off Orion’s capabilities. I asked Meta AI to generate an image, and to come up with recipes based on a handful of ingredients on a shelf in front of me. The latter is a trick I’ve with the Ray-Ban Meta Smart Glasses, except with Orion, Meta AI was also able to project the recipe steps onto the wall in front of me.
I also answered a couple of video calls, including one from a surprisingly lifelike . I watched a YouTube video, scrolled Instagram Reels, and dictated a response to an incoming message. If you’ve used mixed reality headsets, much of this will sound familiar, and a lot of it wasn’t that different from what you can do in VR headsets.
The magic of AR, though, is that everything you see is overlaid onto the world around you and your surroundings are always fully visible. I particularly appreciated this when I got to the gaming portion of the walkthrough. I played a few rounds of a Meta-created game called Stargazer, where players control a retro-looking spacecraft by moving their head to avoid incoming obstacles while shooting enemies with finger tap gestures. Throughout that game, and a subsequent round of AR Pong, I was able to easily keep up a conversation with the people around me while I played. As someone who easily gets motion sick from VR gaming, I appreciated that I never felt disoriented or less aware of my surroundings.
Advertisement
Orion’s displays rely on silicon carbide lenses, micro-LED projectors and waveguides. The actual lenses are clear, though they can dim depending on your environment. One of the most impressive aspects is the 70-degree field of view. It was noticeably wider and more immersive than what I experienced with Snap’s AR Spectacles, which have a 46-degree field of view. At one point, I had three windows open in one multitasking view: Instagram Reels, a video call and a messaging inbox. And while I was definitely aware of the outer limits of the display, I could easily see all three windows without physically moving my head or adjusting my position. It’s still not the all-encompassing AR of sci-fi flicks, but it was wide enough I never struggled to keep the AR content in view.
What was slightly disappointing, though, was the resolution of Orion’s visuals. At 13 pixels per degree, the colors all seemed somewhat muted and projected text was noticeably fuzzy. None of it was difficult to make out, but it was much less vivid than what I saw on , which have a 37 pixels per degree resolution.
Meta’s VP of Wearable Devices, Ming Hua, told me that one of the company’s top priorities is to increase the brightness and resolution of Orion’s displays. She said that there’s already a version of the prototype with twice the pixel density, so there’s good reason to believe this will improve over time. She’s also optimistic that Meta will eventually be able to bring down the costs of its AR tech, eventually reducing it to something “similar to a high end phone.”
What does it mean?
Leaving my demo at Meta’s headquarters, I was reminded of the first time I tried out a prototype of the wireless VR headset that would eventually become known as Quest, back in 2016. Called at the time, it was immediately obvious, even to an infrequent VR user, that the wireless, room-tracking headset was the future of the company’s VR business. Now, it’s almost hard to believe there was a time when Meta’s headsets weren’t fully untethered.
Advertisement
Orion has the potential to be much bigger. Now, Meta isn’t just trying to create a more convenient form factor for mixed reality hobbyists and gamers. It’s offering a glimpse into how it views the future, and what our lives might look like when we’re no longer tethered to our phones.
For now, Orion is still just that: a glimpse. It’s far more complex than anything the company has attempted with VR. Meta still has a lot of work to do before that AR-enabled future can be a reality. But the prototype shows that much of that vision is closer than we think.
Join our daily and weekly newsletters for the latest updates and exclusive content on industry-leading AI coverage. Learn More
While many existing risks and controls can apply to generative AI, the groundbreaking technology has many nuances that require new tactics, as well.
Models are susceptible to hallucinations, or the production of inaccurate content. Other risks include the leaking of sensitive data via a model’s output, tainting of models that can allow for prompt manipulation and biases as a consequence of poor training data selection or insufficiently well-controlled fine-tuning and training.
Ultimately, conventional cyber detection and response needs to be expanded to monitor for AI abuses — and AI should conversely be used for defensive advantage, said Phil Venables, CISO of Google Cloud.
Advertisement
“The secure, safe and trusted use of AI encompasses a set of techniques that many teams have not historically brought together,” Venables noted in a virtual session at the recent Cloud Security AllianceGlobal AI Symposium.
Lessons learned at Google Cloud
Venables argued for the importance of delivering controls and common frameworks so that every AI instance or deployment does not start all over again from scratch.
“Remember that the problem is an end-to-end business process or mission objective, not just a technical problem in the environment,” he said.
Nearly everyone by now is familiar with many of the risks associated with the potential abuse of training data and fine-tuned data. “Mitigating the risks of data poisoning is vital, as is ensuring the appropriateness of the data for other risks,” said Venables.
Advertisement
Importantly, enterprises should ensure that data used for training and tuning is sanitized and protected and that the lineage or provenance of that data is maintained with “strong integrity.”
“Now, obviously, you can’t just wish this were true,” Venables acknowledged. “You have to actually do the work to curate and track the use of data.”
This requires implementing specific controls and tools with security built in that act together to deliver model training, fine-tuning and testing. This is particularly important to assure that models are not tampered with, either in the software, the weights or any of their other parameters, Venables noted.
“If we don’t take care of this, we expose ourselves to multiple different flavors of backdoor risks that can compromise the security and safety of the deployed business or mission process,” he said.
Advertisement
Filtering to fight against prompt injection
Another big issue is model abuse from outsiders. Models may be tainted through training data or other parameters that get them to behave against broader controls, said Venables. This could include adversarial tactics such as prompt manipulation and subversion.
Venables pointed out that there are plenty of examples of people manipulating prompts both directly and indirectly to cause unintended outcomes in the face of “naively defended, or flat-out unprotected models.”
This could be text embedded in images or other inputs in single or multimodal models, with problematic prompts “perturbing the output.”
“Much of the headline-grabbing attention is triggering on unsafe content generation, some of this can be quite amusing,” said Venables.
Advertisement
It’s important to ensure that inputs are filtered for a range of trust, safety and security goals, he said. This should include “pervasive logging” and observability, as well as strong access control controls that are maintained on models, code, data and test data, as well.
“The test data can influence model behavior in interesting and potentially risky ways,” said Venables.
Controlling the output, as well
Users getting models to misbehave is indicative of the need to manage not just the input, but the output, as well, Venables pointed out. Enterprises can create filters and outbound controls — or “circuit breakers” —around how a model can manipulate data, or actuate physical processes.
“It’s not just adversarial-driven behavior, but also accidental model behavior,” said Venables.
Advertisement
Organizations should monitor for and address software vulnerabilities in the supporting infrastructure itself, Venables advised. End-to-end platforms can control the data and the software lifecycle and help manage the operational risk of AI integration into business and mission-critical processes and applications.
“Ultimately here it’s about mitigating the operational risks of the actions of the model’s output, in essence, to control the agent behavior, to provide defensive depth of unintended actions,” said Venables.
He recommended sandboxing and enforcing the least privilege for all AI applications. Models should be governed and protected and tightly shielded through independent monitoring API filters or constructs to validate and regulate behavior. Applications should also be run in lockdown loads and enterprises need to focus on observability and logging actions.
In the end, “it’s all about sanitizing, protecting, governing your training, tuning and test data. It’s about enforcing strong access controls on the models, the data, the software and the deployed infrastructure. It’s about filtering inputs and outputs to and from those models, then finally making sure you’re sandboxing more use and applications in some risk and control framework that provides defense in depth.”
Advertisement
VB Daily
Stay in the know! Get the latest news in your inbox daily
Buying a home has always been complicated. You have to figure out how much money to put down and how that down payment will affect a monthly mortgage bill. Then there are the closing costs and fees. Kevin Bennett launched Further to try to help make the financial process easier to navigate — especially for first-time buyers.
Further is a fintech platform that walks users through the financial side of home buying. The company’s first product, which goes live Friday, is a calculator that shows what people can afford and what their monthly mortgage payments and closing costs could look like, among other metrics based on real-time interest rates.
Unlike other mortgage calculators that you can find on Zillow and LendingTree, Further looks to give users more than the numbers. It tells users how easy it will be for them to find a loan based on their financial status, whether they should wait to buy, or if they should pursue specific types of loans based on their financial profile, among others.
The platform is currently free to use. The company plans to monetize once it releases more product developments but declined to share details.
Advertisement
“A generation ago, our parents bought a $200,000 home with a 20% mortgage, and it was very straightforward,” Bennett said. “There was one kind of mortgage, and that’s what you did and it’s just more complicated. There are lots of kinds of mortgages. There are lots of implications. Homes are much more expensive now, so there’s just a lot more complexity, and it’s a much bigger financial decision.”
Last year Bennett found himself looking for something new to work on after stepping back from Caribou, the auto loan refinancing startup he co-founded in 2016 and where he served as CEO. He knew he wanted to do something else mission-oriented but wasn’t sure where.
He started looking into real estate, a category he said he’s always been fascinated with. The fact that his whole family works in real estate helped, too. He started talking to folks who had purchased their home within the last two years and found a lot of common pain points: People didn’t understand the process and were relying on homemade spreadsheets to try to figure out what they could afford.
Bennett also had a personal experience: He bought and sold a townhouse in his 20s and was surprised to find out he endured a $30,000 loss, despite selling the home for the original purchase price. That’s because he missed out on certain home improvements that could’ve increased the house’s value.
Advertisement
“You can’t hit the undo button once you buy that house,” Bennett said. “It felt like there was a gap in the market. It felt like it was a lot more complicated than it was a generation ago.”
He reached out to his friend Chris Baker, a real estate expert, and former head of product at EasyKnock, about his idea last year. The pair got to work fast. Their first conversation was November 3, 2023. They decided to work together in January, launched the product in April, and raised an undisclosed pre-seed round in June. Now, they are coming out of stealth.
“Our goal is to take care of the complicated jargon and stuff and really help you understand as easily as possible what it is you need to know, with transparency, obviously, but also putting you in the driver’s seat and in control,” he said.
The company’s previously undisclosed pre-seed round raised $4.1 million from investors including Link Ventures, Vesta Ventures, and Fidi Ventures, among others. Bennett said that fundraising wasn’t too challenging, as half of the capital the company raised was from investors who backed him while he was at Caribou. Bennett thinks his track record as a founder made a big difference. The company built its cap table intentionally to include angel investors who have experience in the real estate market, he said.
Advertisement
This kind of financial information and guidance seems like something a Zillow or Redfin would be ripe to copy especially considering Zillow already offers a mortgage calculator and some advice of its own. But Bennett said he wasn’t super concerned about the competition. He said he thinks that many companies either fall on the proptech side or the fintech side and rarely in the middle, as Further does, which gives it more of a moat.
But Further is definitely not the only company that sits between proptech and fintech that is aimed at consumers. Online mortgage startup Better.com, which allows consumers to browse for mortgage options or refinance an existing one, is a good example.
It will likely depend on what Further unveils in its planned Q1 product release that will include more features and capabilities, but Bennett didn’t share too many details just yet. For now, users can use Further to get an idea of what they can afford and what they can expect to pay when buying a house.
“My hope is that we can enable people with the right insights and information to make good decisions and plan for this really big part of their life in a way that gives them confidence, puts them at ease and and lets them focus on, you know, what they really want to focus on, which is kind of that that dream of being a homeowner,” Bennett said.
A newly proposed cosmic speed limit may constrain how fast anything in the universe can grow. Its existence follows from Alan Turing’s pioneering work on theoretical computer science, which opens the intriguing possibility that the structure of the universe is fundamentally linked to nature of computation.
Cosmic limits are not a new idea. While studying the relationship between space and time, Albert Einstein showed that nothing in the universe can exceed the speed of light, as part of his special theory of relativity. Now, Toby Ord at…
The addition of a 4.3-inch color TFT screen makes the new Wyze Scale Ultra one of the brand’s most expensive smart scales to date, but at $43.99, it’s still considerably cheaper than offerings from companies like Withings. It’s available from Wyze directly or from Amazon in white or black.
The Wyze Scale Ultra says it can track 13 different health metrics, including your heart rate, your metabolic age (a comparison of how your body burns calories at rest to others your age), and measurements of fat, muscle, and water.
Previous versions of Wyze’s smart scales featured simple segmented LED displays to display basic information like weight, BMI, and muscle mass, leaving more detailed breakdowns of your health metrics for an accompanying mobile app. The Wyze Scale Ultra can display more data, including how measurements like weight or body fat have fluctuated over time, and it’s customizable, so it only displays what you want it to.
The information displayed on the Wyze Scale Ultra’s full color screen can be customized by each user.Image: Wyze
Advertisement
Like the Wyze Scale X introduced in 2022, the Scale Ultra offers modes for easily weighing pets, babies, or luggage and a pregnancy mode that turns off the weak electrical current used for bioelectric impedance analysis (BIA) as an added safety precaution.
The Wyze Scale Ultra can also be used to weigh pets, children, and luggage.Image: Wyze
Connectivity includes both Bluetooth and Wi-Fi, and the Wyze Scale Ultra can automatically recognize and sync measured health metrics for up to eight different users — either to its mobile app or to the Apple Health, Google Fit, and Fitbit platforms. It’s not rechargeable, however. It runs on four AA batteries, which Wyze says will keep the scale powered for up to nine months.
The outcome of the U.S. presidential election on Nov. 5 won’t affect oil production levels in the short- to medium term, Exxon CEO Darren Woods told CNBC on Friday.
Former President Donald Trump has called for unconstrained oil and gas production to lower energy prices and fight inflation, boiling his energy policy down to three words on the campaign trail: “Drill, baby, drill.”
“I’m not sure how drill, baby, drill translates into policy,” Woods told CNBC’s “Squawk Box” Friday after the largest U.S. oil and gas company reported third-quarter results.
Woods said U.S. shale production does not face constraints from “external restrictions.” The U.S. has produced record amounts of oil and gas during the Biden administration.
Over the past six years, the U.S. has produced more crude oil than any other nation in history, including Saudi Arabia and Russia, according to the Energy Information Administration.
Advertisement
Output in the U.S. is driven by the oil and gas industry deploying technology and investment to generate shareholder returns based on the break-even cost of production, the CEO said.
“Certainly we wouldn’t see a change based on a political change but more on an economic environment,” Woods said. “I don’t think there’s anybody out there that’s developing a business strategy to respond to a political agenda,” he said.
While shale production has not faced constraints on developing new acreage, there are resources in areas like the Gulf of Mexico that have not opened up due to federal permitting, the CEO said.
“That could, for the longer term, open up potential sources of supply,” Wood said. In the short- to medium term, however, unconventional shale resources are available and it’s just a matter of developing them based on market dynamics, he said.
Advertisement
Exxon Mobil shares in 2024.
The vast majority of shale resources in the U.S. are on private land and regulated at the state level, according to an August note from Morgan Stanley. About 25% of oil and 10% of natural gas is produced on federal land and waters subject to permitting, according to Morgan Stanley.
Vice President Kamala Harris opposed fracking during her bid for the 2020 Democratic presidential nomination. She has since reversed that position in an effort to shore up support in the crucial swing state of Pennsylvania, where the natural gas industry is important for the state’s economy.
LiteSpeed Cache, an immensely popular WordPress plugin for site performance optimization, suffered from a vulnerability which allowed threat actors to gain admin status.
With such elevated privileges, they would be able to perform all sorts of malicious activities on the compromised websites.
According to researchers from Patchstack, the vulnerability was discovered in the is_role_simulation function, and it is relatively similar to a different vulnerability that was discovered last summer. The function apparently used a weak security hash check that could be broken with brute force, granting the attackers the ability to abuse the crawler feature and simulate a logged-in administrator.
Who is vulnerable?
There are a few factors that need to align before the vulnerability can be abused, though.
That includes having the crawler turned on, with run duration between 2500 and 4000, and the intervals between runs being set to 2500- 4000. Furthermore, Server Load Limit should be set to 9, Role Simulation to 1 (ID of user with admin role), and Turn every row to OFF except Administrator should be activated.
Advertisement
The vulnerability is now tracked as CVE-2024-50550, and has a severity score of 8.1 (high severity). It was already patched, with the version 6.5.2 of the plugin being the earliest clean one. LiteSpeed Cache is one of the most popular plugins of its kind, with more than six million active installations.
There is no talk of any evidence of in-the-wild abuse, so chances are cybercrooks have not picked up on the vulnerability in the past.
Sign up to the TechRadar Pro newsletter to get all the top news, opinion, features and guidance your business needs to succeed!
However, now that the patch is public, it’s only a matter of time before they start scanning for vulnerable websites. Currently, almost three-quarters (72.1%) of all LiteSpeed Cache websites are running the latest version, 6.5, with 6.7% running 6.4, and a notable 21.2% running “other” versions. Therefore, at least 27.6% of sites could be targeted, which is more than 1.6 million.
You must be logged in to post a comment Login