Connect with us

Technology

NYT Strands today: hints, spangram and answers for Sunday, November 10

Published

on

NYT Strands today: hints, spangram and answers for Saturday, September 21

Strands is a brand new daily puzzle from the New York Times. A trickier take on the classic word search, you’ll need a keen eye to solve this puzzle.

Like Wordle, Connections, and the Mini Crossword, Strands can be a bit difficult to solve some days. There’s no shame in needing a little help from time to time. If you’re stuck and need to know the answers to today’s Strands puzzle, check out the solved puzzle below.

How to play Strands

You start every Strands puzzle with the goal of finding the “theme words” hidden in the grid of letters. Manipulate letters by dragging or tapping to craft words; double-tap the final letter to confirm. If you find the correct word, the letters will be highlighted blue and will no longer be selectable.

If you find a word that isn’t a theme word, it still helps! For every three non-theme words you find that are at least four letters long, you’ll get a hint — the letters of one of the theme words will be revealed and you’ll just have to unscramble it.

Advertisement

Every single letter on the grid is used to spell out the theme words and there is no overlap. Every letter will be used once, and only once.

Each puzzle contains one “spangram,” a special theme word (or words) that describe the puzzle’s theme and touches two opposite sides of the board. When you find the spangram, it will be highlighted yellow.

The goal should be to complete the puzzle quickly without using too many hints.

Hint for today’s Strands puzzle

Today’s theme is “Nice fit”

Advertisement

Here’s a hint that might help you: specific types of outfits.

Today’s Strand answers

NYT Strands logo.
NYT

Today’s spanagram

We’ll start by giving you the spangram, which might help you figure out the theme and solve the rest of the puzzle on your own:

Today’s Strands answers

  • JUMP
  • CIVIL
  • STRONG
  • LEISURE
  • BIRTHDAY
  • BUSINESS






Source link

Advertisement
Continue Reading
Advertisement
Click to comment

You must be logged in to post a comment Login

Leave a Reply

Technology

Strava adds Night and Weekly Heatmaps to its fitness app

Published

on

Strava adds Night and Weekly Heatmaps to its fitness app

Strava, a popular app for tracking fitness activities, is expanding its Hatmaps feature to help improve the safety of its users. The update should be especially useful now for users in the Northern Hemisphere, which is heading into winter with reduced daylight.

The new Night and Weekly Heatmaps were announced by the San Francisco-based company on Wednesday and are available to all Strava subscribers. As the name of the feature suggests, the Heatmaps show where Strava users are choosing to exercise, with dark thick lines showing well-used routes, and light thin lines showing less popular ones.

First up, the new Night Heatmaps feature is ideal for those who are doing their activities in the late evening or early morning hours, when there’s less light. They show the most popular areas for outdoor activities from sunset to sunrise, helping athletes to better plan their outdoor activities during this time frame. If it’s a new area for you, you may also want to cross-check the Night Heatmap data with Google Street View images to get a better understanding of the place.

Weekly Heatmaps, on the other hand, show data for recent heat from the last seven days so that users can see which trails and roads are currently active, particularly during seasonal transitions when conditions may be impacted by weather.

Advertisement

“Our global community powers ourHeatmaps and now we’ve made it easier for our community members to build routes with confidence, regardless of the season or time of day,” Matt Salazar, Strava’s chief product officer, said in Wednesday’s announcement about the new features. “We are continually improving our mapping technology to make human-powered movement easier for all skill levels.”

Strava has also shared a useful at-a-glance guide to all four of its Heatmaps, Night, Weekly, Global, and Personal:

Night (new): Discover the most frequented areas between sunset and sunrise; ideal for evening or early morning users.

Weekly (new): Stay updated with the latest data from the past seven days; perfect for adjusting plans around seasonal changes or unexpected closures.

Advertisement

Global (existing): Viewable by anyone regardless of whether you have a Strava account, the Global Heatmap allows you to see what areas are most popular around the world based on community activity uploads.

Personal (existing): A one-of-a-kind illustration showing the record of everywhere you’ve logged a GPS activity. This heatmap is private and only available to you.



Advertisement




Source link

Continue Reading

Technology

Google’s new AI model is here to help you learn

Published

on

Google's new AI model is here to help you learn

Google’s Gemini is useful as an educational tool to help you study for that exam. However, Gemini is sort of the “Everything chatbot” that’s useful for just about everything. Well, Google has a new model for people looking for more of a robust educational tool. Google calls it Learn About, and it could give other tools a run for their money.

Say what you want about Google’s AI, the company has been hard at work making AI tools centered around teaching rather than cheating. For example, it has tools in Android studio that guides programmers and help them learn coding. Also, we can’t forget about NotebookLM. This is the tool that takes your uploaded educational content and helps you digest it. We can’t forget abou the Audio Overviews feature that turns your uploaded media into a live podcast-style educational discussion.

So, Google has a strong focus on education with its AI tools. Let’s just hope that other companies will follow suit.

Google’s new AI tool is called Learn About

This tool is pretty self-explanatory, as it focus on giving you more text-book style explanations for your questions. Rather than simply giving you answers, this tool will go the extra mile to be more descriptive with its explanation. Along with that, Learn About will also provide extra context on the subject and give you other educational material on it.

Advertisement

Google achieved this by using a totally different model to power this tool. Rather than using the Gemini model, Ask About uses a model called LearnLM. At this point, we don’t really know much about this model, but we know that Google steered it more towards providing academic answers.

Gemini’s answer vs. Learn About’s answers

We tested it out by asking what pulsars are, and we compared the answer to what Gemini gave us for the same question. Gemini delivered a pretty fleshed-out explanation in the form of a few paragraphs. It also snagged a few pictures from the internet and pasted the link to a page at the bottom. This is good for a person who’s casually looking up a definition. Maybe that person isn’t looking to learn the ins and outs of what a pulsar is.

There was one issue with Gemini’s answer; one of the images that it pasted was an image of a motorcycle. It pasted an image of the Bajaj Pulsar 150. So, while it technically IS a pulsar, a motorcycle shares very few similarities with massive rapidly spinning balls of superheated plasma billions of miles away from Earth.

What about Learn About?

Learn About also gave an explanation in the form of a few paragraphs;  however, Learn About’s explanation was shorter. It makes up for it by producing more extraneous material. Along with images, it provided three links (one of which was a YouTube video) and chips with commands like Simplify, Go deeper, and Get images (more on the chips below).

Advertisement

Under the chips, you’ll see suggestions of other queries that you can put in for additional context. Lastly, in textbook style, you’ll see colored blocks with additional content. For example, there’s a Why it matters block and a Stop & think block.

Chips

Going back to the aforementioned chips, selecting Simplify and Get images are axiomatic enough. Tapping/clicking on the Go Deeper chip is a bit more interesting. It brought up an Interactive List consisting of a selection of additional queries that will provide extra information about pulsars. Each query you select will bring up even more information.

Google Ask about 6

Textbook blocks

Think about the textbooks you used in school, and you’ll be familiar with these blocks. These blocks come in different colors. The Why it matters block tells you why this information is important. Next, the Stop & think block seems to give you a little bit of tangential information. It asks a question and has a button to reveal the answer. It’s a way to get you to think outside of the box a bit.

There’s a Build your vocab box that introduces you to a relevant term and shows you a dictionary-style definition of it. This is a term that the reader is most likely not familiar with.

Advertisement

The next block we encountered was the Test your knowledge block. This one has a quiz-style question and it gives you two options. Other subject matters might have more choices, but this is what we got in our usage.

We also saw a Common misconception block. This one pretty much explains itself.

Bottom bar

At the very bottom of the screen, you’ll see a bar with some additional chips. One chip should show the title of the current subject, and Tapping/clicking on it will bring up a floating window with additional topic suggestions. In our case, we also saw the interactive list that we saw previously. This one will show the list in a floating window.

One issue

So, do you remember when Gemini gave us the image of the motorcycles? Well, while the majority of Learn About’s images were relevant to the subject, it still retrieved two images of the motorcycles. As comical as it is, it shows that Google’s AI still has a ways to go before it’s perfect. However, barring that little mishap, Learn About runs as smoothly as the motorcycle it’s surfacing pictures of.

Advertisement

Use it today!

You can use Learn About today if you want to try it out. Just go to the Learn About website Learn About website, and you’ll be able to try it out. Just know that, as with most Google services, the availability will depend on your region. We were able to access it in the U.S. in English. Just know that you may not have it in regions that Google typically overlooks.

You can use it regardless of if you’re a free or paid user. Please note that Learn About is technically an experiment. This means that Google only put this on the market for testing. Google could potentially lock this behind a paywall after the beta testing phase. Just know that this feature could disappear down the line. So, you’ll want to get in and use it while you can.

Source link

Advertisement
Continue Reading

Technology

GOG’s preservation label highlights classic games it’s maintaining for modern hardware

Published

on

GOG's preservation label highlights classic games it's maintaining for modern hardware

GOG is launching an effort to help make older video games playable on modern hardware. The will label the classic titles that the platform has taken steps to adapt in order to make them compatible with contemporary computer systems, controllers and screen resolutions, all while adhering to its DRM-free policy. The move could bring new life to games of decades past, just as GOG did two years ago with a refresh of . So far, 92 games have received the preservation treatment.

“Our guarantee is that they work and they will keep working,” the company says in the video announcing the initiative.

Preservation has been a hot topic as more games go digital only. Not only are some platforms disk drives by default, but ownership over your library is more ephemeral than it seems. After all, most game purchases are , and licenses can be revoked (as The Crew players know ).

Source link

Advertisement
Continue Reading

Technology

How Microsoft’s next-gen BitNet architecture is turbocharging LLM efficiency

Published

on

How Microsoft's next-gen BitNet architecture is turbocharging LLM efficiency

Join our daily and weekly newsletters for the latest updates and exclusive content on industry-leading AI coverage. Learn More


One-bit large language models (LLMs) have emerged as a promising approach to making generative AI more accessible and affordable. By representing model weights with a very limited number of bits, 1-bit LLMs dramatically reduce the memory and computational resources required to run them.

Microsoft Research has been pushing the boundaries of 1-bit LLMs with its BitNet architecture. In a new paper, the researchers introduce BitNet a4.8, a new technique that further improves the efficiency of 1-bit LLMs without sacrificing their performance.

The rise of 1-bit LLMs

Traditional LLMs use 16-bit floating-point numbers (FP16) to represent their parameters. This requires a lot of memory and compute resources, which limits the accessibility and deployment options for LLMs. One-bit LLMs address this challenge by drastically reducing the precision of model weights while matching the performance of full-precision models.

Advertisement

Previous BitNet models used 1.58-bit values (-1, 0, 1) to represent model weights and 8-bit values for activations. This approach significantly reduced memory and I/O costs, but the computational cost of matrix multiplications remained a bottleneck, and optimizing neural networks with extremely low-bit parameters is challenging. 

Two techniques help to address this problem. Sparsification reduces the number of computations by pruning activations with smaller magnitudes. This is particularly useful in LLMs because activation values tend to have a long-tailed distribution, with a few very large values and many small ones.  

Quantization, on the other hand, uses a smaller number of bits to represent activations, reducing the computational and memory cost of processing them. However, simply lowering the precision of activations can lead to significant quantization errors and performance degradation.

Furthermore, combining sparsification and quantization is challenging, and presents special problems when training 1-bit LLMs. 

Advertisement

“Both quantization and sparsification introduce non-differentiable operations, making gradient computation during training particularly challenging,” Furu Wei, Partner Research Manager at Microsoft Research, told VentureBeat.

Gradient computation is essential for calculating errors and updating parameters when training neural networks. The researchers also had to ensure that their techniques could be implemented efficiently on existing hardware while maintaining the benefits of both sparsification and quantization.

BitNet a4.8

BitNet a4.8 addresses the challenges of optimizing 1-bit LLMs through what the researchers describe as “hybrid quantization and sparsification.” They achieved this by designing an architecture that selectively applies quantization or sparsification to different components of the model based on the specific distribution pattern of activations. The architecture uses 4-bit activations for inputs to attention and feed-forward network (FFN) layers. It uses sparsification with 8 bits for intermediate states, keeping only the top 55% of the parameters. The architecture is also optimized to take advantage of existing hardware.

“With BitNet b1.58, the inference bottleneck of 1-bit LLMs switches from memory/IO to computation, which is constrained by the activation bits (i.e., 8-bit in BitNet b1.58),” Wei said. “In BitNet a4.8, we push the activation bits to 4-bit so that we can leverage 4-bit kernels (e.g., INT4/FP4) to bring 2x speed up for LLM inference on the GPU devices. The combination of 1-bit model weights from BitNet b1.58 and 4-bit activations from BitNet a4.8 effectively addresses both memory/IO and computational constraints in LLM inference.”

Advertisement

BitNet a4.8 also uses 3-bit values to represent the key (K) and value (V) states in the attention mechanism. The KV cache is a crucial component of transformer models. It stores the representations of previous tokens in the sequence. By lowering the precision of KV cache values, BitNet a4.8 further reduces memory requirements, especially when dealing with long sequences. 

The promise of BitNet a4.8

Experimental results show that BitNet a4.8 delivers performance comparable to its predecessor BitNet b1.58 while using less compute and memory.

Compared to full-precision Llama models, BitNet a4.8 reduces memory usage by a factor of 10 and achieves 4x speedup. Compared to BitNet b1.58, it achieves a 2x speedup through 4-bit activation kernels. But the design can deliver much more.

“The estimated computation improvement is based on the existing hardware (GPU),” Wei said. “With hardware specifically optimized for 1-bit LLMs, the computation improvements can be significantly enhanced. BitNet introduces a new computation paradigm that minimizes the need for matrix multiplication, a primary focus in current hardware design optimization.”

Advertisement

The efficiency of BitNet a4.8 makes it particularly suited for deploying LLMs at the edge and on resource-constrained devices. This can have important implications for privacy and security. By enabling on-device LLMs, users can benefit from the power of these models without needing to send their data to the cloud.

Wei and his team are continuing their work on 1-bit LLMs.

“We continue to advance our research and vision for the era of 1-bit LLMs,” Wei said. “While our current focus is on model architecture and software support (i.e., bitnet.cpp), we aim to explore the co-design and co-evolution of model architecture and hardware to fully unlock the potential of 1-bit LLMs.”


Source link
Continue Reading

Technology

Starfish Space raises $29M to launch satellite-servicing spacecraft missions

Published

on

render of Starfish spacecraft on orbit

Starfish Space has closed a new tranche of funding led by a major defense tech investor as it looks to launch three full-size satellite servicing and inspection spacecraft in 2026. 

The Washington-based startup’s Otter spacecraft is designed for two primary missions: extending the operational life of expensive satellites in geostationary orbit (GEO) and disposing of defunct satellites in low Earth orbit (LEO). It’s a series of capabilities that have never been available for satellite operators, who launch their satellites with the expectation that they’ll only have a limited span of useful life. 

The aim, as Starfish CEO and co-founder Austin Link put it in a recent interview, is to “make it affordable enough that the benefits of having your satellite serviced outweigh the costs.”

The $29 million round was led by Shield Capital, a venture firm focused on funding technologies that will affect U.S. national security. It has participated in just a handful of other deals in the space industry. The round also includes participation from new investors Point72 Ventures, Booz Allen Ventures, Aero X Ventures, Trousdale Ventures, TRAC VC, and existing investors Munich Re Ventures, Toyota Ventures, NFX, and Industrious Ventures.

Advertisement

“You start a company because you want to build satellites, not because you want to fundraise,” Link told TechCrunch. Link founded Starfish in 2019 with Trevor Bennett after the pair worked as flight sciences engineers at Blue Origin. They raised $7 million in 2021 and $14 million two years later. Starfish launched its first demonstration mission, a sub-scale spacecraft fittingly called Otter Pup, last summer. 

Although that mission did not quite go according to plan, Starfish has racked up several wins since then, including three separate contracts for full-size Otter spacecraft. That includes a $37.5 million deal with the U.S. Space Force for a first-of-its-kind docking and maneuvering mission with a defense satellite in GEO and a contract with major satellite communications company Intelsat for life extension services. The third contract, a $15 million NASA mission to inspect multiple defunct satellites in LEO, was announced while Starfish was in the middle of fundraising, Link said. 

Starfish purposefully set out to find investors that had experience helping their portfolio companies navigate selling to the government, Link said. “The government is a customer that it sometimes can be harder to scale with, so having investors that understood the process a little better … we thought they’d be good additions to our cap table.”

Link added that the company is seeing a “fairly even split” in demand between government and commercial customers. 

Advertisement

Satellite servicing, life extension, and satellite disposal are “exciting first steps,” Link said, but they’re stepping stones on the way to developing a broader suite of capabilities for even more ambitious missions on orbit. 

“Along the way, we end up with this set of autonomy and robotics technologies and capabilities and datasets that allow us to go eventually do broadly a set of complex robotic or servicing or ISAM-type missions in space that maybe stretch a little beyond what we do with the Otter,” he said. “I think a lot of those are a long ways off, and not necessarily where our focus is right now … but some of the effort that goes into the Otter today and is funded through this funding round, and some of the growth there leads to a longer term where Starfish Space can have a broad impact on the way that humans go out into the universe.”  

Source link

Continue Reading

Technology

Apple updates Logic Pro with new sounds and search features

Published

on

Apple updates Logic Pro with new sounds and search features

Apple today announced some minor updates to Logic Pro for both the Mac and the iPad, including the ability to search for plug-ins and sources and the addition of more analog-simulating sounds.

In Logic Pro for Mac 11.1 and Logic Pro for iPad 2.1, you can now reorder channel strips and plug-ins in the mixer and plug-in windows to make it easier to organize the layout of an audio mix.

As for the new sounds, Apple added a library of analog synthesizer samples called Modular Melodies, akin to the Modular Rhythms pack already found in Logic.

A more exciting sonic addition is the new Quantec Room Simulator (QRS) plug-in, which emulates the vintage digital reverb hardware of the same name, found in professional recording studios all over the world. Apple has acquired the technology for the classic QRS model and the later YardStick models to integrate into this software.

Advertisement

I wish the QRS plug-in looked like the real life reverb unit
Image: Apple

Specific to Logic Pro for Mac, you are now able to share a song to the Mac’s Voice Memos app — which may be a great feature for when Voice Memos gets that multitrack option on the iPhone in iOS 18.2

Added to the iPad version of Logic Pro is the ability to add your own local third party sample folders to the browser window, to make it easier to bring external audio files into tracks and sampler plug-ins.

These upgrades are small for current Logic users, but they do overall make the digital audio workplace easier to use and adds to the plethora of useful tools for no additional cost. Users will have access to upgrade to Logic Pro for Mac 11.1 and Logic Pro for iPad 2.1 today.

Advertisement

Source link

Continue Reading

Trending

Copyright © 2024 WordupNews.com