Connect with us

Technology

13 companies from YC Demo Day 1 that are worth paying attention to

Published

on

13 companies from YC Demo Day 1 that are worth paying attention to

Famed Silicon Valley startup accelerator Y Combinator on Wednesday kicked off its two-day “Demo Day” event that showcases what the most recent YC batch, S24, companies are building.

Unsurprisingly, AI companies dominated the day, with startups looking to apply the technology to problems like estate planning and settlements, Elayne; automating clinical trial data, Baseline AI; and helping companies get goods through customs, Passage.

Sectors like fintech, healthcare, and web3, which dominated YC cohorts of the past, were noticeably quieter, or completely absent, from Wednesday’s presentation.

Here are the companies worth paying attention to from the first day of Demo Day. Spoiler alert: Pretty much all use AI.

Advertisement

What it does: Automates moving baggage at airports with robots

Why it’s a fave: This seems like an ideal use case for robots, considering that collecting and moving baggage at airports is an entirely manual process, which can also be dangerous. This may also be technology that airports would actually be willing to pay for.

What it does: AI automation of clinical trial documents

Why it’s a fave: I’m a fan of anything that is aiming to make clinical trials work better and run faster, considering how important they are in the process of getting new drugs and treatments to market. The company claims it can save companies $18 million in costs and lost revenue, which seems like a notable improvement.

Advertisement

What it does: AI-powered estate planning and settlements

Why it’s a fave: As someone who has watched a family member navigate this process, I’m glad someone is building a better solution. Plus, the fact that Elayne is looking to reach consumers through their employers is a smart way to get more people thinking about this before they have to.

What it does: Automated testing for AI voice agents

Why it’s a fave: There are so many startups building customer support AI systems, but do they work? I think Hamming’s strategy of testing out these AI customer service bots is a needed service in this growing ecosystem.

Advertisement

What it does: Data centers in space

Why it’s a fave: This company stood out because it seems like an extreme moonshot, and yet it’s already landed customers and is launching a demonstrator satellite next year. The concept of using solar energy to power data centers may be one we might want to consider doing on Earth, too.

What it does: Helps cities optimize transit

Why it’s a fave: Ontra Mobility’s quest to help local governments better utilize their public transit options is a solid one. Most cities don’t have the budget to expand public transit options despite population growth, so figuring out a smarter way to utilize what options they already have makes sense.

Advertisement

What it does: AI-assisted customs support

Why it’s a fave: Considering how easy it is for consumers to get packages held up by customs, I can only imagine how complicated the importing process is for companies moving a lot of goods across the border all the time.

What it does: AI Price optimization

Why it’s a fave: This is a super interesting approach to ecommerce pricing. Promi’s AI looks to help companies offer data-informed fluctuating discounts to customers that change based on interest and activity. This makes a lot of sense.

Advertisement

What it does: TurboTax for building rebates

Why it’s a fave: Personally I’m a fan of any company that helps consumers or other companies unlock the government incentives they are eligible for. I like RetroFix’s approach in particular because it’s unlocking government money for contractors to make buildings more sustainable.

What it does: Automates government approvals for construction projects

Why it’s a fave: This is the kind of application AI was made for. SchemeFlow’s software helps construction companies automate technical reports shrinking the process to minutes. Further impressive, the young company has already generated reports for more than 400 construction projects.

Advertisement

What it does: Synthetic datasets for vision models

Why it’s a fave: There is only so much quality data available for large language models to train on, which leaves many LLM companies tempted to get data from sources they shouldn’t — or aren’t allowed to. Help stop AI companies from illegally scraping data? Sounds like a good goal to me.

What it does: Network of in-space refueling stations

Why it’s a fave: The space industry is booming; many entrepreneurs are looking to build and send satellites, rockets, and other devices up into space. Building a company that services this growing economy seems like a smart strategy.

Advertisement

What it does: Helps businesses become employee owned

Why it’s a fave: The company’s mission to help companies transition into employee owned is a novel one. Selling a company to its employees helps create wealth for the employees and generally results in a bigger payout for the seller. Sounds like a win-win.

Source link

Advertisement
Continue Reading
Advertisement
Click to comment

You must be logged in to post a comment Login

Leave a Reply

Servers computers

100+ Computer Science Concepts Explained

Published

on

100+ Computer Science Concepts Explained



Learn the fundamentals of Computer Science with a quick breakdown of jargon that every software engineer should know. Over 100 technical concepts from the CS curriculum are explained to provide a foundation for programmers.

#compsci #programming #tech

🔗 Resources

– Computer Science https://undergrad.cs.umd.edu/what-computer-science
– CS101 Stanford https://online.stanford.edu/courses/soe-ycscs101-sp-computer-science-101
– Controversial Developer Opinions https://youtu.be/goy4lZfDtCE
– Design Patterns https://youtu.be/tv-_1er1mWI

Advertisement

🔥 Get More Content – Upgrade to PRO

Upgrade to Fireship PRO at https://fireship.io/pro
Use code lORhwXd2 for 25% off your first payment.

🎨 My Editor Settings

– Atom One Dark
– vscode-icons
– Fira Code Font

Advertisement

🔖 Topics Covered

Turning Machine
CPU
Transistor
Bit
Byte
Character Encoding ASCII
Binary
Hexadecimal
Nibble
Machine Code
RAM
Memory Address
I/O
Kernel (Drivers)
Shell
Command Line Interface
SSH
Mainframe
Programming Language
Abstraction
Interpreted
Compiled
Executable
Data Types
Variable
Dynamic Typing
Static Typing
Pointer
Garbage Collector
int
signed / unsigned
float
Double
Char
string
Big endian
Little endian
Array
Linked List
Set
Stack
Queue
Hash
Tree
Graph
Nodes and Edges
Algorithms
Functions
Return
Arguments
Operators
Boolean
Expression
Statement
Conditional Logic
While Loop
For Loop
Iterable
Void
Recursion
Call Stack
Stack Overflow
Base Condition
Big-O
Time Complexity
Space Complexity
Brute Force
Divide and conquer
Dynamic Programming
Memoization
Greedy
Dijkstra’s Shortest Path
Backtracking
Declarative
Functional Language
Imperative
Procedural Language
Multiparadigm
OOP
Class
Properties
Methods
Inheritance
Design Patterns
Instantiate
Heap Memory
Reference
Threads
Parallelism
Concurrency
Bare Metal
Virtual Machine
IP Address
URL
DNS
TCP
Packets.
SSL
HTTP
API
Printers .

source

Continue Reading

Technology

Women of color running for Congress are attacked disproportionately on X, report finds

Published

on

Women of color running for Congress are attacked disproportionately on X, report finds

Women of color running for Congress in 2024 have faced a disproportionate number of attacks on X compared with other candidates, according to a new report from the nonprofit Center for Democracy and Technology (CDT) and the University of Pittsburgh.

The report sought to “compare the levels of offensive speech and hate speech that different groups of Congressional candidates are targeted with based on race and gender, with a particular emphasis on women of color.” To do this, the report’s authors analyzed 800,000 tweets that covered a three-month period between May 20 and August 23 of this year. That dataset represented all posts mentioning a candidate running for Congress with an account on X.

The report’s authors found that more than 20 percent of posts directed at Black and Asian women candidates “contained offensive language about the candidate.” It also found that Black women in particular were targeted with hate speech more often compared with other candidates.

“On average, less than 1% of all tweets that mentioned a candidate contained hate speech,” the report says. “However, we found that African-American women candidates were more likely than any other candidate to be subject to this type of post (4%).” That roughly lines up with X’s recent transparency report — the since Elon Musk took over the company — which said that rule-breaking content accounts for less than 1 percent of all posts on its platform.

Advertisement

Notably, the CDT’s report analyzed both hate speech — which ostensibly violates X’s policies — and “offensive speech,” which the report defined as “words or phrases that demean, threaten, insult, or ridicule a candidate.” While the latter category may not be against X’s rules, the report notes that the volume of suck attacks could still deter women of color from running for office. It recommends that X and other platforms take “specific measures” to counteract such effects.

“This should include clear policies that prohibit attacks against someone based on race or gender, greater transparency into how their systems address these types of attacks, better reporting tools and means for accountability, regular risk assessments with an emphasis on race and gender, and privacy preserving mechanisms for independent researchers to conduct studies using their data. The consequences of the status-quo where women of color candidates are targeted with significant attacks online at much higher rates than other candidates creates an immense barrier to creating a truly inclusive democracy.”

Source link

Advertisement
Continue Reading

Technology

DeepMind’s SCoRe shows LLMs can use their internal knowledge to correct their mistakes

Published

on

DeepMind's SCoRe shows LLMs can use their internal knowledge to correct their mistakes

Join our daily and weekly newsletters for the latest updates and exclusive content on industry-leading AI coverage. Learn More


While large language models (LLMs) are becoming increasingly effective at complicated tasks, there are many cases where they can’t get the correct answer on the first try. This is why there is growing interest in enabling LLMs to spot and correct their mistakes, also known as “self-correction.” However, current attempts at self-correction are limited and have requirements that often cannot be met in real-world situations.

In a new paper, researchers at Google DeepMind introduce Self-Correction via Reinforcement Learning (SCoRe), a novel technique that significantly improves the self-correction capabilities of LLMs using only self-generated data. SCoRe can be a valuable tool for making LLMs more robust and reliable and opens new possibilities for enhancing their reasoning and problem-solving abilities.

The importance of self-correction in LLMs

“Self-correction is a capability that greatly enhances human thinking,” Aviral Kumar, research scientist at Google DeepMind, told VentureBeat. “Humans often spend more time thinking, trying out multiple ideas, correcting their mistakes, to finally then solve a given challenging question, as opposed to simply in one-shot producing solutions for challenging questions. We would want LLMs to be able to do the same.”

Advertisement

Ideally, an LLM with strong self-correction capabilities should be able to review and refine its own answers until it reaches the correct response. This is especially important because LLMs often possess the knowledge needed to solve a problem internally but fail to use it effectively when generating their initial response.

“From a fundamental ML point of view, no LLM is expected to solve hard problems all within zero-shot using its memory (no human certainly can do this), and hence we want LLMs to spend more thinking computation and correct themselves to succeed on hard problems,” Kumar said.

Previous attempts at enabling self-correction in LLMs have relied on prompt engineering or fine-tuning models specifically for self-correction. These methods usually assume that the model can receive external feedback on the quality of the outputs or has access to an “oracle” that can guide the self-correction process.

These techniques fail to use the intrinsic self-correction capabilities of the model. Supervised fine-tuning (SFT) methods, which involve training a model to fix the mistakes of a base model, have also shown limitations. They often require oracle feedback from human annotators or stronger models and do not rely on the model’s own knowledge. Some SFT methods even require multiple models during inference to verify and refine the answer, which makes it difficult to deploy and use them.

Advertisement

Additionally, DeepMind’s research shows that while SFT methods can improve a model’s initial responses, they do not perform well when the model needs to revise its answers over multiple steps, which is often the case with complicated problems.

“It might very well happen that by the end of training the model will know how to fix the base model’s mistakes but might not have enough capabilities to detect its own mistakes,” Kumar said.

Another challenge with SFT is that it can lead to unintended behavior, such as the model learning to produce the best answer in the first attempt and not changing it in subsequent steps, even if it’s incorrect.

“We found behavior of SFT trained models largely collapses to this ‘direct’ strategy as opposed to learning how to self-correct,” Kumar said.

Advertisement

Self-correction through reinforcement learning

DeepMind SCoRe
DeepMind SCoRe framework (source: arXiv)

To overcome the limitations of previous approaches, the DeepMind researchers turned to reinforcement learning (RL). 

“LLMs today cannot do [self-correction], as is evident from prior studies that evaluate self-correction. This is a fundamental issue,” Kumar said. “LLMs are not trained to look back and introspect their mistakes, they are trained to produce the best response given a question. Hence, we started building methods for self-correction.”

SCoRe trains a single model to both generate responses and correct its own errors without relying on external feedback. Importantly, SCoRe achieves this by training the model entirely on self-generated data, eliminating the need for external knowledge.

Previous attempts to use RL for self-correction have mostly relied on single-turn interactions, which can lead to undesirable outcomes, such as the model focusing solely on the final answer and ignoring the intermediate steps that guide self-correction.

“We do see… ‘behavior collapse’ in LLMs trained to do self-correction with naive RL. It learned to simply ignore the instruction to self-correct and produce the best response out of its memory, in zero-shot, without learning to correct itself,” Kumar said.

Advertisement

To prevent behavior collapse, SCoRe uses a two-stage training process with regularization techniques. The first stage replaces SFT with a process that optimizes correction performance while ensuring that the model’s initial attempts remain close to the base model’s outputs.

The second stage employs multi-turn RL to optimize reward at both the initial and subsequent attempts while incorporating a reward bonus that encourages the model to improve its responses from the first to the second attempt.

“Both the initialization and the reward bonus ensure that the model cannot simply learn to produce the best first-attempt response and only minorly edit it,” the researchers write. “Overall, SCoRe is able to elicit knowledge from the base model to enable positive self-correction.”

SCoRe in action

The DeepMind researchers evaluated SCoRe against existing methods that use self-generated data for self-correction training. They focused on math and coding tasks, using benchmarks such as MATH, MBPP, and HumanEval.

Advertisement
DeepMind SCoRe vs other self-correct methods
DeepMind SCoRe outperforms other self-correct methods in multi-step correction. it also learns to avoid switching correct answers during the correction phase (source: arXiv)

The results showed that SCoRe significantly improved the self-correction capabilities of Gemini 1.0 Pro and 1.5 Flash models. For example, SCoRe achieved a 15.6% absolute gain in self-correction on the MATH benchmark and a 9.1% gain on the HumanEval benchmark in comparison to the base model, beating other self-correction methods by several percentage points.

The most notable improvement was in the model’s ability to correct its mistakes from the first to the second attempt. SCoRe also considerably reduced the instances where the model mistakenly changed a correct answer to an incorrect one, indicating that it learned to apply corrections only when necessary.

Furthermore, SCoRe proved to be highly efficient when combined with inference-time scaling strategies such as self-consistency. By splitting the same inference budget across multiple rounds of correction, SCoRe enabled further performance gains.

DeepMind SCoRe inference-time scaling
SCoRe (green line) enables LLMs to make better use of inference-time scaling techniques (source: arXiv)

While the paper primarily focuses on coding and reasoning tasks, the researchers believe that SCoRe can be beneficial for other applications as well.

“You could imagine teaching models to look back at their outputs that might potentially be unsafe and improve them all by themselves, before showing it to the user,” Kumar said.

The researchers believe that their work has broader implications for training LLMs and highlights the importance of teaching models how to reason and correct themselves rather than simply mapping inputs to outputs. 

Advertisement

Source link
Advertisement
Continue Reading

Servers computers

GME Stock: DRS to ComputerShare Why It Should Matter to You

Published

on

GME Stock: DRS to ComputerShare Why It Should Matter to You



This video goes over why ComputerShare matters and what it actually means to register your shares in your own name. All in all, just providing my GME update and documenting the journey of my own personal investment in GameStop.

This channel does NOT provide financial advice. I do not provide financial advice. Only sharing my thoughts and opinions, always do your own research!

Instagram:

Twitter:


source

Continue Reading

Technology

Numa raises $32M to bring AI and automation to car dealerships

Published

on

Numa raises $32M to bring AI and automation to car dealerships

Sometimes, a pivot ends up being the smartest decision company leaders can make. See Netflix’s pivot from DVDs to streaming, or Corning’s pivot from lightbulbs to touchscreens.

The list of extremely successful startup pivots goes on. And on. And on.

A less-prominent (but by no means failed) pivot is Numa’s. Its co-founders killed the startup’s original conversational AI product to instead sell customer service automation tools. Not just any tools, though — these tools are targeted at auto dealerships.

That sounds like a highly specific niche, but it’s been profitable, according to Tasso Roumeliotis, Numa’s CEO. The company closed a $32 million Series B round in September.

Advertisement

“We were early to build AI and conversational commerce,” Roumeliotis told TechCrunch in an interview. “But we decided to focus our AI entirely on the automotive vertical after identifying enormous opportunity in that space.”

Roumeliotis co-founded Numa in 2017 with Andy Ruff, Joel Grossman, and Steven Ginn. Grossman hails from Microsoft, where he helped ship headliner products like Windows XP, as well as a few less recognizable ones like MSN Explorer. Ruff, another Microsoft veteran, led the team that created the first Outlook for Mac client.

Numa is actually the co-founders’ second venture together. Roumeliotis, Grossman, Ginn, and Ruff previously started Location Labs, a family-focused security company that AVG bought for $220 million 10 years ago.

What rallied the old crew behind Numa, Roumeliotis says, was a shared belief in the potential of “thoughtfully applied” AI to transform entire industries. “The market is full of AI and automation point solutions or broad, unfocused tools,” he said. “Numa offers an end-to-end solution that prioritizes the needs of the customer: car dealerships.”

Advertisement

The U.S. has more than 17,000 new-car dealerships, representing a $1.2 trillion industry. Yet many dealerships struggle to manage customer service requests. Per one survey, a third of dealers miss at least a fifth of their incoming calls.

Poor responsiveness leads to low customer service scores, which in turn hurt sales. But Numa can prevent things from getting that bad — or so Roumeliotis claims — by tackling the low-hanging fruit.

Image Credits: Numa

Numa uses AI to automate tasks such as “rescuing” missed calls and booking service appointments. For example, if a customer rings a dealership but hangs up immediately afterward, Numa can send a follow-up text or automatically place a reminder call. The platform can also give customers status updates on ongoing service, and facilitate trade-ins by collecting any necessary information ahead of time.

“Many dealerships still rely on legacy systems that are inefficient and lack integration with modern, AI-driven platforms,” Roumeliotis said. “Today’s consumers expect fast, seamless interactions across all platforms. Dealerships struggle to meet these expectations, especially in areas like real-time communication, service updates, and personalized experiences, which AI can help address.”

Other small-time automation vendors (e.g., Brooke.ai, Stella AI) provide products designed to ease dealerships’ customer service burdens. Tech giants, meanwhile, sell a range of generic solutions to automate away customer service. But Roumeliotis argues that Numa stands out because it understands how workflows within dealerships impact the end-customer experience.

Advertisement

“Dealership service leaders and employees are running around constantly, handling customers in person, going out to check on cars and parts, dealing with ringing phones, and balancing coordination with co-workers,” Roumeliotis said. “Numa brings that all together in a way intentionally designed with AI and the user inside the dealership to drive how the platform works rather than the other way around.”

Roumeliotis asserts Numa has another advantage in its in-house models, which drive the platform’s automations. He said the models were trained on datasets from OEMs and dealership systems as well as conversation data between dealerships and clients.

Were each one of these clients, OEMs, and dealerships informed that their data would be used to train Numa’s models? Roumeliotis declined to say. “Numa’s models are bootstrapped by a feedback loop between dealerships, customers interacting with dealerships, and the usage of Numa to facilitate this,” he said.

Numa
Image Credits: Numa

That answer probably won’t satisfy privacy-conscious folk, but it’s seemingly immaterial to many dealerships. Numa has 600 customers across the U.S. and Canada, including the largest retail auto dealership in the world. Roumeliotis claims Numa is “just about” cash-flow break-even.

“We don’t need capital to continue scaling revenue,” he added. “Instead, Numa is using its money to accelerate product development by expanding our team of AI and machine learning engineers, including investing in building AI models for the automotive vertical.” The company currently has 70 employees.

Advertisement

Benefiting Numa in its conquest is the willingness of dealerships to pilot AI to abstract away certain back-office work.

According to a survey by automotive software provider CDK Global last year, 67% of dealerships are using AI to identify sales leads, while 63% have deployed it for service. Those responding to the poll were quite bullish on the tech overall, with close to two-thirds saying that they anticipated positive returns.

Touring Capital and Mitsui, a Japanese conglomerate that’s one of the largest shareholders in automaker Penske, led Numa’s Series B round. Costanoa Ventures, Threshold Ventures, and Gradient, Google’s AI-focused venture fund, also participated in the round. The funding brings Oakland-based Numa’s total raised to $48 million.

Source link

Advertisement

Continue Reading

Servers computers

Full Form of Computer 💻 ||

Published

on

Full Form of Computer 💻 ||



Full Form of Computer 💻 ||
.
.
.
.
.
.
.
.
.
.
.
.
#computer
#fullform
#computerfullform
#computerfullforms
#gk
#gkshorts
#generalknowledge
#important
#importantfullform
#study
#studymotivation
#educational
#educationalvideo
#laptop
#computerfullform .

source

Continue Reading

Trending

Copyright © 2024 WordupNews.com