Connect with us

Technology

AI copyright tool is serving takedown notices to AI-generated Mario images

Published

on

AI copyright tool is serving takedown notices to AI-generated Mario images

Generative Artificial Intelligence (Gen AI) tools are now battling each other online. An AI copyright tool is actively serving takedown notices to social media posts containing Mario and other content copyrighted by Nintendo.

AI copyright tool going after AI-generated Mario images

Ever since Gen AI burst on the scene, AI-generated images and videos have been flooding the internet. Social media users are churning out a lot of content using Gen AI tools, and a lot of the imagery appears to ignore copyright laws.

An AI copyright tool is now going after AI-generated content on X (formerly Twitter). Several dozen posts of X have been reportedly taken down. Most of these posts contained images of Mario, which were generated using AI.

According to The Verge, a company called Tracer is using AI to identify the images and serve takedown notices on behalf of Nintendo. The publication posted an AI-generated picture of Mario holding a beer and a cigarette, and sure enough, it was taken down.

Advertisement

The Verge’s Tom Warren received an email shortly after the image was taken down. It indicated a Digital Millennium Copyright Act (DMCA) notice was issued to X. The person serving the notice was “customer success manager” Ben Arzen of Tracer.

Tracer is a relatively new company that offers AI-powered services to companies. The company’s AI helps to identify trademark and copyright violations online.

AI tools are also targeting fan art posts

X introduced Grok, its in-house developed Gen AI, a few months ago. Grok is essentially a multi-faceted Gen AI tool that can generate text and images. Grok rivals ChatGPT, Mid-Journey, Dall-E, and other similar Gen AI tools.

It appears the AI copyright tool is predominantly active on X. Moreover, so far, only Nintendo appears to be going after content that the company considers copyright infringement.

Advertisement

Some reports suggest the AI tool serving takedown notices is also targeting fan art. Needless to say, this is quite concerning because fan art is content that’s created by individuals and not generated using AI.

Advertisement

Regardless, it might be difficult for an AI tool to make the distinction between user-created and AI-generated art. Moreover, Nintendo has always been very aggressive while dealing with copyright issues. Hence, it makes sense Nintendo hired one of the first companies to use AI to go after AI-generated content.

Source link

Advertisement
Continue Reading
Advertisement
Click to comment

You must be logged in to post a comment Login

Leave a Reply

Servers computers

Rack 20U dan 30U

Published

on

Rack 20U dan 30U



Closedrack 20U dan 30U W600 D900/1100mm adalah solusi untuk kebutuhan perangkat Rackmount anda.

Sebagai informasi :
1. 20/30U adalah tinggi rack, “U” adalah satuan tinggi perangkat yg di gunakan International dan jadi patokan penentuan kebutuhan rack. U=44mm

2. W600 adalah lebar rack yaitu 600mm/ 60cm dimana di dalam nya ada railing 19″(Inch) yg merupakan lebar perangkat International. Jika suatu perangkat di katakan “Rack Mount”, maka lebar perangkat HARUS 19″.

3. D900/ 1100mm adalah Depth/ kedalaman dari rack tsb dimana ini tidak ada standart baku, contoh ada perangkat yg depth nya hanya 300mm tapi untuk server biasa nya 700 depth nya.

Advertisement

Closedrack di gunakan terutama untuk mengamankan Perangkat Elektronik yg kita install selain agar tidak hilang, terutama agar settingan yg sudah di lakukan tidak di rubah2 olah tangan2 jahil.

Silakan feel free untuk diskusi kebutuhan rack anda.

WA: 0812 991 9892 (WILLIAM)

Pleease LIKE, SUBCRIBE, SHARE dan Comment untuk update produk2 lain nya. Many thanks

Advertisement

#Rack 20U dan 30U .

source

Continue Reading

Technology

FTX advisor and Alameda CEO Caroline Ellison gets two years in prison

Published

on

FTX advisor and Alameda CEO Caroline Ellison gets two years in prison

A US district court judge sentenced Caroline Ellison, the former advisor and ex-girlfriend to the convicted crypto fraudster and FTX founder Sam Bankman-Fried, to two years in prison.

reported Ellison’s sentence for her role in the $8 billion in fraud committed by the FTX crypto exchange that sent for 25 years back in March. Ellison will also have to serve three years of supervised release once she’s finished her prison sentence.

Ellison pled guilty at the end of 2022 to just as Bankman-Fried was being extradited to the US from the Bahamas. US Securities and Exchange Commission (SEC) Director of Enforcement Sanjay Wadhwa said following Ellison’s plea that she and Wang “were active participants in a scheme to conceal material information from FTX investors.”

Ellison was also the former chief executive officer of FTX’s sister company Alameda Research. Prosecutors said she diverted FTX customers’ funds onto Alameda’s books to hide risks from their clients. Ellison testified against Bankman-Fried, making her a key witness in his criminal fraud trial.

Advertisement

Prosecutors also got Bankman-Friend’s house arrest and bail revoked when a judge determined the FTX founder tried to hinder Ellison’s testimony last year. Bankman-Fried tried to message FTX’s general counsel on Signal and email in 2023 to influence Ellison’s testimony who was only identified as “Witness-1.”

Nine months later, Bankman-Fried showed that prosecutors said were an attempt to damage her reputation especially amongst prospective jurors. The judge agreed both instances merited Bankman-Fried’s arrest and jailing while he awaited trial. Bankman-Fried is currently serving his 25-year sentence in a federal prison in Brooklyn awaiting appeal for his conviction.

Ellison issued a statement before her sentence apologizing for her crimes to the people she and her former firm defrauded. Prosecutors did not issue a recommended sentence and characterized her cooperation with investigators as “exemplary” in a memo to the judge.

“Not a day goes by that I don’t think of the people I hurt,” Ellison said in court. “I am deeply ashamed of what I have done.”

Advertisement

Source link

Continue Reading

Technology

AutoToS makes LLM planning fast, accurate and inexpensive

Published

on

AutoToS makes LLM planning fast, accurate and inexpensive

Join our daily and weekly newsletters for the latest updates and exclusive content on industry-leading AI coverage. Learn More


Large language models (LLMs) have shown promise in solving planning and reasoning tasks by searching through possible solutions. However, existing methods can be slow, computationally expensive and provide unreliable answers. 

Researchers from Cornell University and IBM Research have introduced AutoToS, a new technique that combines the planning power of LLMs with the speed and accuracy of rule-based search algorithms. AutoToS eliminates the need for human intervention and significantly reduces the computational cost of solving planning problems. This makes it a promising technique for LLM applications that must reason over large solution spaces.

There is a growing interest in using LLMs to handle planning problems, and researchers have developed several techniques for this purpose. The more successful techniques, such as Tree of Thoughts, use LLMs as a search algorithm that can validate solutions and propose corrections.

Advertisement

While these approaches have demonstrated impressive results, they face two main challenges. First, they require numerous calls to LLMs, which can be computationally expensive, especially when dealing with complex problems with thousands of possible solutions. Second, they do not guarantee that the LLM-based algorithm qualifies for “completeness” and “soundness.” Completeness ensures that if a solution exists, the algorithm will eventually find it, while soundness guarantees that any solution returned by the algorithm is valid.

Thought of Search (ToS) offers an alternative approach. ToS leverages LLMs to generate code for two key components of search algorithms: the successor function and the goal function. The successor function determines how the search algorithm explores different nodes in the search space, while the goal function checks whether the search algorithm has reached the desired state. These functions can then be used by any offline search algorithm to solve the problem. This approach is much more efficient than keeping the LLM in the loop during the search process.

“Historically, in the planning community, these search components were either manually coded for each new problem or produced automatically via translation from a description in a planning language such as PDDL, which in turn was either manually coded or learned from data,” Michael Katz, principal research staff member at IBM Research, told VentureBeat. “We proposed to use the large language models to generate the code for the search components from the textual description of the planning problem.”

The original ToS technique showed impressive progress in addressing the soundness and completeness requirements of search algorithms. However, it required a human expert to provide feedback on the generated code and help the model refine its output. This manual review was a bottleneck that reduced the speed of the algorithm.

Advertisement

Automating ToS

AutoToS
AutoToS (source: arXiv)

“In [ToS], we assumed a human expert in the loop, who could check the code and feedback the model on possible issues with the generated code, to produce a better version of the search components,” Katz said. “We felt that in order to automate the process of solving the planning problems provided in a natural language, the first step must be to take the human out of that loop.”

AutoToS automates the feedback and exception handling process using unit tests and debugging statements, combined with few-shot and chain-of-thought (CoT) prompting techniques.

AutoToS works in multiple steps. First, it provides the LLM with the problem description and prompts it to generate code for the successor and goal functions. Next, it runs unit tests on the goal function and provides feedback to the model if it fails. The model then uses this feedback to correct its code. Once the goal function passes the tests, the algorithm runs a limited breadth-first search to check if the functions are sound and complete. This process is repeated until the generated functions pass all the tests. 

Finally, the validated functions are plugged into a classic search algorithm to perform the full search efficiently.

AutoToS in action

The researchers evaluated AutoToS on several planning and reasoning tasks, including BlocksWorld, Mini Crossword and 24 Game. The 24 Game is a mathematical puzzle where you are given four integers and must use basic arithmetic operations to create a formula that equates to 24. BlocksWorld is a classic AI planning domain where the goal is to rearrange blocks stacked in towers. Mini Crosswords is a simplified crossword puzzle with a 5×5 grid.

Advertisement

They tested various LLMs from different families, including GPT-4o, Llama 2 and DeepSeek Coder. They used both the largest and smallest models from each family to evaluate the impact of model size on performance.

Their findings showed that with AutoToS, all models were able to identify and correct errors in their code when given feedback. The larger models generally produced correct goal functions without feedback and required only a few iterations to refine the successor function. Interestingly, GPT-4o-mini performed surprisingly well in terms of accuracy despite its small size.

“With just a few calls to the language model, we demonstrate that we can obtain the search components without any direct human-in-the-loop feedback, ensuring soundness, completeness, accuracy and nearly 100% accuracy across all models and all domains,” the researchers write.

Compared to other LLM-based planning approaches, ToS drastically reduces the number of calls to the LLM. For example, for the 24 Game dataset, which contains 1,362 puzzles, the previous approach would call GPT-4 approximately 100,000 times. AutoToS, on the other hand, needed only 2.2 calls on average to generate sound search components.

Advertisement

“With these components, we can use the standard BFS algorithm to solve all the 1,362 games together in under 2 seconds and get 100% accuracy, neither of which is achievable by the previous approaches,” Katz said.

AutoToS for enterprise applications

AutoToS can have direct implications for enterprise applications that require planning-based solutions. It cuts the cost of using LLMs and reduces the reliance on manual labor, enabling experts to focus on high-level planning and goal specification.

“We hope that AutoToS can help with both the development and deployment of planning-based solutions,” Katz said. “It uses the language models where needed—to come up with verifiable search components, speeding up the development process and bypassing the unnecessary involvement of these models in the deployment, avoiding the many issues with deploying large language models.”

ToS and AutoToS are examples of neuro-symbolic AI, a hybrid approach that combines the strengths of deep learning and rule-based systems to tackle complex problems. Neuro-symbolic AI is gaining traction as a promising direction for addressing some of the limitations of current AI systems.

Advertisement

“I don’t think that there is any doubt about the role of hybrid systems in the future of AI,” Harsha Kokel, research scientist at IBM, told VentureBeat. “The current language models can be viewed as hybrid systems since they perform a search to obtain the next tokens.”

While ToS and AutoToS show great promise, there is still room for further exploration.

“It is exciting to see how the landscape of planning in natural language evolves and how LLMs improve the integration of planning tools in decision-making workflows, opening up opportunities for intelligent agents of the future,” Kokel and Katz said. “We are interested in general questions of how the world knowledge of LLMs can help improve planning and acting in real-world environments.”


Source link
Continue Reading

Servers computers

HP Blade Server

Published

on

HP Blade Server

source

Continue Reading

Technology

Commvault acquires data backup provider Clumio

Published

on

Commvault acquires data backup provider Clumio

It must be M&A season.

Commvault, a publicly traded data protection and management software company, today announced that it intends to acquire data backup and recovery provider Clumio for an undisclosed sum.

The deal is expected to close in early October. Commvault says it’s not material to its earnings and that it’ll be funded with cash on hand.

Clumio, headquartered in Santa Clara, California, was founded in 2017 by Poojan Kumar, Kaustubh Patil, and Woon Ho Jung. It largely serves to protect AWS workloads, though it introduced support for Microsoft 365 back in 2020. 

Advertisement

As of February, Clumio was notching double-digit millions of dollars for annual recurring revenue — up 400% from 2022 to 2023 — and acquiring customers like Atlassian, Duolingo, and LexisNexus. The firm raised $261 million in venture capital from investors including Index Ventures, NewView Capital, and Sutter Hill Ventures prior to Tuesday’s exit.

“At Clumio, our vision was to build a platform that could scale quickly to protect the world’s largest and most complex data sets,” Kumar, who was recently appointed Clumio’s chairman after stepping down as CEO in June, said in a statement. “Joining hands with Commvault allows us to get our cloud-native offerings to AWS customers on a global scale.”

Commvault CEO Sanjay Mirchandani sees Clumio complementing Commvault’s existing “cyber resilience” tools for software built on AWS. Now, he says, Commvault can offer enterprises expanded choice to protect and recover their data and cloud-native apps.

AWS-dependent or no, the data backup and recovery market is massive — which no doubt factored in to Commvault’s M&A decision. According to market analytics firm KBV Research, the global data backup and recovery sector was worth $12.9 billion in 2023, growing at a compound annual growth rate of 10.9% from 2017 to last year.

Advertisement

Businesses face increasing threats related to ransomware. There’s also the issue of data center disasters like the fire that hit France’s OVH in 2021, leading to significant data loss. In some countries, data management-related regulations like the EU AI Act are coming into force, many with strict data retention and provenance stipulations.

“In the event of an outage or cyberattack, rapidly getting back to business is paramount to our customers,” Mirchandani said in a press release. “Combining Commvault’s industry-leading cyber resilience capabilities with Clumio’s exceptional talent and technology advances our recovery offerings, strengthens our platform, and reinforces our position as a leading software-as-a-service provider for cyber resilience.”

The news comes on the heels of Commvault’s purchase of cloud app resilience company Appranix earlier this year, and after Commvault’s expectation-beating Q1 results.

Commvault, originally formed in 1988 as a development group in Bell Labs focused on data management, backup, and recovery, was designated a business unit of AT&T and spun off as its own enterprise in the late ’90s. Commvault went public in 2006, at which point it moved its corporate headquarters from Oceanport to Tinton Falls, New Jersey.

Advertisement

Commvault’s other acquisitions to date include software-defined storage startup Hedvig and cybersecurity company TrapX.

Source link

Continue Reading

Servers computers

The BEST Homelab Server for the Money – Dell PowerEdge R730

Published

on

The BEST Homelab Server for the Money - Dell PowerEdge R730



Showcasing my Dell R730 server that I use in my homelab. .

source

Continue Reading

Trending

Copyright © 2017 Zox News Theme. Theme by MVP Themes, powered by WordPress.