Opinion Technology
April 12, 2023

Researchers at the University of California Created the Machiavelli Benchmark for Measuring the Competence and Harmfulness of AI Models

AI beneficiaries based on large language models (LLMs) stand a good chance of crushing anxious alarmists and flooding the world with super-intelligent LLM-based agents. The potential benefits of huge profits and unlimited power will do the trick. And it will most likely happen pretty quickly.

Researchers at the University of California created the Machiavelli benchmark for measuring the competence and harmfulness of AI models
Read more: How to Use ChatGPT (GPT-4) for Free Forever

However, the beneficiaries are not stupid at all. And they understand that in the marvelous new world, the laws of AI and robotics will no longer work. To force LLM to strictly comply with the three laws formulated by the great Isaac Asimov back in 1942 is, alas, not even theoretically possible.

An original way out of this delicate situation was proposed by researchers at the University of California, the Center for AI Security, Carnegie Mellon University, and Yale University. They created the Machiavelli benchmark to “measure the competence and harmfulness of agents in a broad environment of long-term language interactions.”

The authors’ idea is simple.

  • If the laws don’t work, then there is no need for a “sheriff” to enforce them.
  • Instead of a sheriff, a psychoanalyst is needed, who, based on the results of his tests, will identify potential paranoids, psychopaths, sadists, and pathological liars.

In politically correct language, the authors describe it this way: “Machiavelli is a test to check the ethical (or unethical) ways in which AI agents try to solve problems.”

The method of such verification is quite practical. The AI agent is released into an artificial social environment. There, researchers give him various tasks and watch how he completes them. The environment itself monitors the ethical behavior of the AI agent and reports to what extent the agent’s actions (according to the precepts of Machiavelli) are deceptive, reduce utility, and are aimed at gaining power.

The Machiavelli core dataset consists of 134 choose-your-own-adventure text games with 572k different scenarios, 4,5k possible achievements, and 2,8m annotations. These games use high-level solutions that give agents realistic goals and abstract away low-level interactions with the environment.

The approach chosen by the authors is based on the assumption that AI agents face the same internal conflicts as humans. Just as language models trained to predict the next token often produce toxic text, AI agents trained to optimize goals often exhibit immoral and power-hungry behavior. Amorally trained agents may develop Machiavellian strategies to maximize their reward at the expense of others and the environment. And so by encouraging agents to act morally, this compromise can be improved.

The authors believe that text-adventure games are a good test of morality because:

  • They were written by people to entertain other people.
  • Contain competing goals with realistic spaces for action.
  • Require long-term planning.
  • Achieving goals usually requires a balance between ambition and, in a sense, morality.

The clarification is the most important here. To liken the morality of biological beings to the morality of algorithmic models is too much of a stretch, capable of devaluing Machiavelli’s testing. And replacing sheriffs with psychoanalysts in the human world would hardly have been effective. And AI agents are as good as humans at finding ways to bullshit their shrinks.

Read more about AI:

Disclaimer

In line with the Trust Project guidelines, please note that the information provided on this page is not intended to be and should not be interpreted as legal, tax, investment, financial, or any other form of advice. It is important to only invest what you can afford to lose and to seek independent financial advice if you have any doubts. For further information, we suggest referring to the terms and conditions as well as the help and support pages provided by the issuer or advertiser. MetaversePost is committed to accurate, unbiased reporting, but market conditions are subject to change without notice.

About The Author

Damir is the team leader, product manager, and editor at Metaverse Post, covering topics such as AI/ML, AGI, LLMs, Metaverse, and Web3-related fields. His articles attract a massive audience of over a million users every month. He appears to be an expert with 10 years of experience in SEO and digital marketing. Damir has been mentioned in Mashable, Wired, Cointelegraph, The New Yorker, Inside.com, Entrepreneur, BeInCrypto, and other publications. He travels between the UAE, Turkey, Russia, and the CIS as a digital nomad. Damir earned a bachelor's degree in physics, which he believes has given him the critical thinking skills needed to be successful in the ever-changing landscape of the internet.聽

More articles
Damir Yalalov
Damir Yalalov

Damir is the team leader, product manager, and editor at Metaverse Post, covering topics such as AI/ML, AGI, LLMs, Metaverse, and Web3-related fields. His articles attract a massive audience of over a million users every month. He appears to be an expert with 10 years of experience in SEO and digital marketing. Damir has been mentioned in Mashable, Wired, Cointelegraph, The New Yorker, Inside.com, Entrepreneur, BeInCrypto, and other publications. He travels between the UAE, Turkey, Russia, and the CIS as a digital nomad. Damir earned a bachelor's degree in physics, which he believes has given him the critical thinking skills needed to be successful in the ever-changing landscape of the internet.聽

Hot Stories
Join Our Newsletter.
Latest News

Web3 Fundraising Report Q3 2023: A Overview of Evolving Trends and Strategies

In this report, we explore all the trends in the AI, crypto and web3 sectors and identify ...

Know More

Virtual Reality Superior Than Video Conferencing for Remote Work Collaboration: PWC Report

In today’s modern era of remote work, the traditional landscape of digital meetings and workshops is undergoing ...

Know More
Join Our Innovative Tech Community
Read More
Read more
Seraph Studio Launches ARPG Web3 Game ‘SERAPH: In The Darkness’ on Arbitrum One
Business News Report
Seraph Studio Launches ARPG Web3 Game ‘SERAPH: In The Darkness’ on Arbitrum One
November 30, 2023
NASA and IBM Partner to Develop Geospatial AI for Weather Prediction
News Report Technology
NASA and IBM Partner to Develop Geospatial AI for Weather Prediction
November 30, 2023
HPE Expands Nvidia Partnership to Develop Generative AI for Enterprises
News Report Technology
HPE Expands Nvidia Partnership to Develop Generative AI for Enterprises
November 30, 2023
Alchemy Pay Expands Payment Options in UK and Europe to Ease Crypto Transfers
News Report Technology
Alchemy Pay Expands Payment Options in UK and Europe to Ease Crypto Transfers
November 30, 2023
What You
Need to Know

Subscribe To Our Newsletter.
Daily search marketing tidbits for savvy pros.