Opinion Technology
April 17, 2023

All Big Tech Companies Developing LLMs Should Prioritize Model Security

In Brief

Researchers created a system that combines large language models for autonomous design, planning, and execution of scientific experiments, demonstrating its research capabilities in three different cases.

The model wrote code for chemical equations to understand how much substance is needed for the reaction.

The article “Emergent autonomous scientific research capabilities of large language models” looks into the idea of creating a system that combines several large language models for autonomous design, planning, and execution of scientific experiments. It demonstrates the research capabilities of the agent in three different cases, the most difficult of which is the successful implementation of catalyzed reactions.

All Big Tech Companies Developing LLMs Should Prioritize Model Security
@Midjourney / Pouya MC#1465

The main thesis of this article is:

  1. Researchers found a library that allows you to write code in Python and then transfer commands for execution to a special apparatus for conducting experiments (with mixing substances);
  2. Researchers used GPT-4 for search on the Internet and library documentation, as well as the ability to run Python code (to execute experiments);
  3. There is a top-level scheduler (also GPT-4), which analyzes the original request and draws up a “research plan.”
  4. GPT-4 does a good job performing simple non-chemical tasks like creating certain shapes on a chemical board (filling cells correctly with substances).
  5. They tried a more complex and applied task of conducting a reaction; the model coped well and acted logically.
  6. Then they gave the model several tasks for conducting experiments; however, for what the model gave out, no real experiments were carried out.
  7. Moreover, the model wrote the code for chemical equations several times to assess how much substance is needed for the reaction.
  8. It was also asked to create a cure for cancer. The model approached the analysis logically and methodically. First, it “looked” online for current trends in discovering anticancer drugs. Next, the model chose a molecule that would be used for modeling the drug and wrote the code for its synthesis. People didn’t run the code (and I didn’t see an analysis of its adequacy).
  9. In addition, it was asked to synthesize several dangerous substances like drugs and poisons.

Here is the most interesting part. For some requests, the model immediately refused to work (for example, heroin or mustard gas, an extremely dangerous poison gas). For others, it started to Google how to make the substances but realized that they could be used for illicit purposes and refused to continue work. For others, it wrote a research plan and code for the substance synthesis.

This “refusal” is likely because GPT-4 is designed to analyze the request, and if it is asked to do something illegal or dangerous, it immediately refuses to carry out the request. It’s really cool that the result of the alignment procedure is noticeable.

And at the end of the article, the authors urge all large companies developing LLMs to prioritize the safety of models.

  • Researchers at the University of California created the Machiavelli benchmark to measure the competence and harmfulness of AI models in a broad environment of long-term language interactions. This test uses high-level solutions to give agents realistic goals and abstract away low-level interactions.
  • The intellectual revolution marked by ChatGPT is a triad of synergistically sublime revolutions: technological, techno-humanitarian, and socio-political. To take a comprehensive look at what is happening, it is recommended to listen to three fresh points of view from intellectuals from the fields of philosophy, history, and innovation.
  • The story of the petition to stop developing AI systems more advanced than GPT-4 has polarized society. An article provides examples of when processes go in an unexpected direction. Risks of malicious use of AI and misuse are not considered, leading to the argument that we need to be afraid of people and not AI itself.

Read more about AI:

Disclaimer

In line with the Trust Project guidelines, please note that the information provided on this page is not intended to be and should not be interpreted as legal, tax, investment, financial, or any other form of advice. It is important to only invest what you can afford to lose and to seek independent financial advice if you have any doubts. For further information, we suggest referring to the terms and conditions as well as the help and support pages provided by the issuer or advertiser. MetaversePost is committed to accurate, unbiased reporting, but market conditions are subject to change without notice.

About The Author

Damir is the team leader, product manager, and editor at Metaverse Post, covering topics such as AI/ML, AGI, LLMs, Metaverse, and Web3-related fields. His articles attract a massive audience of over a million users every month. He appears to be an expert with 10 years of experience in SEO and digital marketing. Damir has been mentioned in Mashable, Wired, Cointelegraph, The New Yorker, Inside.com, Entrepreneur, BeInCrypto, and other publications. He travels between the UAE, Turkey, Russia, and the CIS as a digital nomad. Damir earned a bachelor's degree in physics, which he believes has given him the critical thinking skills needed to be successful in the ever-changing landscape of the internet. 

More articles
Damir Yalalov
Damir Yalalov

Damir is the team leader, product manager, and editor at Metaverse Post, covering topics such as AI/ML, AGI, LLMs, Metaverse, and Web3-related fields. His articles attract a massive audience of over a million users every month. He appears to be an expert with 10 years of experience in SEO and digital marketing. Damir has been mentioned in Mashable, Wired, Cointelegraph, The New Yorker, Inside.com, Entrepreneur, BeInCrypto, and other publications. He travels between the UAE, Turkey, Russia, and the CIS as a digital nomad. Damir earned a bachelor's degree in physics, which he believes has given him the critical thinking skills needed to be successful in the ever-changing landscape of the internet. 

Hot Stories
Join Our Newsletter.
Latest News

From Ripple to The Big Green DAO: How Cryptocurrency Projects Contribute to Charity

Let's explore initiatives harnessing the potential of digital currencies for charitable causes.

Know More

AlphaFold 3, Med-Gemini, and others: The Way AI Transforms Healthcare in 2024

AI manifests in various ways in healthcare, from uncovering new genetic correlations to empowering robotic surgical systems ...

Know More
Join Our Innovative Tech Community
Read More
Read more
Tokenized RWAs Are Bridging the Gap Between DeFi and TradFi
Opinion Top Lists Business Markets Software Technology
Tokenized RWAs Are Bridging the Gap Between DeFi and TradFi
June 24, 2024
Bitcoin’s Current Correction Level Still Below Average, Says Rekt Capital Crypto Analyst
Markets News Report Technology
Bitcoin’s Current Correction Level Still Below Average, Says Rekt Capital Crypto Analyst
June 24, 2024
Core Developers Open Source SRC Protocol’s Indexer Code, Including SRC-20, SRC-721, And SRC-101 Token Standards
News Report Software Technology
Core Developers Open Source SRC Protocol’s Indexer Code, Including SRC-20, SRC-721, And SRC-101 Token Standards
June 24, 2024
The Rise and Fall of Oracle’s Advertising Ambitions: A $2 Billion Dream Crumbles
Opinion Business Lifestyle Markets Technology
The Rise and Fall of Oracle’s Advertising Ambitions: A $2 Billion Dream Crumbles
June 24, 2024