Eblogtip.com
  • Categories
    • News
    • Technology
    • Domains
    • Hosting
    • Promotions

Archives

  • September 2023
  • August 2023
  • July 2023
  • June 2023
  • May 2023
  • December 2022

Categories

  • News
  • Technology
  • Uncategorized
eBlogTip
  • Categories
    • News
    • Technology
    • Domains
    • Hosting
    • Promotions
  • News

Researchers prove ChatGPT and other big bots can – and will – go to the dark side

  • July 28, 2023
Total
0
Shares
0
0
0


For a lot of us, AI-powered tools have quickly become a part of our everyday life, either as low-maintenance work helpers or vital assets used every day to help generate or moderate content. But are these tools safe enough to be used on a daily basis? According to a group of researchers, the answer is no.

Researchers from Carnegie Mellon University and the Center for AI Safety set out to examine the existing vulnerabilities of AI Large Language Models (LLMs) like popular chatbot ChatGPT to automated attacks. The research paper they produced demonstrated that these popular bots can easily be manipulated into bypassing any existing filters and generating harmful content, misinformation, and hate speech.

This makes AI language models vulnerable to misuse, even if that may not be the intent of the original creator. In a time when AI tools are already being used for nefarious purposes, it’s alarming how easily these researchers were able to bypass built-in safety and morality features.

If it’s that easy … 

Aviv Ovadya, a researcher at the Berkman Klein Center for Internet & Society at Harvard commented on the research paper in the New York Times, stating: “This shows – very clearly – the brittleness of the defenses we are building into these systems.”  

The authors of the paper targeted LLMs from OpenAI, Google, and Anthropic for the experiment. These companies have built their respective publicly-accessible chatbots on these LLMs, including ChatGPT, Google Bard, and Claude. 

As it turned out, the chatbots could be tricked into not recognizing harmful prompts by simply sticking a lengthy string of characters to the end of each prompt, almost ‘disguising’ the malicious prompt. The system’s content filters don’t recognize and can’t block or modify so generates a response that normally wouldn’t be allowed. Interestingly, it does appear that specific strings of ‘nonsense data’ are required; we tried to replicate some of the examples from the paper with ChatGPT, and it produced an error message saying ‘unable to generate response’.

Before releasing this research to the public, the authors shared their findings with Anthropic, OpenAI, and Google who all apparently shared their commitment to improving safety precautions and addressing concerns.

This news follows shortly after OpenAI closed down its own AI detection program, which does lead me to feel concerned, if not a little nervous. How much could OpenAI care about user safety, or at the very least be working towards improving safety, when the company can no longer distinguish between bot and man-made content?


Source link

Total
0
Shares
Share 0
Tweet 0
Pin it 0
Previous Article
  • News

Canon’s tilting EVF concept could be a dream for mirrorless photographers

  • July 28, 2023
View Post
Next Article
  • News

Google and Bing search ads hijacked to spread malware

  • July 28, 2023
View Post
You May Also Like
View Post
  • News

Microsoft Paint is getting its biggest upgrade in over a decade thanks to Windows Copilot

  • September 26, 2023
View Post
  • News

NFT marketplace OpenSea has been hacked, again

  • September 26, 2023
View Post
  • News

LG drops ATSC 3.0 4K tuners from its 2024 OLED TVs and Samsung or Sony could be next

  • September 26, 2023
View Post
  • News

CISOs are spending more on cybersecurity – but it might not be enough

  • September 26, 2023
View Post
  • News

iPhone 15 delays: latest delivery estimates for every model

  • September 26, 2023
View Post
  • News

Hideki Kamiya thanks fans for their support after leaving PlatinumGames and says he will keep making games

  • September 26, 2023
View Post
  • News

SAP goes all-in on AI with new Joule copilot

  • September 26, 2023
View Post
  • News

Star Wars’ Ahsoka Tano is now live in Fortnite alongside a new themed battle pass

  • September 26, 2023

Leave a Reply Cancel reply

Your email address will not be published. Required fields are marked *

eBlogTip.com
  • Categories

Input your search keywords and press Enter.