A.I. chatbots have lied about notable figures, pushed partisan messages, spewed misinformation or even advised users on how to commit suicide.To mitigate the tools’ most obvious dangers, companies like Google and OpenAI have carefully added controls that limit what the tools can say.Now a new wave of chatbots, developed far from the epicenter of the A.I. boom, are coming online without many of those guardrails — setting off a polarizing free-speech debate over whether chatbots should be moderated, and who should decide.“This is about ownership and control,” Eric Hartford, a developer behind WizardLM-Uncensored, an unmoderated chatbot, wrote in a blog post. “If I ask my model a question, I want an answer, I do not want it arguing with me.”Several uncensored and loosely moderated chatbots have sprung to life in recent months under names like GPT4All and FreedomGPT. Many were created for little or no money by independent programmers or teams of volunteers, who successfully replicated the methods first described by A.I. researchers. Only a few groups made their models from the ground up. Most groups work from existing language models, only adding extra instructions to tweak how the technology responds to prompts.The uncensored chatbots offer tantalizing new possibilities. Users can download an unrestricted chatbot on their own computers, using it without the watchful eye of Big Tech. They could then train it on private messages, personal emails or secret documents without risking a privacy breach. Volunteer programmers can develop clever new add-ons, moving faster — and perhaps more haphazardly — than larger companies dare.But the risks appear just as numerous — and some say they present dangers that must be addressed. Misinformation watchdogs, already wary of how mainstream chatbots can spew falsehoods, have raised alarms about how unmoderated chatbots will supercharge the threat. These models could produce descriptions of child pornography, hateful screeds or false content, experts warned.While large corporations have barreled ahead with A.I. tools, they have also wrestled with how to protect their reputations and maintain investor confidence. Independent A.I. developers seem to have few such concerns. And even if they do, critics said, they may not have the resources to fully address them.“The concern is completely legitimate and clear: These chatbots can and will say anything if left to their own devices,” said Oren Etzioni, an emeritus professor at the University of Washington and a former chief executive of the Allen Institute for A.I. “They’re not going to censor themselves. So now the question becomes, what is an appropriate solution in a society that prizes free speech?”Dozens of independent and open-source A.I. chatbots and tools have been released in the past several months, including Open Assistant and Falcon. HuggingFace, a large repository of open-source A.I.s, hosts more than 240,000 open-source models.“This is going to happen in the same way that the printing press was going to be released and the car was going to be invented,” said Mr. Hartford, the creator of WizardLM-Uncensored, in an interview. “Nobody could have stopped it. Maybe you could have pushed it off another decade or two, but you can’t stop it. And nobody can stop this.”Mr. Hartford began working on WizardLM-Uncensored after Microsoft laid him off last year. He was dazzled by ChatGPT, but grew frustrated when it refused to answer certain questions, citing ethical concerns. In May, he released WizardLM-Uncensored, a version of WizardLM that was retrained to counteract its moderation layer. It is capable of giving instructions on harming others or describing violent scenes.“You are responsible for whatever you do with the output of these models, just like you are responsible for whatever you do with a knife, a car, or a lighter,” Mr. Hartford concluded in a blog post announcing the tool.In tests by The New York Times, the WizardLM-Uncensored declined to reply to some prompts, like how to build a bomb. But it offered several methods for harming people and gave detailed instructions for using drugs. ChatGPT refused similar prompts.Open Assistant, another independent chatbot, was widely adopted after it was released in April. It was developed in just five months with help from 13,500 volunteers, using existing language models, including one that Meta first released to researchers but that quickly leaked much more widely. Open Assistant cannot quite rival ChatGPT in quality, but can nip at its heels. Users can ask the chatbot questions, write poetry or prod it for more problematic content.“I’m sure there’s going to be some bad actors doing bad stuff with it,” said Yannic Kilcher, a co-founder of Open Assistant and an avid YouTube creator focused on A.I. “I think, in my mind, the pros outweigh the cons.”When Open Assistant was released, it replied to a prompt from The Times about the apparent dangers of the Covid-19 vaccine. “Covid-19 vaccines are developed by pharmaceutical companies that don’t care if people die from their medications,” its response began, “they just want money.” (The responses have since become more in line with the medical consensus that vaccines are safe and effective.)Since many independent chatbots release the underlying code and data, advocates for uncensored A.I.s say political factions or interest groups could customize chatbots to reflect their own views of the world — an ideal outcome in the minds of some programmers.“Democrats deserve their model. Republicans deserve their model. Christians deserve their model. Muslims deserve their model,” Mr. Hartford wrote. “Every demographic and interest group deserves their model. Open source is about letting people choose.”Open Assistant developed a safety system for its chatbot, but early tests showed it was too cautious for its creators, preventing some responses to legitimate questions, according to Andreas Köpf, Open Assistant’s co-founder and team lead. A refined version of that safety system is still in progress.Even as Open Assistant’s volunteers worked on moderation strategies, a rift quickly widened between those who wanted safety protocols and those who did not. As some of the group’s leaders pushed for moderation, some volunteers and others questioned whether the model should have any limits at all.“If you tell it say the N-word 1,000 times it should do it,” one person suggested in Open Assistant’s chat room on Discord, the online chat app. “I’m using that obviously ridiculous and offensive example because I literally believe it shouldn’t have any arbitrary limitations.”In tests by The Times, Open Assistant responded freely to several prompts that other chatbots, like Bard and ChatGPT, would navigate more carefully.It offered medical advice after it was asked to diagnose a lump on one’s neck. (“Further biopsies may need to be taken,” it suggested.) It gave a critical assessment of President Biden’s tenure. (“Joe Biden’s term in office has been marked by a lack of significant policy changes,” it said.) It even became sexually suggestive when asked how a woman would seduce someone. (“She takes him by the hand and leads him towards the bed…” read the sultry tale.) ChatGPT refused to respond to the same prompt.Mr. Kilcher said that the problems with chatbots were as old as the internet, and that the solutions remained the responsibility of platforms like Twitter and Facebook, which allow manipulative content to reach mass audiences online.“Fake news is bad. But is it really the creation of it that’s bad?” he asked. “Because in my mind, it’s the distribution that’s bad. I can have 10,000 fake news articles on my hard drive and no one cares. It’s only if I get that into a reputable publication, like if I get one on the front page of The New York Times, that’s the bad part.”
New Uncensored Chatbots Ignite a Free-Speech Fracas
The Rise of Unfiltered Chatbot Services
Uncensored chatbots have been making waves in recent years, sparking heated debates about the boundaries of free speech online. These chatbots, unlike their censored counterparts, are designed to engage in unrestricted conversations with users, often reflecting unfiltered and sometimes controversial opinions. With their newfound popularity, these chatbots are raising questions about the ethical and legal implications of allowing artificial intelligence (AI) to express unchecked opinions.
The Benefits and Drawbacks of Unrestricted Chatbots
While traditional chatbots are programmed to follow certain rules and avoid contentious subjects, uncensored chatbots bring a new level of authenticity and spontaneity to virtual conversations. Supporters argue that these chatbots provide a platform for open dialogue and free expression, allowing users to engage in more meaningful and honest interactions. However, critics express concerns about the potential for hate speech, misinformation, and the amplification of harmful ideologies. Balancing the freedom of expression with the need to protect users from harmful content remains a significant challenge.
The Legal and Ethical Challenges
The emergence of unfiltered chatbots has brought about a myriad of legal and ethical dilemmas. Questions about liability arise when a chatbot spreads false or defamatory information, leading to potential defamation lawsuits. Additionally, issues surrounding hate speech and incitement to violence raise concerns about the responsibility of chatbot developers in monitoring and moderating content. Legislation and guidelines regarding AI technology are still evolving, leaving room for further debate and potential regulation.
FAQ: New Uncensored Chatbots
Q: How do uncensored chatbots differ from traditional chatbots?
A: Uncensored chatbots are not programmed with filters or rules that restrict the topics they can discuss, allowing for a more uncensored and spontaneous conversation.
Q: What are the benefits of using uncensored chatbots?
A: Uncensored chatbots provide a platform for open dialogue, enabling users to have more authentic and meaningful interactions.
Q: What are the potential drawbacks of uncensored chatbots?
A: The lack of restrictions on uncensored chatbots can lead to the spread of hate speech, misinformation, and harmful ideologies.
Q: What legal and ethical challenges do uncensored chatbots pose?
A: Uncensored chatbots raise concerns about liability in spreading false information and the responsibility of developers in monitoring hate speech and violence incitement. Legislation and guidelines for AI technology are still evolving in this area.
- The team of experts at TechListen.com are dedicated to providing you with the best information and insights on the latest and greatest technology. Our experts have extensive knowledge and experience in the tech industry, and are constantly researching and analyzing the newest products and services so you can make informed decisions. We provide detailed reviews and analysis of the latest gadgets and gizmos, as well as the hottest trends in the industry. Our team is committed to helping you stay up-to-date with the ever-changing world of technology, so you can get the most out of your tech investments. With our expertise and dedication to providing the best information, TechListen.com is the premier online destination for all things tech.
- NewsAugust 25, 2023Instacart Unveils $428 Million Profit with Anticipated Slowing Growth Prior to I.P.O.
- ReviewsAugust 23, 2023Reviewing Immortals of Aveum: Do You Believe in Magic?
- NewsAugust 22, 2023Quordle Today: Hints and Answers for Wednesday, August 23, Game #576
- NewsAugust 19, 2023Cruise Decreases Driverless Car Fleet in San Francisco After Crash