AIs are more likely to mislead people if trained on human feedback 

You May Be Interested In:‘We got stuck in puddles’: skiers upset by lack of snow on Swedish slopes


Striving to come up with answers that please humans may make chatbots more likely to pull the wool over our eyes

JuSun/Getty Images

Giving AI chatbots human feedback on their responses seems to make them better at giving convincing, but wrong, answers.

The raw output of large language models (LLMs), which power chatbots like ChatGPT, can contain biased, harmful or irrelevant information, and their style of interaction can seem unnatural to humans. To get around this, developers often get people to evaluate a model’s responses and then fine-tune it based on this feedback.

share Paylaş facebook pinterest whatsapp x print

Similar Content

Tesla posts bigger-than-expected loss, bigger-than-expected revenue [Updated]
Tesla posts bigger-than-expected loss, bigger-than-expected revenue [Updated]
New Scientist. Science news and long reads from expert journalists, covering developments in science, technology, health and the environment on the website and the magazine.
Experts say US flights are safe now but flag warning signs to look for
DeepSeek logo on a smartphone against a blue streaming-code tech-ish background.
DeepSeek iOS app sends data unencrypted to ByteDance-controlled servers
For the first time, beloved IDE Jetbrains Rider will be available for free
For the first time, beloved IDE Jetbrains Rider will be available for free
Man in dark pinstripe suit holding open a briefcase, from which papers are flying out and covering his face.
CEO of AI ad-tech firm pledging “world free of fraud” sentenced for fraud
What can governments do about online disinformation from abroad?
What can governments do about online disinformation from abroad?
The News Spectrum | © 2024 | News