Scientists should beware using artificial intelligence in research as the technology may have been trained using faulty or biased datasets, a Nobel laureate has said.
While researchers have been quick to embrace the new technology, Avram Hershko, who won the Nobel Prize in Chemistry in 2004, told a scientific conference in Germany that he had concerns over the uncritical way in which some scientists accepted findings generated by聽large language models such as ChatGPT.
鈥淲e have to know what datasets it is using 鈥 it should be transparent,鈥 Professor Hershko, who is based at Technion Israel Institute of Technology, told the annual聽Lindau Nobel Laureate Meeting,聽a gathering of 40 Nobel prizewinners and hundreds of young scientists.
The apparent certainty of results and recommendations generated by large language models trained on millions of scientific journals and books should also be addressed, said the Hungarian-Israeli biochemist. 鈥淚t should say what the margin of certainty is鈥nd this should be part of regulation,鈥 he added.
探花视频
However, Michael Levitt, a University of Stanford structural biologist who won the chemistry Nobel in 2013, said scientists were often required to work with less-than-certain data and should be able to critically assess ChatGPT-generated data in a similar way.
鈥淲e expect data to be wrong all the time 鈥 science is very good at filtering these [bad results] out,鈥 he said.
探花视频
Medical scientists often did not have full access to results behind discoveries, treatments or drugs but were still able to use less-than-complete datasets in a productive way, continued Professor Levitt.
鈥淲hen the first Covid vaccines were designed, getting hold of the data was impossible 鈥 we often have to work with systems that are corrupted,鈥 he said, adding: 鈥淲e can use AI without having all the details.鈥
That said, Professor Levitt believed the most commercially successful AI models would be those that were transparent about the data聽they were reading because consumers preferred to understand more about聽their聽processes.
鈥淭hat鈥檚 one reason why transparency will come鈥ot because [these companies] want to be particularly open,鈥 he said.
探花视频
The panel on AI鈥檚 likely impact on research also raised concerns over the 鈥渙ne-way鈥 transfer of data flows from academia and scientific literature into the large language models owned by private companies, which help to make the AI models more powerful. In contrast, industry was not required to make its data open, limiting the ability of academia to develop comparable tools, the panel heard.
On this issue, Professor Hershko said that 鈥渋nformation should be shared and AI should be open鈥.
鈥淭he only way to get around this problem is to make data open access so everyone can use it,鈥 he said.
However, Aaron Ciechanover, who shared the Nobel Prize with Professor Hershko for their work on cell degradation, said this was unrealistic.
探花视频
鈥淚 would like data to be open but it will not be [possible],鈥 he said, adding that drug companies and health insurance companies would fear that their data would be raided by their commercial rivals if it was made public.
鈥淚f you open up your data, other companies will make billions of dollars from it 鈥 not just from one thing, it would be endless,鈥 said the Israeli biologist on the ways that the data would be exploited.
探花视频
Instead, companies who held data should be allowed to 鈥渕ake one type of money鈥 in terms of using the information and 鈥渢hen the data is open鈥, he said.
Register to continue
Why register?
- Registration is free and only takes a moment
- Once registered, you can read 3 articles a month
- Sign up for our newsletter
Subscribe
Or subscribe for unlimited access to:
- Unlimited access to news, views, insights & reviews
- Digital editions
- Digital access to 罢贬贰鈥檚 university and college rankings analysis
Already registered or a current subscriber?








