Cong Lu has long been fascinated by how to use technology to make his job as a research scientist more efficient. But his latest project takes the idea to an extreme.
Lu, who is a postdoctoral research and teaching fellow at the University of British Columbia, is part of a team building an âAI Scientistâ with the ambitious goal of creating an AI-powered system that can autonomously do every step of the scientific method.
âThe AI Scientist automates the entire research lifecycle, from generating novel research ideas, writing any necessary code, and executing experiments, to summarizing experimental results, visualizing them, and presenting its findings in a full scientific manuscript,â says a write-up on the projectâs website. The AI system even attempts a âpeer review,â of the research paper, which essentially brings in another chatbot to check the work of the first.
An initial version of this AI Scientist has already been released â anyone can download the code for free. And plenty of people have. It did the coding equivalent of going viral, with more than 7,500 people liking the project on the code library GitHub.
To Lu, the goal is to accelerate scientific discovery by letting every scientist effectively add Ph.D.-level assistants to quickly push boundaries, and to âdemocratizeâ science by making it easier to conduct research.
âIf we scale up this system, it could be one of the ways that we truly scale scientific discovery to thousands of underfunded areas,â he says. âA lot of times the bottleneck is on good personnel and years of training. What if we could deploy hundreds of scientists on your pet problems and have a go at it?â
But he admits there are plenty of challenges to the approach â such as preventing the AI systems from âhallucinating,â as generative AI in general is prone to do.
And if it works, the project raises a host of existential questions about what role human researchers â the workforce that powers much of higher education â would play in the future.
The project comes at a moment where other scientists are raising concerns about the role of AI in research.
A paper out this month, for instance, found that AI chatbots are already being used to create fabricated research papers that are showing up in Google Scholar, often on contentious topics like climate research.
And as tech firms continue to release more-powerful chatbots to the public â like the new version of ChatGPT put out by OpenAI this month â prominent AI experts are raising fresh concerns that AI systems could leap guardrails in ways that threaten global safety. After all, part of âdemocratizing researchâ could lead to greater risk of weaponizing science.
It turns out the bigger question may be whether the latest AI technology is even capable of making novel scientific breakthroughs by automating the scientific process, or thereâs something uniquely human about the endeavor.
Checking for Errors
The field of machine learning â the only field the AI Scientist tool is designed for so far â may be uniquely suited for automation.
For one thing, it is highly structured. And even when humans do the research, all of the work happens on a computer.
âFor anything that requires a wet lab or hands-on stuff, weâve still got to wait for our robotic assistants to show up,â Lu says.
But the researcher says that pharmaceutical companies have already done significant work to automate the process of drug discovery, and he believes AI could take those measures further.
One practical challenge for the AI Scientist project has been avoiding AI hallucinations. For instance, Lu says that because large language models continually generate the next character or âtokenâ based on probability derived from training data, there are times when such systems might produce errors when copying data. For instance, the AI Scientist might enter 7.1 when the correct number in a dataset was 9.2, he says.
To prevent that, his team is using a non-AI system when moving some data, and having the system ârigorously check through all of the numbers,â to detect any errors and correct them. He says a second version of the teamâs system that they expect to release later this year will be more accurate than the current one when it comes to handling data.
Even in the current version, the projectâs website boasts that the AI Scientist can carry out research far cheaper than human Ph.D.s can, estimating that a research paper can be created â from idea generation to writing and peer review â for about $15 in computing costs.
Does Lu worry that the system will put researchers like himself out of work?
âWith the current capabilities of AI systems, I don’t think so,â says Lu. âI think right now it’s mainly an extremely powerful research assistant that can help you take the first steps and early explorations on all the ideas that you never had time for, or even help you brainstorm and investigate a few ideas on a new topic for you.â
Down the road, if the tool improves, though, Lu admits it could eventually raise tougher questions for the role of human researchers. Though in that context research will not be the only thing transformed by advanced AI tools. For now, though, he sees it as what he calls a âforce multiplier.â
âItâs just like how code assistants now let anyone very simply code up a mobile game app or a new website,â he says.
The projectâs leaders have put in guardrails on the kinds of projects it can attempt, to prevent the system from becoming an AI mad scientist.
âWe donât really want loads of new viruses or lots of different ways to make bombs,â he says.
And theyâve limited the AI Scientist to a maximum of running two or three hours at a time, he says, âso we have control of it,â noting that thereâs only so much âhavoc it could wreak in that time.â
Multiplying Bad Science?
As the use of AI tools spreads rapidly, some scientists worry that they could be used to actually hinder scientific progress by flooding the web with fabricated papers.
When researcher Jutta Haider, a professor of librarianship, information, education and IT at the Swedish School of Library and Information Science, went looking on Google Scholar for papers with AI-fabricated results, she was surprised at how many she found.
âBecause it was really badly produced ones,â she explains, noting that the papers were clearly not written by a human. âJust simple proofreading should have eliminated those.â
She says she expects there are many more AI-fabricated papers that her team did not detect. âItâs the tip of the iceberg,â she says, since AI is getting more sophisticated, so it will be increasingly difficult to tell if something was human- or AI-written.
One problem, she says, is that it is easy to get a paper listed in Google Scholar, and if you are not a researcher yourself, it may be difficult to tell reputable journals and articles from those created by bad actors trying to spread misinformation or add fabricated work to their CV and hope no one checks where it is published.
âBecause of the publish-or-perish paradigm that rules academia, you can’t make a career without publishing a lot,â Haider says. âBut some of the papers are really bad, so nobody will probably make a career with those ones that we found.â
She and her colleagues are calling on Google to do more to scan for AI-fabricated articles and other junk science. âWhat I really recommend Google Scholar do is hire a team of librarians to figure out how to change it,â she adds. âIt isnât transparent. We donât know how it populates the index.â
EdSurge reached out to Google officials but got no response.
Lu, of the AI Scientist project, says that junk science papers have been a problem for a while, and he shares the concern that AI could make the phenomenon more pervasive. âWe recommend whenever you run the AI Scientist system, that anything that is AI-generated should be watermarked so it is verifiably AI-generated and it cannot be passed off as a real submission,â he says.
And he hopes that AI can actually be used to help scan existing research â whether written by humans or bots â to ferret out problematic work.
But Is It Science?
While Lu says the AI Scientist has already produced some useful results, it remains unclear whether the approach can lead to novel scientific breakthroughs.
âAI bots are really good thieves in many ways,â he says. âThey can copy anyoneâs art style. But could they invent a new art style that hasnât been seen before? Itâs hard to say.â
He says there is a debate in the scientific community about whether major discoveries come from a pastiche of ideas over time or involve unique acts of human creativity and genius.
âFor instance, were Einsteinâs ideas new, or were those ideas in the air at the time?â he wonders. âOften the right idea has been staring us in the face the whole time.â
The consequences of the AI Scientist will hinge on that philosophical question.
For Haider, the Swedish scholar, sheâs not worried about AI ever usurping her job.
âThereâs no point for AI to be doing science,â she says. âScience comes from a human need to understand â an existential need to want to understand â the world.â
âMaybe there will be something that mimics science,â she concludes, âbut itâs not science.â