Home / Health / AI & Science: How AI Agents Performed at a Landmark Conference

AI & Science: How AI Agents Performed at a Landmark Conference

AI & Science: How AI Agents Performed at a Landmark Conference
Kathryn Hulick 2025-10-24 16:00:00

In ⁢a first, a scientific⁤ conference welcomed paper submissions from any area of science, but with one catch: AI had to ‌do most‌ of​ the work. Called Agents4Science 2025, the Oct. 22 virtual event focused on the ‍work of artificial intelligence agents — systems that pair large language models with other tools or databases to perform multistep tasks.

From formulating hypotheses to analyzing data and ‌providing the first round⁢ of peer reviews, AI agents⁢ took the lead. Human reviewers then ‌stepped in to assess the top submissions. ‍In all, ‌48 papers out ‌of ‌314 made ‍the cut. Each had to ‍detail how people and AI collaborated on ‍every​ stage of the research⁤ and writing process.

“We’re seeing this interesting paradigm shift,” said James Zou,a computer scientist at Stanford University who co-organized ⁣the conference. “People are starting to explore using AI‌ as a co-scientist.”

Most scientific journals and meetings currently ban AI coauthors ⁣and prohibit peer ⁢reviewers from relying on AI. These policies aim to avoid hallucinations and other issues related ​to AI use. ‌However, this⁣ approach makes it tough⁣ to learn how good AI is‍ at science. That’s what Agents4Science aimed to‌ explore, Zou said, calling ‌the conference an ‌experiment, with all the materials publicly available for anyone to study.

At the virtual meeting, humans ‌presented ​AI-assisted work ‌spanning fields such as economics, biology and engineering. Min min Fong, an economist at the University⁣ of⁢ California, berkeley,​ and her ⁢team collaborated‍ with AI to‌ study ⁤car-towing data from San Francisco. Their study found that waiving high⁣ towing fees helped low-income⁢ people keep their vehicles.

“AI was really great at helping us with computational acceleration,” Fong said. But, she found,‌ “you have to be really careful when working with AI.”

As an ⁢example,the AI kept ‍citing⁣ the wrong date for when ​San ⁢francisco’s rule ‌waiving towing fees went ​into effect.⁣ Fong had to check⁣ this ​in the original source to discover the error. “The core scientific ‍work still remains human-driven,” she said.

Also Read:  Radiopharmaceuticals: Logistics, Challenges & Future Research

for Risa Wechsler, a computational astrophysicist at Stanford who helped review submissions, the results were⁣ mixed. ‍The papers she saw were technically correct, she said, “but they ⁤were neither ⁢interesting nor important.” she ​was excited about the potential of AI for ⁢research​ but remained unconvinced that today’s agents can “design ‍robust​ scientific ‍questions.”‌ And, she added, the technical skill of AI can “mask poor ⁤scientific judgment.”

Still, the event included some glimmers​ of hope for⁣ the⁣ future of ⁤AI in science.⁤ Silvia Terragni, a machine learning engineer at the company Upwork in San Francisco,‌ said⁤ that​ she gave chatgpt some context about the kinds of problems her company deals with and asked the bot to propose paper ideas. “One ⁢of these was⁢ the winner,” ‍she said, selected as ⁢one of the three ​top papers in the conference. ⁤It was a​ study about using AI ⁤reasoning‌ in a job marketplace. “I think [AI] can actually come up with novel ideas,” ⁢she ​said.

Leave a Reply