Google's 'co-scientist' AI tool: A collaborator, or Reviewer #2?
Last week, Google announced an AI "co-scientist" with training inspired by the scientific method that's supposed to be able to help researchers surface new hypotheses. The system is more than just a regurgitative LLM, with several AI "agent" systems embedded within it to allow it to "reason."
However, when I think back to the multi-dimensional streams of information I had to sort through in my PhD to come up with hypotheses for what was going on with my samples, I remain skeptical that an AI would be able to pull that together in a meaningful way. As I understand it from the co-scientist paper, Google's tool is limited to evaluating text and not raw data, more like getting a peer review back than sitting down with a colleague.
Are you part of Google's Trusted Tester Program? Have you used the tool? Let me know about your experience: aiprognosis@statnews.com
A STAT exclusive: Elicit raises $22M for an AI research tool for scientists
Relatedly, AI research startup Elicit has raised a $22 million series A at a $100 million valuation for its AI research tool specifically designed for researchers, in addition to its $9 million in seed funding. Company executives told me that Elicit already has grown to 500,000 monthly users through word of mouth. Those include research teams at the NIH, NHS, and major pharma companies, which they couldn't name publicly.
The Elicit tool summarizes scientific literature and provides quotations from the sources it cites. The company has a free tier as well as individual and team subscriptions. Its ambitions are a little more down-to-earth than the Google's or OpenAI's tools for researchers.
How is Elicit different from competing deep research AI tools? Tools like OpenAI's and Perplexity's are trained using reinforcement learning, said Elicit co-founder and chief operating officer Jungwon Byun, where a human rates the model's output and that rating is used as feedback for further training.
The problem with that approach "is that they are basically saying, 'Hey, powerful AI system, here is a reward metric, a feedback signal. Do whatever it takes to make this number go up or down and use whatever techniques you want to get there,'" said Byun. "We think this is going to get riskier and riskier for really high-stakes research and reasoning, because already what happens is you start to see that the models do things that don't make sense," such as switching between languages and code, which makes it hard for humans to trace the AI's logic, she said.
How was Elicit designed? Byun said it's designed with research in mind, asking how humans put together meta-analyses and systematic literature reviews, and automating those processes with AI.
How does Elicit deal with the fact that lots of scientific research is paywalled? Users can add papers from their library to Elicit's tool, and the company is looking into ways to leverage researchers' subscriptions to journals. But right now, the tool operates off of open access papers and abstracts.
"You still have to figure out which paywalled papers are you going to pay for and read, and the ability to go through hundreds or thousands of them is still going to save you a lot of time," said Byun. "I think there's a lot more to be done before these tools can really lead to incredible breakthroughs. But I think right now they're already saving people a lot of time and are probably superhuman in being able to summarize all the relevant literature, to figure out what's most worth digging into more."
A 'ChatGPT for medicine' unicorn
Last week, AI medical information startup OpenEvidence announced a $75 million investment from Sequoia Capital that valued the company at $1 billion. If you're not familiar, OpenEvidence is like an AI-powered UpToDate (alternatively, a ChatGPT for medicine) that summarizes clinical evidence for health care providers. It recently announced a partnership with the New England Journal of Medicine's publisher that allows it to pull from all of the NEJM family content from 1990 forward.
Is the $1 billion valuation realistic, especially for a tool that's free for health care professionals? Of course, UpToDate leader Peter Bonis thinks "that there is a bit of a hype cycle going on, at least for the valuations just across the [AI] sector," though he wouldn't comment on OpenEvidence specifically.
UpToDate last week announced an integration with health care AI assistant Corti, which is similar to a partnership the company announced with Abridge last fall. Though UpToDate's content is still written by humans, getting that information pulled in to AI-generated clinical note platforms helps clinicians make better decisions, Bonis told me. Eventually, the company hopes that this information can even surface in the middle of a visit where an AI scribe is being used.
No comments