AI in materials science: promise and pitfalls of automated discovery

Are you ready to bring more awareness to your brand? Consider becoming a sponsor for The AI Impact Tour. Learn more about the opportunities here.


Last week, a team of researchers from University of California Berkeley published a highly anticipated paper in the journal Nature describing an “autonomous laboratory” or “A-Lab” that aimed to use artificial intelligence (AI) and robotics to speed up the discovery and synthesis of new materials. 

Dubbed a “self-driving lab,” the A-Lab presented an ambitious vision of what an AI-powered system could reach in scientific research when equipped with the latest techniques in computational modeling, machine learning, automation and natural language processing.

Diagram showing how the A-Lab works: UC Berkeley/Nature

However, within days of publication, doubts began to emerge about some of the key claims and results presented in the paper. 

Robert Palgrave, an inorganic chemistry and materials science professor at University College London with decades of encounter in X-ray crystallography, raised a series of technical concerns on X (formerly Twitter) about inconsistencies he noticed in the data and analysis provided as evidence for the A-Lab’s purported successes. 

VB Event

The AI Impact Tour

Connect with the enterprise AI community at VentureBeat’s AI Impact Tour coming to a city near you!

 


Learn More

In particular, Palgrave argued that the phase identification of synthesized materials conducted by the A-Lab’s AI via powder X-ray diffraction (XRD) appeared to be seriously flawed in a number of cases and that some of the newly synthesized materials were already discovered.

AI’s promising attempts — and their pitfalls

Palgrave’s concerns, which he aired in an interview with VentureBeat and a pointed letter to Nature, revolve around the AI’s interpretation of  XRD data – a technique akin to taking a molecular fingerprint of a material to grasp its structure.

visualize XRD as a high-tech camera that can snap pictures of atoms in a material. When X-rays hit the atoms, they scatter, creating patterns that scientists can read, admire using shadows on a wall to ascertain a source object’s shape. 

Similarly to how children use hand shadows to copy shapes of animals, scientists make models of materials, and then see if those models produce similar X-ray patterns to the ones they measured. 

Palgrave pointed out that the AI’s models didn’t match the actual patterns, suggesting the AI might have gotten a bit too creative with its interpretations.

Palgrave argued this represented such a fundamental failure to confront basic standards of evidence for identifying new materials that the paper’s central thesis — that 41 novel synthetic inorganic solids had been produced — could not be upheld. 

In a letter to Nature, Palgrave detailed over a slew of examples where the data simply did not preserve the conclusions drawn. In some cases, the calculated models provided to match XRD measurements differed so dramatically from the actual patterns that “serious doubts exist over the central claim of this paper, that new materials were produced.” 

Although he remains a proponent of AI use in the sciences, Palgrave questions whether such an undertaking could realistically be performed fully autonomously with current technology. “Some level of human verification is still needed,” he contends.

Palgrave didn’t mince words: “The models that they make are in some cases completely different to the data, not even a little bit close, admire utterly, completely different.” His message? The AI’s autonomous efforts might have missed the mark, and a human touch could have steered it right.

The human touch in AI’s ascent

Responding to the wave of skepticism, Gerbrand Ceder, the head of the Ceder Group at Berkeley, stepped into the fray with a LinkedIn post

Ceder acknowledged the gaps, saying, “We appreciate his feedback on the data we shared and aim to address [Palgrave’s] specific concerns in this response.” Ceder admitted that while A-Lab laid the groundwork, it still needed the discerning eye of human scientists.

Ceder’s update included new evidence that supported the AI’s success in creating compounds with the right ingredients. However, he conceded, “a human can perform a higher-quality [XRD] refinement on these samples,” recognizing the AI’s current limitations. 

Ceder also reaffirmed that the paper’s objective was to “display what an autonomous laboratory can reach” — not claim perfection. And upon review, more comprehensive analysis methods were still needed.

The conversation spilled back over to social media, with Palgrave and Princeton Professor Leslie Schoop weighing in on the Ceder Group’s response Their back-and-forth highlighted a key takeaway: AI is a promising tool for material science’s future, but it’s not ready to go solo.

The next steps from Palgrave and his team is a re-analysis of the XRD results, with an eye to produce a much more thorough description of what compounds were actually synthesized.

Navigating the AI-human partnership in science

For those in executive and corporate leadership roles, this experiment is a case research in the potential and limitations of AI in scientific research. It illustrates the importance of marrying AI’s speed with the meticulous oversight of human experts.

The key lessons are clear: AI can revolutionize research by handling the heavy lifting, but it can’t yet replicate the nuanced judgment of seasoned scientists. The experiment also underscores the value of peer review and transparency in research, as expert critiques from Palgrave and Schoop have highlighted areas for improvement.

Looking ahead, the future involves a synergistic blend of AI and human intelligence. Despite its flaws, the Ceder group’s experiment has sparked an essential conversation about AI’s role in advancing science. It’s a reminder that while technology can push boundaries, it’s the wisdom of human encounter that ensures we’re moving in the right direction.
This experiment stands as both a testament to AI’s potential in material science and a cautionary tale. It’s a rallying cry for researchers and tech innovators to refine AI tools, ensuring they’re reliable partners in the quest for knowledge. The future of AI in science is indeed luminous, but it will shine its brightest when guided by the hands of those who have a deep understanding of the world’s complexities.

VentureBeat’s mission is to be a digital town square for technical decision-makers to gain knowledge about transformative enterprise technology and transact. unearth our Briefings.

Source link