Google Play icon

Computer equal to or better than humans at indexing science

Posted December 2, 2014

In 1997, IBM’s Deep Blue computer beat chess wizard Garry Kasparov. This year, a computer system developed at the University of Wisconsin-Madison equaled or bested scientists at the complex task of extracting data from scientific publications and placing it in a database that catalogs the results of tens of thousands of individual studies.

Photo: Christopher Ré

Christopher Ré


“We demonstrated that the system was no worse than people on all the things we measured, and it was better in some categories,” says Christopher Ré, who guided the software development for a project while a UW professor of computer sciences.

The development, described in the current issue of PLoS, marks a milestone in the quest to rapidly and precisely summarize, collate and index the vast output of scientists around the globe, says first author Shanan Peters, a professor of geoscience at UW-Madison.

Photo: Shanan Peters

Shanan Peters


Peters and colleagues set up the faceoff between PaleoDeepDive, their new machine reading system, and the human scientists who had manually entered data into thePaleobiology Database. This repository, compiled by hundreds of researchers, is the destination for data from paleontology studies funded by the National Science Foundation and other agencies internationally.

The knowledge produced by paleontologists is fragmented into hundreds of thousands of publications. Yet many research questions require what Peters calls a “synthetic approach: For example, how many species were on the planet at any given time?”

Teaming up with Ré, who is now at Stanford University, and UW-Madison computer sciences professor Miron Livny, the group built on the DeepDive machine reading system and the HTCondor distributed job management system to create PaleoDeepDive. “We were lucky that Miron Livny brought the high throughput computing capabilities of the UW-Madison campus to bear,” says Peters. “Getting started required a million hours of computer time.”

PaleoDeepDive mimics the human activities needed to assemble the Paleobiology Database. “We extracted the same data from the same documents and put it into the exact same structure as the human researchers, allowing us to rigorously evaluate the quality of our system, and the humans,” Peters says.

“We demonstrated that the system was no worse than people on all the things we measured, and it was better in some categories.”

Christopher Ré

Instead of trying to divine the single correct meaning, the tactic was to “to look at the entire problem of extraction as a probabilistic problem,” says Ré, who credits much of the heavy lifting to UW-Madison Ph.D. candidate Ce Zhang.

Computers often have trouble deciphering even simple-sounding statements, Ré says. Ré imagines a study containing the terms “Tyrannosaurus rex” and “Alberta, Canada.” Is Alberta where the fossil was found, or where it is stored? “We take a more relaxed approach: There is some chance that these two are related in this manner, and some chance they are related in that manner.”

In these large-data tasks, PaleoDeepDive has a major advantage, Peters says. “Information that was manually entered into the Paleobiology Database by humans cannot be assessed or enhanced without going back to the library and re-examining original documents. Our machine system, on the other hand, can extend and improve results essentially on the fly as new information is added.”

Further advantages can result from improvements in the computer tools. “As we get more feedback and data, it will do a better job across the board,” Peters says.

“Our machine system, on the other hand, can extend and improve results essentially on the fly as new information is added.”

Shanan Peters

The machine-reading trial required access to tens of thousands of articles, saysJacquelyn Crinion, assistant director of licensing and acquisitions services at theUW–Madison General Library System. And the download volume threatened logjams in document delivery. Eventually, Elsevier gave the UW-Madison team broad access to 10,000 downloads per week.

As text- and data-mining takes off, Crinion says the library system and publishers will adapt. “The challenge for all of us is to provide specialized services for researchers while continuing to meet the core needs of the vast majority of our customers.”

The Paleobiology Database has already generated hundreds of studies about the history of life, Peters says. “Ultimately, we hope to have the ability to create a computer system that can do almost immediately what many geologists and paleontologists try to do on a smaller scale over a lifetime: read a bunch of papers, arrange a bunch of facts, and relate them to one another in order to address big questions.”

Source: WISC, by David Tenenbaum

Featured news from related categories:

Technology Org App
Google Play icon
87,522 science & technology articles

Most Popular Articles

  1. An 18 carat gold nugget made of plastic (January 13, 2020)
  2. Anti Solar Cells: A Photovoltaic Cell That Works at Night (February 3, 2020)
  3. Toyota Raize is a new cool compact SUV that we will not see in this part of the world (November 24, 2019)
  4. Nuclear waste could be recycled for diamond battery power (January 21, 2020)
  5. Physicist Proposes a Testable Theory Stating that Information has Mass and could Account for Universe s Dark Matter (January 24, 2020)

Follow us

Facebook   Twitter   Pinterest   Tumblr   RSS   Newsletter via Email