Bill Hart-Davidson, MSU’s associate dean for graduate education, suggests that the likely answer is “text-mining,” the use of computer programs to analyze large collections of documents to generate data. When I called Hart-Davidson, I suggested that the East Lansing Sci-Hub scraper might be someone from his own research team. But he laughed and said that he had no idea who it was. But he understands why the scraper goes to Sci-Hub even though MSU subscribes to the downloaded journals. For his own research on the linguistic structure of scientific discourse, Hart-Davidson obtained more than 100 years of biology papers the hard way—legally with the help of the publishers. “It took an entire year just to get permission,” says Thomas Padilla, the MSU librarian who did the negotiating. And once the hard drive full of papers arrived, it came with strict rules of use. At the end of each day of running computer programs on it from an offline computer, Padilla had to walk the resulting data across campus on a thumb drive for analysis with Hart-Davidson.