Love what Google is doing here: "We believe that the field could benefit from a large, standard set with benchmarks for easy comparison and experiments with new modeling techniques. To that end, we are releasing scripts that convert a set of public data into a language model consisting of over a billion words, with standardized training and test splits ... [and] the processed data in one convenient location, along with the training and test data. This will make it much easier for the research community to quickly reproduce results, and we hope will speed up progress on these tasks. The field needs a new and better standard benchmark. Currently, researchers report from a set of their choice, and results are very hard to reproduce because of a lack of a standard in preprocessing. We hope that this will solve both those problems."