I need to analyze a large data set that is distributed as a lz4 compressed JSON file.
The compressed file is almost 1TB. I'd prefer not to uncompress it to disk due to cost. Each "record" in the dataset is very small, but it is obviously not feasible to read the entire data set into memory.
Any advice on how to iterate through records in this large lz4 compressed JSON file in Python 2.7?