cat hn-uri.txt | xargs -P 100 -n 100 wget --quiet

Our latest project on Sizzle is a visualization of the Top 10k Posts of All Time on Hacker News To create the visualization, we first needed to collect the data.I noticed that there was an old copy of the hacker news dataset available on Big Query . But I needed an up-to-date copy, so I looked into the Hacker News Firebase API The API allows you to get each item by Id. You can start by retrieving the current Max ID, then walking backwards from there. (Items my be stories, comments, etc., it's the same API endpoints for all types of items.)There is no rate limit, so I created the following script that will generate a text file with 10MM lines containing all of the URIs to retrieve. (we will then feed this file into wget using xargs)Note: 10MM items was ~5 years worth of data.Script to create the 10MM line file of URIs to retrieve:That script takes about 10 minutes to produce a file that's around 560MB in size.After the file is generated, you can feed it to wget using xargs to retrieve all the URIs.ex:wget will save the result of each GET request in a separate file with the format {id}.jsonCaution: That command took just over 30 hours to complete on my macbook. (it also killed Finder a couple times and I had to disable spotlight on the folder I was saving all the .json files to)I found that it can be difficult to work with 10MM files in a single directory on my mac, so I will try to save you the trouble.// EDIT: links have been removedHere are a couple copies of the dataset I retrieved:1. Here is a zip of the directory containing 10MM json files. (4GB)2. Here is a SQL Server backup of a database containing a single table, that contains one record per json file. (2GB)// EDIT: links have been removed// EDIT:you can get a nightly up-to-date copy of the dataset here: https://bigquery.cloud.google.com/table/bigquery-public-data:hacker_news.full?tab=details Hope this helps!Aaron