Currently when we request for a page and the website is down or unreachable, the scraper fails. It should instead just log an error and move to the next website to scrape.
We are currently only scraping the Standard Newspaper's stories but we should also scrape data on other newsrooms that are doing financial, business, and budget stories.
Currently we have each news outlet as it's own separate JSON.
For our purposes of having the stories displayed on https://taxclock.codeforkenya.org/ We should combine them into a single data/news.json that is sorted by time and limited to 7 stories.