txdeathrow_scraper / Week 6 / due 2018-02-21 23:59

Scraping the HTML from the TX Death Row page and turning it into data. A followup to the assignment at: txdeathrow_check / Week 5 / due 2018-02-12 23:59

Assignment notes and extensive discussion here:

https://github.com/compciv/homeworkhome/tree/master/txdeathrow_scraper

Requirements

Basically, pass the tests!

When I visit your Github.com repo page

I expect your Github repo at compciv-2018-SUNETID repo to have the following subfolder:

compciv-2018-SUNETID/week-06/txdeathrow_scraper/

On this subfolder’s page, I would expect the file tree to look like this:

├── data_helper.py
├── format_helper.py
├── scraper.py

Note that data_helper.py is given to you as a file to include in your homework directory). Note that the above listing does not include tests or the setup_hw.py files, which are used for the logistics of this homework but not for the actual work.

When I clone your Github repo

If I were to clone your repo onto my own computer, e.g.

$ git clone https://github.com/GITHUBID/compciv-2018-SUNETID.git

I would expect your homework subfolder to look like this:

compciv-2018-SUNETID/
└── week-06/
    └── txdeathrow_scraper/
        ├── data_helper.py
        ├── format_helper.py
        └── scraper.py

(note that data_helper.py is given to you as a file to include in your homework directory)

Command-line setup

Creating the directory and making it your current working directory:

$ cd ~/Desktop/compciv-2018-SUNETID
$ mkdir week-06
$ mkdir week-06/txdeathrow_scraper
$ cd week-06/txdeathrow_scraper

The repo has a section on the shell command for downloading the allt he starter and test files:

[Setup via the command-line](https://github.com/compciv/homeworkhome/tree/master/txdeathrow_scraper#cli-setup)