I'm in the process of creating a project in R. I've done projects in the past (I've have read some of the best practices for project/folder organization i.e. using
usethis) and always find myself trying to remember how I did something, or where data came from. It's time to improve on this.
I have a pretty basic question. A lot of the data that I use gets scraped via API. I'd like to store the raw untouched data, in addition to the modified data based on the raw data.
I want to do the following:
- Save Raw Data via Scrape
- Save "Tidy Dataset"
- Document both the Raw Data Scrape, and Tidy Data
After reading Hadley's book on packages, I think I understand the process for saving the modified data. I was planning on documenting the data in an .r script in the
data-raw folder that I created.
A few questions:
- Should I save the raw scraped data in
- Do I document both the data scrape and the modified data in the same
- Can I create 1 master
.Rfile that documents all of my data-sets (I have over 10 scrapes that I do), or do I need to have 1 file per data-set?