Skip to content

nationalparkservice/QC-tips-and-tricks

Repository files navigation

QC-tips-and-tricks

This is a small collection of scripts for establishing quality control (QC'ing) over datasets. These techniques can help you to clean up data prior to publishing datasets. Roughly equivalent code examples are given in R tidyverse and Python pandas. As you will see, these scripts are extremely simple, and they can be used over and over again on different datasets.

The motivation for including both R and Python is partly to reach a wider audience, but it's also to demostrate that these techniques are essentially language independent. They're based on the common "set" based methods used to QC data in SQL and relational databases. Most rectangular datasets are, in fact, true sets in the sense of relational database theory. The requirement for a true set is that every record is unique. With the development of R tidyverse and Python pandas, you can use these set operations to tame even the most unruly dataframes.

The main methods for QC'ing data can be expressed in SQL, R and Python. These methods include the functional equivalents of SELECT, UNIQUE, GROUP BY, JOIN and several other column-based operations. Both R and Python also bring data visualizations to the table, something SQL cannot do. For continuous data, for example, ggplot histograms work well with the arrange() function to look at the tails of continuous numeric data.

Example data - Electric Vehicle Population Data

The rationale of this project is to take a publicly available dataset, the Electric Vehicle Population Data from Washington State, and introduce all kinds of horrible inconsistencies and errors into the data. That defines the problem domain. The solutions will be to hunt down and locate those problems as quickly and efficiently as possible. The goal is to make QC'ing systematic and perhaps somewhat tedious, rather than chaotic, frustrating and difficult. (NOTE - the dataset used here is a small subset of the actual electric vehicle dataset. The complete dataset is available at DATA.GOV).

How to use these scripts

If you are familiar with Git and GitHub, you can simply "git clone" the URL for this repository. If you are new to GitHub, go to the Code button at the top of this page and select download ZIP. The ZIP file will contain everything.

Note - this is currently a work-in-progress so please feel free to send questions, comments or code fixes to me at gareth_rowell@nps.gov.

Outline of scripts

1_getting_started
1_1_lay_of_the_land
1_2_setting_data_types
1_3_stray_commas
1_4_additional_columns

2_exploring_the_dataset
2_1_one_big_problem_many_small_problems
2_2_meaningful_groups
2_3_weird_variables 
 

3_non_numeric_columns
3_1_typos_and_categorical_errors
3_2_Using_left_joins 
3_3_distinct_and_dates

4_missing_values
4_1_detecting_missing_values
4_2_resolving_missing_data

5_lookups_and_uniqueness
5_1_finding_duplicates 
5_2_testing_uniqueness
5_3_multivariable_uniqueness

6_joins
6_1_join_types
6_2_primary_key_tests
6_3_foreign_key_tests
6_4_one_to_many_errors
6_5_many_to_many_errors

7_histograms_arrange
7_1_continuous_variables
7_2_examine_tails
7_3_arrange_and_tails
7_4_associated_variables

About

Simple methods for QC'ing datasets in R and Python

Resources

Stars

Watchers

Forks

Releases

No releases published

Packages

No packages published