Skip to content

Commit

Permalink
🍪📝 update documentation with directory and data structure
Browse files Browse the repository at this point in the history
  • Loading branch information
JoseRZapata committed Mar 28, 2024
1 parent e23273a commit a13581b
Show file tree
Hide file tree
Showing 5 changed files with 19 additions and 12 deletions.
1 change: 1 addition & 0 deletions docs/data_schema.md
Original file line number Diff line number Diff line change
@@ -0,0 +1 @@
--8<-- "{{cookiecutter.repo_name}}/data/README.md"
1 change: 1 addition & 0 deletions docs/directory_hierarchy.md
Original file line number Diff line number Diff line change
@@ -0,0 +1 @@
--8<-- "README.md:104:182"
3 changes: 2 additions & 1 deletion docs/index.md
Original file line number Diff line number Diff line change
@@ -1 +1,2 @@
--8<-- "README.md"
--8<-- "README.md::20"
--8<-- "README.md:52"
8 changes: 6 additions & 2 deletions mkdocs.yml
Original file line number Diff line number Diff line change
Expand Up @@ -19,12 +19,15 @@ theme:

nav:
- 📖 Introduction: 'index.md'
- 🗃️ Project structure:
- 🗂️ Directory Hierarchy: 'directory_hierarchy.md'
- 🗄️ Data Schema: 'data_schema.md'
- 🔑 Setup Tokens: 'setup_tokens.md'
- ⚙️ Pre-commit configuration: 'pre-commit.md'
- ⚙️ Pre-commit configuration: 'pre-commit.md'
- 🚀 Github Actions:
- Pre-commit_autoupdate: 'github_actions/gh_action_pre-commit-autoupdate.md'
- 💻 VSCode configuration: 'vscode.md'
- 🏠 Local Dev setup: 'local_setup.md'
- 🛠️ Local Dev setup: 'local_setup.md'

markdown_extensions:
- pymdownx.snippets:
Expand All @@ -39,3 +42,4 @@ markdown_extensions:
pygments_lang_class: true
- pymdownx.inlinehilite
- pymdownx.superfences
- tables
18 changes: 9 additions & 9 deletions {{cookiecutter.repo_name}}/data/README.md
Original file line number Diff line number Diff line change
Expand Up @@ -4,16 +4,16 @@ layered data-engineering convention

![layeded data engineering](https://docs.kedro.org/en/0.18.6/_images/data_engineering_convention.png)

| ****Folder in data**** | ****Description**** |
| `Folder in data` | `Description` |
| ---------------------- | --- |
| ****raw**** | initial start of the pipeline, containing the sourced data model(s) that should never be changed, it forms your single source of truth to work from. these data models are typically un-typed in most cases e.g. csv, but this will vary from case to case |
| ****intermediate**** | optional data model(s), which are introduced to type your raw data model(s), e.g. converting string based values into their current typed representation |
| ****primary**** | domain specific data model(s) containing cleansed, transformed and wrangled data from either raw or intermediate, which forms your layer that you input into your feature engineering |
| ****feature**** | analytics specific data model(s) containing a set of features defined against the primary data, which are grouped by feature area of analysis and stored against a common dimension |
| ****model input**** | analytics specific data model(s) containing all feature data against a common dimension and in the case of live projects against an analytics run date to ensure that you track the historical changes of the features over time |
| ****models**** | stored, serialised pre-trained machine learning models |
| ****model output**** | analytics specific data model(s) containing the results generated by the model based on the model input data |
| ****reporting**** | reporting data model(s) that are used to combine a set of primary, feature, model input and model output data used to drive the dashboard and the views constructed. it encapsulates and removes the need to define any blending or joining of data, improve performance and replacement of presentation layer without having to redefine the data models |
| `raw` | initial start of the pipeline, containing the sourced data model(s) that should never be changed, it forms your single source of truth to work from. these data models are typically un-typed in most cases e.g. csv, but this will vary from case to case |
| `intermediate` | optional data model(s), which are introduced to type your raw data model(s), e.g. converting string based values into their current typed representation |
| `primary` | domain specific data model(s) containing cleansed, transformed and wrangled data from either raw or intermediate, which forms your layer that you input into your feature engineering |
| `feature` | analytics specific data model(s) containing a set of features defined against the primary data, which are grouped by feature area of analysis and stored against a common dimension |
| `model input` | analytics specific data model(s) containing all feature data against a common dimension and in the case of live projects against an analytics run date to ensure that you track the historical changes of the features over time |
| `models` | stored, serialised pre-trained machine learning models |
| `model output` | analytics specific data model(s) containing the results generated by the model based on the model input data |
| `reporting` | reporting data model(s) that are used to combine a set of primary, feature, model input and model output data used to drive the dashboard and the views constructed. it encapsulates and removes the need to define any blending or joining of data, improve performance and replacement of presentation layer without having to redefine the data models |

## References

Expand Down

0 comments on commit a13581b

Please sign in to comment.