Skip to content

A comprehensive cricket statistics pipeline built using Google Cloud services. The pipeline involves retrieving data via the Cricbuzz API, and leveraging various cloud technologies such as Cloud Composer, Cloud Storage, Cloud Functions, Dataflow, and more to process and manage the data efficiently.

Notifications You must be signed in to change notification settings

sofiasawczenko/ETL_cricket_stats

Repository files navigation

Building a Cricket Statistics Pipeline with Google Cloud Services

Constructed a comprehensive cricket statistics pipeline using Google Cloud services. From retrieving data via the Cricbuzz API to crafting a dynamic Looker Studio dashboard, each phase contributes to the seamless flow of data for analysis and visualization.

Architecture

Architecture

Data Retrieval with Python and Cricbuzz API

The foundation of our project begins with Python’s prowess in interfacing with APIs. We’ll delve into the methods of fetching cricket statistics from the Cricbuzz API, harnessing the power of Python to gather the required data efficiently.

Storing Data in Google Cloud Storage (GCS)

Once the data is obtained, our next step involves preserving it securely in the cloud. We’ll explore how to store this data in a CSV format within Google Cloud Storage (GCS), ensuring accessibility and scalability for future processing.

Creating a Cloud Function Trigger

With our data safely stored, we proceed to set up a Cloud Function that acts as the catalyst for our pipeline. This function triggers upon file upload to the GCS bucket, serving as the initiator for our subsequent data processing steps.

Execution of the Cloud Function

Within the Cloud Function, intricate code is crafted to precisely trigger a Dataflow job. We’ll meticulously handle triggers and pass the requisite parameters to seamlessly initiate the Dataflow job, ensuring a smooth flow of data processing.

Dataflow Job for BigQuery

The core of our pipeline lies in the Dataflow job. Triggered by the Cloud Function, this job orchestrates the transfer of data from the CSV file in GCS to BigQuery. We’ll meticulously configure the job settings to ensure optimal performance and accurate data ingestion into BigQuery.

Looker Dashboard Creation

Finally, we’ll explore the potential of BigQuery as a data source for Looker Studio. Configuring Looker to connect with BigQuery, we’ll create a visually compelling dashboard. This dashboard will serve as the visualization hub, enabling insightful analysis based on the data loaded from our cricket statistics pipeline.

About

A comprehensive cricket statistics pipeline built using Google Cloud services. The pipeline involves retrieving data via the Cricbuzz API, and leveraging various cloud technologies such as Cloud Composer, Cloud Storage, Cloud Functions, Dataflow, and more to process and manage the data efficiently.

Topics

Resources

Stars

Watchers

Forks

Releases

No releases published

Packages

No packages published

Contributors 3

  •  
  •  
  •