From bcf838536094e3c6caa1215d079b57d200caf4a1 Mon Sep 17 00:00:00 2001 From: Sergii Mikhtoniuk Date: Wed, 21 Aug 2024 17:44:49 -0700 Subject: [PATCH] Tweak web3 demo notebook --- .../01 - Introduction.ipynb | 10 +++++----- .../01 - Working with Web3 data.ipynb | 4 ++-- 2 files changed, 7 insertions(+), 7 deletions(-) diff --git a/images/demo/user-home/01 - Kamu Basics (COVID-19 example)/01 - Introduction.ipynb b/images/demo/user-home/01 - Kamu Basics (COVID-19 example)/01 - Introduction.ipynb index b10582478..7add6be22 100644 --- a/images/demo/user-home/01 - Kamu Basics (COVID-19 example)/01 - Introduction.ipynb +++ b/images/demo/user-home/01 - Kamu Basics (COVID-19 example)/01 - Introduction.ipynb @@ -167,12 +167,12 @@ " engine: datafusion # Kamu supports many data processing engines like DataFusion, Spark, Flink...\n", " query: |\n", " select\n", - " to_timestamp(\"Reported_Date\") as reported_date,\n", - " \"Classification_Reported\" as classification,\n", + " to_timestamp(Reported_Date) as reported_date,\n", + " Classification_Reported as classification,\n", " id,\n", - " \"HA\" as ha,\n", - " \"Sex\" as sex,\n", - " \"Age_Group\" as age_group\n", + " HA as ha,\n", + " Sex as sex,\n", + " Age_Group as age_group\n", " from input\n", " # How to combine newly-ingested data with data that is already in the dataset\n", " merge:\n", diff --git a/images/demo/user-home/02 - Web3 Data (Ethereum trading example)/01 - Working with Web3 data.ipynb b/images/demo/user-home/02 - Web3 Data (Ethereum trading example)/01 - Working with Web3 data.ipynb index 9591c226a..6406cb3a2 100644 --- a/images/demo/user-home/02 - Web3 Data (Ethereum trading example)/01 - Working with Web3 data.ipynb +++ b/images/demo/user-home/02 - Web3 Data (Ethereum trading example)/01 - Working with Web3 data.ipynb @@ -228,7 +228,7 @@ "metadata": {}, "source": [ "# Getting datasets from IPFS\n", - "The only thing better than doing an easy task is if **someone else did id for you**... and that's the whole point of [Open Data Fabric network](https://docs.kamu.dev/odf/) - doing the data prep work once and sharing results with the whole world.\n", + "The only thing better than doing an easy task is if **someone else did it for you**... and that's the whole point of [Open Data Fabric network](https://docs.kamu.dev/odf/) - doing the data prep work once and sharing results with the whole world.\n", "\n", "Pulling data that is already in the newtwork is super easy.\n", "\n", @@ -260,7 +260,7 @@ "- `kamu` uses the CID to download the entire dataset block-by-block\n", "- Next time you do `kamu pull net.rocketpool.reth.tokens-minted` only the new blocks will be downloaded (a minimal update)\n", "\n", - "So we are pulling **public ledger** data that is parsed into **analytical data format** by **verifiable code** and stored in a **globally-decentralized file system** as a **near-real-time data stream**.\n", + "So we are pulling **public ledger** data that is parsed into **analytical data format** by **verifiable code** and stored in a **globally-decentralized file system** as a **near-real-time data stream**.\n", "\n", "...Neat!\n", "\n",