Skip to content

Latest commit

 

History

History
92 lines (55 loc) · 4.21 KB

README.md

File metadata and controls

92 lines (55 loc) · 4.21 KB

The Zero-to-Tinybird Workshop

This repository is a companion piece to the 'Zero-to-Tinybird` 90-minute workshop. The intended audience of this workshop are people who have no experience with Tinybird and are looking to learn the basics with 90 minutes of hands on content. The intent of this workshop is to teach the basics of creating Data Sources, building Pipes, designing Materialized Views, deploying API Endpoints with dynamic query parameters and providing introductions to the Tinybird command-line interface (CLI).

This repository includes resources to help attendees find documentation and other content focused on these topics. It also includes the Data Project used in the workshop. Workshop content is based on a real-time stream of mocked stock prices and a supporting dimensional table of company metadata.

Workshop sections

Prerequisites

  • A free Tinybird account. Navigate to tinybird.co/signup and create a free account. Create a new Workspace (name it whatever you want).
  • Basic knowledge of SQL, querying concepts, and the fundamentals of API endpoint design.

Other resources:

What are we building?

For the workshop project we will start off and building on two data sources:

  • company_info dimensional table with company metadata, including full name, creation date, economic sector, and stock symbol.

  • event_stream live stream of stock prices for a set of ~80 mock companies. These prices are reported every few seconds and published on a Kafka-based stream hosted on Confluent Cloud.

Here is an example time-series

This project includes the Python script used to generate the real-time data stream.

Our intial Tinybird data flow will look like this:

Data flow diagram

Here we have the two Data Sources, and three data 'pipelines' based on them. These Tinybird Pipes illustrate fundamental SQL transformations: filtering, aggregating, and joining data sources.

Then we will create a Materialized View to provide hourly statistics:

Data flow diagram

What topics will we be covering?

  • Creating Data Sources

    • Introduction to native connectors: Kafka/Confluent streams, AWS S3, BigQuery, and Snowflake.

    • Importing dimensional/fact tables.

    • Connecting to a Confluent stream of real-time (mocked) stock prices.

      See more here.

  • Building data analysis pipelines

    • Getting started by developing SQL queries in Playgrounds.

    • Building our first Pipe and publishing an API Endpoint.

    • Building Pipes that filter, aggregate and join Data Sources.

    • Creating dynamic request parameters.

      See more here.

  • Introduction to Materialized Views

    • Purpose and example use cases.
    • Improving performance by filtering and aggregating at ingest time, not query time.
    • Temporal rollups.
    • State/Merge functions.

    See more here.

  • Introduction to the Tinybird CLI

    • Installing the CLI.

    • Exploring our Workspace with the CLI.

    • Touring available commands and preparing for managing projects under version control.

      See more here.

Workshop deck

The first workshop was held on February 13, 2024.

Current workshop deck