How to build data pipelines with Airbyte | Modern Data Stack with Airbyte | Open Source | Airbyte

แชร์
ฝัง
  • เผยแพร่เมื่อ 5 ม.ค. 2025

ความคิดเห็น • 39

  • @BiInsightsInc
    @BiInsightsInc  2 ปีที่แล้ว +3

    Setup required to follow this ETL or ELT pipeline video:
    PostgreSQL Setup: th-cam.com/video/fjYiWXHI7Mo/w-d-xo.html&t
    SQL Server Setup: th-cam.com/video/e5mvoKuV3xs/w-d-xo.html&t
    Original ETL pipeline video: th-cam.com/video/dfouoh9QdUw/w-d-xo.html&t

  • @andressaszx
    @andressaszx 3 หลายเดือนก่อน

    Amazing video, my friend. Thank you for providing this to us. This video showed me what I really need to do, completely hands on. Thank you from Brazil!

  • @alisahibqadimov7659
    @alisahibqadimov7659 10 หลายเดือนก่อน +1

    Hi, tutorial is good. I have been trying Airbyte for almost 1 month. And I can say that it is not good, even really bad for some purposes. Connectors are very, very slow. I deployed it on local machine, docker, Kubernetes same for all of them. Even it is bad with if you have enabled your CDC on source and trying to move some data to destination. 10 rows are loaded in 4 minutes. Good way is that WRITE YOUR OWN CODE.

    • @BiInsightsInc
      @BiInsightsInc  10 หลายเดือนก่อน +1

      Thanks for stopping by. Some of the Airbyte connectors are in beta mood and they do need work. But in my experience they perform way better. I am able to process 232,776 rows under one minute. Anyways, if you want to perform ETL with Python then I also covered that here: th-cam.com/video/dfouoh9QdUw/w-d-xo.html

  • @MochSalmanR1295
    @MochSalmanR1295 2 ปีที่แล้ว

    Thanks fot the tutorial. it helps me undestand airbyte better.

  • @abdullahmusheer4238
    @abdullahmusheer4238 ปีที่แล้ว +2

    when i hit docker-compose up, i get no configuration file provided:not found, what could be the issue?

    • @BiInsightsInc
      @BiInsightsInc  ปีที่แล้ว

      Your Dockerfile name is incorrect or it has an extension. Correct the name and it should work.

  • @saadlechhb3702
    @saadlechhb3702 7 หลายเดือนก่อน +1

    Hello,when i hit docker-compose up, i get no configuration file provided:not found, and when i tried to transfer another yaml file from different source in github to myfolder i get invalid spec: workspace:: empty section between colons, and i don't know how to solve the problem

    • @BiInsightsInc
      @BiInsightsInc  7 หลายเดือนก่อน

      You want to make sure you have docker and docker compose installed. Also, make sure you are in the right directory.

  • @VigyaanJyoti
    @VigyaanJyoti ปีที่แล้ว +1

    I have to load data from SQL Server(Onpremise) to Azure SQL for 100 different customer sources. They all are using same database structure. Is there a dynamic way to create pipelines so that I don't have to do it manually 100 times?? Or Can I create just 1 generic pipeline and change source connection dynamically. Destination (AZure SQl) is anyways same for all.

    • @BiInsightsInc
      @BiInsightsInc  ปีที่แล้ว

      Hello , you can use the Octavia CLI to achieve this. Airbyte provides Configuration as Code (CaC) in YAML and a command line interface (Octavia CLI) to manage resource configurations. Octavia CLI provides commands to import, edit, and apply Airbyte resource configurations: sources, destinations, and connections. I'd advise to look into Octavia as you can manipulate the yaml stored configurations.

  • @hoanglam2814
    @hoanglam2814 2 ปีที่แล้ว

    Amazing as usual!

  • @fernandomaximoferreira1067
    @fernandomaximoferreira1067 7 หลายเดือนก่อน

    Awesome tutorial.

  • @Shataakshi_Ganesh
    @Shataakshi_Ganesh หลายเดือนก่อน

    data transformation is not coming for me.. is it a custom ?

    • @BiInsightsInc
      @BiInsightsInc  หลายเดือนก่อน

      Yes, the transaformation are part of the customization option. This is an EL (Extract and Load) tool. Transformation can be carried out by another tool or SQL. As I mentioned you can use dbt. I will cover this in a future video.

  • @muddashir
    @muddashir 2 ปีที่แล้ว

    Very informative!

  • @machinelearningfromscratch9564
    @machinelearningfromscratch9564 ปีที่แล้ว

    Great video, thanks!

  • @aamirshabeer8648
    @aamirshabeer8648 ปีที่แล้ว

    Hi, I need to get data from Twitch and export that to Local or S3 using AirByte please help me?

    • @BiInsightsInc
      @BiInsightsInc  ปีที่แล้ว +1

      Check if Airbyte has a Twitch connector, establish a connection and remaining process should say the same.

  • @p4palani
    @p4palani ปีที่แล้ว

    Hi, when i tried to work on small schemas(having few tables) i am able to configure connection and able to push the data to snowflake. But when i tried to use big schemas, it always throwing some errors. I am using Redshift as source
    so is there any way to overcome this? what size data Airbyte can move at once?

    • @BiInsightsInc
      @BiInsightsInc  ปีที่แล้ว

      For large datasets you will need to scale Airbyte. Scaling Airbyte is a matter of ensuring that the Docker container or Kubernetes Pod running the jobs has sufficient resources to execute its work. We are mainly concerned with Sync jobs when thinking about scale. Sync jobs sync data from sources to destinations and are the majority of jobs run. Sync jobs use two workers. One worker reads from the source; the other worker writes to the destination. Here is a docs from Airbyte on scaling with recommendation: docs.airbyte.com/operator-guides/scaling-airbyte/

  • @azizaalkuatova9527
    @azizaalkuatova9527 10 หลายเดือนก่อน

    Hi i have a problem with setting up local postgre as destination, it gives error Discovering schema failed
    common.error even if trying with csv, what is the problem, did you have such errors?

    • @BiInsightsInc
      @BiInsightsInc  10 หลายเดือนก่อน

      You may need to increase the timeout configured for the server. Take a look at the following post with similar issue: discuss.airbyte.io/t/failed-to-load-schema-in-discovery-schema-timeout-in-loadbalancer/2665/8

  • @SMCGPRA
    @SMCGPRA 2 ปีที่แล้ว

    Can we do just data transfer between dbs with airbyte creating tables

    • @BiInsightsInc
      @BiInsightsInc  2 ปีที่แล้ว +1

      Yes we can transfer data between databases using Airbyte. Airbyte will create the tables for you in the target environment.

    • @SMCGPRA
      @SMCGPRA 2 ปีที่แล้ว

      @@BiInsightsInc thank you

  • @geethanshr
    @geethanshr 7 หลายเดือนก่อน

    how to install airbyte without git?

    • @BiInsightsInc
      @BiInsightsInc  7 หลายเดือนก่อน

      Simply download their repo from GitHub as a zip file and extract it. Now install using docker.

  • @DerDudeHH
    @DerDudeHH 11 หลายเดือนก่อน

    ?? Where is the "T" in ETL ?
    That's just an ELT Pipeline

    • @BiInsightsInc
      @BiInsightsInc  11 หลายเดือนก่อน +1

      This is the EL part of the ELT. The "T" is carried out with dbt. Here is the link to the whole series:
      hnawaz007.github.io/mds.html
      Here is how you navigate the site: th-cam.com/video/pjiv6j7tyxY/w-d-xo.html

  • @geethanshr
    @geethanshr 7 หลายเดือนก่อน

    Why didn't you use docker?

    • @BiInsightsInc
      @BiInsightsInc  7 หลายเดือนก่อน

      I am using Docker to build an Airbyte container.

  • @STEVEN4841
    @STEVEN4841 ปีที่แล้ว

    But..is not dangerous to give your credentials to tool open source?.. because with that information...your data is totally expose 😢...

    • @BiInsightsInc
      @BiInsightsInc  ปีที่แล้ว +1

      Yes, it is not a good practice to share your credentials. You can change the credentials to your liking and keep them confidential.

  • @jackignatev
    @jackignatev 10 หลายเดือนก่อน

    not mac 💕

    • @BiInsightsInc
      @BiInsightsInc  หลายเดือนก่อน +1

      Docker is OS agnostic. You can deploy it on any platform that runs a docker engine.

  • @huyvu4741
    @huyvu4741 4 หลายเดือนก่อน

    free?

    • @BiInsightsInc
      @BiInsightsInc  4 หลายเดือนก่อน

      Yes, there is an open source version!