Azure Data Factory Managed Airflow | Set up Airflow & Trigger Data Factory Pipelines | Simple steps

แชร์
ฝัง
  • เผยแพร่เมื่อ 2 มิ.ย. 2024
  • In this video we see step by step how to set up Managed Airflow in Data Factory and execute Data Factory pipelines.
    The DAG code can be found at: github.com/apostolos1927/mana...
    Follow me on social media:
    LinkedIn: / apostolos-athanasiou-9...
    GitHub: github.com/apostolos1927/
    Facebook: profile.php?...
    00:00 - Intro
    01:30 - Service Principal registration
    04:00 - Create Airflow connection to Data Factory
    08:37 - Run Simple Data Factory Pipeline from Airflow
    17:40 - Second DAG example
    22:42 - Conclusion

ความคิดเห็น • 16

  • @arjunpalitphotography
    @arjunpalitphotography วันที่ผ่านมา

    Very comprehensive and basics thanks for the tutorial

  • @arnabdutta462
    @arnabdutta462 15 วันที่ผ่านมา

    Thanks a lot for creating this video. For the orchestration. Adf triggers are the only option. With Airflow orchestration we customize on the scheduling part.

  • @user-io8it9ke3e
    @user-io8it9ke3e 4 หลายเดือนก่อน

    It was simple, but I'm new to Azure and spent a lot of time on it.
    Thank you so much for this video

  • @saadjoiya4338
    @saadjoiya4338 4 หลายเดือนก่อน

    Great content and step by step explanantion!

  • @humans34
    @humans34 9 หลายเดือนก่อน +2

    Great! Thanks for the content

  • @lili23880
    @lili23880 4 หลายเดือนก่อน

    really good this video! Very clear explanation and example

  • @iaslink
    @iaslink 7 หลายเดือนก่อน

    First i want to thank you for the content. Been searching a lot for this and you are the only guy who made it clear and concise.
    I have some questions:
    1) Why do i need to connect the managed airflow with datafactory if the managed airflow enviroment was created in the datafactory? That seems strange.
    2) Let's say i don't want to manage any adf pipeline using the managed airflow enviroment, i just created it to execute DAG's that carry transformations... In this case do i really need to connect it with datafactory or just run the DAGs in the airflow?
    3) How the pricing works? I read in the documentation the pricing but not understand it would you be so kind to explain that to me? Is it based on the amount of hours the managed airflow is running?

    • @AthanasiouApostolos
      @AthanasiouApostolos  7 หลายเดือนก่อน +1

      Thank you mate, I appreciate your support!
      1) well, I am not sure if I understand the question correctly but you do have to navigate from the Data Factory UI to the Airflow environment, this is how they built it.
      2) Yes you can execute DAGs that carry transformations outside of Data Factory. For example you can transform data which are in an Azure Database without involving Data Factory. But you need to set up the connection to the Azure SQL database.
      3)The pricing is pretty standard learn.microsoft.com/en-us/azure/data-factory/airflow-pricing you are charged based on the hour and the node size. Small node = 0.5 per hour * 24 hours = 12 dollars per day.

  • @datalearningsihan
    @datalearningsihan 28 วันที่ผ่านมา

    I have an etl process in place in the ADF. In our team, we wanted to implement the table and views transformation and implementation with dbt core. We were wondering if we could orchestrate the dbt with Azure. If so, then how? One of the approaches I could think of was to use Azure Managed Airflow Instance. But, will it allow us to install astronomer cosmos? I have never implemented dbt this way before, so needed to know if this would be the right approach or is there anything else you would suggest me?

    • @AthanasiouApostolos
      @AthanasiouApostolos  28 วันที่ผ่านมา +1

      Unfortunately I haven't tried this approach either so I cannot tell. It seems astronomer cosmos works well with Apache airflow (github.com/astronomer/astronomer-cosmos) so in theory it should work with Azure Managed Airflow instance too. That being said, I haven't tried it, better give it a try and see.

  • @chittaranjanpradhan5290
    @chittaranjanpradhan5290 9 หลายเดือนก่อน

    Good content and nicely explained.could you please share how the airflow job can be triggered automatically.ij the video example you have ran the airflow job manually,how can we do it automatically

    • @AthanasiouApostolos
      @AthanasiouApostolos  9 หลายเดือนก่อน +1

      You either trigger the job manually as I did or you let the scheduler do it. When you define the DAG there is a parameter called schedule. You can specify the exact time using cron syntax if you like. See here for more details: hevodata.com/learn/trigger-airflow-dags/

  • @saleem801
    @saleem801 5 หลายเดือนก่อน

    I have found that managed airflow requires adding drivers for ODBC connections to be made, but the CLI cannot be accessed for the installation of drivers.
    Additionally, I cannot get a sight of memory allocation or add to it. When working with 6gb CSVs my dag breaks :)

  • @Cloudintelligence7
    @Cloudintelligence7 8 หลายเดือนก่อน +1

    Good explanation.. Can we schedule different pipelines from different data factory instances ?
    Example:
    Datafactory1_Pipeline1 ->> Datfactory2_Pipeline2

    • @AthanasiouApostolos
      @AthanasiouApostolos  8 หลายเดือนก่อน

      I haven't tried it yet so I can't provide a definite answer. I don't know if you can achieve it directly from the dag but certainly you can achieve it by triggering an ADF pipeline from the DAG which triggers another ADF pipeline (on a different Data Factory) using Web Activity.

  • @davidchoi1655
    @davidchoi1655 2 หลายเดือนก่อน

    Would it be simpler to use AD authentication as it guaranttes single sign on?