Databricks Asset Bundles: Advanced Examples

แชร์
ฝัง
  • เผยแพร่เมื่อ 5 ก.ย. 2024

ความคิดเห็น • 23

  • @asuretril867
    @asuretril867 16 วันที่ผ่านมา

    Thanks a lot Dustin... Really appreciate it :)

  • @NoahPitts713
    @NoahPitts713 2 หลายเดือนก่อน

    Exciting stuff! Will definitely be trying to implement this in my future work!

  • @pytalista
    @pytalista 20 วันที่ผ่านมา

    Thanks for the video. It helped me a lot in my YT channel.

  • @bartsimons6325
    @bartsimons6325 23 วันที่ผ่านมา

    Great video Dustin! Especially on the advanced configuration of the databricks.yaml.
    I'd like to hear your opinion on the /src in the root of the folder. If you're team/organisation is used to work with a mono repo it would be great to have all common packages in the root, however, if you're more of a polyrepo kinda team/organisation, building and hosting the packages remotely (i.e. Nexus or something) could be a better approach in my opinion. Or am I missing something?
    How would you deal with a job where task 1 and task 2 have source code with conflicting dependencies?

  • @ameliemedem1918
    @ameliemedem1918 2 หลายเดือนก่อน

    Thanks a lot, @DustinVannoy for this great presentation! I have a question: which is the better approach for project structuration: one bundle yml config file for all my sub-projects or each sub-project have its own Databricks and bundle yml file? Thanks again :)

  • @gardnmi
    @gardnmi 2 หลายเดือนก่อน

    Loving bundles so far. Only issue so far I've had is the databricks vscode extension seems to be modifying my bundles yml file behind the scenes. For example when I attach to a cluster in the extension it will override my job cluster to use that attached cluster when I deploy to the dev target in development mode.

    • @DustinVannoy
      @DustinVannoy  2 หลายเดือนก่อน

      Which version of the extension are you on, 1.3.0?

    • @gardnmi
      @gardnmi 2 หลายเดือนก่อน

      ​@@DustinVannoyYup, I did have it on a pre release which I thought was the issue but switched back to 1.3.0 and the "feature" persisted.

  • @deepakpatil5059
    @deepakpatil5059 8 วันที่ผ่านมา

    Great content!! I am trying to deploy the same job into different environments DEV/QA/PRD. I want to override parameters passed to the job from variable-group defined on the Azure DevOps portal. Can you please suggest how to proceed on this?

    • @DustinVannoy
      @DustinVannoy  4 วันที่ผ่านมา +1

      The part that references variables group PrdVariables shows how you set different variables and values depending on target environment.
      - stage: toProduction
      variables:
      - group: PrdVariables
      condition: |
      eq(variables['Build.SourceBranch'], 'refs/heads/main')
      In the part where you deploy the bundle, you can pass in variable values. See the docs for how that can be set. docs.databricks.com/en/dev-tools/bundles/settings.html#set-a-variables-value

  • @DataMyselfAI
    @DataMyselfAI 24 วันที่ผ่านมา

    Is there a way for python wheel tasks to combine the functionality we had without serverless to use:
    libraries: - whl../dist/*.whl so that the wheel gets deployed automatically with using serverless?
    As if I am trying to include environments for serverless I can't longer specify libraries for the wheel task (and therefore it is not deployed automatically) and I also need to hardcode my path for the wheel in the workspace.
    Could not find an example for that so far.
    All the best,
    Thomas

    • @DustinVannoy
      @DustinVannoy  4 วันที่ผ่านมา

      Are you trying to install the wheel in a notebook task, so you are required to install with %pip install?
      If you include the artifact section it should build and upload the wheel regardless of usage in a taks. You can predict the path within the .bundle deploy if you aren't setting mode: development, but I've been uploading it to a specific workspace or volume location.
      As environments for serverless evolve I may come back wtih more examples of how those should be used.

  • @etiennerigaud7066
    @etiennerigaud7066 2 หลายเดือนก่อน

    Great video ! Is there a way to overide variables defined in the databricks.yml in each of the job yml definition so that the variable has a different value for that job only ?

    • @DustinVannoy
      @DustinVannoy  4 วันที่ผ่านมา

      If value is the same for a job across all targets you wouldn't use a variable. To override job values you would set those in the target section which I always include in databricks.yml.

  • @dreamsinfinite83
    @dreamsinfinite83 หลายเดือนก่อน

    how do you change the Catalog Name specific to an environment?

    • @DustinVannoy
      @DustinVannoy  17 วันที่ผ่านมา

      I would use a bundle variable and set it in the target overrides, then reference it anywhere you need it.

  • @fortheknowledge145
    @fortheknowledge145 2 หลายเดือนก่อน

    Can we integrate Azure pipelines + DAB for ci cd implementation?

    • @DustinVannoy
      @DustinVannoy  2 หลายเดือนก่อน +2

      Are you referring to Azure DevOps CI pipelines? You can do that and I am considering a video on that since it has been requested a few times.

    • @fortheknowledge145
      @fortheknowledge145 2 หลายเดือนก่อน

      @@DustinVannoy yes, thank you!

    • @felipeporto4396
      @felipeporto4396 หลายเดือนก่อน

      @@DustinVannoy Please, can you do that? hahaha

    • @DustinVannoy
      @DustinVannoy  17 วันที่ผ่านมา +1

      Video showing Azure DevOps Pipeline is published!
      th-cam.com/video/ZuQzIbRoFC4/w-d-xo.html

  • @9829912595
    @9829912595 2 หลายเดือนก่อน

    Once the code is deployed it gets uploaded in the shared folder can't we store that some where else like an artifact or storage account because there are chances that someone may deleted that bundle from shared folder. It is always like with databricks deployment before and after asset bundles.

    • @DustinVannoy
      @DustinVannoy  2 หลายเดือนก่อน

      You can set permissions on the workspace folder and I recommend also having it all checked into version control such as GitHub in case you ever need to recover an older version.