Spark Runtime Architecture (Cluster Mode) |

แชร์
ฝัง
  • เผยแพร่เมื่อ 2 ธ.ค. 2024

ความคิดเห็น • 13

  • @2412_Sujoy_Das
    @2412_Sujoy_Das 9 หลายเดือนก่อน +1

    Needed this one badly... Thanks Naresh

  • @RangaM-t2s
    @RangaM-t2s 4 หลายเดือนก่อน +1

    very good job done

  • @mitrasubham
    @mitrasubham 4 หลายเดือนก่อน +1

    Nice explained

  • @PavanKumar-vi7hd
    @PavanKumar-vi7hd 9 หลายเดือนก่อน

    Hi Naresh
    your way of explanation is excellent.
    this is first time i understand spark architeecture is very easy way in Cluster Mode

    • @cleverstudies
      @cleverstudies  9 หลายเดือนก่อน

      Thank you Pavan.❤

  • @pankajchikhalwale8769
    @pankajchikhalwale8769 8 หลายเดือนก่อน

    Please make a short video on the relationship between stages, node, executor, dataframe/dataset/RDD, and core, partition, and task. Want to know what consists what ? And what contains what.

  • @sagararora-e1x
    @sagararora-e1x 8 หลายเดือนก่อน

    How cluster manager can create any of the nearest worker as application master because the configuration of the master can be different. So will it not go and create master to the machine that is configured for master role. With memory allocated to master depending on the type of task.

  • @TechGuyTelugu
    @TechGuyTelugu 7 หลายเดือนก่อน

    What about the node manager do in this architecture

  • @vivekmerugu6711
    @vivekmerugu6711 8 หลายเดือนก่อน

    Hi Naresh,
    Thank you so much,
    This help me a lot. ❤
    Naresh, I've execute my spark application in cluster mode (yarn) in emr cluster,
    My spark application is failing with an exception saying application master container failed 2 times, exists with 137 code.
    This exceptional is occuring for only one dataset which I'm processing with spark application.
    For other datasets, my spark application is working fine.
    The dataset for which spark application is failing having large input payload, ( one record with 25000+ characters ).
    I tried increasing the driver memory and executor memory, now this time , I'm getting an exception while deserialization of input payload.
    Any suggestions how to resolve this issue.
    It will be helpful, please

  • @sohaibshah1771
    @sohaibshah1771 หลายเดือนก่อน

    In your Spark Architecture the driver is created on worker node while in other Architecture I can see the driver on Master Node Why? Thanks

    • @cleverstudies
      @cleverstudies  หลายเดือนก่อน +1

      In cluster mode, both the driver program and the Spark driver run on one of the worker nodes in the cluster. The cluster manager selects a worker node to host the driver.
      FYI: The driver program is the full application you write and submit, while the Spark driver is a component that runs within the driver program to manage the distributed execution of your code.

  • @sohaibshah1771
    @sohaibshah1771 หลายเดือนก่อน

    what will happen if a worker node goes down where we have created driver?

    • @cleverstudies
      @cleverstudies  หลายเดือนก่อน +1

      In cluster mode, if the worker node hosting the driver goes down, the entire Spark application will fail because the driver is responsible for coordinating the job, and Spark has no built-in recovery for driver failure.