Install Apache PySpark on Windows PC | Apache Spark Installation Guide

แชร์
ฝัง
  • เผยแพร่เมื่อ 31 ก.ค. 2024
  • In this lecture, we're going to setup Apache Spark (PySpark) on Windows PC where we have installed JDK, Python, Hadoop and Apache Spark. Please find the below installation links/steps:
    PySpark installation steps on MAC: sparkbyexamples.com/pyspark/h...
    Apache Spark Installation links:
    1. Download JDK: www.oracle.com/in/java/techno...
    2. Download Python: www.python.org/downloads/
    3. Download Spark: spark.apache.org/downloads.html
    Winutils repo link: github.com/steveloughran/winu...
    Environment Variables:
    HADOOP_HOME- C:\hadoop
    JAVA_HOME- C:\java\jdk
    SPARK_HOME- C:\spark\spark-3.3.1-bin-hadoop2
    PYTHONPATH- %SPARK_HOME%\python;%SPARK_HOME%\python\lib\py4j-0.10.9-src;%PYTHONPATH%
    Required Paths:
    %SPARK_HOME%\bin
    %HADOOP_HOME%\bin
    %JAVA_HOME%\bin
    Also check out our full Apache Hadoop course:
    • Big Data Hadoop Full C...
    ----------------------------------------------------------------------------------------------------------------------
    Apache Spark Installation links:
    1. Download JDK: www.oracle.com/in/java/techno...
    2. Download Python: www.python.org/downloads/
    3. Download Spark: spark.apache.org/downloads.html
    -------------------------------------------------------------------------------------------------------------
    Also check out similar informative videos in the field of cloud computing:
    What is Big Data: • What is Big Data? | Bi...
    How Cloud Computing changed the world: • How Cloud Computing ch...
    What is Cloud? • What is Cloud Computing?
    Top 10 facts about Cloud Computing that will blow your mind! • Top 10 facts about Clo...
    Audience
    This tutorial has been prepared for professionals/students aspiring to learn deep knowledge of Big Data Analytics using Apache Spark and become a Spark Developer and Data Engineer roles. In addition, it would be useful for Analytics Professionals and ETL developers as well.
    Prerequisites
    Before proceeding with this full course, it is good to have prior exposure to Python programming, database concepts, and any of the Linux operating system flavors.
    -----------------------------------------------------------------------------------------------------------------------
    Check out our full course topic wise playlist on some of the most popular technologies:
    SQL Full Course Playlist-
    • SQL Full Course
    PYTHON Full Course Playlist-
    • Python Full Course
    Data Warehouse Playlist-
    • Data Warehouse Full Co...
    Unix Shell Scripting Full Course Playlist-
    • Unix Shell Scripting F...
    -----------------------------------------------------------------------------------------------------------------------Don't forget to like and follow us on our social media accounts:
    Facebook-
    / ampcode
    Instagram-
    / ampcode_tutorials
    Twitter-
    / ampcodetutorial
    Tumblr-
    ampcode.tumblr.com
    -----------------------------------------------------------------------------------------------------------------------
    Channel Description-
    AmpCode provides you e-learning platform with a mission of making education accessible to every student. AmpCode will provide you tutorials, full courses of some of the best technologies in the world today. By subscribing to this channel, you will never miss out on high quality videos on trending topics in the areas of Big Data & Hadoop, DevOps, Machine Learning, Artificial Intelligence, Angular, Data Science, Apache Spark, Python, Selenium, Tableau, AWS , Digital Marketing and many more.
    #pyspark #bigdata #datascience #dataanalytics #datascientist #spark #dataengineering #apachespark

ความคิดเห็น • 378

  • @ipheiman3658
    @ipheiman3658 ปีที่แล้ว +3

    This worked so well for me :-) The pace is great and your explanations are clear. I am so glad i came across this, thanks a million! 😄 I have subscribed to your channel!!

  • @sisterkeys
    @sisterkeys 10 หลายเดือนก่อน +3

    What I was doing in 2 days, you narrowed to 30 mins!! Thank you!!

    • @ampcode
      @ampcode  6 หลายเดือนก่อน

      Thank you so much! Subscribe for more content 😊

  • @eloscarc5782
    @eloscarc5782 3 หลายเดือนก่อน

    Your video helped me understand it better than other videos, now the other videos make sense. This was not as convoluted as I thought.

  • @susmayonzon9198
    @susmayonzon9198 ปีที่แล้ว +2

    Excellent! Thank you for making this helpful lecture! You relieved my headache, and I did not give up.

    • @ampcode
      @ampcode  ปีที่แล้ว

      Thank you so much!

    • @moathmtour1798
      @moathmtour1798 ปีที่แล้ว +1

      hey , which version of hadoop did you install because the 2.7 wasn't available

  • @neeleshgaikwad6387
    @neeleshgaikwad6387 ปีที่แล้ว +2

    Very helpful video. Just by following the steps you mentioned I could run the spark on my windows laptop. Thanks a lot for making this video!!

    • @ampcode
      @ampcode  ปีที่แล้ว

      Thank you so much!😊

    • @iniyaninba489
      @iniyaninba489 8 หลายเดือนก่อน

      @@ampcode bro I followed every step you said, but in CMD when I gave "spark-shell", it displayed " 'spark-shell' is not recognized as an internal or external command,
      operable program or batch file." Do you know how to solve this?

    • @sssssshreyas
      @sssssshreyas 2 หลายเดือนก่อน

      @@iniyaninba489 add same path in User Variables Path also, just like how u added in System Variables Path

  • @indianintrovert281
    @indianintrovert281 2 หลายเดือนก่อน +16

    Those who are facing problems like 'spark-shell' is not recognized as an internal or external command
    On command prompt write 'cd C:\Spark\spark-3.5.1-bin-hadoop3\bin' use your own spark filepath(include bin too)
    And then write spark-shell or pyspark (It finally worked for me, hope it works for you too)
    If it worked, like this so that more people benefit from this

    • @SharinH
      @SharinH 2 หลายเดือนก่อน +1

      It worked .. Thank you

    • @jagjodhsingh2358
      @jagjodhsingh2358 2 หลายเดือนก่อน +1

      It worked, thanks :)

    • @Manishamkapse
      @Manishamkapse 2 หลายเดือนก่อน +1

      Thank you 😊 so much it worked

    • @Manishamkapse
      @Manishamkapse 2 หลายเดือนก่อน

      Thank you 😊 so much it worked

    • @vishaltanwar2238
      @vishaltanwar2238 2 หลายเดือนก่อน

      why did we get this error?

  • @cloudandsqlwithpython
    @cloudandsqlwithpython ปีที่แล้ว +1

    Great ! got SPARK working on Windows 10 -- Good work !

    • @ampcode
      @ampcode  6 หลายเดือนก่อน

      Thank you so much! Subscribe for more content 😊

  • @yashusachdeva
    @yashusachdeva 6 หลายเดือนก่อน

    It worked, my friend. The instructions were concise and straightforward.

    • @DecemberBorn12
      @DecemberBorn12 3 ชั่วโมงที่ผ่านมา

      can we connect ?

  • @ragisatyasai2469
    @ragisatyasai2469 ปีที่แล้ว +1

    Thank for sharing this. Beautifully explained.

    • @ampcode
      @ampcode  ปีที่แล้ว

      Glad it was helpful!

  • @joshizic6917
    @joshizic6917 8 หลายเดือนก่อน +6

    how is your spark shell running from your users directory?
    its not running for me

  • @saswatarakshit9488
    @saswatarakshit9488 11 หลายเดือนก่อน

    Great Video, awesome comments for fixing issues

    • @ampcode
      @ampcode  6 หลายเดือนก่อน

      Thank you so much! Subscribe for more content 😊

  • @alulatafere6008
    @alulatafere6008 หลายเดือนก่อน

    Thank you! It is clear and much helpful!! from Ethiopia

  • @nedvy
    @nedvy ปีที่แล้ว +1

    Great video! It helped me a lot. Thank you ❤

    • @ampcode
      @ampcode  ปีที่แล้ว +1

      Thank you so much!

  • @juanmiguelvargascortes9933
    @juanmiguelvargascortes9933 ปีที่แล้ว

    Excellent video!!! Thanks for your help!!!

    • @ampcode
      @ampcode  6 หลายเดือนก่อน

      Thank you so much! Subscribe for more content 😊

  • @HamsiniRamesh-ig6ih
    @HamsiniRamesh-ig6ih 3 หลายเดือนก่อน

    This video was great! Thanks a lot

  • @ramnisanthsimhadri3161
    @ramnisanthsimhadri3161 2 หลายเดือนก่อน +3

    I am not able to find the package type: pre-build for Apache Hadoop 2.7 in the drop-down. FYI - my spark release versions that i can see in the spark releases are 3.4.3 and 3.5.1.

  • @Saravanan_G_Official
    @Saravanan_G_Official 2 หลายเดือนก่อน +2

    is there any thing wrong with the latest version of the python and spark 3.3.1 ?
    i am still getting the error

  • @pratikshyapriyadarshini4677
    @pratikshyapriyadarshini4677 6 หลายเดือนก่อน

    Very Helpful.. Thankyou

  • @user-tr9pz1je7g
    @user-tr9pz1je7g ปีที่แล้ว

    Very helpful, thanks!

    • @ampcode
      @ampcode  6 หลายเดือนก่อน

      Thank you so much! Subscribe for more content 😊

  • @veerabadrappas3158
    @veerabadrappas3158 ปีที่แล้ว +1

    Excellent Video.., Sincere Thank You

    • @ampcode
      @ampcode  ปีที่แล้ว

      Thank you!

  • @davidk7212
    @davidk7212 ปีที่แล้ว +1

    Very helpful, thank you.

    • @ampcode
      @ampcode  ปีที่แล้ว

      Thank you so much!

  • @chrominux5272
    @chrominux5272 5 หลายเดือนก่อน

    Very useful, thanks :D

  • @gosmart_always
    @gosmart_always 9 หลายเดือนก่อน

    Every now and then we receive alert from Oracle to upgrade JDK. Do we need to upgrade our JDK version? If we upgrade, will it impact running of spark.

  • @metaviation
    @metaviation ปีที่แล้ว +1

    very clear one thank you

    • @ampcode
      @ampcode  ปีที่แล้ว

      Thank you!

  • @prashanthnm3406
    @prashanthnm3406 หลายเดือนก่อน

    Thanks bro fixed it after struggling for 2 days 2 nights 2hours 9mins.

    • @nickcheruiyot9069
      @nickcheruiyot9069 หลายเดือนก่อน

      Hello, I have been trying to install it for some days too, I keep getting an error when I try to run the spark shell command is not recognized any suggestions?

  • @user-vq4oz9oc5o
    @user-vq4oz9oc5o ปีที่แล้ว

    Brilliant, Thanks a ton

    • @ampcode
      @ampcode  6 หลายเดือนก่อน

      Thank you so much! Subscribe for more content 😊

  • @theeewebdev
    @theeewebdev 8 หลายเดือนก่อน

    and when downloading the spark a set of files came to download not the tar file

  • @somanathking4694
    @somanathking4694 3 หลายเดือนก่อน

    This works as smooth as butter. Be patient that's it! Once set up done, no looking back.

    • @SUDARSANCHAKRADHARAkula
      @SUDARSANCHAKRADHARAkula 3 หลายเดือนก่อน

      Bro, which version of spark & winutils you've downloaded. I took 3.5.1 and hadoop-3.0.0/bin/winutils but not worked

    • @meriemmouzai2147
      @meriemmouzai2147 2 หลายเดือนก่อน

      @@SUDARSANCHAKRADHARAkula same for me!

  • @theeewebdev
    @theeewebdev 8 หลายเดือนก่อน

    i have fallowed all these steps and installed those 3 and created paths too, but when i go to check in the command prompt... its not working.. error came... can anyone help me please to correct this

  • @shankarikarunamoorthy4391
    @shankarikarunamoorthy4391 2 หลายเดือนก่อน

    sir, spark version is available with Hadoop 3.0 only. Spark-shell not recognized as internal or external command. Please do help.

  • @pooja1899
    @pooja1899 ปีที่แล้ว +1

    Thank you for sharing this video

    • @ampcode
      @ampcode  ปีที่แล้ว

      Most welcome!

  • @Adhikash015
    @Adhikash015 ปีที่แล้ว +1

    Bhai, bro, Brother, Thank you so much for this video

    • @ampcode
      @ampcode  ปีที่แล้ว

      Thank you so much!

  • @pulkitdikshit9474
    @pulkitdikshit9474 3 หลายเดือนก่อน

    hi i installed but when I restarted my pc it is no longer running from cmd? what might be the issue?

  • @rayudusunkavalli2318
    @rayudusunkavalli2318 5 หลายเดือนก่อน +4

    i did every step you have said, but still spark is not working

  • @danieljosephs
    @danieljosephs 5 หลายเดือนก่อน

    Very helpful video

  • @nikhilupmanyu8804
    @nikhilupmanyu8804 5 หลายเดือนก่อน

    Hi, Thanks for the steps. I am unable to see Web UI after installing pyspark. It gives This URL can't be reached. Kindly help

  • @amitkumarpatel7762
    @amitkumarpatel7762 4 หลายเดือนก่อน +2

    I have followed whole instruction but when I am running spark -shell is not recognised

  • @sicelovilane5391
    @sicelovilane5391 ปีที่แล้ว +1

    You are the best. Thanks!

  • @basanthaider3238
    @basanthaider3238 8 หลายเดือนก่อน

    I have an issue with the pyspark it's not working and it's related to java class I can't realy understant what is wrong ???

  • @sibrajbanerjee6297
    @sibrajbanerjee6297 หลายเดือนก่อน +1

    I am getting a message of 'spark-version' is not recognized as an internal or external command,
    operable program or batch file. This is after setting up the path in environment variables for PYSPARK_HOME.

  • @nftmobilegameshindi8392
    @nftmobilegameshindi8392 4 หลายเดือนก่อน +4

    spark shell not working

  • @prateektripathi3834
    @prateektripathi3834 7 หลายเดือนก่อน +4

    Did Everything as per the video, still getting this error : The system cannot find the path specified. on using spark-shell

    • @srishtimadaan03
      @srishtimadaan03 2 หลายเดือนก่อน

      On command prompt write 'cd C:\Spark\spark-3.5.1-bin-hadoop3\bin' use your own spark filepath(include bin too)
      And then write spark-shell or pyspark (It finally worked for me, hope it works for you too)

  • @ed_oliveira
    @ed_oliveira 6 หลายเดือนก่อน +1

    Thank you!
    👍

    • @ampcode
      @ampcode  6 หลายเดือนก่อน

      Thank you so much! Subscribe for more content 😊

  • @Kartik-vy1rh
    @Kartik-vy1rh ปีที่แล้ว +1

    Video is very helpful. Thanks for sharing

    • @ampcode
      @ampcode  ปีที่แล้ว

      Thank you so much!

  • @nagalakshmip8725
    @nagalakshmip8725 3 หลายเดือนก่อน

    I'm getting spark- shell is not recognised as an internal or external command, operable program or batch file

  • @manasa3097
    @manasa3097 11 หลายเดือนก่อน

    This really worked for me...I have completed spark installation but when I'm trying to quit from the scala the cmd is not working and it's showing the error as 'not found'.. can you please help me on this...

  • @NileshKumar9975
    @NileshKumar9975 ปีที่แล้ว +1

    very helpful video

    • @ampcode
      @ampcode  ปีที่แล้ว

      Thank you so much!

  • @Manapoker1
    @Manapoker1 ปีที่แล้ว +1

    the only tutorial that worked for me.....

    • @ampcode
      @ampcode  ปีที่แล้ว

      Thank you so much!

  • @Manoj-ed3lj
    @Manoj-ed3lj หลายเดือนก่อน

    installed successfully but when i am checking hadoop version, i am getting an like hadoop is not recognized as internal or external command

  • @user-oy8gu5cs9j
    @user-oy8gu5cs9j ปีที่แล้ว +1

    ERROR Shell: Failed to locate the winutils binary in the hadoop binary path
    I am getting above error while running spark or pyspark session.
    I have ensured that winutils file is present in C:\hadoop\bin

    • @ampcode
      @ampcode  ปีที่แล้ว

      Could you please let me know if your all the env variables are set properly?

  • @sanketraut8462
    @sanketraut8462 5 หลายเดือนก่อน

    how to set up com.jdbc.mysql.connector using jar file, actually I am getting the same error that its not found while working in pyspark.

  • @arnoldochris5082
    @arnoldochris5082 8 หลายเดือนก่อน +5

    Ok guys this is how to do it, incase you are having problems👇
    1.) I used the latest version 3.5.0, (Pre-built for apache hadoop 3.3 or later) - downloaded it.
    2.) Extracted the zip file just as done, the first time it gave me a file, not a folder but a .rar file which winrar could not unzip, so I used 7-zip and it finally extracted to a folder that had the bins and all the other files.
    3.) In the system variables he forgot to edit the path variables and to add %SPARK_HOME%\bin.
    4.) Downloaded winutils.exe for hadoop 3.0.0 form the link provided in the video.
    5.) Added it the same way but c>Hadoop>bin>winutils.exe
    6.) Then edit the user variables as done then do the same to the path %HADOOP_HOME%\bin
    Reply for any parts you might have failed to understand🙂

    • @MANALROGUI
      @MANALROGUI 8 หลายเดือนก่อน

      What do you mean for the 3rd step ?

    • @stay7485
      @stay7485 7 หลายเดือนก่อน

      Thanks

    • @ampcode
      @ampcode  6 หลายเดือนก่อน

      Thank you so much 😊

    • @sarahq6497
      @sarahq6497 2 หลายเดือนก่อน +1

      Hello, I had to use the latest version as well, but I'm not able to make it work, I followed the tutorial exactly :(

  • @ganeshkalaivani6250
    @ganeshkalaivani6250 ปีที่แล้ว +1

    can any one please help...last two days tried to install spark and give correct variable path but still getting system path not speicifed

    • @ampcode
      @ampcode  ปีที่แล้ว

      Sorry for late reply. Could you please check if your spark-shell is running properly from the bin folder. If yes I guess there are some issues with your env variables only. Please let me know.

  • @AmreenKhan-dd3lf
    @AmreenKhan-dd3lf 12 วันที่ผ่านมา

    Apache 2.7 option not available during spark download. Can we choose Apache Hadoop 3.3 and later ( scala2.13) as package type during download

  • @edu_tech7594
    @edu_tech7594 ปีที่แล้ว +1

    my Apache hadoop which i downloaded previously is version 3.3.4 eventhough i should choose pre-built for Apache Hadoop 2.7?

    • @sriram_L
      @sriram_L 11 หลายเดือนก่อน

      Same doubt bro.
      Did u install now

  • @Karansingh-xw2ss
    @Karansingh-xw2ss 10 หลายเดือนก่อน +2

    i'm facing this issue can anyone help me to fix this 'spark-shell' is not recognized as an internal or external command,
    operable program or batch file'.

    • @nikhilupmanyu8804
      @nikhilupmanyu8804 5 หลายเดือนก่อน

      Try to add direct path at System Environment. It will fix the issue

  • @bramhanaskari3152
    @bramhanaskari3152 ปีที่แล้ว +1

    you haven't give solution for that warn procfsMetricsGetter exception is there any solution for that ?

    • @ampcode
      @ampcode  ปีที่แล้ว

      Sorry for late response. This could happen in windows only and can be safely ignored. Could you please confirm if you’re able to kick off spark-shell and pyspark?

  • @badnaambalak364
    @badnaambalak364 6 หลายเดือนก่อน +1

    I followed the steps & Installed JDK 17, spark 3.5 and python 3.12 when I am trying to use map function I am getting an Py4JJavaError: An error occurred while calling z:org.apache.spark.api.python.PythonRDD.collectAndServe Error please someone help me

  • @rakeshkandula2318
    @rakeshkandula2318 8 หลายเดือนก่อน +2

    Hi, i followed exact steps (installed spark 3.2.4 as that is the only version available for hadoop 2.7). Spark-shell command is working but pyspark is thrwing errors.
    if anyone has fix to this please help me.
    Thanks

    • @thedataguyfromB
      @thedataguyfromB 7 หลายเดือนก่อน

      Step by step solution
      th-cam.com/video/jO9wZGEsPRo/w-d-xo.htmlsi=aaITbbN7ggnczQTc

  • @rakeshd3250
    @rakeshd3250 7 หลายเดือนก่อน

    not working for me i set up everything except hadoop version came with 3.0

  • @juliocesarcabanillas2433
    @juliocesarcabanillas2433 11 หลายเดือนก่อน

    Love you dude

    • @ampcode
      @ampcode  6 หลายเดือนก่อน

      Thank you so much! Subscribe for more content 😊

  • @user-vb7im1jb1b
    @user-vb7im1jb1b 11 หลายเดือนก่อน

    Thanks for this video. For learning purposes on my own computer, do I need to install apache.spark (spark-3.4.1-bin-hadoop3.tgz) to be able to run spark scripts/notebooks, or just pip install pyspark on my python environment?

    • @practicemail3227
      @practicemail3227 2 หลายเดือนก่อน

      Hi, I'm in the same boat, can you tell me what did you do. I'm also learning currently and have no idea.

  • @riptideking
    @riptideking 4 หลายเดือนก่อน +2

    'pyspark' is not recognized as an internal or external command,
    operable program or batch file.
    getting this error and tried it for whole day and same issue.

    • @srishtimadaan03
      @srishtimadaan03 2 หลายเดือนก่อน

      On command prompt write 'cd C:\Spark\spark-3.5.1-bin-hadoop3\bin' use your own spark filepath(include bin too)
      And then write spark-shell or pyspark (It finally worked for me, hope it works for you too)

  • @vennilagunasekhar5460
    @vennilagunasekhar5460 ปีที่แล้ว

    Thank you so much

    • @ampcode
      @ampcode  6 หลายเดือนก่อน

      Thank you so much! Subscribe for more content 😊

  • @harshithareddy5087
    @harshithareddy5087 6 หลายเดือนก่อน +3

    I don't have the option for Hadoop 2.7 what to choose now???

    • @LLM_np
      @LLM_np 5 หลายเดือนก่อน

      did you get any solution?
      please let me know

  • @anastariq1310
    @anastariq1310 ปีที่แล้ว +1

    After entering pyspark in cmd it shows "The system cannot find the path specified. Files\Python310\python.exe was unexpected at this time" please help me resolve it

    • @mahamudullah_yt
      @mahamudullah_yt ปีที่แล้ว

      i face the same problem. is there any solution

  • @nagarajgotur
    @nagarajgotur ปีที่แล้ว +2

    spark-shell is working for me, pyspark is not working from home directory, getting error 'C:\Users\Sana>pyspark
    '#' is not recognized as an internal or external command,
    operable program or batch file.'
    But when I go to python path and run the cmd pyspark is working. I have setup the SPARK_HOME and PYSPARK_HOME environment variables. Could you please help me. Thanks

    • @ampcode
      @ampcode  ปีที่แล้ว

      Sorry for late response. Could you please also set PYSPARK_HOME as well to your python.exe path. I hope this will solve the issue😅👍

    • @bintujose1981
      @bintujose1981 ปีที่แล้ว

      @@ampcode nope. Same error

  • @user-uc7qf6uf5c
    @user-uc7qf6uf5c 8 หลายเดือนก่อน +1

    Great thanks

    • @ampcode
      @ampcode  6 หลายเดือนก่อน

      Thank you so much! Subscribe for more content 😊

  • @DevSharma_31
    @DevSharma_31 11 หลายเดือนก่อน

    I am getting this error while running spark-shell or pyspark "java.lang.IllegalAccessError: class org.apache.spark.storage.StorageUtils$ (in unnamed module @0x46fa7c39) cannot access class sun.nio.ch.DirectBuffer (in module java.base) because module java.base does not export sun.nio.ch to unnamed module @0x46fa7c39" I tried all version of java as well as spark, Please help

  • @matheswaranp9574
    @matheswaranp9574 หลายเดือนก่อน

    Thanks a Lot.

  • @user-ef9vh7qz9h
    @user-ef9vh7qz9h ปีที่แล้ว

    java.lang.IllegalAccessException: final field has no write access:
    I'm getting this error while running the code
    when I run the same code in another system it is getting executed.
    Any idea?

  • @Jerriehomie
    @Jerriehomie ปีที่แล้ว +2

    Getthing this error: WARN ProcfsMetricsGetter: Exception when trying to compute pagesize, as a result reporting of ProcessTree metrics is stopped. People have mentioned to use python fodler path which I have as you have mentioned but still.

    • @bukunmiadebanjo9684
      @bukunmiadebanjo9684 ปีที่แล้ว +1

      I found a fix for this. Change your python path to that of anaconda(within the environment variable section of this video) and use your anaconda command prompt instead. No errors will pop up again.

    • @ampcode
      @ampcode  ปีที่แล้ว

      Sorry for late response. Could you please let me know if you are still facing this issue and also confirm if you’re able to open spark-shell?

    • @shivalipurwar7205
      @shivalipurwar7205 ปีที่แล้ว +1

      @@bukunmiadebanjo9684 Hi Adebanjo, my error got resolved with you solution. Thanks for your help!

  • @abhinavtiwari6186
    @abhinavtiwari6186 ปีที่แล้ว +1

    where is that git repository link? Its not there in the description box below

    • @ampcode
      @ampcode  ปีที่แล้ว +1

      Extremely sorry for that. I have added it in the description as well as pasting it here.
      GitHUB: github.com/steveloughran/winutils
      Hope this is helpful! :)

  • @jeremychaves2269
    @jeremychaves2269 ปีที่แล้ว

    thanks dude!

    • @ampcode
      @ampcode  6 หลายเดือนก่อน

      Thank you so much! Subscribe for more content 😊

  • @ankushv2642
    @ankushv2642 7 หลายเดือนก่อน

    Did not work for me. At last when I typed the pyspark in command prompt, it did not worked.

  • @saikrishnareddy3474
    @saikrishnareddy3474 10 หลายเดือนก่อน +2

    I’m little confused on how to setup the PYTHONHOME environment variable

    • @thedataguyfromB
      @thedataguyfromB 7 หลายเดือนก่อน

      Step by step
      th-cam.com/video/jO9wZGEsPRo/w-d-xo.htmlsi=aaITbbN7ggnczQTc

  • @itsshehri
    @itsshehri ปีที่แล้ว +1

    hey pyspark isnt working at my pc. I did everything how you asked. Can you help please

    • @ampcode
      @ampcode  ปีที่แล้ว

      Sorry for late response. Could you please also set PYSPARK_HOME env variable to the python.exe path. I guess this’ll do the trick😅👍

  • @ganeshkalaivani6250
    @ganeshkalaivani6250 ปีที่แล้ว +1

    FileNotFoundError: [WinError 2] The system cannot find the file specified getting this error even i have installed all required intalliation

    • @ampcode
      @ampcode  ปีที่แล้ว

      Sorry for late reply. I hope your issue is resolved. If not we can have a connect and discuss further on it!

  • @Cardinal_Seen
    @Cardinal_Seen 10 หลายเดือนก่อน

    Thank you. :D

    • @ampcode
      @ampcode  6 หลายเดือนก่อน

      Thank you so much! Subscribe for more content 😊

  • @sanchitabhattacharya353
    @sanchitabhattacharya353 5 หลายเดือนก่อน +1

    while launching the spark-shell getting the following error, any idea??
    WARN jline: Failed to load history
    java.nio.file.AccessDeniedException: C:\Users\sanch\.scala_history_jline3

    • @DecemberBorn12
      @DecemberBorn12 2 ชั่วโมงที่ผ่านมา

      resolved hua ?

  • @ismailcute1584
    @ismailcute1584 5 หลายเดือนก่อน +3

    Thank you so much for this video. Unfortunately, I couldn't complete this - getting this erros C:\Users\Ismahil>spark-shell
    'cmd' is not recognized as an internal or external command,
    operable program or batch file. please help

  • @syamprasad8295
    @syamprasad8295 10 หลายเดือนก่อน +1

    while selecting a package type for spark, Hadoop 2.7 is not available now. Only Hadoop 3.3 and later is available. And winutils 3.3 is not available at the link provided at the git. What to do now? can I download Hadoop 3.3 version and can proceed with winutils2.7 ? Pls help.. Thanks In Advacnce

    • @ShivamRai-xo8fu
      @ShivamRai-xo8fu 9 หลายเดือนก่อน

      I got same issue

    • @ankitgupta5446
      @ankitgupta5446 7 หลายเดือนก่อน

      100 % working solution
      th-cam.com/video/jO9wZGEsPRo/w-d-xo.htmlsi=lzXq4Ts7ywqG-vZg

  • @sriramsivaraman4100
    @sriramsivaraman4100 ปีที่แล้ว +2

    Hello when I try to run the command spark_shell as a local user its not working (not recognized as an internal or external command) and it only works if I use it as an administratror. Can you please help me solve this? Thanks.

    • @ampcode
      @ampcode  ปีที่แล้ว

      Sorry for late response. Could you please try once running the same command from the spark/bin directory and let me know. I guess there might be some issues with your environment vatiables🤔

    • @dishantgupta1489
      @dishantgupta1489 ปีที่แล้ว

      @@ampcode followed each and every step of video still getting not recognised as an internal or external command error

    • @ayonbanerjee1969
      @ayonbanerjee1969 ปีที่แล้ว

      ​@@dishantgupta1489 open fresh cmd prompt window and try after you save the environment variables

    • @obulureddy7519
      @obulureddy7519 ปีที่แล้ว

      In Environment Variables you give the paths in Users variable Admin. NOT IN System variables

  • @prajakta-dh7fc
    @prajakta-dh7fc 2 หลายเดือนก่อน

    'spark' is not recognized as an internal or external command,
    operable program or batch file. its not working for me i have follow all the steps but its still not working waiting for solution

  • @manikantaperumalla2197
    @manikantaperumalla2197 หลายเดือนก่อน

    java,python and spark should be in same directory?

  • @Nathisri
    @Nathisri 8 หลายเดือนก่อน +1

    I have some issues in launching python & pyspark. I need some help. Can you pls help me?

  • @KarpagamE-fu5xn
    @KarpagamE-fu5xn 9 หลายเดือนก่อน

    how to clear this problem,
    The system cannot find the path specified.

    • @anshuman7559
      @anshuman7559 3 หลายเดือนก่อน

      Hey, did you get it resolved? Please let me know how to fix this issue.

  • @Mralbersan
    @Mralbersan 3 หลายเดือนก่อน

    I can't see Pre-Built for Apache Hadoop 2.7 on the spark website

    • @meriemmouzai2147
      @meriemmouzai2147 2 หลายเดือนก่อน

      same problem for me! I tried the "3.3 and later" version with the "winutils/hadoop-3.0.0/bin", but it didn't work

  • @khushboojain3883
    @khushboojain3883 11 หลายเดือนก่อน +1

    Hi, I have installed Hadoop 3.3 (the lastest one) as 2.7 was not available. But while downloading winutils, we don't have for Hadoop 3.3 in repository. Where do i get it from?

    • @sriram_L
      @sriram_L 11 หลายเดือนก่อน

      Same here.Did u get it now?

    • @khushboojain3883
      @khushboojain3883 11 หลายเดือนก่อน

      @@sriram_L yes, u can directly get it from google by simply mention the Hadoop version for which u want winutils. I hope this helps.

    • @hritwikbhaumik5622
      @hritwikbhaumik5622 9 หลายเดือนก่อน

      @@sriram_L it still not working for me though

  • @moathmtour1798
    @moathmtour1798 ปีที่แล้ว +1

    hello, which Hadoop Version should i install since the 2.7 is not available anymore ? thanks in advance

    • @ampcode
      @ampcode  ปีที่แล้ว

      You can go ahead and install the latest one as well. no issues!

    • @venkatramnagarajan2302
      @venkatramnagarajan2302 9 หลายเดือนก่อน

      @@ampcode Will the utils file still be 2.7 version ?

  • @shahrahul5872
    @shahrahul5872 ปีที่แล้ว +1

    on apache spark's installation page, under choose a package type, the 2.7 version seem to not be any option anymore as on 04/28/2023. What to do?

    • @shahrahul5872
      @shahrahul5872 ปีที่แล้ว +2

      I was able to get around this by copying manually the URL of the site you were opened up to after selecting the 2.7th version from the dropdown. Seems like they have archived it.

    • @ampcode
      @ampcode  ปีที่แล้ว

      Sorry for late reply. I hope your issue is resolved. If not we can discuss further on it!

  • @infamousprince88
    @infamousprince88 15 วันที่ผ่านมา

    I'm still unable to get this to work. I've been trying to solve this problem for nearly 2 weeks

  • @kchavan67
    @kchavan67 8 หลายเดือนก่อน +1

    Hi, following all the steps given in video, I am still getting error as "cannot recognize spark-shell as internal or external command" @Ampcode

    • @psychoticgoldphish5797
      @psychoticgoldphish5797 8 หลายเดือนก่อน

      I was having this issue as well, when I added the %SPARK_HOME%\bin, %HADOOP_HOME%\bin and %JAVA_HOME%\bin to the User variables (top box, in the video he shows doing system, bottom box) it worked. Good luck.

    • @thedataguyfromB
      @thedataguyfromB 7 หลายเดือนก่อน

      Step by step spark + PySpark in pycharm solution video
      th-cam.com/video/jO9wZGEsPRo/w-d-xo.htmlsi=aaITbbN7ggnczQTc

  • @syafiq3420
    @syafiq3420 ปีที่แล้ว +1

    how did you download the apache spark in zipped file? mine was downloaded as tgz file

    • @ampcode
      @ampcode  ปีที่แล้ว

      Sorry for late response. You’ll get both options on their official website. Could you please check if you are using the right link?

    • @georgematies2521
      @georgematies2521 ปีที่แล้ว

      @@ampcode There is no way now to download the zip file, only tgz.

  • @Analystmate
    @Analystmate 10 หลายเดือนก่อน

    C:\Users\lavdeepk>spark-shell
    'spark-shell' is not recognized as an internal or external command,
    operable program or batch file.
    Not working

    • @syamprasad8295
      @syamprasad8295 10 หลายเดือนก่อน

      which winutil file did u download? Its Hadoop 2.7 or later version?

  • @akira.19.9
    @akira.19.9 11 หลายเดือนก่อน

    muy util !!

    • @ampcode
      @ampcode  6 หลายเดือนก่อน

      Thank you so much! Subscribe for more content 😊

  • @user-gc6ku9mp3d
    @user-gc6ku9mp3d ปีที่แล้ว +6

    Hi, I completed the process step by step and everything else is working but when I run 'spark-shell' , it shows - 'spark-shell' is not recognized as an internal or external command,
    operable program or batch file. Do you know what went wrong?

    • @viniciusfigueiredo6740
      @viniciusfigueiredo6740 ปีที่แล้ว +1

      I'm having this same problem, the command only works if I run CMD as an administrator. Did you manage to solve it?

    • @hulkbaiyo8512
      @hulkbaiyo8512 11 หลายเดือนก่อน

      @@viniciusfigueiredo6740 same as you, run as administrator works

    • @shivamsrivastava4337
      @shivamsrivastava4337 11 หลายเดือนก่อน

      @@viniciusfigueiredo6740 same issue is happening with me

    • @RohitRajKodimala
      @RohitRajKodimala 11 หลายเดือนก่อน

      @@viniciusfigueiredo6740same issue for me did u fix it?

    • @santaw
      @santaw 8 หลายเดือนก่อน +1

      Anyone solved this?

  • @nikhilchavan7741
    @nikhilchavan7741 8 หลายเดือนก่อน

    'spark-shell' is not recognized as an internal or external command,
    operable program or batch file.-- Getting this error

    • @nayanagrawal9878
      @nayanagrawal9878 8 หลายเดือนก่อน

      Do everything that he said but not in User Variables but in System variables. I was facing the same problem but then I did the same in system variables and my spark started running.

  • @laxman0457
    @laxman0457 10 หลายเดือนก่อน +2

    i have followed all your steps,still i'm facing an issue.
    'spark2-shell' is not recognized as an internal or external command

    • @nayanagrawal9878
      @nayanagrawal9878 8 หลายเดือนก่อน

      Do everything that he said but not in User Variables but in System variables. I was facing the same problem but then I did the same in system variables and my spark started running.

    • @thedataguyfromB
      @thedataguyfromB 7 หลายเดือนก่อน

      Step by step spark + PySpark in pycharm solution video
      th-cam.com/video/jO9wZGEsPRo/w-d-xo.htmlsi=aaITbbN7ggnczQTc

  • @karthikeyinikarthikeyini380
    @karthikeyinikarthikeyini380 9 หลายเดือนก่อน +1

    hadoop 2.7 tar file is not available in the link

    • @ankitgupta5446
      @ankitgupta5446 7 หลายเดือนก่อน

      100 % working solution
      th-cam.com/video/jO9wZGEsPRo/w-d-xo.htmlsi=lzXq4Ts7ywqG-vZg