Indeed Jobs Web Scraping Save to CSV

แชร์
ฝัง
  • เผยแพร่เมื่อ 6 พ.ย. 2024

ความคิดเห็น • 238

  • @franklinokech
    @franklinokech 3 ปีที่แล้ว +46

    Great tutorial John, just a quick fix on the for loop, you forgot to add the i to the extract function: Made the changes to this
    for i in range(0, 41, 10):
    print(f'Getting page {i} ')
    c = extract(i)
    transform(c)

    • @JohnWatsonRooney
      @JohnWatsonRooney  3 ปีที่แล้ว +8

      Great thank you!

    • @sosentreprises9411
      @sosentreprises9411 2 ปีที่แล้ว +1

      Hi everyone,
      I had the following error message :
      File "/Users/admin/Downloads/Test.py", line 36, in
      transform (C)
      File "/Users/admin/Downloads/Test.py", line 22, in transform
      summary = item.find('div', class_='job-snippet').text.strip().replace('
      ','') #trouver sommaire et remplacer new lines with noth'
      UnboundLocalError: local variable 'item' referenced before assignment
      ANY HELP ?

  • @vvvvv432
    @vvvvv432 2 ปีที่แล้ว +8

    That's an excellent video for the following reasons:
    -- the flow of the tutorial is really smooth,
    -- the explanation is excellent so you can easily adjust the classes that existed in the time of the video to the current ones
    -- and the iterations are detailed so every step is easy to understand.
    Thank you so much for this video! Greetings from Greece! 🇬🇷

  • @JulianFerguson
    @JulianFerguson 4 ปีที่แล้ว +17

    I am very surprised you only have 1500 views. This is one of the best webscraping tutorials i have come across. Can you do one for rightmove or zoopla?

    • @thenoobdev
      @thenoobdev 2 ปีที่แล้ว +1

      Heheh already at 42k today 😁 well deserved

  • @igordc16
    @igordc16 2 ปีที่แล้ว +3

    Worked flawlessly, I just had to edit a few things, like the classes and the tags. Nothing wrong with the code, just the indeed website that changed since you posted this video. Thanks!

  • @davidberrien9711
    @davidberrien9711 2 ปีที่แล้ว +9

    Hello, John. I have just started learning Python, and I'm trying to use it automate some daily tasks, and web scraping is my current "class". I really enjoy watching your workflow. I love watching the incremental development of the program as you work your way through. You are very fluent in the language, as well as the various libraries you demonstrate. I am still at the stage where I have to look up the syntax of tuples and dictionaries... (Is it curly brace or brackets? commas or colons?) so I find myself staring in amazement as two or three lines of code blossom into 20, and this wondrously effective program is completed in minutes... I am envious of your skill, and I wanted to let you know I appreciate your taking the time to share your knowledge. I find your content compelling. Sometimes I forget to click the like button before moving on to the next vid, so sorry about that. I just have to go watch it again, just to make sure I leave a like... Your work is very inspiring to me as a noob. I aspire to the same type of fluency as you demonstrate so charmingly. Thanks again.

    • @JohnWatsonRooney
      @JohnWatsonRooney  2 ปีที่แล้ว +5

      Hi David! Thank you very much for the comment, I really appreciate it. It's always great to hear that the content I make is helping out. learning programming is a skill and will take time, but if you stick with it, things click and then no doubt you'll be watching my videos and commenting saying I could have done it better! (which is also absolutely fine) John

  • @rukon8887
    @rukon8887 2 ปีที่แล้ว +3

    John amazing tutorial and skills, love the way you slip in sometimes different method on going about. Hope you getting big bucks for your expertise. Keep them video coming.

  • @Why_So_Saad
    @Why_So_Saad 2 ปีที่แล้ว +1

    This channel has helped me a lot. Everything I know about web scrapping is thanks to John and his to-the-point tutorials.

  • @hi_nesh
    @hi_nesh 3 หลายเดือนก่อน +1

    Honestly, This channel is marvelous. It has helped me a lot. 'a lot' is even an understatement

  • @afrodeveloper3929
    @afrodeveloper3929 4 ปีที่แล้ว +2

    Your style of code is so beautiful and easy to follow.

  • @ibrahimseck8520
    @ibrahimseck8520 2 ปีที่แล้ว +2

    I couldn't thank you enough for this tutorial...I am following a Python course on Udemy for the moment, and I found the section on web scraping incomplete...I followed this tutorial and it's brilliant...The indeed page is quite different including the html code, but the logic stays the same...I will put my code in the comments, it might be of interest especially for people using indeed in french

    • @oximine
      @oximine 2 ปีที่แล้ว +1

      Any update on your code bud? I'm trying to scrape indeed right now and the html looks very different than what's in the video

    • @cryptomoonmonk
      @cryptomoonmonk 2 ปีที่แล้ว +2

      @@oximine Yea, indeed changed their code. I had a rough time figuring it out.
      the job title is no longer in the 'a' tag in the new html.
      At 9:13 in the video you need to use:
      ...
      divs = soup.find_all('div', class_='heading4')
      for item in divs:
      title = item.find('h2').text
      print(title) ...
      Reason being, Indeed now has the title of each job title within an h2 element which is in the class starting with heading4.
      So the code searches for the class heading4, once it finds it will search for the 'title' item in the h2 element
      Just look at the html and see where the "title" of the job search is in the new code.
      One thing is for sure, once you figure this out and understand it, you understand what's going on.

    • @michealdmouse
      @michealdmouse 2 ปีที่แล้ว

      @@cryptomoonmonk The code works. Thank you for sharing.

  • @sujithchennat
    @sujithchennat 2 ปีที่แล้ว +4

    Good work John, please use the variable i in the extract function to avoid duplicate results

  • @eligr8523
    @eligr8523 ปีที่แล้ว +2

    Thank you. You saved my entire semester!

  • @martpagente7587
    @martpagente7587 4 ปีที่แล้ว +3

    Very thankful to your videos John, we support your channel and your popular now in youtube, I wish you can make video also scraping LinkIn or Zillow website, these are demands from freelance sites

    • @JohnWatsonRooney
      @JohnWatsonRooney  4 ปีที่แล้ว +2

      Sure I can have a look at the best way to scrape those sites

    • @expat2010
      @expat2010 4 ปีที่แล้ว

      @@JohnWatsonRooney That would be great and don't forget the github link when you do. :)

  • @ιυ_αα-ξ5σ
    @ιυ_αα-ξ5σ 2 ปีที่แล้ว

    Hey. Seriously. Thank you. I just downloaded soft and I can CLEARLY see why your vid was recomnded. You're an aweso intro into

  • @datasciencewithshaxriyor7153
    @datasciencewithshaxriyor7153 2 ปีที่แล้ว +1

    bro with your help i have finished my project

  • @dmytrodavydenko7467
    @dmytrodavydenko7467 ปีที่แล้ว +1

    Great tutorial! Nice and easy flow of code! As a beginner programmer, I really enjoyed this video! Thank you a lot!

  • @lokeswarreddyvalluru5918
    @lokeswarreddyvalluru5918 2 ปีที่แล้ว

    This man is from another planet .....

  • @mrremy8
    @mrremy8 2 ปีที่แล้ว

    Dude, thanks so much. You deserve much more views and likes. I didn't understand scraping one bit before this.

  • @rob5820
    @rob5820 2 ปีที่แล้ว +2

    Cheers! I'd love an updated version of this. It seems They've changed it. I have a project due soon for which I'd like to scrape Indeed as the project is a job search app.

    • @JohnWatsonRooney
      @JohnWatsonRooney  2 ปีที่แล้ว +2

      Thanks, I did a new version not that long ago the code is on my GitHub (jhnwr)

    • @rob5820
      @rob5820 2 ปีที่แล้ว +1

      @@JohnWatsonRooney Unreal. Thanks for the quick reply too.

  • @aliazadi9509
    @aliazadi9509 3 ปีที่แล้ว +1

    I just did webscraping on this website and youtube recomanded this video!🤣

  • @MyFukinBass
    @MyFukinBass 2 ปีที่แล้ว +1

    Damn this was top quality my man, thank you!

  • @vijayaraghavankraman
    @vijayaraghavankraman 4 ปีที่แล้ว +1

    Sir I became a great fan of u. Really interesting. A great skill to explain things in a better way to understand. Thanks a lot

  • @kmgmunges
    @kmgmunges 3 ปีที่แล้ว +1

    keep up the good work those lines of code and the logic is sure fire.

  • @OBPagan
    @OBPagan 3 ปีที่แล้ว +2

    You sir are a true legend. This taught me so much! I really appreciate it!

  • @hassanabdelalim
    @hassanabdelalim ปีที่แล้ว +3

    Hi great , but follow the same steps but i get 403 response not 200 , any help

  • @Eckister
    @Eckister ปีที่แล้ว +2

    your video has helped me a lot, thank you!

  • @lebudosh2275
    @lebudosh2275 3 ปีที่แล้ว +1

    Hello John,
    Thank you for the good work.
    It would be nice to be able to be see how the job descriptions can be added to the data collected from the webpage as well.

  • @benatakaan613
    @benatakaan613 ปีที่แล้ว +1

    Amazing content and teaching style! Thank you.

  • @anayajutt335
    @anayajutt335 2 ปีที่แล้ว +1

    Ima download it thanks for sharing!!

  • @caiomenudo
    @caiomenudo 4 ปีที่แล้ว +2

    dude, you're awesome. Thank you for this.
    Nice guitars btw

  • @CodePhiles
    @CodePhiles 3 ปีที่แล้ว +3

    Good Job, but in the loop you forgot to add "i" in extract function , the data were replication of first page, thanks a lot, plus more option to make location and job titile as parameter as well

    • @nathantyrell4898
      @nathantyrell4898 3 ปีที่แล้ว

      can you explain where to add the i in the extrct function? im dealing with this very problem right now

    • @CodePhiles
      @CodePhiles 3 ปีที่แล้ว

      @@nathantyrell4898 see at time 18:43 in line#35 just make it ..... c = extract(i) instead of c = extract(0)

    • @therealwatcher
      @therealwatcher 2 ปีที่แล้ว

      Do you know how I could extract the full job description ? since the url changes based on the selected job.

  • @gabrielalabi4385
    @gabrielalabi4385 3 ปีที่แล้ว +1

    Thank a lot, really helpful.
    I'll love to see how to automate applying to them 🤔🤔🤔

  • @SamiKhan-fd8gn
    @SamiKhan-fd8gn 2 ปีที่แล้ว +2

    Hello John, great video but unfortunately I keep getting 403 from indeed instead of 200 so not working for me.

  • @rajuchegoni108
    @rajuchegoni108 ปีที่แล้ว +2

    Hi John, how did u customize the output path, i tried so many experiments but it did not work. can u help me with that?

  • @jonathanfriz4410
    @jonathanfriz4410 4 ปีที่แล้ว +1

    Like usual very helpful John. Than you!

  • @ajinkyapehekar8985
    @ajinkyapehekar8985 ปีที่แล้ว +2

    I hope this message finds you well. I wanted to reach out and let you know that I've been trying to interact with your video, but I keep receiving a 403 response instead of the expected 200 response.
    I have checked my code, and it seems that I am setting the User-Agent header correctly to mimic a browser request. However, despite these efforts, I am still encountering the 403 error. I wanted to ask if there's anything specific I should be aware of or if there are any additional steps I need to take to ensure proper access to your video.
    I appreciate your time and any guidance you can provide to help me resolve this issue. Thank you for creating such valuable content, and I look forward to your response.

  • @nikoprabowo6551
    @nikoprabowo6551 2 ปีที่แล้ว +1

    I think its the best tutorial!!!! big thanks

  • @glennmacrae3831
    @glennmacrae3831 2 ปีที่แล้ว +2

    This is great, thanks!

  • @alexeyi451
    @alexeyi451 2 ปีที่แล้ว +1

    Great job, neat explanations! Thanks a lot!

  • @ritiksaxena7515
    @ritiksaxena7515 2 ปีที่แล้ว +1

    really thanks for this wonderful work

  • @ertanman
    @ertanman 2 ปีที่แล้ว +1

    GREAT VIDEO !!! Thank you very much

  • @tenminutetokyo2643
    @tenminutetokyo2643 2 ปีที่แล้ว +1

    That's nuts!

  • @sayyadsalman9132
    @sayyadsalman9132 4 ปีที่แล้ว +1

    Thanks for the video john! It was really helpful.

  • @kammelna
    @kammelna 2 ปีที่แล้ว

    Thanks John for your valuable efforts
    In my case I wanna scrape data inside each container where there is a table of info then loop over every link in the page
    So I need to click the link of first job for example and get data from a table and so on so forth for the rest of the page
    It would be highly appreciated if you could consider similar case in your next vids.
    Cheers

  • @anthonyb5625
    @anthonyb5625 2 ปีที่แล้ว +2

    Great tutorial thanks

  • @shayanhdry6224
    @shayanhdry6224 2 ปีที่แล้ว +1

    god of scraping

  • @jakepartridge6701
    @jakepartridge6701 2 ปีที่แล้ว +1

    This is brilliant, thank you!

  • @theprimecoder4981
    @theprimecoder4981 3 ปีที่แล้ว

    I really appreciate this video, you thought me a lot. Keep up the good work

  • @dewangbhavsar6025
    @dewangbhavsar6025 3 ปีที่แล้ว

    Great videos. Very helpful in learning scraping. Nicely done. Thanks!

  • @GudusSeb
    @GudusSeb 11 หลายเดือนก่อน +1

    Any idea how I can render/display the response data on a browser using HTML instead of saving it into CSV?
    Your Aswere is much appreciated. Thanks.

  • @thecodfather7109
    @thecodfather7109 4 ปีที่แล้ว +2

    Thank you 🙏🏼

  • @visualdad9453
    @visualdad9453 2 ปีที่แล้ว

    Great tutorial! thank you John

  • @raji_creation155
    @raji_creation155 ปีที่แล้ว +1

    Hi John, I want to know how to solve 403 error in scrapy.If u know please give explanation.

  • @irfankalam509
    @irfankalam509 4 ปีที่แล้ว +1

    Nice and very informative video. keep going!

  • @GreatestOfAllTimes0
    @GreatestOfAllTimes0 3 หลายเดือนก่อน

    is there a way to get the emails of the company?

  • @saifali4107
    @saifali4107 2 ปีที่แล้ว

    Hi John,
    Thanks for this wonderful video. I am following the steps but struggling with getting Company reviews the same way. Can not seem to find the right div class. Could you please help there.

  • @jfk-rm9sn
    @jfk-rm9sn ปีที่แล้ว +1

    Hi John, Tks for a great video! I am studying Python with your video, yet it keep ending up with 403 message. Any plans to update the tutorial?Thank you:)

  • @samiulhuda
    @samiulhuda 2 ปีที่แล้ว +1

    Can't get the 200, tried lots of mimic headers, cookies. But no results. Any advice?

  • @sanketnawale1938
    @sanketnawale1938 3 ปีที่แล้ว +1

    Thanks! It was really helpful.

  • @gihonglee6167
    @gihonglee6167 2 ปีที่แล้ว +2

    I followed your guide and edit a few lines of code so that I can scrap the whole job description.
    It worked well, but after 15 pages or so, I faced a captcha page and was unable to scrap anymore.
    I watched your user-agent video and changed the user-agent, still no luck.
    Is there any way I can scrap again?

    • @therealwatcher
      @therealwatcher 2 ปีที่แล้ว

      how were you able to get the full job description? Doesn't the url changes for each selected job id

  • @eligr8523
    @eligr8523 ปีที่แล้ว +1

    Hi. How can I scrape multiple pages? Can I just define another function to scrape another page? Ideally I would like to add all the information to one database using sqlite.

  • @ramkumarrs1170
    @ramkumarrs1170 3 ปีที่แล้ว +1

    Awesome tutorial!

  • @JulianFerguson
    @JulianFerguson 4 ปีที่แล้ว +2

    I know you mentioned using a while loop to run through more pages. Could you give an example of how this might look like?

  • @loganpaul8699
    @loganpaul8699 3 ปีที่แล้ว +1

    Such a great video!

  • @Free.Education786
    @Free.Education786 3 ปีที่แล้ว +1

    How to GRAB job listening email addresses to e-mail CV in BULK at ONCE???. Thanks

  • @nguyettran6118
    @nguyettran6118 ปีที่แล้ว +2

    hi,
    why my object divs = soup.find_all("div", class_ = "jobsearch...") returns empty list?

  • @truptymaipadhy7387
    @truptymaipadhy7387 8 หลายเดือนก่อน +1

    After coding with the same code it's showing me 403 error can any one help me

  • @dominicmuturi5369
    @dominicmuturi5369 3 ปีที่แล้ว

    great content hopefully more videos to come

  • @AtifShafiinheritance
    @AtifShafiinheritance 3 ปีที่แล้ว +1

    really good for lead generation ty

  • @alibaba2746
    @alibaba2746 3 ปีที่แล้ว

    Can u please teach us how to Automate or Scrape Facebook too. Thank u again bro for ur valuable teachings. GBU

  • @daniel76900
    @daniel76900 3 ปีที่แล้ว +1

    really, really good content!!

  • @guillaumejames4222
    @guillaumejames4222 2 ปีที่แล้ว +1

    Great coding structure and explanations. However, the website underlying CSS structure has change and as a results the codes no longer works. Is there a work around?

    • @JohnWatsonRooney
      @JohnWatsonRooney  2 ปีที่แล้ว +2

      Thanks. I haven’t revisited this one yet, I’m sure there is a way I will look into doing an update!

    • @guillaumejames4222
      @guillaumejames4222 2 ปีที่แล้ว

      @@JohnWatsonRooney Many thanks for you kind reply. Indeed, it would very practical to see the code adaptation to the new CSS structure.

  • @arsalraza3997
    @arsalraza3997 2 ปีที่แล้ว

    GREAT. Can you tell me how to go inside these jobs urls? how to get jobs urls!?

  • @yazanrizeq7537
    @yazanrizeq7537 3 ปีที่แล้ว +1

    You are awesome!!! Def Subscribing

  • @looijiahao2359
    @looijiahao2359 2 ปีที่แล้ว

    hi John , great tutorial , how would you add the time function in this particular set of code .

  • @oximine
    @oximine 2 ปีที่แล้ว +2

    Hi John! Great Video, however could you please update or make a new video to scrape indeed in present day? The website's html is very different now and the same code doesn't work.
    Would really appreciate it!

    • @JohnWatsonRooney
      @JohnWatsonRooney  2 ปีที่แล้ว +1

      Sure, I’ve actually rewritten it recently I could put out a helper video soon

    • @oximine
      @oximine 2 ปีที่แล้ว +2

      Appreciate you responding! I have also been getting a 403 status code despite trying out multiple User Agents. Being a python noob i could really use that helper video! Ty!

    • @JohnWatsonRooney
      @JohnWatsonRooney  2 ปีที่แล้ว +1

      @@oximine OK cool, I'll see if I can put it in for next week

    • @eleojoadegbe
      @eleojoadegbe 2 ปีที่แล้ว

      Thanks Oxamine for this question

  • @jaysonp9426
    @jaysonp9426 ปีที่แล้ว +1

    Great video minus the single letter variables

  • @ashu60071
    @ashu60071 4 ปีที่แล้ว +1

    Thanks 🙏🏻 you so so much. Actually can’t thank you enough.

  • @ansarisaami5196
    @ansarisaami5196 3 ปีที่แล้ว +1

    its so helpful brother

  • @Hgrewssauujdkhvcjjipp
    @Hgrewssauujdkhvcjjipp 2 ปีที่แล้ว +1

    Cool 👍

  • @kamaleshpramanik7645
    @kamaleshpramanik7645 3 ปีที่แล้ว +1

    Thank you very much Sir ...

  • @absoluteRandom69
    @absoluteRandom69 3 ปีที่แล้ว +1

    Hello John, I'm not able to crawl the website because of captcha. Who should I handle it?

  • @julianangelsotelo4757
    @julianangelsotelo4757 2 ปีที่แล้ว +1

    I got a 403 on my status code, does anyone know any potential solutions?
    Thanks!

  • @peterh7842
    @peterh7842 2 ปีที่แล้ว +1

    I am new to this so wondered if you can point me somewhere that shows how to set things up before you start typing i.e. this seems to be visual studio but how do you set that up for python? - I am stuck before even starting! : )😞

    • @JohnWatsonRooney
      @JohnWatsonRooney  2 ปีที่แล้ว +2

      Sure no worries I have a video on my channel for setting up Python, VS code and everything else you need to get to this point, I’m sure it will help you if you look for it on my channel page!

    • @peterh7842
      @peterh7842 2 ปีที่แล้ว

      @@JohnWatsonRooney Thanks John - very kind 🙂 I will have a look again tonight

  • @hosseinmohit
    @hosseinmohit ปีที่แล้ว

    Hi John
    I appreciate your great tutorial.
    I have a quick question. I know this video is from two years ago. I ran the first part today, and I got the 403 code. The Indeed website blocks me from getting data. Would you happen to have any suggestions for me to resolve my issue? A newer method? Your answer would help me a lot.

    • @ajewoledamilola7008
      @ajewoledamilola7008 ปีที่แล้ว +1

      I'm also tying the same thing and I am getting an error 403. Please help resolve this.

  • @Didanihaaaa
    @Didanihaaaa 3 ปีที่แล้ว +1

    very neat!

  • @ALANAMUL
    @ALANAMUL 4 ปีที่แล้ว

    Thanks for video....realy useful content

  • @raph6709
    @raph6709 2 ปีที่แล้ว +1

    Thanks

  • @benimustikoaji1393
    @benimustikoaji1393 3 ปีที่แล้ว +1

    It returns javascript not an html tag

  • @akhil2001
    @akhil2001 2 ปีที่แล้ว

    Is there any tips you can offer?

  • @Dev-zr8si
    @Dev-zr8si 3 ปีที่แล้ว +1

    This is amazing

  • @prashanthchandrasekar1026
    @prashanthchandrasekar1026 3 ปีที่แล้ว +1

    Thank u so much🙏

  • @lbayout2775
    @lbayout2775 3 ปีที่แล้ว +1

    perfect class

  • @whysotipsy
    @whysotipsy 2 ปีที่แล้ว +2

    Shouldn’t it be c = extract(i) instead of c = extract(0)
    Can someone please explain?

    • @JohnWatsonRooney
      @JohnWatsonRooney  2 ปีที่แล้ว +1

      Yes I think it should be - I made a mistake on this on I’m the for loop which meant the same result was repeated instead of getting the next

    • @whysotipsy
      @whysotipsy 2 ปีที่แล้ว

      @@JohnWatsonRooney Thank you! My friend pointed this out. Your videos are amazing! Keep rocking Buddy!

  • @therealhustle4629
    @therealhustle4629 2 ปีที่แล้ว +1

    Hello John, thank you very much for your Tutorial !
    I wanted to ask you if you know how its possible to get the Business owner & Phone number (from the website) on my Scraping list ?

    • @JohnWatsonRooney
      @JohnWatsonRooney  2 ปีที่แล้ว

      Hey thank you. Yes if the data is there on the site it should be possible using the same methods

    • @therealhustle4629
      @therealhustle4629 2 ปีที่แล้ว +1

      @@JohnWatsonRooney thank you very much for your response, in 95% of the leads there is no business owner in the article, is there a alternative ? Im doing cold calling in Germany

  • @FadeOutLetters
    @FadeOutLetters 3 ปีที่แล้ว

    Amazing video! Do you have a copy of the code you used in this video anywhere?

  • @necrodrucifver
    @necrodrucifver 4 หลายเดือนก่อน +1

    hello it still working on 2024? i have 0 results how about and update?

    • @JohnWatsonRooney
      @JohnWatsonRooney  4 หลายเดือนก่อน

      Afraid this doesn’t work anymore

  • @hanman5195
    @hanman5195 4 ปีที่แล้ว +3

    @john watson Rooney - Thnx for nice Video :) .
    Please share code as well on github :)

  • @dkaleka
    @dkaleka 3 ปีที่แล้ว +1

    Hi John, nice job. Do you work as a freelancer? How can I contact you?

    • @JohnWatsonRooney
      @JohnWatsonRooney  3 ปีที่แล้ว

      my email is on my main youtube page if you want to reach out

  • @pavelerokhin1512
    @pavelerokhin1512 ปีที่แล้ว +1

    cool!