Web Scraping to CSV | Multiple Pages Scraping with BeautifulSoup

แชร์
ฝัง
  • เผยแพร่เมื่อ 31 ธ.ค. 2024

ความคิดเห็น •

  • @skye0613
    @skye0613 10 หลายเดือนก่อน +14

    Finally, a video i can understand and doesn't make me feel dumb.
    Thank you good sir!

    • @FIBONACCIVEGA
      @FIBONACCIVEGA 6 หลายเดือนก่อน +1

      same for me.

  • @abdulwali4920
    @abdulwali4920 2 ปีที่แล้ว +18

    literally the best video on webscrapping....i have watched hundreds of videos but this is the best.

    • @Pythonology
      @Pythonology  2 ปีที่แล้ว +1

      Thank you very much Abdul Wali for your nice words. Very encouraging :)

  • @mohammedtofa431
    @mohammedtofa431 8 หลายเดือนก่อน +3

    I have been searching for a video like this forever. Thank God I landed on your page. Really wonderful and amazing video showing step by step. You are a living legend. Just subscribed as well.

  • @satishm5260
    @satishm5260 ปีที่แล้ว +2

    This is the most on point tutorial I ever watched. No bullshit, no jargon, Just pure knowledge. Thank you Sir, I learnt a lot from this small video.

  • @appleestrada6810
    @appleestrada6810 11 หลายเดือนก่อน +1

    This is by far THE best and easiest to understand explanation I’ve heard about using python to scrape data. Thank you for your effort in creating this video. You got a new subscriber!

  • @knoxholst67901
    @knoxholst67901 2 ปีที่แล้ว +12

    I just wanted to let you know that I really enjoyed this video. I was feeling like learning python was stupid. Then I found you doing a cool project and it was easy to follow. I am inspired again thank you.

  • @homeoffice3524
    @homeoffice3524 ปีที่แล้ว +3

    Finally! Really clean and easy to follow scraping video.

  • @SandeepGurjar-ko5ju
    @SandeepGurjar-ko5ju ปีที่แล้ว +2

    Your work needs to be appreciated man. The way you explain things in a calm, composed and soothing voice. The simplicity of the tutorial indicates your grasp on the web scraping. Thank you.

    • @Pythonology
      @Pythonology  ปีที่แล้ว

      Appreciate it, Sandeep.

  • @zedascouve2
    @zedascouve2 ปีที่แล้ว +1

    Absolutely jaw-dropping the power of web scraping. Congrats for the wonderful and comprehensive video. Waiting for more!!!

  • @marcoschwarz4640
    @marcoschwarz4640 ปีที่แล้ว +4

    I watch a lot of videos about programming and most of them are really good. However, this really is a standout piece. The way it combines theory and practice is second none. Well done,sir

  • @mattbaker6475
    @mattbaker6475 10 หลายเดือนก่อน +1

    Great video!
    You're not getting enough credit for how well this is made.

  • @samuelebojielu855
    @samuelebojielu855 ปีที่แล้ว +2

    Thank you so much for this video. I have watched several web scraping videos but this is absolutely the best so far.

  • @jungjaehyun8625
    @jungjaehyun8625 5 หลายเดือนก่อน +1

    Multiple pages start at 21:20

  • @wanangwalumbaningwira3871
    @wanangwalumbaningwira3871 ปีที่แล้ว +1

    ooooooooooooooooooooohhhh I really love this video you saved me big time. This is really outstanding, well detailed and you explainations are very logical and clear

  • @muhammadmubeen8452
    @muhammadmubeen8452 ปีที่แล้ว +1

    Thank you very much sir I was watching many tutorials and was getting confused to understand the html structure then I found your video you explained everything beautifully I completed my project successfully thanks a lot sir❤

  • @randomperson-gp2wd
    @randomperson-gp2wd ปีที่แล้ว

    Thanks a lot. With some basic level in python and 0 background in programming, I was able to successfully do a project for my master thesis related to media coverage about certain topic.

  • @SkillsGear
    @SkillsGear 8 หลายเดือนก่อน

    Finally, a video that puts paid courses to shame! Hats off to you for the great tutorial! You just did not explain, the way you went back and forth helped me understand a lot. Kudos! Could not resist the urge to hit the like and subscribe button. Will definitely visit your channel for more guides and tutorials! ♥

  • @lesotv1056
    @lesotv1056 ปีที่แล้ว

    OMG, I am so impress thank you so much for this wonderful lesson. I cant believe I got this for free . God bless you.

  • @nox6687
    @nox6687 3 หลายเดือนก่อน

    Wish I could like this twice, I had a web scraping class that didn't explain this as well as you did in half an hour

  • @Mr.Rahul-s-18
    @Mr.Rahul-s-18 11 หลายเดือนก่อน

    Such a GreaT Explanation Dear
    JusT Love it😘
    Love From India 🇮🇳NamsTe🙏

  • @hirrog3865
    @hirrog3865 6 หลายเดือนก่อน

    After so much searching I finally get a video that is so easy to grasp on scrapping from multiple pages. Thank you

  • @extropiantranshuman
    @extropiantranshuman ปีที่แล้ว +1

    wow - best tutorial so far on beautifulsoup! Thank you!

  • @sundy93
    @sundy93 9 หลายเดือนก่อน

    This video is such a relief ,absolutely the best material about scraping ! Thank you so much!

  • @ibrahimaboelsoud7881
    @ibrahimaboelsoud7881 ปีที่แล้ว +1

    24:53 what a vim move 😄.

  • @Raulenz
    @Raulenz ปีที่แล้ว

    Thanks, great video. Excellent explanation and great english.

  • @catchiemitakata2142
    @catchiemitakata2142 8 หลายเดือนก่อน

    The best tut on web scraping. Very beginner friendly. Keep it up

  • @KamleshShirsath-n7b
    @KamleshShirsath-n7b ปีที่แล้ว +1

    Thanks a lot for this detailed video. Hoping to see more video's like this.

  • @everyany4161
    @everyany4161 ปีที่แล้ว +1

    Thank you so much for this video! It's literally an answered prayer for me. 🙏

  • @AbsSak
    @AbsSak ปีที่แล้ว

    I just loved it . I used to think web scraping is too hard but when I saw your video it is like so simple that even 10 years old can also understand. Simply great job 👏

  • @raedkm
    @raedkm ปีที่แล้ว

    Thanknyou for this clear and easy to follow video.

  • @jeinnerabdel
    @jeinnerabdel ปีที่แล้ว +3

    Great video! I wrote the code while you were explaining it and I kinda grasped the idea behind what you were doing. The only thing I don't understand is about the indentation and how it affects the for structure. In other languages, you end the for with some code and nest them like any while-do or if-else-endif type of stuff. I also thought that Python was like Javascript where data would automatically be translated on each variable based on it's content Var1 = Here you go (text) or Var1 = 12 (num) but as I saw on your example, you have to transform data into numbers even if they are actually numbers already. Interesting!

  • @pannusingchauhan6883
    @pannusingchauhan6883 ปีที่แล้ว

    this is just crazyyy. loved the tutorial

  • @nairchandnichandran7503
    @nairchandnichandran7503 ปีที่แล้ว +1

    Wonderful!. Simple and concise🥰

  • @mohamadehtheshams2527
    @mohamadehtheshams2527 ปีที่แล้ว +1

    Great Sir Today I learned how to do web scrapping .. Nicely Explained 👍. Please make more content

  • @rambhupal8870
    @rambhupal8870 2 ปีที่แล้ว +2

    Its just really awesome and very easy to understand also and i have submitted this as a mini project. Thank You brother.

  • @dawnS33ker
    @dawnS33ker ปีที่แล้ว

    This video is an absolute gem. Thank you for this..

  • @MKOladapo
    @MKOladapo ปีที่แล้ว

    Great lesson...Very resourceful

  • @glowtechvisuals6176
    @glowtechvisuals6176 5 หลายเดือนก่อน

    I like the way you teach while talking. make me understand. Tank you very much

  • @BenjaminBalaj-yp8zk
    @BenjaminBalaj-yp8zk ปีที่แล้ว

    You explain everything very clearly. Everything makes sense now!

  • @danielomoregie7603
    @danielomoregie7603 ปีที่แล้ว +1

    This is the best web scrapping video on the internet

  • @dimi_ga
    @dimi_ga ปีที่แล้ว

    Thank you very much, you helped me a lot with your vid. 🙏

  • @ephraimmotho887
    @ephraimmotho887 2 ปีที่แล้ว +2

    Great tutorial yet again... This channel is so valuable for people who want to learn programming but do not have the money to go to school for it... Are there any other similar channels on youtube or outside the platform (websites etc.) that offer such great value but may not be popular? Please reply if you even have one suggestion. It is really helpful

  • @joelngige5776
    @joelngige5776 ปีที่แล้ว

    Fantastic video on web scraping

  • @David-c3h3o
    @David-c3h3o ปีที่แล้ว +1

    You really did make a great video.Thank you

  • @janp3648
    @janp3648 ปีที่แล้ว

    Omg .. this is such a perfect, informative, easy to understand explanation ! Thx a lot.

  • @debasishrajbehera
    @debasishrajbehera 2 ปีที่แล้ว +2

    thank you so much sir....................i learned a lot .....its so helpful to me..🙏

  • @wawy0
    @wawy0 หลายเดือนก่อน

    i love your way to get into my brain you are awesome how easily explain with no feed dumb or crazy this tut is the best tut ever is best then the tut i been put my money on it thank you very much friend :)

  • @informativecontent4778
    @informativecontent4778 9 หลายเดือนก่อน

    paginaton starts from 21:30

  • @osoriomatucurane9511
    @osoriomatucurane9511 ปีที่แล้ว

    So awesome! Consice & crystal clear! You are absolutely a legend.❤

  • @AsmaaMGaber
    @AsmaaMGaber 11 หลายเดือนก่อน

    very informative video, thank you for your efforts.
    I use jupyter notebook and I wrote the exact code, yet it doesn't scrap all pages and it scrap only the last number in range. do you have any idea what could cause this error?

  • @ghashanti
    @ghashanti ปีที่แล้ว +2

    This was great content. You made web scraping super easy.

  • @Birwi
    @Birwi 11 หลายเดือนก่อน

    Really helpful, thank you!

  • @netacklee
    @netacklee ปีที่แล้ว

    Wow, this video is so helpful, thank you!

  • @bashirahsalami3408
    @bashirahsalami3408 ปีที่แล้ว +2

    Thank you very much Pythonology. This was well-explained and very easy to understand.

  • @AdamDoe-br5tw
    @AdamDoe-br5tw ปีที่แล้ว

    Great tutorial Thanks, now what if the pages have different/variable names like site/brand/VariableBrandname and I only have a list of the pages
    How to set the "i" variable to look into a set of "variablebrandname"?

  • @richardharris-to3mc
    @richardharris-to3mc ปีที่แล้ว +1

    top notch. I managed to follow this, so thankyou!

  • @rodolfog.6826
    @rodolfog.6826 ปีที่แล้ว +1

    What happens if i get a 403 response I think is Forbidden Access ??

    • @Pythonology
      @Pythonology  ปีที่แล้ว

      Reasons for 403:
      1- The URL you are trying to scrape is forbidden, and you need to be authorized to access it.
      2- The website detects that you are scraper and returns a 403 Forbidden HTTP Status Code as a ban page (the website could be protected by Cloudflare for example).

  • @pasindumadushan474
    @pasindumadushan474 ปีที่แล้ว +1

    Thank You So much SIr!

  • @NikhilPawar151
    @NikhilPawar151 ปีที่แล้ว +1

    Great video. I really like the way you explain the concepts. Everything working fine and easy to understand

  • @svenGPT
    @svenGPT ปีที่แล้ว

    Just found it and love it. Thank you!

  • @payamsarwar5611
    @payamsarwar5611 ปีที่แล้ว +1

    Very productive, thank you

  • @mavisadjei4192
    @mavisadjei4192 ปีที่แล้ว

    Thank you very much. Is there a good book you can recommend?

  • @izainonline
    @izainonline 6 หลายเดือนก่อน

    Great Tutorial can we scrap the secure or can say not allowed scrapping text.

  • @ItIsWell2222
    @ItIsWell2222 ปีที่แล้ว

    and i want to access the text in span tag
    and this span tag is within a 1i tag
    pls how can i go about it
    because I tried using spag tag it not giving right text

  • @AmanKumar-xw5bs
    @AmanKumar-xw5bs ปีที่แล้ว

    wonderful sir ! Learnt a lot

  • @aspirant9511
    @aspirant9511 ปีที่แล้ว +1

    Very well explained ...thank u..

  • @ItIsWell2222
    @ItIsWell2222 ปีที่แล้ว

    pls in a situation whereby i have multiple p tag and I want the text of the second p tag and no class or attrs to differentiate it
    pls how can i go about it

  • @georgewilliams9422
    @georgewilliams9422 4 หลายเดือนก่อน

    Thank you! Very clear and useful!

  • @annicesullivan
    @annicesullivan 8 หลายเดือนก่อน

    Nice tutorial on scraping multiple pages to CSV with BeautifulSoup! Any tips on reliable proxies for handling large scraping jobs like this? Heard Proxy-Store offers specialized scraping packages, anyone tried them out?

  • @sabrinaazwa759
    @sabrinaazwa759 7 หลายเดือนก่อน +1

    thanks alot for the detailed tutorial!!!

  • @newperspective9
    @newperspective9 ปีที่แล้ว

    I saw what I need to saw, Thank you!!!

  • @czr372
    @czr372 2 ปีที่แล้ว

    Thanks a lot for all this web scrapping tutorials! Ill try to do my own scraps now!

  • @Fisayoxx.x
    @Fisayoxx.x ปีที่แล้ว

    Thank you so much for this. Thank you

  • @tundebadmus7702
    @tundebadmus7702 ปีที่แล้ว

    You explain really well.. keep it up

  • @littleboyspace8940
    @littleboyspace8940 5 หลายเดือนก่อน +1

    title = article.find('img').attrs['alt']
    star = article.find('p')['class']
    Could you explain why we need .attrs for title when you can ditect access it like you did with the class tag?
    title = article.find('img')['alt']
    star = article.find('p')['class']
    I have try this and it work the same. Is there benefit to use attrs?
    Thank You

  • @saharyarmohammadtoosky
    @saharyarmohammadtoosky ปีที่แล้ว +1

    Thank you very much, very good and detailed explanation

  • @nadareham9784
    @nadareham9784 ปีที่แล้ว

    Hello I am struggling with something . Can you help me . I can't see ol but all started with section how to run soup to grab exact data .

  • @chiedozieehirim5659
    @chiedozieehirim5659 ปีที่แล้ว +1

    Thanks a lot. It helped solve a problem.
    I have a question though.
    How do u handle 403 and 503 status_codes errors when scrapping a website?

    • @Pythonology
      @Pythonology  ปีที่แล้ว

      403 and 503 status code errors, indicate that the server is refusing to fulfill the request. To handle these errors, you can use the requests library to make the request and check the status code.
      One way to handle these errors is to use try-except blocks to catch the error and handle it appropriately. For example, you could include a sleep function to wait a certain amount of time before trying again, or you could implement a retry loop to keep trying until the request is successful. Another approach is to use a library like requests-html which has a built-in support for handling these errors and retrying failed requests automatically. Also, you can use a User-Agent in the headers to make the request appear as if it was coming from a browser instead of a scraper, as some websites block requests from known scraper IPs and user-agents.

  • @lesteravecilla77
    @lesteravecilla77 ปีที่แล้ว

    Super Thanks for this video. It is very clear and useful for people who like start web scraping like me. Good job and keep it up! 👏🙂

  • @dreamer14492
    @dreamer14492 9 หลายเดือนก่อน

    you pulled the name from the image alternative tag sometimes image alternative can be anything instead pulling the title from the h3 tag and title attribute will be better in my opinion

  • @matteos9783
    @matteos9783 2 ปีที่แล้ว

    Thanks, but what if I want follow the subpage of every book and extract the informations in these pages? I mean before I extract the informations in the first page, then go into every subpage of the books and finally grab the pages informations

  • @fpldaily5172
    @fpldaily5172 2 ปีที่แล้ว

    how do i webscrape the page and the content in the page eg, your video extracts the title, price etc but lets say I also want to extract the book page and the content after the book the book page
    like some e-commerce sites show the products name, price etc but when I click the page it shows decscriptions, reviews , and more pictures of the product how do I extract that aswell?
    Thanks mahn I like your work!

  • @gur0004
    @gur0004 2 ปีที่แล้ว +1

    lovely stuff. I thoroughly enjoyed it.

  • @bkkaan52
    @bkkaan52 ปีที่แล้ว

    how can i save a png from a page?

  • @DSTtools-p4j
    @DSTtools-p4j 9 หลายเดือนก่อน

    This is very cool
    Keep.it up bro...

  • @evoplayz7136
    @evoplayz7136 ปีที่แล้ว

    Would you please make a video on how to scrape the data inside of each link

  • @mynshk
    @mynshk ปีที่แล้ว

    Many thanks for your demonstration! :D

  • @gadomix3989
    @gadomix3989 ปีที่แล้ว

    Thank you for a great video realy its cool project ever seen

  • @abrarbaltii
    @abrarbaltii ปีที่แล้ว

    How to store these data into a database like mongodb ...kindly make a video on it also it would be a great help.

  • @PP-pd3pf
    @PP-pd3pf ปีที่แล้ว

    Thank you for the great content.

  • @swatisingh-yw1fw
    @swatisingh-yw1fw ปีที่แล้ว

    it was very helpful video, keep on making such video.

  • @ryansmithza
    @ryansmithza ปีที่แล้ว

    Exactly what I was looking for! Thanks!

  • @jarnoojasaar
    @jarnoojasaar 2 ปีที่แล้ว +1

    What if I want the parser to click into every book and get some info from each book page?

    • @Pythonology
      @Pythonology  ปีที่แล้ว

      To handle clicks, it is better to use Selenium or Scrapy than BeautifulSoup.

  • @khaledhamdy9944
    @khaledhamdy9944 ปีที่แล้ว

    Why this NameError: " i is not defined" -- for i is the variable in the link

  • @mister_maxwell
    @mister_maxwell ปีที่แล้ว

    Very very good, I learned so much new and interesting stuff.

  • @Vish_27-v8x
    @Vish_27-v8x 10 หลายเดือนก่อน

    Could you please show how to web scrap target product reviews?

  • @AnulithaVardhanMarripeddiVijay
    @AnulithaVardhanMarripeddiVijay 9 หลายเดือนก่อน

    I have the prices in tags and soup.find ignores it all together. Any idea how to handle that?

  • @ReadyF0RHeady
    @ReadyF0RHeady หลายเดือนก่อน

    For my project i want to scrape two websites about different index values for various countries and put it into one database eg. freedom house index and index of economic freedom for germany and other countries...im not sure how merge this data into one database

  • @olubukola107
    @olubukola107 2 ปีที่แล้ว +1

    Thank you 🙏 so easy to understand and helpful

  • @ayoubcharbaji884
    @ayoubcharbaji884 9 หลายเดือนก่อน

    hi sir, well i have a question about the page numbers, if i'm working with for exemple three websites and i don't know how many pages they've got so what should i do to make my code scrape all the products ?