How to use read and write streams in node.js for BIG CSV files

แชร์
ฝัง
  • เผยแพร่เมื่อ 8 มี.ค. 2022
  • An overview on how to use a read and write stream in node when you want to read in a very large csv file and process it quickly.
    ------------
    🔔 Newsletter eepurl.com/hnderP
    💬 Discord / discord
    📁. GitHub github.com/codyseibert/youtube

ความคิดเห็น • 26

  • @sameerbharadwaj11
    @sameerbharadwaj11 ปีที่แล้ว +2

    Thank You! This helped a lot with writing 20mil records to solr.

    • @WebDevCody
      @WebDevCody  ปีที่แล้ว

      Oh wow this actually helped? I’m glad 😂

  • @officialabdulrehman
    @officialabdulrehman ปีที่แล้ว

    This was very helpful. Thank you

  • @bipinshukla5222
    @bipinshukla5222 ปีที่แล้ว

    Awesome explaination

  • @CarlosHonorioBriones
    @CarlosHonorioBriones 2 ปีที่แล้ว

    nice! This is good to know. Thanks!

  • @LarsRyeJeppesen
    @LarsRyeJeppesen 9 หลายเดือนก่อน

    Super stuff, thanks. Subbed

  • @rorycawley
    @rorycawley 2 ปีที่แล้ว

    Great video, thanks

  • @SeibertSwirl
    @SeibertSwirl 2 ปีที่แล้ว

    Good job babe!!!

  • @YOUdudex
    @YOUdudex ปีที่แล้ว

    Cool video, thanks subscribed

  • @user-xw6jr7ue4b
    @user-xw6jr7ue4b ปีที่แล้ว

    thank u, buddy!

  • @pranjaldoorwar9743
    @pranjaldoorwar9743 2 ปีที่แล้ว +3

    This might be helpful to me i am trying to read csv data and insert it into mongodb first before parsing it

  • @lucanthony2209
    @lucanthony2209 2 ปีที่แล้ว

    Hello @vJunkie,
    Thanks for the great content.
    CAn you make a video using csv-parser & fs libraries to read & write csv files ?

  • @PhuTran-nb7me
    @PhuTran-nb7me 2 ปีที่แล้ว +3

    Hello Sir, I have a question, much appreciated If you can address my doubt. Do chunks come in sequence, in other words, does the read stream wait for the current chunk to finish the operations before it emit the "data" event to process the next chunk? You have a potentially long running iteration there and more importantly, each chunk is interdependent to each other as the last If statement is used to assign the last incomplete row to "unprocessed" variable for the next chunk to use.

    • @WebDevCody
      @WebDevCody  2 ปีที่แล้ว +2

      Yes they should come in order.

    • @empty8537empty
      @empty8537empty 2 ปีที่แล้ว +2

      It should be in order..which is taken care by your transport layer..

  • @martinmingosuarez8690
    @martinmingosuarez8690 2 ปีที่แล้ว +15

    there is a core module called readline that works with streams to read one line at a time, no need to write the logic yourself, otherwise great explanation and video. Cheers

    • @veremox
      @veremox ปีที่แล้ว

      im not an expert but a chunk is not necessarily a line...

    • @oscarmamani5996
      @oscarmamani5996 ปีที่แล้ว

      martinmingosuarez8690 thank you for this comment! it help me

    • @LegendKopper
      @LegendKopper 7 หลายเดือนก่อน

      -----The topic "NodeJS streams---->
      (your head)

  • @mtuchidev
    @mtuchidev ปีที่แล้ว

    How do you pass a chunk size ?

  • @haohong5141
    @haohong5141 ปีที่แล้ว

    node version?

  • @boople2snoot
    @boople2snoot 2 ปีที่แล้ว

    them jump cuts

    • @boople2snoot
      @boople2snoot 2 ปีที่แล้ว

      also no sub, your variable is overWatermark, you're waiting for an event emitter if not overWatermark, the function therefore must return true if it is not over the watermark and var should be named underWatermark. Or it would run if not overWatermark, which is under the watermark, which? shit you just ramble, bye.

  • @himaliamit8698
    @himaliamit8698 2 ปีที่แล้ว

    if you show us the format of the csv file it would have made much more sense.

    • @majidshah48
      @majidshah48 2 ปีที่แล้ว

      i think opening the file will take time or it might just crash

  • @ViktosB
    @ViktosB 4 หลายเดือนก่อน

    Hi, I really enjoy your videos. One of the most productive publisher on TH-cam. But in that video the name "overWatermark" is highly misleading. I struggled understand the whole structure only because of that naming. This would be better if you use instead "canWrite". canWrite is true when we can write and false when the buffer is full. Then we stop -> !canWrite.
    const canWrite = writeStream.write(`${i},
    `);
    if (!canWrite) {
    await new Promise((resolve) => writeStream.once('drain', resolve));
    }
    }
    I might be wrong of course.
    Other than that - great job again. Continue watching you content 🙂