This is excellent and valuable knowledge sharing... Easily one can make out these trainings are coming out of personal deep hands-on experience and not the mere theory ..Great work
Earlier I watched some videos regarding this topic ,no one can explained in this way ,I am glad to see this video,now clearly understood spark optimization techniques
Awesome explanation of the optimisation techniques. If possible please create a video to cover the realtime challenges which you faced in your project and the solution you provided. That will be really helpful.
thanks a lot , i have case where someother modules write parquet file , i need to process in my module by reading it, so how should i apply bucketing on that day ...can it be possible without writing ???
@@BigDataThoughts thanks! Can you build an end to end project or some mini project where one can see how and where these properties arte getting implemented? Just watching these in silos only give half knowledge. Thanks.
how i join small table with big table but i want to fetch all the data in small table like the small table is 100k record and large table is 1 milion record df = smalldf.join(largedf, smalldf.id==largedf.id , how = 'left_outerjoin') it makes out of memory and i cant do broadcast the small df idont know why what is best case here pls help
This is excellent and valuable knowledge sharing... Easily one can make out these trainings are coming out of personal deep hands-on experience and not the mere theory ..Great work
thanks
Earlier I watched some videos regarding this topic ,no one can explained in this way ,I am glad to see this video,now clearly understood spark optimization techniques
Thanks
perfectly went into my brain, what a clean explanation.
can you please do videos on Pyspark from scratch.
It's very very easy to understand whatever you explained, thank you so much
Thanks namrata
what a wonderfull explanation to the point... thank you
Thanks
best video on internet for spark performace...
Thanks harsh
I only have one word for this video
"Awesome!!"
Thanks
Very Good vedio, awesome work. To the point and one can understand easily
Thanks sahil
Great expectations Mam ... eagerly waiting for your upcoming video
Thanks Rajni
Very good explanation keep create more videos in spark
Thanks
Neat and clean explanation. Looking forward to the videos on Spark Optimization.
Thanks Gunasekaran
Thanks for sharing. Very informative .
Thanks karan
Very informative 👏👍
Keep such videos coming
Many thanks for making such informative video
This is excellent maam. Looking forward to watching more videos from you
Thanks Chethan
Thanks a lot mam for making these videos. These are extremely useful. One of the best videos I have come across.
Thanks Ram
Another much needed video on Spark optimizations. point to point. Thank you very much for the video.
Thanks Kiran
Thanks for video ma'am., you made it very simple to understand.. waiting for more video's on this topic and spark
Thanks Arjun
th-cam.com/video/snPYj3TqM1g/w-d-xo.html you can chk this too. There are others videos on spark that i have posted
Nice video it was really good 👍🏻 Thank you
Thanks
Very well explained. Loving your videos!❤
Thanks
REALLY helped me get better at my work
Thanks
All yours tutorials are too good!
One of the best explanation on TH-cam 😊
Awesome explanation of the optimisation techniques. If possible please create a video to cover the realtime challenges which you faced in your project and the solution you provided. That will be really helpful.
Tnks Amigo
It's very helpful
thanks
Awesome Shreya, if possible could you please upload realtime challenges which we faced in realtime environment
Thanks Khader
So nicely explained Shreya..
Thanks Abhishek
Great explaination !!
Thanks Puneet
Great explanation
Thanks
Ultimate 👏👏👏
Thanks
Good explanation 😊
Thanks
Amazing Job Shreya... Keep it Up..
Thanks Ayush
Nice video mam
good optimization tips
Thanks Neethu
so nice it helped a lot
Thanks tanushree
Excellent.
If count is not adviced, how can we count the number of rows in data frame?
❤
thanks a lot , i have case where someother modules write parquet file , i need to process in my module by reading it, so how should i apply bucketing on that day ...can it be possible without writing ???
where do I get practical session on this optimization technique of spark..?
very well explained. but you are telling everything i know to exclude. ?? We need count, group,agg
Coalesce doesn't do Shuffle and that's why it's less expensive than repartition. I believe.
It does but not as much as repartition. Repartition does entire data shuffle as it can reduce or increase no of partitions.
@@BigDataThoughts thanks! Can you build an end to end project or some mini project where one can see how and where these properties arte getting implemented? Just watching these in silos only give half knowledge. Thanks.
What are configurations file in spark ?
Plz anyone can ans me please
Keep the good work #Prinetechs.
I can clearly see all the good reviews about you man…I never believed my account can fixed after 7 months hahaha
how i join small table with big table but i want to fetch all the data in small table like
the small table is 100k record and large table is 1 milion record
df = smalldf.join(largedf, smalldf.id==largedf.id , how = 'left_outerjoin')
it makes out of memory and i cant do broadcast the small df idont know why what is best case here pls help
Keep the good work #Prinetechs.
I can clearly see all the good reviews about you man…I never believed my account can fixed after 7 months hahaha
Keep the good work #Prinetechs.
I can clearly see all the good reviews about you man…I never believed my account can fixed after 7 months hahaha