Hello Sir, while executing the partition_pdf() function I'm gettint this error "OSError: No such file or directory: '/root/nltk_data/tokenizers/punkt/PY3_tab'"
I was also getting this error, but it is resolved now. First, upgrade the NLTK library: !𝐩𝐢𝐩 𝐢𝐧𝐬𝐭𝐚𝐥𝐥 --𝐮𝐩𝐠𝐫𝐚𝐝𝐞 𝐧𝐥𝐭𝐤 and make sure nltk version should be 3.9.1, to check version run this: 𝐩𝐫𝐢𝐧𝐭(𝐧𝐥𝐭𝐤.__𝐯𝐞𝐫𝐬𝐢𝐨𝐧__)
thanks for your great video. I just found an error from NLTK while running your code on colab. ntlk 3.8 on colab would return error. please update to 3.9.
Great video! Thank you! Have you checked to parse multiple pdf files simultaneously and not parse one by one? Actually, i noticed that when i parse multiple pdf files, some images are ommited. Have you noticed something similar? Thank you again!
please mention in title or description whether if the tutorial uses any paid feature. would help save time as we can select apt videos instead of watching the video to know that. thanks
the way you explain in depth and clear the concept beyond just the running method or create an app is just something very unique. highly appreciated.
thanks
Wonderful.. Great Video.. Waiting for the upcoming videos..
Very soon
Thank you so much for this great explanation sir
Great video sir thank you sir was very helpfull in solving problem
It’s Magical 🧙
Great Video sir Thank you
Hello Sir,
while executing the partition_pdf() function I'm gettint this error "OSError: No such file or directory: '/root/nltk_data/tokenizers/punkt/PY3_tab'"
I was also getting this error, but it is resolved now.
First, upgrade the NLTK library:
!𝐩𝐢𝐩 𝐢𝐧𝐬𝐭𝐚𝐥𝐥 --𝐮𝐩𝐠𝐫𝐚𝐝𝐞 𝐧𝐥𝐭𝐤
and make sure nltk version should be 3.9.1, to check version run this:
𝐩𝐫𝐢𝐧𝐭(𝐧𝐥𝐭𝐤.__𝐯𝐞𝐫𝐬𝐢𝐨𝐧__)
Awesome video....waiting for more
Coming soon
Great explanation sir❤❤
thanks for your great video. I just found an error from NLTK while running your code on colab.
ntlk 3.8 on colab would return error. please update to 3.9.
thanks
Can I use unstructured library to extract video data? If no can you please do the video for this. Thanks.
Very Very Good Thanks a Lot !
You are welcome!
Great video! Thank you! Have you checked to parse multiple pdf files simultaneously and not parse one by one? Actually, i noticed that when i parse multiple pdf files, some images are ommited. Have you noticed something similar? Thank you again!
my dataset was small it might be the issue you can solve it using simple python logic
please mention in title or description whether if the tutorial uses any paid feature. would help save time as we can select apt videos instead of watching the video to know that. thanks
Wonderful! when is next video release?
soon...
Sir can you tell us where you use the termtorial or where you got this code are getting this code please please please sir its so necessary
Love it.
If executing in windows environment, what shall be the set of libraries. (As in the video it is executed in ubuntu)
using pip or anaconda you can download the package or use executable file
Can we Extract Data from chart or graph image ?? please answer.......if possible, please tell me way >>>>
Did you figure out ?
Sir please use ollama / langchain for rag, I don’t have open ai paid!
sure
classy of its own...
Sir Please try to use opensource LLMs and models . We don't have paid ones.
noted
thanks for the video...the code link is no longer there...
check with the updated github link
Sir what if the pdf cotains bullet points, I checked that the bullet points are not being detected. What to do sir?
you can write manual code to provide bullet point to your data
please add versions of the libraries used in the notebook. lot of time many errors coming because of version mismatch, Thanks
noted
the github link for Notebook is giving 404 error
Check with my GitHub link
how to get cj pdf?
how to load muliple pdfs and other type of files
checkout my rag playlist
404 - page not found notebook link not working
check with my github open the mention github link in the description
@@sunnysavita10 I checked it again, I am not able to attach the screenshot. You are requested to look into it.
where is the pdf /content/data/cj.pdf
github.com/sunnysavita10/Indepth-GENAI/tree/main/cj