I have a doubt. If the question doesn't needs any chat history the same process will repeat right? So its actually wasting tokens? Is there any ways to solve this??. And also how to manage session in real-life like in production there may be several users right so how to handle that situation? Pls reply
What I found problematic here is the usage of continuous chat history, in such a way as you said, we're wasting tokens, there was `ConversationSummaryMemory` in the LangChain framework to provide a summary of the chat history but sadly LangChain has discontinued support for that and advises users to migrate to LangGraph persistence which is a little complicated. Secondly, I don't think there's any way to decide beforehand if a user query requires the chat history context, so, that's why it's more robust to at least provide a summary of the chat history as an additional context for more accurate responses.
A session ID in generative AI typically refers to a unique identifier assigned to a user session. It tracks the interaction context between the user and the AI model, allowing the model to maintain continuity and reference previous inputs, ensuring coherent responses throughout the session.
Thank you very very much champ🏆🔥
Thankyou 😊
Thats very great session, I have been searching for the best playlist for quite some time, and I guess this playlist will prove to be very helpful💓
Glad you liked it
Thank you so much for putting in so much effort into these please keep them coming💗💗
sure why not thankyou
Great session keep going ❤❤❤
Thank you so much Sir, It was a long awaited one for me 😇😇
Most welcome 😊
Great session Sunny 🎉🎉🎉
BEST EXPLAINATION!
Thank you so much sir
I am looking for this only 🙏🙏🙏
Most welcome
This is done with few shot learning prompt right ?
i did not get you
How can we do the same thing with agent executor class? And how to check whether the context of the rag is passed or not
soon i will create a video ontop of agentic RAG with memory then you will get a idea
As a beginner how to start langchain form your channel?
start with langchain crash course
Sir please make a Deep learning playlist
soon will uploaded
I have a doubt. If the question doesn't needs any chat history the same process will repeat right? So its actually wasting tokens? Is there any ways to solve this??. And also how to manage session in real-life like in production there may be several users right so how to handle that situation? Pls reply
What I found problematic here is the usage of continuous chat history, in such a way as you said, we're wasting tokens, there was `ConversationSummaryMemory` in the LangChain framework to provide a summary of the chat history but sadly LangChain has discontinued support for that and advises users to migrate to LangGraph persistence which is a little complicated. Secondly, I don't think there's any way to decide beforehand if a user query requires the chat history context, so, that's why it's more robust to at least provide a summary of the chat history as an additional context for more accurate responses.
@@sougaaat true langraph documentation itself is very complicated. But one thing that i done is message trimming to reduce token usage.
@@anandukc4709 Indeed. It's really irritating how fast they introduce breaking changes.
Which LLM should I use sir to run locally ?
watch my ollama video it is having plenty of options
this is great,but what is session_id still a mystery for me.can you shed some light? some answer and link.
A session ID in generative AI typically refers to a unique identifier assigned to a user session. It tracks the interaction context between the user and the AI model, allowing the model to maintain continuity and reference previous inputs, ensuring coherent responses throughout the session.
@@tomtom-it4co thanks for sharing.bdw how can we track that?
@@tomtom-it4co where can we get that from?
it could be any random value
@@sunnysavita10 okh got it ,bdw how to track for each user when there is 1000s of user??