ycopie
ycopie
  • 54
  • 82 652
How does OpenAI Tokenizer Work - Tiktoken Tutorial
Video explains how OpenAI's tokenizer divides text into smaller units called tokens. It covers how Tiktoken can be used for getting tokenization for OpenAI models. This is useful to count token and maybe make a decision in project whether to route to a different bigger model or even do some processing on text like chunking.
Code: github.com/oppasource/ycopie/blob/main/GenAI_OpenAI_LangChain/3.%20OpenAI_Tokenizer_Using_Tiktoken/OpenAI%20Tokenizer%20Using%20Tiktoken.ipynb
GenAI with OpenAI and LangChain playlist: th-cam.com/play/PL0gCgnksEIeG5BlAOpogwf8xgPVvrwo20.html
Linkedin: www.linkedin.com/in/yash-agrawal-a22597162/
มุมมอง: 15

วีดีโอ

LangChain: What, Why and OpenAI Integration
มุมมอง 114 ชั่วโมงที่ผ่านมา
In this video, we dive into LangChain, an innovative framework designed to streamline the integration of language models like OpenAI's into your applications. Learn what LangChain is, why it's a game-changer for developers, and how to utilize OpenAI through LangChain. Code: github.com/oppasource/ycopie/blob/main/GenAI_OpenAI_LangChain/2. OpenAI_Using_LangChain_Initial_Setup/OpenAI Using LangCha...
OpenAI API Explained: Python SDK, API Key Setup, Parameters and Pricing
มุมมอง 229 ชั่วโมงที่ผ่านมา
In this tutorial, learn how to integrate the OpenAI API with Python using the official SDK. We'll cover everything from setting up your API key to understanding input/output parameters, and we'll also dive into the pricing structure. Code: github.com/oppasource/ycopie/tree/main/GenAI_OpenAI_LangChain/1. OpenAI_API_Initial_Setup GenAI with OpenAI and LangChain playlist: th-cam.com/play/PL0gCgnks...
GenAI using OpenAI and LangChain | Series Intro and What is LLM based System
มุมมอง 4914 ชั่วโมงที่ผ่านมา
With this video we start a series on GenAI using OpenAI and LangChain. We’ll outline what to expect and dive into building practical AI applications like chatbots and agent-based systems. Along the way, we’ll also explore fine-tuning LLMs. We'll build applications with both the OpenAI Python SDK and LangChain, comparing the two approaches to better understand their differences. Excited to share...
RAG implementation using Llama-2 model
มุมมอง 1.9K9 หลายเดือนก่อน
In this video we discuss how the RAG pipeline works along with semantic search to answer questions from any document using LLM. We specifically use the all-mpnet-base-v2 as embedding model and ChromaDB as our vector database. We then use Llama-2-7b-chat model to finally generate the answer given question and retrieved context. Code: github.com/oppasource/ycopie/blob/main/LLM_Series/RAG_with_Lla...
Building Semantic Retriever with Vector Database for RAG | Discussion around Embeddings
มุมมอง 6999 หลายเดือนก่อน
In this video we see how Vector DB can be helpful in low latency retrieval for semantic search using Embeddings. Such retrieval is useful in frameworks like RAG for question answering. We also discuss a little about embeddings and how similairty search works by calculating the distances. How we can make descission on which embedding model we can use. Code: github.com/oppasource/ycopie/blob/main...
What is Retrieval Augmented Generation (RAG) and how is it useful
มุมมอง 6759 หลายเดือนก่อน
We explore the synergy of RAG (Retrieval-Augmented Generation) with LLM (Large Language Models) in this video, tackling outdated knowledge and context limitations. It is very helpful in doing QA or chatting with latest or private documents to be sure that LLMs does not hallucinate. LLM Series Complete Playlist: th-cam.com/play/PL0gCgnksEIeGgNUVUKOQN0pt34fQ05ESO.html Linkedin: www.linkedin.com/i...
Prompt Engineering using Llama-2 model
มุมมอง 5K10 หลายเดือนก่อน
In this video we see how we can engineer prompts to get desired responses from LLMs. We cover following prompting techniques: 1. Zero Shot Prompting 2. Few Shot Prompting 3. Chain-of-Thought Prompting (CoT) 4. Chat Prompting Code: github.com/oppasource/ycopie/blob/main/LLM_Series/Prompt_Engineering/prompt_engineering.ipynb Prompting Guide reference: www.promptingguide.ai Difference between Llam...
What is the difference between Llama-2 and Llama-2-Chat model?
มุมมอง 3.4K10 หลายเดือนก่อน
In this video, we compare the main differences between Llama-2 and its fine-tuned variant, Llama-2-Chat. The discussion extends to other language models like GPT and ChatGPT. We delve into the advantages of using the fine-tuned Llama-2-Chat model for question answering, highlighting the challenges faced by the base Llama-2. The video provides insights on when to use each model based on specific...
Step-by-step guide on how to setup and run Llama-2 model locally
มุมมอง 40K10 หลายเดือนก่อน
In this video we look at how to run Llama-2-7b model through hugginface and other nuances around it: 1. Getting Access to Llama Model via Meta and Hugging Face: Learn how to obtain access to the Llama language model through Meta and Hugging Face platforms. 2. Downloading and Running Llama-2-7b Locally: Follow step-by-step instructions on downloading the llama-2-7b model and running it on your l...
A basic introduction to LLM | Ideas behind ChatGPT
มุมมอง 1.1K10 หลายเดือนก่อน
This is a fascinating exploration into the world of Large Language Models (LLM) as we delve into the ideas behind ChatGPT. In this video, we give a basic introduction to LLMs, unravelling the intricate workings of these powerful language models. Whether you're a tech enthusiast or just curious about cutting-edge AI, this video offers insights into the fascinating realm of language models and th...
N-gram Language Modeling | Theory, Math, Code
มุมมอง 18K3 ปีที่แล้ว
Video introduces you to building a simple n-gram based language model. Link for code and slides: github.com/oppasource/ycopie/tree/main/N-gram Language Modeling
8. for loops (C programming tutorial part: 8/42)
มุมมอง 459 ปีที่แล้ว
8. for loops (C programming tutorial part: 8/42)
9. while loops (C programming tutorial part: 9/42)
มุมมอง 349 ปีที่แล้ว
9. while loops (C programming tutorial part: 9/42)
13. decision making (C programming tutorial part: 13/42)
มุมมอง 699 ปีที่แล้ว
13. decision making (C programming tutorial part: 13/42)
15. more on decisions (C programming tutorial part: 15/42)
มุมมอง 159 ปีที่แล้ว
15. more on decisions (C programming tutorial part: 15/42)
18. array (C programming tutorial part: 18/42)
มุมมอง 319 ปีที่แล้ว
18. array (C programming tutorial part: 18/42)
17. switch (C programming tutorial part: 17/42)
มุมมอง 279 ปีที่แล้ว
17. switch (C programming tutorial part: 17/42)
16. if else ladder (C programming tutorial part: 16/42)
มุมมอง 2779 ปีที่แล้ว
16. if else ladder (C programming tutorial part: 16/42)
14. nested if (C programming tutorial part: 14/42)
มุมมอง 239 ปีที่แล้ว
14. nested if (C programming tutorial part: 14/42)
10. do while loops (C programming tutorial part: 10/42)
มุมมอง 309 ปีที่แล้ว
10. do while loops (C programming tutorial part: 10/42)
12. application of loops (C programming tutorial part: 12/42)
มุมมอง 379 ปีที่แล้ว
12. application of loops (C programming tutorial part: 12/42)
20. 2Dimension array (C programming tutorial part: 20/42)
มุมมอง 359 ปีที่แล้ว
20. 2Dimension array (C programming tutorial part: 20/42)
11. nested for loops (C programming tutorial part: 11/42)
มุมมอง 209 ปีที่แล้ว
11. nested for loops (C programming tutorial part: 11/42)
19. application of array (C programming tutorial part: 19/42)
มุมมอง 6839 ปีที่แล้ว
19. application of array (C programming tutorial part: 19/42)
22. strings (C programming tutorial part: 22/42)
มุมมอง 229 ปีที่แล้ว
22. strings (C programming tutorial part: 22/42)
23. introduction to string library (C programming tutorial part: 23/42)
มุมมอง 299 ปีที่แล้ว
23. introduction to string library (C programming tutorial part: 23/42)
21. character data type (C programming tutorial part: 21/42)
มุมมอง 239 ปีที่แล้ว
21. character data type (C programming tutorial part: 21/42)
24. strcpy and strcat (C programming tutorial part: 24/42)
มุมมอง 1.5K9 ปีที่แล้ว
24. strcpy and strcat (C programming tutorial part: 24/42)
25. functions (C programming tutorial part: 25/42)
มุมมอง 499 ปีที่แล้ว
25. functions (C programming tutorial part: 25/42)

ความคิดเห็น

  • @dpsai3632
    @dpsai3632 6 วันที่ผ่านมา

    I am also waiting

  • @swakathumamakeshwaran8881
    @swakathumamakeshwaran8881 6 วันที่ผ่านมา

    Great! Awaiting for the next videos. Thanks for your efforts.

  • @Ravi-b8i
    @Ravi-b8i 28 วันที่ผ่านมา

    thanks bro, really helpfull video

  • @data_science_tutorial
    @data_science_tutorial หลายเดือนก่อน

    nice explanatation

  • @hemav5258
    @hemav5258 2 หลายเดือนก่อน

    Your video is very good.. easy to learn the concept easily..thank you sir..Please provided the detailed example for grammar based LM

  • @RoobiniB-i7b
    @RoobiniB-i7b 2 หลายเดือนก่อน

    Very nice 👍

  • @JosueK.
    @JosueK. 2 หลายเดือนก่อน

    hi with a 16GB RAM i am not able, to save embedding in the collection what could be the issue?What are your system configuration, my kernel is crashing

  • @rakeshkumarrout2629
    @rakeshkumarrout2629 3 หลายเดือนก่อน

    can i usethis in vscode?

  • @jatindhiman448
    @jatindhiman448 3 หลายเดือนก่อน

    Really great explanation..... But i am stuck in a problem of getting space on gpu. If i tried this on google collab ,the free version gets collapsed due to all memory usage. Pls suggest me for this solution or list the name of small models that are under 12gb of space & are used for prompting purpose.

  • @mayowaogundipe375
    @mayowaogundipe375 3 หลายเดือนก่อน

    Thanks for your time... Please may I ask how to download coda toolkit on my laptop to support GPU support. The code for Coda or cpu is not working on my laptop

  • @dr_einsteinn
    @dr_einsteinn 3 หลายเดือนก่อน

    is the code working on any app platform like streamlit?

  • @niklasweiss2557
    @niklasweiss2557 3 หลายเดือนก่อน

    I currently have the problem that it only says "Loading widget..." when I try to run the code and doesn't display the progress bar. Do you possibly know how to fix this?

  • @swxin9
    @swxin9 3 หลายเดือนก่อน

    System Configuration used?

  • @vasanthnagkv5654
    @vasanthnagkv5654 4 หลายเดือนก่อน

    Thanks! this was my first AI development video watch.

  • @INGLERAJKAMALRAJENDRA
    @INGLERAJKAMALRAJENDRA 4 หลายเดือนก่อน

    @3:58 Fine-tuning of a LLM is done via such chat or instructions

  • @RoobiniB-i7b
    @RoobiniB-i7b 4 หลายเดือนก่อน

    nlp concepts you are the best to explain super bro

  • @taijingshen4503
    @taijingshen4503 4 หลายเดือนก่อน

    Do we need to request access from the repo owner for llama 2 7b chat hf used here?

  • @khaitruong9831
    @khaitruong9831 5 หลายเดือนก่อน

    Great video. Thank you ycopie!

  • @jennilthiyam1261
    @jennilthiyam1261 5 หลายเดือนก่อน

    hi. thank you for this video. i am new to this field. i wanna ask where is the model located? i have downloaded llama 2 70B from meta. i got it after signing up and from a link they have provided through mail. if you are working in llama-2 i think you might know what i am talking about. i am new here so i am confused. what model can i used to work on my local system .

  • @jennilthiyam1261
    @jennilthiyam1261 5 หลายเดือนก่อน

    what will we do if i need interactive mode, like having conversation like we do with chatgpt

  • @litttlemooncream5049
    @litttlemooncream5049 5 หลายเดือนก่อน

    really subtle! subscribed

  • @litttlemooncream5049
    @litttlemooncream5049 5 หลายเดือนก่อน

    love your username lol

  • @IITianGamer640
    @IITianGamer640 5 หลายเดือนก่อน

    bro just found your vedio for this topic and you teched the best all others are just shit bro thanks a lot if you see this message

  • @AlainFavre-n4k
    @AlainFavre-n4k 5 หลายเดือนก่อน

    Very instructive and easy to follow

  • @lesstalkeatmore9441
    @lesstalkeatmore9441 5 หลายเดือนก่อน

    how to fine tune with our own data sets, like answer the pdf of our own data sets.

  • @pradachan
    @pradachan 5 หลายเดือนก่อน

    i'm new to LLM and i just wanted to know that you need all these access for using llama, but when you'd use ollama you just put "ollama run llama2" in the terminal, so whats the difference? they can access it without any explicit access from meta??

  • @abutareqrony4513
    @abutareqrony4513 6 หลายเดือนก่อน

    Great explanation bro. Thanks

  • @gaspardtissandier3204
    @gaspardtissandier3204 6 หลายเดือนก่อน

    Great video, and it is indeed the right translation to french :)

  • @weelianglien687
    @weelianglien687 6 หลายเดือนก่อน

    Thank you for this hands-on! Initially I tried it on my laptop which although has an NVDA GEFORCE GTX, it can't run very well. Eventually I have to run it on Colab (T4 GPU), though not with adding the following lines to help with the GPU usage (just sharing) : !pip install accelerate from accelerate import Accelerator accelerator = Accelerator() device = accelerator.device

    • @22nd.of.may.
      @22nd.of.may. 3 หลายเดือนก่อน

      my model needs 16.2gb of GPU, which in colab is limited to 15gb, do you have any way to fix that?

  • @sachinworld_
    @sachinworld_ 6 หลายเดือนก่อน

    ValueError: You are trying to offload the whole model to the disk. Please use the `disk_offload` function instead. i got this error

  • @karthikr4183
    @karthikr4183 6 หลายเดือนก่อน

    You just got a new subscriber. Good job with the explanations.

  • @tanumoymajumdar874
    @tanumoymajumdar874 6 หลายเดือนก่อน

    Langchain in related to RAG ? I am very new.

  • @bhaavamritdhaara
    @bhaavamritdhaara 6 หลายเดือนก่อน

    Very helpful. Thanks a lot for this.

  • @sambitbehura6456
    @sambitbehura6456 7 หลายเดือนก่อน

    Hi Yash. I have been trying to run your code on Colab. I have been facing issues while running model = AutoModelForCausalLM.from_pretrained( "meta-llama/Llama-2-7b-chat-hf", device_map='auto' ) I tried passing access token, but getting errors. Your guidance would be greatly appreciated

    • @ycopie1126
      @ycopie1126 7 หลายเดือนก่อน

      Can you please post the error

    • @sambitbehura6456
      @sambitbehura6456 7 หลายเดือนก่อน

      @@ycopie1126 ------------------------------------- CODE --------------------------------------------- model_id = "meta-llama/Llama-2-7b-chat-hf" model = AutoModelForCausalLM.from_pretrained(model_id, load_in_4bit=True, device_map="auto") tokenizer = AutoTokenizer.from_pretrained(model_id) -----------------------------------ERROR-------------------------------------------------------------------------- The `load_in_4bit` and `load_in_8bit` arguments are deprecated and will be removed in the future versions. Please, pass a `BitsAndBytesConfig` object in `quantization_config` argument instead. --------------------------------------------------------------------------- ImportError Traceback (most recent call last) <ipython-input-26-93e675faebf6> in <cell line: 7>() 5 model_id = "meta-llama/Llama-2-7b-chat-hf" 6 ----> 7 model = AutoModelForCausalLM.from_pretrained(model_id, load_in_4bit=True, device_map="auto") 8 tokenizer = AutoTokenizer.from_pretrained(model_id) 2 frames /usr/local/lib/python3.10/dist-packages/transformers/quantizers/quantizer_bnb_4bit.py in validate_environment(self, *args, **kwargs) 60 def validate_environment(self, *args, **kwargs): 61 if not (is_accelerate_available() and is_bitsandbytes_available()): ---> 62 raise ImportError( 63 "Using `bitsandbytes` 8-bit quantization requires Accelerate: `pip install accelerate` " 64 "and the latest version of bitsandbytes: `pip install -i pypi.org/simple/ bitsandbytes`" ImportError: Using `bitsandbytes` 8-bit quantization requires Accelerate: `pip install accelerate` and the latest version of bitsandbytes: `pip install -i pypi.org/simple/ bitsandbytes` --------------------------------------------------------------------------- NOTE: If your import is failing due to a missing package, you can manually install dependencies using either !pip or !apt. To view examples of installing some common dependencies, click the "Open Examples" button below.

  • @CarolinaHernandez-zt6li
    @CarolinaHernandez-zt6li 7 หลายเดือนก่อน

    Do you offer any paid consulting? I’m stuck on an installation error.

  • @harsh2014
    @harsh2014 7 หลายเดือนก่อน

    Thanks for this discussion !

  • @mohammedmujtabaahmed490
    @mohammedmujtabaahmed490 7 หลายเดือนก่อน

    ConnectionError: ('Connection aborted.', ConnectionResetError(10054, 'An existing connection was forcibly closed by the remote host', None, 10054, None)) bro iam getting this error when running on jupyter notebook. please help.

  • @mohammedmujtabaahmed490
    @mohammedmujtabaahmed490 7 หลายเดือนก่อน

    ConnectionError: ('Connection aborted.', ConnectionResetError(10054, 'An existing connection was forcibly closed by the remote host', None, 10054, None)) bro iam getting this error when running on jupyter notebook. please help.

  • @raahul42
    @raahul42 7 หลายเดือนก่อน

    How are you loading a 7b model locally? can I know your RAM

  • @hellobunny12
    @hellobunny12 7 หลายเดือนก่อน

    thank you 🙏

  • @hellobunny12
    @hellobunny12 7 หลายเดือนก่อน

    thank you for sharing your knowledge its very helpful

  • @samirait-abbou5954
    @samirait-abbou5954 7 หลายเดือนก่อน

    good job!

  • @fabiotarocasalino257
    @fabiotarocasalino257 7 หลายเดือนก่อน

    good vid

  • @sumandas829
    @sumandas829 7 หลายเดือนก่อน

    hf represents human feedback not hugging face

    • @ycopie1126
      @ycopie1126 7 หลายเดือนก่อน

      You can follow this discussion: github.com/facebookresearch/llama/issues/612 The model card has small difference which states that it's hugging face format.

    • @sumandas829
      @sumandas829 7 หลายเดือนก่อน

      Extremely sorry for doubting, just thought hf should mean human feedback, again I am wrong, sorry for that, good job

    • @ycopie1126
      @ycopie1126 7 หลายเดือนก่อน

      No worries at all. Happy you put it in comments so it would help other people as well 😄

  • @flareonspotify
    @flareonspotify 8 หลายเดือนก่อน

    very good and simple

  • @mj_cta
    @mj_cta 8 หลายเดือนก่อน

    The model takes lot of time to respond , like 3minutes for 50 tokens. Any suggestions to how we can improve the timing, would be much appreciated. I am running it on V100 on Databricks. Enough Compute and memory is present.

  • @mj_cta
    @mj_cta 8 หลายเดือนก่อน

    24:39 - Fun part of the video, good luck Yash ! Thanks for the video.

  • @rastapopolous8446
    @rastapopolous8446 8 หลายเดือนก่อน

    nice tutorial but how would you do to wait for the prompt.. so we can enter the prompt like what is capital in Indisa and press enter.. then the model should reply.. how to do it.

  • @thamilarasan4048
    @thamilarasan4048 8 หลายเดือนก่อน

    Please share you system specs, specially about GPU you are using

  • @SpartanDemiGod
    @SpartanDemiGod 8 หลายเดือนก่อน

    Can you please tell me your PC specs ?