Introduction to large language models

แชร์
ฝัง
  • เผยแพร่เมื่อ 2 มิ.ย. 2024
  • Enroll in this course on Google Cloud Skills Boost → goo.gle/3nXSmLs
    Large Language Models (LLMs) and Generative AI intersect and they are both part of deep learning. Watch this video to learn about LLMs, including use cases, Prompt Tuning, and GenAI development tools.
    Subscribe to Google Cloud Tech → goo.gle/GoogleCloudTech
  • วิทยาศาสตร์และเทคโนโลยี

ความคิดเห็น • 109

  • @EKOLegend
    @EKOLegend ปีที่แล้ว +143

    The mere fact that every large player in this space has videos teaching people about these things means this is super super serious.

    • @ChatGTA345
      @ChatGTA345 ปีที่แล้ว +10

      Or that it is a massive massive waste of time and effort

    • @zappy9880
      @zappy9880 ปีที่แล้ว +10

      @@ChatGTA345 unlikely. 1 or 2 small companies pursuing this tech with such ambition could be a waste of time. But if all the big players are investing their time and money in this tech, then it has to be something very real and very serious

    • @ChatGTA345
      @ChatGTA345 ปีที่แล้ว +5

      @@zappy9880 I don't think that necessarily follows. The industry has followed so many hype waves before. The competitive advantage is actually not to do what everyone else does

    • @padmakalluri181
      @padmakalluri181 11 หลายเดือนก่อน

      @@ChatGTA345 😊😊😊😊😊😊❤

    • @espiya5557
      @espiya5557 11 หลายเดือนก่อน +2

      @@ChatGTA345 well, it is a waste depends on how you use it. but can really be useful in several fields if you know how to use it and how you fine-tune it. just treat it as some sort of assisting tool as of now, and not as something that you actually use as some sort of definitive source of knowledge.

  • @sarahsalt3689
    @sarahsalt3689 9 หลายเดือนก่อน +7

    Thank you for making this available to the general public!

  • @davidcottrell1308
    @davidcottrell1308 ปีที่แล้ว +20

    Fantastic presentation...and...(I LOVE THIS) NO ANNOYING BACKING TRACK!! Thank you, Google!

  • @genlu6137
    @genlu6137 8 หลายเดือนก่อน +33

    Appreciate the valuable content! Sharing some key takeaways of the video and I hope this can help someone out.
    1) 00:50 - Large language models (LLMs) are general purpose language models that can be pre-trained and fine-tuned for specific purposes.
    LLMs are trained for general purposes to solve common language problems, and then tailored to solve specific problems in different fields.
    2) 02:04 - Large language models have enormous size and parameter count.
    The size of the training data set can be at the petabyte scale, and the parameter count refers to the memories and knowledge learned by the machine during training.
    3) 03:01 - Pre-training and fine-tuning are key steps in developing large language models.
    Pre-training involves training a large language model for general purposes with a large data set, while fine-tuning involves training the model for specific aims with a much smaller data set.
    4) 03:15 - Large language models offer several benefits.
    They can be used for different tasks, require minimal field training data, and their performance improves with more data and parameters.
    5) 08:50 - Prompt design and prompt engineering are important in large language models.
    Prompt design involves creating a clear, concise, and informative prompt for the desired task, while prompt engineering focuses on improving performance.
    6) 13:43 - Generative AI Studio and Generative AI App Builder are tools for exploring and customizing generative AI models.
    Generative AI Studio provides pre-trained models, tools for fine-tuning and deploying models, and a community forum for collaboration.
    7) 14:52 - Palm API and Vertex AI provide tools for testing, tuning, and deploying large language models.
    Palm API allows testing and experimenting with large language models and gen AI tools, while Vertex AI offers task-specific Foundation models and parameter efficient tuning methods.
    This takeaway note is made with the Notable app (getnotable.ai).

  • @richardglady3009
    @richardglady3009 8 หลายเดือนก่อน

    Thank you. I understood about half (optimistically) of it. I subscribed to the channel hoping to start from the beginning and understanding more. My ultimate goal: a LLM Librarian, combining the catalog of a library with results from internet search engine, giving the deepest answer possible.

  • @joseperez-ig5yu
    @joseperez-ig5yu ปีที่แล้ว +24

    Finding answers to questions has become so much easier now with new tech. I have never been good at writing code, so this is a welcome change as far as I'm concerned! Look forward to more progress in technology.

    • @yenda12
      @yenda12 11 หลายเดือนก่อน +3

      Be careful in the world to come being reliant on these AIs without developing any specific field will make you obsolete in future society

  • @JonathanPoczatek
    @JonathanPoczatek ปีที่แล้ว +12

    Can't wait to see demos at GoogleIO

  • @fred-nyanokwi
    @fred-nyanokwi 2 หลายเดือนก่อน

    This is one of the educative sessions I've come across

  • @yabadab8609
    @yabadab8609 ปีที่แล้ว +30

    Actually, really helpful, thank you Google.
    Wondering how far this technology will go in the next couple of years, if it's this far already in a couple of months.

    • @webgpu
      @webgpu ปีที่แล้ว +1

      we all wondering too ;)

  • @robertcormia7970
    @robertcormia7970 11 หลายเดือนก่อน +9

    This was fantastic! While I've been watching The Full Stack LLM Bootcamp, I'm not technically strong enough to start there, and will use these Google Cloud Tech videos as a means to "jumpstart" my knowledge of LLM and Generative AI. This is a great general primer for students and colleagues!

    • @mostafatouny8411
      @mostafatouny8411 9 หลายเดือนก่อน +2

      Thanks for referencing Full Stack LLM Bootcamp, A great resource I was not aware of.

  • @dariannwankwo9126
    @dariannwankwo9126 6 หลายเดือนก่อน +10

    Minor Correction @ 2:14. "In ML, parameters are often called hyperparameters." In ML, parameters and hyperparameters can exist simultaneously and serve two different purposes. One can think of hyperparameters as the set of knobs that the designer has direct influence to change as they see fit (whether algorithmically or manually). As for the parameters of a model, one can think of it as the set of knobs that are learned directly from the data. For hyperparameters, you specify them prior to the training step; while the training step proceeds, the parameters of the model are being learned.

  • @PayzdeMontEtteineneyeUtlingele
    @PayzdeMontEtteineneyeUtlingele ปีที่แล้ว +2

    proximity and stream for seek time reduction..memory in case reduced latency, can also be optimized for seek time and pattern analysis.

  • @cassianocominetti7784
    @cassianocominetti7784 5 หลายเดือนก่อน

    Very comprehensive video! Thank you guys!

  • @sasasunshine6
    @sasasunshine6 ปีที่แล้ว +5

    Great video! Thank you!!

  • @arnoldpraesent174
    @arnoldpraesent174 9 หลายเดือนก่อน

    very well and understandable explained... good job!

  • @henri8903
    @henri8903 ปีที่แล้ว +10

    Thank you John. I believe you conflated model parameters and hyperparameters at 2:16. As far as I know, these are two different concepts.

    • @fierce10
      @fierce10 9 หลายเดือนก่อน +3

      Yes, they are different conceptually. Parameters are directly applied/calculated in the hypothesis or model; while, hyperparameters are somewhat heuristically decided based on what works. For example if you were figuring out how to get from home to office, the path details maybe calculated directly by the GPS, but the time at which you leave maybe heuristically decided by you. Another example of a hyperparameter can be how many backup cameras you choose to add should the main camera fail on a robot, there is no 'correct' number, it's more of a cost or design choice. In an ML transformer, choosing the number of encoders or decoders can be a hyperparameter. The parameters would be learned from the language training in the LLM.

    • @03timboy
      @03timboy 7 หลายเดือนก่อน +1

      Agreed, two totally different things. It's not great that the video encourages this confusion.

  • @malacca498
    @malacca498 8 หลายเดือนก่อน

    Always great to learn from GCT!

  • @BrandonLee-ik8kw
    @BrandonLee-ik8kw ปีที่แล้ว +9

    2:47 You mentioned the parameters are hyper parameters is incorrect and confusing

  • @sidindian1982
    @sidindian1982 10 หลายเดือนก่อน

    Excellent Presentation Sir ... truly i admire it 😍😍😍😍

  • @BREAKKWISS
    @BREAKKWISS ปีที่แล้ว +2

    Thank you for teaching.

  • @user-mh4sz7yu1p
    @user-mh4sz7yu1p ปีที่แล้ว +1

    great content! make me feel like an expert now💯

  • @higiniofuentes2551
    @higiniofuentes2551 6 หลายเดือนก่อน

    Wow!
    Thank you for this very useful video so well explained!

  • @bakerkawesa
    @bakerkawesa ปีที่แล้ว +10

    Great explainer. I'm a little less anxious about AI taking our jobs.

    • @near_.
      @near_. 11 หลายเดือนก่อน +2

      1980s or so, there were telephone operator who connects those STD lines.
      Now they are vanished but their next gen kids are employed in another market.
      That's how innovation works!!

  • @YHK_YT
    @YHK_YT ปีที่แล้ว +2

    Time to start my own

  • @luminouswolf7117
    @luminouswolf7117 11 หลายเดือนก่อน +2

    If you define the problem you are trying to solve first
    Then reason from their
    Wouldn’t it be more efficient?

  • @jeganathanmanickam6604
    @jeganathanmanickam6604 4 หลายเดือนก่อน

    Very Informative - Thanks for sharing 😊 prompt design and prompt engineering would take make the conversation more realistic and accurate.

  • @DrJanpha
    @DrJanpha ปีที่แล้ว

    is it true that AI models like ChatGPT or Bard are fed in with codes (like programming languages) as well?

  • @lengsolace6074
    @lengsolace6074 ปีที่แล้ว +1

    helpful for me,tks google

  • @user-xv8dn4nm5k
    @user-xv8dn4nm5k 4 หลายเดือนก่อน +1

    Thank for sharing👍

  • @MeenakshiSharma-ss2ir
    @MeenakshiSharma-ss2ir 10 หลายเดือนก่อน

    At 4:50 I did not understand the third point that the speaker made i.e. "Orchestrated distributes computation for accelerators". Can someone please explain?

  • @Minimalrevolt-m83
    @Minimalrevolt-m83 ปีที่แล้ว +3

    Waow! Such an eye-opening knowledge!🤓

  • @reinventingai
    @reinventingai ปีที่แล้ว

    Very slim on the prompt engineering education. This is a very important skill!

  • @CiscoZero
    @CiscoZero ปีที่แล้ว

    Exciting stuff.

  • @lifeofdean3647
    @lifeofdean3647 ปีที่แล้ว +7

    can u share awesome slides ?

  • @eriqfromimo2570
    @eriqfromimo2570 4 หลายเดือนก่อน

    Nice one!

  • @pradeepbhatt4857
    @pradeepbhatt4857 11 หลายเดือนก่อน

    where can i access gen ai studio and build apps?

  • @deeplearningpartnership
    @deeplearningpartnership ปีที่แล้ว +5

    Nice.

  • @jaykef
    @jaykef 11 หลายเดือนก่อน +1

    please provide link to the slides

  • @alexanderwilliams6527
    @alexanderwilliams6527 2 หลายเดือนก่อน

    0:57 What do pre-trainned and fine-tuned llms means? Good analogy with dogs.

  • @artus198
    @artus198 ปีที่แล้ว +2

    What does 540 billion parameters mean , and how do you pass those to your model ? What kind of computational processing power is needed for this ?

    • @strider806
      @strider806 ปีที่แล้ว

      You don't have to pass the parameters. In Llm you just send the data as text and it must be able to tokenize the text.

    • @MrAmgadHasan
      @MrAmgadHasan ปีที่แล้ว +3

      You first instantiate the model with randomly generated parameters (540B in this case) and use lots and lots and lots of data to make the model "learn" and modify these parameters so they are better. For llms, you need hundreds of powerful gpus and you need weeks or months to train such massive models. Falcon 40B which is a state of the art open source model with 40B parameters was trained for two months.

    • @artus198
      @artus198 ปีที่แล้ว

      @@MrAmgadHasan chatgpt was trained for about 2 years , there are 2 seperate models within chatgpt , one to understand context, the other to predict the text 🤪

    • @MrAmgadHasan
      @MrAmgadHasan ปีที่แล้ว

      @@artus198 chatgpt is not a pure LLM. It was finetuned using multiple instructions datsets and RLHF. I was talking about training pure LLMs

  • @coryrandolph8501
    @coryrandolph8501 ปีที่แล้ว +11

    This is a great overview video thank you.
    Do you have a reference for how to host open-sourced LLM's in Vertex AI (or other GCP tools)? Overall I'm looking for GCP tools and ways for turning open-source LLM's into API's to be used within our native cloud instance.

    • @B2M2948
      @B2M2948 ปีที่แล้ว

      You lost any semblance of an answer from @Google Cloud Tech the second you said "open-source"

    • @coryrandolph8501
      @coryrandolph8501 ปีที่แล้ว

      @@B2M2948 lol.
      I still want to host the Open source thing on their platform so I thought there might be a shot.

    • @andrestorres7343
      @andrestorres7343 6 หลายเดือนก่อน

      ​@@coryrandolph8501did you ever figure it out?

    • @coryrandolph8501
      @coryrandolph8501 6 หลายเดือนก่อน

      @@andrestorres7343 Yes, but it was really pricey since you have to host the underlying infrastructure. Usually large GPU virtual machines and on GCP depending on model size it was $2k - $5k per month to host an open source model.
      We are sticking with the API version of the big models because of this.

  • @littlebrit
    @littlebrit ปีที่แล้ว +4

    What is the legal status now of LLM models trained on proprietary data ?

  • @amirkhalesi5294
    @amirkhalesi5294 8 หลายเดือนก่อน

    For the fellow beginners:
    PETM is also called PEFT

  • @jeanpeuplu3862
    @jeanpeuplu3862 7 หลายเดือนก่อน

    I have an urgent question (school related) -> is LLM part of NLP? Is an LLM always an NLP model? Or can an LLM be another kind of model? "L" for Language in both kinds of models. Both in AI. Both for language.
    A colleague says LLM is not necessarily an NLP model but then I did not understand LLM and/or NLP and my oral exam is in few days omg

    • @jeanpeuplu3862
      @jeanpeuplu3862 7 หลายเดือนก่อน +1

      Also, BERT is Transformer but not an LLM, right? Transformer can be LLM or not, right?

  • @LavaCreeperPeople
    @LavaCreeperPeople หลายเดือนก่อน +1

    Cool!

  • @PayzdeMontEtteineneyeUtlingele
    @PayzdeMontEtteineneyeUtlingele ปีที่แล้ว +1

    pattern analysis with causal.

  • @jackyang7401
    @jackyang7401 20 วันที่ผ่านมา

    i love it

  • @PayzdeMontEtteineneyeUtlingele
    @PayzdeMontEtteineneyeUtlingele ปีที่แล้ว +1

    you can use a new drive architecture sought via gpu pixels for proximity stream like to not need large.lamguage models, and use multi factor checks to reduce need of a lot of data..thank me now.

    • @PayzdeMontEtteineneyeUtlingele
      @PayzdeMontEtteineneyeUtlingele ปีที่แล้ว +1

      proximity and stream for seek time reduction..memory in case reduced latency, can also be optimized for seek time and pattern analysis.

  • @imranhossain3504
    @imranhossain3504 11 หลายเดือนก่อน

    I'm with you

  • @JaySmith91
    @JaySmith91 ปีที่แล้ว

    What's a TPU V4 Pod? Sounds like a Turboencabulator, or?

    • @MrAmgadHasan
      @MrAmgadHasan ปีที่แล้ว

      It's a custom built computer chip developed by google to perform matrix operations and train deep learning models. Think of them as gpus specialized for deep learning.

    • @MrAmgadHasan
      @MrAmgadHasan ปีที่แล้ว

      A pod is "rack" of tens or hundred tpu/gpu.

  • @josephvanname3377
    @josephvanname3377 ปีที่แล้ว +3

    Reversible computing is still the future regardless of whether people would like to admit it or not.

    • @IForgetWhatISay
      @IForgetWhatISay ปีที่แล้ว

      What do you mean

    • @josephvanname3377
      @josephvanname3377 ปีที่แล้ว

      @@IForgetWhatISay I mean that in the future, all computers will be reversible computers. Reversible computation will take over AI.

    • @IForgetWhatISay
      @IForgetWhatISay ปีที่แล้ว +2

      @@josephvanname3377 ah I just looked it up. Interesting. Thanks

  • @farukkara3929
    @farukkara3929 6 หลายเดือนก่อน

    Can I have these slides please?

  • @artie5172
    @artie5172 8 หลายเดือนก่อน

    Do LLM charge money for using them

  • @ninaz2735
    @ninaz2735 ปีที่แล้ว

    11:45 Can anybody explain the difference between these two prompts?

    • @johnnycyberseed
      @johnnycyberseed 6 หลายเดือนก่อน

      I understand the message of this slide to be not about prompt design, but AI response: that if the app in which the model is embedded first instructs the model to describe the process to get to an answer and THEN feed that back in with the original prompt, that the quality of the final response improves.

  • @ChatGTA345
    @ChatGTA345 ปีที่แล้ว

    I've been extremely frustrated in my interactions with chatbots, they never seem to tell the truth and it's getting harder and harder to tell what's true from what's not. I honestly like regular Google searches much more!

  • @s.ackermann5498
    @s.ackermann5498 11 หลายเดือนก่อน

    whats the name of the last circle at th-cam.com/video/zizonToFXDs/w-d-xo.html ?

  • @jumarkpelismino5632
    @jumarkpelismino5632 10 หลายเดือนก่อน

    Can users teach AI?

  • @satkotech
    @satkotech 2 หลายเดือนก่อน

    RIP Bard, gone so young..

  • @robertsutkowski3170
    @robertsutkowski3170 10 หลายเดือนก่อน

    Google 👍

  • @aliwafaafif
    @aliwafaafif ปีที่แล้ว

    Anybody who read this comment, you'd want to type this prompt in Chat-GPT or Bard: "I have 15 liter jug, 10 liter jug, and 5 liter jug. How do I measure 5 liters of water?" ---> See what they answer

  • @tombombadil9123
    @tombombadil9123 8 หลายเดือนก่อน

    Citizen Kane9 :D

  • @aditeepathak8894
    @aditeepathak8894 ปีที่แล้ว +2

    Is it just me or the quality of google training videos has gone down?

    • @MrAmgadHasan
      @MrAmgadHasan ปีที่แล้ว

      Yes. They made a mistake when they described parameters as hyper parameters and the chain of thought part wasn't clear.

  • @hussienalsafi1149
    @hussienalsafi1149 ปีที่แล้ว +2

    😊😊😊😊😊😊😊☺️☺️☺️☺️👌👌👌👌

  • @jebuskmiest
    @jebuskmiest 10 หลายเดือนก่อน +1

    So a prompt engineer is anyone with common sense?

  • @enmedallo
    @enmedallo ปีที่แล้ว

    Why so few comments

  • @IrynaCherednychenko
    @IrynaCherednychenko 18 วันที่ผ่านมา

    passed

  • @christianstill.6654
    @christianstill.6654 3 หลายเดือนก่อน

    We are creating our own prison...

  • @akj3344
    @akj3344 ปีที่แล้ว +1

    This felt more like advertisement for Bard. Not very helpful.

    • @MrAmgadHasan
      @MrAmgadHasan ปีที่แล้ว

      It is both advertising for bard and helpful too.

  • @iskalasrinivas5640
    @iskalasrinivas5640 11 หลายเดือนก่อน +2

    Really helpful video, but dont understand why it's called intelligent because it cannot discover something on its own

  • @fk_torty
    @fk_torty 2 หลายเดือนก่อน

    Was this long? YES.
    Did I learn? YES.
    Did I want to sleep? YES.
    Did I sleep before the end? NO.
    A WIN