Mitigating LLM Hallucinations with a Metrics-First Evaluation Framework

แชร์
ฝัง
  • เผยแพร่เมื่อ 24 ก.ค. 2024
  • Join in on this workshop where we will showcase some powerful metrics to evaluate the quality of the inputs and outputs with a focus on both RAG and fine-tuning use cases. In the context of LLMs, “hallucination” refers to a phenomenon where the model generates text that is incorrect, nonsensical, or not real. Since LLMs are not databases or search engines, they would not cite where their response is based on. These models generate text as an extrapolation from the prompt you provided.
    What attendees can expect to takeaway from the workshop:
    -Deep dive into research-backed metrics to evaluate the quality of the inputs (data quality, RAG context quality, etc) and outputs (hallucinations) while building LLM powered applications.
    -Evaluation and experimentation framework while prompt engineering with RAG, as well as while fine-tuning with your own data
    -Demo led practical guide to building guardrails and mitigating hallucinations while building LLM powered applications
    To access the slides, please click here:
    docs.google.com/presentation/...
    To read the academic paper, please click here:
    www.rungalileo.io/blog/chainpoll
    To see these concepts in action, take a look at the Hallucination Index here: www.rungalileo.io/hallucinati...
    This event is inspired by DeepLearning.AI’s GenAI short courses, created in collaboration with AI companies across the globe. Our courses help you learn new skills, tools, and concepts efficiently within 1 hour.
    www.deeplearning.ai/short-cou...
    About Galileo
    At Galileo we are building the first algorithm-powered LLMOps Platform for the enterprise. Galileo provides ML teams with an intelligent ML data bench to collaboratively improve data quality across their model workflows - from pre-training, to post-production. Galileo is currently powering ML teams across the Fortune 500 as well as startups across multiple industries.
    Speakers:
    Vikram Chatterji, Co-founder and CEO at Galileo
    / vikram-chatterji
    Atindriyo Sanyal, Co-founder and CTO at Galileo
    / atinsanyal
  • บันเทิง

ความคิดเห็น • 20

  • @HonestGraduate
    @HonestGraduate 9 หลายเดือนก่อน +1

    Thank you for the presentation and demo!

  • @ajeethkumar6296
    @ajeethkumar6296 หลายเดือนก่อน

    Thanks for the clear cut explanation

  • @user-wz5rd6vg2r
    @user-wz5rd6vg2r 9 หลายเดือนก่อน +15

    The real contribution seems to be the prompt they used to generate the CoT and the metric value... Could you share the code used for the metric and the prompt for ChatPGT?

  • @KokkeOP
    @KokkeOP 9 หลายเดือนก่อน +2

    The paper and the Slides are both in the description, guys. :) read.

  • @purvislewies3118
    @purvislewies3118 9 หลายเดือนก่อน +1

    Blessed love...givethanks...Cape Town

  • @user-wz5rd6vg2r
    @user-wz5rd6vg2r 9 หลายเดือนก่อน +4

    Nice talk! Could you please share the notebook?

  • @danteblink
    @danteblink 9 หลายเดือนก่อน +1

    Do you think human intervention in the evaluation process is going to last? It seems its a process that LLMs could achieve by themselves in the near future.

  • @JuliusOpusprofundum
    @JuliusOpusprofundum 9 หลายเดือนก่อน +1

  • @senderlapin
    @senderlapin 9 หลายเดือนก่อน +2

    Я из России. Спасибо за вебинар.

  • @zaursamedov8906
    @zaursamedov8906 9 หลายเดือนก่อน +3

    Guys would u be able to drop the notebook please?

    • @hcrespo3
      @hcrespo3 9 หลายเดือนก่อน +4

      I'm also interested, thanks

  • @komalmistry7284
    @komalmistry7284 9 หลายเดือนก่อน

    Could someone share the link to the paper that was mentioned here "ChainPoll" , I believe.

    • @Deeplearningai
      @Deeplearningai  9 หลายเดือนก่อน

      It is in the video description!

  • @davidvilla2402
    @davidvilla2402 9 หลายเดือนก่อน +1

    I don't know how bt I searched the n word and it came up