How Codeium Breaks Through the Ceiling for Retrieval: Kevin Hou

แชร์
ฝัง
  • เผยแพร่เมื่อ 29 ก.ย. 2024
  • Codeium is trailblazing the next frontier in retrieval and hint: it’s not just embeddings. Learn what the next generation of retrieval looks like and how 1M+ developers are already leveraging this superpower using the Codeium IDE plugin for AI autocomplete, chat, and search. We’ll dive deep into how existing benchmarks are failing us, what it takes to serve our custom models at scale, and what the future of AI-assisted software development looks like.
    Recorded live in San Francisco at the AI Engineer World's Fair. See the full schedule of talks at www.ai.enginee... & join us at the AI Engineer World's Fair in 2025! Get your tickets today at ai.engineer/2025
    About Kevin
    A full stack engineer by trade and a creator by heart. Enjoys the process of creation whether it be in the physical (woodshop, blacksmithing, circuity) or the digital (software engineering, photography, and film).
    Currently building AI-powered dev tools at Codeium (Exafunction). Previously a tech lead manager at Nuro self-driving. Received a computer science engineering degree from Princeton University with certificates in entrepreneurship and statistics and machine learning.

ความคิดเห็น • 23

  • @codeiumdev
    @codeiumdev 2 หลายเดือนก่อน +33

    Shoutout to the AI Engineer World's Fair for having us and Kevin Hou! We're constantly improving our code retrieval system and working on tough problems at the cutting edge of AI. We're hiring across all roles!

    • @raulhernandez2010
      @raulhernandez2010 หลายเดือนก่อน

      Accept my internship application 😭

    • @ZeZa1515
      @ZeZa1515 หลายเดือนก่อน

      Is the commit message bank dataset available anywhere for personal use?

  • @KevinHou22
    @KevinHou22 2 หลายเดือนก่อน +24

    Thank you for having me! Always great to share what we’ve learned with the world.

    • @StephenRayner
      @StephenRayner หลายเดือนก่อน +2

      Just stopped hacking away on a Saturday on a coding assistant I am building to cut the lawn. So glad I clicked on this video.
      The talk was excellent, really enjoyed the insight into how you are your team are tackling the context issue.

    • @KevinHou22
      @KevinHou22 หลายเดือนก่อน

      @@StephenRayner Sounds like a great Saturday. Thanks for watching! Glad you were able to learn something new

    • @CryptoMaN_Rahul
      @CryptoMaN_Rahul หลายเดือนก่อน +1

      hey where can i read more about this?? i want to use similar logic in my project

  • @TimothyJoh
    @TimothyJoh 2 หลายเดือนก่อน +9

    Brilliant talk. As an engineering leader this approach is way more powerful than the state of the art from the larger companies. Appreciate the deep dive and I can see all the deep thinking going into your product.

    • @KevinHou22
      @KevinHou22 2 หลายเดือนก่อน

      Much appreciated 🙂 glad you enjoyed the talk

  • @muhannadobeidat
    @muhannadobeidat 2 หลายเดือนก่อน +3

    Thanks for the tech and for AIE for making the video available with such great production quality.
    One thing I missed is what do you train on? Especially that generated code will have ownership and IP concerns/issues, is this custom trained on organization private codebase?

    • @rohanp26
      @rohanp26 2 หลายเดือนก่อน +3

      They don't train on GPL or non-permissive code. You can use Codeium in your organization (aka Codeium doesn't claim IP over generated code). Source: Codeium FAQ

  • @shafikhan2
    @shafikhan2 หลายเดือนก่อน +1

    @KevinHou22 - We are looking for AI experts interested in part-time collaboration. If you're interested in contributing to cutting-edge AI projects for a startup, let's connect!

    • @JoJoseM-kb3mf
      @JoJoseM-kb3mf หลายเดือนก่อน

      Hi Shafikhan can we connect?

  • @JoJoseM-kb3mf
    @JoJoseM-kb3mf หลายเดือนก่อน

    Hello. This sounds like a great approach. I just have a question about scalability up/down when using bare metal?

  • @gnsdgabriel
    @gnsdgabriel หลายเดือนก่อน

    What do you think about this idea? A sort of deep augmented retrieval.
    Step 1: User Query
    Description: The process begins with the user submitting a query, representing their information need or question.
    Benefit: This step ensures the system receives the user’s intent clearly and initiates the interaction based on their specific requirements.
    Step 2: Generate Query Variations
    Description: The language model generates a series of query variations based on its underlying understanding of the original query.
    Benefit: By creating variations, the system captures different interpretations and aspects of the user’s intent, increasing the likelihood of retrieving relevant information from diverse sources.
    Step 3: Pattern Search Execution
    Description: Instead of performing a semantic search, the system conducts a pattern search using the query variations to identify possible or probable vector patterns in the knowledge base.
    Benefit: This approach allows the system to discover less obvious connections and related content, potentially uncovering information that might be overlooked by traditional semantic searches.
    Step 4: Retrieve Full Extracts
    Description: The system retrieves full relevant sections from the knowledge base based on the identified patterns.
    Benefit: Extracting comprehensive sections rather than isolated snippets provides a richer context, enhancing the model’s ability to generate accurate and contextually rich responses.
    Step 5: Input Extracts into Context Window
    Description: The retrieved extracts are directly input into the context window of the language model.
    Benefit: By referencing specific information, the model can draw upon current and relevant data, mitigating the limitations of its internal, potentially outdated knowledge.
    Step 6: Internal Reasoning and Query Optimization
    Description: The system performs an internal reasoning step to refine and optimize the original query based on the insights gained from the extracts.
    Benefit: This iterative refinement aligns the query more closely with the available information, improving the accuracy and relevance of the system’s responses.
    Step 7: New Inference Instance
    Description: A new inference instance is initiated using both the optimized query and the relevant extracts.
    Benefit: By leveraging fresh insights and a refined query, the model can produce a more informed and accurate output, effectively integrating the retrieved knowledge.
    Step 8: Generate Output
    Description: The language model generates the final output based on the optimized query and the contextual extracts.
    Benefit: This step culminates the process, delivering a response that integrates the user’s original intent with the most relevant and current information, thereby enhancing the quality of the output.

  • @Player-oz2nk
    @Player-oz2nk 2 หลายเดือนก่อน +2

    Cursor plus codeium is BEAST

    • @maxwelljiang4729
      @maxwelljiang4729 2 หลายเดือนก่อน +4

      whoa :0 what's your setup here? curious what cursor provides that codeium's missing, and vice versa

    • @sarolovito2838
      @sarolovito2838 หลายเดือนก่อน

      @@maxwelljiang4729 I agree with you: cursor + codeium is awesome

  • @ForChiddlers
    @ForChiddlers หลายเดือนก่อน

    What is this? I came for knowledge and got an 18 minute ad with sales blafasel buzzwords

    • @trejohnson7677
      @trejohnson7677 หลายเดือนก่อน

      im unsure of what wasn't expressed clearly? did u stop to think that possibly the "buzzwords" were things that went over ur head? they said "we got our own shit and the other guys are using someone else's shit".

    • @ForChiddlers
      @ForChiddlers หลายเดือนก่อน

      @@trejohnson7677 are they now Training their llms or are they just ragging? How does the pipeline look? Do they do it supervised or unsupervised ? Do they use an open source llm and do a lora training on it?

  • @ScriptGurus1
    @ScriptGurus1 หลายเดือนก่อน +1

    Lots of abstraction wordsalad signifying nothing.