The Ultimate Writing Challenge: Longwriter Tackles 10,000 Words In One Sitting

แชร์
ฝัง
  • เผยแพร่เมื่อ 20 ก.ย. 2024

ความคิดเห็น • 26

  • @Arcticwhir
    @Arcticwhir 21 วันที่ผ่านมา +9

    tried telling it to write a 400 line game of snake, it generated a plan (beginning was partially a story), then it output a simple code output, then a plan for more features, then code output, this repeated 4 or 5 times - the end result before getting timed out was about 100 lines of okayish python code. I know it wasnt made for this - but just wanted to share

  • @paullopez_ai
    @paullopez_ai 22 วันที่ผ่านมา +1

    Great video, Sam! I appreciate your clarity and consistency of your videos

  • @unclecode
    @unclecode 20 วันที่ผ่านมา +1

    I love how they generated the data; it's similar to how I used to create long content, starting with an outline, then building section by section, with summaries to keep things on track. What's fascinating is their explanation, showing how you can train a model for any output. I'd love to see a comparison between this method and using a model to quickly generate and merge content for each part. Which one gives better quality and integrity? That's an interesting topic to explore.

  • @Arcticwhir
    @Arcticwhir 21 วันที่ผ่านมา +1

    Tsinghua University consistently has amazing AI research papers, such a strong research department!

    • @samwitteveenai
      @samwitteveenai  15 วันที่ผ่านมา

      totally agree, lots of cool stuff coming from there

  • @novantha1
    @novantha1 22 วันที่ผ่านมา +1

    I'm not sure if this would be in your wheelhouse, but the paper "Automated Design of Agentic Systems" discussed how to automatically (!) produce agentic workflows to achieve a variety of capabilities (they showed solving the MMLU for example, I believe), where an agentic system designs an agentic system to solve your problem. I think that would be quite a fun video and a good followup if you were to show an example of it being deployed for generating synthetic data on a topic. I'm pretty sure something like that will be pretty important in the future, and I could even imagine designing an agentic system to "grade" the outputs of the automatically produced agent for achieving "soft targets" that don't have explicit successes or failures, like writing "interesting" articles, or writing "well-written" fiction and so on.

    • @andydataguy
      @andydataguy 22 วันที่ผ่านมา

      This sounds epic please do this 🙌🏾

    • @samwitteveenai
      @samwitteveenai  15 วันที่ผ่านมา +1

      This sound interesting let me check it out. I did make an Agent that can make CrewAI bots pretty easily.
      Just Googled it and I see it is citing AI Scientist which I have also worked on making a LangGraph version of.
      Thanks for the tip much appreciated, I will give the paper a read etc.

  • @reza2kn
    @reza2kn 22 วันที่ผ่านมา

    Wonderful video! 😍had seen the project but was too lazy to play with it myself. I'm reading up on Knowledge Distillation, specifically for translation tasks from the new Flash model to tiny ones, to develop tiny MT models that > SOTA. This idea of long-generations might come in handy there as well.

  • @TechnoMageCreator
    @TechnoMageCreator 22 วันที่ผ่านมา

    In my experience it's all about context. Kinda like a reverese dominos. I've been able to do it with ChatGPT-4o since it has memeory. Fist you add to memory important information. Generate a structure and add some details, generate chapter by chapter and than have it generate everything together, most i generated was about 20 min without user interaction and about 10000 lines of vode, pressed about 20 times continue generating. With every generation start to run much much slower. Since I generated like this for months manual mode I got to write down my entire process step by step. Trying now to emultae with software and multiple AI, file read/update/edit/delete. The idea is to have the user input a text, AI searches the existing documents and always create an aditional task list in order to access next file.

  • @bseddonmusic1
    @bseddonmusic1 22 วันที่ผ่านมา +5

    Interesting but what's the use case for explicitly generating long documents? Are there people who are going sit with an LLM and choose to spend hours reading long output? Maybe book publishers wanting to get away from unreliable authors? Maybe authors who have a deadline and need to get some copy to their publisher? I can understand the benefit of long output in the context of coding where real world applications are 10's of thousands of line long but is code output possible from a model like this?

    • @vanshpundir6648
      @vanshpundir6648 22 วันที่ผ่านมา +4

      Actually this is helpful when you want to generate alot of data in a single LLM call. Otherwise we need to make multiple LLM call.

    • @samwitteveenai
      @samwitteveenai  22 วันที่ผ่านมา +6

      Lots of people ask for long outputs for SEO and lots of other uses etc. For me personally I am interested in long outputs for editing long form text. Imagine something like Grammarly but where it can totally change elements of content not just grammar or spelling correction. Image things like please change this character from male to female and correct all references as an obvious example or rewrite this doc removing all political bias and make it neutral. Currently these are very hard for things that go beyond a 1000 words etc. I totally agree about the long form code btw. You could train a model to do that in a similar way.

    • @sammcj2000
      @sammcj2000 22 วันที่ผ่านมา +2

      It can be useful for coding, refactoring large code bases etc if your available tooling doesn’t handle chunking requests

    • @cemtekesin9033
      @cemtekesin9033 22 วันที่ผ่านมา

      I had a project where I had to categorize thousands of risk titles to first group them together and rewrite them to capture the risk groups consistently (data ingestion quality issue) . I was surprised to see how output token limit can be such a limiting factor.

    • @rishabnandi9593
      @rishabnandi9593 21 วันที่ผ่านมา

      Company FAQ and CS Training docs often are over 10k and someone somewhere in the ladder has to read them

  • @finlay422
    @finlay422 22 วันที่ผ่านมา

    Interested to see how this fine tuning can be applied to summarization. Maybe this can solve the issue of large documents being condensed into a few sentences.

  • @MeinDeutschkurs
    @MeinDeutschkurs 22 วันที่ผ่านมา

    You nailed it! 🎉

  • @andydataguy
    @andydataguy 22 วันที่ผ่านมา

    This is awesome!

  • @NetZeroEarth
    @NetZeroEarth 22 วันที่ผ่านมา

    🔥 🔥 🔥

  • @micbab-vg2mu
    @micbab-vg2mu 22 วันที่ผ่านมา

    Nice:)

  •  22 วันที่ผ่านมา

    Wow! This is an amazing way to create synthetic data! Today I saw this eye-opening video th-cam.com/video/FJTZP7ZdQf0/w-d-xo.html from David Shapiro based on the research "Textbooks Are All You Need
    " and it looks like LongWriter can be the solution to mastering specialized LLM.

  • @anishbhanushali
    @anishbhanushali 22 วันที่ผ่านมา

    ah ....they could have named the model "wan shi tong" ( he who knows 10,000 things)

  • @claytoncarroll2309
    @claytoncarroll2309 14 วันที่ผ่านมา

    poor video quality

  • @JNET_Reloaded
    @JNET_Reloaded 22 วันที่ผ่านมา +1

    how to use this with ollama locally?

    • @samwitteveenai
      @samwitteveenai  22 วันที่ผ่านมา +2

      It should be able to be converted to gguf for Ollama. You will need quite a bit of RAM though to generate the long outputs.