Thanks a lot. However, I got "NameError: name 'load_dataset' is not defined: at the top of "Load Datasets and Train" section. The path is correct and your copy seemed to work just fine???
Also without the comparison to the original non-fined-tuned Llama, the effectiveness of this fine tuning says nothing. It's like adding a new mechanical part to your car and measure max speed without making the same measurement before the installation
The two models are similar, except that if using Llama model, you need to login hugging face to download, if using NousResearch model, there is no need.
Want to connect?
💼Consulting: calendly.com/engineerprompt/consulting-call
🦾 Discord: discord.com/invite/t4eYQRUcXB
☕ Buy me a Coffee: ko-fi.com/promptengineering
|🔴 Join Patreon: Patreon.com/PromptEngineering
I was under the impression that the openai terms of service prevents you from using GPT4 to train another model.
only if your new model competes with them
It's really good content
Thank you!
Thanks a lot. However, I got "NameError: name 'load_dataset' is not defined: at the top of "Load Datasets and Train" section. The path is correct and your copy seemed to work just fine???
Also without the comparison to the original non-fined-tuned Llama, the effectiveness of this fine tuning says nothing. It's like adding a new mechanical part to your car and measure max speed without making the same measurement before the installation
What was the purpose of trying to use NousResearch model instead of the original Llama model by Meta?
The two models are similar, except that if using Llama model, you need to login hugging face to download, if using NousResearch model, there is no need.
Very Nice !
Thank you very much!
nice, is it also possible to run a LLM on Visual Studio Code instead of Jupiter?
Yes, the process will be similar
I wonder if soon an llm could be taught to train itself on whatever it wanted or was told to learn
youd have to have a lot of money or a very powerful architecture
it also is important to note, it doesn't want, feel, or need anything unless you somehow program that into it.
Would you be able to fine-tune a 7B parameter model locally on a MacBook Pro M2 with enough RAM? Really curious to see if it’s possible…
I haven’t had luck with it, so probably not
@@engineerprompt did you come across a solution?
NousResearch/llama-2-7b-chat-hf is no longer available. What model would you recommend instead?
If you have the access token then use the meta original version
thank you for all this great job , and where I can find discord group
In the description of the video :)
Is it possible to use any kind of source code(python, c++ or c#) as a dataset to train llama 2 or any other LLM? Thank you!
Yes, but you will need a relatively large dataset for it to train properly.
Nothing to do with a "Dataset".
This is "how to train ChatGPT to respond to a specific prompt"
first comment
السنة من أعطى قيمة لهذه الطائفة