I tried out brev for my machine learning course, love the options for the payment system where i have the option to cut if off after X dollars, low prices, and ui looks awesome. I know it said it somewhere but it took me a minute to realize that my Jupyter notebook takes around 4 minutes to launch so for blind ppl like me I’d put some more text saying Jupyter notebook will be created in X minutes. Love this vid and outreach- I’ll keep watching Baxate
WOW!!! I'm so glad I found this information. This is incredible! I'm utilizing the llava model now in my robot project and just like you said, llava out of the box is not a good fit, especially for the computer vision/classification experiments I'm doing now. Can't wait to try out this fine-tuning process for my robot project.
Hello bro , after running the deepspeed script there is no file with name mm_projector.bin is generated which is required in merging process but a non_lora_trainable.bin is generated
is it possible that the model can tell you there a picture was taken(geographic), based on probability, and purely focuses on this, because you give him the information in fintune( im a beginner)
Came from TikTok! But I have no experience w AIs but am surely going to dive in to train a model for my startup application. Do you think this model could be trained to estimate macros from an image, let’s say in buckets or ranges, after identifying the food itself?
Great video:) Can you please comment on the dataset size? The one you used consists of roughly 9k samples. How many samples are needed to have a decent lora fine-tune? I've heard that with LLMs you can achieve much even with only a few examples. Is it the case for LLava as well? Please share any more information you can on the dataset creation. thanks!
Hey, I had a query regarding generating the custom dataset using gpt 4, shown at the very beginning. It seems it does not generate json file with the exact format necessary for LLaVA
Cool demo, thank you. Could you share some examples of training data? That new model is great. Can you share it on Hugingface? How big did it end up being for inference purposes,
Hey Paul! Here is the documentation for the model on hugging face: huggingface.co/docs/transformers/en/model_doc/llava Here is the training dataset: huggingface.co/datasets/Multimodal-Fatima/OK-VQA_train Here is the testing dataset: huggingface.co/datasets/Multimodal-Fatima/OK-VQA_test Note that we did not create the model, nor the training or testing datasets! We are simply using them as an example here
@brev-dev Yeah I feel so too. Finetuning is more useful when you have a novel category or more deeper classification. For example general model identifies the dog, but the fine-tuned identify the breed as well. Then you need to have the tagged dataset of dog breeds.
Hello, I followed your guides and got the error message in Gradio as error Could not parse server response: SyntaxError: Unexpected token 'I', "Internal S"... is not valid JSON What would be a matter? Could you give an advice? Thank you
@@brev-dev thanks for replying :) Carters insta story said “beginners guide” so i thought it would be a intro to machine learning or something But after seeing this video … A beginner WOULD NEVER be able to comprehend a single sentence in this video 😂😂
@@aamirshaikh2100 I am not Carter (or in any way related to the channel, lol) but if you tell me what is your starting point I can send you some resources or more specific questions, I'm not a pro yet but it may be useful if you're coming from zero
You know bro is a A level engineer when he can explain stuff soo easily
Baxate, you’re the goat. For a beginner like myself, that was a very useful video
glad you found it useful!
I tried out brev for my machine learning course, love the options for the payment system where i have the option to cut if off after X dollars, low prices, and ui looks awesome. I know it said it somewhere but it took me a minute to realize that my Jupyter notebook takes around 4 minutes to launch so for blind ppl like me I’d put some more text saying Jupyter notebook will be created in X minutes.
Love this vid and outreach- I’ll keep watching Baxate
Thank you so much for the kind comment! I will bring that feedback to the team :).
WOW!!! I'm so glad I found this information. This is incredible! I'm utilizing the llava model now in my robot project and just like you said, llava out of the box is not a good fit, especially for the computer vision/classification experiments I'm doing now. Can't wait to try out this fine-tuning process for my robot project.
Big man ting yeh. Looking good brev
Bax ate with this one!
yurrr
Best guide/insights on fine tuning I’ve seen. Subscribed 🔥
Splendid! follow your tutoial, I succeeded in training a llava on medical image, next step, however, how should we evaluate it?
Hello bro , after running the deepspeed script there is no file with name mm_projector.bin is generated which is required in merging process but a non_lora_trainable.bin is generated
is it possible that the model can tell you there a picture was taken(geographic), based on probability, and purely focuses on this, because you give him the information in fintune( im a beginner)
Thank you, this is great
bro the goat
Very useful! Subscribed indeed 🙂
For this use case, why didn't you just use prompt engineering (using a very specific prompt) to give you the same output?
Came from TikTok! But I have no experience w AIs but am surely going to dive in to train a model for my startup application. Do you think this model could be trained to estimate macros from an image, let’s say in buckets or ranges, after identifying the food itself?
yes, absolutely! That is a perfect use case
Great video:) Can you please comment on the dataset size? The one you used consists of roughly 9k samples. How many samples are needed to have a decent lora fine-tune? I've heard that with LLMs you can achieve much even with only a few examples. Is it the case for LLava as well? Please share any more information you can on the dataset creation. thanks!
Hey, I had a query regarding generating the custom dataset using gpt 4, shown at the very beginning. It seems it does not generate json file with the exact format necessary for LLaVA
man you are awesome!
Hi @Brev, Can you please help me with the configuration of the laptop.
Cool demo, thank you. Could you share some examples of training data? That new model is great. Can you share it on Hugingface? How big did it end up being for inference purposes,
Hey Paul! Here is the documentation for the model on hugging face:
huggingface.co/docs/transformers/en/model_doc/llava
Here is the training dataset:
huggingface.co/datasets/Multimodal-Fatima/OK-VQA_train
Here is the testing dataset:
huggingface.co/datasets/Multimodal-Fatima/OK-VQA_test
Note that we did not create the model, nor the training or testing datasets! We are simply using them as an example here
How can I to finetune with videos taggeds?
And how can I use LlaVa model but not with images, only with video inputs.
Wouldn't it have been simpler to feed the fluffy text to llama3 to come up with the summary?
Have you used this link? I'm reporting an error when loading the dataset now, if you can please take a look . thank you
Can you show how to fine-tune VILA models from Nvidia?
Do I have to buy credits to follow along?
How you labeled the training datset images ? Can u give a sample
i see you're finetuning LLaVA 1.5 is it possible to use this notebook for 1.6 too?
Wouldn't prompting the LLM in various scenarios in the application code be enough to get the right response? I am not clear on fine-tuning.
@brev-dev Yeah I feel so too. Finetuning is more useful when you have a novel category or more deeper classification. For example general model identifies the dog, but the fine-tuned identify the breed as well. Then you need to have the tagged dataset of dog breeds.
Hello, I followed your guides and got the error message in Gradio as
error
Could not parse server response: SyntaxError: Unexpected token 'I', "Internal S"... is not valid JSON
What would be a matter?
Could you give an advice?
Thank you
Please answer me
Answer
came from ig
Came from IG
Thank you
twin served meat 🔥🔥
The notebook please?
Is this video for elementary school ?
baxate !
Lol what clickbait headline I thought you were going to quantize the model or something , instead you're running it on 4*a100 in the cloud lol
In what world is this a “beginner friendly machine learning guide”? What💀💀💀😂😂
let me know where you struggled! I tried to explain the concepts at a high level and run the cells as they were written.
@@brev-dev thanks for replying :)
Carters insta story said “beginners guide” so i thought it would be a intro to machine learning or something
But after seeing this video …
A beginner WOULD NEVER be able to comprehend a single sentence in this video 😂😂
@@aamirshaikh2100 This is Carter :). I will keep that in mind and maybe make a dedicated intro to machine learning video!
@@brev-dev thanks for taking the time 💓
@@aamirshaikh2100 I am not Carter (or in any way related to the channel, lol) but if you tell me what is your starting point I can send you some resources or more specific questions, I'm not a pro yet but it may be useful if you're coming from zero