Image segmentation with Yolov8 custom dataset | Computer vision tutorial

แชร์
ฝัง
  • เผยแพร่เมื่อ 13 มิ.ย. 2024
  • Code: github.com/computervisioneng/...
    Download a semantic segmentation dataset from the Open Images Dataset v7 in the format you need to train a model with Yolov8: bit.ly/47PUfff
    0:00 Intro
    0:30 Dataset
    3:09 Data annotation
    13:56 Data structure
    17:26 Train
    27:47 Validate model
    40:26 Predict
    #computervision #semanticsegmentation #objectdetection #python #yolov8

ความคิดเห็น • 208

  • @WelcomeToMyLife888
    @WelcomeToMyLife888 ปีที่แล้ว +8

    Another awesome tutorial, showing all the necessary steps! Wish you all the best.

  • @dmitrium12
    @dmitrium12 ปีที่แล้ว +2

    This is a very cool manual, thank you for it, this is exactly what I wanted to see. I have always been surprised in your channel that you post all these materials and videos for free, because you could sell them.

    • @ComputerVisionEngineer
      @ComputerVisionEngineer  ปีที่แล้ว +1

      Hey, glad you find it helpful! I really enjoy sharing my knowledge of computer vision with everyone! 😃💪 Selling courses is not a bad idea, though. Maybe I will do it in the future. 😊

  • @ComputerVisionEngineer
    @ComputerVisionEngineer  6 หลายเดือนก่อน

    Did you enjoy this video? Try my premium courses! 😃🙌😊
    ● End-To-End Computer Vision: Build and Deploy a Video Summarization API bit.ly/3tyQX0M
    ● Hands-On Computer Vision in the Cloud: Building an AWS-based Real Time Number Plate Recognition System bit.ly/3RXrE1Y
    ● Machine Learning Entrepreneur: How to start your entrepreneurial journey as a freelancer and content creator bit.ly/4bFLeaC
    All my premium courses are available to the Computer Vision Experts in my Patreon. 😉
    www.patreon.com/ComputerVisionEngineer

  • @vishalpahuja2967
    @vishalpahuja2967 ปีที่แล้ว

    Wow! was waiting for this video.
    Thank you!

  • @AlexFernando1996
    @AlexFernando1996 4 หลายเดือนก่อน

    I must say that only your videos are helping me with computer vision project. All others do not work. Thank you, from Serbia

  • @sugarbycand2845
    @sugarbycand2845 ปีที่แล้ว

    Great tutorial, everything explained very well. You saved me :D

  • @user-ke8ll1to7b
    @user-ke8ll1to7b ปีที่แล้ว

    Thank you so much for your kindness tutorial! Hope you everything well!!!!!!

  • @FazriGading
    @FazriGading หลายเดือนก่อน

    Wow. you are amazing bro. Thank you so much for teaching me this!!!

  • @adhammahmoud5574
    @adhammahmoud5574 23 วันที่ผ่านมา

    Thank you so much for this wonderful video. It helped me so much.

  • @user-xe9mw8uo8t
    @user-xe9mw8uo8t 9 หลายเดือนก่อน

    Hi! thanks for the video, it's helping too much!!!
    I have a question,
    What version of tensorboard and numpy do you use?

  • @4Tjohny
    @4Tjohny 9 หลายเดือนก่อน

    Thank you. Awesome tutorial.

  • @aakashbhosale9140
    @aakashbhosale9140 11 หลายเดือนก่อน

    Thanks for the detailed explanation.

  • @bbb-xu7wx
    @bbb-xu7wx ปีที่แล้ว +1

    thank you for the video again amazing job

  • @neptunelearning9249
    @neptunelearning9249 3 หลายเดือนก่อน

    Nice class sir. You explained to finetune YOLO in a simple way. Thank you

  • @mohamedkeddache4202
    @mohamedkeddache4202 5 หลายเดือนก่อน

    thanks you are very helpful, keep going bro

  • @simonbaumgartner6612
    @simonbaumgartner6612 หลายเดือนก่อน

    Great video! I wonder if you could speak to the data vizualization. How do you create the masks? And say I only have the Yolov8 annotation format (.png and corresponding .txt), any recommendation on how to visualize it by chance?

  • @user-gl4eb6nw4f
    @user-gl4eb6nw4f 10 หลายเดือนก่อน

    When I'm exporting the annotated files using segmentation mask 1.1, the zip I'm getting is just a single text file. Any idea what else I can do?

  • @ahmadfaraz1288
    @ahmadfaraz1288 3 หลายเดือนก่อน

    Hey, thanks for the tutorials. I am new to computer vision. Currently, I am preparing a data set on the dry wall construction process. We have to detect dry wall stages like: Stud Installation, Gypsum Panelling, electrical works and plastering. However, I have some confusion about labelling the data. My question is: do I need to label each object in the image at the same time? Or should I focus on a single object in each image? Besides, we have only 250 images from the construction sites; are these enough for training?

  • @vladimirkuzmenkov7285
    @vladimirkuzmenkov7285 7 หลายเดือนก่อน

    Good stuff, thanks

  • @inquisitiverakib5844
    @inquisitiverakib5844 11 หลายเดือนก่อน

    how can we get .json file from this annotated image which will carry the co-ordinates of the polygon mask as text format???

  • @amjoode2
    @amjoode2 4 วันที่ผ่านมา

    When creating the labels, do you devide by the dimensions of the mask or to the whole image?
    I am trying to adapt your label creation process to handle multiple masks in one image.

  • @logos1396
    @logos1396 ปีที่แล้ว

    Amazing video 👍🏻👍🏻👍🏻

  • @aurum9864
    @aurum9864 ปีที่แล้ว +1

    Thanks for the video! Do you have a masks_to_polygons script that would also work for multiple segmentation classes? Or do you know where I would find one? Have been looking for ages..

    • @ComputerVisionEngineer
      @ComputerVisionEngineer  ปีที่แล้ว

      I don't have a multiclass masks_to_polygons script, but I think you could create one taking my one class script as baseline. Maybe chatgpt can help you adapting the script to multiclass. 💪💪

    • @quillaja
      @quillaja 6 หลายเดือนก่อน

      I'd annotate my images in Inkscape or Illustrator, using paths as the masks, save it to SVG, then just convert SVG to YOLO format. Straight text-to-text conversion, more or less. All the info you need to normalize the vertices is in the SVG. To create multiple classes, you could group the classes, or probably the better thing would be to give each path a custom xml attribute for the object class.

  • @pavanmhamsa
    @pavanmhamsa หลายเดือนก่อน

    Thanks a Lot

  • @zaidahmed4069
    @zaidahmed4069 5 หลายเดือนก่อน

    Can you help me with exporting this model because I'm facing an error while exporting the model into tflite or pb format

  • @user-be7cb6cn9y
    @user-be7cb6cn9y ปีที่แล้ว

    Great job, man! Thanks a lot! Btw, does this segmebtation project work in c2.capread? Or can i use it for segmentation objects in video ?

    • @ComputerVisionEngineer
      @ComputerVisionEngineer  ปีที่แล้ว

      C2.capread? What do you mean?

    • @user-be7cb6cn9y
      @user-be7cb6cn9y ปีที่แล้ว

      @@ComputerVisionEngineer oh, sorry, im mistaken. Does it works with cv2.VideoCapture?

    • @ComputerVisionEngineer
      @ComputerVisionEngineer  ปีที่แล้ว

      Sure! You can read frames using cv2.VideoCapture and then input each frame into the model to get the mask 🙌

    • @user-be7cb6cn9y
      @user-be7cb6cn9y ปีที่แล้ว

      @@ComputerVisionEngineer thanks for getting back to me! So I'll try to use it on livestream)

  • @joshteixeira6750
    @joshteixeira6750 10 หลายเดือนก่อน

    You are a hero

  • @guillemcobos1987
    @guillemcobos1987 ปีที่แล้ว

    Hello! I found your video very interesting, and it's helping me a lot in my new job as a vision engineer. I managed to train the duck-segmenting algorithm following your steps - amazingly clear! I can see how it makes some batch predictions in the 'runs' folder for some of the images in 'val'. However, when I import the model 'last.pt' and I try to make predictions, I consistently get 'no detections' and 'masks: None'. Do you know what could be going on? Thanks a million😊

    • @ComputerVisionEngineer
      @ComputerVisionEngineer  ปีที่แล้ว

      Hey Guillem, I am glad the video is helping you in your new job! 😃 How many epochs did you train the model for? Are you using the exact same dataset as I use in the video?

    • @sarthakdas815
      @sarthakdas815 9 หลายเดือนก่อน

      @@ComputerVisionEngineer Hi getting the same error as stated above. No masks are gnerated for the predictions after 10 ephors using the same data set and code you have given not sure whats going wrong.

    • @ebrahim-nourmohammadi
      @ebrahim-nourmohammadi 5 หลายเดือนก่อน

      @@sarthakdas815 I faced the same issue but solved it. That was because I used the masks that are in the "SegmentationObject" folder, we should use the masks that are in the "SegmentationClass" folder.

  • @rosemutegi8830
    @rosemutegi8830 6 หลายเดือนก่อน

    Getting this error what could be the issue.
    IndexError: list index out of range when i train my .yaml file

  • @lewislord9341
    @lewislord9341 10 หลายเดือนก่อน

    please sir i have an error and i didnt find the solution
    its not possible to find the images, i dont know if the problem is on my config file. i have a big problem with it

  • @shafagh_projects
    @shafagh_projects 5 หลายเดือนก่อน

    Thank you so much for your lovely content. Indeed, it is very informative. However, I have a question about data handling. I noticed in the images folder you shared, the duck images are in both the train and val folders. Shouldn't it be that the train folder contains only duck images and the val folder is with non-duck images? Looking forward to your clarification. Thanks!

    • @ComputerVisionEngineer
      @ComputerVisionEngineer  5 หลายเดือนก่อน +1

      The val folder contains the validation data, this is how you validate the model. If the model detects ducks, it's appropriate to use images with ducks as validation data.

    • @shafagh_projects
      @shafagh_projects 5 หลายเดือนก่อน

      many thanks for your prompt response but I have a big challenge of using a webcam to detect ducks using the following line:
      model.predict(source=0, show=True, conf=0.2)
      it has a huge lag.
      can you help me how to resolve this to be real-time detection?

  • @faizalbarrisi7254
    @faizalbarrisi7254 ปีที่แล้ว +3

    can you give me a step, on how to download the dataaset from the openimages and the annotation mask

    • @ialbornoz
      @ialbornoz 8 หลายเดือนก่อน

      I have the same question

  • @enricobovo2184
    @enricobovo2184 ปีที่แล้ว

    Hi and thank you for your video! I have noticed that, when annotating a dataset of multiple images with CVAT with two labels, the export phase goes wrong and not all the segmentation masks are created correctly. Some of them contain for example just one class of objects even though I had previously annotated objects of different classes in that picture. Do you know how to solve it? Is there any other annotation tool that allows to export the images of the segmented masks?
    Thank you

    • @ComputerVisionEngineer
      @ComputerVisionEngineer  ปีที่แล้ว

      Hey, I just tried to do it and everything seems fine with a couple of images. I annotated two labels and exported it as 'Segmentation mask 1.1', are you using this export format?

    • @mithactatus
      @mithactatus 10 หลายเดือนก่อน

      you can just manually correct the .txt files. first number in the file represents the class. all of them might be 0 in your case.

  • @Abhinavnair1103
    @Abhinavnair1103 หลายเดือนก่อน

    hey, your way is working perfectly!!, but when I am taking multiple objects it is classifying all of them as one label. I believe the problem is in masks_to_polygon.py, I did the same thing as instructed by you for config.yaml. Can you tell me where I can be wrong

  • @lopezbryan7589
    @lopezbryan7589 3 หลายเดือนก่อน

    please , how can i get the code at 11:40 in the video? type in by myself?

  • @muhteguhadhiputra6041
    @muhteguhadhiputra6041 ปีที่แล้ว

    penjelasan yang bagus👍👍

  • @alphagenerativeai
    @alphagenerativeai 2 หลายเดือนก่อน

    How can I labeled 2 classes and configured training. We hope to receive feedback from you! Thank you, from Viet Nam

  • @mpfmax0
    @mpfmax0 ปีที่แล้ว

    would the segment anything model be better for this task? I'm trying to segment plants from a herbarium collection, they are full dried plants pressed on white paper sheets and scanned into digital images, but there is a paper label with collection data and a stamp getting in the way of my automatic segmentation attempts. Im a bit confused on what would be the best method to accomplish the task of extracting the plant from the background ( I also may want to segment pieces of the plant, like leaves, flowers, stems). So far it seems to me the best method would be to train YOLO to detect the plant and draw a bounding box around it and then use SAM to make a mask of the plant inside the box (or multiple masks for the pieces of the plant) . Does this make sense?

    • @ComputerVisionEngineer
      @ComputerVisionEngineer  ปีที่แล้ว

      It makes sense, although you would need a yolo model trained on detecting plants, do you have one?

    • @mpfmax0
      @mpfmax0 11 หลายเดือนก่อน

      @@ComputerVisionEngineer I do now. I trained it to draw bounding boxes around the plants using your other tutorial video, it performs really well. Now I'm going to try to use the bounding boxes as prompts for SAM (segment anything model) to extract detailed masks of the plants. Wish me luck!

    • @ComputerVisionEngineer
      @ComputerVisionEngineer  11 หลายเดือนก่อน

      @@mpfmax0 Good luck! Let me know how it goes! 😃

  • @penguinie4325
    @penguinie4325 ปีที่แล้ว

    hello! thank you for your video! I have a question regarding using the prediction to predict segmentation from the image. From my results, It states it indicates 2 ducks in my image (which has 2 ducks) however, the outcome only displays 1 image segmentation. What should I do if I want both image segmentations to be predicted? Thank you!

    • @ComputerVisionEngineer
      @ComputerVisionEngineer  ปีที่แล้ว

      Hey, when you say the outcome only displays 1 image segmentation you mean it only covers one of the two ducks?

    • @penguinie4325
      @penguinie4325 11 หลายเดือนก่อน

      ​@@ComputerVisionEngineer yes, can the code detect 2 ducks instead? or is it only for one duck segmentation detection...

  • @usamasherazi4598
    @usamasherazi4598 7 หลายเดือนก่อน

    can i use CVAT Anotation website for yoloV5

  • @yassinebouchoucha
    @yassinebouchoucha 7 หลายเดือนก่อน

    @18:50 starting with `epoch=1` to make sure everything is well established is a Golden Rule !

  • @nomaanrizvi6561
    @nomaanrizvi6561 5 หลายเดือนก่อน

    i need the code for polygons to mask...can i get it?..please

  • @user-ww2zf9tq6o
    @user-ww2zf9tq6o 6 หลายเดือนก่อน

    I have a question, I see you used the masked label (txt) data for training, What is the process to train the model directly using mask and original samples without any txt data on YoloV8? I have mask image but don't have any text data.

    • @ComputerVisionEngineer
      @ComputerVisionEngineer  6 หลายเดือนก่อน +1

      You need to convert the masks into the txt files in order to train the model with yolov8. I have a Python script in this project's github repository that may help you to do that. 🙌

    • @mdfaysalahamed4605
      @mdfaysalahamed4605 6 หลายเดือนก่อน

      Thank you brother. I found the file named as "masks_to_polygons". ❤

  • @govindaagrawal816
    @govindaagrawal816 ปีที่แล้ว

    Just a query, If I wanted to train it on multiple classes, how would I go about editing the config file?

    • @ComputerVisionEngineer
      @ComputerVisionEngineer  ปีที่แล้ว

      Edit the 'nc' field to your number of classes, and edit the 'names' field so it contains all your class names. In case of multiclass segmentation you also need to edit your masks. 🙌

  • @zeinabelsharkawy9014
    @zeinabelsharkawy9014 ปีที่แล้ว

    Hi, thank you for this video. can you convert the Yolo label to a binary mask?

  • @mscussiatto
    @mscussiatto 7 หลายเดือนก่อน

    Thank you for the tutorial, it's one of the best i've seen in yolo. Would you be able to provide me some support on how to get the RGB masks from the inferences crop results? Cheers

    • @ComputerVisionEngineer
      @ComputerVisionEngineer  7 หลายเดือนก่อน

      Do you mean how to crop the original rgb image in the region given by the predicted mask?

    • @anisiobinzubechi
      @anisiobinzubechi 5 หลายเดือนก่อน

      This is one of the most informative I have seen on this topic. Yes, I would also like to know how to crop out the original region given by the predicted mask. I have been having a hard time with that
      @@ComputerVisionEngineer

    • @ComputerVisionEngineer
      @ComputerVisionEngineer  5 หลายเดือนก่อน

      @@anisiobinzubechi I will try to make a video about it soon.

  • @revuriakhil3148
    @revuriakhil3148 8 หลายเดือนก่อน

    Is it mandatory to convert masks to polygon or we can directly do labeling in polygon template and can we train that

    • @ComputerVisionEngineer
      @ComputerVisionEngineer  8 หลายเดือนก่อน

      Converting masks to polygon is necessary in order to do semantic segmentation with yolov8. 🙌

  • @khoahuynh2809
    @khoahuynh2809 6 หลายเดือนก่อน

    My dataset has two classes and after using your python file to convert, I found that it just has only one class in txt file (class which labels 0) although in the image has clearly two objects in two classes. How can I fix this error?

  • @samb23692
    @samb23692 หลายเดือนก่อน +1

    HI, I have a project wherein, I have to segment multiple classes, how do i go about it? What changes do I need to make in the code?

    • @user-uk4xc7rr7o
      @user-uk4xc7rr7o 2 วันที่ผ่านมา

      actually this one bothers me to

  • @juliogomez6065
    @juliogomez6065 ปีที่แล้ว +1

    Hola Felipe, estuve trabajando sobre los archivos que nos compartes, lo adapté a mis necesidades, previamente hice todo el etiquetado en CVAT, pero me queda una duda ya que el training al parecer no me está funcionando: En el archivo "config.yaml", hay dos líneas que no explicaste: "nc:1" (que supongo es la cantidad de classes generadas en CVAT, y la línea "names:['...'] (Supongo que son los nombres asignados a las classes en CVAT). El problema es que asumiendo esto, lo adapto a mi necesidad (nc:7 names: ['Sin arandela', 'Arandela OK', 'Arandela rota', ...ETC ], y en el archivo run - weights, con el collage de imágenes que me arroja, solo me aparece la primera etiqueta, siendo "Sin arandela". ¿Es posible que me digas qué puedo estar haciendo mal? Hice 100 epochs, donde en TRAIN tengo 187 fotografías, y en VAL tengo 46 fotografías.

    • @ComputerVisionEngineer
      @ComputerVisionEngineer  ปีที่แล้ว

      Hola Julio, 187 + 46 imagenes parecen pocas para entrenar un algoritmo de este tipo, especialmente considerando que tenes 7 clases. Adaptaste las mascaras para trabajar con 7 clases?

    • @juliogomez6065
      @juliogomez6065 ปีที่แล้ว +1

      @@ComputerVisionEngineer Hola, Felipe. Cuantas consideras que pueden ser una buena base para desarrollar un buen script? El problema es que en este momento soy únicamente yo en el proyecto, por lo que no puedo tener una cantidad muy grande, al menos hasta demostrar resultados y que me asignen una persona adicional. Sobre adaptar las máscaras, no entiendo a qué te refieres, apenas inicio en este mundo. No sé si responda tu pregunta, pero a lo largo del etiquetado en CVAT, hice uso de todas las etiquetas, habiendo aproximadamente de 2 a 3 etiquetas por imagen.

    • @ComputerVisionEngineer
      @ComputerVisionEngineer  ปีที่แล้ว

      @@juliogomez6065 El tutorial de este video es para semantic segmentation de una sola clase. Para segmentación multi-clase hay que estudiar la documentación de yolov8 para ver cómo hacer las máscaras. Las máscaras que usé en el video son binarias y solo sirven para segmentación de una sola clase (blanco=objeto, negro=fondo). Sobre la cantidad de imágenes, todo depende... pero te sugiriía al menos unas miles, por ejemplo en este video uso ~3000 imágenes para hacer segmentación de una sola clase.

  • @user-fv9mr1cy3s
    @user-fv9mr1cy3s 10 หลายเดือนก่อน

    thanks for the response earlier, how can I expend it to a data set with multiple segmentation classes?

    • @ComputerVisionEngineer
      @ComputerVisionEngineer  10 หลายเดือนก่อน

      You would need to edit the config file and create the annotations accordingly. I may do a video about multiclass detection in the future.

  • @user-rj1pf9xl5e
    @user-rj1pf9xl5e 10 หลายเดือนก่อน +1

    Hi, I was just wondering how you installed the specific segmentation masks of ducks from open images, I couldn't figure it out on my own. I would greatly appreciate it if you lend me a hand

    • @ComputerVisionEngineer
      @ComputerVisionEngineer  10 หลายเดือนก่อน

      Hey, take a look at the 'Annotations and metadata' section of this page storage.googleapis.com/openimages/web/download_v7.html 🙌

    • @user-rj1pf9xl5e
      @user-rj1pf9xl5e 10 หลายเดือนก่อน

      @@ComputerVisionEngineer Should I only download the mask data? If so, after that have you written a script which creates a list of the image ID's and extracts the necessary annotations. Like in your object detection video, could you share it with me?

    • @user-rj1pf9xl5e
      @user-rj1pf9xl5e 10 หลายเดือนก่อน

      I have resorted to another solution after I couldn't figure out the exact thing I wanted to make: I downloaded all the zip files containing all the masks, all mask files have a notation like __.png for example: 00a3d94534a1b356_m0k4j_97c014cd.png I then wrote a script in python to filter only the masks with the wanted class label utilising multiprocessing and multithreading, the script saves the files as .png omitting the other parts. After running the script, I was able to obtain the binary masks only for my selected class. Now I'm going to annotate them using the polygonization script which you've provided, I might modify it a little. Anyway, thanks for your help!

    • @ComputerVisionEngineer
      @ComputerVisionEngineer  10 หลายเดือนก่อน

      @@user-rj1pf9xl5e Glad you solved it!

    • @abhisheknegi2888
      @abhisheknegi2888 23 วันที่ผ่านมา

      @@user-rj1pf9xl5e hey can you please share the script

  • @santarosajoe2
    @santarosajoe2 4 หลายเดือนก่อน

    it might seem simple to you, but I'd love to see just one example of utilizing a webcam.
    It seems using the webcam is the most interesting use of YOLO, instead of cycling thru a bunch of still jpg images.

    • @ComputerVisionEngineer
      @ComputerVisionEngineer  4 หลายเดือนก่อน

      Ok, thank you for your feedback, next time I work with Yolo I will use a Webcam. 🙌

  • @gizem3166
    @gizem3166 10 หลายเดือนก่อน

    Hi, thanks for the video, its helping too much. Can we crop the segmented object instead of taking a mask? Do you have another video for this?

    • @ComputerVisionEngineer
      @ComputerVisionEngineer  10 หลายเดือนก่อน

      Hey, yes it is possible. I don't have another video for that.

  • @muhammadgulfam1869
    @muhammadgulfam1869 2 หลายเดือนก่อน

    In the config file, what is is "nc" variable referring to? Also what are the instructions for the "names" variable, does it contains object names? do we decide what we want to call the object?

    • @ComputerVisionEngineer
      @ComputerVisionEngineer  2 หลายเดือนก่อน

      nc is the number of classes, yes you can name the objects whatever you want, the names won't affect the training process 🙌

    • @muhammadgulfam1869
      @muhammadgulfam1869 2 หลายเดือนก่อน

      @@ComputerVisionEngineer Thank you

  • @craftman147100
    @craftman147100 11 หลายเดือนก่อน

    I noticed that images\train has around 1800 files - unlike in the video, and labels\train has 3965 files. Is that an issue?

    • @ComputerVisionEngineer
      @ComputerVisionEngineer  11 หลายเดือนก่อน

      Hey, you should have the same number of images and label files. Yolov8 will probably trigger an error in any other case.

  • @Jugeenias
    @Jugeenias 11 หลายเดือนก่อน

    What could be the reason for the following issue:
    for j, mask in enumerate(result.masks.data):
    AttributeError: 'list' object has no attribute 'masks'

    • @ComputerVisionEngineer
      @ComputerVisionEngineer  11 หลายเดือนก่อน

      If no detections were found result may be an empty list. Print result and see how it looks like. Let me know how it goes. 🙌

  • @saikiran3964
    @saikiran3964 ปีที่แล้ว

    Is there any reference on how to save the segmented object as it's own image?

  • @jacobmars1902
    @jacobmars1902 หลายเดือนก่อน

    could i use A111 inpaint anything to create the masks for the training?

    • @ComputerVisionEngineer
      @ComputerVisionEngineer  หลายเดือนก่อน

      Not sure if I understand, but no, I don't think you could do that.

    • @jacobmars1902
      @jacobmars1902 หลายเดือนก่อน

      @@ComputerVisionEngineer ok, thanks

  • @jesusmtz29
    @jesusmtz29 8 หลายเดือนก่อน

    do you need an account with cvat even when you host it locally?

    • @ComputerVisionEngineer
      @ComputerVisionEngineer  8 หลายเดือนก่อน

      If I am not mistaken when hosted locally you don't need an account with cvat, but each user needs to create an account in your locally hosted cvat app. 🙌

  • @sto2779
    @sto2779 7 หลายเดือนก่อน

    8:33 - Its called duck feet. The arms of the duck is the wings.

    • @ComputerVisionEngineer
      @ComputerVisionEngineer  7 หลายเดือนก่อน +1

      Oh right the arms are the wings. 'Duck feet', ok, noted. Thank you! 🙌

  • @chispun2
    @chispun2 ปีที่แล้ว

    That part of the duck is just called the webbed foot (palmeado). I have some questions I would like to ask you, computer vision related, because I don't know who else could I ask

    • @ComputerVisionEngineer
      @ComputerVisionEngineer  ปีที่แล้ว +1

      Oh, webbed foot! 🦆 Cool, thank you!
      Sure, you can ask me on discord. 💪🙌

    • @chispun2
      @chispun2 ปีที่แล้ว

      @@ComputerVisionEngineer No sabía lo del discord! Allá voy! Gracias!!

  • @stevegreen5638
    @stevegreen5638 8 หลายเดือนก่อน

    what is speed on segmentation, will I be able to use it on live video?

    • @ComputerVisionEngineer
      @ComputerVisionEngineer  8 หลายเดือนก่อน

      You can make it work on ~real time if not mistaken, let me know how it goes. 🙌

  • @vikashkumar-cr7ee
    @vikashkumar-cr7ee ปีที่แล้ว

    Dear Tutor,
    Greetings! I am getting the following error while running the code at time stamp 44:35
    'for j, mask in enumerate(result.masks.data):
    AttributeError: 'NoneType' object has no attribute 'data'
    Can you please help me out ?

    • @ComputerVisionEngineer
      @ComputerVisionEngineer  ปีที่แล้ว

      Hey, it may be that you are not detecting any objects in that image. Have you tried with other images?

    • @vikashkumar-cr7ee
      @vikashkumar-cr7ee ปีที่แล้ว

      @@ComputerVisionEngineer I tried on three or four other images in val set, but I am getting the same error

    • @user-ip6kr5bo5f
      @user-ip6kr5bo5f ปีที่แล้ว

      Same problem for me

  • @santarojoe1
    @santarojoe1 5 หลายเดือนก่อน

    and would be cool if had small cv2 script that you waive a picture of a duck under cam, and it highlights the duck.

  • @I77AGIC
    @I77AGIC ปีที่แล้ว

    But how do we export from cvat into yolo format if we have more than just one class?

    • @ComputerVisionEngineer
      @ComputerVisionEngineer  ปีที่แล้ว

      I will try to make a video about multi class image segmentation in the future 🙌

    • @I77AGIC
      @I77AGIC ปีที่แล้ว

      I ended up exporting as COCO then using roboflow to convert to YOLO format. Now I'm just using roboflow to annotate. I like it better and it actually lets me export to YOLOv8.

  • @fawazmirza4646
    @fawazmirza4646 10 หลายเดือนก่อน

    I followed everything exactly but for some reason my val_batch0_pred has no segmentations on it. Even though the val_batch0_labels is segmented perfectly. I think this is probably the reason why I'm getting "AttributeError: 'NoneType' object has no attribute 'data'" when I try running the code. The object I'm trying to detect and the images given are very simple and easy, the model should not be struggling with this at all. What can I do?

    • @ComputerVisionEngineer
      @ComputerVisionEngineer  10 หลายเดือนก่อน

      Hey, are you using the same dataset as I did in the video? How many epochs are you training?

    • @fawazmirza4646
      @fawazmirza4646 10 หลายเดือนก่อน

      @@ComputerVisionEngineer no I'm using my own data set which is alot smaller, because I don't have many images of the thing I'm trying to detect, because it's of a proprietary ph indicator test so not many images exist, and so getting more is not an option.
      I have 6 images for training, and 4 for validation. I tried with 10, 50 and 100 epochs but still not a single detection on val_batch0_pred

    • @fawazmirza4646
      @fawazmirza4646 10 หลายเดือนก่อน

      @@ComputerVisionEngineer I've seen other people on github who have had more images and everything have the same issue, but non of them really got an answer. Or at least not one that is relevant in my case.
      My validation pictures and very similar to the training ones so the model should have no issues, idk what's wrong.

    • @ComputerVisionEngineer
      @ComputerVisionEngineer  10 หลายเดือนก่อน

      @@fawazmirza4646 oh I see. 10 images is usually not enough to train this type of model. Training for that many epochs on 6 images will produce overfitting.

    • @fawazmirza4646
      @fawazmirza4646 10 หลายเดือนก่อน

      So what do you suggest I do with the small data I have? What machine learning method, if any should I try? Or is there a way to make yolov8 work for my case?

  • @mithactatus
    @mithactatus 10 หลายเดือนก่อน +1

    I am predicting watermelons, pineapples and blackberries. My model can predict the objects but call them all watermelons. Do you have any idea?

    • @ComputerVisionEngineer
      @ComputerVisionEngineer  9 หลายเดือนก่อน

      Take a look at your training data. Perhaps you need to train the model with more data. 🙌

    • @mithactatus
      @mithactatus 9 หลายเดือนก่อน

      i fixed it. it was because of the mask transformation to yolo files. the txt files had all 0 as the class (the very first number of the .txt file). and i manually changed 0's according to the images with the same name @@ComputerVisionEngineer

  • @OceanAye
    @OceanAye 8 หลายเดือนก่อน

    Thanks a lot for the tutorial, however, I seem to run into the same problems as @dmitrium12. Somehow the runs/segment/train file does not mask predictions and thus the graphs with train/loss and val/loss is just a dot in the middle of the grapth.
    I have used your dataset and followed every step.

    • @OceanAye
      @OceanAye 8 หลายเดือนก่อน

      sorry meant @guillemcobos1987

    • @ComputerVisionEngineer
      @ComputerVisionEngineer  8 หลายเดือนก่อน

      If evaluation plots are only a dot in the middle of the graph it means you are training for only 1 epoch. Increase the number of epochs and you should be able to see a different plot. 🙌

  • @larafischer420
    @larafischer420 6 หลายเดือนก่อน

    I have a question: How do you download the images from google datasets? Can you make a video explaining that process? Seems like a dumb process, but I really don't know how to do that

    • @ComputerVisionEngineer
      @ComputerVisionEngineer  6 หลายเดือนก่อน

      I am currently preparing a Python script to download a semantic segmentation dataset from the google open images dataset. It will be available in my Patreon soon. 🙌

    • @larafischer420
      @larafischer420 6 หลายเดือนก่อน

      @@ComputerVisionEngineer

  • @adithin740
    @adithin740 หลายเดือนก่อน

    i am not able to download dataset

  • @kagadevishal5008
    @kagadevishal5008 6 หลายเดือนก่อน

    what if there are more than 1 classes, will the same method to convert to polygon will work?

    • @ComputerVisionEngineer
      @ComputerVisionEngineer  6 หลายเดือนก่อน +1

      If there are more than one classes, the same script will not work, you would need to adjust it to deal with multi class masks. 🙌

    • @kagadevishal5008
      @kagadevishal5008 6 หลายเดือนก่อน

      @@ComputerVisionEngineer any Idea how can we do that, I tried but not able to find concrete solution.

  • @marsrover2754
    @marsrover2754 10 หลายเดือนก่อน

    @ComputerVisionEngineer Can you make a video on finetune SAM model(Segment Anything Model) on custom dataset.

  • @thisurawz
    @thisurawz ปีที่แล้ว +1

    hello, im new to computer vision and I have a question. what is the most suitable algorithm/s or method/s for image steganalysis to detect the changed pixels in the stego image? i want to segment only the changed pixels in the stego image? can I use semantic segmentation also for this kind of problem?

    • @ComputerVisionEngineer
      @ComputerVisionEngineer  ปีที่แล้ว

      Hey, I don't think that is a problem you can solve with semantic segmentation 🤔, but you can try! 😃 Regarding what are the most suitable methods for image steganalysis, I recommend you do a {Google, Github, Google Scholar} search, it is a field I haven't been involved in. 💪💪

    • @thisurawz
      @thisurawz ปีที่แล้ว

      @@ComputerVisionEngineer Okay, i'll search. thank you. btw, I really appreciate your effort in making really valuable videos related to CV for free. I learned so much from your channel. this is one of the best channels with real-world implementations for CV that I've seen on TH-cam. keep going 💪💪!!

    • @ComputerVisionEngineer
      @ComputerVisionEngineer  ปีที่แล้ว

      @@thisurawz 😃 Thank you so much for your support! 💪🙌

  • @lobo5727
    @lobo5727 4 หลายเดือนก่อน

    no attribute called data... Error

  • @santarojoe1
    @santarojoe1 5 หลายเดือนก่อน

    Very well done , however my code still errors that .
    The only things you aren't explaining very well is WHAT goes inside each of the directories. You've explained that its " images to train the model" and " to validate the model", however I cant tell if IMAGES\TRAIN contains
    1) images the masks were trained from
    2) images of the masks
    3) or bulk unknown images to be analyzed
    IMAGES\VAL
    you said contains "images to validate the training model" , however I dont know which images those might be - 1)original bulk of all ducks known and unknown, 2) images of masks 3) or just the images that were used to create the masks)
    ...all same above questions with
    LABELS\TRAIN
    LABELS\VAL (you never mentioned that any files are inserted into this directory, or if this is the output )
    Then,
    are any of the folders we created empty?
    finally, it would be great to see the results found in your runs\detect\train folder.

    • @ComputerVisionEngineer
      @ComputerVisionEngineer  5 หลายเดือนก่อน

      Hi, you can download the data from the github repository if I am not mistaken.

  • @vishalpahuja2967
    @vishalpahuja2967 ปีที่แล้ว

    Hi,
    Can you make a video of getting output such as on your thumbnail?
    Thank you!

    • @ComputerVisionEngineer
      @ComputerVisionEngineer  ปีที่แล้ว +1

      Hey, next time I make a video about semantic segmentation I will make the output to look like that 💪

  • @user-ip6kr5bo5f
    @user-ip6kr5bo5f ปีที่แล้ว

    Hello my weights folder is empty how can I overcome this problem it may be an issue of my train coding

    • @ComputerVisionEngineer
      @ComputerVisionEngineer  ปีที่แล้ว +1

      Probably your training process is not being completed. Do you see any error? Have you tried to train the model from a google colab?

    • @user-ip6kr5bo5f
      @user-ip6kr5bo5f ปีที่แล้ว

      @@ComputerVisionEngineer thank you very much no I didn't try using colab I used Spyder

    • @user-ip6kr5bo5f
      @user-ip6kr5bo5f ปีที่แล้ว

      Thank you I used colab and I got the weights

    • @user-ip6kr5bo5f
      @user-ip6kr5bo5f ปีที่แล้ว

      I trained model in colab then I downloaded but when I use the Spyder for prediction using that last weight model got an error called no attribute 'data' how can I solve this

    • @user-ip6kr5bo5f
      @user-ip6kr5bo5f ปีที่แล้ว

      Hello can we use Spyder for this without using colab

  • @akifakbulut765
    @akifakbulut765 ปีที่แล้ว

    Each frame reading interval is 53 milliseconds in ultralytics, how can we reduce this interval to 33 milliseconds.

    • @ComputerVisionEngineer
      @ComputerVisionEngineer  ปีที่แล้ว

      Hey, do you mean the inference is taking 53 ms per frame?

    • @akifakbulut765
      @akifakbulut765 ปีที่แล้ว

      @@ComputerVisionEngineer Yes, the elapsed time between each frame is 53 milliseconds

    • @ComputerVisionEngineer
      @ComputerVisionEngineer  ปีที่แล้ว

      @@akifakbulut765 are you using a GPU?

    • @akifakbulut765
      @akifakbulut765 ปีที่แล้ว

      @@ComputerVisionEngineer No

    • @ComputerVisionEngineer
      @ComputerVisionEngineer  ปีที่แล้ว

      @@akifakbulut765 using a GPU would be a good way to try to reduce the execution time, if you don't have a GPU in your local computer you could consider using something like an EC2 instance from AWS.

  • @joydeepkundu509
    @joydeepkundu509 ปีที่แล้ว

    Can anyone share the dataset?

  • @oi4252
    @oi4252 ปีที่แล้ว

    have you seen meta's SAM

    • @ComputerVisionEngineer
      @ComputerVisionEngineer  ปีที่แล้ว +1

      I have! Although I haven't tested so far. I should make a video about it later on! 💪

    • @oi4252
      @oi4252 ปีที่แล้ว

      @@ComputerVisionEngineer thank you for your videos. they are truly amazing and very very educative and fun to watch

    • @ComputerVisionEngineer
      @ComputerVisionEngineer  ปีที่แล้ว +1

      @@oi4252 😊 I am so happy you enjoy them!

  • @vishalpahuja2967
    @vishalpahuja2967 ปีที่แล้ว +1

    Here if we want to infer on an image , so how to do it?
    I tried doing:
    from ultralytics import YOLO
    # Load a model
    model = YOLO('/content/yolov8n-seg.pt') # load an official model
    model = YOLO('/content/runs/segment/train/weights/best.pt') # load a custom model
    # Predict with the model
    results = model('image.jpg') # predict on an image
    Output:
    image 1/1 /content/gdrive/MyDrive/segmentation/data/images/train/11-03-22-ROHAN SANGHVI-DAUGHTER_S BEDROOM_page-0001.jpg: 480x640 (no detections), 10.6ms
    Speed: 0.6ms preprocess, 10.6ms inference, 0.5ms postprocess per image at shape (1, 3, 640, 640)
    it ran successfully but cannot see where is it saved.
    if the code is wrong pls update on how i can change it for inferencing on a single image?
    Thank you.

    • @ComputerVisionEngineer
      @ComputerVisionEngineer  ปีที่แล้ว

      Hey, take a look at the tutorial. In the last chapter I show you how to make predictions with the model you trained. 🙌

    • @vishalpahuja2967
      @vishalpahuja2967 ปีที่แล้ว +1

      @@ComputerVisionEngineer yes after doing that masks can be seen of that shapes , what to do if i want want my segmentation on my test image or actual image so that bounding box can be seen in my output with segmented part?

    • @ComputerVisionEngineer
      @ComputerVisionEngineer  ปีที่แล้ว +1

      @@vishalpahuja2967 I see, you would like a visualization as the one in the thumbnail, right? img + mask on top + bounding box, is that it? you can visualize the mask on top of the image by applying an overlay, take a look on how to do that, and about the bounding box take a look at my video on object detection with yolov8 + object tracking, the first part is about how to get bounding boxes with a yolov8 model and how to draw the bounding box on the image. 💪🙌

  • @I77AGIC
    @I77AGIC ปีที่แล้ว

    great info but this could have easily been done in 15-20 minutes and still covered all the necessary info. Otherwise awesome job

  • @afolabiowoloye804
    @afolabiowoloye804 4 หลายเดือนก่อน

    Hi, nice video, you hit it.
    Please I have an issue with the prediction
    After running the prediction code, I got this
    "AttributeError"
    "---> 13 for j, mask in enumerate(result.masks.data):"
    "'NoneType' object has no attribute 'data'"

    • @ComputerVisionEngineer
      @ComputerVisionEngineer  4 หลายเดือนก่อน +1

      Hi, try to see the content of result, maybe you are not detecting any masks. Also, maybe something has changed in a more recent version of ultralytics Python package, try downgrading to a previous version.

    • @afolabiowoloye804
      @afolabiowoloye804 4 หลายเดือนก่อน

      @@ComputerVisionEngineer Thanks for the response, I later discovered that have "(no detections)". However, I used the same image from the validation dataset.
      Please, if you don't mind, you can suggest another tutorial video.

    • @afolabiowoloye804
      @afolabiowoloye804 4 หลายเดือนก่อน

      @ComputerVisionEngineer Thanks for the response, I later discovered that have "(no detections)". However, I used the same image from the validation dataset.
      Please, if you don't mind, you can suggest another tutorial video.@@ComputerVisionEngineer

    • @ComputerVisionEngineer
      @ComputerVisionEngineer  4 หลายเดือนก่อน

      @@afolabiowoloye804 another tutorial for semantic segmentation with yolov8? I made another video where I trained the model in a Google colab.

    • @afolabiowoloye804
      @afolabiowoloye804 4 หลายเดือนก่อน

      @@ComputerVisionEngineer thanks for your prompt response, it means a lot to me. Kindly help me with the video. Thanks

  • @Bukuboys
    @Bukuboys 9 หลายเดือนก่อน

    What could be the reason for the following issue:
    for j, mask in enumerate(result.masks.data):
    AttributeError: 'NoneType' object has no attribute 'data'

  • @jay_9070
    @jay_9070 4 หลายเดือนก่อน

    Is the validation dataset a separate data set?? Is it necessary?

    • @ComputerVisionEngineer
      @ComputerVisionEngineer  4 หลายเดือนก่อน

      It is not absolutely necessary, but it is a good practice to use a different dataset as validation set.