- 72
- 15 367
ARJUNTHEPROGRAMMER
India
เข้าร่วมเมื่อ 2 ธ.ค. 2019
“Formal education will make you a living; self-education will make you a fortune.”
Text-to-3D Environment Generation AI Tool | Robotics Simulation | AuraML
🚀 AuraML's Text-to-3D Environment Generation AI Tool!
This cutting-edge technology turns simple text prompts into immersive 3D environments, making it easier than ever to create realistic scenarios for a variety of applications.
🌟 Why This Matters: In our latest demo, we demonstrate a powerful robotics use case-training robotic arms in diverse 3D environments. With our AI tool, users can effortlessly generate scenarios to simulate real-world challenges, enabling comprehensive testing and training.
🔍 Key Features:
- Intuitive Interface: No 3D modelling experience is needed. Just enter a text prompt and watch the transformation!
- Versatile Applications: Ideal for robotics testing, game development, architectural visualization, and more.
- Rapid Prototyping: Quickly generate multiple scenarios to identify potential failure points in robotic operations.
🎯 Use Case Spotlight: Picture a robotic arm navigating a factory floor with obstacles or performing precise tasks in a cluttered setting. Our tool allows engineers and developers to visualize and optimize these scenarios before they reach the physical stage, saving both time and cost.
👉 Ready to see the magic? Try it out today and let your creativity soar!
Contact us through our website: www.auraml.com
#robotics #simulation #textto3d #genai #training #testing #roboticarm #reinforcementlearning #auraml #syntheticdata
This cutting-edge technology turns simple text prompts into immersive 3D environments, making it easier than ever to create realistic scenarios for a variety of applications.
🌟 Why This Matters: In our latest demo, we demonstrate a powerful robotics use case-training robotic arms in diverse 3D environments. With our AI tool, users can effortlessly generate scenarios to simulate real-world challenges, enabling comprehensive testing and training.
🔍 Key Features:
- Intuitive Interface: No 3D modelling experience is needed. Just enter a text prompt and watch the transformation!
- Versatile Applications: Ideal for robotics testing, game development, architectural visualization, and more.
- Rapid Prototyping: Quickly generate multiple scenarios to identify potential failure points in robotic operations.
🎯 Use Case Spotlight: Picture a robotic arm navigating a factory floor with obstacles or performing precise tasks in a cluttered setting. Our tool allows engineers and developers to visualize and optimize these scenarios before they reach the physical stage, saving both time and cost.
👉 Ready to see the magic? Try it out today and let your creativity soar!
Contact us through our website: www.auraml.com
#robotics #simulation #textto3d #genai #training #testing #roboticarm #reinforcementlearning #auraml #syntheticdata
มุมมอง: 96
วีดีโอ
3D Reconstruction using Gaussian Splatting | AuraML Office Terrace | Product Demo |
มุมมอง 2893 หลายเดือนก่อน
www.auraml.com/ 🚀 3D Environment Reconstruction using Gaussian Splatting 🚀 We're thrilled to announce a groundbreaking demonstration of our latest product that leverages 3D reconstruction through Gaussian splatting! This innovative approach allows us to construct detailed environments and segment the environment to configure it according to your requirements. Here's what you can expect from the...
AuraML Synthetic Dataset Generation Platform Demo | Warehouse Env | Safety Monitoring
มุมมอง 2384 หลายเดือนก่อน
www.auraml.com/ Configure 3D environment and generate synthetic image datasets - Change Lighting - Add objects with variations - Add robots using splines - Add workers: males and females - Configure Camera Properties - Configure Annotation settings: COCO, YOLO
InstructScene: Instruction-Driven 3D Indoor Scene Synthesis with Semantic Graph Prior | Paper + Code
มุมมอง 1775 หลายเดือนก่อน
GitHub: github.com/chenguolin/InstructScene Project: chenguolin.github.io/projects/InstructScene/ Paper: arxiv.org/abs/2402.04717 Dataset: huggingface.co/datasets/chenguolin/InstructScene_dataset/tree/main My CheckPoints for Semantic Graph and Layout Decoder (Bedroom): github.com/chenguolin/InstructScene/issues/9 A novel generative framework that integrates a semantic graph prior and a layout d...
3D Gaussian Splatting with Unreal Engine 5 | Train, View, Edit | 12th July 2024 |
มุมมอง 1406 หลายเดือนก่อน
Project Page: repo-sam.inria.fr/fungraph/3d-gaussian-splatting/ Research Paper: repo-sam.inria.fr/fungraph/3d-gaussian-splatting/3d_gaussian_splatting_low.pdf Original Repo: github.com/graphdeco-inria/gaussian-splatting Windows Setup: github.com/jonstephens85/gaussian-splatting-Windows Gaussian Splat Editor: playcanvas.com/supersplat/editor Unreal Plugin: github.com/xverse-engine/XV3DGS-UEPlugi...
Mistoline - SDXL ControlNet | Generative AI - Image Generation | 28th June 2024
มุมมอง 2096 หลายเดือนก่อน
GitHub: github.com/arjuntheprogrammer/MistoLine MistoLine is a versatile and robust SDXL-ControlNet model developed by TheMistoAI that can adapt to any type of line art input, demonstrating high accuracy and excellent stability - It can generate high-quality images (with a short side greater than 1024px) based on user-provided line art of various types, including hand-drawn sketches, different ...
Text to 3D Environment Generation - Part 2 | Llama 3 Finetuning - Unsloth | AuraML | 18th June 2024
มุมมอง 1436 หลายเดือนก่อน
github.com/auraml/UE5PCGTools github.com/auraml/llm_text_to_3d
Text to 3D Environment Generation - Part 1 | Llama 3 Finetuning - Unsloth | AuraML | 14th June 2024
มุมมอง 957 หลายเดือนก่อน
GitHub: github.com/auraml/llm_text_to_3d Dataset Generation: github.com/auraml/UE5PCGTools Use case: Racks Cluster Generation Input - Text description to build 3D racks cluster Output - List of transforms which can be used to spawn different objects in the scene. Sample Input: Create a 3D racks cluster system with 2 rows and 5 columns of racks. Each rack consists of 3 rows and 4 columns of tray...
Finetune Google Flan T5 Base - Text Classification | IMDB Dataset | 13th June 2024
มุมมอง 4137 หลายเดือนก่อน
GitHub: github.com/arjuntheprogrammer/finetune-flan-t5-text-classification GPU RAM: 11.4 GB Train Dataset Size: 25K Training Time: 28 mins Epochs: 2 Loss: 0.0786 F1: 94.9518
Procedural Racks Cluster System | Unreal Engine 5 | Using Python | AuraML | 11th June 2024 |
มุมมอง 1397 หลายเดือนก่อน
Procedural Racks Cluster using Python GitHub: github.com/auraml/UE5PCGTools Feel free to raise PR to extend this procedural system for indoor environments. LinkedIn: www.linkedin.com/in/arjuntheprogrammer/
Finetune CodeLlama 7B HF LLM Model | Alpaca 20K Dataset | 31st May 2024 |
มุมมอง 3877 หลายเดือนก่อน
GitHub: github.com/arjuntheprogrammer/finetune_codellama_alpaca20K GPU TYPE: A100 80GB VRAM USED: 11.6GB Training Dataset: HuggingFaceH4/CodeAlpaca_20K Dataset Link: huggingface.co/datasets/HuggingFaceH4/CodeAlpaca_20K Dataset Size: 20K rows (train: 18K, test: 2K) Pretrained Model: codellama/CodeLlama-7b-hf Pretrained Model Link: huggingface.co/codellama/CodeLlama-7b-hf Finetuning Training Time...
Finetune and Deploy Mistral 7B LLM Model on AWS Sagemaker | QLoRA | 29th May 2024 |
มุมมอง 1.4K7 หลายเดือนก่อน
GitHub: github.com/arjuntheprogrammer/sagemaker_finetune_mistral7B_and_deploy/ Model: huggingface.co/mistralai/Mistral-7B-v0.1 Dataset: huggingface.co/datasets/databricks/databricks-dolly-15k QLora: arxiv.org/abs/2305.14314 Sagemaker: aws.amazon.com/sagemaker/
Preprocessing Unstructured Data - Part 5 | LLM RAG BOT | Unstructured IO | 26th May 2024
มุมมอง 1867 หลายเดือนก่อน
GitHub: github.com/arjuntheprogrammer/llm_preprocess_unstructured_data/tree/main/L6:RAGBot
Preprocessing Unstructured Data - Part 1 | Normalizing Data | Unstructured IO | 26th May 2024
มุมมอง 1877 หลายเดือนก่อน
Github: github.com/arjuntheprogrammer/llm_preprocess_unstructured_data/tree/main/L2:Normalizing
Preprocessing Unstructed Data - Part 2 | Metadata Extraction and Chunking | Unstructured IO | 26th M
มุมมอง 1327 หลายเดือนก่อน
Github: github.com/arjuntheprogrammer/llm_preprocess_unstructured_data/tree/main/L3:MetadataExtraction
Preprocessing Unstructured Data - Part 4 | Extracting Tables | Unstructured IO | 26th May 2024
มุมมอง 2857 หลายเดือนก่อน
Preprocessing Unstructured Data - Part 4 | Extracting Tables | Unstructured IO | 26th May 2024
Preprocessing Unstructured Data - Part 3 | Preprocessing PDFs and Images | Unstructured IO
มุมมอง 1457 หลายเดือนก่อน
Preprocessing Unstructured Data - Part 3 | Preprocessing PDFs and Images | Unstructured IO
ChatDev | Communicative Agents for Software Development | 25th May 2024 |
มุมมอง 8537 หลายเดือนก่อน
ChatDev | Communicative Agents for Software Development | 25th May 2024 |
Multi Agent LLM | Resume Builder | Part6 | Crew AI | 24th May 2024 |
มุมมอง 1.8K7 หลายเดือนก่อน
Multi Agent LLM | Resume Builder | Part6 | Crew AI | 24th May 2024 |
Multi Agent LLM | Financial Analysis | Part 5 | Crew AI | 23rd May 2024 |
มุมมอง 5147 หลายเดือนก่อน
Multi Agent LLM | Financial Analysis | Part 5 | Crew AI | 23rd May 2024 |
Multi Agent LLM | Event Planning | Part 4 | Crew AI | 23rd May 2024 |
มุมมอง 2347 หลายเดือนก่อน
Multi Agent LLM | Event Planning | Part 4 | Crew AI | 23rd May 2024 |
Multi Agent LLM | Customer Outreach Campaign | Part 3 | Crew AI | | 23rd May 2024 |
มุมมอง 1997 หลายเดือนก่อน
Multi Agent LLM | Customer Outreach Campaign | Part 3 | Crew AI | | 23rd May 2024 |
Multi Agent LLM | Customer Support | Part 2 | Crew AI |21st May 2024 |
มุมมอง 3627 หลายเดือนก่อน
Multi Agent LLM | Customer Support | Part 2 | Crew AI |21st May 2024 |
Multi Agent LLM | Research and Write Article | Part 1 | Crew AI | 21st May 2024 |
มุมมอง 7957 หลายเดือนก่อน
Multi Agent LLM | Research and Write Article | Part 1 | Crew AI | 21st May 2024 |
How to Finetune Llama 3 with Custom Data | Unsloth | 21st May 2024 |
มุมมอง 4237 หลายเดือนก่อน
How to Finetune Llama 3 with Custom Data | Unsloth | 21st May 2024 |
Chatbot | Prompt Engineering | OpenAI | 21st May 2024 |
มุมมอง 297 หลายเดือนก่อน
Chatbot | Prompt Engineering | OpenAI | 21st May 2024 |
Evaluation | LLM Finetuning Part 4 | 16th May 2024 | mp4
มุมมอง 337 หลายเดือนก่อน
Evaluation | LLM Finetuning Part 4 | 16th May 2024 | mp4
InstructionFinetuning | LLM Finetuning Part 1 | 16th May 2024 |
มุมมอง 217 หลายเดือนก่อน
InstructionFinetuning | LLM Finetuning Part 1 | 16th May 2024 |
LLM Training | LLM Finetuning Part 3 | 16th May 2024 |
มุมมอง 207 หลายเดือนก่อน
LLM Training | LLM Finetuning Part 3 | 16th May 2024 |
DataPreparation | LLM Finetuning Part 2 | 16th May 2024 |
มุมมอง 407 หลายเดือนก่อน
DataPreparation | LLM Finetuning Part 2 | 16th May 2024 |
It's one of the project which we can definitely add in our resume. Thank you Arjun.
Hi Arjun, Thanks for this great tutorial. I am facing one issue. hoping if you could help. I am working on a use case and as per requirement i have to use offline LLM model, hence using an offline model (mistral-7b-instruct-v0.2.Q4_K_M.gguf). I am also using crewAI to build my multiagent system. I am keep on getting this error : "Number of tokens (85083) exceeded maximum context length (32000)". 32k context window is max for model i am using. Without changing model (given i have limited resource to handle large model) what i can do to overcome this issue.. any work around you would like to suggest.? thanks and would really appreciate your response.
Hello sir I have followed your video, i started with the notebook example, however i got stuck when i run this cell [ from transformers import AutoTokenizer model_id = "mistralai/Mistral-7B-v0.1" tokenizer = AutoTokenizer.from_pretrained(model_id, use_auth_token=True) ] This is the error i am getting: Exception: data did not match any variant of untagged enum PyPreTokenizerTypeWrapper at line 40 column 3 I did not change anything though. Please help me out.
did you find a solution? I was having the same issue and I installed the tokenizer locally and used the path to the files, instead of model_id. Don't know if it's the best solution
Hii bro great video but can you tell me that if i upload a llama model on sagemaker and only use it through API gateway then i am going to charge for per API request i will charge for 24/7 as llm is deployed there on sage maker
hi , isn't this AR type ?
yes, you can say this is AR but the point that we have highlighted here is that using your mobile camera, you can quickly build your own 3D environment which can be then used to generated dataset for training your computer vision models with better accuracy. Currently there is no tool out there which lets you modify and edge cases to your current scene where the model is failing.
@@arjuntheprogrammer So, can one just upload a video and this would give us a proper 3d reconstruction of the space? Is this available as an API? I might want to use this for our startup
if can't market , you can't sell :)
Thank you
What's the advantage of using CrewAI vs a couple of API calls to OpenAI?
CrewAI provides a higher-level abstraction on top of the OpenAI API, making it easier to build AI applications without needing to manage all the low-level details
@@arjuntheprogrammer Thanks! Very clear.
Nice work, Sir ! Very informative
project\StableVITON-master\download\VITONHD.ckpt project\StableVITON-master\output i am getting this error what to doo ?
will you please make a detailed video how you are running the program, from scratch like after cloning github repo what we have to do, and how we have to connect to output. Please!!!!!!!!!!!!🥺🥺
in the requirements, I couldn't install the triton package and its coming like this ERROR: Could not find a version that satisfies the requirement triton==2.0.0 (from versions: none) ERROR: No matching distribution found for triton==2.0.0
me too
can ypu add api to the arxiv,ieee, nature,conferences , etc ? how to do it? how to control the number of references? and length
Great video! thank you very much! :), As another question, I've been trying to use llama 3 7b, but for some reason it seems quite large for it to fit with my data, any idea what instance might be best for llama 3 7b? and now llama 3.1 7b
you can try any 24 GB VRAM GPU - AWS A10G or GCP L4
@@arjuntheprogrammer I have another question, if I wanted to fine tune for TaskType.QUESTION_ANS, what modifications should I make to the code? something in particular to the way I store my training data?
Great Video
thanks
bro did u train the model first?
No, I didn't train it. They have released their checkpoints: kaistackr-my.sharepoint.com/personal/rlawjdghek_kaist_ac_kr/_layouts/15/onedrive.aspx?id=%2Fpersonal%2Frlawjdghek%5Fkaist%5Fac%5Fkr%2FDocuments%2FStableVITON&ga=1
@@arjuntheprogrammer Above link is expired.
how can i load my finetuned model after the save please
from transformers import AutoModelForSequenceClassification, AutoTokenizer # Specify the model ID or path model_id = "hf_username/repo_name" # Load the model and tokenizer model = AutoModelForSequenceClassification.from_pretrained(model_id) tokenizer = AutoTokenizer.from_pretrained(model_id)
@@arjuntheprogrammer i need to merge the adaptater with the base model?
Why have you copied all the Notebooks from Deep learning AI course, by crew AI? Try to create your own content.
Hi, I have used their notebooks so that it becomes easier of any one to get started quickly without going through the entire course. Also in my other videos, you can see I have created my own content and explain the whole training process.
Outanding twice! Best tut of its kind :) Let me buy you a coffee.
Sure. buymeacoffee.com/arjungupta
Truly helpful... but I wanted to know that i have my own custom data in the form of 2 txt files how am i supposed to fine tune them with the model. Do i need to convert them to a csv format? Or is there another way to proceed?
yes, better to convert it into a CSV file to give it a structure and then load it up. Code Example: from datasets import load_dataset dataset = load_dataset("csv", data_files="path/to/your_file.csv")
Outstanding! But mistral-finetune github has a different approach and in less than one hour the job is done, what are the main differences or advantages? Thanks!
They might be using a different dataset to fine-tune the model and maybe a bigger GPU to train faster.
How to install unstructured io as docker container ?
docker pull downloads.unstructured.io/unstructured-io/unstructured:latest
Can you explain what the flags repaint and unpair are? I was able to run the code. but results generated are same as previous models photos
Can you say why this is coming: Loaded model config from [None] Traceback (most recent call last): File "C:\Users\dev\Downloads\StableVITON\inference.py", line 110, in <module> main(args) File "C:\Users\dev\anaconda3\envs\StableVITON\lib\site-packages\torch\utils\_contextlib.py", line 115, in decorate_context return func(*args, **kwargs) File "C:\Users\dev\Downloads\StableVITON\inference.py", line 66, in main for batch_idx, batch in enumerate(dataloader): File "C:\Users\dev\anaconda3\envs\StableVITON\lib\site-packages\torch\utils\data\dataloader.py", line 634, in _next_ data = self._next_data() File "C:\Users\dev\anaconda3\envs\StableVITON\lib\site-packages\torch\utils\data\dataloader.py", line 1346, in _next_data return self._process_data(data) File "C:\Users\dev\anaconda3\envs\StableVITON\lib\site-packages\torch\utils\data\dataloader.py", line 1372, in _process_data data.reraise() File "C:\Users\dev\anaconda3\envs\StableVITON\lib\site-packages\torch\_utils.py", line 644, in reraise raise exception ValueError: Caught ValueError in DataLoader worker process 0. Original Traceback (most recent call last): File "C:\Users\dev\anaconda3\envs\StableVITON\lib\site-packages\torch\utils\data\_utils\worker.py", line 308, in _worker_loop data = fetcher.fetch(index) File "C:\Users\dev\anaconda3\envs\StableVITON\lib\site-packages\torch\utils\data\_utils\fetch.py", line 51, in fetch data = [self.dataset[idx] for idx in possibly_batched_index] File "C:\Users\dev\anaconda3\envs\StableVITON\lib\site-packages\torch\utils\data\_utils\fetch.py", line 51, in <listcomp> data = [self.dataset[idx] for idx in possibly_batched_index] File "C:\Users\dev\Downloads\StableVITON\dataset.py", line 245, in _getitem_ transformed_image = self.transform_crop_person( File "C:\Users\dev\anaconda3\envs\StableVITON\lib\site-packages\albumentations\core\composition.py", line 255, in _call_ self._check_args(**data) File "C:\Users\dev\anaconda3\envs\StableVITON\lib\site-packages\albumentations\core\composition.py", line 324, in _check_args raise ValueError(msg) ValueError: Key agn is not in available keys.
Did you fix this?
Looks like the error is raised from the StableVITON package itself. Not sure, I will require more context to understand the problem.
I am getting same problem.
superrr.. I only worry about training /fine tuning cost to LLM
Hi, you can also use spot instances provided by AWS. I was able to save upto 70% training cost using spot instances for another model I trained in the past.
Nice tutorial! Thanks for that. How would you distribute this over a multi GPU instance, which might be needed for 8x7b (Instruct or Base), for example?
docs.aws.amazon.com/sagemaker/latest/dg/distributed-training.html did you try the config params provided by AWS Sagemaker (SMDDP library)? such as: distribution={"pytorchddp": {"enabled": True}}
what is the recommended gpu to run it on please, thaank you
StableVITON authors recommend using a powerful GPU like the RTX 3090 for best results. I used my inhouse RTX 4090.
Great Video! Concise and to the point 👏
Glad you liked it!!
👏👏👍
thanks
hello, im trying to run this but i cant get it to work without setting the CUDA visible devices. how do i do it?
Before setting CUDA_VISIBLE_DEVICES, it is helpful to know the available GPUs on your system. try to use the following statement in python: os.environ["CUDA_VISIBLE_DEVICES"] = "0" Replace "0" with the indices of the GPUs you want to use.
can i get ur contact information
Happy Diwali!! Wonderful celebration 🎉
same to you
Happy Diwali 🎉
same to you
How does it compare with COLMAP in terms of accuracy and compute?
Before running the train.py command for this video as specified in their Github Repo(github.com/graphdeco-inria/gaussian-splatting?tab=readme-ov-file#running), we have to actually prepare the COLMAP dataset. So I see COLMAP as part of the pipeline for generating Gaussian Splats. But recently, I came across new paper: oasisyang.github.io/colmap-free-3dgs/ - which removes the colmap to create 3DGS.
👏👏👏
thanks
Is amazon sagemaker free
No, it is paid.
Truly inspiring sir👍🏻 Best wishes
thanks
Really inspiring
thanks
Really nice video ... Very detailed
thanks
Detailed and well explained, Thanks for the video 👍
thanks
Excellent.
thanks
Very good attempt Arjun!
Thanks TSAI. Learnt so much from your classes.
Please can you send me the code
it is a very old project, you can search for yolo based detection colab file on google