Thanks so much. I just read through the privateGPT py files, and I don't have any python experience. This video helped me understand the reasoning behind that code.
thanks glad you got something out of it! fine-tuning in general is definitely something I'd like to cover if I can, as it's likely something most productized uses of LLMs will benefit from...
This is amazing! Thank you so much! It is not some "black box" that so many of these doomsayers bleat incessantly, but rather the closest we can really get to a superpower for those who thoroughly understand them!
I'm not a doomsayer, but I agree with the "black box" tag that AI models get. I don't understand how this video proved to you that it is not a "black box". The guy making the video was often literally surprised by the answers he got from the language model. And when you're unable to predict the answer, that's basically a "black box" by definition. You probably think that "black box" means we don't understand how the language models work. That is not correct. What is meant by the phrase is, that we don't know what those models will return. I hope i clarified things for you.
Your REPL is a manual ML evaluation. Many devs don't like manual stuff and so eventually we will all realize that figuring out if it works (like asking it the questions the base model flubbed) is manual testing. It's sort of similar to functional testing. "I just coded up something but how do I know it works?". We have a choice to make on whether to quickly manually test it with the massive trade-off of: you cannot possibly ask it every question. Put another way: ask it 10,000 questions every time you change something (code/scripts/model/parameters). Give it a score. The score is still not enough without context. The model itself it not just bigger for the sake of being bigger but would be improved and measured. We would be attracted to better performing models. In other words, we expect to consume software and models that are well-functioning and performant. Those things use automation. We've been doing this for a long time with plain old software. In ML, this is the new bit. The data and how the result is not deterministic. This is different than plain software.
you'd definitely need one or the other (this particular example uses pytorch but I think you can just replace "pt" with "tf" for tensorflow). Not sure about any of the specifics on getting it working in Windows though.
Hi, thanks for the nice video! A question, in your Rust video you refer to, you load libtorch, or pytorch. However if you build that program in Rust and give it to someone, would they not also have to install libtorch on their machine and set their environment? And if so, do you know of a way to embed libtorch or pytorch?
I wanted to clear up something about the example. When the tokenizer separated the word "undoubtedly" into two tokens, "_un" and "doubtedly," you mentioned it was mainly for efficiency. However, what I learned is that it's called subword tokenization strategy, and it helps the model handle words not found in its training data. Is not that the purpose ?
I think that's another aspect of it. But it's also my understanding (could be wrong) that it's also to simplify the model (and potentially improve performance) by having it learn the semantic meaning of the "root" or a word, instead of having to independently learn the semantic meaning of each slightly different flavor of a word
@@codetothemoon I dug deep and found something interesting and you were right about the performance point , he is a summary of what I found: the large language models typically have a fixed vocabulary size due to memory constraints. Subword tokenization enables the model to represent a more significant number of unique words by encoding them as subword tokens. This helps in handling a wide range of words while keeping the vocabulary manageable. Also, another goal is to reduce the overall number of tokens in a text, making the input sequences shorter. This leads to more efficient memory utilization and faster computation during training and inference
Awesome tutorial. Would it be possible to make a video on how to use one of these models to further fine train it on our own corpus of content? Sort of make it learn additional things on top, is that too complicated?
How can i load an list of "intentions" for a language model to classify ? Like when i send some question like: "Please turn on the lights" i would like that the language model classify the "turn_on" intentions as true with the entity "lights"
This is excelent. Just the other day I was wondering if I could take a markdown wiki (or obsidian vault) and feed it to a LM to ask questions about it. Specifically I wanted to take a messy wiki from work and giving it to a LM and then asking the LM questions. Then we can improve the wiki and get an ever better chatbot that can answer questions. Then can do something like feed it the python and django docs as a baseline.
yes, but this will be done automatically the first time you run the program (if written the way we do here). By default, the Transformers package completely abstracts away the process of downloading the model
spoke too soon. Base model works nicely for the [blue] sky but the common pet question gets the spokesman for the Associated Press involved: Enter something: What is a common pet for humans to have? ["W: I'm sorry, but I'm not sure what happened. I'm not sure what happened."] ['h - h - h'] ['spokesman for the Associated Press said the agency was unable to comment on the matter.'] ['t t t t t t t t t t t t t t t t t t t t t t t t t t t t t t t t t t t t t t t t t t t t t t t t t t t t t t t t t t t t t t t t t t t t t t t t t t t t t t t t t t t t t t t t t t t t t t t t t t t t'] [' '] ['i. i. i. i. i. i. i. i. i. i. i. i. i. i. i. i. i. i. i. i. i. i. i. i. i. i. i. i. i. i. i. i. i. i. i. i. i. i. i. i. i. i. i. i. i. i. i. i. i. i. i. i. i. i. i. i. i. i. i. i. i. i. i. i. i. i. i']
Interestingly this bad (t5-large) model behavior only happens when run inside your FOR loop structure. When a FOR loop uses a function call then the large model gets the answers right even for Apollo13 when the question is more specific, i.e. aborted landing.
I'm willing to bet it could be fixed with fine tuning, especially if you have some idea of the specific types of questions the model will be expected to answer accurately. The other option is to use larger flavors of the same model. You should be able to get what you're going for using one or both of those strategies
The way I'm using it in this video - the answer is no. There are models specifically trained for this task though, check this out huggingface.co/models?pipeline_tag=conversational&sort=downloads
how can i add my own corpus to a base model and weighting it? i have a big searchable database but want to make it "transformer searchable". Does such a solution exist? Pls someone point me in the right direction. Thank You
I know this tutorial wanted to show how to save on the expenses and run a model on your own machine. But ironically, the outcome actually showed me what we're paying OpenAI for. These models get already very basic things wrong, really. I don't know if it's much of use with these models, yet.
the idea is to combine this idea with our local data and then it becomes powerful. we would do that via Llama_index. Just not too many videos out there on how to do this.
They are pretty capable with no fine tuning, it’s very possible they might already do what you need without any additional training. But fine tuning can definitely make them better - I’m up for making a video about that if folks are interested
@@codetothemoon I'm thinking like site functionality, so a user is asking a question about how to use the site in some way, or some rules about the site. If it's just a general model I don't see the point of every site having one.
@@theLowestPointInMyLife right - actually you can cover quite a bit of ground by having an app-specific prompts that you interpolate user-specific values into. Using this approach you can generate app-specific output without actually doing any fine tuning to the model itself. That may or may not be sufficient depending on what you're doing.
not sure if I'll make it to a TypeScript version, though I imagine such a thing would be immensely popular. I am likely going to make a chatbot webapp in 100% Rust - it might be fun to do a followup of the same webapp using something like SvelteKit - we'll see
actually its doom emacs! Which has evil mode (vim bindings) enabled by default. I use the vim motions because I feel like they make me more productive, and I use emacs mostly because of org mode - if there was no org mode I might be using Neovim.
I suspect tokenization stems from cognitive psychology and morphemes. If a token is a meaning unit then a morpheme is the smallest possible meaning unit. I.e., un + happy, un changes the meaning of the word.
@@codetothemoon Yep! I lecture in cognitive psychology and there are a lot more parallels. I did a class with a huggingFace GPT model where you could change the token length to illustrate meaningful sentence production/comprehension (i.e., Wernicke's aphasia). I am fairly certain that an interdisciplinary field of academia will spawn in the next few years that marry machine learning and cognitive psychology. I'm sure there's the odd journal here and there but I think it'll take the place of conventional cognitive psychology.
How do I get the max amount of tokens supported by each model? I'm trying to build a generic scraper that scrapes the given documentation of an app, finds all relevant REST APIs, webhooks etc along with the relevant details. I would then convert them into embeddings and insert into a database. I'll then search on these semantically. What I'm struggling with right now is how to find the structure of the documentation. I think a LLM model would help with that but with most, I feel that I'll have a problem with the max number of tokens when I pass the HTML to them. I have access to GPT-4 but AFAIK, the 32k model hasn't been released yet. And even if I had access to it, I'll run into cost constraints.
great point! I think I originally mentioned this in the video but I may have left it out of the final cut. But the one we're using in the video (T5 Flan) is Apache 2.0 and thus should be fine to use for commercial purposes.
@@codetothemoon Yes, the ones you showed are save. Just wanted to point it out… ran in a start-up last week that proudly pitched their product and than were devastated that a superficial tech DD reveled that they break licenses left and right. Licenses always have been the annoying accounting part everyone dreads, but for ML models it’s currently 10x as confusing. And probably the only reason, why the „you can actually run stuff locally“ needs extra care in that department. Anyway: Great video as always Mr Moon !
They teach how to import pretrained data but never show how to train models and create system of neural network layers to use it as normal dev would all that AI stuff is just monopoly on computation patterns
pretrained models are an incredible resource, as the vast majority of developers don't have the resources or budget to perform such trainings. I'm not sure what you mean by "create a system of neural network layers to use as a normal dev would", as the models in this video are built upon neural networks
Thank you, however, believe most of the famous tech corps where aware of this space, prior 2017 paper (~7-5 yrs), and was masking this technology from the public space, though they used public dataset, Public should really seek for accountability using a proper legal framework, if only politicians were skillful, knowledgeable and talent enough, least they could seek from these tech corps is to provide cloud platform workspace for closer to a zero pricing scenario.
Thanks for the video 👍. I just can't stop wondering, how come such an impossible poor, slow and inconvenient programming language with the development environment from the past century became a standard for machine learning tasks... It doesn't even have method names autocomplete, it doesn't show the full list of object's methods available, no parameters list, no types - just nothing. 💩 Crazy. return_tensors="pt" 🤦
Just a bit of food for thought ... all CPU chips ONLY understand how to manipulate bits and bytes in registers. EVERY computer language is simply a way to create those bit patterns to feed the hardware chip. Each language chooses how much abstraction to use and how the abstractions are interpreted. Some do it this way and some do it that way, but the end result in ALL cases is a sequence of machine (binary) codes that the Assembly language for the chip can write directly with NO abstractions except, perhaps, some mnemonics (memory aid) to avoid memorizing all the binary values associated with each machine instruction. Programming, paradoxically, is much easier in Assembly language than all others because you've gotten rid of all abstractions and interpretations that you have to wrap your head around. Of course, you still have the abstractions needed for solving the problem you are tackling but you don't have abstractions within abstractions. That's what makes computer programming challenging. Just sayin' ... 😉
I've never encountered a problem that was easier to solve in assembly than a higher level language. The abstractions of higher level languages are there to reduce our cognitive load, and they are incredibly effective at doing so
@@codetothemoon Absolutely, I agree. Abstracting things is important in many fields, not just programming. I was principally interested in peeling the onion to show why something like programming CAN be made more difficult because of abstractions but, of course, we have to understand that creating abstractions is an art in itself. It can be done well or done poorly. The effectiveness of using assembly language also depends on the programmer having a rock-solid understanding of the underlying electronics hardware. As an electronics engineer, I have had to work with many chips that didn't have compilers available so all coding had to be done in assembly and carefully mapped out with only the most rudimentary abstractions. This is also true of the microcoding that exists internally in every chip to interpret the binary machine code instructions themselves.
just came here for the thumbnail. we're not "coders" a trained monkey can code. we are programmers. if you dont understand the difference, it's the same difference between writing and typing. yeah......typing/coding is just hitting the keys in order to form the words. writing/programming is thinking. not even gonna watch the video. the thumnail earned it a dislike.
Been self learning Arduino & Python on YT for several projects over the past few years. But, I have absolutely no idea what is being said in this video. Must be a different language than English. 🤷♂ I'll circle back around in another couple of years, after gaining some more smarts.
Possibly the best tutorial on how to plug and play with the models. Much appreciated, I gained valuable insight. Respect!✊
thanks, glad you got something out of it!
this
Agreed
Thank you. This was a very impressive, hype free, presentation. 10 out of 10.
Thanks for the best plug and play model video out there. Super short and right to the point.
thanks for watching, glad you got something out of it!
This is a really great tutorial! The obvious followup would be fine tuning an LLM with your own data. Or maybe using LangChain to do it?
thanks, and I agree! I'd love to make videos on both of those topics.
@@codetothemoono it then 😁 (great video)
*I just found myself a Goldmine, definitely subscribing*
Thanks so much. I just read through the privateGPT py files, and I don't have any python experience. This video helped me understand the reasoning behind that code.
Wow man, this is so good and simple! Didn't know it were this easy to get a model and run it!
nice, really happy you got something out of the video!
One of the best video ive found on this topic ✅
thank you so much!
Wow man you have put so much effort into this tutorial!
I love it, thank you :)
i did, thanks for watching, glad you got something out of it!
As always, very informative and clear. Explains the essentials.
thanks, glad you liked it!
This was really great! Would love to see something on how to LoRA train a model for text generation. Would be really cool!
thanks glad you got something out of it! fine-tuning in general is definitely something I'd like to cover if I can, as it's likely something most productized uses of LLMs will benefit from...
This is amazing! Thank you so much! It is not some "black box" that so many of these doomsayers bleat incessantly, but rather the closest we can really get to a superpower for those who thoroughly understand them!
You're very welcome, glad you got something out of it! 😎
I'm not a doomsayer, but I agree with the "black box" tag that AI models get. I don't understand how this video proved to you that it is not a "black box". The guy making the video was often literally surprised by the answers he got from the language model. And when you're unable to predict the answer, that's basically a "black box" by definition.
You probably think that "black box" means we don't understand how the language models work. That is not correct. What is meant by the phrase is, that we don't know what those models will return. I hope i clarified things for you.
Thank you for your content
I would like to ask you to make a video on the web app and the new developments in that regard.
thanks for watching! I will definitely be making more content around Rust full stack web applications.
Your REPL is a manual ML evaluation. Many devs don't like manual stuff and so eventually we will all realize that figuring out if it works (like asking it the questions the base model flubbed) is manual testing. It's sort of similar to functional testing. "I just coded up something but how do I know it works?". We have a choice to make on whether to quickly manually test it with the massive trade-off of: you cannot possibly ask it every question. Put another way: ask it 10,000 questions every time you change something (code/scripts/model/parameters). Give it a score. The score is still not enough without context.
The model itself it not just bigger for the sake of being bigger but would be improved and measured. We would be attracted to better performing models. In other words, we expect to consume software and models that are well-functioning and performant. Those things use automation.
We've been doing this for a long time with plain old software. In ML, this is the new bit. The data and how the result is not deterministic. This is different than plain software.
bruh, this video is amazing! do you have recomendations on prompt engineering courses for developers?
thank you! I don't have any recommendations unfortunately. That's something I've been meaning to dive deeper into as well.
For running transformers on windows machine - do I need pytorch or tensorflow or both installed?
you'd definitely need one or the other (this particular example uses pytorch but I think you can just replace "pt" with "tf" for tensorflow). Not sure about any of the specifics on getting it working in Windows though.
I’m curious about your vim workflow. Nice colorings etc. also how do you use the keyboard to open a penv terminal automatically sized side by side?
thanks, actually its doom emacs! To open the python terminal you can just use M-x then select "run python". I'm using the doom-monokai-pro theme.
Hi, thanks for the nice video! A question, in your Rust video you refer to, you load libtorch, or pytorch. However if you build that program in Rust and give it to someone, would they not also have to install libtorch on their machine and set their environment? And if so, do you know of a way to embed libtorch or pytorch?
Good question, I'm actually not sure!
I wanted to clear up something about the example. When the tokenizer separated the word "undoubtedly" into two tokens, "_un" and "doubtedly," you mentioned it was mainly for efficiency. However, what I learned is that it's called subword tokenization strategy, and it helps the model handle words not found in its training data.
Is not that the purpose ?
I think that's another aspect of it. But it's also my understanding (could be wrong) that it's also to simplify the model (and potentially improve performance) by having it learn the semantic meaning of the "root" or a word, instead of having to independently learn the semantic meaning of each slightly different flavor of a word
@@codetothemoon I dug deep and found something interesting and you were right about the performance point , he is a summary of what I found:
the large language models typically have a fixed vocabulary size due to memory constraints. Subword tokenization enables the model to represent a more significant number of unique words by encoding them as subword tokens. This helps in handling a wide range of words while keeping the vocabulary manageable.
Also, another goal is to reduce the overall number of tokens in a text, making the input sequences shorter. This leads to more efficient memory utilization and faster computation during training and inference
Awesome tutorial. Would it be possible to make a video on how to use one of these models to further fine train it on our own corpus of content? Sort of make it learn additional things on top, is that too complicated?
Yeah this would be greattttt
search for PriavateGPT and you will find your answer!
How can i load an list of "intentions" for a language model to classify ? Like when i send some question like: "Please turn on the lights" i would like that the language model classify the "turn_on" intentions as true with the entity "lights"
This is excelent. Just the other day I was wondering if I could take a markdown wiki (or obsidian vault) and feed it to a LM to ask questions about it.
Specifically I wanted to take a messy wiki from work and giving it to a LM and then asking the LM questions. Then we can improve the wiki and get an ever better chatbot that can answer questions. Then can do something like feed it the python and django docs as a baseline.
nice, yes I think what you describe is definitely achievable!
so to use it do we have to download the model first?
yes, but this will be done automatically the first time you run the program (if written the way we do here). By default, the Transformers package completely abstracts away the process of downloading the model
your example works nicely in Jupyter notebook after making sure the working version of 'transformers-4.28.0.dev0' is installed.
spoke too soon. Base model works nicely for the [blue] sky but the common pet question gets the spokesman for the Associated Press involved:
Enter something: What is a common pet for humans to have?
["W: I'm sorry, but I'm not sure what happened. I'm not sure what happened."]
['h - h - h']
['spokesman for the Associated Press said the agency was unable to comment on the matter.']
['t t t t t t t t t t t t t t t t t t t t t t t t t t t t t t t t t t t t t t t t t t t t t t t t t t t t t t t t t t t t t t t t t t t t t t t t t t t t t t t t t t t t t t t t t t t t t t t t t t t t']
[' ']
['i. i. i. i. i. i. i. i. i. i. i. i. i. i. i. i. i. i. i. i. i. i. i. i. i. i. i. i. i. i. i. i. i. i. i. i. i. i. i. i. i. i. i. i. i. i. i. i. i. i. i. i. i. i. i. i. i. i. i. i. i. i. i. i. i. i. i']
Interestingly this bad (t5-large) model behavior only happens when run inside your FOR loop structure. When a FOR loop uses a function call then the large model gets the answers right even for Apollo13 when the question is more specific, i.e. aborted landing.
How can I fix errors the model makes, like the ones with the Apollo question?
I'm willing to bet it could be fixed with fine tuning, especially if you have some idea of the specific types of questions the model will be expected to answer accurately. The other option is to use larger flavors of the same model. You should be able to get what you're going for using one or both of those strategies
Just a question, do these language models hold context for previous questions and replies?
The way I'm using it in this video - the answer is no. There are models specifically trained for this task though, check this out huggingface.co/models?pipeline_tag=conversational&sort=downloads
@@codetothemoon hey thanks, appreciate your time to reply back.
I would love if you did a video taking this one step further and shopw how to connect to our local data using llama index
actually even scary, how simple everything is
13:49 There were two: Apollo 8 and 10
Subscribed. Thank you
fantastic, really happy to have you on board!
The diagram in minute 9:30 reminds me when I tried to learn SVMs.
pyhton IDE you're showing is within hugging face?
actually I'm using doom emacs in this video!
how can i add my own corpus to a base model and weighting it? i have a big searchable database but want to make it "transformer searchable". Does such a solution exist? Pls someone point me in the right direction. Thank You
Great video, thanks!
thanks for watching!
Isn't this fire 🔥 ship narrator?
Sounds like him!
Is this your new channel?
I know this tutorial wanted to show how to save on the expenses and run a model on your own machine. But ironically, the outcome actually showed me what we're paying OpenAI for. These models get already very basic things wrong, really. I don't know if it's much of use with these models, yet.
the idea is to combine this idea with our local data and then it becomes powerful. we would do that via Llama_index. Just not too many videos out there on how to do this.
Excellent!
thank you, glad you liked it!
What use are these models to your own app/site? Wouldn't you need to train them on your own data that's relevant to your site?
They are pretty capable with no fine tuning, it’s very possible they might already do what you need without any additional training. But fine tuning can definitely make them better - I’m up for making a video about that if folks are interested
@@codetothemoon I'm thinking like site functionality, so a user is asking a question about how to use the site in some way, or some rules about the site. If it's just a general model I don't see the point of every site having one.
@@theLowestPointInMyLife right - actually you can cover quite a bit of ground by having an app-specific prompts that you interpolate user-specific values into. Using this approach you can generate app-specific output without actually doing any fine tuning to the model itself. That may or may not be sufficient depending on what you're doing.
@@codetothemoon Please create a video about fine-tuning if you can. Thanks for this one. Great stuff as always.
Same here
Can I use tensorflow instead of pytorch?
Is this possible on a node environment?Thanks.
do you fed the AI with all your channel and now it does spit out the whole video for you? it's super cool those edit tho
I wish! Definitely spent many hours editing this one 😎
Dude, this is amazing
Nice, glad you got something out of it!
Typescript pleeze, or is that not cool this week?
not sure if I'll make it to a TypeScript version, though I imagine such a thing would be immensely popular. I am likely going to make a chatbot webapp in 100% Rust - it might be fun to do a followup of the same webapp using something like SvelteKit - we'll see
Hey you use vim, it's kind of unexpected since the context is AI and stuff. That's great! 👍
But can I ask why though?
actually its doom emacs! Which has evil mode (vim bindings) enabled by default. I use the vim motions because I feel like they make me more productive, and I use emacs mostly because of org mode - if there was no org mode I might be using Neovim.
thank you so much for this
thanks for watching, glad you got something out of it!
I realized something, the tokens is what we also see in The Matrix digital rain ? 🙂
maybe!
Thank you!
thanks for watching, glad you got something out of it!
Impressive!
thanks, glad you liked it!
I: "In whose legion fights the most formidable armed combatant?" O: "british army" 🤦♂ Great video, CttM.
thank you!
Have you found one model to be particularly good for coding?
I haven't experimented with enough open source models to have a good answer to this, sorry!
Subscribed. Just include a few practical projects. Most people just want to replicate. How to make useful things which might work in real world....
nice, very happy to have you here! I do plan to make more videos about practical applications of language models. Stay tuned...
I love this one
glad you liked it!
I love the random Rust shoutout
not entirely random - much of my audience comes from my Rust content 🦀 gotta build a bit of a bridge there!
@@codetothemoon I'm a rustacean, I loved it.
I suspect tokenization stems from cognitive psychology and morphemes. If a token is a meaning unit then a morpheme is the smallest possible meaning unit. I.e., un + happy, un changes the meaning of the word.
I suspect this is true! it's interesting how advances in language models might wind up providing more hints about how the human brain works...
@@codetothemoon Yep! I lecture in cognitive psychology and there are a lot more parallels. I did a class with a huggingFace GPT model where you could change the token length to illustrate meaningful sentence production/comprehension (i.e., Wernicke's aphasia). I am fairly certain that an interdisciplinary field of academia will spawn in the next few years that marry machine learning and cognitive psychology. I'm sure there's the odd journal here and there but I think it'll take the place of conventional cognitive psychology.
2:24 tasks of NLP
How do I get the max amount of tokens supported by each model?
I'm trying to build a generic scraper that scrapes the given documentation of an app, finds all relevant REST APIs, webhooks etc along with the relevant details. I would then convert them into embeddings and insert into a database. I'll then search on these semantically.
What I'm struggling with right now is how to find the structure of the documentation. I think a LLM model would help with that but with most, I feel that I'll have a problem with the max number of tokens when I pass the HTML to them. I have access to GPT-4 but AFAIK, the 32k model hasn't been released yet. And even if I had access to it, I'll run into cost constraints.
🚨 Keep in mind, that many open source models DO NOT have commercial licensing
great point! I think I originally mentioned this in the video but I may have left it out of the final cut. But the one we're using in the video (T5 Flan) is Apache 2.0 and thus should be fine to use for commercial purposes.
@@codetothemoon Yes, the ones you showed are save. Just wanted to point it out… ran in a start-up last week that proudly pitched their product and than were devastated that a superficial tech DD reveled that they break licenses left and right. Licenses always have been the annoying accounting part everyone dreads, but for ML models it’s currently 10x as confusing. And probably the only reason, why the „you can actually run stuff locally“ needs extra care in that department. Anyway: Great video as always Mr Moon !
Before this Video, AI was blackbox for me, now i know basics, how it works
nice, really happy you got something out of the video!
Disappointed it didn't know about Apollo 10!
They teach how to import pretrained data but never show how to train models and create system of neural network layers to use it as normal dev would
all that AI stuff is just monopoly on computation patterns
pretrained models are an incredible resource, as the vast majority of developers don't have the resources or budget to perform such trainings. I'm not sure what you mean by "create a system of neural network layers to use as a normal dev would", as the models in this video are built upon neural networks
@@codetothemoon to show how to build NN matrixes with predictors for specific models like shape recognition or sound recognition
Thank you, however, believe most of the famous tech corps where aware of this space, prior 2017 paper (~7-5 yrs), and was masking this technology from the public space, though they used public dataset, Public should really seek for accountability using a proper legal framework, if only politicians were skillful, knowledgeable and talent enough, least they could seek from these tech corps is to provide cloud platform workspace for closer to a zero pricing scenario.
Awesome video, thanks, this helps!
Not a fan of clickbait title though
Thanks for the feedback! What do you think a better title would be?
@John *flan*-t5
I thought it was less clickbait and more tongue-in-cheek (in retrospect -- I didn't have the context until I watched the video).
@@John-wh9oh flan is an amazing dessert 🍮!
@@John-wh9ohflan is the name of the library primarily used in this video, which also happens to be the name of a dessert
** and * are called 'splat' operators when they're used this way: th-cam.com/video/tL1zltXuHO8/w-d-xo.html
thanks for pointing this out!
🎉
what in the hell is that line number schema? Lollll. Great video though!
Interesting, but let's see you do it from scratch in C.
Great presentation, but the presenter needs, IMHO, to speak more slowly
thanks and thanks for the feedback!
Thanks for the video 👍.
I just can't stop wondering, how come such an impossible poor, slow and inconvenient programming language with the development environment from the past century became a standard for machine learning tasks... It doesn't even have method names autocomplete, it doesn't show the full list of object's methods available, no parameters list, no types - just nothing. 💩 Crazy. return_tensors="pt" 🤦
Just a bit of food for thought ... all CPU chips ONLY understand how to manipulate bits and bytes in registers. EVERY computer language is simply a way to create those bit patterns to feed the hardware chip. Each language chooses how much abstraction to use and how the abstractions are interpreted. Some do it this way and some do it that way, but the end result in ALL cases is a sequence of machine (binary) codes that the Assembly language for the chip can write directly with NO abstractions except, perhaps, some mnemonics (memory aid) to avoid memorizing all the binary values associated with each machine instruction. Programming, paradoxically, is much easier in Assembly language than all others because you've gotten rid of all abstractions and interpretations that you have to wrap your head around. Of course, you still have the abstractions needed for solving the problem you are tackling but you don't have abstractions within abstractions. That's what makes computer programming challenging. Just sayin' ... 😉
I've never encountered a problem that was easier to solve in assembly than a higher level language. The abstractions of higher level languages are there to reduce our cognitive load, and they are incredibly effective at doing so
@@codetothemoon Absolutely, I agree. Abstracting things is important in many fields, not just programming. I was principally interested in peeling the onion to show why something like programming CAN be made more difficult because of abstractions but, of course, we have to understand that creating abstractions is an art in itself. It can be done well or done poorly. The effectiveness of using assembly language also depends on the programmer having a rock-solid understanding of the underlying electronics hardware. As an electronics engineer, I have had to work with many chips that didn't have compilers available so all coding had to be done in assembly and carefully mapped out with only the most rudimentary abstractions. This is also true of the microcoding that exists internally in every chip to interpret the binary machine code instructions themselves.
just came here for the thumbnail.
we're not "coders" a trained monkey can code.
we are programmers.
if you dont understand the difference, it's the same difference between writing and typing.
yeah......typing/coding is just hitting the keys in order to form the words. writing/programming is thinking.
not even gonna watch the video. the thumnail earned it a dislike.
Been self learning Arduino & Python on YT for several projects over the past few years. But, I have absolutely no idea what is being said in this video. Must be a different language than English. 🤷♂
I'll circle back around in another couple of years, after gaining some more smarts.
Thanks!
wow thank you so much for the support! much appreciated! really happy you got something out of the video.