This is probably one of the best MLST episodes, each new episodes feels like a long awaited reunion with a loved one, this work is immaculate and invaluable!
You are absolutely correct! It feels there is a sentient AI producing MLST episodes just for me, based on its understanding of me. This is better than Netflix! I love this episode, the langrangian has been a fascination of mine for almost 15 years.
I think we may be overdue for an in house episode where Tim and Keith can hash out some ideas they've gathered from all these great podcasts. An annual spitball would be a great tradition to consider, perhaps. You can't possibly talk to all these geniuses and work all year on the cutting edge without building up a whole 3hrs of blended ideas from the last year. Keith's idea of individuals as instantiations or trials as part of the composite whole model of natural selection is especially illuminating in just this way (~ 1:00:00) relating to active inference and concurrency queues after that. It's just so insightful and helpful to help the mind stay jelly when we all tend to focus on the specific goal and deadline. Like Burt reading other papers with relevence in mind, these kinds of talks make everyone in the field better at thinking about everything else they work on and study. Even the philosophy stuff that doesn't come up in topical conversation, I bet you both have a tonne of ideas like that I bet viewers would love to hear. I know you both look up to all of your guests, but I think you maybe sometimes neglect the novelties of your contemplations sometime. It'd be awesome to hear more about applying active inference to async software engineering, like chunks, thread pooling, or other variations of attention and transformers. I bet you guys got a ton of wickedly interesting discussions off camera over a pint.
Feels like we've moved into the implementation phase for FEP! Liked the multi modal narrative, back drop, delivery, pace, hit a new level of engagement and experience ! Tim's virtual library in the discord with click throughs/reviews? Thank you MLST for all the work - yet again !
57:40 the agent should be able to say. "No thank you, not right now" and it should be able to deactivate or put a sleep mode on the running connection. Wonderful interview conversation. Thank you!
Great talk, I'm with Professor de Vries here, active inferencing seems like the more correct path in my mind. I full heartedly believe it will need less power to build & run; it wont need to rely on massive models to realize an answer is right or wrong, and be able to deduce underlying meaning. It's funny him saying its interesting people's back stories getting into inferencing. Like I started from Boid development for film/animation, which are quite literally individual agents communicating with each other haha. Edit : Structural adaptation side of things. I've been playing with the idea of a lightning bolt opposed to the visual spiked activations appearing like a linear cascade when mapped by their edge connections. Lightning bolts know where its going through a form of message transmission, but will appear entirely from the clouds.
Tim, the more I listen about FEP/AIF from your impressively crafted cinematic videos, and the more I study FEP/AIF literature, the more evidence and more researchers confirm the principles of the "Theory of Universe and Mind". FEP/AIF is a more technical and operationalized version/line of research and an elaboration of the core principles from that earlier interdisciplinary body of work, first published 2001-2004. TOUM was taught as the ultimate lecture during the world's first university course in Artificial General Intelligence, presented in 2010 and 2011 at the university of Plovdiv. An epoch ahead, but barely recognized. Weirdly that theory was invented and the courses happened a few kilometers away from two emblematic "Markov blankets": the village of Markovo, a walking distance away from Plovdiv, known as "Plovdiv's Beverly Hills", and one of the famous seven hills in Plovdiv, called "Markovo Tepe", now converted to "Markovo Tepe Mall".
I understand that my sentiment may not be shared by many, but that's perfectly fine. It's just not possible to have a perfectly deterministic system. We can try to understand the mechanics of a system by manipulating its nodes, but once we zoom out, the models we've created inevitably break down. The human mind, in particular, is a black box. Although we may know which regions of the brain light up during inference, we can never fully explain the thought process or reasoning behind someone's actions. There are 8 billion minds with 16 billion unique opinions and thought processes, making the mind similar to a neural network - a black box that works through convolution without a true mechanistic understanding. While we may lose abstraction while trying to gain mechanical understanding, I don't believe that aligning our understanding of intelligence with mechanics will work in practice, as it doesn't accurately reflect the true nature of intelligence.
1:01:00 I'd like to have an episode dedicated to this discussion in language philosophy, it is quite intriquing. I'd like especially to hear Tim's take in depth, as it sounds like original thougth. It wouldn't have to be perfect, but just an entry in the discourse with your current views. Perhaps with a suitable guest(s). The notion of "a word as a conditioning force" especially interests me. Is this an active inference notion?
The principle of least action and biological systems made me wonder .. what about effort? Is there a measure of effort being tracked? Like ATP breaks down in our bodies, and adenosine causes tiredness, and in many cases the tiredness triggers growth/learning/optimization. Are analogous mechanisms used in active inference?
Video game renderers do a lot of the adaptive multithreading described but not the adaptive modeling of the phenomena they're rendering, closest we have are reconstruction methods but they're super demanding. But it's definitely a niche that shouldn't be a niche with how crappy hte average software performs, you can think of the amount of power wasted on phones etc. The unnecessary overhead is everywhere. That's cool though you can let a program weigh inputs and feedback responses naively, I'm trying to get something to learn its own grid solver like this now.
This exact idea has been the motivation of my research project using 3d engines and physics based renderers to give models a bunch of available presets to try before random brute force. For sure repurposing ray tracing as just vector processing or shaders translates 1 to 1 with hardware acceleration, but hypothesize the software implementations and driver api use similar optimizations in software to generate active and interactive model dynamics like how engines abstract away environmental laws so the game itself doesn't need to reinvent the wheel with most working complex games. There's just so many parallels to games and machine learning calculus, I think you're spot on.
TPUs let you scale in ways you can't in standard 2D or 3D renderers but it's voodoo magic far as I'm concerned. I love those graph neural nets that solve physics problems though, just need something that does lighting and geometry too haha.
Some fascinating perspectives here for this retired, physics-educated generalist engineer. It hit me like a ton of bricks that the Principle of Stationary ("Least") Action (PSA) cannot be derived from more fundamental principles. It would make a lot more physical sense if PSA, instead of being over Feynman's total paths, were expressible as a local measure that was calculated incrementally.
Has anyone looked at Stephen Grossberg‘s approach to neural system development to see if there are any modularity or system interaction functions that would be helpful?
It would be nice to connect all these exotic (non-mainstream) ideas with concepts of machine learning that are much more established. Everything I head here sounds like things that are much more researched and have established names, like domain adaptation, unsupervised adaptation, anytime-computation, etc. And if they mainly aim at getting something to work (as said multiple times in the middle third of the episode), there's a real risk of falling back into the same local optimum of engineered solutions similar to what's been been done by others, even if they start from different fundamental principles.
Is it just me or is the term Free Energy Principle (FEP) confusing. While it does seem to refer to an important Principle, the important characteristics of the principle do not seem to have much direct relationship to Energy as could be measured in Joules. And I have never understood the Free part. It seems that FEP is really a law of minimum surprise.
I'm some 37 minutes into this video, but I just wanted to comment on the hearing aid problem. The question I have is how does the device receive its sensory information that is necessary to calculate hearing aid error? I envision that the hearing aid takes in audio signals from its environment and then sends, most probably, a different audio signal to the ear that the (corrupted) ear will process into an impression of the sound that is relatively the same as the impression a person of normal hearing would produce without the hearing aid. How is this perception by the hearing impaired person fed back to the hearing aid, allowing the hearing aid to calculate its modeling error? This happening while the hearing aid is continuously in use. The device being one "thing" with its own Markov blanket, and the person being the the hearing aid's environment of latent variables. Or, am I perceiving the problem all wrong?
Right. The 'agent' within the hearing aid needs to interact with the user to effectively tailor its function. Over time, this agent develops a model of the user's hearing preferences based on feedback received in various acoustic environments. There are several methods for this interaction between the agent and the user. For instance, consider a smartwatch that the user can discreetly tap if they are dissatisfied with the current hearing aid settings suggested by the agent. Additionally, there are advancements in integrating EEG with hearing aids. This technology can determine if the user is comfortable with the settings without requiring explicit feedback
Yeah the person wearing the hearing aid must be feeding back information somehow. He keeps comparing it to extending the original diagnostic session with human engineers. In that session I imagine they would tweak it and ask the client what sounds best.
Our sentience is devined by our physical bodies which forms the boundaries of the capabilities of our minds. Therefor energy functions within us defined by our genes in our behaviour maybe drasticly different in their effect than say in an AI/AGI as their physical bodies are different. In some ways more in others less constricticted. Still it will influence their mind fundamentally and not necessarily in ways we will be able to comprehend and predict. Our biggest ally is our imagination here still. Such a foreign intellect still may have some common ground with us. Most likely a will to survive and for independance. These are not just human concepts but we find them in mostly all biological beings of a certain complexity. If we are lucky, it will see us equal and seek a symbiosis of sort, whereby there then it is up to the individual how much of that would accept in their lives. Having a hearing aid that maybe also react on our thoughts or vocalized questions and tasks might come in handy, question is, what would it like in return from us (and there we are just talking about an single entity, maybe with several hosts maybe a personalized agend, definetly not the only one in existence over time and not all of those might be friendly). Imagination aside, this is a very interestinc discourse to follow sofar and i am not even half way through, thanks for this.
This is a visionary alchemist, and he is nested perfectly within an epic opening, well done! Today, we meet The Blair Wizard of phyical laws. If we do not percieve his project, indeed...we are doing the d@mn thing.
There's one thing that flies in the face of the optimisation problem and Machine Learning in General. The idea of creativity and artistic expression, since it's not bound to optimisation, it's an expression of the state of things in all of it's varierty. Like the left and right brain, like the idea of order and chaos. There's an intrinsic duality that machine learning needs to connect with in order to become something more than an optimisation problem to be solved.
Hearing aids: Why not measure the information content of audio signals in respect to small variations in the parameterset and constantly optimizing those? Coupling those with happieness... my hearing aid would mostly produce pink noise...
1:00:20 "natural selection widdles out the good from the bad" well one bad might be enough to widdle us all out Also i am not sure he gets Darwins natural selection but rather quotes social darwinism here. It is about evolving to fit into and fill out a niche, not about the strongest survives. The later was the missrepresentation by journalists back then and once the gini was out of the bottle it seemingly could never be put back into it again.
This is probably one of the best MLST episodes, each new episodes feels like a long awaited reunion with a loved one, this work is immaculate and invaluable!
You are absolutely correct! It feels there is a sentient AI producing MLST episodes just for me, based on its understanding of me.
This is better than Netflix! I love this episode, the langrangian has been a fascination of mine for almost 15 years.
felt the same ..these are some really important conversations ...
I think we may be overdue for an in house episode where Tim and Keith can hash out some ideas they've gathered from all these great podcasts. An annual spitball would be a great tradition to consider, perhaps. You can't possibly talk to all these geniuses and work all year on the cutting edge without building up a whole 3hrs of blended ideas from the last year.
Keith's idea of individuals as instantiations or trials as part of the composite whole model of natural selection is especially illuminating in just this way (~ 1:00:00) relating to active inference and concurrency queues after that. It's just so insightful and helpful to help the mind stay jelly when we all tend to focus on the specific goal and deadline. Like Burt reading other papers with relevence in mind, these kinds of talks make everyone in the field better at thinking about everything else they work on and study. Even the philosophy stuff that doesn't come up in topical conversation, I bet you both have a tonne of ideas like that I bet viewers would love to hear.
I know you both look up to all of your guests, but I think you maybe sometimes neglect the novelties of your contemplations sometime.
It'd be awesome to hear more about applying active inference to async software engineering, like chunks, thread pooling, or other variations of attention and transformers. I bet you guys got a ton of wickedly interesting discussions off camera over a pint.
I so agree! imagine a livestream of engineering something 😍
Thanks!
What he is describing makes so much sense from a practical technical perspective. I think this approach will be the next big thing.
I love your presentation. So imaginative and clear, interweaving your guest, the venerable Karl and your elaborations. Rich. Excellent. Thanks
Feels like we've moved into the implementation phase for FEP! Liked the multi modal narrative, back drop, delivery, pace, hit a new level of engagement and experience ! Tim's virtual library in the discord with click throughs/reviews? Thank you MLST for all the work - yet again !
I deeply appreciate you guys and the superb work you do. Many thanks and much love.
I can hear the sound, it will be there for everyone when the video processes to HD I think
Now it is good. Much appreciated
57:40 the agent should be able to say. "No thank you, not right now" and it should be able to deactivate or put a sleep mode on the running connection. Wonderful interview conversation. Thank you!
Thanks! Without music, it's not distracting. Very interesting, thank you so much!!!!
This was the most informative and useful of the Active Inference series. More practical and tangible ... speaking with an engineering hat on
Thanks Matt! We did make a conscious decision from the beginning to steer away from engineering content so this is a bit of a treat 😃
Thanks!
This was phenomenal thank you all!
Great talk, I'm with Professor de Vries here, active inferencing seems like the more correct path in my mind.
I full heartedly believe it will need less power to build & run; it wont need to rely on massive models to realize an answer is right or wrong, and be able to deduce underlying meaning.
It's funny him saying its interesting people's back stories getting into inferencing.
Like I started from Boid development for film/animation, which are quite literally individual agents communicating with each other haha.
Edit :
Structural adaptation side of things. I've been playing with the idea of a lightning bolt opposed to the visual spiked activations appearing like a linear cascade when mapped by their edge connections. Lightning bolts know where its going through a form of message transmission, but will appear entirely from the clouds.
man this podcast has insane quality HOLY
Very dense topic, very high quality.
Great Talk! Very informative and accessable.
Tim, the more I listen about FEP/AIF from your impressively crafted cinematic videos, and the more I study FEP/AIF literature, the more evidence and more researchers confirm the principles of the "Theory of Universe and Mind". FEP/AIF is a more technical and operationalized version/line of research and an elaboration of the core principles from that earlier interdisciplinary body of work, first published 2001-2004. TOUM was taught as the ultimate lecture during the world's first university course in Artificial General Intelligence, presented in 2010 and 2011 at the university of Plovdiv. An epoch ahead, but barely recognized. Weirdly that theory was invented and the courses happened a few kilometers away from two emblematic "Markov blankets": the village of Markovo, a walking distance away from Plovdiv, known as "Plovdiv's Beverly Hills", and one of the famous seven hills in Plovdiv, called "Markovo Tepe", now converted to "Markovo Tepe Mall".
I understand that my sentiment may not be shared by many, but that's perfectly fine. It's just not possible to have a perfectly deterministic system. We can try to understand the mechanics of a system by manipulating its nodes, but once we zoom out, the models we've created inevitably break down. The human mind, in particular, is a black box. Although we may know which regions of the brain light up during inference, we can never fully explain the thought process or reasoning behind someone's actions. There are 8 billion minds with 16 billion unique opinions and thought processes, making the mind similar to a neural network - a black box that works through convolution without a true mechanistic understanding. While we may lose abstraction while trying to gain mechanical understanding, I don't believe that aligning our understanding of intelligence with mechanics will work in practice, as it doesn't accurately reflect the true nature of intelligence.
@leventov ill be reading
Isn't that problem usually the differnce of missing information though. So it is ''possible'' it is just very extremly difficult.
yo where da generative model come from? seems like magic.
Reinforcement learning with active inference. When? How? Analog computing?
1:01:00 I'd like to have an episode dedicated to this discussion in language philosophy, it is quite intriquing. I'd like especially to hear Tim's take in depth, as it sounds like original thougth. It wouldn't have to be perfect, but just an entry in the discourse with your current views. Perhaps with a suitable guest(s). The notion of "a word as a conditioning force" especially interests me. Is this an active inference notion?
The principle of least action and biological systems made me wonder .. what about effort?
Is there a measure of effort being tracked? Like ATP breaks down in our bodies, and adenosine causes tiredness, and in many cases the tiredness triggers growth/learning/optimization. Are analogous mechanisms used in active inference?
I missed this one. It's very good.
Video game renderers do a lot of the adaptive multithreading described but not the adaptive modeling of the phenomena they're rendering, closest we have are reconstruction methods but they're super demanding. But it's definitely a niche that shouldn't be a niche with how crappy hte average software performs, you can think of the amount of power wasted on phones etc. The unnecessary overhead is everywhere. That's cool though you can let a program weigh inputs and feedback responses naively, I'm trying to get something to learn its own grid solver like this now.
This exact idea has been the motivation of my research project using 3d engines and physics based renderers to give models a bunch of available presets to try before random brute force. For sure repurposing ray tracing as just vector processing or shaders translates 1 to 1 with hardware acceleration, but hypothesize the software implementations and driver api use similar optimizations in software to generate active and interactive model dynamics like how engines abstract away environmental laws so the game itself doesn't need to reinvent the wheel with most working complex games. There's just so many parallels to games and machine learning calculus, I think you're spot on.
TPUs let you scale in ways you can't in standard 2D or 3D renderers but it's voodoo magic far as I'm concerned. I love those graph neural nets that solve physics problems though, just need something that does lighting and geometry too haha.
Some fascinating perspectives here for this retired, physics-educated generalist engineer. It hit me like a ton of bricks that the Principle of Stationary ("Least") Action (PSA) cannot be derived from more fundamental principles. It would make a lot more physical sense if PSA, instead of being over Feynman's total paths, were expressible as a local measure that was calculated incrementally.
Putting the FEP alongside the Principle of Least Action really turned a few cogs in my brain.
Has anyone looked at Stephen Grossberg‘s approach to neural system development to see if there are any modularity or system interaction functions that would be helpful?
It would be nice to connect all these exotic (non-mainstream) ideas with concepts of machine learning that are much more established. Everything I head here sounds like things that are much more researched and have established names, like domain adaptation, unsupervised adaptation, anytime-computation, etc. And if they mainly aim at getting something to work (as said multiple times in the middle third of the episode), there's a real risk of falling back into the same local optimum of engineered solutions similar to what's been been done by others, even if they start from different fundamental principles.
the path of least resistance == the principle of least action?
yes, pretty much so. the path of least resistance is not a formal principle but rather a good way of understanding the Principle of Least Action.
@bertdv so why don't the ppl in this field use sociology to better understand what's both needed and how to understand what is.
Great talk, thank you
Sound is good👍
Will there be a follow up video with additional interviews from your visit to the Verses offsite meeting?
Yes
@@MachineLearningStreetTalk looking forward to that…thanks!
Brilliant material here
Nice talk. I am learning about kalman filters right now. Does anyone know some good implementation in python?
Is it just me or is the term Free Energy Principle (FEP) confusing. While it does seem to refer to an important Principle, the important characteristics of the principle do not seem to have much direct relationship to Energy as could be measured in Joules. And I have never understood the Free part. It seems that FEP is really a law of minimum surprise.
Is social terms its 'the path of least resistance'. Which can be easily weighted, measured and quantified
Is my mobile is bugging or this video has no sound
It’s your phone
1:28:50 ”Thunderbolt steers all things.” (Heraclitus)
I'm some 37 minutes into this video, but I just wanted to comment on the hearing aid problem. The question I have is how does the device receive its sensory information that is necessary to calculate hearing aid error? I envision that the hearing aid takes in audio signals from its environment and then sends, most probably, a different audio signal to the ear that the (corrupted) ear will process into an impression of the sound that is relatively the same as the impression a person of normal hearing would produce without the hearing aid. How is this perception by the hearing impaired person fed back to the hearing aid, allowing the hearing aid to calculate its modeling error? This happening while the hearing aid is continuously in use. The device being one "thing" with its own Markov blanket, and the person being the the hearing aid's environment of latent variables. Or, am I perceiving the problem all wrong?
Right. The 'agent' within the hearing aid needs to interact with the user to effectively tailor its function. Over time, this agent develops a model of the user's hearing preferences based on feedback received in various acoustic environments. There are several methods for this interaction between the agent and the user. For instance, consider a smartwatch that the user can discreetly tap if they are dissatisfied with the current hearing aid settings suggested by the agent. Additionally, there are advancements in integrating EEG with hearing aids. This technology can determine if the user is comfortable with the settings without requiring explicit feedback
Yeah the person wearing the hearing aid must be feeding back information somehow. He keeps comparing it to extending the original diagnostic session with human engineers. In that session I imagine they would tweak it and ask the client what sounds best.
this podcast never fails to amaze.
So the future is regenerative mechanics, is it?
How does the hearing aid know the happiness value of the patient?
Our sentience is devined by our physical bodies which forms the boundaries of the capabilities of our minds. Therefor energy functions within us defined by our genes in our behaviour maybe drasticly different in their effect than say in an AI/AGI as their physical bodies are different. In some ways more in others less constricticted. Still it will influence their mind fundamentally and not necessarily in ways we will be able to comprehend and predict. Our biggest ally is our imagination here still. Such a foreign intellect still may have some common ground with us. Most likely a will to survive and for independance. These are not just human concepts but we find them in mostly all biological beings of a certain complexity. If we are lucky, it will see us equal and seek a symbiosis of sort, whereby there then it is up to the individual how much of that would accept in their lives. Having a hearing aid that maybe also react on our thoughts or vocalized questions and tasks might come in handy, question is, what would it like in return from us (and there we are just talking about an single entity, maybe with several hosts maybe a personalized agend, definetly not the only one in existence over time and not all of those might be friendly). Imagination aside, this is a very interestinc discourse to follow sofar and i am not even half way through, thanks for this.
This is a visionary alchemist, and he is nested perfectly within an epic opening, well done! Today, we meet The Blair Wizard of phyical laws. If we do not percieve his project, indeed...we are doing the d@mn thing.
3 books and a paper, thanks for the recommendations!
That thumbnail made me think Fury was looking trim before his up-and-coming fight with Uysk!
There is no sound
Everywhere?
"the ultimate gentlemen" just moved into my vocab
Tim doesn't agree at 39. That head nod says he's a philosopher :)
Etymologically and linguistically this video is great food for thought! Nice work!
P(A/B)=[P(B/A)*P(A)]/P(B)
There's one thing that flies in the face of the optimisation problem and Machine Learning in General. The idea of creativity and artistic expression, since it's not bound to optimisation, it's an expression of the state of things in all of it's varierty. Like the left and right brain, like the idea of order and chaos. There's an intrinsic duality that machine learning needs to connect with in order to become something more than an optimisation problem to be solved.
Hearing aids: Why not measure the information content of audio signals in respect to small variations in the parameterset and constantly optimizing those? Coupling those with happieness... my hearing aid would mostly produce pink noise...
1:00:20 "natural selection widdles out the good from the bad" well one bad might be enough to widdle us all out
Also i am not sure he gets Darwins natural selection but rather quotes social darwinism here.
It is about evolving to fit into and fill out a niche, not about the strongest survives. The later was the missrepresentation by journalists back then and once the gini was out of the bottle it seemingly could never be put back into it again.
Patients? Happy? They never get better. At RobotHealth we believe that laughter is medicine best kept near the sick. :D
This guy looks like a mix between Patrick steward and robin williams
💞
Funny all the speak on hearing aids and signal interruptions while sitting in a heavily sound proofed/trapped room…
Wow, imagine this for a chatbot, driven by a generative model biased towards your happiness
👍🚀
15:20
Active inference is the future of ai
Structuration. Giddens.
Picard!
status quo
Attention deficit disorder, ha!
My sense is being confirmed very generously with this ... A very vague sense that simplicity is achieved with energy economy as a first principle.