FYI you can dramatically increase the quality of the resulting CSV file by having chat GPT smooth out the curves BEFORE you even import. Just upload the CSV file to GPT 4o and say "The following CSV file contains many columns atributes pertaining to a 3d animation. When applied, these values shake erratically. I was wondering if you could smooth these values out for me for a smooth animation. Each row represents the next frame in a 30 frames per second animation" Then just download the file it creats and B A M even though those instructions are a little iffiy, every time it creates a great results for me, removing the jitter completely.
Bro, you are a live saver, was going to buy a second hand iPhone. You saved me, you don't know, how happy I am right now!!! Spending months and months to find how to animate the face, finally I can. Thank you bro
Dear friend! What you showed is just super! You are saving my life! You pulled me out of such a swamp! I spent days on the Internet scrolling for hours, watching useless tutorials on paid programs and plugins with terribly complex settings that don’t really work, until I came across your video. But it turned out to be so simple and easy, just three clicks of the mouse. Well, I want to say that there is one very important nuance: in order for the animation not to tremble or twitch, you need to prepare the source video well. That is, the face must be on a white background. There should not be a face on this screen, the face should be like in the passport photos. You don’t have to put your face directly into the camera, you have to shoot from afar. The video needs to be run through an editor, contrast, gamma, and brightness should be added to it in black and white so that the eyebrows and pupils of the lips look good on a white background. And it is desirable if it is possible to record video at 60 frames per second, so that the program calculates not 24 frames, but 60. Then the animation is rendered perfectly and nothing twitches or shakes. A huge heartfelt thank you for your help!
Two questions. Firstly how do I bake it to control rig properly so I can combine it with other animations and blend it with the body animation? And secondly I’m having a small issue where if u look to the side in the animation the metahuman looks down instead…
ooo, I didn't have those issues, to bake to the control rig I just did the normal workflow (but sometimes it gets woonky), the look down issue I need to test it a bit more, didn't have much time to do so.
Extremely usefull when you do not own an apple gadget... Thanks a lot. Any fixes in view for mouth being a bit funky ? How do you smooth curves ? By hand or is there any automation in UE for doing so ?
I've seen people looking a lot for this, I have some ideias to aproach it but still need some time to test it! If I get something usable going I will do it for sure!
It works, but once imported into unreal, it skips 15 frames of data after every 15 frames. So if you have a 100 frame animation it will skip 16-30, 46-60, 76-90 .. Any possible solution ?
@@spaceorcastudioThank you for the reply. I am looking forward to your next upload regarding realtime face capture. You should know that I solved the problem of frame skipping by recording the video in 60fps instead of 30.
Thank you for this video, but is it possible to have some technical details and sorry for my broken English, it's Google trad, I'm French. Kind regards, Jean-Pascal
@@spaceorcastudio Ok, looking forward to your new video and new method, if you can use livelink in real time in the animation blueprint, you can also freely change the strength of each attribute of the blendshape
It's funny they use the iPhone exclusively for face capture... No one uses iPhones this the lack of any facial animations from any independent animator. I wish they'd notice that.
Yea, mainly from what I know it's because of the depth sensor. But anyway on the other hand there is xr animator that can send facial data (I have a video but still need some work, check the offworld live youtube channel) Nvidia launched an sdk for face capture using rtx gpus ( thats how iclone can capture faces with accu face). One the same point as the accu face for face capture (using rtx gpu) you can try Ifacial mocap ( it doesn't support metahumans yet) but it works with blender and unreal if you use arkit 52 blend shapes! Anything hit me up on discord lucasrodrigo
Sir I have a Sci-fi short story. But there are no computer systems. I want to make a short film with my friends through Mobile. It needs vfx help. Can you do it ?.
Very interesting solution. Yeah its not great, but very useful. For people that want to have a better result, its possible to use this result as a starting point, I think. As you said, combine with other solutions too. thanks for sharing!!! Lol, I was one that said its was not good 😅, but I see now its value. It helps and its free. Thats great !!!
FYI you can dramatically increase the quality of the resulting CSV file by having chat GPT smooth out the curves BEFORE you even import. Just upload the CSV file to GPT 4o and say "The following CSV file contains many columns atributes pertaining to a 3d animation. When applied, these values shake erratically. I was wondering if you could smooth these values out for me for a smooth animation. Each row represents the next frame in a 30 frames per second animation" Then just download the file it creats and B A M
even though those instructions are a little iffiy, every time it creates a great results for me, removing the jitter completely.
wow, interesting, thanks for the info!
it says i reached the limit for advanced data analysis
you are a legend
Thank you so much for this
Bro, you are a live saver, was going to buy a second hand iPhone. You saved me, you don't know, how happy I am right now!!! Spending months and months to find how to animate the face, finally I can. Thank you bro
Dude, you are a hero especially offering it for free. Thanks!
Straight to the point and very helpful. I'll give it a try. Thank you so much.
There is a extremely useful tool. Thanks for the tutorial.
How do I move a Live Link track to another Sequence in Unreal Engine 5?
Works great, thank you!
Muito obrigado mano, tava precisando disso pras cutscenes do meu game indie 😁
Dear friend!
What you showed is just super!
You are saving my life! You pulled me out of such a swamp!
I spent days on the Internet scrolling for hours, watching useless tutorials on paid programs and plugins with terribly complex settings that don’t really work, until I came across your video.
But it turned out to be so simple and easy, just three clicks of the mouse.
Well, I want to say that there is one very important nuance: in order for the animation not to tremble or twitch, you need to prepare the source video well. That is, the face must be on a white background. There should not be a face on this screen, the face should be like in the passport photos. You don’t have to put your face directly into the camera, you have to shoot from afar.
The video needs to be run through an editor, contrast, gamma, and brightness should be added to it in black and white so that the eyebrows and pupils of the lips look good on a white background.
And it is desirable if it is possible to record video at 60 frames per second, so that the program calculates not 24 frames, but 60. Then the animation is rendered perfectly and nothing twitches or shakes.
A huge heartfelt thank you for your help!
hey I'm glad it helped! also thanks for the infor, I'm very busy and couldn't update the workflow to some new fancy stuff.
For some reason this programm exists with error when I load video
WOO! Bless and thanks for that! 🙏🤟
Thank you so much for your video! This is a real game changer! Im wondering if there the same for body movement!
Two questions. Firstly how do I bake it to control rig properly so I can combine it with other animations and blend it with the body animation? And secondly I’m having a small issue where if u look to the side in the animation the metahuman looks down instead…
ooo, I didn't have those issues, to bake to the control rig I just did the normal workflow (but sometimes it gets woonky), the look down issue I need to test it a bit more, didn't have much time to do so.
Bro I am new please help me I have same problem to bake animation please guide me how to do
I tried everything and it's the best I got. I can't buy an iPhone. Thank you! You rules!
you can also try xr animator, it's a bit tricky but it works in realtime :)
@@spaceorcastudio thanks!
Extremely usefull when you do not own an apple gadget... Thanks a lot. Any fixes in view for mouth being a bit funky ? How do you smooth curves ? By hand or is there any automation in UE for doing so ?
mainly try to smooth the curves from the mounth on the graph editor inside unreal!
Thank you so much bruh.
can u pls say how to bake this to control rig ? i want to export this to maya
how to combine body and face anim in sequencer?
thank you, i will be try
amazing 🤩🤩🤩
can you do Free Face Mocap from any video
I've seen people looking a lot for this, I have some ideias to aproach it but still need some time to test it!
If I get something usable going I will do it for sure!
looks promising, i tried it just now, and it always looks like the input-data was taken at a VERY cold place....
shake shake shake the head ^^
haha, try to smooth the curves a bit in the graph editor
thanks bro .
i found some thing like this for many time.
again thanks.
Very valuable information mate
But How to do this for custom charecter with ARKit morphs?
just use the same file on your character, the csv file is in arkit format
It works, but once imported into unreal, it skips 15 frames of data after every 15 frames. So if you have a 100 frame animation it will skip 16-30, 46-60, 76-90 .. Any possible solution ?
I think it's a problem with the tracking software, but no worries soon I will post another face cap method that is super flexibe and also realtime!
@@spaceorcastudioThank you for the reply. I am looking forward to your next upload regarding realtime face capture. You should know that I solved the problem of frame skipping by recording the video in 60fps instead of 30.
oooo great, that is awesome!@@j27vivek
Nice bruh, but we can't live🤧.
For some reason for me the file outputs without any keyframes... I've tested with different videos and nothing seems to work..
I have no idea why, I need to make an updated video about face cap, but I'm 100% out of time right now :(
how the hell did u import it ???? i cant use it as a level sq*
1:28 of the video (livelink face importer plugin)
how to use it with custom character
you can check this video:
th-cam.com/video/L8RckB9ZRJE/w-d-xo.html
your character has to have the apple arkit blendshapes and you are good to go!
14 year old boy and he already have a beard..
Wild hormones
@@spaceorcastudio at least bread 🍞
He's been doing unreal videos for 14 years lol
Can tbis be used live? Or just rexorded? Thanks!
not with this method, take a look at xr animator.
There is a video on the channel
Thank you for this video, but is it possible to have some technical details and sorry for my broken English, it's Google trad, I'm French. Kind regards, Jean-Pascal
What technical details?
That's it, problem solved, thank you for your response and your video. I just subscribed and liked. Kind regards. Jean-Pascal
@@tcmtheatredelacomedie8054 just import it, you need to csv plugin to import in unreal
you are best
Is there a way to live stream this data through livelink and modify it in the UE animation blueprint?
not with this workflow, with the new one I will make a video about probably we can do it!
@@spaceorcastudio Ok, looking forward to your new video and new method, if you can use livelink in real time in the animation blueprint, you can also freely change the strength of each attribute of the blendshape
bro my animation is jittery is there any fix for that
Try looking at the curves in unreal, and smooth them out a bit
its saying data table
1:28 of the video (livelink face importer plugin)
It's funny they use the iPhone exclusively for face capture... No one uses iPhones this the lack of any facial animations from any independent animator. I wish they'd notice that.
Yea, mainly from what I know it's because of the depth sensor. But anyway on the other hand there is xr animator that can send facial data (I have a video but still need some work, check the offworld live youtube channel)
Nvidia launched an sdk for face capture using rtx gpus ( thats how iclone can capture faces with accu face).
One the same point as the accu face for face capture (using rtx gpu) you can try Ifacial mocap ( it doesn't support metahumans yet) but it works with blender and unreal if you use arkit 52 blend shapes!
Anything hit me up on discord lucasrodrigo
Sir I have a sci-fi story. I want to make it as a short film. But I have no money and no software. Can you tell me how to do it?
now this depends on your hardware, what are your specs?@@THEKINGMAKER9041
Sir I have a Sci-fi short story. But there are no computer systems. I want to make a short film with my friends through Mobile. It needs vfx help. Can you do it ?.
Oooo, without a pc I have no idea how you could do it
my result is shaking a lot,,,, anyone knows the solution??
did you solve that problem ?
@@alessandrosirsi5902 unfortunately... not yet
oo sorry for the delay, I'm not using this tool anymore, I was going to do an update video but I'm out of time right now
why we got space empty .? its a stop ? how fix that ?
where? when? I need more information
@@spaceorcastudio when you put the sequence in timeline , you can see it in his video _ _ _ _ _ _ its like that nlt in continue why ?
ooo now I undestand, I have no idea actually@@MickaelSchaack
thank u alot
can i import this in blneder ??
Probably with some addon that supports importing csv files :)
Thank u
does the script work on linux too?
I have no idea, I actually need to record the new method
@@spaceorcastudio what new method?
@@GameDevGeeks for face capture, it's more or less the same but it runs at "realtime"
@spaceorcastudio does it work with any video or do we need true depth sensor with an IPhone?
@@GameDevGeeks works with a webcam :)
Very interesting solution. Yeah its not great, but very useful.
For people that want to have a better result, its possible to use this result as a starting point, I think. As you said, combine with other solutions too.
thanks for sharing!!!
Lol, I was one that said its was not good 😅, but I see now its value. It helps and its free. Thats great !!!
haha, all good man, I'm going to do somemore test calibrating the values in unreal, the mouth is a bit funky
i had bought iphone x yesterday....😂
🤣🤣 you can capture in realtime and with way more precision now :)
Nice bruh, but we can't live🤧.
Nice bruh, but we can't live🤧.