Comments (9)
3D facial tracking is used to obtain 3D facial landmarks and head poses. Deca is an alternative option to get these parameters. In fact, any parametric 3D facial reconstruction method is considerable. Camera calibration is used to achieve better reconstruction.
from livespeechportraits.
Thanks,I still have some questions about the two preprocessed data. How can I get the id_scale.mat and shoulder_ Point3d (I know this is calculated through LK optical flow, but how to get the key points of 3D)
from livespeechportraits.
id & scale parameters are results of 3D facial tracking.
To get 3D shoulder points, we first detect 2D shoulder points using LK flow and reconstruct 3D shoulder points by assuming a billboard model where the depth is the mean 3D facial depth for each frame.
from livespeechportraits.
thank you again ! I trained the audio2feature and audio2headpose modules. Here are some questions I hope you can answer
-
How can I get a linear velocity term in Δ Pt audio2headpose
-
Whether tracked 3D in audio2feature is normalized? I used [0,1] normalization. It seems that there are some problems
-
In the article, proj in Fig. 2 stands for camera projection. How can I integrate the mouth headpose and eye information through this? My understanding is to convert 3D into 2D through the camera. So the rest of the landmark information adopts the original landmark of the video?
-
Whether HeadPose is relevant with contour(landmark) ?
from livespeechportraits.
- velocity head pose is just the delta speed, velocity_t = pose_t+1 - pose_t. It is an implicit supervision term. You can define it as you will.
- The 3D facial landmarks lie in the linear 3DMM space, I didn't do additional normalization on it. The template face model is normalized in a fixed space actually for any 3DMM model.
- Camera projection depends on the 3d face tracking algorithm you used. Landmarks like eyes can just be sampled from the dataset, of course in 3D space. All these should be done before the projection.
- Head pose is the tracking results of the reconstruction results. No relationship exists between the pose and the contour landmarks.
from livespeechportraits.
- As for 3D face tracking, whether it is obtained from the original image(original video is 1920*1080)and whether it is affected by camera calibration, If I using deca to calculate the 3D key points for each frame,what is the different with two methods?
- What method is used to obtain GroundTruth of the HeadPose? ( I use OpenFace2.2.0 method to get the HeadPose GroundTruth )
Thank you again for publishing the code. Thank you very much!
Hello, I have two questions, can you help me. I use the method of OpenFace2.2.0 to get the HeadPose GroundTruth.
- How can I get the tracked3D_normalized_pts_fix_contour.npy and 3d_fit_data.npz
- And I know what they mean?
thank you so much for your marvelous work.
from livespeechportraits.
These two files are 3d facial tracking results (3D landmarks, head poses, etc.). Check the inference code and replace them with your tracking results works.
from livespeechportraits.
As I understand, the papers 1 and 2 (mentioned in paragraph 2 of sec 4.1) are used to extract the face landmarks. Did you just use the paper codes, or there are off-the-shelf tools in python?
[1] Automatic Acquisition of High-fidelity Facial Performances Using Monocular Videos.
[2] Face2Face: Real-time Face Capture and Reenactment of RGB Videos.
from livespeechportraits.
As I understand, the papers 1 and 2 (mentioned in paragraph 2 of sec 4.1) are used to extract the face landmarks. Did you just use the paper codes, or there are off-the-shelf tools in python?
[1] Automatic Acquisition of High-fidelity Facial Performances Using Monocular Videos. [2] Face2Face: Real-time Face Capture and Reenactment of RGB Videos.
maybe you can see, foocker/LSP
from livespeechportraits.
Related Issues (20)
- Personalized data generation HOT 3
- What is the meaning of implementing by C++? HOT 1
- 候选照片,一共四张,是基于什么逻辑进行选择的?
- how can i use it in real time? HOT 1
- Does anyone implement the training code of this project? HOT 1
- How to run demo in "Real-time" HOT 1
- 模型得到的矩阵值可以和ARkit进行映射吗?
- RuntimeError: Found no NVIDIA driver on your system.
- Great project, where does the author achieve real-time performance? HOT 2
- 如何生成自己的模型。从哪里导入我的视频素材生成我自己的模型。
- How to train these models in custom dataset? Any documentation? HOT 1
- What tool did you use to create a sketch from a face image, in case i want to train the image to image transition model?
- 73 facial landmarks HOT 1
- FileNotFoundError: [Errno 2] No such file or directory: './data/May\\mean_pts3d.npy' HOT 1
- 数字人技术交流群请联系VX:metahuman668
- GMMLogLoss for training audio2headpose
- training data download
- Is the Released Models Trained on Whole Video Clip?
- code for data processing, training HOT 2
- REAL TIME 哪里去了?不是说好可以根据音频流来实时输出吗? HOT 2
Recommend Projects
-
React
A declarative, efficient, and flexible JavaScript library for building user interfaces.
-
Vue.js
🖖 Vue.js is a progressive, incrementally-adoptable JavaScript framework for building UI on the web.
-
Typescript
TypeScript is a superset of JavaScript that compiles to clean JavaScript output.
-
TensorFlow
An Open Source Machine Learning Framework for Everyone
-
Django
The Web framework for perfectionists with deadlines.
-
Laravel
A PHP framework for web artisans
-
D3
Bring data to life with SVG, Canvas and HTML. 📊📈🎉
-
Recommend Topics
-
javascript
JavaScript (JS) is a lightweight interpreted programming language with first-class functions.
-
web
Some thing interesting about web. New door for the world.
-
server
A server is a program made to process requests and deliver data to clients.
-
Machine learning
Machine learning is a way of modeling and interpreting data that allows a piece of software to respond intelligently.
-
Visualization
Some thing interesting about visualization, use data art
-
Game
Some thing interesting about game, make everyone happy.
Recommend Org
-
Facebook
We are working to build community through open source technology. NB: members must have two-factor auth.
-
Microsoft
Open source projects and samples from Microsoft.
-
Google
Google ❤️ Open Source for everyone.
-
Alibaba
Alibaba Open Source for everyone
-
D3
Data-Driven Documents codes.
-
Tencent
China tencent open source team.
from livespeechportraits.