Giter VIP home page Giter VIP logo

smirk's People

Contributors

filby89 avatar georgeretsi avatar

Stargazers

 avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar

Watchers

 avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar

smirk's Issues

error occurred when I generated demo.py

Hello!
Thank you for Amazing work!
I know it's a stupid question,
An error occurred when I generated demo.py : Missing file pretrained_models/SMIRK_em1.pt

Error while pre-processing and missing assets files

Hi,
Thanks for sharing this amazing repo 😄
I was trying to retrain the model with a couple of modifications.
In the processes I faced a couple of hitches while trying to run the pre-processing code on the mentioned datasets
Would help me a lot if you can guide me through them

Errors:

  1. In apply_mediapipe_to_dataset.py, inside the def preprocess_sample() function,
    you used model_asset_path='assets/face_landmarker.task' but the file is not provided within the assets folder.
    Could you share a link from where I can download this / update the repo with this file ?
  2. In apply_fan_to_dataset.py I noticed you os.walk() the root, store the paths and then in L36 we loop on those pairs
    In the process we were also storing .mp4 and .avi but trying to read them with cv2.imread() in L44.
    This would in the end throw None type has no shape attribuite error. How do I handle this case / video datasets ?

Generic Doubts:

  1. Since LRS3 is not available anymore, I was trying to make it work for LRS2. There were 2 subsets in it - pretrain and main. Should I combine them together and then preprocess it ?
  2. I have not used MEAD dataset before but from the config files, I noticed that the code uses MEAD_front and MEAD_sides
    But the configs do not have provision landmark paths for MEAD_sides. Are they already provided with the dataset ?
  3. I noticed that the output flame params from the model just has 50 expression params. I wanted retrain the model so that it outputs 100 expr params instead. For this, is it sufficient to just modify in the config and retrain or should I replace the flame implementation as a whole ?

Some questions about SMIRK

Hello George!
     Thank you for your work and open-source contributions; it's indeed very enlightening.  I have a few simple questions about your work:
      1) Besides the common encoding of 3DMM params such as shape, expression, global pose+jaw pose, and caw, an additional 'eyelip' parameter is used in smirk. I'm not sure if this corresponds to the eye_pose parameter in the generic FLAME model. Has the FLAME model used in your open-source code been modified as shown in Figures 1 and 2, or does this parameter serve an additional purpose? Can it be used for decoding with the generic FLAME model?
      2) For the same input image (as shown in Figure 3), SMIRK's reconstruction shows a better eye closure effect compared to DECA and Emoca, as shown in Figures 4 and 5. Since SMIRK does not provide code for generating meshes, I used the decoded FLAME vertices from SMIRK combined with the generic FLAME faces to generate an OBJ file. Although SMIRK's reconstruction is closer to the input, a noticeable issue is the overlapping of eyelids and eyeballs, which is not an isolated case (as seen in the zoom in of image). What could be causing this problem, and is there a solution?
image

                                                                         fig1: the 3DMM params in smirk

image

                                                                         fig2: the 3DMM params in deca and emoca

image
fig3: the input image

image
image

                            fig4: the smirk result

image
image

                        fig5: the emoca result

Adding a Tracker to SMIRK

Since the pre-training is based on MICA, Is there a way to use Metrical_Tracker that comes with MICA, with SMIRK?

If there is no direct way and need a bit of development effort, I am happy to do it and send a PR. Just need directions on how to go about it.

3D Face Tracking General Question

Hi, I'm looking forward to giving this a test drive. This is a general question about 3DMM's and face tracking. Having read this board, It sounds like smirk won't yet produce smooth results on video input. Outside of EMOCAv2 and MICA, are you aware of any repo's that have pushed that work further?

I've seen FlawlessAI's new paper improving results on 2D and 3D landmarks, but that's not publicly available code.

about smirk_generator

Hi!I'm very interested in smirk_generator.

it could re-generate the entire face very well.I noticed that the covered part is not completely black. It randomly samples the relevant information of the original image. I want to know what the training process of this model is.

image

Can I obtain the mesh files?

The current results exist as monochromatic rendered mesh images. How can I directly obtain the mesh source files, for instance, stored in a common 3D file format like OBJ or others? Additionally, the paper mentions using FLAME as the facial prior, but the results do not appear to have FLAME topology. Is additional post-processing required?

Customizing mask.

Hello, would you like to let me know how to create mask to reconstruct only mouth area not full face?
In addition to this we can reconstruct lips part with different jaw pose?
Looking forward to hearing from you!
Thanks

Different jaw-pose FLAME parameter

Hello, everybody.
I have got jaw_pose parameter using smirk.
([[[ 0.0809, -0.0012, -0.0506]]], device='cuda:0')

But jaw_pose in metrical_tacker has different format.
([[1., 0., 0., 0., 1., 0.]], device='cuda:0', requires_grad=True)

Would you like to let me know which format is correct and how to convert with each other?
Thanks

Can a fixed value be used for the camera CAM parameter here?

Here are the obtained pose and cam. Does the cam here refer to the camera parameters? Is it possible to select a camera parameter with a front-facing orientation?
image
Can the following be made to face forward?
image
hope you can answer my question, thank you.

Input as a video?

Hello!

Great work!

Is there a way to input videos/track SMIRK over frames?

Recommend Projects

  • React photo React

    A declarative, efficient, and flexible JavaScript library for building user interfaces.

  • Vue.js photo Vue.js

    🖖 Vue.js is a progressive, incrementally-adoptable JavaScript framework for building UI on the web.

  • Typescript photo Typescript

    TypeScript is a superset of JavaScript that compiles to clean JavaScript output.

  • TensorFlow photo TensorFlow

    An Open Source Machine Learning Framework for Everyone

  • Django photo Django

    The Web framework for perfectionists with deadlines.

  • D3 photo D3

    Bring data to life with SVG, Canvas and HTML. 📊📈🎉

Recommend Topics

  • javascript

    JavaScript (JS) is a lightweight interpreted programming language with first-class functions.

  • web

    Some thing interesting about web. New door for the world.

  • server

    A server is a program made to process requests and deliver data to clients.

  • Machine learning

    Machine learning is a way of modeling and interpreting data that allows a piece of software to respond intelligently.

  • Game

    Some thing interesting about game, make everyone happy.

Recommend Org

  • Facebook photo Facebook

    We are working to build community through open source technology. NB: members must have two-factor auth.

  • Microsoft photo Microsoft

    Open source projects and samples from Microsoft.

  • Google photo Google

    Google ❤️ Open Source for everyone.

  • D3 photo D3

    Data-Driven Documents codes.