If you are a Applied Scientist (Audio-3D Vision) with experience, please read on!
Top Reasons to Work with Us
We are an exciting and rapidly growing startup operating at the intersection of AI and film making.
We have an exciting technology that uses generative AI to make our users look like their favorite actor or actress is speaking another language, without subtitles or voice-overs, making it indistinguishable from the original performance using automated visual translation.
We have been awarded Awarded TIME Best Inventions of the Year and also won Best Neural Network Award at AI Tech Awards!
What You Will Be Doing
As an applied scientist on the audio and video synthesis team at Flawless, you will work with a close-knit, passionate group of world-class researchers tackling some of the most challenging problems in deep learning, including Audio-driven 3D facial animation, GAN models for visual speech synthesis, speech-based modeling, multi-modal fusion for audio-visual learning, and a lot more.
Putting the scientists complex solutions in to production as it pertains to matching the sounds of actors voices in different languages and expanding their vocabulary to sound the same - and matching that sounds to the movement of the lips and facial shifts
Solve challenging problems in deep learning, including putting RS Solutions in production focused on Audio-driven 3D facial animation, GAN models for visual speech synthesis, speech-based modeling, multi-modal fusion for audio-visual learning, and a lot more
Emphasis in Audio-driven 3D facial animation
Generative models for visual speech synthesis
Speech-based modeling with deep neural networks
Recurrent and visual attention models for audio-video processing
What You Need for this Position
Minimum Qualifications
-MS in Computer Vision, Computer Graphics, Speech Recognition and Synthesis, Machine Learning, or related field
-2+ years of applied research experience in:
-Audio-driven 3D facial animation
-Generative models for visual speech synthesis
-Speech-based modeling with deep neural networks
-Recurrent and visual attention models for audio-video processing
-3+ years of experience in Python with proficiency in deep learning frameworks such as PyTorch or Tensorflow
Preferred Qualifications
-MS (with experience) or Ph.D. in Computer Vision, Computer Graphics, Speech Synthesis, Machine Learning, or related field
-Experience developing tools or solutions at scale for multi-modal or audio-driven visual systems in AR/VR or VFX
-Experience with Python, C++ or CUDA
-Experience with cloud platforms such as GCP, AWS, or Azure
What's In It for You
-Autonomy - You'll own your work from start to finish
-Influence - You'll impact major research decisions
-Publication - You'll be encouraged to publish work through collaborations with researchers
-Learning - Youll push the state-of-the-art with the best in the world
-Impact - Your input genuinely matters
-Workspace flexibility
-Stock Options
-Comprehensive medical, dental, and vision insurance
-401(k) plan
So, if you are a Applied Scientist (Audio-3D Vision) with experience, please apply today or apply directly to liana.pryor@cybercoders.com!
Applicants must be authorized to work in the U.S.