echoinside – Telegram
echoinside
107 subscribers
834 photos
65 videos
41 files
933 links
ML in computer graphics and random stuff.
Any feedback: @fogside
Download Telegram
This media is not supported in your browser
VIEW IN TELEGRAM
MeshTalk: 3D Face Animation from Speech using Cross-Modality Disentanglement
[Facebook Reality Labs, Carnegie Mellon University]
* pdf
This paper presents a generic method for generating full facial 3D animation from speech. Existing approaches to audio-driven facial animation exhibit uncanny or static upper face animation, fail to produce accurate and plausible co-articulation or rely on person-specific models that limit their scalability. To improve upon existing models, we propose a generic audio-driven facial animation approach that achieves highly realistic motion synthesis results for the entire face. At the core of our approach is a categorical latent space for facial animation that disentangles audio-correlated and audio-uncorrelated information based on a novel cross-modality loss. Our approach ensures highly accurate lip motion, while also synthesizing plausible animation of the parts of the face that are uncorrelated to the audio signal, such as eye blinks and eye brow motion.
#speech2animation
StylePeople: A Generative Model of Fullbody Human Avatars
[Samsung AI Center, SkolTech]
* pdf, abs
We propose a new type of full-body human avatars, which combines parametric mesh-based body model with a neural texture. We show that with the help of neural textures, such avatars can successfully model clothing and hair, which usually poses a problem for mesh-based approaches. We also show how these avatars can be created from multiple frames of a video using backpropagation. We then propose a generative model for such avatars that can be trained from datasets of images and videos of people. The generative model allows us to sample random avatars as well as to create dressed avatars of people from one or few images.#avatars #neural_rendering #gan #smpl #3d
This media is not supported in your browser
VIEW IN TELEGRAM
Geometry-Free View Synthesis: Transformers and no 3D Priors
* pdf, abs
* code
Is a geometric model required to synthesize novel views from a single image? Being bound to local convolutions, CNNs need explicit 3D biases to model geometric transformations. In contrast, we demonstrate that a transformer-based model can synthesize entirely novel views without any hand-engineered 3D biases. This is achieved by (i) a global attention mechanism for implicitly learning long-range 3D correspondences between source and target views, and (ii) a probabilistic formulation necessary to capture the ambiguity inherent in predicting novel views from a single image, thereby overcoming the limitations of previous approaches that are restricted to relatively small viewpoint changes. We evaluate various ways to integrate 3D priors into a transformer architecture. However, our experiments show that no such geometric priors are required and that the transformer is capable of implicitly learning 3D relationships between images.
Regularizing Generative Adversarial Networks under Limited Data
[Google research, UC Merced, Waymo, Yonsei University]
* pdf, abs
* code
The success of the GAN models hinges on a large amount of training data. This work proposes a regularization approach for training robust GAN models on limited data. We theoretically show a connection between the regularized loss and an f-divergence called LeCam-divergence, which we find is more robust under limited training data. Extensive experiments on several benchmark datasets demonstrate that the proposed regularization scheme 1) improves the generalization performance and stabilizes the learning dynamics of GAN models under limited training data, and 2) complements the recent data augmentation methods. These properties facilitate training GAN models to achieve state-of-the-art performance when only limited training data of the ImageNet benchmark is available.
- related work
#gan #limited_data
Recently I discovered Toronto Geometry Colloquium channel.
They also have twitter and website.
"It is a weekly hour-long webseries showcasing geometry processing research, including a 10-min opener speaker and a 50-min headliner in the style of live comedy!"
I have to say, this is an amazing series. I will share some featured talks I found there in the following posts.
- episode from the poster
COALESCE: Component Assembly by Learning to Synthesize Connections
[Simon Fraser University, Adobe research]
* pdf, abs
* youtube talk
Kitbashing with deep learning.
We introduce COALESCE, the first data-driven framework for component-based shape assembly which employs deep learning to synthesize part connections. To handle geometric and topological mismatches between parts, we remove the mismatched portions via erosion, and rely on a joint synthesis step, which is learned from data, to fill the gap and arrive at a natural part joint. Given a set of input parts extracted from different objects, COALESCE automatically aligns them and synthesizes plausible joints to connect the parts into a coherent 3D object represented by a mesh. The joint synthesis network, designed to focus on joint regions, reconstructs the surface between the parts by predicting an implicit shape representation that agrees with existing parts, while generating a smooth and topologically meaningful connection.
#3d #implicit_geometry
Cycles-X rendering engine is available in an experimental branch and it works significantly faster than cycles on CPU and GPU.
There is much be done. We expect it will take at least 6 months until this work is part of an official Blender release.
- Official blog post
#blender
This media is not supported in your browser
VIEW IN TELEGRAM
New Fb Oculus avatars
Appearing first in three games for Quest.
- source
Oculus is beginning to roll out redesigned avatars that are more expressive and customizable than those that launched in 2016.
By the end of 2021, Oculus will have opened its new avatar SDK to all developers, and these VR personas will be supported in Facebook Horizon, the company’s own expansive social VR playground. Though, games are just one application for these refreshed avatars. Oculus says the avatar you create will eventually appear in some form within the Facebook app, Messenger, Instagram, and more, but only if you choose to.
#avatars #VR
This media is not supported in your browser
VIEW IN TELEGRAM
Softwrap - Dynamics For Retopology.
- available on blendermarket
Softwrap works by running a custom softbody simulation while snapping in a way similar to the shrinkwrap modifier.
#simulation #physics #tools #blender
Forwarded from Denis Sexy IT 🤖
This media is not supported in your browser
VIEW IN TELEGRAM
Логичное продолжение нейронки которая стала популярной из-за того как клево она оживляет фотографии с лицами: двигающиеся ЧБ-фотографии, портреты, мемы, все это результат работы алгоритма который называется First Order Model.

Несмотря на то, что алгоритм хорошо работает, оживлять им что-то кроме лиц довольно сложно, хоть он это и поддерживает — «гличи» и помарки создают довольно неприятный эффект.

И вот, спасибо группе ученых, скоро мемы можно будет оживлять в полный рост — новый алгоритм уже может понимать какие именно части тела как бы двигались на фотографии исходя из исходного видео – сделал нарезку с видео, там все понятно (живая фигурка особенно криповая получилась).

Страница проекта:
https://snap-research.github.io/articulated-animation/
(код проекта выложат попозже)
Sketch-based Normal Map Generation with Geometric Sampling
* pdf, abs
Normal map is an important and efficient way to represent complex 3D models. A designer may benefit from the auto-generation of high quality and accurate normal maps from freehand sketches in 3D content creation. This paper proposes a deep generative model for generating normal maps from users’ sketch with geometric sampling. Our generative model is based on Conditional Generative Adversarial Network with the curvature-sensitive points sampling of conditional masks. This sampling process can help eliminate the ambiguity of generation results as network input. In addition, we adopted a U-Net structure discriminator to help the generator be better trained. It is verified that the proposed framework can generate more accurate normal maps.
#gan #sketch
NVIDIA Omniverse Audio2Face is now available in open beta. Unfortunately, it works only on Windows rn. And it requires some RTX gpu. For some reason I think that this kind of product would be much more consumer friendly as a web app like Mixamo or MetaHuman creator.
- download
- tutorial
Audio2Face simplifies animation of a 3D character to match any voice-over track, whether you’re animating characters for a game, film, real-time digital assistants, or just for fun. You can use the app for interactive real-time applications or as a traditional facial animation authoring tool. Run the results live or bake them out, it’s up to you.
#speech2animation