Given input speech signal, GaussianSpeech can synthesize photorealistic 3D-consistent talking human head avatars. Realistic and high-quality animations can be generated, including mouth interiors such as teeth, wrinkles, and specularities in the eyes.
We learn to generate large displacements for parametric head models, such as long hair, with high level of detail. The displacements can be added to an arbitrary head for animation and semantic editing.
EG3D is a powerful {z, camera}->image generative model, but inverting EG3D (finding a corresponding z for a given image) is not always trivial. We propose a fully-convolutional encoder for EG3D based on the observation that predicting both z code and tri-planes is beneficial. TriPlaneNet also works for videos and in real time (check out the Live Demo).
Can the power of generative models provide us with the face recognition on steroids? Random collections of faces + StyleGAN are the secret sauce. We release the collections themselves, as well as a new fairness-concerned testing benchmark.
By taking a simple selfie-like capture by a smartphone, one can easily create a relightable 3D head portrait. The system is based on Neural Point-Based Graphics.
Given RGB(D) images and a point cloud reconstruction of a scene, our neural network generates extreme novel views of the scene which look highly photoreal.
An advanced version of "Optic disc and cup segmentation methods..." (see below), where a segmentation is performed by a U-Net stacked multiple times, and a validation is performed on large amounts of data provided by UCSF.
Automatic segmentation of two organs on an eye fundus image allows medical doctors to make more accurate early diagnosis of glaucoma and evaluate its progression over time.
The webpage template was borrowed from the exciting page of Jon Barron.