In this episode of the Talking Papers Podcast, I hosted Hsueh-Ti Derek Liu to chat about his paper "Learning Smooth Neural Functions via Lipschitz Regularizationโ, published in SIGGRAPH 2022.ย
In this paper, they took on the unique task of enforcing smoothness on Neural Fields (modelled as a neural network). They do this by introducing a regularization term that forces the Lipschitz constant of the network to be very small. They show the performance of their method on shape interpolation, extrapolation and partial shape reconstruction from 3D point clouds. I mostly like the fact that it is implemented in only 4 lines of code.ย
Derek will soon complete his PhD at the University of Toronto and will start a research scientist position at Roblox Research. This work was done when he was interning at NVIDIA. During our chat, I had the pleasure to discover that Derek is one of the few humans on the plant that has the ability to take a complicated idea and explain it in a simple and easy-to-follow way. His strong background in geometry processing makes this paper, which is well within the learning domain, very unique in the current paper landscape. It was a pleasure recording this episode with him.ย
AUTHORS
Hsueh-Ti Derek Liu, Francis Williams, Alec Jacobson, Sanja Fidler, Or Litany
ABSTRACT
Neural implicit fields have recently emerged as a useful representation for 3D shapes. These fields are commonly represented as neural networks which map latent descriptors and 3D coordinates to implicit function values. The latent descriptor of a neural field acts as a deformation handle for the 3D shape it represents. Thus, smoothness with respect to this descriptor is paramount for performing shape-editing operations. In this work, we introduce a novel regularization designed to encourage smooth latent spaces in neural fields by penalizing the upper bound on the field's Lipschitz constant. Compared with prior Lipschitz regularized networks, ours is computationally fast, can be implemented in four lines of code, and requires minimal hyperparameter tuning for geometric applications. We demonstrate the effectiveness of our approach on shape interpolation and extrapolation as well as partial shape reconstruction from 3D point clouds, showing both qualitative and quantitative improvements over existing state-of-the-art and non-regularized baselines.
RELATED PAPERS
๐DeepSDF
๐Neural Fields (collection of works)
๐Sorting Out Lipschitz Function Approximation
LINKS AND RESOURCES
๐ปProject website
๐ Paper
๐ปCode
To stay up to date with Derek's latest research, follow him on:
๐จ๐ปโ๐Personal Page
๐จ๐ปโ๐Google Scholar
Recorded on May 30th 2022.
CONTACT
If you would like to be a guest, sponsor or just share your thoughts, feel free to reach out via email
๐งSubscribe on your favourite podcast app: https://talking.papers.podcast.itzikbs.com
๐งSubscribe to our mailing list: http://eepurl.com/hRznqb
๐ฆFollow us on Twitter: https://twitter.com/talking_papers
๐ฅYouTube Channel: https://bit
Dateline NBC
Current and classic episodes, featuring compelling true-crime mysteries, powerful documentaries and in-depth investigations. Follow now to get the latest episodes of Dateline NBC completely free, or subscribe to Dateline Premium for ad-free listening and exclusive bonus content: DatelinePremium.com
24/7 News: The Latest
The latest news in 4 minutes updated every hour, every day.
Therapy Gecko
An unlicensed lizard psychologist travels the universe talking to strangers about absolutely nothing. TO CALL THE GECKO: follow me on https://www.twitch.tv/lyleforever to get a notification for when I am taking calls. I am usually live Mondays, Wednesdays, and Fridays but lately a lot of other times too. I am a gecko.