Jonathan Lorraine

lorraine2

AI & ML interests

machine learning, computer vision, generative AI

Recent Activity

View all activity

Organizations

Social Post Explorers's profile picture

lorraine2's activity

posted an update 9 days ago
view post
Post
505
๐Ÿ”Š New NVIDIA paper: Audio-SDS ๐Ÿ”Š

We adapt Score Distillation Sampling (SDS), originally developed for text-to-3D generation, to audio diffusion models, allowing us to reuse large pretrained models for new text-guided parametric audio tasks such as source separation, physically informed impact synthesis, and more.

๐Ÿ”Ž Project Page: https://research.nvidia.com/labs/toronto-ai/Audio-SDS/
๐Ÿ“– Full Paper: https://arxiv.org/abs/2505.04621

Check out more from NVIDIAโ€™s Spatial Intelligence Lab here: https://research.nvidia.com/labs/toronto-ai/

This project was led by the great work of Jessie Richter-Powell, along with Antonio Torralba.

Notably, we find a new and exciting use case for Stable Audio Open ๐Ÿš€
posted an update 5 months ago
view post
Post
2013
๐Ÿฆ™New NVIDIA paper: LLaMA-Mesh ๐Ÿฆ™

We enable large language models to generate and understand 3D meshes by representing them as text and fine-tuning. This unifies the 3D and text modalities in a single model and preserves language abilities, unlocking conversational 3D creation with mesh understanding.

๐Ÿ”Ž Project Page: https://research.nvidia.com/labs/toronto-ai/LLaMA-Mesh/
๐Ÿ•น๏ธ Interactive Demo: Zhengyi/LLaMA-Mesh (courtesy of HuggingFace and Gradio)
๐Ÿ“– Full Paper: https://arxiv.org/abs/2411.09595
๐Ÿ‘จโ€๐Ÿ’ปCode: https://github.com/nv-tlabs/LLaMa-Mesh
๐Ÿ’พ Model Checkpoint: Zhengyi/LLaMA-Mesh
๐Ÿงฉ Blender Addon: https://github.com/huggingface/meshgen (courtesy of Dylan Ebert)
๐ŸŽฅ 5-min Overview Video: https://youtu.be/eZNazN-1lPo?si=-idQa5aaceVw0Bbj (courtesy of AI Papers Academy)
reacted to their post with ๐Ÿ‘€ 5 months ago
view post
Post
1218
New NVIDIA paper: โšก Multi-student Diffusion Distillation for Better One-step Generators โšก

Do you want to make your diffusion models (a) run in a single step, (b) run with a smaller model, and (c) have improved quality simultaneously? Check out our multi-student distillation (MSD) method, which is simple and applicable to most diffusion models! The only catch is now we have to distill (and store) a mixture-of-expert student generators.

Explore the MSD project page to learn more: https://research.nvidia.com/labs/toronto-ai/MSD/

Work led by Yanke Song along with Weili Nie, Karsten Kreis and James Lucas

Check out more work from the Toronto AI Lab here: https://research.nvidia.com/labs/toronto-ai/
  • 1 reply
ยท
posted an update 5 months ago
view post
Post
1218
New NVIDIA paper: โšก Multi-student Diffusion Distillation for Better One-step Generators โšก

Do you want to make your diffusion models (a) run in a single step, (b) run with a smaller model, and (c) have improved quality simultaneously? Check out our multi-student distillation (MSD) method, which is simple and applicable to most diffusion models! The only catch is now we have to distill (and store) a mixture-of-expert student generators.

Explore the MSD project page to learn more: https://research.nvidia.com/labs/toronto-ai/MSD/

Work led by Yanke Song along with Weili Nie, Karsten Kreis and James Lucas

Check out more work from the Toronto AI Lab here: https://research.nvidia.com/labs/toronto-ai/
  • 1 reply
ยท
reacted to their post with ๐Ÿ‘€ 6 months ago
view post
Post
294
New NeurIPS paper: โ€œTraining Data Attribution via Approximate Unrollingโ€

Ever wondered how individual data points influence AI decisions? ๐Ÿค” We explore how specific training data pieces affect machine learning models' behavior, which can be crucial for making AI systems more transparent, trustworthy, and fair.

Our method, SOURCE, bridges the gap between implicit differentiation and unrolling approaches, combining computational efficiency with flexibility making it suitable for non-converged models and multi-stage training pipelines.

๐Ÿ“„ Full paper: https://openreview.net/pdf?id=3NaqGg92KZ

Juhan Bae led along with Wu Lin and Roger Grosse.

Supported by the University of Toronto, Vector Institute, NVIDIA, and Anthropic
posted an update 6 months ago
view post
Post
294
New NeurIPS paper: โ€œTraining Data Attribution via Approximate Unrollingโ€

Ever wondered how individual data points influence AI decisions? ๐Ÿค” We explore how specific training data pieces affect machine learning models' behavior, which can be crucial for making AI systems more transparent, trustworthy, and fair.

Our method, SOURCE, bridges the gap between implicit differentiation and unrolling approaches, combining computational efficiency with flexibility making it suitable for non-converged models and multi-stage training pipelines.

๐Ÿ“„ Full paper: https://openreview.net/pdf?id=3NaqGg92KZ

Juhan Bae led along with Wu Lin and Roger Grosse.

Supported by the University of Toronto, Vector Institute, NVIDIA, and Anthropic
posted an update 9 months ago
view post
Post
565
๐Ÿšจ Code now available for "Using Large Language Models for Hyperparameter Optimization" at https://github.com/michaelrzhang/LLM-HyperOpt ๐Ÿšจ

TLDR: You can just ask LLMs which hyperparameters to use, and it works pretty well! You can even directly optimize your modelโ€™s code as a hyperparameter with this.

Check out the paper at https://arxiv.org/abs/2312.04528 - with Michael Zhang, Nishkrit Desai, Juhan Bae, and Jimmy Ba
reacted to their post with ๐Ÿ‘๐Ÿ”ฅ 9 months ago
view post
Post
2714
โšก My PhD thesis, โ€œScalable Nested Optimization for Deep Learning,โ€ is now on arXiv! โšก

tl;dr: We develop various optimization tools with highlights, including:
ยท Making the momentum coefficient complex for adversarial games like GANs.
ยท Optimizing millions of hyperparameters using implicit differentiation.
ยท Tuning hyperparameters using hypernetworks.
ยท Differentiably finding bifurcations in optimization for diverse solutions.

https://arxiv.org/abs/2407.01526
  • 3 replies
ยท
reacted to their post with ๐Ÿ‘ 9 months ago
view post
Post
1709
New NVIDIA GTC24 paper ๐ŸŽŠ

We generate high-quality 3D assets in only 400ms from text by combining (a) amortized optimization for speed, (b) surface rendering for quality, and (c) 3D data for robustness.

โ˜• LATTE3D project details: https://research.nvidia.com/labs/toronto-ai/LATTE3D/
  • 2 replies
ยท
reacted to their post with ๐Ÿš€๐Ÿค—๐Ÿ”ฅ 9 months ago
view post
Post
2396
New #NVIDIA paper: Improving Hyperparameter Optimization with Checkpointed Model Weights

Hyperparameter optimization often dominates the cost of model design. So, we want cheap surrogate functions that approximate model performance to guide our search. Existing methods can train on optimization metadata โ€“ like a trajectory of losses โ€“ to build these surrogates.

In our work, we add the ability to train our hyperparameter optimization surrogates on checkpointed model weights with a graph metanetwork. This allows us to leverage a large, pre-existing source of information that can featurize the architecture, dataset, losses, and optimization procedure.

๐Ÿ”Project page: https://research.nvidia.com/labs/toronto-ai/FMS/
๐Ÿ‘จโ€๐Ÿ’ป Code for reproduction: https://github.com/NVlabs/forecasting-model-search
๐Ÿ“„ Full Paper: https://arxiv.org/abs/2406.18630

Our project was a collaboration between NVIDIAโ€™s Toronto AI Lab and the TAO team.

Check out more work from Toronto AI Lab here: https://research.nvidia.com/labs/toronto-ai/

You can view the TAO toolkit here: https://developer.nvidia.com/tao-toolkit
reacted to their post with โค๏ธ๐Ÿš€ 9 months ago
view post
Post
1709
New NVIDIA GTC24 paper ๐ŸŽŠ

We generate high-quality 3D assets in only 400ms from text by combining (a) amortized optimization for speed, (b) surface rendering for quality, and (c) 3D data for robustness.

โ˜• LATTE3D project details: https://research.nvidia.com/labs/toronto-ai/LATTE3D/
  • 2 replies
ยท
posted an update 9 months ago
view post
Post
2714
โšก My PhD thesis, โ€œScalable Nested Optimization for Deep Learning,โ€ is now on arXiv! โšก

tl;dr: We develop various optimization tools with highlights, including:
ยท Making the momentum coefficient complex for adversarial games like GANs.
ยท Optimizing millions of hyperparameters using implicit differentiation.
ยท Tuning hyperparameters using hypernetworks.
ยท Differentiably finding bifurcations in optimization for diverse solutions.

https://arxiv.org/abs/2407.01526
  • 3 replies
ยท
posted an update 11 months ago
view post
Post
2396
New #NVIDIA paper: Improving Hyperparameter Optimization with Checkpointed Model Weights

Hyperparameter optimization often dominates the cost of model design. So, we want cheap surrogate functions that approximate model performance to guide our search. Existing methods can train on optimization metadata โ€“ like a trajectory of losses โ€“ to build these surrogates.

In our work, we add the ability to train our hyperparameter optimization surrogates on checkpointed model weights with a graph metanetwork. This allows us to leverage a large, pre-existing source of information that can featurize the architecture, dataset, losses, and optimization procedure.

๐Ÿ”Project page: https://research.nvidia.com/labs/toronto-ai/FMS/
๐Ÿ‘จโ€๐Ÿ’ป Code for reproduction: https://github.com/NVlabs/forecasting-model-search
๐Ÿ“„ Full Paper: https://arxiv.org/abs/2406.18630

Our project was a collaboration between NVIDIAโ€™s Toronto AI Lab and the TAO team.

Check out more work from Toronto AI Lab here: https://research.nvidia.com/labs/toronto-ai/

You can view the TAO toolkit here: https://developer.nvidia.com/tao-toolkit
replied to their post about 1 year ago
view reply

We include a narrated 30s summary video here and, additionally, on our project webpage, a video demonstrating our model's usage and a 3-minute video overview explaining our method.

posted an update about 1 year ago
view post
Post
1709
New NVIDIA GTC24 paper ๐ŸŽŠ

We generate high-quality 3D assets in only 400ms from text by combining (a) amortized optimization for speed, (b) surface rendering for quality, and (c) 3D data for robustness.

โ˜• LATTE3D project details: https://research.nvidia.com/labs/toronto-ai/LATTE3D/
  • 2 replies
ยท