Add model card for VGGT
Browse filesThis PR adds a model card to the repository. The model card includes the relevant pipeline tag, license, and links to the paper and code. It also includes a brief overview of the model.
README.md
CHANGED
@@ -2,8 +2,14 @@
|
|
2 |
tags:
|
3 |
- model_hub_mixin
|
4 |
- pytorch_model_hub_mixin
|
|
|
|
|
|
|
5 |
---
|
6 |
|
7 |
-
|
8 |
-
|
9 |
-
|
|
|
|
|
|
|
|
2 |
tags:
|
3 |
- model_hub_mixin
|
4 |
- pytorch_model_hub_mixin
|
5 |
+
library_name: pytorch
|
6 |
+
pipeline_tag: image-to-3d
|
7 |
+
license: mit
|
8 |
---
|
9 |
|
10 |
+
Visual Geometry Grounded Transformer (VGGT, CVPR 2025) is a feed-forward neural network that directly infers all key 3D attributes of a scene, including extrinsic and intrinsic camera parameters, point maps, depth maps, and 3D point tracks, **from one, a few, or hundreds of its views, within seconds**.
|
11 |
+
|
12 |
+
Paper: [VGGT: Visual Geometry Grounded Transformer](https://huggingface.co/papers/2503.11651)
|
13 |
+
Code: https://github.com/facebookresearch/vggt
|
14 |
+
Project Page: https://vgg-t.github.io/
|
15 |
+
Demo: https://huggingface.co/spaces/facebook/vggt
|