This media is not supported in your browser
VIEW IN TELEGRAM
โบFACET: Fairness in Computer Visionโบ
๐#META AI opens a large, publicly available dataset for classification, detection & segmentation. Potential performance disparities & challenges across sensitive demographic attributes
๐Review https://t.ly/mKn-t
๐Paper arxiv.org/pdf/2309.00035.pdf
๐Dataset https://facet.metademolab.com/
๐#META AI opens a large, publicly available dataset for classification, detection & segmentation. Potential performance disparities & challenges across sensitive demographic attributes
๐Review https://t.ly/mKn-t
๐Paper arxiv.org/pdf/2309.00035.pdf
๐Dataset https://facet.metademolab.com/
๐ฅ10โค6๐4๐1
This media is not supported in your browser
VIEW IN TELEGRAM
๐ฅ๐ฅ #META's DINOv2 is now commercial! ๐ฅ๐ฅ
๐Universal features for image classification, instance retrieval, video understanding, depth & semantic segmentation. Now suitable for commercial.
๐Review https://t.ly/LNrGy
๐Paper arxiv.org/pdf/2304.07193.pdf
๐Code github.com/facebookresearch/dinov2
๐Demo dinov2.metademolab.com/
๐Universal features for image classification, instance retrieval, video understanding, depth & semantic segmentation. Now suitable for commercial.
๐Review https://t.ly/LNrGy
๐Paper arxiv.org/pdf/2304.07193.pdf
๐Code github.com/facebookresearch/dinov2
๐Demo dinov2.metademolab.com/
๐ฅ15๐3โค1๐คฏ1๐ฑ1
This media is not supported in your browser
VIEW IN TELEGRAM
โ๏ธ Relighted 3D Hands ๐ค
๐#META unveils Re:InterHand: a large dataset of relighted 3D interacting hands
๐Review https://t.ly/I1dQk
๐Paper arxiv.org/pdf/2310.17768.pdf
๐Project mks0601.github.io/ReInterHand
๐Data github.com/mks0601/ReInterHand
๐#META unveils Re:InterHand: a large dataset of relighted 3D interacting hands
๐Review https://t.ly/I1dQk
๐Paper arxiv.org/pdf/2310.17768.pdf
๐Project mks0601.github.io/ReInterHand
๐Data github.com/mks0601/ReInterHand
๐คฏ8โค1๐ฑ1
This media is not supported in your browser
VIEW IN TELEGRAM
๐ Emu: image edit / video gen. ๐
๐#Meta the new SOTA in text-to-video generation and instruction-based image editing
๐ Review https://t.ly/PMTBc
๐ Paper (images): https://lnkd.in/eVadH-QS
๐ Project https://lnkd.in/eG8eWUJY
๐ Paper (video): https://lnkd.in/eVadH-QS
๐ Project https://lnkd.in/eu6Zu6gp
๐#Meta the new SOTA in text-to-video generation and instruction-based image editing
๐ Review https://t.ly/PMTBc
๐ Paper (images): https://lnkd.in/eVadH-QS
๐ Project https://lnkd.in/eG8eWUJY
๐ Paper (video): https://lnkd.in/eVadH-QS
๐ Project https://lnkd.in/eu6Zu6gp
๐ฅ8๐คฏ2๐1๐ฑ1
This media is not supported in your browser
VIEW IN TELEGRAM
๐ชRT Humanoid from Head-Mounted Sensors๐ช
๐#META (+CMU) announced SimXR, a method for controlling a simulated avatar from info obtained from AR/VR headsets
๐Review https://t.ly/Si2Mp
๐Paper arxiv.org/pdf/2403.06862.pdf
๐Project www.zhengyiluo.com/SimXR/
๐#META (+CMU) announced SimXR, a method for controlling a simulated avatar from info obtained from AR/VR headsets
๐Review https://t.ly/Si2Mp
๐Paper arxiv.org/pdf/2403.06862.pdf
๐Project www.zhengyiluo.com/SimXR/
โค12โก1๐1
This media is not supported in your browser
VIEW IN TELEGRAM
๐งคHOT3D Hand/Object Tracking๐งค
๐#Meta opens a novel egocentric dataset for 3D hand & object tracking. A new benchmark for vision-based understanding of 3D hand-object interactions. Dataset available ๐
๐Review https://t.ly/cD76F
๐Paper https://lnkd.in/e6_7UNny
๐Data https://lnkd.in/e6P-sQFK
๐#Meta opens a novel egocentric dataset for 3D hand & object tracking. A new benchmark for vision-based understanding of 3D hand-object interactions. Dataset available ๐
๐Review https://t.ly/cD76F
๐Paper https://lnkd.in/e6_7UNny
๐Data https://lnkd.in/e6P-sQFK
๐ฅ9โค3๐3๐2๐คฏ1
This media is not supported in your browser
VIEW IN TELEGRAM
๐ฅ๐ฅ SAM v2 is out! ๐ฅ๐ฅ
๐#Meta announced SAM 2, the novel unified model for real-time promptable segmentation in images and videos. 6x faster, it's the new SOTA by a large margin. Source Code, Dataset, Models & Demo released under permissive licenses๐
๐Review https://t.ly/oovJZ
๐Paper https://t.ly/sCxMY
๐Demo https://sam2.metademolab.com
๐Project ai.meta.com/blog/segment-anything-2/
๐Models github.com/facebookresearch/segment-anything-2
๐#Meta announced SAM 2, the novel unified model for real-time promptable segmentation in images and videos. 6x faster, it's the new SOTA by a large margin. Source Code, Dataset, Models & Demo released under permissive licenses๐
๐Review https://t.ly/oovJZ
๐Paper https://t.ly/sCxMY
๐Demo https://sam2.metademolab.com
๐Project ai.meta.com/blog/segment-anything-2/
๐Models github.com/facebookresearch/segment-anything-2
๐ฅ27โค10๐คฏ4๐2๐พ1
This media is not supported in your browser
VIEW IN TELEGRAM
๐ EFM3D: 3D Ego-Foundation ๐
๐#META presents EFM3D, the first benchmark for 3D object detection and surface regression on HQ annotated egocentric data of Project Aria. Datasets & Code released๐
๐Review https://t.ly/cDJv6
๐Paper arxiv.org/pdf/2406.10224
๐Project www.projectaria.com/datasets/aeo/
๐Repo github.com/facebookresearch/efm3d
๐#META presents EFM3D, the first benchmark for 3D object detection and surface regression on HQ annotated egocentric data of Project Aria. Datasets & Code released๐
๐Review https://t.ly/cDJv6
๐Paper arxiv.org/pdf/2406.10224
๐Project www.projectaria.com/datasets/aeo/
๐Repo github.com/facebookresearch/efm3d
๐ฅ9โค2๐2โก1๐1๐1
This media is not supported in your browser
VIEW IN TELEGRAM
๐ฅ CoTracker3 by #META is out! ๐ฅ
๐#Meta (+VGG Oxford) unveils CoTracker3, a new tracker that outperforms the previous SoTA by a large margin using only the 0.1% of the training data ๐คฏ๐คฏ๐คฏ
๐Review https://t.ly/TcRIv
๐Paper arxiv.org/pdf/2410.11831
๐Project cotracker3.github.io/
๐Code github.com/facebookresearch/co-tracker
๐#Meta (+VGG Oxford) unveils CoTracker3, a new tracker that outperforms the previous SoTA by a large margin using only the 0.1% of the training data ๐คฏ๐คฏ๐คฏ
๐Review https://t.ly/TcRIv
๐Paper arxiv.org/pdf/2410.11831
๐Project cotracker3.github.io/
๐Code github.com/facebookresearch/co-tracker
โค14๐ฅ3๐คฏ3๐พ2๐1๐ฑ1๐1
This media is not supported in your browser
VIEW IN TELEGRAM
โ๏ธ Universal Relightable Avatars โ๏ธ
๐#Meta unveils URAvatar, photorealistic & relightable avatars from phone scan with unknown illumination. Stunning results!
๐Review https://t.ly/U-ESX
๐Paper arxiv.org/pdf/2410.24223
๐Project junxuan-li.github.io/urgca-website
๐#Meta unveils URAvatar, photorealistic & relightable avatars from phone scan with unknown illumination. Stunning results!
๐Review https://t.ly/U-ESX
๐Paper arxiv.org/pdf/2410.24223
๐Project junxuan-li.github.io/urgca-website
โค11๐ฅ5โก1๐1
This media is not supported in your browser
VIEW IN TELEGRAM
โค๏ธโ๐ฅ Uncommon object in #3D โค๏ธโ๐ฅ
๐#META releases uCO3D, a new object-centric dataset for 3D AI. The largest publicly-available collection of HD videos of objects with 3D annotations that ensures full-360โฆ coverage. Code & data under CCA 4.0๐
๐Review https://t.ly/Z_tvA
๐Paper https://arxiv.org/pdf/2501.07574
๐Project https://uco3d.github.io/
๐Repo github.com/facebookresearch/uco3d
๐#META releases uCO3D, a new object-centric dataset for 3D AI. The largest publicly-available collection of HD videos of objects with 3D annotations that ensures full-360โฆ coverage. Code & data under CCA 4.0๐
๐Review https://t.ly/Z_tvA
๐Paper https://arxiv.org/pdf/2501.07574
๐Project https://uco3d.github.io/
๐Repo github.com/facebookresearch/uco3d
โค11โก2๐2๐1๐1๐คฉ1๐พ1
This media is not supported in your browser
VIEW IN TELEGRAM
โ๏ธ Relightable Full-Body Avatars โ๏ธ
๐#Meta unveils the first approach ever to jointly model the relightable appearance of the body, face, and hands of drivable avatars.
๐Review https://t.ly/kx9gf
๐Paper arxiv.org/pdf/2501.14726
๐Project neuralbodies.github.io/RFGCA
๐#Meta unveils the first approach ever to jointly model the relightable appearance of the body, face, and hands of drivable avatars.
๐Review https://t.ly/kx9gf
๐Paper arxiv.org/pdf/2501.14726
๐Project neuralbodies.github.io/RFGCA
โค3๐3๐ฅ3โก1๐คฏ1๐ข1๐คฉ1
This media is not supported in your browser
VIEW IN TELEGRAM
๐ฅ VideoJAM: #META's Video-Model (SOTA) ๐ฅ
๐#META's VideoJAM: the new SOTA (by large margin) in motion coherence for video generation, much better than SORA! A strong motion prior into any video-gen model. Impressive results, no code announced๐ฅฒ
๐Review https://shorturl.at/id7Bt
๐Paper https://arxiv.org/pdf/2502.02492
๐Project https://hila-chefer.github.io/videojam-paper.github.io/
๐#META's VideoJAM: the new SOTA (by large margin) in motion coherence for video generation, much better than SORA! A strong motion prior into any video-gen model. Impressive results, no code announced๐ฅฒ
๐Review https://shorturl.at/id7Bt
๐Paper https://arxiv.org/pdf/2502.02492
๐Project https://hila-chefer.github.io/videojam-paper.github.io/
๐ฅ9โค4๐1๐1๐คฉ1
This media is not supported in your browser
VIEW IN TELEGRAM
๐ค META Human-Robot ๐ค
๐#META PARTNR: novel benchmark for Planning And Reasoning Tasks in humaN-Robot collaboration. The largest benchmark of its kind: 100,000+ natural language tasks, spanning 60 houses and 5,819 unique objects. Code & Data (๐ค) under MIT๐
๐Review https://t.ly/zcN0K
๐Paper arxiv.org/pdf/2411.00081
๐Repo github.com/facebookresearch/partnr-planner
๐คData huggingface.co/datasets/ai-habitat/partnr_episodes
๐#META PARTNR: novel benchmark for Planning And Reasoning Tasks in humaN-Robot collaboration. The largest benchmark of its kind: 100,000+ natural language tasks, spanning 60 houses and 5,819 unique objects. Code & Data (๐ค) under MIT๐
๐Review https://t.ly/zcN0K
๐Paper arxiv.org/pdf/2411.00081
๐Repo github.com/facebookresearch/partnr-planner
๐คData huggingface.co/datasets/ai-habitat/partnr_episodes
๐ฅ9๐คฉ2โค1๐1๐1
This media is not supported in your browser
VIEW IN TELEGRAM
๐ฒ๏ธ VGG Transformer ๐ฒ๏ธ
๐VGGT by VGG & #META (#CVPR2025) is a feed-forward neural net. that directly infers all key 3D attributes of a scene within seconds. Code released๐
๐Review https://t.ly/WoWXL
๐Paper https://arxiv.org/pdf/2503.11651
๐Project https://vgg-t.github.io/
๐Code github.com/facebookresearch/vggthttps://t.ly/WoWXL
๐VGGT by VGG & #META (#CVPR2025) is a feed-forward neural net. that directly infers all key 3D attributes of a scene within seconds. Code released๐
๐Review https://t.ly/WoWXL
๐Paper https://arxiv.org/pdf/2503.11651
๐Project https://vgg-t.github.io/
๐Code github.com/facebookresearch/vggthttps://t.ly/WoWXL
๐คฏ26๐11๐ฅ6โค2๐คฉ1
This media is not supported in your browser
VIEW IN TELEGRAM
๐ฆ DINOv3 is out ๐ฆ
๐#Meta unveils DINOv3! A novel foundation model outperforming the previous SOTAs in computer vision. Code & weights released under DINOv3 License๐
๐Review https://t.ly/-S3ZL
๐Paper https://t.ly/ervOT
๐Project https://lnkd.in/dHFf3esd
๐Repo https://lnkd.in/dPxhDxAq
๐คHF https://lnkd.in/dWGudY2i
๐#Meta unveils DINOv3! A novel foundation model outperforming the previous SOTAs in computer vision. Code & weights released under DINOv3 License๐
๐Review https://t.ly/-S3ZL
๐Paper https://t.ly/ervOT
๐Project https://lnkd.in/dHFf3esd
๐Repo https://lnkd.in/dPxhDxAq
๐คHF https://lnkd.in/dWGudY2i
โค43๐ฅ13๐2๐1๐พ1
This media is not supported in your browser
VIEW IN TELEGRAM
๐ซATLAS: SOTA Human Model๐ซ
๐#META presents ATLAS, a novel high-fidelity body model learned from 600k high-res. scans captured using 240 synchronized cams. Code announced, to be released๐
๐Review https://t.ly/0hHud
๐Paper arxiv.org/pdf/2508.15767
๐Project jindapark.github.io/projects/atlas/
๐Repo TBA
๐#META presents ATLAS, a novel high-fidelity body model learned from 600k high-res. scans captured using 240 synchronized cams. Code announced, to be released๐
๐Review https://t.ly/0hHud
๐Paper arxiv.org/pdf/2508.15767
๐Project jindapark.github.io/projects/atlas/
๐Repo TBA
๐ฅ7โค6๐1๐1
This media is not supported in your browser
VIEW IN TELEGRAM
โค๏ธโ๐ฅPHD: Personalized 3D Humansโค๏ธโ๐ฅ
๐ETH & #Meta unveil PHD, a novel approach for personalized 3D human mesh recovery (HMR) and body fitting that leverages user-specific shape information. Code & models to be released๐
๐Review https://t.ly/IeRhH
๐Paper https://arxiv.org/pdf/2508.21257
๐Project https://phd-pose.github.io/
๐Repo TBA
๐ETH & #Meta unveil PHD, a novel approach for personalized 3D human mesh recovery (HMR) and body fitting that leverages user-specific shape information. Code & models to be released๐
๐Review https://t.ly/IeRhH
๐Paper https://arxiv.org/pdf/2508.21257
๐Project https://phd-pose.github.io/
๐Repo TBA
โค7๐ฅ2๐1