CVPR 2024 論文和代碼速遞! 2024.4.16
共 3710字,需瀏覽 8分鐘
·
2024-04-16 17:05
推薦
微信交流群現(xiàn)已有2000+從業(yè)人員交流群,歡迎進(jìn)群交流學(xué)習(xí),微信:nvshenj125
B站最新成果demo分享地址:https://space.bilibili.com/288489574
頂會(huì)工作整理Github repo:https://github.com/DWCTOD/CVPR2023-Papers-with-Code-Demo
工作整理
CVPR 2024
Updated on : 16 Apr 2024
total number : 22
No More Ambiguity in 360° Room Layout via Bi-Layout Estimation
論文/Paper: http://arxiv.org/pdf/2404.09993
代碼/Code: None
One-Click Upgrade from 2D to 3D: Sandwiched RGB-D Video Compression for Stereoscopic Teleconferencing
論文/Paper: http://arxiv.org/pdf/2404.09979
代碼/Code: None
Video2Game: Real-time, Interactive, Realistic and Browser-Compatible Environment from a Single Video
論文/Paper: http://arxiv.org/pdf/2404.09833
代碼/Code: None
3D Face Tracking from 2D Video through Iterative Dense UV to Image Flow
論文/Paper: http://arxiv.org/pdf/2404.09819
代碼/Code: None
FSRT: Facial Scene Representation Transformer for Face Reenactment from Factorized Appearance, Head-pose, and Facial Expression Features
論文/Paper: http://arxiv.org/pdf/2404.09736
代碼/Code: None
The revenge of BiSeNet: Efficient Multi-Task Image Segmentation
論文/Paper: http://arxiv.org/pdf/2404.09570
代碼/Code: None
Learning Tracking Representations from Single Point Annotations
論文/Paper: http://arxiv.org/pdf/2404.09504
代碼/Code: None
SparseOcc: Rethinking Sparse Latent Representation for Vision-Based Semantic Occupancy Prediction
論文/Paper: http://arxiv.org/pdf/2404.09502
代碼/Code: None
TCCT-Net: Two-Stream Network Architecture for Fast and Efficient Engagement Estimation via Behavioral Feature Signals
論文/Paper: http://arxiv.org/pdf/2404.09474
代碼/Code: https://github.com/vedernikovphoto/tcct_net
PhyScene: Physically Interactable 3D Scene Synthesis for Embodied AI
論文/Paper: http://arxiv.org/pdf/2404.09465
代碼/Code: None
Contrastive Mean-Shift Learning for Generalized Category Discovery
論文/Paper: http://arxiv.org/pdf/2404.09451
代碼/Code: None
The 8th AI City Challenge
論文/Paper: http://arxiv.org/pdf/2404.09432
代碼/Code: None
DetCLIPv3: Towards Versatile Generative Open-vocabulary Object Detection
論文/Paper: http://arxiv.org/pdf/2404.09216
代碼/Code: None
Coreset Selection for Object Detection
論文/Paper: http://arxiv.org/pdf/2404.09161
代碼/Code: None
Exploring Explainability in Video Action Recognition
論文/Paper: http://arxiv.org/pdf/2404.09067
代碼/Code: None
PracticalDG: Perturbation Distillation on Vision-Language Models for Hybrid Domain Generalization
論文/Paper: http://arxiv.org/pdf/2404.09011
代碼/Code: None
MMA-DFER: MultiModal Adaptation of unimodal models for Dynamic Facial Expression Recognition in-the-wild
論文/Paper: http://arxiv.org/pdf/2404.09010
代碼/Code: None
MCPNet: An Interpretable Classifier via Multi-Level Concept Prototypes
論文/Paper: http://arxiv.org/pdf/2404.08968
代碼/Code: None
AMU-Tuning: Effective Logit Bias for CLIP-based Few-shot Learning
論文/Paper: http://arxiv.org/pdf/2404.08958
代碼/Code: https://github.com/tju-sjyj/amu-tuning
Label-free Anomaly Detection in Aerial Agricultural Images with Masked Image Modeling
論文/Paper: http://arxiv.org/pdf/2404.08931
代碼/Code: None
`Eyes of a Hawk and Ears of a Fox': Part Prototype Network for Generalized Zero-Shot Learning
論文/Paper: http://arxiv.org/pdf/2404.08761
代碼/Code: None
Exploring Text-to-Motion Generation with Human Preference
論文/Paper: http://arxiv.org/pdf/2404.09445
代碼/Code: None
