default search action
ACCV 2024: Hanoi, Vietnam - Part III
- Minsu Cho, Ivan Laptev, Du Tran, Angela Yao, Hongbin Zha:
Computer Vision - ACCV 2024 - 17th Asian Conference on Computer Vision, Hanoi, Vietnam, December 8-12, 2024, Proceedings, Part III. Lecture Notes in Computer Science 15474, Springer 2025, ISBN 978-981-96-0907-9
Vision and Language
- Hazel Doughty, Fida Mohammad Thoker, Cees G. M. Snoek:
LocoMotion: Learning Motion-Focused Video-Language Representations. 3-24 - Aozhu Chen, Hazel Doughty, Xirong Li, Cees G. M. Snoek:
Beyond Coarse-Grained Matching in Video-Text Retrieval. 25-43 - Rabin Adhikari, Safal Thapaliya, Manish Dhakal, Bishesh Khanal:
TuneVLSeg: Prompt Tuning Benchmark for Vision-Language Segmentation Models. 44-62 - Ragav Sachdeva, Gyungin Shin, Andrew Zisserman:
Tails Tell Tales: Chapter-Wide Manga Transcriptions with Character Names. 63-80 - Junyu Xie, Tengda Han, Max Bain, Arsha Nagrani, Gül Varol, Weidi Xie, Andrew Zisserman:
AutoAD-Zero: A Training-Free Framework for Zero-Shot Audio Description. 81-97 - Qiuhui Chen, Yi Hong:
MedBLIP: Bootstrapping Language-Image Pretraining from 3D Medical Images and Texts. 98-113 - Erdong Hu, Longteng Guo, Tongtian Yue, Zijia Zhao, Shuning Xue, Jing Liu:
OneDiff: A Generalist Model for Image Difference Captioning. 114-130 - Ting-Yu Chu, Yong-Xiang Lin, Ching-Chun Huang, Kai-Lung Hua:
Enhancing Anchor-Based Weakly Supervised Referring Expression Comprehension with Cross-Modality Attention. 131-147 - Jaakko Kainulainen, Zixin Guo, Jorma Laaksonen:
Diffusion-Based Multimodal Video Captioning. 148-165 - Kazuki Matsuda, Yuiga Wada, Komei Sugiura:
Deneb: A Hallucination-Robust Automatic Evaluation Metric for Image Captioning. 166-182
Video Analysis and Event Recognition
- Jiayan Song, Renjie Pan, Jun Zhou, Hua Yang:
M-RAT: a Multi-grained Retrieval Augmentation Transformer for Image Captioning. 185-203 - Yuxuan Liu, Yasuhisa Fujii, Xinru Zhu, Kayoko Nohara:
Fine-Tuning Large Language Models for Automatic Font Skeleton Generation: Exploration and Analysis. 204-219 - Yun-Zhu Song, Yi-Syuan Chen, Tzu-Ling Lin, Bei Liu, Jianlong Fu, Hong-Han Shuai:
Capture Concept Through Comparison: Vision-and-Language Representation Learning with Intrinsic Information Mining. 220-238 - Yuyan Shi, Chenyi Jiang, Run Shi, Haofeng Zhang:
Do They Share the Same Tail? Learning Individual Compositional Attribute Prototype for Generalized Zero-Shot Learning. 239-256 - Haichen He, Weibin Liu, Weiwei Xing:
BiEfficient: Bidirectionally Prompting Vision-Language Models for Parameter-Efficient Video Recognition. 257-274 - Toby Perrett, Tengda Han, Dima Damen, Andrew Zisserman:
It's Just Another Day: Unique Video Captioning by Discriminitive Prompting. 275-293 - Seungjun Oh, Hyunmo Yang, Eunbyung Park:
Parameter-Efficient Instance-Adaptive Neural Video Compression. 294-311 - Sunghyun Ahn, Youngwan Jo, Kijung Lee, Sanghyun Park:
VideoPatchCore: An Effective Method to Memorize Normality for Video Anomaly Detection. 312-328 - Shibo Gao, Peipei Yang, Linlin Huang:
Scene-Adaptive SVAD Based On Multi-modal Action-Based Feature Extraction. 329-346 - Yash Bhalgat, Vadim Tschernezki, Iro Laina, João F. Henriques, Andrea Vedaldi, Andrew Zisserman:
3D-Aware Instance Segmentation and Tracking in Egocentric Videos. 347-364 - Jaesung Huh, Andrew Zisserman:
Character-Aware Audio-Visual Subtitling in Context. 365-383 - Saptarshi Sinha, Alexandros Stergiou, Dima Damen:
Every Shot Counts: Using Exemplars for Repetition Counting in Videos. 384-402 - Shreyank N. Gowda, Davide Moltisanti, Laura Sevilla-Lara:
Continual Learning Improves Zero-Shot Action Recognition. 403-421 - Yonatan Dinai, Avraham Raviv, Nimrod Harel, Donghoon Kim, Ishay Goldin, Niv Zehngut:
TAPS: Temporal Attention-Based Pruning and Scaling for Efficient Video Action Recognition. 422-438 - Nhat-Minh Nguyen, Tien-Dung Mai, Duy-Dinh Le:
Text Query to Web Image to Video: A Comprehensive Ad-Hoc Video Search. 439-453 - Shreyank N. Gowda, Laura Sevilla-Lara:
Telling Stories for Common Sense Zero-Shot Action Recognition. 454-471
manage site settings
To protect your privacy, all features that rely on external API calls from your browser are turned off by default. You need to opt-in for them to become active. All settings here will be stored as cookies with your web browser. For more information see our F.A.Q.