default search action
Andrew Zisserman
Person information
- affiliation: University of Oxford, UK
Refine list
refinements active!
zoomed in on ?? of ?? records
view refined list in
export refined list as
showing all ?? records
2020 – today
- 2025
- [j106]Ioana Croitoru, Simion-Vlad Bogolin, Marius Leordeanu, Hailin Jin, Andrew Zisserman, Yang Liu, Samuel Albanie:
TeachText: CrossModal text-video retrieval through generalized distillation. Artif. Intell. 338: 104235 (2025) - 2024
- [j105]Jaesung Huh, Joon Son Chung, Arsha Nagrani, Andrew Brown, Jee-weon Jung, Daniel Garcia-Romero, Andrew Zisserman:
The VoxCeleb Speaker Recognition Challenge: A Retrospective. IEEE ACM Trans. Audio Speech Lang. Process. 32: 3850-3866 (2024) - [c498]Ragav Sachdeva, Gyungin Shin, Andrew Zisserman:
Tails Tell Tales: Chapter-Wide Manga Transcriptions with Character Names. ACCV (3) 2024: 63-80 - [c497]Junyu Xie, Tengda Han, Max Bain, Arsha Nagrani, Gül Varol, Weidi Xie, Andrew Zisserman:
AutoAD-Zero: A Training-Free Framework for Zero-Shot Audio Description. ACCV (3) 2024: 81-97 - [c496]Toby Perrett, Tengda Han, Dima Damen, Andrew Zisserman:
It's Just Another Day: Unique Video Captioning by Discriminitive Prompting. ACCV (3) 2024: 275-293 - [c495]Junyu Xie, Charig Yang, Weidi Xie, Andrew Zisserman:
Moving Object Segmentation: All You Need is SAM (and Flow). ACCV (10) 2024: 291-308 - [c494]Yash Bhalgat, Vadim Tschernezki, Iro Laina, João F. Henriques, Andrea Vedaldi, Andrew Zisserman:
3D-Aware Instance Segmentation and Tracking in Egocentric Videos. ACCV (3) 2024: 347-364 - [c493]Jaesung Huh, Andrew Zisserman:
Character-Aware Audio-Visual Subtitling in Context. ACCV (3) 2024: 365-383 - [c492]Carl Doersch, Pauline Luc, Yi Yang, Dilara Gokay, Skanda Koppula, Ankush Gupta, Joseph Heyward, Ignacio Rocco, Ross Goroshin, João Carreira, Andrew Zisserman:
BootsTAP: Bootstrapped Training for Tracking-Any-Point. ACCV (2) 2024: 483-500 - [c491]Ragav Sachdeva, Andrew Zisserman:
The Manga Whisperer: Automatically Generating Transcriptions for Comics. CVPR 2024: 12967-12976 - [c490]Mark Hamilton, Andrew Zisserman, John R. Hershey, William T. Freeman:
Separating the "Chirp" from the "Chat": Self-supervised Visual Grounding of Sound and Language. CVPR 2024: 13117-13127 - [c489]Pinelopi Papalampidi, Skanda Koppula, Shreya Pathak, Justin Chiu, Joe Heyward, Viorica Patraucean, Jiajun Shen, Antoine Miech, Andrew Zisserman, Aida Nematzadeh:
A Simple Recipe for Contrastively Pre-Training Video-First Encoders Beyond 16 Frames. CVPR 2024: 14386-14397 - [c488]Jacob Chalk, Jaesung Huh, Evangelos Kazakos, Andrew Zisserman, Dima Damen:
TIM: A Time Interval Machine for Audio-Visual Action Recognition. CVPR 2024: 18153-18163 - [c487]Tengda Han, Max Bain, Arsha Nagrani, Gül Varol, Weidi Xie, Andrew Zisserman:
AutoAD III: The Prequel - Back to the Pixels. CVPR 2024: 18164-18174 - [c486]Guanqi Zhan, Chuanxia Zheng, Weidi Xie, Andrew Zisserman:
Amodal Ground Truth and Completion in the Wild. CVPR 2024: 28003-28013 - [c485]João Carreira, Michael King, Viorica Patraucean, Dilara Gokay, Catalin Ionescu, Yi Yang, Daniel Zoran, Joseph Heyward, Carl Doersch, Yusuf Aytar, Dima Damen, Andrew Zisserman:
Learning from One Continuous Video Stream. CVPR 2024: 28751-28761 - [c484]Yash Bhalgat, Iro Laina, João F. Henriques, Andrew Zisserman, Andrea Vedaldi:
N2F2: Hierarchical Scene Understanding with Nested Neural Feature Fields. ECCV (59) 2024: 197-214 - [c483]Junyu Xie, Weidi Xie, Andrew Zisserman:
Appearance-Based Refinement for Object-Centric Motion Segmentation. ECCV (39) 2024: 238-256 - [c482]Charig Yang, Weidi Xie, Andrew Zisserman:
Made to Order: Discovering Monotonic Temporal Changes via Self-supervised Video Ordering. ECCV (74) 2024: 268-286 - [c481]Bruno Korbar, Yongqin Xian, Alessio Tonioni, Andrew Zisserman, Federico Tombari:
Text-Conditioned Resampler For Long Form Video Understanding. ECCV (86) 2024: 271-288 - [c480]Akam Rahimi, Triantafyllos Afouras, Andrew Zisserman:
Voicevector: Multimodal Enrolment Vectors for Speaker Separation. ICASSP Workshops 2024: 785-789 - [c479]Bruno Korbar, Jaesung Huh, Andrew Zisserman:
Look, Listen and Recognise: Character-Aware Audio-Visual Subtitling. ICASSP 2024: 2975-2979 - [c478]Vladimir Iashin, Weidi Xie, Esa Rahtu, Andrew Zisserman:
Synchformer: Efficient Synchronization From Sparse Cues. ICASSP 2024: 5325-5329 - [c477]Andreea-Maria Oncescu, João F. Henriques, Andrew Zisserman, Samuel Albanie, A. Sophia Koepke:
A Sound Approach: Using Large Language Models to Generate Audio Descriptions for Egocentric Text-Audio Retrieval. ICASSP 2024: 7300-7304 - [c476]Emmanuelle Bourigault, Amir Jamaludin, Andrew Zisserman:
3D Spine Shape Estimation from Single 2D DXA. MICCAI (5) 2024: 3-13 - [c475]Robin Y. Park, Rhydian Windsor, Amir Jamaludin, Andrew Zisserman:
Automated Spinal MRI Labelling from Reports Using a Large Language Model. MICCAI (5) 2024: 101-111 - [i234]Ragav Sachdeva, Andrew Zisserman:
The Manga Whisperer: Automatically Generating Transcriptions for Comics. CoRR abs/2401.10224 (2024) - [i233]Bruno Korbar, Jaesung Huh, Andrew Zisserman:
Look, Listen and Recognise: Character-Aware Audio-Visual Subtitling. CoRR abs/2401.12039 (2024) - [i232]Vladimir Iashin, Weidi Xie, Esa Rahtu, Andrew Zisserman:
Synchformer: Efficient Synchronization from Sparse Cues. CoRR abs/2401.16423 (2024) - [i231]Carl Doersch, Yi Yang, Dilara Gokay, Pauline Luc, Skanda Koppula, Ankush Gupta, Joseph Heyward, Ross Goroshin, João Carreira, Andrew Zisserman:
BootsTAP: Bootstrapped Training for Tracking-Any-Point. CoRR abs/2402.00847 (2024) - [i230]Andreea-Maria Oncescu, João F. Henriques, Andrew Zisserman, Samuel Albanie, A. Sophia Koepke:
A SOUND APPROACH: Using Large Language Models to generate audio descriptions for egocentric text-audio retrieval. CoRR abs/2402.19106 (2024) - [i229]Yash Bhalgat, Iro Laina, João F. Henriques, Andrew Zisserman, Andrea Vedaldi:
N2F2: Hierarchical Scene Understanding with Nested Neural Feature Fields. CoRR abs/2403.10997 (2024) - [i228]Debidatta Dwibedi, Vidhi Jain, Jonathan Tompson, Andrew Zisserman, Yusuf Aytar:
FlexCap: Generating Rich, Localized, and Flexible Captions in Images. CoRR abs/2403.12026 (2024) - [i227]Jacob Chalk, Jaesung Huh, Evangelos Kazakos, Andrew Zisserman, Dima Damen:
TIM: A Time Interval Machine for Audio-Visual Action Recognition. CoRR abs/2404.05559 (2024) - [i226]Junyu Xie, Charig Yang, Weidi Xie, Andrew Zisserman:
Moving Object Segmentation: All You Need Is SAM (and Flow). CoRR abs/2404.12389 (2024) - [i225]Tengda Han, Max Bain, Arsha Nagrani, Gül Varol, Weidi Xie, Andrew Zisserman:
AutoAD III: The Prequel - Back to the Pixels. CoRR abs/2404.14412 (2024) - [i224]Charig Yang, Weidi Xie, Andrew Zisserman:
Made to Order: Discovering monotonic temporal changes via self-supervised video ordering. CoRR abs/2404.16828 (2024) - [i223]Charles Raude, K. R. Prajwal, Liliane Momeni, Hannah Bull, Samuel Albanie, Andrew Zisserman, Gül Varol:
A Tale of Two Languages: Large-Vocabulary Continuous Sign Language Recognition from Spoken Language Supervision. CoRR abs/2405.10266 (2024) - [i222]Mark Hamilton, Andrew Zisserman, John R. Hershey, William T. Freeman:
Separating the "Chirp" from the "Chat": Self-supervised Visual Grounding of Sound and Language. CoRR abs/2406.05629 (2024) - [i221]Niki Amini-Naieni, Tengda Han, Andrew Zisserman:
CountGD: Multi-Modal Open-World Counting. CoRR abs/2407.04619 (2024) - [i220]Skanda Koppula, Ignacio Rocco, Yi Yang, Joseph Heyward, João Carreira, Andrew Zisserman, Gabriel Brostow, Carl Doersch:
TAPVid-3D: A Benchmark for Tracking Any Point in 3D. CoRR abs/2407.05921 (2024) - [i219]Junyu Xie, Tengda Han, Max Bain, Arsha Nagrani, Gül Varol, Weidi Xie, Andrew Zisserman:
AutoAD-Zero: A Training-Free Framework for Zero-Shot Audio Description. CoRR abs/2407.15850 (2024) - [i218]Debidatta Dwibedi, Yusuf Aytar, Jonathan Tompson, Andrew Zisserman:
OVR: A Dataset for Open Vocabulary Temporal Repetition Counting in Videos. CoRR abs/2407.17085 (2024) - [i217]Ragav Sachdeva, Gyungin Shin, Andrew Zisserman:
Tails Tell Tales: Chapter-Wide Manga Transcriptions with Character Names. CoRR abs/2408.00298 (2024) - [i216]Yash Bhalgat, Vadim Tschernezki, Iro Laina, João F. Henriques, Andrea Vedaldi, Andrew Zisserman:
3D-Aware Instance Segmentation and Tracking in Egocentric Videos. CoRR abs/2408.09860 (2024) - [i215]Jaesung Huh, Joon Son Chung, Arsha Nagrani, Andrew Brown, Jee-weon Jung, Daniel Garcia-Romero, Andrew Zisserman:
The VoxCeleb Speaker Recognition Challenge: A Retrospective. CoRR abs/2408.14886 (2024) - [i214]Jaesung Huh, Andrew Zisserman:
Character-aware audio-visual subtitling in context. CoRR abs/2410.11068 (2024) - [i213]Toby Perrett, Tengda Han, Dima Damen, Andrew Zisserman:
It's Just Another Day: Unique Video Captioning by Discriminative Prompting. CoRR abs/2410.11702 (2024) - [i212]Robin Y. Park, Rhydian Windsor, Amir Jamaludin, Andrew Zisserman:
Automated Spinal MRI Labelling from Reports Using a Large Language Model. CoRR abs/2410.17235 (2024) - [i211]Debidatta Dwibedi, Yusuf Aytar, Jonathan Tompson, Pierre Sermanet, Andrew Zisserman:
A Short Note on Evaluating RepNet for Temporal Repetition Counting in Videos. CoRR abs/2411.08878 (2024) - [i210]Piyush Bagad, Makarand Tapaswi, Cees G. M. Snoek, Andrew Zisserman:
The Sound of Water: Inferring Physical Properties from Pouring Liquids. CoRR abs/2411.11222 (2024) - [i209]Joseph Heyward, João Carreira, Dima Damen, Andrew Zisserman, Viorica Patraucean:
Perception Test 2024: Challenge Summary and a Novel Hour-Long VideoQA Benchmark. CoRR abs/2411.19941 (2024) - [i208]Oishi Deb, K. R. Prajwal, Andrew Zisserman:
New keypoint-based approach for recognising British Sign Language (BSL) from sequences. CoRR abs/2412.09475 (2024) - 2023
- [j104]Michael P. J. Camilleri, Li Zhang, Rasneer S. Bains, Andrew Zisserman, Christopher K. I. Williams:
Persistent animal identification leveraging non-visual markers. Mach. Vis. Appl. 34(4): 68 (2023) - [c474]Sindhu B. Hegde, Andrew Zisserman:
GestSync: Determining who is speaking without a talking head. BMVC 2023: 506-509 - [c473]Niki Amini-Naieni, Kiana Amini-Naieni, Tengda Han, Andrew Zisserman:
Open-world Text-specifed Object Counting. BMVC 2023: 510 - [c472]Yash Bhalgat, João F. Henriques, Andrew Zisserman:
A Light Touch Approach to Teaching Transformers Multi-view Geometry. CVPR 2023: 4958-4969 - [c471]Tengda Han, Max Bain, Arsha Nagrani, Gül Varol, Weidi Xie, Andrew Zisserman:
AutoAD: Movie Description in Context. CVPR 2023: 18930-18940 - [c470]Jaesung Huh, Jacob Chalk, Evangelos Kazakos, Dima Damen, Andrew Zisserman:
Epic-Sounds: A Large-Scale Dataset of Actions that Sound. ICASSP 2023: 1-5 - [c469]Hala Lamdouar, Weidi Xie, Andrew Zisserman:
The Making and Breaking of Camouflage. ICCV 2023: 832-842 - [c468]Carl Doersch, Yi Yang, Mel Vecerík, Dilara Gokay, Ankush Gupta, Yusuf Aytar, João Carreira, Andrew Zisserman:
TAPIR: Tracking Any Point with per-frame Initialization and temporal Refinement. ICCV 2023: 10027-10038 - [c467]Tengda Han, Max Bain, Arsha Nagrani, Gül Varol, Weidi Xie, Andrew Zisserman:
AutoAD II: The Sequel - Who, When, and What in Movie Audio Description. ICCV 2023: 13599-13609 - [c466]Chuhan Zhang, Ankush Gupta, Andrew Zisserman:
Helping Hands: An Object-Aware Ego-Centric Video Recognition Model. ICCV 2023: 13855-13866 - [c465]Liliane Momeni, Mathilde Caron, Arsha Nagrani, Andrew Zisserman, Cordelia Schmid:
Verbs in Action: Improving verb understanding in video-language models. ICCV 2023: 15533-15545 - [c464]Ragav Sachdeva, Andrew Zisserman:
The Change You Want to See (Now in 3D). ICCV (Workshops) 2023: 2052-2061 - [c463]Prannay Kaul, Weidi Xie, Andrew Zisserman:
Multi-Modal Classifiers for Open-Vocabulary Object Detection. ICML 2023: 15946-15969 - [c462]Max Bain, Jaesung Huh, Tengda Han, Andrew Zisserman:
WhisperX: Time-Accurate Speech Transcription of Long-Form Audio. INTERSPEECH 2023: 4489-4493 - [c461]Emmanuelle Bourigault, Amir Jamaludin, Emma Clark, Jeremy Fairbank, Timor Kadir, Andrew Zisserman:
3D Shape Analysis of Scoliosis. ShapeMI@MICCAI 2023: 271-286 - [c460]Rhydian Windsor, Amir Jamaludin, Timor Kadir, Andrew Zisserman:
Vision-Language Modelling For Radiological Imaging and Reports In The Low Data Regime. MIDL 2023: 53-73 - [c459]Jonathan Campbell, Mitchell Dawson, Andrew Zisserman, Weidi Xie, Christoffer Nellåker:
Deep Facial Phenotyping with Mixup Augmentation. MIUA 2023: 133-144 - [c458]Yash Bhalgat, Iro Laina, João F. Henriques, Andrea Vedaldi, Andrew Zisserman:
Contrastive Lift: 3D Object Instance Segmentation by Slow-Fast Contrastive Fusion. NeurIPS 2023 - [c457]Viorica Patraucean, Lucas Smaira, Ankush Gupta, Adrià Recasens, Larisa Markeeva, Dylan Banarse, Skanda Koppula, Joseph Heyward, Mateusz Malinowski, Yi Yang, Carl Doersch, Tatiana Matejovicova, Yury Sulsky, Antoine Miech, Alexandre Fréchette, Hanna Klimczak, Raphael Koster, Junlin Zhang, Stephanie Winkler, Yusuf Aytar, Simon Osindero, Dima Damen, Andrew Zisserman, João Carreira:
Perception Test: A Diagnostic Benchmark for Multimodal Video Models. NeurIPS 2023 - [c456]Sagar Vaze, Andrea Vedaldi, Andrew Zisserman:
No Representation Rules Them All in Category Discovery. NeurIPS 2023 - [c455]Ragav Sachdeva, Andrew Zisserman:
The Change You Want to See. WACV 2023: 3982-3991 - [i207]Adrià Recasens, Jason Lin, João Carreira, Andrew Jaegle, Luyu Wang, Jean-Baptiste Alayrac, Pauline Luc, Antoine Miech, Lucas Smaira, Ross Hemsley, Andrew Zisserman:
Zorro: the masked multimodal transformer. CoRR abs/2301.09595 (2023) - [i206]Jaesung Huh, Jacob Chalk, Evangelos Kazakos, Dima Damen, Andrew Zisserman:
Epic-Sounds: A Large-scale Dataset of Actions That Sound. CoRR abs/2302.00646 (2023) - [i205]Jaesung Huh, Andrew Brown, Jee-weon Jung, Joon Son Chung, Arsha Nagrani, Daniel Garcia-Romero, Andrew Zisserman:
VoxSRC 2022: The Fourth VoxCeleb Speaker Recognition Challenge. CoRR abs/2302.10248 (2023) - [i204]Max Bain, Jaesung Huh, Tengda Han, Andrew Zisserman:
WhisperX: Time-Accurate Speech Transcription of Long-Form Audio. CoRR abs/2303.00747 (2023) - [i203]Relja Arandjelovic, Alex Andonian, Arthur Mensch, Olivier J. Hénaff, Jean-Baptiste Alayrac, Andrew Zisserman:
Three ways to improve feature alignment for open vocabulary detection. CoRR abs/2303.13518 (2023) - [i202]Tengda Han, Max Bain, Arsha Nagrani, Gül Varol, Weidi Xie, Andrew Zisserman:
AutoAD: Movie Description in Context. CoRR abs/2303.16899 (2023) - [i201]Rhydian Windsor, Amir Jamaludin, Timor Kadir, Andrew Zisserman:
Vision-Language Modelling For Radiological Imaging and Reports In The Low Data Regime. CoRR abs/2303.17644 (2023) - [i200]Liliane Momeni, Mathilde Caron, Arsha Nagrani, Andrew Zisserman, Cordelia Schmid:
Verbs in Action: Improving verb understanding in video-language models. CoRR abs/2304.06708 (2023) - [i199]Viorica Patraucean, Lucas Smaira, Ankush Gupta, Adrià Recasens Continente, Larisa Markeeva, Dylan Banarse, Skanda Koppula, Joseph Heyward, Mateusz Malinowski, Yi Yang, Carl Doersch, Tatiana Matejovicova, Yury Sulsky, Antoine Miech, Alexandre Fréchette, Hanna Klimczak, Raphael Koster, Junlin Zhang, Stephanie Winkler, Yusuf Aytar, Simon Osindero, Dima Damen, Andrew Zisserman, João Carreira:
Perception Test: A Diagnostic Benchmark for Multimodal Video Models. CoRR abs/2305.13786 (2023) - [i198]Niki Amini-Naieni, Kiana Amini-Naieni, Tengda Han, Andrew Zisserman:
Open-world Text-specified Object Counting. CoRR abs/2306.01851 (2023) - [i197]Yash Bhalgat, Iro Laina, João F. Henriques, Andrew Zisserman, Andrea Vedaldi:
Contrastive Lift: 3D Object Instance Segmentation by Slow-Fast Contrastive Fusion. CoRR abs/2306.04633 (2023) - [i196]Prannay Kaul, Weidi Xie, Andrew Zisserman:
Multi-Modal Classifiers for Open-Vocabulary Object Detection. CoRR abs/2306.05493 (2023) - [i195]Carl Doersch, Yi Yang, Mel Vecerík, Dilara Gokay, Ankush Gupta, Yusuf Aytar, João Carreira, Andrew Zisserman:
TAPIR: Tracking Any Point with per-frame Initialization and temporal Refinement. CoRR abs/2306.08637 (2023) - [i194]Jaesung Huh, Max Bain, Andrew Zisserman:
OxfordVGG Submission to the EGO4D AV Transcription Challenge. CoRR abs/2307.09006 (2023) - [i193]Chuhan Zhang, Ankush Gupta, Andrew Zisserman:
Helping Hands: An Object-Aware Ego-Centric Video Recognition Model. CoRR abs/2308.07918 (2023) - [i192]Ragav Sachdeva, Andrew Zisserman:
The Change You Want to See (Now in 3D). CoRR abs/2308.10417 (2023) - [i191]Hala Lamdouar, Weidi Xie, Andrew Zisserman:
The Making and Breaking of Camouflage. CoRR abs/2309.03899 (2023) - [i190]Sindhu B. Hegde, Andrew Zisserman:
GestSync: Determining who is speaking without a talking head. CoRR abs/2310.05304 (2023) - [i189]Guanqi Zhan, Chuanxia Zheng, Weidi Xie, Andrew Zisserman:
What Does Stable Diffusion Know about the 3D Scene? CoRR abs/2310.06836 (2023) - [i188]Tengda Han, Max Bain, Arsha Nagrani, Gül Varol, Weidi Xie, Andrew Zisserman:
AutoAD II: The Sequel - Who, When, and What in Movie Audio Description. CoRR abs/2310.06838 (2023) - [i187]Jianbo Jiao, Mohammad Alsharid, Lior Drukker, Aris T. Papageorghiou, Andrew Zisserman, J. Alison Noble:
Show from Tell: Audio-Visual Modelling in Clinical Settings. CoRR abs/2310.16477 (2023) - [i186]Amir Jamaludin, Timor Kadir, Emma Clark, Andrew Zisserman:
Predicting Spine Geometry and Scoliosis from DXA Scans. CoRR abs/2311.09424 (2023) - [i185]Sagar Vaze, Andrea Vedaldi, Andrew Zisserman:
No Representation Rules Them All in Category Discovery. CoRR abs/2311.17055 (2023) - [i184]João Carreira, Michael King, Viorica Patraucean, Dilara Gokay, Catalin Ionescu, Yi Yang, Daniel Zoran, Joseph Heyward, Carl Doersch, Yusuf Aytar, Dima Damen, Andrew Zisserman:
Learning from One Continuous Video Stream. CoRR abs/2312.00598 (2023) - [i183]Pinelopi Papalampidi, Skanda Koppula, Shreya Pathak, Justin Chiu, Joseph Heyward, Viorica Patraucean, Jiajun Shen, Antoine Miech, Andrew Zisserman, Aida Nematzadeh:
A Simple Recipe for Contrastively Pre-training Video-First Encoders Beyond 16 Frames. CoRR abs/2312.07395 (2023) - [i182]Junyu Xie, Weidi Xie, Andrew Zisserman:
Appearance-based Refinement for Object-Centric Motion Segmentation. CoRR abs/2312.11463 (2023) - [i181]Bruno Korbar, Yongqin Xian, Alessio Tonioni, Andrew Zisserman, Federico Tombari:
Text-Conditioned Resampler For Long Form Video Understanding. CoRR abs/2312.11897 (2023) - [i180]Joseph Heyward, João Carreira, Dima Damen, Andrew Zisserman, Viorica Patraucean:
Perception Test 2023: A Summary of the First Challenge And Outcome. CoRR abs/2312.13090 (2023) - [i179]Guanqi Zhan, Chuanxia Zheng, Weidi Xie, Andrew Zisserman:
Amodal Ground Truth and Completion in the Wild. CoRR abs/2312.17247 (2023) - 2022
- [j103]Gül Varol, Liliane Momeni, Samuel Albanie, Triantafyllos Afouras, Andrew Zisserman:
Scaling Up Sign Spotting Through Sign Language Dictionaries. Int. J. Comput. Vis. 130(6): 1416-1439 (2022) - [j102]Manuel J. Marín-Jiménez, Vicky Kalogeiton, Pablo Medina-Suarez, Andrew Zisserman:
LAEO-Net++: Revisiting People Looking at Each Other in Videos. IEEE Trans. Pattern Anal. Mach. Intell. 44(6): 3069-3081 (2022) - [j101]Kai Han, Sylvestre-Alvise Rebuffi, Sébastien Ehrhardt, Andrea Vedaldi, Andrew Zisserman:
AutoNovel: Automatically Discovering and Learning Novel Visual Categories. IEEE Trans. Pattern Anal. Mach. Intell. 44(10): 6767-6781 (2022) - [j100]Triantafyllos Afouras, Joon Son Chung, Andrew W. Senior, Oriol Vinyals, Andrew Zisserman:
Deep Audio-Visual Speech Recognition. IEEE Trans. Pattern Anal. Mach. Intell. 44(12): 8717-8727 (2022) - [c454]Chuhan Zhang, Ankush Gupta, Andrew Zisserman:
Is an Object-Centric Video Representation Beneficial for Transfer? ACCV (4) 2022: 379-397 - [c453]Olivia Wiles, João Carreira, Iain Barr, Andrew Zisserman, Mateusz Malinowski:
Compressed Vision for Efficient Video Understanding. ACCV (7) 2022: 679-695 - [c452]Guanqi Zhan, Weidi Xie, Andrew Zisserman:
A Tri-Layer Plugin to Improve Occluded Detection. BMVC 2022: 250 - [c451]Chang Liu, Yujie Zhong, Andrew Zisserman, Weidi Xie:
CounTR: Transformer-based Generalised Visual Counting. BMVC 2022: 370 - [c450]Vladimir Iashin, Weidi Xie, Esa Rahtu, Andrew Zisserman:
Sparse in Space and Time: Audio-visual Synchronisation with Trainable Selectors. BMVC 2022: 395 - [c449]K. R. Prajwal, Hannah Bull, Liliane Momeni, Samuel Albanie, Gül Varol, Andrew Zisserman:
Weakly-supervised Fingerspelling Recognition in British Sign Language Videos. BMVC 2022: 609 - [c448]Tengda Han, Weidi Xie, Andrew Zisserman:
Turbo Training with Token Dropout. BMVC 2022: 622 - [c447]