


default search action
ACCV 2022: Macao, China - Part IV
Face and Gesture
- Puning Yang

, Huaibo Huang
, Zhiyong Wang
, Aijing Yu
, Ran He
:
Confidence-Calibrated Face Image Forgery Detection with Contrastive Representation Distillation. 3-19 - Han Chen

, Yuzhen Lin
, Bin Li
:
Exposing Face Forgery Clues via Retinex-Based Image Enhancement. 20-34 - Mingqiang Chen, Lizhe Liu, Xiaohao Chen, Siyu Zhu:

GB-CosFace: Rethinking Softmax-Based Face Recognition from the Perspective of Open Set Classification. 35-51 - Tianyi Wu

, Yusuke Sugano
:
Learning Video-Independent Eye Contact Segmentation from In-the-Wild Videos. 52-70 - Viresh Ranjan, Minh Hoai Nguyen:

Exemplar Free Class Agnostic Counting. 71-87 - Chaoyu Zhao, Jianjun Qian, Shumin Zhu, Jin Xie, Jian Yang:

Emphasizing Closeness and Diversity Simultaneously for Deep Face Representation. 88-104 - Li-Chen Cheng

, Shu-Chuan Hsu, Pin-Hua Lee, Hsiu-Chieh Lee, Che-Hsien Lin, Jun-Cheng Chen, Chih-Yu Wang:
KinStyle: A Strong Baseline Photorealistic Kinship Face Synthesis with an Optimized StyleGAN Encoder. 105-120 - Jiahe Wang

, Heyan Ding, Shangfei Wang:
Occluded Facial Expression Recognition Using Self-supervised Learning. 121-136 - Nan Gao

, Zhi Zeng
, Guixuan Zhang
, Shuwu Zhang
:
Heterogeneous Avatar Synthesis Based on Disentanglement of Topology and Rendering. 137-152
Pose and Action
- Zhimin Gao

, Peitao Wang, Pei Lv, Xiaoheng Jiang, Qidong Liu, Pichao Wang, Mingliang Xu, Wanqing Li:
Focal and Global Spatial-Temporal Transformer for Skeleton-Based Action Recognition. 155-171 - Rui Hang, MinXian Li:

Spatial-Temporal Adaptive Graph Convolutional Network for Skeleton-Based Action Recognition. 172-188 - Ziyi Zhao, Sena Kiciroglu, Hugues Vinzant, Yuan Cheng, Isinsu Katircioglu

, Mathieu Salzmann, Pascal Fua:
3D Pose Based Feedback for Physical Exercises. 189-205 - Xu Zheng, Yali Zheng

, Shubing Yang:
Generating Multiple Hypotheses for 3D Human Mesh and Pose Using Conditional Generative Adversarial Nets. 206-222 - Na Ye

, Xing Zhang
, Dawei Yan
, Wei Dong
, Qingsen Yan
:
SCOAD: Single-Frame Click Supervision for Online Action Detection. 223-238 - Simon Giebenhain, Urs Waldmann

, Ole Johannsen
, Bastian Goldluecke
:
Neural Puppeteer: Keypoint-Based Neural Rendering of Dynamic Shapes. 239-256 - Yue Zhu

, David Picard
:
Decanus to Legatus: Synthetic Training for 2D-3D Human Pose Lifting. 257-274 - Xiaolin Zhai

, Zhengxi Hu
, Dingye Yang
, Lei Zhou
, Jingtai Liu
:
Social Aware Multi-modal Pedestrian Crossing Behavior Prediction. 275-290 - Haoyuan Gao, Yifaan Zhang, Linhui Sun, Jian Cheng:

Action Representing by Constrained Conditional Mutual Information. 291-306 - Lei Wang

, Piotr Koniusz
:
Temporal-Viewpoint Transportation Plan for Skeletal Few-Shot Action Recognition. 307-326
Video Analysis and Event Recognition
- Xiaolin Zhai

, Zhengxi Hu
, Dingye Yang
, Lei Zhou
, Jingtai Liu
:
Spatial Temporal Network for Image and Skeleton Based Group Activity Recognition. 329-346 - Zhiyi Gao, Yonghong Hou, Wanqing Li, Zihui Guo, Bin Yu:

Learning Using Privileged Information for Zero-Shot Action Recognition. 347-362 - Hang Guo, Zhengxi Hu, Jingtai Liu:

MGTR: End-to-End Mutual Gaze Detection with Transformer. 363-378 - Chuhan Zhang, Ankush Gupta, Andrew Zisserman:

Is an Object-Centric Video Representation Beneficial for Transfer? 379-397 - Zutong Li, Lei Yang:

DCVQE: A Hierarchical Transformer for Video Quality Assessment. 398-416 - Fangqiu Yi, Yanfeng Yang, Tingting Jiang:

Not End-to-End: Explore Multi-Stage Architecture for Online Surgical Phase Recognition. 417-432 - Zhi-Song Liu, Robin Courant, Vicky Kalogeiton

:
FunnyNet: Audiovisual Learning of Funny Moments in Videos. 433-450 - Adriano Fragomeni, Michael Wray

, Dima Damen:
ConTra: (Con)text (Tra)nsformer for Cross-Modal Video Retrieval. 451-468 - Zhimeng Huang

, Chuanmin Jia
, Shanshe Wang
, Siwei Ma
:
A Compressive Prior Guided Mask Predictive Coding Approach for Video Analysis. 469-484 - Jonghwan Mun, Minchul Shin, Gunsoo Han, Sangho Lee, Seongsu Ha, Joonseok Lee, Eun-Sol Kim:

BaSSL: Boundary-aware Self-Supervised Learning for Video Scene Segmentation. 485-501 - Li Li, Liansheng Zhuang, Shenghua Gao, Shafei Wang:

HaViT: Hybrid-Attention Based Vision Transformer for Video Classification. 502-517
Vision and Language
- Yang Zhao, Jiaguo Yu, Shengbin Liao, Zheng Zhang

, Haofeng Zhang:
From Sparse to Dense: Semantic Graph Evolutionary Hashing for Unsupervised Cross-Modal Retrieval. 521-536 - Yizhao Gao, Zhiwu Lu:

SST-VLM: Sparse Sampling-Twice Inspired Video-Language Model. 537-553 - Zhenxing Zheng, Gaoyun An, Shan Cao

, Zhaoqilin Yang
, Qiuqi Ruan:
PromptLearner-CLIP: Contrastive Multi-Modal Action Representation Learning with Context Optimization. 554-570 - Ruonan Zhang, Gaoyun An:

Causal Property Based Anti-conflict Modeling with Hybrid Data Augmentation for Unbiased Scene Graph Generation. 571-587 - Peijie Chen, Qi Li, Saad Biaz, Trung Bui, Anh Nguyen:

gScoreCAM: What Objects Is CLIP Looking At? 588-604 - Vinh Tran, Niranjan Balasubramanian, Minh Hoai:

From Within to Between: Knowledge Distillation for Cross Modality Retrieval. 605-622 - Nasib Ullah, Partha Pratim Mohanta:

Thinking Hallucination for Video Captioning. 623-640 - Jianxiang Dong

, Zhaozheng Yin:
Boundary-Aware Temporal Sentence Grounding with Adaptive Proposal Refinement. 641-657 - Bingjia Li, Jie Wang, Minyi Zhao, Shuigeng Zhou:

Two-Stage Multimodality Fusion for High-Performance Text-Based Visual Question Answering. 658-674 - Huyen Thi Thanh Tran, Takayuki Okatani:

Bright as the Sun: In-depth Analysis of Imagination-Driven Image Captioning. 675-691 - Yuanchao Zheng

, Xiaowei Zhang
:
Heterogeneous Interactive Learning Network for Unsupervised Cross-Modal Retrieval. 692-707
Biometrics
- Ming Wang, Beibei Lin, Xianda Guo, Lincheng Li, Zheng Zhu, Jiande Sun, Shunli Zhang, Yu Liu, Xin Yu

:
GaitStrip: Gait Recognition via Effective Strip-Based Feature Representations and Multi-level Framework. 711-727 - Haipeng Ming, Wenhuan Lu, Wei Zhang:

Soft Label Mining and Average Expression Anchoring for Facial Expression Recognition. 728-744 - Shreya Ghosh

, Abhinav Dhall, Munawar Hayat, Jarrod Knibbe
:
'Labelling the Gaps': A Weakly Supervised Automatic Eye Gaze Estimation. 745-763

manage site settings
To protect your privacy, all features that rely on external API calls from your browser are turned off by default. You need to opt-in for them to become active. All settings here will be stored as cookies with your web browser. For more information see our F.A.Q.


Google
Google Scholar
Semantic Scholar
Internet Archive Scholar
CiteSeerX
ORCID














