{"id":605,"date":"2023-06-18T19:27:01","date_gmt":"2023-06-18T19:27:01","guid":{"rendered":"https:\/\/todaysainews.com\/index.php\/2023\/06\/18\/google-at-cvpr-2023-google-ai-blog\/"},"modified":"2025-04-27T07:33:22","modified_gmt":"2025-04-27T07:33:22","slug":"google-at-cvpr-2023-google-ai-blog","status":"publish","type":"post","link":"https:\/\/todaysainews.com\/index.php\/2023\/06\/18\/google-at-cvpr-2023-google-ai-blog\/","title":{"rendered":"Google at CVPR 2023 \u2013 Google AI Blog"},"content":{"rendered":"<p> [ad_1]<br \/>\n<\/p>\n<p>\nThis week marks the beginning of the premier annual <a href=\"https:\/\/cvpr2023.thecvf.com\/\">Computer Vision and Pattern Recognition<\/a> conference (CVPR 2023), held in-person in Vancouver, BC (with additional virtual content). As a leader in computer vision research and a <a href=\"https:\/\/cvpr2023.thecvf.com\/Conferences\/2023\/Sponsors\">Platinum Sponsor<\/a>, <a href=\"https:\/\/research.google\/\">Google Research<\/a> will have a strong presence across CVPR 2023 with 90 papers being presented at the <a href=\"https:\/\/cvpr2023.thecvf.com\/Conferences\/2023\/AcceptedPapers\">main conference<\/a> and active involvement in over 40 conference <a href=\"https:\/\/cvpr2023.thecvf.com\/Conferences\/2023\/workshop-list\">workshops<\/a> and <a href=\"https:\/\/cvpr2023.thecvf.com\/Conferences\/2023\/tutorial-list\">tutorials<\/a>.\n<\/p>\n<p>\nIf you are attending CVPR this year, please stop by our booth to chat with our researchers who are actively exploring the latest techniques for application to various areas of <a href=\"https:\/\/research.google\/pubs\/?area=machine-perception\">machine perception<\/a>. Our researchers will also be available to talk about and demo several recent efforts, including on-device ML applications with <a href=\"https:\/\/developers.google.com\/mediapipe\">MediaPipe<\/a>, strategies for differential privacy, neural radiance field technologies and much more.\n<\/p>\n<p>\nYou can also learn more about our research being presented at CVPR 2023 in the list below (Google affiliations in <strong>bold<\/strong>).\n<\/p>\n<div style=\"margin-left: 20px;\">\n<p>\n<a href=\"https:\/\/arxiv.org\/pdf\/2211.09682.pdf\">AligNeRF: High-Fidelity Neural Radiance Fields via Alignment-Aware Training<\/a><br \/>\n<br \/><em>Yifan Jiang*, <strong>Peter Hedman<\/strong>, <strong>Ben Mildenhall<\/strong>, Dejia Xu, <strong>Jonathan T. Barron<\/strong>, Zhangyang Wang, Tianfan Xue*<\/em>\n<\/p>\n<p>\n<a href=\"https:\/\/openaccess.thecvf.com\/content\/CVPR2023\/papers\/Kania_BlendFields_Few-Shot_Example-Driven_Facial_Modeling_CVPR_2023_paper.pdf\">BlendFields: Few-Shot Example-Driven Facial Modeling<\/a><br \/>\n<br \/><em>Kacper Kania, Stephan Garbin, <strong>Andrea Tagliasacchi<\/strong>, Virginia Estellers, Kwang Moo Yi, Tomasz Trzcinski, Julien Valentin, Marek Kowalski<\/em>\n<\/p>\n<p>\n<a href=\"https:\/\/arxiv.org\/pdf\/2304.00583.pdf\">Enhancing Deformable Local Features by Jointly Learning to Detect and Describe Keypoints<\/a><br \/>\n<br \/><em>Guilherme Potje, Felipe Cadar, <strong>Andre Araujo<\/strong>, Renato Martins, Erickson Nascimento<\/em>\n<\/p>\n<p>\n<a href=\"https:\/\/openaccess.thecvf.com\/content\/CVPR2023\/papers\/Zhou_How_Can_Objects_Help_Action_Recognition_CVPR_2023_paper.pdf\">How Can Objects Help Action Recognition?<\/a><br \/>\n<br \/><em><strong>Xingyi Zhou<\/strong>, <strong>Anurag Arnab<\/strong>, <strong>Chen Sun<\/strong>, <strong>Cordelia Schmid<\/strong><\/em>\n<\/p>\n<p>\n<a href=\"https:\/\/arxiv.org\/pdf\/2304.12652.pdf\">Hybrid Neural Rendering for Large-Scale Scenes with Motion Blur<\/a><br \/>\n<br \/><em>Peng Dai, <strong>Yinda Zhang<\/strong>, Xin Yu, Xiaoyang Lyu, Xiaojuan Qi<\/em>\n<\/p>\n<p>\n<a href=\"https:\/\/arxiv.org\/pdf\/2303.14396.pdf\">IFSeg: Image-Free Semantic Segmentation via Vision-Language Model<\/a><br \/>\n<br \/><em>Sukmin Yun, Seong Park, <strong>Paul Hongsuck Seo<\/strong>, Jinwoo Shin<\/em>\n<\/p>\n<p>\n<a href=\"https:\/\/openaccess.thecvf.com\/content\/CVPR2023\/papers\/Chen_Learning_From_Unique_Perspectives_User-Aware_Saliency_Modeling_CVPR_2023_paper.pdf\">Learning from Unique Perspectives: User-Aware Saliency Modeling<\/a> (see <a href=\"https:\/\/ai.googleblog.com\/2023\/06\/enabling-delightful-user-experiences.html\">blog post<\/a>)<br \/>\n<br \/><em>Shi Chen*, <strong>Nachiappan Valliappan<\/strong>, <strong>Shaolei Shen<\/strong>, <strong>Xinyu Ye<\/strong>, <strong>Kai Kohlhoff<\/strong>, <strong>Junfeng He<\/strong><\/em>\n<\/p>\n<p>\n<a href=\"https:\/\/arxiv.org\/pdf\/2211.09117.pdf\">MAGE: MAsked Generative Encoder to Unify Representation Learning and Image Synthesis<\/a><br \/>\n<br \/><em>Tianhong Li*,<strong> Huiwen Chang<\/strong>, Shlok Kumar Mishra,<strong> Han Zhang<\/strong>, Dina Katabi,<strong> Dilip Krishnan<\/strong><\/em>\n<\/p>\n<p>\n<a href=\"https:\/\/arxiv.org\/pdf\/2303.17603.pdf\">NeRF-Supervised Deep Stereo<\/a><br \/>\n<br \/><em>Fabio Tosi, <strong>Alessio Tonioni<\/strong>, Daniele Gregorio, Matteo Poggi<\/em>\n<\/p>\n<p>\n<a href=\"https:\/\/openaccess.thecvf.com\/content\/CVPR2023\/papers\/Suhail_Omnimatte3D_Associating_Objects_and_Their_Effects_in_Unconstrained_Monocular_Video_CVPR_2023_paper.pdf\">Omnimatte3D: Associating Objects and their Effects in Unconstrained Monocular Video<\/a><br \/>\n<br \/><em>Mohammed Suhail, <strong>Erika Lu<\/strong>, <strong>Zhengqi Li<\/strong>, <strong>Noah Snavely<\/strong>, Leon Sigal, <strong>Forrester Cole<\/strong><\/em>\n<\/p>\n<p>\n<a href=\"https:\/\/arxiv.org\/pdf\/2211.15654.pdf\">OpenScene: 3D Scene Understanding with Open Vocabularies<\/a><br \/>\n<br \/><em><strong>Songyou Peng<\/strong>, <strong>Kyle Genova<\/strong>, <strong>Chiyu Jiang<\/strong>, <strong>Andrea Tagliasacchi<\/strong>, <strong>Marc Pollefeys<\/strong>, <strong>Thomas Funkhouser<\/strong><\/em>\n<\/p>\n<p>\n<a href=\"https:\/\/arxiv.org\/pdf\/2302.08504.pdf\">PersonNeRF: Personalized Reconstruction from Photo Collections<\/a><br \/>\n<br \/><em>Chung-Yi Weng,<strong> Pratul Srinivasan<\/strong>, <strong>Brian Curless<\/strong>, <strong>Ira Kemelmacher-Shlizerman<\/strong><\/em>\n<\/p>\n<p>\n<a href=\"https:\/\/openaccess.thecvf.com\/content\/CVPR2023\/papers\/Saito_Prefix_Conditioning_Unifies_Language_and_Label_Supervision_CVPR_2023_paper.pdf\">Prefix Conditioning Unifies Language and Label Supervision<\/a><br \/>\n<br \/><em>Kuniaki Saito*, <strong>Kihyuk Sohn<\/strong>, <strong>Xiang Zhang<\/strong>, <strong>Chun-Liang Li<\/strong>, <strong>Chen-Yu Lee<\/strong>, Kate Saenko, <strong>Tomas Pfister<\/strong><\/em>\n<\/p>\n<p>\n<a href=\"https:\/\/openaccess.thecvf.com\/content\/CVPR2023\/papers\/Piergiovanni_Rethinking_Video_ViTs_Sparse_Video_Tubes_for_Joint_Image_and_CVPR_2023_paper.pdf\">Rethinking Video ViTs: Sparse Video Tubes for Joint Image and Video Learning<\/a> (see <a href=\"https:\/\/ai.googleblog.com\/2023\/05\/sparse-video-tubes-for-joint-video-and.html\">blog post<\/a>)<br \/>\n<br \/><em><strong>AJ Piergiovanni<\/strong>, <strong>Weicheng Kuo<\/strong>, <strong>Anelia Angelova<\/strong><\/em>\n<\/p>\n<p>\n<a href=\"https:\/\/arxiv.org\/pdf\/2304.01194.pdf\">Burstormer: Burst Image Restoration and Enhancement Transformer<\/a><br \/>\n<br \/><em>Akshay Dudhane, Syed Waqas Zamir, Salman Khan, Fahad Shahbaz Khan,<strong> Ming-Hsuan Yang<\/strong><\/em>\n<\/p>\n<p>\n<a href=\"https:\/\/arxiv.org\/pdf\/2211.15774.pdf\">Decentralized Learning with Multi-Headed Distillation<\/a><br \/>\n<br \/><em><strong>Andrey Zhmoginov<\/strong>, <strong>Mark Sandler<\/strong>, <strong>Nolan Miller<\/strong>, <strong>Gus Kristiansen<\/strong>, <strong>Max Vladymyrov<\/strong><\/em>\n<\/p>\n<p>\n<a href=\"https:\/\/arxiv.org\/pdf\/2304.02163.pdf\">GINA-3D: Learning to Generate Implicit Neural Assets in the Wild<\/a><br \/>\n<br \/><em>Bokui Shen, Xinchen Yan, Charles R. Qi, Mahyar Najibi, Boyang Deng,<strong> Leonidas Guibas<\/strong>, Yin Zhou, Dragomir Anguelov<\/em>\n<\/p>\n<p>\n<a href=\"https:\/\/arxiv.org\/pdf\/2304.11846.pdf\">Grad-PU: Arbitrary-Scale Point Cloud Upsampling via Gradient Descent with Learned Distance Functions<\/a><br \/>\n<br \/><em>Yun He, <strong>Danhang Tang<\/strong>, <strong>Yinda Zhang<\/strong>, Xiangyang Xue, Yanwei Fu<\/em>\n<\/p>\n<p>\n<a href=\"https:\/\/arxiv.org\/pdf\/2212.11042.pdf\">Hi-LASSIE: High-Fidelity Articulated Shape and Skeleton Discovery from Sparse Image Ensemble<\/a><br \/>\n<br \/><em>Chun-Han Yao*, Wei-Chih Hung, <strong>Yuanzhen Li<\/strong>, <strong>Michael Rubinstein<\/strong>, <strong>Ming-Hsuan Yang<\/strong>, <strong>Varun Jampani<\/strong><\/em>\n<\/p>\n<p>\n<a href=\"https:\/\/arxiv.org\/pdf\/2212.00653.pdf\">Hyperbolic Contrastive Learning for Visual Representations beyond Objects<\/a><br \/>\n<br \/><em>Songwei Ge, Shlok Mishra,<strong> Simon Kornblith<\/strong>, <strong>Chun-Liang Li, <\/strong>David Jacobs<\/em>\n<\/p>\n<p>\n<a href=\"https:\/\/arxiv.org\/pdf\/2210.09276.pdf\">Imagic: Text-Based Real Image Editing with Diffusion Models<\/a><br \/>\n<br \/><em>Bahjat Kawar*,<strong> Shiran Zada<\/strong>, <strong>Oran Lang<\/strong>, <strong>Omer Tov<\/strong>, <strong>Huiwen Chang<\/strong>, <strong>Tali Dekel<\/strong>, <strong>Inbar Mosseri<\/strong>, <strong>Michal Irani<\/strong><\/em>\n<\/p>\n<p>\n<a href=\"https:\/\/arxiv.org\/pdf\/2305.02743.pdf\">Incremental 3D Semantic Scene Graph Prediction from RGB Sequences<\/a><br \/>\n<br \/><em>Shun-Cheng Wu, <strong>Keisuke Tateno<\/strong>, Nassir Navab, <strong>Federico Tombari<\/strong><\/em>\n<\/p>\n<p>\n<a href=\"https:\/\/arxiv.org\/pdf\/2303.00575.pdf\">IPCC-TP: Utilizing Incremental Pearson Correlation Coefficient for Joint Multi-Agent Trajectory Prediction<\/a><br \/>\n<br \/><em>Dekai Zhu, Guangyao Zhai, Yan Di, <strong>Fabian Manhardt<\/strong>, Hendrik Berkemeyer, Tuan Tran, Nassir Navab, <strong>Federico Tombari<\/strong>, Benjamin Busam<\/em>\n<\/p>\n<p>\n<a href=\"https:\/\/arxiv.org\/pdf\/2211.10844.pdf\">Learning to Generate Image Embeddings with User-Level Differential Privacy<\/a><br \/>\n<br \/><strong><em>Zheng Xu, Maxwell Collins, Yuxiao Wang, Liviu Panait, Sewoong Oh, Sean Augenstein, Ting Liu, Florian Schroff, H. Brendan McMahan<\/em><\/strong>\n<\/p>\n<p>\n<a href=\"https:\/\/arxiv.org\/pdf\/2304.05866.pdf\">NoisyTwins: Class-Consistent and Diverse Image Generation Through StyleGANs<\/a><br \/>\n<br \/><em>Harsh Rangwani, Lavish Bansal, Kartik Sharma,<strong> Tejan Karmali<\/strong>, <strong>Varun Jampani<\/strong>, Venkatesh Babu Radhakrishnan<\/em>\n<\/p>\n<p>\n<a href=\"https:\/\/arxiv.org\/pdf\/2211.09794.pdf\">NULL-Text Inversion for Editing Real Images Using Guided Diffusion Models<\/a><br \/>\n<br \/><em>Ron Mokady*, Amir Hertz*, <strong>Kfir Aberman<\/strong>, <strong>Yael Pritch<\/strong>, Daniel Cohen-Or*<\/em>\n<\/p>\n<p>\n<a href=\"https:\/\/arxiv.org\/pdf\/2211.14020.pdf\">SCOOP: Self-Supervised Correspondence and Optimization-Based Scene Flow<\/a><br \/>\n<br \/><em>Itai Lang*,<strong> Dror Aiger<\/strong>, <strong>Forrester Cole<\/strong>, <strong>Shai Avidan<\/strong>, <strong>Michael Rubinstein<\/strong><\/em>\n<\/p>\n<p>\n<a href=\"https:\/\/arxiv.org\/pdf\/2211.11674.pdf\">Shape, Pose, and Appearance from a Single Image via Bootstrapped Radiance Field Inversion<\/a><br \/>\n<br \/><em>Dario Pavllo*,<strong> David Joseph Tan<\/strong>, <strong>Marie-Julie Rakotosaona<\/strong>, <strong>Federico Tombari<\/strong><\/em>\n<\/p>\n<p>\n<a href=\"https:\/\/arxiv.org\/pdf\/2212.12902.pdf\">TexPose: Neural Texture Learning for Self-Supervised 6D Object Pose Estimation<\/a><br \/>\n<br \/><em>Hanzhi Chen, <strong>Fabian Manhardt<\/strong>, Nassir Navab, Benjamin Busam<\/em>\n<\/p>\n<p>\n<a href=\"https:\/\/openaccess.thecvf.com\/content\/CVPR2023\/papers\/Zhu_TryOnDiffusion_A_Tale_of_Two_UNets_CVPR_2023_paper.pdf\">TryOnDiffusion: A Tale of Two UNets<\/a><br \/>\n<br \/><em>Luyang Zhu*, <strong>Dawei Yang<\/strong>, <strong>Tyler Zhu<\/strong>, <strong>Fitsum Reda<\/strong>, <strong>William Chan<\/strong>, <strong>Chitwan Saharia<\/strong>, <strong>Mohammad Norouzi<\/strong>, <strong>Ira Kemelmacher-Shlizerman<\/strong><\/em>\n<\/p>\n<p>\n<a href=\"https:\/\/arxiv.org\/pdf\/2210.03112.pdf\">A New Path: Scaling Vision-and-Language Navigation with Synthetic Instructions and Imitation Learning<\/a><br \/>\n<br \/><em>Aishwarya Kamath*, <strong>Peter Anderson<\/strong>, <strong>Su Wang<\/strong>, Jing Yu Koh*, <strong>Alexander Ku<\/strong>, <strong>Austin Waters<\/strong>, Yinfei Yang*, <strong>Jason Baldridge<\/strong>, <strong>Zarana Parekh<\/strong><\/em>\n<\/p>\n<p>\n<a href=\"https:\/\/arxiv.org\/pdf\/2212.08045.pdf\">CLIPPO: Image-and-Language Understanding from Pixels Only<\/a><br \/>\n<br \/><em><strong>Michael Tschannen<\/strong>, <strong>Basil Mustafa<\/strong>, <strong>Neil Houlsby<\/strong><\/em>\n<\/p>\n<p>\n<a href=\"https:\/\/arxiv.org\/pdf\/2305.04745.pdf\">Controllable Light Diffusion for Portraits<\/a><br \/>\n<br \/><em><strong>David Futschik<\/strong>, <strong>Kelvin Ritland<\/strong>, <strong>James Vecore<\/strong>, <strong>Sean Fanello<\/strong>, <strong>Sergio Orts-Escolano<\/strong>, <strong>Brian Curless<\/strong>, <strong>Daniel S\u00fdkora<\/strong>, <strong>Rohit Pandey<\/strong><\/em>\n<\/p>\n<p>\n<a href=\"https:\/\/openaccess.thecvf.com\/content\/CVPR2023\/papers\/Vasconcelos_CUF_Continuous_Upsampling_Filters_CVPR_2023_paper.pdf\">CUF: Continuous Upsampling Filters<\/a><br \/>\n<br \/><em><strong>Cristina Vasconcelos<\/strong>, <strong>Cengiz Oztireli<\/strong>, <strong>Mark Matthews<\/strong>, <strong>Milad Hashemi<\/strong>, <strong>Kevin Swersky<\/strong>, <strong>Andrea Tagliasacchi<\/strong><\/em>\n<\/p>\n<p>\n<a href=\"https:\/\/arxiv.org\/pdf\/2212.01758.pdf\">Improving Zero-Shot Generalization and Robustness of Multi-modal Models<\/a><br \/>\n<br \/><em>Yunhao Ge*, <strong>Jie Ren<\/strong>, <strong>Andrew Gallagher<\/strong>, <strong>Yuxiao Wang<\/strong>, <strong>Ming-Hsuan Yang<\/strong>, <strong>Hartwig Adam<\/strong>, <strong>Laurent Itti<\/strong>, <strong>Balaji Lakshminarayanan<\/strong>, <strong>Jiaping Zhao<\/strong><\/em>\n<\/p>\n<p>\n<a href=\"https:\/\/arxiv.org\/pdf\/2303.09665.pdf\">LOCATE: Localize and Transfer Object Parts for Weakly Supervised Affordance Grounding<\/a><br \/>\n<br \/><em>Gen Li, <strong>Varun Jampani<\/strong>, <strong>Deqing Sun<\/strong>, Laura Sevilla-Lara<\/em>\n<\/p>\n<p>\n<a href=\"https:\/\/arxiv.org\/pdf\/2303.03361.pdf\">Nerflets: Local Radiance Fields for Efficient Structure-Aware 3D Scene Representation from 2D Supervision<\/a><br \/>\n<br \/><em><strong>Xiaoshuai Zhang<\/strong>, <strong>Abhijit Kundu<\/strong>, <strong>Thomas Funkhouser<\/strong>, <strong>Leonidas Guibas<\/strong>, <strong>Hao Su<\/strong>, <strong>Kyle Genova<\/strong><\/em>\n<\/p>\n<p>\n<a href=\"https:\/\/arxiv.org\/pdf\/2212.01762.pdf\">Self-Supervised AutoFlow<\/a><br \/>\n<br \/><em><strong>Hsin-Ping Huang<\/strong>, <strong>Charles Herrmann<\/strong>, <strong>Junhwa Hur<\/strong>, <strong>Erika Lu<\/strong>, <strong>Kyle Sargent<\/strong>, <strong>Austin Stone<\/strong>, <strong>Ming-Hsuan Yang<\/strong>, <strong>Deqing Sun<\/strong><\/em>\n<\/p>\n<p>\n<a href=\"https:\/\/openaccess.thecvf.com\/content\/CVPR2023\/papers\/Chen_Train-Once-for-All_Personalization_CVPR_2023_paper.pdf\">Train-Once-for-All Personalization<\/a><br \/>\n<br \/><em>Hong-You Chen*,<strong> Yandong Li<\/strong>, <strong>Yin Cui<\/strong>, <strong>Mingda Zhang<\/strong>, Wei-Lun Chao,<strong> Li Zhang<\/strong><\/em>\n<\/p>\n<p>\n<a href=\"https:\/\/arxiv.org\/pdf\/2302.14115.pdf\">Vid2Seq: Large-Scale Pretraining of a Visual Language Model for Dense Video Captioning<\/a> (see <a href=\"https:\/\/ai.googleblog.com\/2023\/03\/vid2seq-pretrained-visual-language.html\">blog post<\/a>)<br \/>\n<br \/><em>Antoine Yang*, <strong>Arsha Nagrani<\/strong>, <strong>Paul Hongsuck Seo<\/strong>, Antoine Miech, <strong>Jordi Pont-Tuset<\/strong>, Ivan Laptev, Josef Sivic, <strong>Cordelia Schmid<\/strong><\/em>\n<\/p>\n<p>\n<a href=\"https:\/\/arxiv.org\/pdf\/2303.14302.pdf\">VILA: Learning Image Aesthetics from User Comments with Vision-Language Pretraining<\/a><br \/>\n<br \/><em><strong>Junjie Ke<\/strong>, <strong>Keren Ye<\/strong>, <strong>Jiahui Yu<\/strong>, <strong>Yonghui Wu<\/strong>, <strong>Peyman Milanfar<\/strong>, <strong>Feng Yang<\/strong><\/em>\n<\/p>\n<p>\n<a href=\"https:\/\/arxiv.org\/pdf\/2211.11152.pdf\">You Need Multiple Exiting: Dynamic Early Exiting for Accelerating Unified Vision Language Model<\/a><br \/>\n<br \/><em>Shengkun Tang, <strong>Yaqing Wang<\/strong>, Zhenglun Kong, Tianchi Zhang, Yao Li, Caiwen Ding, Yanzhi Wang, <strong>Yi Liang<\/strong>, Dongkuan Xu<\/em>\n<\/p>\n<p>\n<a href=\"https:\/\/arxiv.org\/pdf\/2301.05211.pdf\">Accidental Light Probes<\/a><br \/>\n<br \/><em>Hong-Xing Yu, Samir Agarwala, <strong>Charles Herrmann<\/strong>, <strong>Richard Szeliski<\/strong>, <strong>Noah Snavely, <\/strong>Jiajun Wu, <strong>Deqing Sun<\/strong><\/em>\n<\/p>\n<p>\n<a href=\"https:\/\/arxiv.org\/pdf\/2207.09653.pdf\">FedDM: Iterative Distribution Matching for Communication-Efficient Federated Learning<\/a><br \/>\n<br \/><em>Yuanhao Xiong, Ruochen Wang, Minhao Cheng,<strong> Felix Yu<\/strong>, Cho-Jui Hsieh<\/em>\n<\/p>\n<p>\n<a href=\"https:\/\/arxiv.org\/pdf\/2212.08013.pdf\">FlexiViT: One Model for All Patch Sizes<\/a><br \/>\n<br \/><em><strong>Lucas Beyer<\/strong>, <strong>Pavel Izmailov<\/strong>, <strong>Alexander Kolesnikov<\/strong>, <strong>Mathilde Caron<\/strong>, <strong>Simon Kornblith<\/strong>, <strong>Xiaohua Zhai<\/strong>, <strong>Matthias Minderer<\/strong>, <strong>Michael Tschannen<\/strong>, <strong>Ibrahim Alabdulmohsin<\/strong>, <strong>Filip Pavetic<\/strong><\/em>\n<\/p>\n<p>\n<a href=\"https:\/\/arxiv.org\/pdf\/2210.03087.pdf\">Iterative Vision-and-Language Navigation<\/a><br \/>\n<br \/><em>Jacob Krantz, Shurjo Banerjee, Wang Zhu, Jason Corso,<strong> Peter Anderson<\/strong>, Stefan Lee, Jesse Thomason<\/em>\n<\/p>\n<p>\n<a href=\"https:\/\/arxiv.org\/pdf\/2206.08010.pdf\">MoDi: Unconditional Motion Synthesis from Diverse Data<\/a><br \/>\n<br \/><em>Sigal Raab, Inbal Leibovitch, Peizhuo Li,<strong> Kfir Aberman<\/strong>, Olga Sorkine-Hornung, Daniel Cohen-Or<\/em>\n<\/p>\n<p>\n<a href=\"https:\/\/arxiv.org\/pdf\/2303.03369.pdf\">Multimodal Prompting with Missing Modalities for Visual Recognition<\/a><br \/>\n<br \/><em>Yi-Lun Lee,<strong> Yi-Hsuan Tsai<\/strong>, Wei-Chen Chiu,<strong> Chen-Yu Lee<\/strong><\/em>\n<\/p>\n<p>\n<a href=\"https:\/\/openaccess.thecvf.com\/content\/CVPR2023\/papers\/Wang_Scene-Aware_Egocentric_3D_Human_Pose_Estimation_CVPR_2023_paper.pdf\">Scene-Aware Egocentric 3D Human Pose Estimation<\/a><br \/>\n<br \/><em>Jian Wang, Diogo Luvizon, Weipeng Xu, Lingjie Liu,<strong> Kripasindhu Sarkar<\/strong>, Christian Theobalt<\/em>\n<\/p>\n<p>\n<a href=\"https:\/\/arxiv.org\/pdf\/2304.06247.pdf\">ShapeClipper: Scalable 3D Shape Learning from Single-View Images via Geometric and CLIP-Based Consistency<\/a><br \/>\n<br \/><em>Zixuan Huang,<strong> Varun Jampani<\/strong>, Ngoc Anh Thai,<strong> Yuanzhen Li<\/strong>, Stefan Stojanov, James M. Rehg<\/em>\n<\/p>\n<p>\n<a href=\"https:\/\/arxiv.org\/pdf\/2304.05173.pdf\">Improving Image Recognition by Retrieving from Web-Scale Image-Text Data<\/a><br \/>\n<br \/><em><strong>Ahmet Iscen<\/strong>, <strong>Alireza Fathi<\/strong>, <strong>Cordelia Schmid<\/strong><\/em>\n<\/p>\n<p>\n<a href=\"https:\/\/arxiv.org\/pdf\/2304.00341.pdf\">JacobiNeRF: NeRF Shaping with Mutual Information Gradients<\/a><br \/>\n<br \/><em>Xiaomeng Xu, Yanchao Yang, Kaichun Mo, Boxiao Pan, Li Yi,<strong> Leonidas Guibas<\/strong><\/em>\n<\/p>\n<p>\n<a href=\"https:\/\/arxiv.org\/pdf\/2304.01436.pdf\">Learning Personalized High Quality Volumetric Head Avatars from Monocular RGB Videos<\/a><br \/>\n<br \/><em>Ziqian Bai*,<strong> Feitong Tan<\/strong>, <strong>Zeng Huang<\/strong>, <strong>Kripasindhu Sarkar<\/strong>, <strong>Danhang Tang<\/strong>, <strong>Di Qiu<\/strong>, <strong>Abhimitra Meka<\/strong>, <strong>Ruofei Du<\/strong>, <strong>Mingsong Dou<\/strong>, <strong>Sergio Orts-Escolano<\/strong>, <strong>Rohit Pandey<\/strong>, Ping Tan,<strong> Thabo Beeler<\/strong>, <strong>Sean Fanello<\/strong>, <strong>Yinda Zhang<\/strong><\/em>\n<\/p>\n<p>\n<a href=\"https:\/\/arxiv.org\/pdf\/2301.08556.pdf\">NeRF in the Palm of Your Hand: Corrective Augmentation for Robotics via Novel-View Synthesis<\/a><br \/>\n<br \/><em>Allan Zhou, Mo Jin Kim, Lirui Wang,<strong> Pete Florence<\/strong>, <strong>Chelsea Finn<\/strong><\/em>\n<\/p>\n<p>\n<a href=\"https:\/\/arxiv.org\/pdf\/2302.03084.pdf\">Pic2Word: Mapping Pictures to Words for Zero-Shot Composed Image Retrieval<\/a><br \/>\n<br \/><em>Kuniaki Saito*,<strong> Kihyuk Sohn<\/strong>, <strong>Xiang Zhang<\/strong>, <strong>Chun-Liang Li<\/strong>, <strong>Chen-Yu Lee<\/strong>, <strong>Kate Saenko<\/strong>, <strong>Tomas Pfister<\/strong><\/em>\n<\/p>\n<p>\n<a href=\"https:\/\/arxiv.org\/pdf\/2303.13582.pdf\">SCADE: NeRFs from Space Carving with Ambiguity-Aware Depth Estimates<\/a><br \/>\n<br \/><em><strong>Mikaela Uy<\/strong>, <strong>Ricardo Martin Brualla<\/strong>, <strong>Leonidas Guibas<\/strong>, <strong>Ke Li<\/strong><\/em>\n<\/p>\n<p>\n<a href=\"https:\/\/arxiv.org\/pdf\/2212.06820.pdf\">Structured 3D Features for Reconstructing Controllable Avatars<\/a><br \/>\n<br \/><em><strong>Enric Corona<\/strong>, <strong>Mihai Zanfir<\/strong>, <strong>Thiemo Alldieck<\/strong>, <strong>Eduard Gabriel Bazavan<\/strong>, <strong>Andrei Zanfir<\/strong>, <strong>Cristian Sminchisescu<\/strong><\/em>\n<\/p>\n<p>\n<a href=\"https:\/\/arxiv.org\/pdf\/2211.09119.pdf\">Token Turing Machines<\/a><br \/>\n<br \/><em><strong>Michael S. Ryoo<\/strong>, <strong>Keerthana Gopalakrishnan<\/strong>, <strong>Kumara Kahatapitiya<\/strong>, <strong>Ted Xiao<\/strong>, <strong>Kanishka Rao<\/strong>, <strong>Austin Stone<\/strong>, <strong>Yao Lu<\/strong>, <strong>Julian Ibarz<\/strong>, <strong>Anurag Arnab<\/strong><\/em>\n<\/p>\n<p>\n<a href=\"https:\/\/arxiv.org\/pdf\/2212.10957.pdf\">TruFor: Leveraging All-Round Clues for Trustworthy Image Forgery Detection and Localization<\/a><br \/>\n<br \/><em>Fabrizio Guillaro, Davide Cozzolino,<strong> Avneesh Sud<\/strong>, <strong>Nicholas Dufour, <\/strong>Luisa Verdoliva<\/em>\n<\/p>\n<p>\n<a href=\"https:\/\/arxiv.org\/pdf\/2302.07685.pdf\">Video Probabilistic Diffusion Models in Projected Latent Space<\/a><br \/>\n<br \/><em>Sihyun Yu,<strong> Kihyuk Sohn, <\/strong>Subin Kim, Jinwoo Shin<\/em>\n<\/p>\n<p>\n<a href=\"https:\/\/arxiv.org\/pdf\/2210.00990.pdf\">Visual Prompt Tuning for Generative Transfer Learning<\/a><br \/>\n<br \/><em><strong>Kihyuk Sohn<\/strong>, <strong>Yuan Hao<\/strong>, <strong>Jose Lezama<\/strong>, <strong>Luisa Polania<\/strong>, <strong>Huiwen Chang<\/strong>, <strong>Han Zhang<\/strong>, <strong>Irfan Essa<\/strong>, <strong>Lu Jiang<\/strong><\/em>\n<\/p>\n<p>\n<a href=\"https:\/\/arxiv.org\/pdf\/2303.17811.pdf\">Zero-Shot Referring Image Segmentation with Global-Local Context Features<\/a><br \/>\n<br \/><em>Seonghoon Yu,<strong> Paul Hongsuck Seo<\/strong>, Jeany Son<\/em>\n<\/p>\n<p>\n<a href=\"https:\/\/arxiv.org\/pdf\/2303.16501.pdf\">AVFormer: Injecting Vision into Frozen Speech Models for Zero-Shot AV-ASR<\/a> (see <a href=\"https:\/\/ai.googleblog.com\/2023\/06\/avformer-injecting-vision-into-frozen.html\">blog post<\/a>)<br \/>\n<br \/><em><strong>Paul Hongsuck Seo<\/strong>, <strong>Arsha Nagrani<\/strong>, <strong>Cordelia Schmid<\/strong><\/em>\n<\/p>\n<p>\n<a href=\"https:\/\/arxiv.org\/pdf\/2304.03285.pdf\">DC2: Dual-Camera Defocus Control by Learning to Refocus<\/a><br \/>\n<br \/><em><strong>Hadi Alzayer<\/strong>, <strong>Abdullah Abuolaim<\/strong>, <strong>Leung Chun Chan<\/strong>, <strong>Yang Yang<\/strong>, <strong>Ying Chen Lou<\/strong>, <strong>Jia-Bin Huang<\/strong>, <strong>Abhishek Kar<\/strong><\/em>\n<\/p>\n<p>\n<a href=\"https:\/\/openaccess.thecvf.com\/content\/CVPR2023\/papers\/Tripathi_Edges_to_Shapes_to_Concepts_Adversarial_Augmentation_for_Robust_Vision_CVPR_2023_paper.pdf\">Edges to Shapes to Concepts: Adversarial Augmentation for Robust Vision<\/a><br \/>\n<br \/><em>Aditay Tripathi*,<strong> Rishubh Singh<\/strong>, Anirban Chakraborty,<strong> Pradeep Shenoy<\/strong><\/em>\n<\/p>\n<p>\n<a href=\"https:\/\/arxiv.org\/pdf\/2212.09898.pdf\">MetaCLUE: Towards Comprehensive Visual Metaphors Research<\/a><br \/>\n<br \/><em><strong>Arjun R. Akula<\/strong>, <strong>Brendan Driscoll<\/strong>, <strong>Pradyumna Narayana<\/strong>, <strong>Soravit Changpinyo<\/strong>, <strong>Zhiwei Jia<\/strong>, <strong>Suyash Damle<\/strong>, <strong>Garima Pruthi<\/strong>, <strong>Sugato Basu<\/strong>, <strong>Leonidas Guibas<\/strong>, <strong>William T. Freeman<\/strong>, <strong>Yuanzhen Li<\/strong>, <strong>Varun Jampani<\/strong><\/em>\n<\/p>\n<p>\n<a href=\"https:\/\/arxiv.org\/pdf\/2212.13824.pdf\">Multi-Realism Image Compression with a Conditional Generator<\/a><br \/>\n<br \/><em><strong>Eirikur Agustsson<\/strong>, <strong>David Minnen<\/strong>, <strong>George Toderici<\/strong>, <strong>Fabian Mentzer<\/strong><\/em>\n<\/p>\n<p>\n<a href=\"https:\/\/arxiv.org\/pdf\/2212.03267.pdf\">NeRDi: Single-View NeRF Synthesis with Language-Guided Diffusion as General Image Priors<\/a><br \/>\n<br \/><em>Congyue Deng, Chiyu Jiang, Charles R. Qi, Xinchen Yan, Yin Zhou,<strong> Leonidas Guibas<\/strong>, Dragomir Anguelov<\/em>\n<\/p>\n<p>\n<a href=\"https:\/\/arxiv.org\/pdf\/2212.12053.pdf\">On Calibrating Semantic Segmentation Models: Analyses and an Algorithm<\/a><br \/>\n<br \/><em>Dongdong Wang,<strong> Boqing Gong<\/strong>, Liqiang Wang<\/em>\n<\/p>\n<p>\n<a href=\"https:\/\/arxiv.org\/pdf\/2303.13515.pdf\">Persistent Nature: A Generative Model of Unbounded 3D Worlds<\/a><br \/>\n<br \/><em>Lucy Chai,<strong> Richard Tucker<\/strong>, <strong>Zhengqi Li<\/strong>, Phillip Isola,<strong> Noah Snavely<\/strong><\/em>\n<\/p>\n<p>\n<a href=\"https:\/\/arxiv.org\/pdf\/2303.13662.pdf\">Rethinking Domain Generalization for Face Anti-spoofing: Separability and Alignment<\/a><br \/>\n<br \/><em>Yiyou Sun*,<strong> Yaojie Liu<\/strong>, <strong>Xiaoming Liu<\/strong>, Yixuan Li,<strong> Wen-Sheng Chu<\/strong><\/em>\n<\/p>\n<p>\n<a href=\"https:\/\/arxiv.org\/pdf\/2303.13277.pdf\">SINE: Semantic-Driven Image-Based NeRF Editing with Prior-Guided Editing Field<\/a><br \/>\n<br \/><em>Chong Bao,<strong> Yinda Zhang<\/strong>, Bangbang Yang, Tianxing Fan, Zesong Yang, Hujun Bao, Guofeng Zhang, Zhaopeng Cui<\/em>\n<\/p>\n<p>\n<a href=\"https:\/\/arxiv.org\/pdf\/2303.15533.pdf\">Sequential Training of GANs Against GAN-Classifiers Reveals Correlated &#8220;Knowledge Gaps&#8221; Present Among Independently Trained GAN Instances<\/a><br \/>\n<br \/><em><strong>Arkanath Pathak<\/strong>, <strong>Nicholas Dufour<\/strong><\/em>\n<\/p>\n<p>\n<a href=\"https:\/\/arxiv.org\/pdf\/2211.16991.pdf\">SparsePose: Sparse-View Camera Pose Regression and Refinement<\/a><br \/>\n<br \/><em>Samarth Sinha, Jason Zhang,<strong> Andrea Tagliasacchi<\/strong>, Igor Gilitschenski, David Lindell<\/em>\n<\/p>\n<p>\n<a href=\"https:\/\/openaccess.thecvf.com\/content\/CVPR2023\/papers\/Yao_Teacher-Generated_Spatial-Attention_Labels_Boost_Robustness_and_Accuracy_of_Contrastive_Models_CVPR_2023_paper.pdf\">Teacher-Generated Spatial-Attention Labels Boost Robustness and Accuracy of Contrastive Models<\/a><br \/>\n<br \/><em>Yushi Yao,<strong> Chang Ye<\/strong>, <strong>Gamaleldin F. Elsayed<\/strong>, <strong>Junfeng He<\/strong><\/em>\n<\/p>\n<\/div>\n<div style=\"margin-left: 20px;\">\n<p>\n<a href=\"https:\/\/cv4mr.github.io\/\">Computer Vision for Mixed Reality<\/a><br \/>\n<br \/>Speakers include: <strong><em>Ira Kemelmacher-Shlizerman<\/em><\/strong>\n<\/p>\n<p>\n<a href=\"https:\/\/cvpr2023.wad.vision\/\">Workshop on Autonomous Driving (WAD)<\/a><br \/>\n<br \/>Speakers include: <strong><em>Chelsea Finn<\/em><\/strong>\n<\/p>\n<p>\n<a href=\"https:\/\/multimodal-content-moderation.github.io\/\">Multimodal Content Moderation (MMCM)<\/a><br \/>\n<br \/>Organizers include: <strong><em>Chris Bregler<\/em><\/strong><br \/>\n<br \/>Speakers include: <strong><em>Mevan Babakar<\/em><\/strong>\n<\/p>\n<p>\n<a href=\"https:\/\/mcv-workshop.github.io\/#updates\">Medical Computer Vision (MCV)<\/a><br \/>\n<br \/>Speakers include: <strong><em>Shekoofeh Azizi<\/em><\/strong>\n<\/p>\n<p>\n<a href=\"https:\/\/sites.google.com\/corp\/view\/vand-cvpr23\/home\">VAND: Visual Anomaly and Novelty Detection<\/a><br \/>\n<br \/>Speakers include: <em><strong>Yedid Hoshen<\/strong>, <strong>Jie Ren<\/strong><\/em>\n<\/p>\n<p>\n<a href=\"https:\/\/struco3d.github.io\/cvpr2023\/\">Structural and Compositional Learning on 3D Data<\/a><br \/>\n<br \/>Organizers include: <strong><em>Leonidas Guibas<\/em><\/strong><br \/>\n<br \/>Speakers include: <em><strong>Andrea Tagliasacchi<\/strong>, <strong>Fei Xia<\/strong>, <strong>Amir Hertz<\/strong><\/em>\n<\/p>\n<p>\n<a href=\"https:\/\/sites.google.com\/corp\/view\/fgvc10\">Fine-Grained Visual Categorization (FGVC10)<\/a><br \/>\n<br \/>Organizers include: <em><strong>Kimberly Wilber<\/strong>, <strong>Sara Beery<\/strong><\/em><br \/>\n<br \/>Panelists include: <strong><em>Hartwig Adam<\/em><\/strong>\n<\/p>\n<p>\n<a href=\"https:\/\/sites.google.com\/corp\/view\/xrnerf\/\">XRNeRF: Advances in NeRF for the Metaverse<\/a><br \/>\n<br \/>Organizers include: <strong><em>Jonathan T. Barron<\/em><\/strong><br \/>\n  <br \/>Speakers include: <strong><em>Ben Poole<\/em><\/strong>\n<\/p>\n<p>\n<a href=\"https:\/\/sites.google.com\/corp\/view\/omnilabel-workshop-cvpr23\/overview\">OmniLabel: Infinite Label Spaces for Semantic Understanding via Natural Language<\/a><br \/>\n<br \/>Organizers include: <em><strong>Golnaz Ghiasi<\/strong>, <strong>Long Zhao<\/strong><\/em><br \/>\n  <br \/>Speakers include: <strong><em>Vittorio Ferrari<\/em><\/strong>\n<\/p>\n<p>\n<a href=\"https:\/\/holistic-video-understanding.github.io\/workshops\/cvpr2023.html\">Large Scale Holistic Video Understanding<\/a><br \/>\n<br \/>Organizers include: <strong><em>David Ross<\/em><\/strong><br \/>\n  <br \/>Speakers include: <strong><em>Cordelia Schmid<\/em><\/strong>\n<\/p>\n<p>\n<a href=\"https:\/\/nice.lgresearch.ai\/\">New Frontiers for Zero-Shot Image Captioning Evaluation (NICE)<\/a><br \/>\n<br \/>Speakers include: <strong><em>Cordelia Schmid<\/em><\/strong>\n<\/p>\n<p>\n<a href=\"https:\/\/ccd2023.github.io\/\">Computational Cameras and Displays (CCD)<\/a><br \/>\n<br \/>Organizers include: <strong><em>Ulugbek Kamilov<\/em><\/strong><br \/>\n<br \/>Speakers include: <strong><em>Mauricio Delbracio<\/em><\/strong>\n<\/p>\n<p>\n<a href=\"https:\/\/gazeworkshop.github.io\/2023\/\">Gaze Estimation and Prediction in the Wild (GAZE)<\/a><br \/>\n<br \/>Organizers include: <strong><em>Thabo Beele<\/em><\/strong><br \/>\n  <br \/>Speakers include: <strong><em>Erroll Wood<\/em><\/strong>\n<\/p>\n<p>\n<a href=\"https:\/\/sites.google.com\/corp\/view\/fgahi2023\/home\">Face and Gesture Analysis for Health Informatics (FGAHI)<\/a><br \/>\n<br \/>Speakers include: <strong><em>Daniel McDuff<\/em><\/strong>\n<\/p>\n<p>\n<a href=\"https:\/\/www.cv4animals.com\/\">Computer Vision for Animal Behavior Tracking and Modeling (CV4Animals)<\/a><br \/>\n<br \/>Organizers include: <strong><em>Sara Beery<\/em><\/strong><br \/>\n<br \/>Speakers include: <strong><em>Arsha Nagrani<\/em><\/strong>\n<\/p>\n<p>\n<a href=\"https:\/\/sites.google.com\/corp\/view\/cvpr2023-3d-vision-robotics\">3D Vision and Robotics<\/a><br \/>\n<br \/>Speakers include: <strong><em>Pete Florence<\/em><\/strong>\n<\/p>\n<p>\n<a href=\"https:\/\/sites.google.com\/corp\/view\/cvpr2023-3d-vision-robotics\">End-to-End Autonomous Driving: Perception, Prediction, Planning and Simulation (E2EAD)<\/a><br \/>\n<br \/>Organizers include: <strong><em>Anurag Arnab<\/em><\/strong>\n<\/p>\n<p>\n<a href=\"https:\/\/opendrivelab.com\/e2ead\/cvpr23\">End-to-End Autonomous Driving: Emerging Tasks and Challenges<\/a><br \/>\n<br \/>Speakers include: <strong><em>Sergey Levine<\/em><\/strong>\n<\/p>\n<p>\n<a href=\"https:\/\/mula-workshop.github.io\/\">Multi-Modal Learning and Applications (MULA)<\/a><br \/>\n<br \/>Speakers include: <strong><em>Aleksander Ho\u0142y\u0144ski<\/em><\/strong>\n<\/p>\n<p>\n<a href=\"https:\/\/sites.google.com\/view\/sdas2023\/\">Synthetic Data for Autonomous Systems (SDAS)<\/a><br \/>\n<br \/>Speakers include: <strong><em>Lukas Hoyer<\/em><\/strong>\n<\/p>\n<p>\n<a href=\"https:\/\/sites.google.com\/corp\/view\/vdu-cvpr23\">Vision Datasets Understanding<\/a><br \/>\n<br \/>Organizers include: <strong><em>Jos\u00e9 Lezama<\/em><\/strong><br \/>\n<br \/>Speakers include: <strong><em>Vijay Janapa Reddi<\/em><\/strong>\n<\/p>\n<p>\n<a href=\"https:\/\/sites.google.com\/corp\/view\/ieeecvf-cvpr2023-precognition\/\">Precognition: Seeing Through the Future<\/a><br \/>\n<br \/>Organizers include: <strong><em>Utsav Prabhu<\/em><\/strong>\n<\/p>\n<p>\n<a href=\"https:\/\/cvlai.net\/ntire\/2023\/\">New Trends in Image Restoration and Enhancement (NTIRE)<\/a><br \/>\n<br \/>Organizers include: <strong><em>Ming-Hsuan Yang<\/em><\/strong>\n<\/p>\n<p>\n<a href=\"https:\/\/generative-vision.github.io\/workshop-CVPR-23\/\">Generative Models for Computer Vision<\/a><br \/>\n<br \/>Speakers include: <em><strong>Ben Mildenhall<\/strong>, <strong>Andrea Tagliasacchi<\/strong><\/em>\n<\/p>\n<p>\n<a href=\"https:\/\/robustart.github.io\/\">Adversarial Machine Learning on Computer Vision: Art of Robustness<\/a><br \/>\n<br \/>Organizers include: <strong><em>Xinyun Chen<\/em><\/strong><br \/>\n  <br \/>Speakers include: <strong><em>Deqing Sun<\/em><\/strong>\n<\/p>\n<p>\n<a href=\"https:\/\/sites.google.com\/corp\/view\/wmf2023\/home\">Media Forensics<\/a><br \/>\n<br \/>Speakers include: <strong><em>Nicholas Carlini<\/em><\/strong>\n<\/p>\n<p>\n<a href=\"http:\/\/taodataset.org\/workshop\/cvpr23\/\">Tracking and Its Many Guises: Tracking Any Object in Open-World<\/a><br \/>\n<br \/>Organizers include: <strong><em>Paul Voigtlaender<\/em><\/strong>\n<\/p>\n<p>\n<a href=\"https:\/\/scene-understanding.com\/\">3D Scene Understanding for Vision, Graphics, and Robotics<\/a><br \/>\n<br \/>Speakers include: <strong><em>Andy Zeng<\/em><\/strong>\n<\/p>\n<p>\n<a href=\"https:\/\/www.es.ele.tue.nl\/cvpm23\/\">Computer Vision for Physiological Measurement (CVPM)<\/a><br \/>\n<br \/>Organizers include: <strong><em>Daniel McDuff<\/em><\/strong>\n<\/p>\n<p>\n<a href=\"https:\/\/ibug.doc.ic.ac.uk\/resources\/cvpr-2023-5th-abaw\/\">Affective Behaviour Analysis In-the-Wild<\/a><br \/>\n<br \/>Organizers include: <strong><em>Stefanos Zafeiriou<\/em><\/strong>\n<\/p>\n<p>\n<a href=\"https:\/\/sites.google.com\/corp\/view\/ec3v-cvpr2023\/home\">Ethical Considerations in Creative Applications of Computer Vision (EC3V)<\/a><br \/>\n<br \/>Organizers include: <em><strong>Rida Qadri<\/strong>, <strong>Mohammad Havaei<\/strong>, <strong>Fernando Diaz<\/strong>, <strong>Emily Denton<\/strong>, <strong>Sarah Laszlo<\/strong>, <strong>Negar Rostamzadeh<\/strong>, <strong>Pamela Peter-Agbia<\/strong>, <strong>Eva Kozanecka<\/strong><\/em>\n<\/p>\n<p>\n<a href=\"https:\/\/vizwiz.org\/workshops\/2023-workshop\/\">VizWiz Grand Challenge: Describing Images and Videos Taken by Blind People<\/a><br \/>\n<br \/>Speakers include: <strong><em>Haoran Qi<\/em><\/strong>\n<\/p>\n<p>\n<a href=\"https:\/\/sites.google.com\/corp\/view\/ecv23\/home\">Efficient Deep Learning for Computer Vision<\/a> (see <a href=\"https:\/\/ai.googleblog.com\/2023\/06\/speed-is-all-you-need-on-device.html\">blog post<\/a>)<br \/>\n<br \/>Organizers include: <em><strong>Andrew Howard<\/strong>, <strong>Chas Leichner<\/strong><\/em><br \/>\n  <br \/>Speakers include: <strong><em>Andrew Howard<\/em><\/strong>\n<\/p>\n<p>\n<a href=\"https:\/\/sites.google.com\/corp\/view\/vcdw2023\/\">Visual Copy Detection<\/a><br \/>\n<br \/>Organizers include: <strong><em>Priya Goyal<\/em><\/strong>\n<\/p>\n<p>\n<a href=\"https:\/\/3dmv2023.github.io\/\">Learning 3D with Multi-View Supervision (3DMV)<\/a><br \/>\n<br \/>Speakers include: <strong><em>Ben Poole<\/em><\/strong>\n<\/p>\n<p>\n<a href=\"https:\/\/image-matching-workshop.github.io\/\">Image Matching: Local Features and Beyond<\/a><br \/>\n<br \/>Organizers include: <strong><em>Eduard Trulls<\/em><\/strong>\n<\/p>\n<p>\n<a href=\"https:\/\/vision4allseason.net\/\">Vision for All Seasons: Adverse Weather and Lightning Conditions (V4AS)<\/a><br \/>\n<br \/>Organizers include: <strong><em>Lukas Hoyer<\/em><\/strong>\n<\/p>\n<p>\n<a href=\"https:\/\/sites.google.com\/corp\/view\/t4v-cvpr23\">Transformers for Vision (T4V)<\/a><br \/>\n<br \/>Speakers include: <em><strong>Cordelia Schmid<\/strong>, <strong>Huiwen Chang<\/strong><\/em>\n<\/p>\n<p>\n<a href=\"https:\/\/sites.google.com\/corp\/view\/academic-cv\/\">Scholars vs Big Models \u2014 How Can Academics Adapt?<\/a><br \/>\n<br \/>Organizers include: <strong><em>Sara Beery<\/em><\/strong><br \/>\n<br \/>Speakers include: <em><strong>Jonathan T. Barron<\/strong>, <strong>Cordelia Schmid<\/strong><\/em>\n<\/p>\n<p>\n<a href=\"http:\/\/www.scan-net.org\/cvpr2023workshop\/\">ScanNet Indoor Scene Understanding Challenge<\/a><br \/>\n<br \/>Speakers include: <strong><em>Tom Funkhouser<\/em><\/strong>\n<\/p>\n<p>\n<a href=\"https:\/\/cvmi-workshop.github.io\/new.html\">Computer Vision for Microscopy Image Analysis<\/a><br \/>\n<br \/>Speakers include: <strong><em>Po-Hsuan Cameron Chen<\/em><\/strong>\n<\/p>\n<p>\n<a href=\"https:\/\/embeddedvisionworkshop.wordpress.com\/\">Embedded Vision<\/a><br \/>\n<br \/>Speakers include: <strong><em>Rahul Sukthankar<\/em><\/strong>\n<\/p>\n<p>\n<a href=\"https:\/\/sightsound.org\/\">Sight and Sound<\/a><br \/>\n<br \/>Organizers include: <em><strong>Arsha Nagrani<\/strong>, <strong>William Freeman<\/strong><\/em>\n<\/p>\n<p>\n<a href=\"https:\/\/ai4cc.net\/\">AI for Content Creation<\/a><br \/>\n<br \/>Organizers include: <em><strong>Deqing Sun<\/strong>, <strong>Huiwen Chang<\/strong>, <strong>Lu Jiang<\/strong><\/em>\n<\/p>\n<p>\nSpeakers include: <em><strong>Ben Mildenhall<\/strong>, <strong>Tim Salimans<\/strong>, <strong>Yuanzhen Li<\/strong><\/em>\n<\/p>\n<p>\n<a href=\"https:\/\/computer-vision-in-the-wild.github.io\/cvpr-2023\/https:\/\/computer-vision-in-the-wild.github.io\/cvpr-2023\/\">Computer Vision in the Wild<\/a><br \/>\n<br \/>Organizers include: <em><strong>Xiuye Gu<\/strong>, <strong>Neil Houlsby<\/strong><\/em><br \/>\n  <br \/>Speakers include: <em><strong>Boqing Gong<\/strong>, <strong>Anelia Angelova<\/strong><\/em>\n<\/p>\n<p>\n<a href=\"https:\/\/vispr-workshop.github.io\/\">Visual Pre-Training for Robotics<\/a><br \/>\n<br \/>Organizers include: <strong><em>Mathilde Caron<\/em><\/strong>\n<\/p>\n<p>\n<a href=\"https:\/\/sites.google.com\/corp\/view\/omnicv2023\/home\">Omnidirectional Computer Vision<\/a><br \/>\n<br \/>Organizers include: <strong><em>Yi-Hsuan Tsai<\/em><\/strong>\n<\/p>\n<\/div>\n<p>[ad_2]<br \/>\n<br \/><a href=\"http:\/\/ai.googleblog.com\/2023\/06\/google-at-cvpr-2023.html\">Source link <\/a><\/p>\n","protected":false},"excerpt":{"rendered":"<p>[ad_1] This week marks the beginning of the premier annual Computer Vision and Pattern Recognition conference (CVPR 2023),<\/p>\n","protected":false},"author":2,"featured_media":606,"comment_status":"closed","ping_status":"open","sticky":false,"template":"","format":"standard","meta":{"footnotes":""},"categories":[20],"tags":[],"class_list":["post-605","post","type-post","status-publish","format-standard","has-post-thumbnail","hentry","category-google-ai"],"_links":{"self":[{"href":"https:\/\/todaysainews.com\/index.php\/wp-json\/wp\/v2\/posts\/605","targetHints":{"allow":["GET"]}}],"collection":[{"href":"https:\/\/todaysainews.com\/index.php\/wp-json\/wp\/v2\/posts"}],"about":[{"href":"https:\/\/todaysainews.com\/index.php\/wp-json\/wp\/v2\/types\/post"}],"author":[{"embeddable":true,"href":"https:\/\/todaysainews.com\/index.php\/wp-json\/wp\/v2\/users\/2"}],"replies":[{"embeddable":true,"href":"https:\/\/todaysainews.com\/index.php\/wp-json\/wp\/v2\/comments?post=605"}],"version-history":[{"count":1,"href":"https:\/\/todaysainews.com\/index.php\/wp-json\/wp\/v2\/posts\/605\/revisions"}],"predecessor-version":[{"id":2772,"href":"https:\/\/todaysainews.com\/index.php\/wp-json\/wp\/v2\/posts\/605\/revisions\/2772"}],"wp:featuredmedia":[{"embeddable":true,"href":"https:\/\/todaysainews.com\/index.php\/wp-json\/wp\/v2\/media\/606"}],"wp:attachment":[{"href":"https:\/\/todaysainews.com\/index.php\/wp-json\/wp\/v2\/media?parent=605"}],"wp:term":[{"taxonomy":"category","embeddable":true,"href":"https:\/\/todaysainews.com\/index.php\/wp-json\/wp\/v2\/categories?post=605"},{"taxonomy":"post_tag","embeddable":true,"href":"https:\/\/todaysainews.com\/index.php\/wp-json\/wp\/v2\/tags?post=605"}],"curies":[{"name":"wp","href":"https:\/\/api.w.org\/{rel}","templated":true}]}}