![IEEE/CVF Convention on Laptop Imaginative and prescient and Sample Recognition (CVPR) 2024 IEEE/CVF Convention on Laptop Imaginative and prescient and Sample Recognition (CVPR) 2024](https://mlr.cdn-apple.com/media/Home_1200x630_48225d82e9.png)
Apple is sponsoring the IEEE/CVF Convention on Laptop Imaginative and prescient and Sample Recognition (CVPR), which is happening in individual from June 17 to 21 in Seattle, Washington. CVPR is the annual pc imaginative and prescient occasion comprising the principle convention and several other co-located workshops and quick programs. Beneath is the schedule of our sponsored workshops and occasions at CVPR 2024.
Schedule
Cease by the Apple sales space within the Arch Constructing, Exhibit Corridor Degree 4, sales space #1905, from 10:30am – 6:30pm PST June 19 and 20; 10:00am – 3:00pm PST on June 21.
Monday, June 17
- ORAL & POSTER PRESENTATION
- Workshop on Image Matching: Local Features & Beyond 2024
- 1:00pm PST – 5:45pm PST, Summit 323
- Affine-based Deformable Attention and Selective Fusion for Semi-dense Matching
- Hongkai Chen, Zixin Luo, Ray Tian, Aron Wang, Lei (VE) Zhou, Xuyang Bai, Mingmin Zhen, Tian Fang, Yanghai Tsin, David McKinnon, Lengthy Quan (The Hong Kong College of Science of Know-how)
Tuesday, June 18
- WORKSHOP
- LatinX in CV (LXCV) at CVPR 2024
- 8:30am PST – 6:00pm PST, Arch 203
- Marcel Santos, Conor O’Brien and Angus Choi are representing Apple on the Latin X workshop occasions.
Wednesday, June 19
- POSTER
- HUGS: Human Gaussian Splatting
- 10:30am PST – 12:00pm PST, #32, Poster Session 1 & Exhibit Corridor (Arch 4A-E)
- Muhammed Kocabas (Max Planck Institute for Clever Methods), Rick Chang, James Gabriel, Oncel Tuzel, Anurag Ranjan
- POSTER
- Direct2.5: Diverse Text-to-3D Generation via Multi-view 2.5D Diffusion
- 5:15pm PST – 6:45pm PST, #382, Poster Session 2 & Exhibit Corridor (Arch 4A-E)
- Yuanxun Lu (Nanjing College), Jingyang Zhang, Shiwei Li, Tian Fang, David McKinnon, Yanghai Tsin, Lengthy Quan (The Hong Kong College of Science and Know-how), Xun Cao (Nanjing College), Yao Yao (Nanjing College)
Thursday, June 20
Friday, June 21
Accepted Papers
Affine-based Deformable Attention and Selective Fusion for Semi-dense Matching
Hongkai Chen, Zixin Luo, Ray Tian, Aron Wang, Lei (VE) Zhou, Xuyang Bai, Mingmin Zhen, Tian Fang, Yanghai Tsin, David McKinnon, Lengthy Quan (The Hong Kong College of Science of Know-how)
Direct2.5: Diverse Text-to-3D Generation via Multi-view 2.5D Diffusion
Yuanxun Lu (Nanjing College), Jingyang Zhang, Shiwei Li, Tian Fang, David McKinnon, Yanghai Tsin, Lengthy Quan (The Hong Kong College of Science and Know-how), Xun Cao (Nanjing College), Yao Yao (Nanjing College)
KPConvX: Modernizing Kernel Point Convolution with Kernel Attention
Hugues Thomas, Hubert Tsai, Tim Barfoot (College of Toronto), Jian (AIML) Zhang
SAM-CLIP: Merging Vision Foundation Models towards Semantic and Spatial Understanding
Haoxiang Wang (College of Illinois Urbana-Champaign), Pavan Kumar Anasosalu Vasu, Fartash Faghri, Raviteja Vemulapalli, Mehrdad Farajtabar, Sachin Mehta, Mohammad Rastegari, Oncel Tuzel, Hadi Pour Ansari
MobileCLIP: Fast Image-Text Models through Multi-Modal Reinforced Training
Pavan Kumar Anasosalu Vasu, Hadi Pour Ansari, Fartash Faghri, Raviteja Vemulapalli, Oncel Tuzel
Probabilistic Speech-Driven 3D Facial Motion Synthesis: New Benchmarks, Methods, and Applications
Karren Yang, Anurag Ranjan, Rick Chang, Raviteja Vemulapalli, Oncel Tuzel
Efficient Diffusion Models without Attention
Jing Nathan Yan (Cornell College), Jiatao Gu, Alexander M. Rush (Cornell College)
HUGS: Human Gaussian Splatting
Muhammed Kocabas (Max Planck Institute for Clever Methods), Rick Chang, James Gabriel, Oncel Tuzel, Anurag Ranjan
HumMUSS: Human Motion Understanding using State Space Models
Arnab Mondal (McGill College), Stefano Alletto, Denis Tome’
Demos
MobileCLIP: Actual-Time Picture-Textual content Models
Wednesday, June 19 – Friday June 21, throughout exhibition hours
Demo reveals zero-shot scene classification operating real-time on an iPhone. Since these fashions align picture and textual content modalities, they will carry out zero-shot picture classification or image-text/text-image retrieval at blazing speeds. The app showcases the analysis work, “MobileCLIP: Quick Picture-Textual content Models by Multi-Modal Strengthened Coaching” being introduced on the identical venue. The app is constructed by David Koski, Megan Maher Welsh with contributions from Hugues Thomas, Mouli Sivapurapu, Jian Zhang.
Circulate Composer for Apple ML
Wednesday, June 19 – Friday June 21, throughout exhibition hours
Demo reveals utilization of Apple ML options on Mac E book Professional and iPad, which leverages a number of applied sciences resembling Imaginative and prescient, CoreML, Core Graphics.
Acknowledgements
Alex Schwing and Philipp Kraehenbuehl are Senior Space Chairs for CVPR 2024.
Alex Toshev, Oncel Tuzel, Mehrdad Farajtabar, Hadi Pour Ansari and Fartash Faghri are Space Chairs for CVPR 2024.
Fartash Faghri, Jason Ren, Jianrui Cai, Jiajia Luo, Jierui Lin, Liangchen Music, Or Dinari, Pavan Kumar Anasosalu Vasu,
Peter Fu, Raviteja Vemulapalli, Haotian Zhang, Hong-You Chen, Wen Shi, Yongzhi Su, Yuyan Li, Trevine Oorloff, Yongxi Lu and Jeff Lai are reviewers for CVPR 2024.
Anshul Shah is a co-organizer for the workshop Learning from Procedural Videos and Language: What is Next?
Jeff Bigham is a co-organizer for the VizWiz Grand Challenge Workshop
Pau Rodriguez Lopez is a co-organizer for the Workshop on Continual Learning in Computer Vision
Jeff Lai has a PhD dissertation chosen for Doctoral Consortium.