Paper registration: April 3 Paper submission deadline: April 5 Paper acceptance notification: May 5 Camera ready date: May 15 Poster/Demo submission deadline: June 30
- Workshop date: July 21
CCD Workshop is part of the CVPR 2017 workshops.
Please see the CVPR webpage for details on venue, accommodations, and other details!
Paper submissions are handled through the workshop’s CMT website: https://cmt3.research.microsoft.com/CCD2017/
The paper submission deadline is Wednesday, April 5th. Submissions can be up to 8 pages in length (excluding reference) prepared using the CVPR-CCD Author Kit. Supplementary material can also be submitted if appropriate. Videos should be in a common format, e.g., MPEG-1, MPEG-4, XviD, or DivX. The submission and review process is double blind; avoid providing any information that may identify the authors. Papers accepted to the workshop will appear in both the CVPR proceedings and on IEEE Xplore.
CCD’s poster & demo session provides an excellent opportunity for attendees to present recently published works or late-breaking results. Note that unlike paper submissions, posters & demos do not have to be original works.
To submit a poster or demo, email us directly at email@example.com with the subject “CCD Poster” by June 30th. Submissions should include a title, authors, affiliations, and a short description of the proposed poster or demo. We also encourage submitting supporting material (e.g. published papers, videos).
|8:30 – 8:45||Welcome and Opening Remarks|
|8:45 – 9:45|| Keynote 1
A Computer Graphics Perspective on Forward and Inverse Light Transport
|9:45 – 10:15||Poster Spotlights|
|10:15 – 10:30||Morning Break|
|10:30 – 11:30||Session 1
Intel RealSense Stereoscopic Depth Cameras
Compressive Light Field Reconstructions using Deep Learning
Invited Talk: One-shot Hyperspectral Imaging using Faced Reflectors
|11:30 – 13:00||Lunch (on your own)|
|13:00 – 14:00||Keynote 2
Computational Display for Virtual and Augmented Reality
|14:00 – 14:45||Session 2
Generating 5D Light Fields in Scattering Media for Representing 3D Images
The Stereoscopic Zoom
|14:45 – 16:00||Afternoon Break + Poster Session|
|16:00 – 16:45||Session 3
Invited Talk: A Wide-Field-of-View Monocentric Light Field Camera
Invited Talk: Inverse-scattering Bridging Micron to Kilometer Scales
|16:45 – 17:00||Closing Remarks|
Matthias Hullin, Professor and Head of Digital Material Appearance Group, University of Bonn
Title: A Computer Graphics Perspective on Forward and Inverse Light Transport
Abstract: A recurring theme in my work is the development of devices and computational methods to measure, simulate, analyse and invert the propagation of light on macroscopic scales. In this talk, I will present a selection of recent research that is linked by the desire to replace specialized and heavyweight optical devices with consumer-grade hardware and computational methods. In particular, I will draw a vision of how computer graphics methodology (forward simulation of light transport, or rendering) could become a useful tool for solving inverse problems in imaging and various remote sensing applications.
Speaker Bio: Matthias Hullin is a professor of Digital Material Appearance at the University of Bonn, a post he took after research stages at the Max Planck Center for Visual Computing and Communication and the University of British Columbia. He obtained a doctorate in computer science from Saarland University (2010) with a dissertation that was awarded the Otto Hahn Medal of the Max Planck Society, and a Diplom in physics from the University of Kaiserslautern (2007). His research is focused on the interface between physics, computer graphics and computer vision.
David Luebke, VP Graphics Research, NVIDIA
Title: Computational Display for Virtual and Augmented Reality
Abstract: Successful wearable displays for virtual & augmented reality face tremendous challenges, including: NEAR-EYE DISPLAY: how to put a display as close to the eye as a pair of eyeglasses, where we cannot bring it into focus? FIELD OF VIEW: how to fill the user’s entire vision with displayed content? RESOLUTION: how to fill that wide field of view with enough pixels, and how to render all of those pixels? A “brute force” display would require something like 10,000×8,000 pixels per eye! BULK: displays should be vanishingly unobtrusive, as light and forgettable as a pair of sunglasses, but the laws of optics dictate that most VR displays today are bulky boxes bigger than ski goggles. FOCUS CUES: today’s VR displays provide binocular display but only a fixed optical depth, thus missing the monocular depth cues from defocus blur and introducing vergence-accommodation conflict. EYEBOX: wearable displays must accommodate a wide range of facial shapes, ideally without being tightly strapped or custom-fitted to the wearer’s head, but many optical designs form a very small exit pupil or exhibit significant distortion across the eyebox.
To overcome these challenges requires understanding and innovation in vision science, optics, display technology, and computer graphics. At NVIDIA Research we have been drawing on all these fields to address (or sidestep!) these challenges, with the goal of moving closer to the “ultimate display”: a diffraction-limited, wide field-of-view, see-through display as light and forgettable as a pair of sunglasses. I will describe several computational display VR/AR prototypes in which we have co-designed the optics, display, and rendering algorithm with the human visual system to achieve new tradeoffs.
Speaker Bio: David Luebke runs graphics research at NVIDIA Research, which he helped found in 2006 after eight years on the faculty of the University of Virginia. His principal research interests are virtual and augmented reality, ray tracing, and real-time rendering.
Ashok Veeraraghavan, Rice University
Ayan Chakrabarti, TTIC
Belen Masia, Universidad de Zaragoza
Brandon Smith, University of Wisconsin-Madison
Douglas Lanman, Oculus Research
Imari Sato, National Institute of Informatics
Ioannis Gkioulekas, Carnegie Mellon University
Jason Holloway, Columbia University
Kaushik Mitra, IIT Madras
Koppal Sanjeev, University of Florida
Kristin Dana, Rutgers University
Oliver Cossairt, Northwestern University
Roarke Horstmeyer, California Institute of Technology
Srikumar Ramalingam, University of Utah
Suren Jayasuriya, Carnegie Mellon University
Tomas Pajdla, Czech Technical University in Prague
Yasuhiro Mukaigawa, NAIST
Yuichi Taguchi, MERL