IMPORTANT DATES
Dates currently being confirmed; check back.
 

2022
Call for Papers Announced 2 May
Journal-first (JIST/JPI) Submissions

∙ Submission site Opens 2 May 
∙ Journal-first (JIST/JPI) Submissions Due 1 Aug
∙ Final Journal-first manuscripts due 28 Oct
Conference Papers Submissions
∙ Abstract Submission Opens 1 June
∙ Priority Decision Submission Ends 15 July
∙ Extended Submission Ends  19 Sept
∙ FastTrack Conference Proceedings Manuscripts Due 25 Dec 
∙ All Outstanding Proceedings Manuscripts Due
 6 Feb 2023
Registration Opens 1 Dec
Demonstration Applications Due 19 Dec
Early Registration Ends 18 Dec


2023
Hotel Reservation Deadline 6 Jan
Symposium begins
15 Jan


Sponsor




No content found

Stereoscopic Displays and Applications XXXIV

Monday 16 January 2023

Stereoscopic Displays (M1)

Session Chair: Bjorn Sommer, Royal College of Art (United Kingdom)
8:45 – 9:50 AM
Cyril Magnin II

8:45
Conference Welcome

8:50SD&A-383
Evaluating the angular resolution of a simulated light field display in regards to three-dimensionality, motion parallax and viewing experience, Sophie Kergaßner and Jan Fröhlich, Stuttgart Media University (Germany) [view abstract]

 

9:10SD&A-384
CubicSpace for stereo cards: Enabling universal 3D playback on today's display devices, Nicholas Routhier, Mindtrick Innovations Inc. (Canada) [view abstract]

 

9:30SD&A-385
Upgrading the Curtin HIVE’s stereoscopic projection facilities, Andrew J. Woods, Wesley Lamont, and Daniel Adams, Curtin University (Australia) [view abstract]

 



10:20 – 10:50 AM Coffee Break

KEYNOTE: SD&A 1 (M2.1)

Session Chair: Andrew Woods, Curtin University (Australia)
10:50 – 11:50 AM
Cyril Magnin II

SD&A-386
KEYNOTE: The long-awaited arrival of holographic interfaces, Shawn Frayne, Looking Glass Factory (United States) [view abstract]

Inspired by movies in the 80s and 90s, Shawn Frayne has been reaching towards the dream of the hologram for over 20 years. Frayne got his start with a classic laser interference pattern holographic studio he built in high school, followed by training in advanced holographic film techniques at MIT. He has been awarded dozens of patents around the world and serves as co-founder and CEO of Looking Glass Factory, based in Brooklyn and Hong Kong.

 




Stereoscopic Simulation (M2.2)

Session Chair: Andrew Woods, Curtin University (Australia)
11:50 AM – 12:30 PM
Cyril Magnin II

11:50SD&A-387
Estimation of motion sickness in automated vehicles using stereoscopic visual simulation (JIST-first), Yoshihiro Banchi and Takashi Kawai, Waseda University (Japan) [view abstract]

 

12:10
Lenny Lipton - In Memoriam



12:30 – 2:00 PM Lunch

Monday 16 January PLENARY: Neural Operators for Solving PDEs

Session Chair: Robin Jenkin, NVIDIA Corporation (United States)
2:00 PM – 3:00 PM
Cyril Magnin I/II/III

Deep learning surrogate models have shown promise in modeling complex physical phenomena such as fluid flows, molecular dynamics, and material properties. However, standard neural networks assume finite-dimensional inputs and outputs, and hence, cannot withstand a change in resolution or discretization between training and testing. We introduce Fourier neural operators that can learn operators, which are mappings between infinite dimensional spaces. They are independent of the resolution or grid of training data and allow for zero-shot generalization to higher resolution evaluations. When applied to weather forecasting, neural operators capture fine-scale phenomena and have similar skill as gold-standard numerical weather models for predictions up to a week or longer, while being 4-5 orders of magnitude faster.


Anima Anandkumar, Bren professor, California Institute of Technology, and senior director of AI Research, NVIDIA Corporation (United States)

 

Anima Anandkumar is a Bren Professor at Caltech and Senior Director of AI Research at NVIDIA. She is passionate about designing principled AI algorithms and applying them to interdisciplinary domains. She has received several honors such as the IEEE fellowship, Alfred. P. Sloan Fellowship, NSF Career Award, and Faculty Fellowships from Microsoft, Google, Facebook, and Adobe. She is part of the World Economic Forum's Expert Network. Anandkumar received her BTech from Indian Institute of Technology Madras, her PhD from Cornell University, and did her postdoctoral research at MIT and assistant professorship at University of California Irvine.


3:00 – 3:30 PM Coffee Break

EI 2023 Highlights Session

Session Chair: Robin Jenkin, NVIDIA Corporation (United States)
3:30 – 5:00 PM
Cyril Magnin II

Join us for a session that celebrates the breadth of what EI has to offer with short papers selected from EI conferences.

NOTE: The EI-wide "EI 2023 Highlights" session is concurrent with Monday afternoon COIMG, COLOR, IMAGE, and IQSP conference sessions.

 

IQSP-309
Evaluation of image quality metrics designed for DRI tasks with automotive cameras, Valentine Klein, Yiqi LI, Claudio Greco, Laurent Chanas, and Frédéric Guichard, DXOMARK (France) [view abstract]

 

SD&A-224
Human performance using stereo 3D in a helmet mounted display and association with individual stereo acuity, Bonnie Posselt, RAF Centre of Aviation Medicine (United Kingdom) [view abstract]

 

IMAGE-281
Smartphone-enabled point-of-care blood hemoglobin testing with color accuracy-assisted spectral learning, Sang Mok Park1, Yuhyun Ji1, Semin Kwon1, Andrew R. O’Brien2, Ying Wang2, and Young L. Kim1; 1Purdue University and 2Indiana University School of Medicine (United States) [view abstract]

 

AVM-118
Designing scenes to quantify the performance of automotive perception systems, Zhenyi Liu1, Devesh Shah2, Alireza Rahimpour2, Joyce Farrell1, and Brian Wandell1; 1Stanford University and 2Ford Motor Company (United States) [view abstract]

 

VDA-403
Visualizing and monitoring the process of injection molding, Christian A. Steinparz1, Thomas Mitterlehner2, Bernhard Praher2, Klaus Straka1,2, Holger Stitz1,3, and Marc Streit1,3; 1Johannes Kepler University, 2Moldsonics GmbH, and 3datavisyn GmbH (Austria) [view abstract]

 

COIMG-155
Commissioning the James Webb Space Telescope, Joseph M. Howard, NASA Goddard Space Flight Center (United States) [view abstract]

 

HVEI-223
Critical flicker frequency (CFF) at high luminance levels, Alexandre Chapiro1, Nathan Matsuda1, Maliha Ashraf2, and Rafal Mantiuk3; 1Meta (United States), 2University of Liverpool (United Kingdom), and 3University of Cambridge (United Kingdom) [view abstract]

 

HPCI-228
Physics guided machine learning for image-based material decomposition of tissues from simulated breast models with calcifications, Muralikrishnan Gopalakrishnan Meena1, Amir K. Ziabari1, Singanallur Venkatakrishnan1, Isaac R. Lyngaas1, Matthew R. Norman1, Balint Joo1, Thomas L. Beck1, Charles A. Bouman2, Anuj Kapadia1, and Xiao Wang1; 1Oak Ridge National Laboratory and 2Purdue University (United States) [view abstract]

 

3DIA-104
Layered view synthesis for general images, Loïc Dehan, Wiebe Van Ranst, and Patrick Vandewalle, Katholieke University Leuven (Belgium) [view abstract]

 

ISS-329
A self-powered asynchronous image sensor with independent in-pixel harvesting and sensing operations, Ruben Gomez-Merchan, Juan Antonio Leñero-Bardallo, and Ángel Rodríguez-Vázquez, University of Seville (Spain) [view abstract]

 

COLOR-184
Color blindness and modern board games, Alessandro Rizzi1 and Matteo Sassi2; 1Università degli Studi di Milano and 2consultant (Italy) [view abstract]

 


5:00 – 6:15 PM EI 2023 All-Conference Welcome Reception (in the Cyril Magnin Foyer)

SD&A Conference 3D Theater (M4)

Hosts: Eric Kurland, 3-D SPACE (United States); Dan Lawrence, Lightspeed Design, Inc. (United States); John Stern, retired (United States); and Andrew Woods, Curtin University (Australia)
6:00 – 7:30 PM
Cyril Magnin II

This ever-popular session of each year's Stereoscopic Displays and Applications Conference showcases the wide variety of 3D content that is being produced and exhibited around the world. All 3D footage screened in the 3D Theatre Session is shown in high-quality polarized 3D on a large screen. The final program will be announced at the conference and 3D glasses will be provided.



Tuesday 17 January 2023

Stereoscopy in Education and Vergence Accommodation (T1)

Session Chair: Takashi Kawai, Waseda University (Japan)
8:50 – 10:10 AM
Cyril Magnin II

8:50SD&A-388
A review of the VR workshop at the Grand Challenge 2022, Bjorn Sommer, Kam Raoofi, Robert Daniel Philips, Laura Ferrarello, Ashley Hall, and Paul Anderson, Royal College of Art (United Kingdom) [view abstract]

 

9:10SD&A-389
Use of a multi-display, mobile 2D / 3D video wall in surgical operating theatres, Justus F. Ilgner, Thien A. Duong Dinh, Martin Westhofen, and Stephan Hackenberg, RWTH Aachen University (Germany) [view abstract]

 

9:30SD&A-390
Investigating the use of spectacle lenses to alleviate vergence-accommodation mismatch in a stereoscopic remote vision system, Eleanor O'Keefe1, Kirk Moffitt2, Eric Seemiller3, Marc Winterbottom4, and Steven Hadley4; 1KBRwyle, 2Moffitt Consulting, 3EIS, and 4United States Air Force (United States) [view abstract]

 

9:50SD&A-391
Observed optical resolution of light field display: Empirical and simulation results, Yuta Miyanishi, Qi Zhang, Erdem Sahin, and Atanas Gotchev, Tampere University (Finland) [view abstract]

 



10:00 AM – 7:30 PM Industry Exhibition - Tuesday (in the Cyril Magnin Foyer)

10:20 – 10:50 AM Coffee Break

KEYNOTE: SD&A 2 (T2.1)

Session Chair: Nicolas Holliman, King's College London (United Kingdom)
10:50 – 11:50 AM
Cyril Magnin II

This session is jointly sponsored by: Engineering Reality of Virtual Reality 2023, and Stereoscopic Displays and Applications XXXIV.


SD&A-224
KEYNOTE: Human performance using stereo 3D in a helmet mounted display and association with individual stereo acuity, Bonnie Posselt, RAF Centre of Aviation Medicine (United Kingdom) [view abstract]

Wing Commander (Dr) Bonnie Posselt is a medical officer in the RAF (UK) specialising in Aviation and Space Medicine. Bonnie is currently based at the RAF Centre of Aviation Medicine in Bedfordshire, UK, having recently returned from a 3.5year exchange tour to Wright-Patterson Air Force Base in Ohio. While working with the 711th Human Performance Wing and the Air Force Research Laboratory (AFRL) in Ohio, Bonnie undertook a PhD in Helmet Mounted Displays and vision standards in collaboration with the University of Birmingham (UK). Bonnie is a graduate of the University of Manchester, King’s College London, and the International Space University. She is an associate fellow of the Aerospace Medical Association and elected member of the Royal Aeronautical Society.

 




Stereoscopy in VR (T2.2)

Session Chair: Nicolas Holliman, King's College London (United Kingdom)
11:50 AM – 12:30 PM
Cyril Magnin II

This session is jointly sponsored by: Engineering Reality of Virtual Reality 2023, and Stereoscopic Displays and Applications XXXIV.


11:50SD&A-225
Incidence of stereo-blindness in a recent VR distance perception user study, Michael Wiebrands, Andrew J. Woods, and Hugh Riddell, Curtin University (Australia) [view abstract]

 

12:10SD&A-226
Evaluating requirements for design education in a virtual studio environment, Bjorn Sommer1, Ayn Sayuti2, Zidong Lin1, Shefali Bohra1, Emre Kayganaci1, Caroline Yan Zheng1, Chang Hee Lee3, Ashley Hall1, and Paul Anderson1; 1Royal College of Art (United Kingdom), 2Universiti Teknologi MARA (UiTM) (Malaysia), and 3Korea Advanced Institute of Science and Technology (KAIST) (Republic of Korea) [view abstract]

 



12:30 – 2:00 PM Lunch

Tuesday 17 January PLENARY: Embedded Gain Maps for Adaptive Display of High Dynamic Range Images

Session Chair: Robin Jenkin, NVIDIA Corporation (United States)
2:00 PM – 3:00 PM
Cyril Magnin I/II/III

Images optimized for High Dynamic Range (HDR) displays have brighter highlights and more detailed shadows, resulting in an increased sense of realism and greater impact. However, a major issue with HDR content is the lack of consistency in appearance across different devices and viewing environments. There are several reasons, including varying capabilities of HDR displays and the different tone mapping methods implemented across software and platforms. Consequently, HDR content authors can neither control nor predict how their images will appear in other apps.

We present a flexible system that provides consistent and adaptive display of HDR images. Conceptually, the method combines both SDR and HDR renditions within a single image and interpolates between the two dynamically at display time. We compute a Gain Map that represents the difference between the two renditions. In the file, we store a Base rendition (either SDR or HDR), the Gain Map, and some associated metadata. At display time, we combine the Base image with a scaled version of the Gain Map, where the scale factor depends on the image metadata, the HDR capacity of the display, and the viewing environment.


Eric Chan, Fellow, Adobe Inc. (United States)

 

Eric Chan is a Fellow at Adobe, where he develops software for editing photographs. Current projects include Photoshop, Lightroom, Camera Raw, and Digital Negative (DNG). When not writing software, Chan enjoys spending time at his other keyboard, the piano. He is an enthusiastic nature photographer and often combines his photo activities with travel and hiking.


Paul M. Hubel, director of Image Quality in Software Engineering, Apple Inc. (United States)

 

Paul M. Hubel is director of Image Quality in Software Engineering at Apple. He has worked on computational photography and image quality of photographic systems for many years on all aspects of the imaging chain, particularly for iPhone. He trained in optical engineering at University of Rochester, Oxford University, and MIT, and has more than 50 patents on color imaging and camera technology. Hubel is active on the ISO-TC42 committee Digital Photography, where this work is under discussion, and is currently a VP on the IS&T Board. Outside work he enjoys photography, travel, cycling, coffee roasting, and plays trumpet in several bay area ensembles.


3:00 – 3:30 PM Coffee Break

VR Systems and Immersion (T3)

Session Chairs: Margaret Dolinsky, Indiana University (United States) and Ian McDowall, Intuitive Surgical / Fakespace Labs (United States)
3:35 – 5:45 PM
Cyril Magnin II

This session is jointly sponsored by: Engineering Reality of Virtual Reality 2023, and Stereoscopic Displays and Applications XXXIV.


3:35ERVR-213
Display system sharpness modeling and requirement in VR and AR applications, Jiawei Lu, Trisha Lian, and Jerry Jia, Meta (United States) [view abstract]

 

4:05ERVR-214
Tangible extended reality with sensor fusion, Yang Cai, CMU (United States) [view abstract]

 

4:25ERVR-215
Exploring third orality in VR, Semi Ryu, Virginia Commonwealth University (United States) [view abstract]

 

4:45ERVR-216
Immersion, presence and behavioral validity in virtual and augmented environments, Daniel R. Mestre; CNRS and Aix-Marseille University (France) [view abstract]

 

5:05ERVR-217
Immersive security personnel training module for active shooter events, Sharad Sharma and Nishith Mannuru, University of North Texas (United States) [view abstract]

 

5:25ERVR-218
Mobile augmented reality system for object detection, alert, and safety, Sharad Sharma1, Nishith Mannuru1, and Don Engel2; 1University of North Texas and 2University of Maryland, Baltimore County (United States) [view abstract]

 



5:30 – 7:00 PM EI 2023 Symposium Demonstration Session (in the Cyril Magnin Foyer)

Wednesday 18 January 2023

KEYNOTE: AR/VR Special Session 1 (W1)

Session Chair: Alexandre Chapiro, Meta (United States)
9:05 – 10:10 AM
Cyril Magnin II

This session is jointly sponsored by: Engineering Reality of Virtual Reality 2023, Human Vision and Electronic Imaging 2023, and Stereoscopic Displays and Applications XXXIV.



Joint Conference Welcome

HVEI-219
KEYNOTE: Display consideration for AR/VR systems, Ajit Ninan, Reality Labs at Meta (United States) [view abstract]

 

Ajit Ninan is a display industry veteran and led the way to the industry adopting HDR. His inventions & innovations are manifest in millions of shipped HDR TV’s and consumer electronics from multiple companies. He holds 400+ granted patents in imaging and display technology and now works in imaging related to AR/VR at Meta as Senior Director of Applied Perceptual Science and Image Quality. His work spans multiple subjects ranging from Displays, Imaging, Color, Video, Compression, Audio and Networking. His career spans early start-ups to public companies. Ninan is the inventor of the local dimmed quantum dot TV and led the way to the industry adoption of quantum dot displays by working with Vizio, Nanosys and 3M to release the first of its kind R-series QD TV with HDR. He also led the effort with the JPEG committee to standardize JPEG-XT to enable JPEG HDR images. Ninan was inducted as a SMPTE Fellow for his contributions to imaging and standards. The display that caused the world to adopt HDR called the “Pulsar” capable of 4000nits down to .005nits with P3 color in 2010, built by Ninan and his team, has received many awards including the Advanced Imaging Society’s Lumiere award which enabled the development of Dolby Vision and earned Ninan an Emmy.

 


10:00 AM – 3:30 PM Industry Exhibition - Wednesday (in the Cyril Magnin Foyer)

10:20 – 10:50 AM Coffee Break



AR/VR Special Session 2 (W2)

Session Chairs: Nicko Caluya, Ritsumeikan University (Japan) and Alexandre Chapiro, Meta (United States)
10:50 AM – 12:30 PM
Cyril Magnin II

This session is jointly sponsored by: Engineering Reality of Virtual Reality 2023, Human Vision and Electronic Imaging 2023, and Stereoscopic Displays and Applications XXXIV.


10:50HVEI-220
Comparison of AR and VR memory palace quality in second-language vocabulary acquisition (Invited), Xiaoyang Tian, Nicko Caluya, and Damon M. Chandler, Ritsumeikan University (Japan) [view abstract]

 

11:10HVEI-221
Projection mapping for enhancing the perceived deliciousness of food (Invited), Yuichiro Fujimoto, Nara Institute of Science and Technology (Japan) [view abstract]

 

11:30HVEI-222
Real-time imaging processing for low-vision users, Yang Cai, CMU (United States) [view abstract]

 

11:50HVEI-223
Critical flicker frequency (CFF) at high luminance levels, Alexandre Chapiro1, Nathan Matsuda1, Maliha Ashraf2, and Rafal Mantiuk3; 1Meta (United States), 2University of Liverpool (United Kingdom), and 3University of Cambridge (United Kingdom) [view abstract]

 

12:10HVEI-253
A multichannel LED-based lighting approach to improve color discrimination for low vision people, Linna Yang1, Éric Dinet1, Pichayada Katemake2, Alain Trémeau1, and Philippe Colantoni1; 1University Jean Monnet Saint-Etienne (France) and 2Chulalongkorn University (Thailand) [view abstract]

 



12:30 – 2:00 PM Lunch

Wednesday 18 January PLENARY: Bringing Vision Science to Electronic Imaging: The Pyramid of Visibility

Session Chair: Andreas Savakis, Rochester Institute of Technology (United States)
2:00 PM – 3:00 PM
Cyril Magnin I/II/III

Electronic imaging depends fundamentally on the capabilities and limitations of human vision. The challenge for the vision scientist is to describe these limitations to the engineer in a comprehensive, computable, and elegant formulation. Primary among these limitations are visibility of variations in light intensity over space and time, of variations in color over space and time, and of all of these patterns with position in the visual field. Lastly, we must describe how all these sensitivities vary with adapting light level. We have recently developed a structural description of human visual sensitivity that we call the Pyramid of Visibility, that accomplishes this synthesis. This talk shows how this structure accommodates all the dimensions described above, and how it can be used to solve a wide variety of problems in display engineering.


Andrew B. Watson, chief vision scientist, Apple Inc. (United States)

 

Andrew Watson is Chief Vision Scientist at Apple, where he leads the application of vision science to technologies, applications, and displays. His research focuses on computational models of early vision. He is the author of more than 100 scientific papers and 8 patents. He has 21,180 citations and an h-index of 63. Watson founded the Journal of Vision, and served as editor-in-chief 2001-2013 and 2018-2022. Watson has received numerous awards including the Presidential Rank Award from the President of the United States.


3:00 – 3:30 PM Coffee Break

PANEL: AR/VR Special Session (W3.1)

Session Chairs: Nicko Caluya, Ritsumeikan University (Japan) and Alexandre Chapiro, Meta (United States)
Panelists: Alexandre Chapiro, Meta (United States); Yuichiro Fujimoto, Nara Institute of Science and Technology (Japan); Nicolas Holliman, King's College London (United Kingdom); and Ajit Ninan, Reality Labs at Meta (United States)
3:30 – 4:50 PM
Cyril Magnin II

This session is jointly sponsored by: Engineering Reality of Virtual Reality 2023, Human Vision and Electronic Imaging 2023, and Stereoscopic Displays and Applications XXXIV.




DISCUSSION: Wednesday End of Joint Sessions (W3.2)

Session Chair: Damon Chandler, Ritsumeikan University (Japan)
4:50 – 5:30 PM
Cyril Magnin II

This session is jointly sponsored by: Engineering Reality of Virtual Reality 2023, Human Vision and Electronic Imaging 2023, and Stereoscopic Displays and Applications XXXIV.

Please join us for a lively discussion of today's presentations. Participate in an interactive, moderated discussion, where key topics and questions are discussed from many perspectives, reflecting the diverse HVEI community.



5:30 – 7:00 PM EI 2023 Symposium Interactive (Poster) Paper Session (in the Cyril Magnin Foyer)

5:30 – 7:00 PM EI 2023 Meet the Future: A Showcase of Student and Young Professionals Research (in the Cyril Magnin Foyer)

Thursday 19 January 2023

Imaging for XR Workshop 2023

8:45 AM – 6:00 PM
Cyril Magnin I/II

Engage with Others at Imaging for XR.

Held in-person during the Electronic Imaging Symposium and scheduled as the final day for SD&A attendees.


More at XR Imaging Workshop.

6:00 AM – 10:50 PM Coffee Break

No content found

No content found