IMPORTANT DATES
Dates currently being confirmed; check back.
 

2022
Call for Papers Announced 2 May
Journal-first (JIST/JPI) Submissions

∙ Submission site Opens 2 May 
∙ Journal-first (JIST/JPI) Submissions Due 1 Aug
∙ Final Journal-first manuscripts due 28 Oct
Conference Papers Submissions
∙ Abstract Submission Opens 1 June
∙ Priority Decision Submission Ends 15 July
∙ Extended Submission Ends  19 Sept
∙ FastTrack Conference Proceedings Manuscripts Due 25 Dec 
∙ All Outstanding Proceedings Manuscripts Due
 6 Feb 2023
Registration Opens 1 Dec
Demonstration Applications Due 19 Dec
Early Registration Ends 18 Dec


2023
Hotel Reservation Deadline 6 Jan
Symposium begins
15 Jan


Best Paper Sponsor


 

No content found

Imaging Sensors and Systems 2023

Monday 16 January 2023

10:20 – 10:50 AM Coffee Break

12:30 – 2:00 PM Lunch

Monday 16 January PLENARY: Neural Operators for Solving PDEs

Session Chair: Robin Jenkin, NVIDIA Corporation (United States)
2:00 PM – 3:00 PM
Cyril Magnin I/II/III

Deep learning surrogate models have shown promise in modeling complex physical phenomena such as fluid flows, molecular dynamics, and material properties. However, standard neural networks assume finite-dimensional inputs and outputs, and hence, cannot withstand a change in resolution or discretization between training and testing. We introduce Fourier neural operators that can learn operators, which are mappings between infinite dimensional spaces. They are independent of the resolution or grid of training data and allow for zero-shot generalization to higher resolution evaluations. When applied to weather forecasting, neural operators capture fine-scale phenomena and have similar skill as gold-standard numerical weather models for predictions up to a week or longer, while being 4-5 orders of magnitude faster.


Anima Anandkumar, Bren professor, California Institute of Technology, and senior director of AI Research, NVIDIA Corporation (United States)

 

Anima Anandkumar is a Bren Professor at Caltech and Senior Director of AI Research at NVIDIA. She is passionate about designing principled AI algorithms and applying them to interdisciplinary domains. She has received several honors such as the IEEE fellowship, Alfred. P. Sloan Fellowship, NSF Career Award, and Faculty Fellowships from Microsoft, Google, Facebook, and Adobe. She is part of the World Economic Forum's Expert Network. Anandkumar received her BTech from Indian Institute of Technology Madras, her PhD from Cornell University, and did her postdoctoral research at MIT and assistant professorship at University of California Irvine.


3:00 – 3:30 PM Coffee Break

EI 2023 Highlights Session

Session Chair: Robin Jenkin, NVIDIA Corporation (United States)
3:30 – 5:00 PM
Cyril Magnin II

Join us for a session that celebrates the breadth of what EI has to offer with short papers selected from EI conferences.

NOTE: The EI-wide "EI 2023 Highlights" session is concurrent with Monday afternoon COIMG, COLOR, IMAGE, and IQSP conference sessions.

 

IQSP-309
Evaluation of image quality metrics designed for DRI tasks with automotive cameras, Valentine Klein, Yiqi LI, Claudio Greco, Laurent Chanas, and Frédéric Guichard, DXOMARK (France) [view abstract]

 

SD&A-224
Human performance using stereo 3D in a helmet mounted display and association with individual stereo acuity, Bonnie Posselt, RAF Centre of Aviation Medicine (United Kingdom) [view abstract]

 

IMAGE-281
Smartphone-enabled point-of-care blood hemoglobin testing with color accuracy-assisted spectral learning, Sang Mok Park1, Yuhyun Ji1, Semin Kwon1, Andrew R. O’Brien2, Ying Wang2, and Young L. Kim1; 1Purdue University and 2Indiana University School of Medicine (United States) [view abstract]

 

AVM-118
Designing scenes to quantify the performance of automotive perception systems, Zhenyi Liu1, Devesh Shah2, Alireza Rahimpour2, Joyce Farrell1, and Brian Wandell1; 1Stanford University and 2Ford Motor Company (United States) [view abstract]

 

VDA-403
Visualizing and monitoring the process of injection molding, Christian A. Steinparz1, Thomas Mitterlehner2, Bernhard Praher2, Klaus Straka1,2, Holger Stitz1,3, and Marc Streit1,3; 1Johannes Kepler University, 2Moldsonics GmbH, and 3datavisyn GmbH (Austria) [view abstract]

 

COIMG-155
Commissioning the James Webb Space Telescope, Joseph M. Howard, NASA Goddard Space Flight Center (United States) [view abstract]

 

HVEI-223
Critical flicker frequency (CFF) at high luminance levels, Alexandre Chapiro1, Nathan Matsuda1, Maliha Ashraf2, and Rafal Mantiuk3; 1Meta (United States), 2University of Liverpool (United Kingdom), and 3University of Cambridge (United Kingdom) [view abstract]

 

HPCI-228
Physics guided machine learning for image-based material decomposition of tissues from simulated breast models with calcifications, Muralikrishnan Gopalakrishnan Meena1, Amir K. Ziabari1, Singanallur Venkatakrishnan1, Isaac R. Lyngaas1, Matthew R. Norman1, Balint Joo1, Thomas L. Beck1, Charles A. Bouman2, Anuj Kapadia1, and Xiao Wang1; 1Oak Ridge National Laboratory and 2Purdue University (United States) [view abstract]

 

3DIA-104
Layered view synthesis for general images, Loïc Dehan, Wiebe Van Ranst, and Patrick Vandewalle, Katholieke University Leuven (Belgium) [view abstract]

 

ISS-329
A self-powered asynchronous image sensor with independent in-pixel harvesting and sensing operations, Ruben Gomez-Merchan, Juan Antonio Leñero-Bardallo, and Ángel Rodríguez-Vázquez, University of Seville (Spain) [view abstract]

 

COLOR-184
Color blindness and modern board games, Alessandro Rizzi1 and Matteo Sassi2; 1Università degli Studi di Milano and 2consultant (Italy) [view abstract]

 


5:00 – 6:15 PM EI 2023 All-Conference Welcome Reception (in the Cyril Magnin Foyer)

Tuesday 17 January 2023

Sensor Design I (T1)

Session Chairs: Jon McElvain, Dolby Laboratories (United States) and Min-Woong Seo, Samsung Electronics (Republic of Korea)
9:05 – 10:10 AM
Powell I/II

9:05
Conference Welcome

9:10ISS-328
Simulation and design of a burst mode 20Mfps global shutter high conversion gain CMOS image sensor in a standard 180nm CMOS image sensor process using sequential transfer gates, Xin Yue and Eric R. Fossum, Dartmouth College (United States) [view abstract]

 

9:30ISS-329
A self-powered asynchronous image sensor with independent in-pixel harvesting and sensing operations, Ruben Gomez-Merchan, Juan Antonio Leñero-Bardallo, and Ángel Rodríguez-Vázquez, University of Seville (Spain) [view abstract]

 

9:50ISS-330
Highly sensitive mutual-capacitive fingerprint sensor with reference electrode, Junghoon Yang, Sarawut Siracosit, and Sang-Hee Ko Park, Korea Advanced Institute of Science and Technology (Republic of Korea) [view abstract]

 



10:00 AM – 7:30 PM Industry Exhibition - Tuesday (in the Cyril Magnin Foyer)

10:20 – 10:50 AM Coffee Break

KEYNOTE: Innovative Imaging Systems (T2)

Session Chairs: Francisco Imai, Apple Inc. (United States) and Kevin Matherson, Microsoft (United States)
10:50 AM – 12:30 PM
Powell I/II

10:50ISS-331
KEYNOTE: Metaphotonic routers for solid-state imaging: Making every photon count, Peter B. Catrysse, Stanford University (United States) [view abstract]

Dr. Peter B. Catrysse is a Senior Research Scientist in the E. L. Ginzton Laboratory (Stanford University). He holds a PhD and an MSc in Electrical Engineering from Stanford University. With his doctoral research, he pioneered the integration of subwavelength metal optics for color filtering in standard deep-submicron CMOS technology. His recent work focuses on metaphotonics at the interface between fundamental physics and imaging applications. Dr. Catrysse has published more than 120 peer-reviewed papers, presented over 40 invited talks, and has been awarded 8 patents. He was named one of “50 Tech” pioneers by the Belgian Financial Times (2017) and is featured in the top 1% leading Engineering and Technology Scientists on the academic portal Research (2022). Dr. Catrysse is a Fellow of the Optical Society (Optica), a Fellow of the SPIE, a Senior Member of the IEEE, and a Hoover Foundation Brussels Fellow of the BAEF.

 

11:30ISS-332
DiffuserCam: Multi-dimensional lensless imaging (Invited), Laura Waller, University of California, Berkeley (United States) [view abstract]

 

11:50ISS-333
Wide-viewing-zone light-field capturing using Turtleback convex reflector (JIST-first), Hiroaki Yano and Tomohiro Yendo, Nagaoka University of Technology (Japan) [view abstract]

 

12:10ISS-334
Digital camera obscuras, Henry G. Dietz, University of Kentucky (United States) [view abstract]

 




12:30 – 2:00 PM Lunch

Tuesday 17 January PLENARY: Embedded Gain Maps for Adaptive Display of High Dynamic Range Images

Session Chair: Robin Jenkin, NVIDIA Corporation (United States)
2:00 PM – 3:00 PM
Cyril Magnin I/II/III

Images optimized for High Dynamic Range (HDR) displays have brighter highlights and more detailed shadows, resulting in an increased sense of realism and greater impact. However, a major issue with HDR content is the lack of consistency in appearance across different devices and viewing environments. There are several reasons, including varying capabilities of HDR displays and the different tone mapping methods implemented across software and platforms. Consequently, HDR content authors can neither control nor predict how their images will appear in other apps.

We present a flexible system that provides consistent and adaptive display of HDR images. Conceptually, the method combines both SDR and HDR renditions within a single image and interpolates between the two dynamically at display time. We compute a Gain Map that represents the difference between the two renditions. In the file, we store a Base rendition (either SDR or HDR), the Gain Map, and some associated metadata. At display time, we combine the Base image with a scaled version of the Gain Map, where the scale factor depends on the image metadata, the HDR capacity of the display, and the viewing environment.


Eric Chan, Fellow, Adobe Inc. (United States)

 

Eric Chan is a Fellow at Adobe, where he develops software for editing photographs. Current projects include Photoshop, Lightroom, Camera Raw, and Digital Negative (DNG). When not writing software, Chan enjoys spending time at his other keyboard, the piano. He is an enthusiastic nature photographer and often combines his photo activities with travel and hiking.


Paul M. Hubel, director of Image Quality in Software Engineering, Apple Inc. (United States)

 

Paul M. Hubel is director of Image Quality in Software Engineering at Apple. He has worked on computational photography and image quality of photographic systems for many years on all aspects of the imaging chain, particularly for iPhone. He trained in optical engineering at University of Rochester, Oxford University, and MIT, and has more than 50 patents on color imaging and camera technology. Hubel is active on the ISO-TC42 committee Digital Photography, where this work is under discussion, and is currently a VP on the IS&T Board. Outside work he enjoys photography, travel, cycling, coffee roasting, and plays trumpet in several bay area ensembles.


3:00 – 3:30 PM Coffee Break

Image Processing (T3)

Session Chairs: Jon McElvain, Dolby Laboratories (United States) and Nitin Sampat, Edmund Optics, Inc (United States)
3:30 – 4:50 PM
Powell I/II

3:30ISS-335
Panoramic Photoacoustic Computed Tomography (PACT): From small-animal wholebody imaging to human breast cancer diagnosis (Invited), Lei Li, Rice University (United States) [view abstract]

 

3:50ISS-336
Array camera image fusion using physics-aware transformers (JIST-first), Qian Huang, Minghao Hu, and David J. Brady, The University of Arizona (United States) [view abstract]

 

4:10ISS-337
Self-supervised intensity-event stereo matching (JIST-first), Jinjin Gu1, Jinan Zhou2, Ringo S. Chu3, Yan Chen4, Jiawei Zhang4, Xuanye Cheng4, Song Zhang4, and Jimmy S. Ren3,5,6; 1The University of Sydney (Australia), 2The Chinese University of Hong Kong (Hong Kong), 3Sensetime Research HK (Hong Kong), 4SenseTime Research (China), 5Qing Yuan Research Institute (Hong Kong), and 6Shanghai Jiao Tong University (China) [view abstract]

 

4:30ISS-340
Improvement of a facial recognition system based on one shot camera, Médégnonmi E. Houssou1,2, Amadou Tidjani Sanda Mahama1,2, Pierre Gouton1, and Guy Degla2; 1University of Burgundy (France) and 2University of Abomey-Calavi (Benin) [view abstract]

 



5:30 – 7:00 PM EI 2023 Symposium Demonstration Session (in the Cyril Magnin Foyer)

Wednesday 18 January 2023

KEYNOTE: Processing at the Edge (W1)

Session Chairs: Stanley Chan, Purdue University (United States) and Boyd Fowler, OmniVision Technologies (United States)
8:45 – 10:20 AM
Market Street

This session is jointly sponsored by: Computational Imaging XXI, Imaging Sensors and Systems 2023, and the International Image Sensor Society (IISS).


8:45
COIMG/ISS Joint Sessions Welcome

8:50COIMG-177
KEYNOTE: Deep optics: Learning cameras and optical computing systems, Gordon Wetzstein, Stanford University (United States) [view abstract]

 

Gordon Wetzstein is an Associate Professor of Electrical Engineering and, by courtesy, of Computer Science at Stanford University. He is the leader of the Stanford Computational Imaging Lab and a faculty co-director of the Stanford Center for Image Systems Engineering. At the intersection of computer graphics and vision, artificial intelligence, computational optics, and applied vision science, Prof. Wetzstein's research has a wide range of applications in next-generation imaging, wearable computing, and neural rendering systems. Prof. Wetzstein is a Fellow of Optica and the recipient of numerous awards, including an NSF CAREER Award, an Alfred P. Sloan Fellowship, an ACM SIGGRAPH Significant New Researcher Award, a Presidential Early Career Award for Scientists and Engineers (PECASE), an SPIE Early Career Achievement Award, an Electronic Imaging Scientist of the Year Award, an Alain Fournier Ph.D. Dissertation Award as well as many Best Paper and Demo Awards.

 

9:40COIMG-178
Computational photography on a smartphone, Michael Polley, Samsung Research America (United States) [view abstract]

 

10:00COIMG-179
Analog in-memory computing with multilevel RRAM for edge electronic imaging application, Glenn Ge, Teramem Inc. (United States) [view abstract]

 


10:00 AM – 3:30 PM Industry Exhibition - Wednesday (in the Cyril Magnin Foyer)

10:20 – 10:50 AM Coffee Break



Processing at the Edge (W2.1)

Session Chairs: Stanley Chan, Purdue University (United States) and Boyd Fowler, OmniVision Technologies (United States)
10:50 – 11:50 AM
Market Street

This session is jointly sponsored by: Computational Imaging XXI, Imaging Sensors and Systems 2023, and the International Image Sensor Society (IISS).


10:50COIMG-180
Processing of real time, bursty and high compute iToF data on the edge (Invited), Cyrus Bamji, Microsoft Corporation (United States) [view abstract]

 

11:10COIMG-181
A distributed on-sensor compute system in AR/VR devices and neural architecture search (NAS) framework for optimal workload distribution (Invited), Chiao Liu1, Xin Dong2, Ziyun Li1, Barbara De Salvo3, and H. T. Kung2; 1Reality Labs, 2Harvard University, and 3Meta (United States) [view abstract]

 

11:30ISS-182
A 2.2um three-wafer stacked back side illuminated voltage domain global shutter CMOS image sensor, Shimpei Fukuoka, OmniVision (Japan) [view abstract]

 



12:30 – 2:00 PM Lunch

Wednesday 18 January PLENARY: Bringing Vision Science to Electronic Imaging: The Pyramid of Visibility

Session Chair: Andreas Savakis, Rochester Institute of Technology (United States)
2:00 PM – 3:00 PM
Cyril Magnin I/II/III

Electronic imaging depends fundamentally on the capabilities and limitations of human vision. The challenge for the vision scientist is to describe these limitations to the engineer in a comprehensive, computable, and elegant formulation. Primary among these limitations are visibility of variations in light intensity over space and time, of variations in color over space and time, and of all of these patterns with position in the visual field. Lastly, we must describe how all these sensitivities vary with adapting light level. We have recently developed a structural description of human visual sensitivity that we call the Pyramid of Visibility, that accomplishes this synthesis. This talk shows how this structure accommodates all the dimensions described above, and how it can be used to solve a wide variety of problems in display engineering.


Andrew B. Watson, chief vision scientist, Apple Inc. (United States)

 

Andrew Watson is Chief Vision Scientist at Apple, where he leads the application of vision science to technologies, applications, and displays. His research focuses on computational models of early vision. He is the author of more than 100 scientific papers and 8 patents. He has 21,180 citations and an h-index of 63. Watson founded the Journal of Vision, and served as editor-in-chief 2001-2013 and 2018-2022. Watson has received numerous awards including the Presidential Rank Award from the President of the United States.


3:00 – 3:30 PM Coffee Break

KEYNOTE: Sensor Design II (W3)

Session Chairs: Min-Woong Seo, Samsung Electronics (Republic of Korea) and Hari Tagat, Casix (United States)
3:30 – 5:10 PM
Powell I/II

3:30ISS-341
KEYNOTE: Event camera noise and denoising, Tobi Delbrück, Institute of Neuroinformatics, University of Zurich and ETH Zurich (Switzerland) [view abstract]

Tobi Delbrück (IEEE M'99, SM'06, F'13) received his BSc in physics from University of California in 1986 and his PhD from Caltech in 1993 as the first student with the Computation and Neural Systems program with PhD supervisor Carver Mead. He is an ETH Honorary Professor of Physics and Electrical Engineering, and has been with the Institute of Neuroinformatics, University of Zurich and ETH Zurich since 1998. The Sensors Group that he co-directs together with Prof. Shih-Chii Liu works on a broad range of topics covering device physics to computer vision and control, with a theme of efficient neuromorphic processing in hardware. He co-organizes the Telluride Neuromorphic Engineering workshop and has organized live demonstration sessions at ISCAS, NeuIPS, and AICAS and two conference Sessions at ISCAS. Delbrück is past Chair of the IEEE CAS Sensory Systems Technical Committee. He worked on electronic imaging at Arithmos, Synaptics, National Semiconductor, and Foveon and has co-founded 3 companies (Inilabs, Insightness and Inivation). His papers have been awarded 13 IEEE awards and he was named a Fellow of the IEEE Circuits and Systems Society for his work on neuromorphic sensors and processing. He likes to read storybooks, play tennis, and sometimes tries card magic.

 

4:10ISS-342
Quantum efficiency of various miniaturized backside illuminated CMOS pixels under ultraviolet illumination, Nour Fassi1,2, Jean-Pierre Carrère1, Pierre Magnan2, Magali Estribeau2, and Vincent Goiffon2; 1STMicroelectronics and 2ISAE-SUPAERO (France) [view abstract]

 

4:30ISS-343
Color performance of 0.8 um CMOS image sensor with CMY color filters, An-Li Kuo, Pohsiang Wang, Hao-Wei Liu, William Tsai, Chia-Ning Hsu, Chien-Wen Lai, Yu C. Chang, Ching-Chiang Wu, and Ken Wu, VisEra Technologies (Taiwan) [view abstract]

 

4:50ISS-344
Reset noise reduction method in 3-T pixels, Kaitlin M. Anagnost, Xin Yue, and Eric R. Fossum, Dartmouth College (United States) [view abstract]

 




Imaging Sensors and Systems 2023 Interactive (Poster) Paper Session

5:30 – 7:00 PM
Cyril Magnin Foyer

The following works will be presented at the EI 2023 Symposium Interactive (Poster) Paper Session.


ISS-345
DevCAM: An open-source multi-camera development system for embedded vision, Meher Akhil Birlangi, Dominique E. Meyer, and Falko Kuester, University of California, San Diego (United States) [view abstract]

 

ISS-346
Development of DVS evaluation methods from user perspective, Raeyoung Kim, Jun-seok Kim, Junhyuk Park, Paul K.J. Park, Jaeha Park, Chunghwan Park, Inchun Lim, Seongwook Song, and Juhyun Ko, Samsung (Republic of Korea) [view abstract]

 

ISS-347
Implementation of EMVA 1288 Standard Release 4.0 for Characterization of Image Sensors, Megan E. Borek, Imatest, LLC (United States) [view abstract]

 

ISS-348
On quantization of convolutional neural networks for image signal processor, Youngil Seo, Dongpan Lim, Jeongguk Lee, and Seongwook Song, Samsung Electronics (Republic of Korea) [view abstract]

 



5:30 – 7:00 PM EI 2023 Symposium Interactive (Poster) Paper Session (in the Cyril Magnin Foyer)

5:30 – 7:00 PM EI 2023 Meet the Future: A Showcase of Student and Young Professionals Research (in the Cyril Magnin Foyer)

No content found

No content found