IMPORTANT DATES
Dates currently being confirmed; check back.
 

2022
Call for Papers Announced 2 May
Journal-first (JIST/JPI) Submissions

∙ Submission site Opens 2 May 
∙ Journal-first (JIST/JPI) Submissions Due 1 Aug
∙ Final Journal-first manuscripts due 28 Oct
Conference Papers Submissions
∙ Abstract Submission Opens 1 June
∙ Priority Decision Submission Ends 15 July
∙ Extended Submission Ends  19 Sept
∙ FastTrack Conference Proceedings Manuscripts Due 25 Dec 
∙ All Outstanding Proceedings Manuscripts Due
 6 Feb 2023
Registration Opens 1 Dec
Demonstration Applications Due 19 Dec
Early Registration Ends 18 Dec


2023
Hotel Reservation Deadline 6 Jan
Symposium begins
15 Jan


Sponsor


Best Paper Sponsor




No content found

Image Quality and System Performance XX

Monday 16 January 2023

20th Anniversary: A Tour of Quality Assessment and System Performance (M1)

Session Chair: Mohamed Chaker Larabi, Université de Poitiers (France)
8:45 – 10:20 AM
Cyril Magnin III

8:45
Conference Welcome, Chaker Larabi

8:50
Twenty years in twenty minutes, Peter Burns

9:10IQSP-450
Subjective image quality: Beauty and the Beast in human vision (Invited), Göte S. Nyman, University of Helsinki (Finland) [view abstract]

 

9:45IQSP-451
Displays and lighting: What do they have in common? (Invited), Ingrid Heynderickx, Eindhoven University of Technology (the Netherlands) [view abstract]

 



10:20 – 10:40 AM Coffee Break

20th Anniversary: A Tour of Quality Assessment and System Performance (M2)

Session Chairs: Mohamed Chaker Larabi, Université de Poitiers (France) and Jonathan Phillips, Imatest, LLC (United States)
10:40 AM – 12:30 PM
Cyril Magnin III

10:40IQSP-452
The revolutionary advancement of camera phone image quality (Invited), Jonathan Phillips, Imatest, LLC (United States) [view abstract]

 

11:10IQSP-453
23 years of ISO 12233 resolution measurement (Invited), Dietmar Wueller, Image Engineering GmbH & Co. KG (Germany) [view abstract]

 

11:20IQSP-456
Limits of MTF in practice (Invited), Alexander Braun, Düsseldorf University of Applied Sciences (Germany) [view abstract]

 

11:30IQSP-454
Measuring camera information capacity with slanted-edges (Invited), Norman Koren, Imatest LLC (United States) [view abstract]

 

11:40IQSP-455
From BxU to integrated information capacity, a brief history of MTF based KPIs at DXOMARK (Invited), Laurent Chanas, DxOMark Image Labs (France) [view abstract]

 

11:50
Panel Discussion




12:30 – 2:00 PM Lunch

Monday 16 January PLENARY: Neural Operators for Solving PDEs

Session Chair: Robin Jenkin, NVIDIA Corporation (United States)
2:00 PM – 3:00 PM
Cyril Magnin I/II/III

Deep learning surrogate models have shown promise in modeling complex physical phenomena such as fluid flows, molecular dynamics, and material properties. However, standard neural networks assume finite-dimensional inputs and outputs, and hence, cannot withstand a change in resolution or discretization between training and testing. We introduce Fourier neural operators that can learn operators, which are mappings between infinite dimensional spaces. They are independent of the resolution or grid of training data and allow for zero-shot generalization to higher resolution evaluations. When applied to weather forecasting, neural operators capture fine-scale phenomena and have similar skill as gold-standard numerical weather models for predictions up to a week or longer, while being 4-5 orders of magnitude faster.


Anima Anandkumar, Bren professor, California Institute of Technology, and senior director of AI Research, NVIDIA Corporation (United States)

 

Anima Anandkumar is a Bren Professor at Caltech and Senior Director of AI Research at NVIDIA. She is passionate about designing principled AI algorithms and applying them to interdisciplinary domains. She has received several honors such as the IEEE fellowship, Alfred. P. Sloan Fellowship, NSF Career Award, and Faculty Fellowships from Microsoft, Google, Facebook, and Adobe. She is part of the World Economic Forum's Expert Network. Anandkumar received her BTech from Indian Institute of Technology Madras, her PhD from Cornell University, and did her postdoctoral research at MIT and assistant professorship at University of California Irvine.


3:00 – 3:30 PM Coffee Break

Subjective Quality Assessment (M3)

Session Chair: Sophie Triantaphillidou, University of Westminster (United Kingdom)
3:30 – 4:50 PM
Cyril Magnin III

3:30IQSP-301
Image demosaicing: Subjective analysis and evaluation of image quality metrics, Tawsin Uddin Ahmed, Seyed Ali Amirshahi, and Marius Pedersen, Norwegian University of Science and Technology (Norway) [view abstract]

 

3:50IQSP-302
Age-specific perceptual image quality assessment, Yinan Wang1, Andrei Chubarau1, Tara Akhavan2, Hyunjin Yoo2, and James Clark1; 1McGill University and 2Forvia (Canada) [view abstract]

 

4:10IQSP-303
A method for evaluating camera auto-focusing performance using a transparent display device, Seungwan Jeon, Kichul Park, Sung-Su Kim, and Yitae Kim, Samsung Electronics (Republic of Korea) [view abstract]

 



EI 2023 Highlights Session

Session Chair: Robin Jenkin, NVIDIA Corporation (United States)
3:30 – 5:00 PM
Cyril Magnin II

Join us for a session that celebrates the breadth of what EI has to offer with short papers selected from EI conferences.

NOTE: The EI-wide "EI 2023 Highlights" session is concurrent with Monday afternoon COIMG, COLOR, IMAGE, and IQSP conference sessions.

 

IQSP-309
Evaluation of image quality metrics designed for DRI tasks with automotive cameras, Valentine Klein, Yiqi LI, Claudio Greco, Laurent Chanas, and Frédéric Guichard, DXOMARK (France) [view abstract]

 

SD&A-224
Human performance using stereo 3D in a helmet mounted display and association with individual stereo acuity, Bonnie Posselt, RAF Centre of Aviation Medicine (United Kingdom) [view abstract]

 

IMAGE-281
Smartphone-enabled point-of-care blood hemoglobin testing with color accuracy-assisted spectral learning, Sang Mok Park1, Yuhyun Ji1, Semin Kwon1, Andrew R. O’Brien2, Ying Wang2, and Young L. Kim1; 1Purdue University and 2Indiana University School of Medicine (United States) [view abstract]

 

AVM-118
Designing scenes to quantify the performance of automotive perception systems, Zhenyi Liu1, Devesh Shah2, Alireza Rahimpour2, Joyce Farrell1, and Brian Wandell1; 1Stanford University and 2Ford Motor Company (United States) [view abstract]

 

VDA-403
Visualizing and monitoring the process of injection molding, Christian A. Steinparz1, Thomas Mitterlehner2, Bernhard Praher2, Klaus Straka1,2, Holger Stitz1,3, and Marc Streit1,3; 1Johannes Kepler University, 2Moldsonics GmbH, and 3datavisyn GmbH (Austria) [view abstract]

 

COIMG-155
Commissioning the James Webb Space Telescope, Joseph M. Howard, NASA Goddard Space Flight Center (United States) [view abstract]

 

HVEI-223
Critical flicker frequency (CFF) at high luminance levels, Alexandre Chapiro1, Nathan Matsuda1, Maliha Ashraf2, and Rafal Mantiuk3; 1Meta (United States), 2University of Liverpool (United Kingdom), and 3University of Cambridge (United Kingdom) [view abstract]

 

HPCI-228
Physics guided machine learning for image-based material decomposition of tissues from simulated breast models with calcifications, Muralikrishnan Gopalakrishnan Meena1, Amir K. Ziabari1, Singanallur Venkatakrishnan1, Isaac R. Lyngaas1, Matthew R. Norman1, Balint Joo1, Thomas L. Beck1, Charles A. Bouman2, Anuj Kapadia1, and Xiao Wang1; 1Oak Ridge National Laboratory and 2Purdue University (United States) [view abstract]

 

3DIA-104
Layered view synthesis for general images, Loïc Dehan, Wiebe Van Ranst, and Patrick Vandewalle, Katholieke University Leuven (Belgium) [view abstract]

 

ISS-329
A self-powered asynchronous image sensor with independent in-pixel harvesting and sensing operations, Ruben Gomez-Merchan, Juan Antonio Leñero-Bardallo, and Ángel Rodríguez-Vázquez, University of Seville (Spain) [view abstract]

 

COLOR-184
Color blindness and modern board games, Alessandro Rizzi1 and Matteo Sassi2; 1Università degli Studi di Milano and 2consultant (Italy) [view abstract]

 


5:00 – 6:15 PM EI 2023 All-Conference Welcome Reception (in the Cyril Magnin Foyer)

Tuesday 17 January 2023

KEYNOTE: Perceptual Video Quality 1 (T1)

Session Chairs: Lukáš Krasula, Netflix, Inc. (United States) and Mohamed Chaker Larabi, Université de Poitiers (France)
9:05 – 10:10 AM
Cyril Magnin III

This session is jointly sponsored by: Human Vision and Electronic Imaging 2023, and Image Quality and System Performance XX.



Joint Conference Welcome

HVEI-258
KEYNOTE: Bringing joy to Netflix members through perceptual encoding optimization, Anne Aaron, Netflix, Inc. (United States) [view abstract]

 

As Director of Encoding Technologies, Anne Aaron leads the team responsible for media processing and encoding at Netflix. Her team works on video, audio, images and timed-text, from analysis to processing, encoding, packaging and DRM. On the streaming side, they strive to deliver a compelling viewing experience for millions of Netflix members worldwide, no matter where, how and what they watch. For the Netflix studio, they build media technologies that can improve content production. In her previous role at Netflix, Aaron led the Video Algorithms team. As a team, they researched and deployed innovation in the video encoding space (per-title encoding, video quality assessment and perceptual metrics, shot-based encoding, HDR, next-generation codecs) that benefited Netflix members as well as impacted the rest of the industry. Recent recognitions include: Some recent recognitions: SMPTE 2019 Workflow Systems Medal, Forbes' 2018 America's top women in Tech, Business Insider's 2017 Most powerful female engineers in US tech in 2017.

 


10:00 AM – 7:30 PM Industry Exhibition - Tuesday (in the Cyril Magnin Foyer)

10:20 – 10:50 AM Coffee Break



Perceptual Video Quality 2 (T2)

Session Chairs: Lukáš Krasula, Netflix, Inc. (United States) and Mohamed Chaker Larabi, Université de Poitiers (France)
10:50 AM – 12:30 PM
Cyril Magnin III

This session is jointly sponsored by: Human Vision and Electronic Imaging 2023, and Image Quality and System Performance XX.


10:50HVEI-259
Video quality of video professionals for Video Assisted Referee (VAR), Kjell Brunnström1,2, Anders Djupsjöbacka1, Johsan Billingham3, Katharina Wistel3, Börje Andrén1, Oskars Ozolins1,4, and Nicolas Evans3; 1RISE Research Institutes of Sweden AB (Sweden), 2Mid Sweden University (Sweden), 3Fédération Internationale de Football Association (FIFA) (Switzerland), and 4KTH (Royal Institute of Technology) (Sweden) [view abstract]

 

11:10HVEI-260
User perception for dynamic video resolution change using VVC, Sachin G. Deshpande and Philip Cowan, Sharp (United States) [view abstract]

 

11:30IQSP-261
Proposing more ecologically-valid experiment protocol using YouTube platform, Gabriela Wielgus, Lucjan Janowski, Kamil Koniuch, Mikolaj Leszczuk, and Rafal Figlus, AGH University of Science and Technology (Poland) [view abstract]

 

11:50IQSP-262
Evaluation of motion blur image quality in video frame interpolation, Hai Dinh, Fangwen Tu, Qinyi Wang, Brett Frymire, and Bo Mu, Omnivision Technology (United States) [view abstract]

 

12:10IQSP-263
Subjective video quality for 4K HDR-WCG content using a browser-based approach for “at-home” testing, Lukáš Krasula1, Anustup Choudhury2, Scott Daly2, Zhi Li1, Robin Atkins2, Ludovic Malfait2, and Aditya Mavlankar1; 1Netflix, Inc. and 2Dolby Laboratories, Inc. (United States) [view abstract]

 



12:30 – 2:00 PM Lunch

Tuesday 17 January PLENARY: Embedded Gain Maps for Adaptive Display of High Dynamic Range Images

Session Chair: Robin Jenkin, NVIDIA Corporation (United States)
2:00 PM – 3:00 PM
Cyril Magnin I/II/III

Images optimized for High Dynamic Range (HDR) displays have brighter highlights and more detailed shadows, resulting in an increased sense of realism and greater impact. However, a major issue with HDR content is the lack of consistency in appearance across different devices and viewing environments. There are several reasons, including varying capabilities of HDR displays and the different tone mapping methods implemented across software and platforms. Consequently, HDR content authors can neither control nor predict how their images will appear in other apps.

We present a flexible system that provides consistent and adaptive display of HDR images. Conceptually, the method combines both SDR and HDR renditions within a single image and interpolates between the two dynamically at display time. We compute a Gain Map that represents the difference between the two renditions. In the file, we store a Base rendition (either SDR or HDR), the Gain Map, and some associated metadata. At display time, we combine the Base image with a scaled version of the Gain Map, where the scale factor depends on the image metadata, the HDR capacity of the display, and the viewing environment.


Eric Chan, Fellow, Adobe Inc. (United States)

 

Eric Chan is a Fellow at Adobe, where he develops software for editing photographs. Current projects include Photoshop, Lightroom, Camera Raw, and Digital Negative (DNG). When not writing software, Chan enjoys spending time at his other keyboard, the piano. He is an enthusiastic nature photographer and often combines his photo activities with travel and hiking.


Paul M. Hubel, director of Image Quality in Software Engineering, Apple Inc. (United States)

 

Paul M. Hubel is director of Image Quality in Software Engineering at Apple. He has worked on computational photography and image quality of photographic systems for many years on all aspects of the imaging chain, particularly for iPhone. He trained in optical engineering at University of Rochester, Oxford University, and MIT, and has more than 50 patents on color imaging and camera technology. Hubel is active on the ISO-TC42 committee Digital Photography, where this work is under discussion, and is currently a VP on the IS&T Board. Outside work he enjoys photography, travel, cycling, coffee roasting, and plays trumpet in several bay area ensembles.


3:00 – 3:30 PM Coffee Break

Objective Quality Assessment (T3)

Session Chair: Peter Burns, Rochester Institure of Tech. (United States)
3:30 – 5:30 PM
Cyril Magnin III

3:30IQSP-305
Another look at SSIM image quality metric, Yuriy Reznik, Brightcove, Inc. (United States) [view abstract]

 

3:50IQSP-306
What are we looking at? An investigation on the use of deep learning models for image quality assessment, Ha Thu Nguyen and Seyed Ali Amirshahi, Norwegian University of Science and Technology (Norway) [view abstract]

 

4:10IQSP-307
A framework for the metrification of input image quality in deep networks, Alexandra Psarrou and Sophie Triantaphillidou, University of Westminster (United Kingdom) [view abstract]

 

4:30IQSP-308
Investigating pretrained self-supervised vision transformers for reference-based quality assessment., Kanjar De, Lulea University of Technology (Sweden) [view abstract]

 

4:50IQSP-309
Evaluation of image quality metrics designed for DRI tasks with automotive cameras, Valentine Klein, Theophanis Eleftheriou, Yiqi LI, Emilie Baudin, Claudio Greco, Laurent Chanas, and Frédéric Guichard, DXOMARK (France) [view abstract]

 

5:10IQSP-310
Towards image-computable visual text quality metric with deep neural network, Ling-Qi Zhang1,2, Minjung Kim1, James Hillis1, and Trisha Lian1; 1Meta Reality Labs and 2University of Pennsylvania (United States) [view abstract]

 



5:30 – 7:00 PM EI 2023 Symposium Demonstration Session (in the Cyril Magnin Foyer)

Wednesday 18 January 2023

System Performance (W1)

Session Chair: Jonathan Phillips, Imatest, LLC (United States)
8:50 – 10:10 AM
Cyril Magnin III

8:50IQSP-311
A tool for deriving camera spatial frequency response from natural scenes (NS-SFR), Oliver van Zwanenberg1, Sophie Triantaphillidou1, and Robin B. Jenkin1,2; 1University of Westminster (United Kingdom) and 2NVIDIA Corporation (United States) [view abstract]

 

9:10IQSP-312
Influence of the light source on the image sensor characterization according to EMVA 1288, Ganesh D. Kubina, Max Gäde, and Uwe Artmann, Image Engineering GmbH & Co KG (Germany) [view abstract]

 

9:30IQSP-313
Managing deviant data in spatial frequency response (SFR) measurement by outlier rejection, Peter Burns1 and Don Williams2; 1Burns Digital Imaging and 2Image Science Associates (United States) [view abstract]

 

9:50IQSP-314
Optimization of ISP parameters for low light conditions using a non-linear reference based approach, Shubham Ravindra Alai1, Radhesh Bhat1, and Ajay Basarur2; 1PathPartner Technology - Member of KPIT Group (India) and 2presenter only (United States) [view abstract]

 



10:00 AM – 3:30 PM Industry Exhibition - Wednesday (in the Cyril Magnin Foyer)

10:20 – 10:50 AM Coffee Break

Mobile and Camera Quality Assessment (W2)

Session Chair: Elaine Jin, Rivian Automotive, Inc. (United States)
10:50 AM – 12:30 PM
Cyril Magnin III

10:50IQSP-315
Image quality performance of CMOS image sensor equipped with Nano Prism, Sungho Cha, Samsung Electronics (Republic of Korea) [view abstract]

 

11:10IQSP-316
Noise quality estimation on portraits in realistic controlled scenarios, Nicolas Chahine1,2, Samuel S. Santos3, Sofiene Lahouar1, Ana-Stefania Calarasanu1, Sira Ferradans1, Benoit Pochon1, and Frédéric Guichard1; 1DXOMARK Image Labs, 2INRIA, and 3Parrot (France) [view abstract]

 

11:30IQSP-317
VCX – Version 2023 – The latest transparent and objective mobile phone test scheme, Uwe Artmann1 and Anthony L. Orchard2; 1Image Engineering GmbH & Co KG (Germany) and 2Intel Corporation (United States) [view abstract]

 

11:50IQSP-318
VCX – A transparent and objective test scheme for webcams, Uwe Artmann1 and Anthony L. Orchard2; 1Image Engineering GmbH & Co KG (Germany) and 2Intel Corporation (United States) [view abstract]

 

12:10IQSP-319
Improvement of the flare evaluation and applications in NIR, Elodie Souksava, Emilie Baudin, Claudio Greco, Hoang-Phi Nguyen, Laurent Chanas, and Frédéric Guichard, DxOMark Image Labs (France) [view abstract]

 



12:30 – 2:00 PM Lunch

Wednesday 18 January PLENARY: Bringing Vision Science to Electronic Imaging: The Pyramid of Visibility

Session Chair: Andreas Savakis, Rochester Institute of Technology (United States)
2:00 PM – 3:00 PM
Cyril Magnin I/II/III

Electronic imaging depends fundamentally on the capabilities and limitations of human vision. The challenge for the vision scientist is to describe these limitations to the engineer in a comprehensive, computable, and elegant formulation. Primary among these limitations are visibility of variations in light intensity over space and time, of variations in color over space and time, and of all of these patterns with position in the visual field. Lastly, we must describe how all these sensitivities vary with adapting light level. We have recently developed a structural description of human visual sensitivity that we call the Pyramid of Visibility, that accomplishes this synthesis. This talk shows how this structure accommodates all the dimensions described above, and how it can be used to solve a wide variety of problems in display engineering.


Andrew B. Watson, chief vision scientist, Apple Inc. (United States)

 

Andrew Watson is Chief Vision Scientist at Apple, where he leads the application of vision science to technologies, applications, and displays. His research focuses on computational models of early vision. He is the author of more than 100 scientific papers and 8 patents. He has 21,180 citations and an h-index of 63. Watson founded the Journal of Vision, and served as editor-in-chief 2001-2013 and 2018-2022. Watson has received numerous awards including the Presidential Rank Award from the President of the United States.


3:00 – 3:30 PM Coffee Break

5:30 – 7:00 PM EI 2023 Symposium Interactive (Poster) Paper Session (in the Cyril Magnin Foyer)

5:30 – 7:00 PM EI 2023 Meet the Future: A Showcase of Student and Young Professionals Research (in the Cyril Magnin Foyer)

No content found

No content found