Sponsors








        13 - 17  January, 2019 • Burlingame, California USA

Monday January 14, 2019

Automotive Image Quality

Session Chairs: Patrick Denny, Valeo Vision Systems (Ireland); Stuart Perry, University of Technology Sydney (Australia); and Peter van Beek, Intel Corporation (United States)
8:50 – 10:10 AM
Grand Peninsula Ballroom D

This session is jointly sponsored by: Autonomous Vehicles and Machines 2019, and Image Quality and System Performance XVI.


8:50AVM-026
Updates on the progress of IEEE P2020 Automotive Imaging Standards Working Group, Robin Jenkin, NVIDIA Corporation (United States)

9:10AVM-027
Signal detection theory and automotive imaging, Paul Kane, ON Semiconductor (United States)

9:30AVM-029
Digital camera characterisation for autonomous vehicles applications, Paola Iacomussi and Giuseppe Rossi, INRIM (Italy)

9:50AVM-030
Contrast detection probability - Implementation and use cases, Uwe Artmann1, Marc Geese2, and Max Gäde1; 1Image Engineering GmbH & Co KG and 2Robert Bosch GmbH (Germany)



10:10 – 11:00 AM Coffee Break

12:30 – 2:00 PM Lunch

Monday Plenary

2:00 – 3:00 PM
Grand Peninsula Ballroom D

Autonomous Driving Technology and the OrCam MyEye, Amnon Shashua, President & CEO, Mobileye, an Intel Company, and Senior Vice President of Intel Corporation (United States)

The field of transportation is undergoing a seismic change with the coming introduction of autonomous driving. The technologies required to enable computer driven cars involves the latest cutting edge artificial intelligence algorithms along three major thrusts: Sensing, Planning and Mapping. Prof. Shashua will describe the challenges and the kind of computer vision and machine learning algorithms involved, but will do that through the perspective of Mobileye's activity in this domain. He will then describe how OrCam leverages computer vision, situation awareness and language processing to enable Blind and Visually impaired to interact with the world through a miniature wearable device.

Prof. Amnon Shashua holds the Sachs chair in computer science at the Hebrew University of Jerusalem. His field of expertise is computer vision and machine learning. For his academic achievements he received the MARR prize Honorable Mention in 2001, the Kaye innovation award in 2004, and the Landau award in exact sciences in 2005.

In 1999 Prof. Shashua co-founded Mobileye, an Israeli company developing a system-on-chip and computer vision algorithms for a driving assistance system, providing a full range of active safety features using a single camera. Today, approximately 24 million cars rely on Mobileye technology to make their vehicles safer to drive. In August 2014, Mobileye claimed the title for largest Israeli IPO ever, by raising $1B at a market cap of $5.3B. In addition, Mobileye is developing autonomous driving technology with more than a dozen car manufacturers. The introduction of autonomous driving capabilities is of a transformative nature and has the potential of changing the way cars are built, driven and own in the future. In August 2017, Mobileye became an Intel company in the largest Israeli acquisition deal ever of $15.3B. Today, Prof. Shashua is the President and CEO of Mobileye and a Senior Vice President of Intel Corporation leading Intel's Autonomous Driving Group.

In 2010 Prof. Shashua co-founded OrCam which harnesses computer vision and artificial intelligence to assist people who are visually impaired or blind. The OrCam MyEye device is unique in its ability to provide visual aid to hundreds of millions of people, through a discreet wearable platform. Within its wide-ranging scope of capabilities, OrCam's device can read most texts (both indoors and outdoors) and learn to recognize thousands of new items and faces.


3:00 – 3:30 PM Coffee Break

Printing System Performance

Session Chair: Mylène Farias, University of Brasilia (Brazil)
3:30 – 4:50 PM
Grand Peninsula Ballroom E

3:30IQSP-300
Detection of streaks on printed pages, Runzhe Zhang1, Eric Maggard2, Renee Jessome2, Yousun Bang2, Minki Cho2, and Jan Allebach1; 1Purdue University and 2HP, Inc. (United States)

3:50IQSP-301
Segmentation-based detection of local defects on printed pages, Qiulin Chen1, Eric Maggard2, Renee Jessome2, Yousun Bang3, Minki Cho3, and Jan Allebach1; 1Purdue University (United States), 2HP, Inc. (United States), and 3HP-Korea (Republic of Korea)

4:10IQSP-302
Banding estimation for print quality, Wan-Eih Huang1, Eric Maggard2, Renee Jessome2, Yousun Bang2, Minki Cho2, and Jan Allebach1; 1Purdue University and 2HP, Inc. (United States)

4:30IQSP-303
Blockwise detection of local defects on printed pages, Xiaoyu Xiang1, Eric Maggard2, Renee Jessome2, Yousun Bang2, Minki Cho2, and Jan Allebach1; 1Purdue University and 2HP, Inc. (United States)



5:00 – 6:00 PM All-Conference Welcome Reception

Tuesday January 15, 2019

7:30 – 8:45 AM Women in Electronic Imaging Breakfast

Image Quality Modeling I

Session Chair: Stuart Perry, University of Technology Sydney (Australia)
8:50 – 9:30 AM
Grand Peninsula Ballroom E

8:50IQSP-304
A referenceless image quality assessment based on BSIF, CLBP, LPQ, and LCP texture descriptors, Pedro Garcia Freitas, Luisa Eira, Samuel Santos, and Mylène Farias, University of Brasilia (Brazil)

9:10IQSP-305
Compensating MTF measurements for chart quality limitations, Norman Koren, Imatest LLC (United States)



Image Quality Modeling II

Session Chair: Stuart Perry, University of Technology Sydney (Australia)
9:30 – 10:10 AM
Grand Peninsula Ballroom E

IQSP-306
KEYNOTE: Conscious of streaming (Quality), Alan Bovik, The University of Texas at Austin (United States)

Alan Bovik is the Cockrell Family Regents Endowed Chair Professor at The University of Texas at Austin. He has received many major international awards, including the 2019 IEEE Fourier Award, the 2017 Edwin H. Land Medal from IS&T and OSA, the 2015 Primetime Emmy Award for Outstanding Achievement in Engineering Development from the Academy of Television Arts and Sciences, and the ‘Society’ and ‘Sustained Impact’ Awards of the IEEE Signal Processing Society. His is a Fellow of the IEEE, OSA, and SPIE. His books include The Handbook of Image and Video Processing, Modern Image Quality Assessment, and The Essential Guides to Image and Video Processing. Al co-founded and was the longest-serving Editor-in-Chief of the IEEE Transactions on Image Processing and created the IEEE International Conference on Image Processing in Austin, Texas, in November, 1994.




10:00 AM – 7:30 PM Industry Exhibition

10:10 – 10:50 AM Coffee Break

Display Performance

Session Chair: Nicolas Bonnier, Apple Inc. (United States)
10:50 AM – 12:30 PM
Grand Peninsula Ballroom E

10:50IQSP-307
Combining quality metrics using machine learning for improved and robust HDR image quality assessment, Anustup Choudhury and Scott Daly, Dolby Laboratories, Inc. (United States)

11:10IQSP-308
Subjective evaluations on perceptual image brightness in high dynamic range television, Yoshitaka Ikeda and Yuichi Kusakabe, NHK (Japan Broadcasting Corporation) (Japan)

11:30IQSP-309
Image quality evaluation on an HDR OLED display, Dalin Tian, Lihao Xu, and Ming Ronnier Luo, Zhejiang University (China)

11:50IQSP-310
A comprehensive framework for visual quality assessment of light field tensor displays, Irene Viola1, Keita Takahashi2, Toshiaki Fujii2, and Touradj Ebrahimi1; 1École Polytechnique Fédérale de Lausanne (EPFL) (Switzerland) and 2Nagoya University (Japan)

12:10IQSP-311
Semantic label bias in subjective video quality evaluation: A standardization perspective, Mihai Mitrea1, Rania Bensaied1, and Patrick Le Callet2; 1Institut Mines-Telecom and 2Université de Nantes (France)



12:30 – 2:00 PM Lunch

Tuesday Plenary

2:00 – 3:00 PM
Grand Peninsula Ballroom D

The Quest for Vision Comfort: Head-Mounted Light Field Displays for Virtual and Augmented Reality, Hong Hua, Professor of Optical Sciences, University of Arizona (United States)

Hong Hua will discuss the high promises and the tremendous progress made recently toward the development of head-mounted displays (HMD) for both virtual and augmented reality displays, developing HMDs that offer uncompromised optical pathways to both digital and physical worlds without encumbrance and discomfort confronts many grand challenges, both from technological perspectives and human factors. She will particularly focus on the recent progress, challenges and opportunities for developing head-mounted light field displays (LF-HMD), which are capable of rendering true 3D synthetic scenes with proper focus cues to stimulate natural eye accommodation responses and address the well-known vergence-accommodation conflict in conventional stereoscopic displays.

Dr. Hong Hua is a Professor of Optical Sciences at the University of Arizona. With over 25 years of experience, Dr. Hua is widely recognized through academia and industry as an expert in wearable display technologies and optical imaging and engineering in general. Dr. Hua’s current research focuses on optical technologies enabling advanced 3D displays, especially head-mounted display technologies for virtual reality and augmented reality applications, and microscopic and endoscopic imaging systems for medicine. Dr. Hua has published over 200 technical papers and filed a total of 23 patent applications in her specialty fields, and delivered numerous keynote addresses and invited talks at major conferences and events worldwide. She is an SPIE Fellow and OSA senior member. She was a recipient of NSF Career Award in 2006 and honored as UA Researchers @ Lead Edge in 2010. Dr. Hua and her students shared a total of 8 “Best Paper” awards in various IEEE, SPIE and SID conferences. Dr. Hua received her Ph.D. degree in Optical Engineering from the Beijing Institute of Technology in China in 1999. Prior to joining the UA faculty in 2003, Dr. Hua was an Assistant Professor with the University of Hawaii at Manoa in 2003, was a Beckman Research Fellow at the Beckman Institute of University of Illinois at Urbana-Champaign between 1999 and 2002, and was a post-doc at the University of Central Florida in 1999.


3:00 – 3:30 PM Coffee Break

Special Session on Image Quality in Standardization

Session Chair: Jonathan Phillips, Google Inc. (United States)
3:30 – 4:50 PM
Grand Peninsula Ballroom E

3:30IQSP-312
Study of subjective and objective quality evaluation of 3D point cloud data by the JPEG Committee, Stuart Perry1, Antonio Pinheiro2, Emil Dumic3, and Luis Cruz4; 1University of Technology Sydney (Australia), 2University of Beira Interior (Portugal),, 3University North (Croatia), and 4University of Coimbra (Portugal)

3:50IQSP-313
Reducing the cross-lab variation of image quality metrics, Henry Koren1 and Benjamin Tseng2; 1Imatest LLC and 2Apkudo (United States)

4:10IQSP-314
Adaptive video streaming with current codecs and formats: Extensions to parametric video quality model ITU-T P.1203, Rakesh Rao Ramachandra Rao1, Steve Göring1, Patrick Vogel1, Nicolas Pachatz1, Juan Jose Villamar Villarreal1, Werner Robitza1, Peter List2, Bernhard Feiten2, and Alexander Raake1; 1TU Ilmenau and 2Deutsche Telekom (Germany)

4:30IQSP-315
Visual noise revision for ISO 15739, Dietmar Wueller1, Akira Matsui2, and Naoyah Kato2; 1Image Engineering GmbH & Co. KG (Germany) and 2Sony (Japan)



IQSP of the Future

Panelists: Nicolas Bonnier, Apple Inc. (United States); Peter Burns, Burns Digital Imaging (United States); Mylène Farias, University of Brasilia (Brazil); and Elaine Jin, NVIDIA Corporation (United States)
Panel Moderator: Stuart Perry, University of Technology Sydney (Australia)
4:50 – 5:30 PM
Grand Peninsula Ballroom E



5:30 – 7:30 PM Symposium Demonstration Session

Wednesday January 16, 2019

Camera Image Quality I

Session Chair: Peter Burns, Burns Digital Imaging (United States)
8:50 – 9:30 AM
Grand Peninsula Ballroom E

8:50IQSP-316
Multivariate statistical modeling for image quality prediction, Praful Gupta1, Christos Bampis2, Jack Glover3, Nicholas Paulter3, and Alan Bovik1; 1The University of Texas at Austin, 2Netflix Inc., and 3National Institute of Standards and Technology (United States)

9:10IQSP-317
Image quality assessment using computer vision, Zhi Li1, Palghat Ramesh2,3, and Chu-heng Liu3; 1Purdue University, 2Palo Alto Research Center, and 3Xerox Corporation (United States)



Camera Image Quality II

Session Chair: Peter Burns, Burns Digital Imaging (United States)
9:30 – 10:10 AM
Grand Peninsula Ballroom E

IQSP-318
KEYNOTE: Benchmarking image quality for billions of images, Jonathan Phillips, Google Inc. (United States)

Jonathan Phillips is co-author of Camera Image Quality Benchmarking, a 2018 addition to the Wiley-IS&T Series in Imaging Science and Technology collection. His experience in the imaging industry spans nearly 30 years, having worked at Kodak in both chemical and electronic photography for more than 20 years followed by image scientist positions with NVIDIA and Google. Currently, he is managing a color science team at Google responsible for the display color of the Pixel phone product line. He was awarded the International Imaging Industry Association (I3A) Achievement Award for his groundbreaking work on modeling consumer-facing camera phone image quality, which is now incorporated into the IEEE Standard for Camera Phone Image Quality. Jonathan has been project lead for numerous photography standards published by I3A, IEEE, and ISO. His graduate studies were in color science at Rochester Institute of Technology and his undergraduate studies were in chemistry and music at Wheaton College (IL).




10:00 AM – 3:30 PM Industry Exhibition

10:10 – 10:40 AM Coffee Break

Video Quality

Session Chair: Elaine Jin, NVIDIA Corporation (United States)
10:40 AM – 12:40 PM
Grand Peninsula Ballroom E

10:40IQSP-319
Best practices for imaging system MTF measurement, David Haefner, Stephen Burks, Josh Doe, and Bradley Preece, NVESD (United States)

11:00IQSP-320
Quantify aliasing – A new approach to make resolution measurement more robust, Uwe Artmann, Image Engineering GmbH & Co KG (Germany)

11:20IQSP-321
Subjective analysis of an end-to-end streaming system, Christos Bampis1, Zhi Li1, Ioannis Katsavounidis2, Te-Yuan Huang1, Chaitanya Ekanadham1, and Alan Bovik3; 1Netflix Inc., 2Facebook, Inc., and 3The University of Texas at Austin (United States)

11:40IQSP-322
Saliency-based perceptual quantization method for HDR video quality enhancement, Naty Sidaty, Wassim Hamidouche, Yi Liu, and Olivier Deforges, IETR/INSA (France)

12:00IQSP-323
Subjective and objective quality assessment for volumetric video compression, Emin Zerman, Pan Gao, Cagri Ozcinar, and Aljosa Smolic, Trinity College Dublin (Ireland)

12:20IQSP-324
Analyzing the influence of cross-modal IP-based degradations on the perceived audio-visual quality, Helard Becerra and Mylène Farias, University of Brasilia (Brazil)



12:40 – 2:00 PM Lunch

Wednesday Plenary

2:00 – 3:00 PM
Grand Peninsula Ballroom D

Light Fields and Light Stages for Photoreal Movies, Games, and Virtual Reality, Paul Debevec, Senior Scientist, Google (United States)

Paul Debevec will discuss the technology and production processes behind "Welcome to Light Fields", the first downloadable virtual reality experience based on light field capture techniques which allow the visual appearance of an explorable volume of space to be recorded and reprojected photorealistically in VR enabling full 6DOF head movement. The lightfields technique differs from conventional approaches such as 3D modelling and photogrammetry. Debevec will discuss the theory and application of the technique. Debevec will also discuss the Light Stage computational illumination and facial scanning systems which use geodesic spheres of inward-pointing LED lights as have been used to create digital actor effects in movies such as Avatar, Benjamin Button, and Gravity, and have recently been used to create photoreal digital actors based on real people in movies such as Furious 7, Blade Runner: 2049, and Ready Player One. Th lighting reproduction process of light stages allows omnidirectional lighting environments captured from the real world to be accurately reproduced in a studio, and has recently be extended with multispectral capabilities to enable LED lighting to accurately mimic the color rendition properties of daylight, incandescent, and mixed lighting environments. They have also recently used their full-body light stage in conjunction with natural language processing and automultiscopic video projection to record and project interactive conversations with survivors of the World War II Holocaust.

Paul Debevec is a Senior Scientist at Google VR, a member of GoogleVR's Daydream team, and Adjunct Research Professor of Computer Science in the Viterbi School of Engineering at the University of Southern California, working within the Vision and Graphics Laboratory at the USC Institute for Creative Technologies. Debevec's computer graphics research has been recognized with ACM SIGGRAPH's first Significant New Researcher Award in 2001 for "Creative and Innovative Work in the Field of Image-Based Modeling and Rendering", a Scientific and Engineering Academy Award in 2010 for "the design and engineering of the Light Stage capture devices and the image-based facial rendering system developed for character relighting in motion pictures" with Tim Hawkins, John Monos, and Mark Sagar, and the SMPTE Progress Medal in 2017 in recognition of his achievements and ongoing work in pioneering techniques for illuminating computer-generated objects based on measurement of real-world illumination and their effective commercial application in numerous Hollywood films. In 2014, he was profiled in The New Yorker magazine's "Pixel Perfect: The Scientist Behind the Digital Cloning of Actors" article by Margaret Talbot.


3:00 – 3:30 PM Coffee Break

Immersive QoE

Session Chairs: Kjell Brunnström, RISE Acreo AB (Sweden) and Stuart Perry, University of Technology Sydney (Australia)
3:30 – 5:10 PM
Grand Peninsula Ballroom A

This session is jointly sponsored by: Human Vision and Electronic Imaging 2019, and Image Quality and System Performance XVI.


3:30HVEI-216
Complexity measurement and characterization of 360-degree content, Francesca De Simone1, Jesús Gutiérrez2, and Patrick Le Callet2; 1CWI (the Netherlands) and 2Université de Nantes (France)

3:50HVEI-217
Using 360 VR video to improve the learning experience in veterinary medicine university degree, Esther Guervós1, Jaime Jesús Ruiz2, Pablo Perez2, Juan Alberto Muñoz1, César Díaz3, and Narciso Garcia3; 1Universidad Alfonso X El Sabio, 2Nokia Bell Labs, and 3Universidad Politécnica de Madrid (Spain)

4:10HVEI-218
Quality of Experience of visual-haptic interaction in a virtual reality simulator, Kjell Brunnström1,2, Elijs Dima2, Mattias Andersson2, Mårten Sjöström2, Tahir Qureshi3, and Mathias Johanson4; 1RISE Acreo AB, 2Mid Sweden University, 3HIAB AB, and 4Alkit Communications AB (Sweden)

4:30HVEI-219
Impacts of internal HMD playback processing on subjective quality perception, Frank Hofmeyer, Stephan Fremerey, Thaden Cohrs, and Alexander Raake, Technische Universität Ilmenau (Germany)

4:50IQSP-220
Are people pixel-peeping 360° videos?, Stephan Fremerey1, Rachel Huang2, and Alexander Raake1; 1Technische Universität Ilmenau (Germany) and 2Huawei Technologies Co., Ltd. (China)



Image Quality and System Performance XVI Interactive Posters Session

5:30 – 7:00 PM
The Grove

The following works will be presented at the EI 2019 Symposium Interactive Papers Session.


IQSP-325
An examination of the effects of noise level on methods to determine curvature in range images, Jacob Hauenstein and Timothy Newman, The University of Alabama in Huntsville (United States)

IQSP-326
The characterization of an HDR OLED display, Dalin Tian, Lihao Xu, and Ming Ronnier Luo, Zhejiang Univesity (China)

IQSP-327
Understanding fashion aesthetics: Training a neural network based predictor using likes and dislikes, Rachel Bilbo, Kendal Norman, Zhi Li, and Jan Allebach, Purdue University (United States)



No content found

No content found

 

Important Dates
Call for Papers Announced 1 Mar 2018
Journal-first Submissions Due 30 Jun 2018
Abstract Submission Site Opens 1 May 2018
Review Abstracts Due (refer to For Authors page
 · Early Decision Ends 30 Jun 2018
· Regular Submission Ends 8 Sept 2018
· Extended Submission Ends 25 Sept 2018
 Final Manuscript Deadlines  
 · Fast Track Manuscripts Due 14 Nov 2018 
 · Final Manuscripts Due 1 Feb 2019 
Registration Opens 23 Oct 2018
Early Registration Ends 18 Dec 2018
Hotel Reservation Deadline 3 Jan 2019
Conference Begins 13 Jan 2019


 
View 2019 Proceedings
View 2018 Proceedings
View 2017 Proceedings
View 2016 Proceedings

Conference Chairs
Nicolas Bonnier, Apple Inc. (United States);  Stuart Perry, University of Technology Sydney (Australia)

Program Committee
Alan Bovik, University of Texas at Austin (United States); Peter Burns, Burns Digital Imaging ; Brian Cooper, Lexmark International, Inc. (United States); Luke Cui, Amazon (United States); Mylène Farias, University of Brasilia (Brazil); Susan Farnand, Rochester Institute of Technology (United States); Frans Gaykema, Océ Technologies B.V. (Netherlands); Jukka Häkkinen, University of Helsinki (Finland); Dirk Hertel, E Ink Corporation (United States); Robin Jenkin, NVIDIA Corporation (United States); Elaine Jin, NVIDIA Corporation (United States); Mohamed-Chaker Larabi, University of Poitiers (France); Göte Nyman, University of Helsinki (Finland); Jonathan Phillips, Google Inc. (United States); Sophie Triantaphillidou, University of Westminster (United Kingdom); Clément Viard, DxOMark Image Labs (United States)