PERSONAL Sign in with your SPIE account to access your personal subscriptions or to use specific features such as save to my library, sign up for alerts, save searches, etc.
This PDF file contains the front matter associated with SPIE Proceedings Volume 9396, including the Title Page, Copyright information, Table of Contents, Introduction (if any), and Conference Committee listing.
Access to the requested content is limited to institutions that have purchased or subscribe to SPIE eBooks.
You are receiving this notice because your organization may not have SPIE eBooks access.*
*Shibboleth/Open Athens users─please
sign in
to access your institution's subscriptions.
To obtain this item, you may purchase the complete book in print or electronic format on
SPIE.org.
In this paper, a practical solution for optimal coding parameters using different bandwidth requirements is presented. The
obtained specification is based on the analysis of a large database of more than 10,000 sequences compressed with different
parameters. The obtained parameters can be used both for adaptive streaming or storage optimization.
Access to the requested content is limited to institutions that have purchased or subscribe to SPIE eBooks.
You are receiving this notice because your organization may not have SPIE eBooks access.*
*Shibboleth/Open Athens users─please
sign in
to access your institution's subscriptions.
To obtain this item, you may purchase the complete book in print or electronic format on
SPIE.org.
High-Content Screening (HCS) is a powerful technology for biological research, which relies heavily on the
capabilities for processing and analysis of cell biology images. The quality of the quantification results, obtained
by analysis of hundreds and thousands of images, is crucial for analysis of biological phenomena under study.
Traditionally, a quality control in the HCS refers to the preparation of biological assay, setting up instrumentation,
and analysis of the obtained quantification results, thus skipping an important step of assessment of the image
quality. So far, only few papers have been addressing this issue, but no standard methodology yet exists, that
would allow pointing out images, potentially producing outliers when processed. In this research the importance
of the image quality control for the HCS is emphasized, with the following possible advantages: (a) validation
of the visual quality of the screening; (b) detection of the potentially problem images; (c) more accurate setting
of the processing parameters. For the detection of outlier images the Power Log-Log Slope (PLLS) is applied,
as it is known to be sensitive to the focusing errors, and validated using open data sets. The results show that
PLLS correlates with the cell counting error and, when taken it into account, allows reducing the variance of
measurements. Possible extensions and problems of the approach are discussed.
Access to the requested content is limited to institutions that have purchased or subscribe to SPIE eBooks.
You are receiving this notice because your organization may not have SPIE eBooks access.*
*Shibboleth/Open Athens users─please
sign in
to access your institution's subscriptions.
To obtain this item, you may purchase the complete book in print or electronic format on
SPIE.org.
Modern image processing pipelines (e.g., those used in digital cameras) are full of advanced, highly adaptive filters that
often have a large number of tunable parameters (sometimes > 100). This makes the calibration procedure for these
filters very complex, and the optimal results barely achievable in the manual calibration; thus an automated approach is a
must. We will discuss an information theory based metric for evaluation of algorithm adaptive characteristics
(“adaptivity criterion”) using noise reduction algorithms as an example. The method allows finding an “orthogonal
decomposition” of the filter parameter space into the “filter adaptivity” and “filter strength” directions. This metric can
be used as a cost function in automatic filter optimization. Since it is a measure of a physical “information restoration”
rather than perceived image quality, it helps to reduce the set of the filter parameters to a smaller subset that is easier for
a human operator to tune and achieve a better subjective image quality. With appropriate adjustments, the criterion can
be used for assessment of the whole imaging system (sensor plus post-processing).
Access to the requested content is limited to institutions that have purchased or subscribe to SPIE eBooks.
You are receiving this notice because your organization may not have SPIE eBooks access.*
*Shibboleth/Open Athens users─please
sign in
to access your institution's subscriptions.
To obtain this item, you may purchase the complete book in print or electronic format on
SPIE.org.
This paper presents a novel High Dynamic Range (HDR) tone mapping (TM) system based on sub-band architecture.
Standard wavelet filters of Daubechies, Symlets, Coiflets and Biorthogonal were used to estimate
the proposed system performance in terms of Low Dynamic Range (LDR) image quality and reconstructed
HDR image fidelity. During TM stage, the HDR image is firstly decomposed in sub-bands using symmetrical
analysis-synthesis filter bank. The transform coefficients are then rescaled using a predefined gain map. The
inverse Tone Mapping (iTM) stage is straightforward. Indeed, the LDR image passes through the same sub-band
architecture. But, instead of reducing the dynamic range, the LDR content is boosted to an HDR representation.
Moreover, in our TM sheme, we included an optimization module to select the gain map components that
minimize the reconstruction error, and consequently resulting in high fidelity HDR content. Comparisons with
recent state-of-the-art methods have shown that our method provides better results in terms of visual quality
and HDR reconstruction fidelity using objective and subjective evaluations.
Access to the requested content is limited to institutions that have purchased or subscribe to SPIE eBooks.
You are receiving this notice because your organization may not have SPIE eBooks access.*
*Shibboleth/Open Athens users─please
sign in
to access your institution's subscriptions.
To obtain this item, you may purchase the complete book in print or electronic format on
SPIE.org.
Digital Photography and Image Quality I, Joint Session with Conferences 9396 and 9404
We present a methodology to compare image sensors with traditional Bayer RGB layouts to sensors with alternative
layouts containing white pixels. We focused on the sensors’ resolving powers, which we measured in the form of a
modulation transfer function for variations in both luma and chroma channels. We present the design of the test chart,
the acquisition of images, the image analysis, and an interpretation of results. We demonstrate the approach at the
example of two sensors that only differ in their color filter arrays. We confirmed that the sensor with white pixels and
the corresponding demosaicing result in a higher resolving power in the luma channel, but a lower resolving power in
the chroma channels when compared to the traditional Bayer sensor.
Access to the requested content is limited to institutions that have purchased or subscribe to SPIE eBooks.
You are receiving this notice because your organization may not have SPIE eBooks access.*
*Shibboleth/Open Athens users─please
sign in
to access your institution's subscriptions.
To obtain this item, you may purchase the complete book in print or electronic format on
SPIE.org.
Objective evaluation of digital image quality usually includes analysis of spatial detail in captured images. Although
previously-developed methods and standards have found success in the evaluation of system performance, the systems in
question usually include spatial image processing (e.g. sharpening or noise-reduction), and the results are influenced by
these operations. Our interest, however, is in the intrinsic resolution of the system. By this we mean the performance
primarily defined by the lens and imager, and not influenced by subsequent image processing steps that are invertible.
Examples of such operations are brightness and contrast adjustments, and simple sharpening and blurring (setting aside
image clipping and quantization). While these operations clearly modify image perception, they do not in general change
the fundamental spatial image information that is captured.
We present a method to measure an intrinsic spatial frequency response (SFR) computed from test image(s) for which
spatial operations may have been applied. The measure is intended ‘see through’ operations for which image detail is
retrievable but measure the loss of image resolution otherwise. We adopt a two-stage image capture model. The first
stage includes a locally-stable point-spread function (lens), the integration and sampling by the detector (imager), and
the introduction of detector noise. The second stage comprises the spatial image processing. We describe the validation
of the method, which was done using both simulation and actual camera evaluations.
Access to the requested content is limited to institutions that have purchased or subscribe to SPIE eBooks.
You are receiving this notice because your organization may not have SPIE eBooks access.*
*Shibboleth/Open Athens users─please
sign in
to access your institution's subscriptions.
To obtain this item, you may purchase the complete book in print or electronic format on
SPIE.org.
Digital Photography and Image Quality II, Joint Session with Conferences 9396 and 9404
High noise values and poor signal to noise ratio are traditionally associated to the low light imaging. Still, there are
several other camera quality features which may suffer from low light environment. For example, what happens to the
color accuracy and resolution or how the camera speed behaves in low light? Furthermore, how low light environments
affect to the camera benchmarking and which metrics are the critical ones?
The work contains standard based image quality measurements including noise, color, and resolution measurements in
three different light environments: 1000, 100, and 30 lux. Moreover, camera speed measurements are done. Detailed
measurement results of each quality and speed category are revealed and compared. Also a suitable benchmark algorithm
is evaluated and corresponding score is calculated to find an appropriate metric which characterize the camera
performance in different environments.
The result of this work introduces detailed image quality and camera speed measurements of mobile phone camera
systems in three different light environments. The paper concludes how different light environments influence to the
metrics and which metrics should be measured in low light environment. Finally, a benchmarking score is calculated
using measurement data of each environment and mobile phone cameras are compared correspondingly.
Access to the requested content is limited to institutions that have purchased or subscribe to SPIE eBooks.
You are receiving this notice because your organization may not have SPIE eBooks access.*
*Shibboleth/Open Athens users─please
sign in
to access your institution's subscriptions.
To obtain this item, you may purchase the complete book in print or electronic format on
SPIE.org.
This study presents effective method to provide user’s visual comfort and reduce power consumption through the
optimization of display luminance and gamma curve for using mobile display in low ambient conditions. Visual
assessment was carried out to obtain appropriate display luminance for visual comfort, and gamma curve was optimized
to obtain power saving effect without reducing image quality in lower luminance than in appropriate luminance. This
suggested method was verified through perceptual experiment and power consumption measurement.
Access to the requested content is limited to institutions that have purchased or subscribe to SPIE eBooks.
You are receiving this notice because your organization may not have SPIE eBooks access.*
*Shibboleth/Open Athens users─please
sign in
to access your institution's subscriptions.
To obtain this item, you may purchase the complete book in print or electronic format on
SPIE.org.
A new method for evaluating the perceptual resolution of the actual image has been developed. The characteristic of this
new evaluation method is that the square sum of the cross correlation coefficients becomes the index of the evaluation.
Because the RIT Contrast-Resolution Test Target was a very systematic analysis pattern that contains spatial resolution
and contrast, it was used for this study. The cross correlation coefficient is calculated between the reference image and
the printed image read by a scanner. To provide the precision of the new evaluation method, subjective evaluation was
performed with the actual image. As a result of having confirmed correlation of the evaluation value from subjective
evaluation and the new evaluation method, the correlation was much higher than other evaluation method using a similar
analysis pattern. Furthermore, the sharpness evaluation of the actual image was enabled by applying the evaluation value
of this new method to evaluate the perceptual resolution, that correlation with the subjectivity was high.
Access to the requested content is limited to institutions that have purchased or subscribe to SPIE eBooks.
You are receiving this notice because your organization may not have SPIE eBooks access.*
*Shibboleth/Open Athens users─please
sign in
to access your institution's subscriptions.
To obtain this item, you may purchase the complete book in print or electronic format on
SPIE.org.
Multifunctional printers (MFP) are products that combine the functions of a printer, scanner, and copier. Our goal is to help customers to be able to easily diagnose scanner or print quality issues with their products by developing an automated diagnostic system embedded in the product. We specifically focus on the characterization of scanner motions, which may be defective due to irregular movements of the scan-head. The novel design of our test page and two-stage diagnostic algorithm are described in this paper. The most challenging issue is to evaluate the scanner performance properly when both printer and scanner units contribute to the motion errors. In the first stage called the uncorrected-print-error-stage, aperiodic and periodic motion behaviors are characterized in both the spatial and frequency domains. Since it is not clear how much of the error is contributed by each unit, the scanned input is statistically analyzed in the second stage called the corrected-print-error-stage. Finally, the described diagnostic algorithms output the estimated scan error and print error separately as RMS values of the displacement of the scan and print lines, respectively, from their nominal positions in the scanner or printer motion direction. We validate our test page design and approaches by ground truth obtained from a high-precision, chrome-on-glass reticle manufactured using semiconductor chip fabrication technologies.
Access to the requested content is limited to institutions that have purchased or subscribe to SPIE eBooks.
You are receiving this notice because your organization may not have SPIE eBooks access.*
*Shibboleth/Open Athens users─please
sign in
to access your institution's subscriptions.
To obtain this item, you may purchase the complete book in print or electronic format on
SPIE.org.
Image quality assessment is a very important field in image processing. Human observation is slow and subjective, it
also requires strict environment setup for the psychological test 1. Thus developing algorithms to match desired human
experiments is always in need. Many studies have focused on detecting the fading phenomenon after the materials are
printed, that is to monitor the persistence of the color ink 2-4. However, fading is also a common artifact produced by
printing systems when the cartridges run low. We want to develop an automatic system to monitor cartridge life and
report fading defects when they appear. In this paper, we first describe a psychological experiment that studies the
human perspective on printed fading pages. Then we propose an algorithm based on Color Space Projection and K-means
clustering to predict the visibility of fading defects. At last, we integrate the psychological experiment result with
our algorithm to give a machine learning tool that monitors cartridge life.
Access to the requested content is limited to institutions that have purchased or subscribe to SPIE eBooks.
You are receiving this notice because your organization may not have SPIE eBooks access.*
*Shibboleth/Open Athens users─please
sign in
to access your institution's subscriptions.
To obtain this item, you may purchase the complete book in print or electronic format on
SPIE.org.
Fading is one of the issues of most critical concern for print quality degradation with color laser electro-
photographic printers. Fading occurs when the cartridge is depleted. ISO/IEC 19798:2007(E) specifies a process
for determining the cartridge page yield for a given color electro-photographic printer model. It is based on
repeatedly printing a suite of test pages, followed by visual examination of the sequence of printed diagnostic
pages. But this method is a very costly process since it involves visual examination of a large number of pages.
And also the final decision is based on the visual examination of a specially designed diagnostic page, which is
different than typical office document pages, since it consists of color bars, and contains no text. In this paper,
we propose a new method to autonomously detect the text fading in prints from home or office color printers
using a typical office document page instead of a specially designed diagnostic page. In our method, we scan and
analyze the printed pages to predict where expert observers would judge fading to have occurred in the print
sequence. Our approach is based on a machine-learning framework in which features derived from image analysis
are mapped to a fade point prediction.
Access to the requested content is limited to institutions that have purchased or subscribe to SPIE eBooks.
You are receiving this notice because your organization may not have SPIE eBooks access.*
*Shibboleth/Open Athens users─please
sign in
to access your institution's subscriptions.
To obtain this item, you may purchase the complete book in print or electronic format on
SPIE.org.
Manufacturing imperfections of photoconductor (PC) drums in electrophotographic (EP) printers cause low-
frequency artifacts that could produce objectionable non-uniformities in the final printouts. In this paper, we
propose a technique to detect and quantify PC artifacts. Furthermore, we spatially model the PC drum surface
for dynamic compensation of drum artifacts. After scanning printed pages of flat field areas, we apply a wavelet-
based filtering technique to the scanned images to isolate the PC-related artifacts from other printing artifacts,
based on the frequency, range, and direction of the PC defects. Prior knowledge of the PC circumference
determines the printed area at each revolution of the drum for separate analysis. Applied to the filtered images,
the expectation maximization (EM) algorithm models the PC defects as a mixture of Gaussians. We use the
estimated parameters of the Gaussians to measure the severity of the defect. In addition, a 2-D polynomial fitting
approach characterizes the spatial artifacts of the drum, by analyzing multiple revolutions of printed output.
The experimental results show a high correlation of the modeled artifacts from different revolutions of a drum.
This allows for generating a defect-compensating profile of the defective drum.
Access to the requested content is limited to institutions that have purchased or subscribe to SPIE eBooks.
You are receiving this notice because your organization may not have SPIE eBooks access.*
*Shibboleth/Open Athens users─please
sign in
to access your institution's subscriptions.
To obtain this item, you may purchase the complete book in print or electronic format on
SPIE.org.
In our previous work1 , we presented a block-based technique to analyze printed page uniformity both visually
and metrically. The features learned from the models were then employed in a Support Vector Machine (SVM)
framework to classify the pages into one of the two categories of acceptable and unacceptable quality. In this
paper, we introduce a set of tools for machine learning in the assessment of printed page uniformity. This
work is primarily targeted to the printing industry, specifically the ubiquitous laser, electrophotographic printer.
We use features that are well-correlated with the rankings of expert observers to develop a novel machine
learning framework that allows one to achieve the minimum "false alarm" rate, subject to a chosen "miss" rate.
Surprisingly, most of the research that has been conducted on machine learning does not consider this framework.
During the process of developing a new product, test engineers will print hundreds of test pages, which can
be scanned and then analyzed by an autonomous algorithm. Among these pages, most may be of acceptable
quality. The objective is to find the ones that are not. These will provide critically important information to
systems designers, regarding issues that need to be addressed in improving the printer design. A "miss" is defined
to be a page that is not of acceptable quality to an expert observer that the prediction algorithm declares to
be a "pass". Misses are a serious problem, since they represent problems that will not be seen by the systems
designers. On the other hand, "false alarms" correspond to pages that an expert observer would declare to be of
acceptable quality, but which are flagged by the prediction algorithm as "fails". In a typical printer testing and
development scenario, such pages would be examined by an expert, and found to be of acceptable quality after
all. "False alarm" pages result in extra pages to be examined by expert observers, which increases labor cost.
But "false alarms" are not nearly as catastrophic as "misses", which represent potentially serious problems that
are never seen by the systems developers. This scenario motivates us to develop a machine learning framework
that will achieve the minimum "false alarm" rate subject to a specified "miss" rate. In order to construct such a
set of receiver operating characteristic2 (ROC) curves, we examine various tools for the prediction, ranging from
an exhaustive search over the space of the nonlinear discriminants to a Cost-Sentitive SVM3 framework. We then
compare the curves gained from those methods. Our work shows promise for applying a standard framework to
obtain a full ROC curve when it comes to tackling other machine learning problems in industry.
Access to the requested content is limited to institutions that have purchased or subscribe to SPIE eBooks.
You are receiving this notice because your organization may not have SPIE eBooks access.*
*Shibboleth/Open Athens users─please
sign in
to access your institution's subscriptions.
To obtain this item, you may purchase the complete book in print or electronic format on
SPIE.org.
In the printing industry, electrophotography (EP) is a commonly used technology in laser printers and copiers. In
the EP printing process, there are many rotating components involved in the six major steps: charging, exposure,
development, transfer, fusing, and cleaning. If there is any irregularity in one of the rotating components, repetitive defects, such as isolated bands or spots, will occur on the output of the printer or copier. To troubleshoot
these types of repetitive defect issues, the repeating interval of these isolated bands or spots is an important clue
to locate the irregular rotating component. In our previous work, we have effectively identified the presence of
isolated large pitch bands in the output from EP printers. In this paper, we describe an algorithm to estimate
the repetitive interval of periodic bands, when the data is corrupted by the presence of aperiodic bands, missing
periodic bands, and noise. We will also illustrate the effectiveness and robustness of our method with example
results.
Access to the requested content is limited to institutions that have purchased or subscribe to SPIE eBooks.
You are receiving this notice because your organization may not have SPIE eBooks access.*
*Shibboleth/Open Athens users─please
sign in
to access your institution's subscriptions.
To obtain this item, you may purchase the complete book in print or electronic format on
SPIE.org.
What is the best luminance contrast weighting-function for image quality optimization? Traditionally measured contrast
sensitivity functions (CSFs), have been often used as weighting-functions in image quality and difference metrics. Such
weightings have been shown to result in increased sharpness and perceived quality of test images. We suggest contextual
CSFs (cCSFs) and contextual discrimination functions (cVPFs) should provide bases for further improvement, since
these are directly measured from pictorial scenes, modeling threshold and suprathreshold sensitivities within the context
of complex masking information. Image quality assessment is understood to require detection and discrimination of
masked signals, making contextual sensitivity and discrimination functions directly relevant.
In this investigation, test images are weighted with a traditional CSF, cCSF, cVPF and a constant function. Controlled
mutations of these functions are also applied as weighting-functions, seeking the optimal spatial frequency band
weighting for quality optimization. Image quality, sharpness and naturalness are then assessed in two-alternative forced-choice
psychophysical tests. We show that maximal quality for our test images, results from cCSFs and cVPFs, mutated
to boost contrast in the higher visible frequencies.
Access to the requested content is limited to institutions that have purchased or subscribe to SPIE eBooks.
You are receiving this notice because your organization may not have SPIE eBooks access.*
*Shibboleth/Open Athens users─please
sign in
to access your institution's subscriptions.
To obtain this item, you may purchase the complete book in print or electronic format on
SPIE.org.
The slanted-edge method of measuring the spatial frequency response (SFR) as an approximation of the modulation transfer function (MTF) has become a well known and widely used image quality testing method over
the last 10 years. This method has been adopted by multiple international standards including ISO and IEEE.
Nearly every commercially available image quality testing software includes the slanted-edge method and there
are numerous open-source algorithms available. This method is one of the most important image quality algorithms in use today. This paper explores test conditions and the impacts they have on the stability and precision
of the slanted-edge method as well as details of the algorithm itself. Real world and simulated data are used
to validate the characteristics of the algorithm. Details of the target such as edge angle and contrast ratio are
tested to determine the impact on measurement under various conditions. The original algorithm defines a near
vertical edge so that errors introduced are minor but the theory behind the algorithm requires a perfectly vertical
edge. A correction factor is introduced as a way to compensate for this problem. Contrast ratio is shown to have
no impact on results in an absence of noise.
Access to the requested content is limited to institutions that have purchased or subscribe to SPIE eBooks.
You are receiving this notice because your organization may not have SPIE eBooks access.*
*Shibboleth/Open Athens users─please
sign in
to access your institution's subscriptions.
To obtain this item, you may purchase the complete book in print or electronic format on
SPIE.org.
In this investigation we identify relationships between human and automated face recognition systems with respect to
compression. Further, we identify the most influential scene parameters on the performance of each recognition system.
The work includes testing of the systems with compressed Closed-Circuit Television (CCTV) footage, consisting of
quantified scene (footage) parameters. Parameters describe the content of scenes concerning camera to subject distance,
facial angle, scene brightness, and spatio-temporal busyness. These parameters have been previously shown to affect the
human visibility of useful facial information, but not much work has been carried out to assess the influence they have
on automated recognition systems. In this investigation, the methodology previously employed in the human
investigation is adopted, to assess performance of three different automated systems: Principal Component Analysis,
Linear Discriminant Analysis, and Kernel Fisher Analysis. Results show that the automated systems are more tolerant to
compression than humans. In automated systems, mixed brightness scenes were the most affected and low brightness
scenes were the least affected by compression. In contrast for humans, low brightness scenes were the most affected and
medium brightness scenes the least affected. Findings have the potential to broaden the methods used for testing imaging
systems for security applications.
Access to the requested content is limited to institutions that have purchased or subscribe to SPIE eBooks.
You are receiving this notice because your organization may not have SPIE eBooks access.*
*Shibboleth/Open Athens users─please
sign in
to access your institution's subscriptions.
To obtain this item, you may purchase the complete book in print or electronic format on
SPIE.org.
This work is performed as part of the perceptual validation stage in the stereoscopic visualization of computer-
generated (CG) images of materials (typically car paints) containing sparkling metallic flakes. The perceived material aspect is closely linked to the flake density, depth, and sparkling; in turn, our perception
of an image of said materials is strongly dependent on the image exposure, that is, the amount of light
entering the sensor during the imaging process. Indeed, a high exposure may over saturate the image, reducing discrimination amongst high-luminance flakes, affecting the perceived depth; on the other hand, a
low exposure may reduce image contrast, merging low-luminance flakes with the background, and reducing
perceived flake density and sparkling. In order to choose the right exposure for each CG image, we have
performed a user study where we presented observers with a series of stereoscopic photographs of plates,
taken at different exposures with a radiometrically color-calibrated camera ,5 and asked them to assess each
photograph's similarity to a physical reference. We expect these results to help us find a correlation between
optical settings and visual perception regarding the aforementioned parameters, which we could then use in
the rendering process to obtain the desired material aspect.
Access to the requested content is limited to institutions that have purchased or subscribe to SPIE eBooks.
You are receiving this notice because your organization may not have SPIE eBooks access.*
*Shibboleth/Open Athens users─please
sign in
to access your institution's subscriptions.
To obtain this item, you may purchase the complete book in print or electronic format on
SPIE.org.
QuickEval is a web application for carrying out psychometric scaling experiments. It offers the possibility of
running controlled experiments in a laboratory, or large scale experiment over the web for people all over the
world. It is a unique one of a kind web application, and it is a software needed in the image quality field.
It is also, to the best of knowledge, the first software that supports the three most common scaling methods;
paired comparison, rank order, and category judgement. It is also the first software to support rank order.
Hopefully, a side effect of this newly created software is that it will lower the threshold to perform psychometric
experiments, improve the quality of the experiments being carried out, make it easier to reproduce experiments,
and increase research on image quality both in academia and industry. The web application is available at
www.colourlab.no/quickeval.
Access to the requested content is limited to institutions that have purchased or subscribe to SPIE eBooks.
You are receiving this notice because your organization may not have SPIE eBooks access.*
*Shibboleth/Open Athens users─please
sign in
to access your institution's subscriptions.
To obtain this item, you may purchase the complete book in print or electronic format on
SPIE.org.
We introduce a new image database dedicated to multi-/hyperspectral image quality assessment. A total of nine
scenes representing pseudo-at surfaces of different materials (textile, wood, skin. . . ) were captured by means of
a 160 band hyperspectral system with a spectral range between 410 and 1000nm. Five spectral distortions were
designed, applied to the spectral images and subsequently compared in a psychometric experiment, in order to
provide a basis for applications such as the evaluation of spectral image difference measures. The database can
be downloaded freely from http://www.colourlab.no/cid.
Access to the requested content is limited to institutions that have purchased or subscribe to SPIE eBooks.
You are receiving this notice because your organization may not have SPIE eBooks access.*
*Shibboleth/Open Athens users─please
sign in
to access your institution's subscriptions.
To obtain this item, you may purchase the complete book in print or electronic format on
SPIE.org.
An established way of validating and testing new image quality assessment (IQA) algorithms have been to compare how
well they correlate with subjective data on various image databases. One of the most common measures is to calculate
linear correlation coefficient (LCC) and Spearman’s rank order correlation coefficient (SROCC) against the subjective
mean opinion score (MOS). Recently, databases with multiply distorted images have emerged 1,2. However with
multidimensional stimuli, there is more disagreement between observers as the task is more preferential than that of
distortion detection. This reduces the statistical differences between image pairs. If the subjects cannot distinguish a
difference between some of the image pairs, should we demand any better performance with IQA algorithms? This paper
proposes alternative performance measures for the evaluation of IQA’s for the CID2013 database. One proposed
alternative performance measure is root-mean-square-error (RMSE) value for the subjective data as a function of the
number of observers. The other alternative performance measure is the number of statistical differences between image
pairs. This study shows that after 12 subjects the RMSE value saturates around the level of three, meaning that a target
RMSE value for an IQA algorithm for CID2013 database should be three. In addition, this study shows that the state-of-the-art IQA algorithms found the better image from the image pairs with a probability of 0.85 when the image pairs with
statistically significant differences were taken into account.
Access to the requested content is limited to institutions that have purchased or subscribe to SPIE eBooks.
You are receiving this notice because your organization may not have SPIE eBooks access.*
*Shibboleth/Open Athens users─please
sign in
to access your institution's subscriptions.
To obtain this item, you may purchase the complete book in print or electronic format on
SPIE.org.
In a typical working cycle of image quality assessment, it is common to have a number of human observers
to give perceptual ratings on multiple levels of distortions of selected test images. If additional distortions
need to be introduced into the experiment, the entire subjective experiment must be performed over again in
order to incorporate the additional distortions. However, this would usually consume considerable more time
and resources. Baseline adjustment is one method to extend an experiment with additional distortions without
having to do a full experiment, reducing both the time and resources needed. In this paper, we conduct a study
to verify and evaluate the baseline adjustment method regarding extending an existing subjective experimental
session to another. Our experimental results suggest that the baseline adjustment method can be effective. We
identify the optimal distortion levels to be included in the baselines should be the ones of which the stimulus
combinations produce the minimum standard deviations in the mean adjusted Z-scores over all human observers
in the existing rating session. We also demonstrate that it is possible to reduce the number of baseline stimuli,
so the cost of extending subjective experiments can be optimized. Comparing to conventional researches mainly
focusing on case studies of hypothetical data sets, we perform this research based on the real perceptual ratings
collected from an existing subjective experiment.
Access to the requested content is limited to institutions that have purchased or subscribe to SPIE eBooks.
You are receiving this notice because your organization may not have SPIE eBooks access.*
*Shibboleth/Open Athens users─please
sign in
to access your institution's subscriptions.
To obtain this item, you may purchase the complete book in print or electronic format on
SPIE.org.
This paper focuses on the influence of ambient light on the perceived quality of videos displayed on Liquid Crystal Display (LCD) with local backlight dimming. A subjective test assessing the quality of videos with two backlight dimming methods and three lighting conditions, i.e. no light, low light level (5 lux) and higher light level (60 lux) was organized to collect subjective data. Results show that participants prefer the method exploiting local dimming possibilities to the conventional full backlight but that this preference varies depending on the ambient light level. The clear preference for one method at the low light conditions decreases at the high ambient light, confirming that the ambient light significantly attenuates the perception of the leakage defect (light leaking through dark pixels). Results are also highly dependent on the content of the sequence, which can modulate the effect of the ambient light from having an important influence on the quality grades to no influence at all.
Access to the requested content is limited to institutions that have purchased or subscribe to SPIE eBooks.
You are receiving this notice because your organization may not have SPIE eBooks access.*
*Shibboleth/Open Athens users─please
sign in
to access your institution's subscriptions.
To obtain this item, you may purchase the complete book in print or electronic format on
SPIE.org.
In this paper, we compare four methods of fusing visible RGB and near-infrared (NIR) images to produce a
color output image, using a psychophysical experiment and image fusion quality metrics. The results of the
psychophysical experiment show that two methods are significantly preferred to the original RGB image, and
therefore RGB-NIR image fusion may be useful for photographic enhancement in those cases. The Spectral Edge
method is the most preferred method, followed by the dehazing method of Schaul et al.
We then investigate image fusion metrics which give results correlated with the psychophysical experiment
results. We extend several existing metrics from 2 to 1 to M to N channel image fusion, as well as introducing
new metrics based on output image colorfulness and contrast, and test them on our experimental data. While
none of the individual metrics gives a ranking of the algorithms which exactly matches that of the psychophysical
experiment, through a combination of two metrics we accurately rank the two leading fusion methods.
Access to the requested content is limited to institutions that have purchased or subscribe to SPIE eBooks.
You are receiving this notice because your organization may not have SPIE eBooks access.*
*Shibboleth/Open Athens users─please
sign in
to access your institution's subscriptions.
To obtain this item, you may purchase the complete book in print or electronic format on
SPIE.org.
Infrared (IR) images are representations of the world and have natural features like images in the visible spectrum.
As such, natural features from infrared images support image quality assessment (IQA).1
In this work, we
compare the quality of a set of indoor and outdoor IR images reconstructed from measurement functions formed
by linear combination of their pixels. The reconstruction methods are: linear discrete cosine transform (DCT)
acquisition, DCT augmented with total variation minimization, and compressive sensing scheme. Peak Signal to
Noise Ratio (PSNR), three full-reference (FR), and four no-reference (NR) IQA measures compute the qualities
of each reconstruction: multi-scale structural similarity (MSSIM), visual information fidelity (VIF), information
fidelity criterion (IFC), sharpness identification based on local phase coherence (LPC-SI), blind/referenceless
image spatial quality evaluator (BRISQUE), naturalness image quality evaluator (NIQE) and gradient singular
value decomposition (GSVD), respectively. Each measure is compared to human scores that were obtained by
differential mean opinion score (DMOS) test. We observe that GSVD has the highest correlation coefficients of
all NR measures, but all FR have better performance. We use MSSIM to compare the reconstruction methods
and we find that CS scheme produces a good-quality IR image, using only 30000 random sub-samples and 1000
DCT coefficients (2%). In contrast, linear DCT provides higher correlation coefficients than CS scheme by using
all the pixels of the image and 31000 DCT (47%) coefficients.
Access to the requested content is limited to institutions that have purchased or subscribe to SPIE eBooks.
You are receiving this notice because your organization may not have SPIE eBooks access.*
*Shibboleth/Open Athens users─please
sign in
to access your institution's subscriptions.
To obtain this item, you may purchase the complete book in print or electronic format on
SPIE.org.
In this article the effects of video content on Quality of Experience (QoE) have been presented. Delivery of the
video content with high level of QoE from bandwidth-limited and error-prone network is of crucial importance
for the service providers. Therefore, it is of fundamental importance to analyse the impact of the network
impairments and video content on perceived quality during the QoE metric design. The major contributions of
the article are in the study of i)the impact of network impairments together with video content, ii) impact of the
video content and ii) the impact of video content related parameters: spatial-temporal perceptual information,
video content, and frame size on QoE has been presented. The results show that when the impact of impairments
on perceived quality is low, the quality is significantly influenced by video content, and video content itself also
has a significant impact on QoE. Finally, the results strengthen the need for new parameter characterization, for
better QoE metric design.
Access to the requested content is limited to institutions that have purchased or subscribe to SPIE eBooks.
You are receiving this notice because your organization may not have SPIE eBooks access.*
*Shibboleth/Open Athens users─please
sign in
to access your institution's subscriptions.
To obtain this item, you may purchase the complete book in print or electronic format on
SPIE.org.
In this investigation we study the effects of compression and frame rate reduction on the performance of four video
analytics (VA) systems utilizing a low complexity scenario, such as the Sterile Zone (SZ). Additionally, we identify the
most influential scene parameters affecting the performance of these systems. The SZ scenario is a scene consisting of a
fence, not to be trespassed, and an area with grass. The VA system needs to alarm when there is an intruder (attack)
entering the scene. The work includes testing of the systems with uncompressed and compressed (using H.264/MPEG-4
AVC at 25 and 5 frames per second) footage, consisting of quantified scene parameters. The scene parameters include
descriptions of scene contrast, camera to subject distance, and attack portrayal. Additional footage, including only
distractions (no attacks) is also investigated. Results have shown that every system has performed differently for each
compression/frame rate level, whilst overall, compression has not adversely affected the performance of the systems.
Frame rate reduction has decreased performance and scene parameters have influenced the behavior of the systems
differently. Most false alarms were triggered with a distraction clip, including abrupt shadows through the fence.
Findings could contribute to the improvement of VA systems.
Access to the requested content is limited to institutions that have purchased or subscribe to SPIE eBooks.
You are receiving this notice because your organization may not have SPIE eBooks access.*
*Shibboleth/Open Athens users─please
sign in
to access your institution's subscriptions.
To obtain this item, you may purchase the complete book in print or electronic format on
SPIE.org.
Ultra High Definition (UHD) is a new technology, which main idea is to improve user’s perception of details and
sensation of immersion in comparison with High Definition systems (HD). However, it is important to understand the
influence of the new UHD technical parameters on user’s perception. Hence, to investigate the influence of the viewing
distance, screen size and scene content on perceived video quality and feelings of users, a series of subjective
experiments with four different contents (3 documentaries and 1 sport content) shooted by UHD camera were performed.
These contents were displayed using three different image resolutions (SD, HD, UHD) and two UHD displays (55-inch
and 84-inch). Each subject had to assess content for three different viewing distances (1.5, 3, 4.5 times of the screen
height corresponding to optimal viewing distances of respectively UHD, HD, and close to SD optimal distance). Finally,
72 test conditions were evaluated. For each scene, observers reported their opinion on the perceived video quality using a
5-grade subjective scale. Results have shown that viewing distance has a significant influence on perceived quality.
Moreover the highest MOS was obtained at optimal viewing for UHD, with a small difference between HD an UHD. At
3H and 4.5H, there is no difference from a statistical point of view. Screen size influences the perception of quality but
not in the same way for the three image resolution and three viewing distances.
Access to the requested content is limited to institutions that have purchased or subscribe to SPIE eBooks.
You are receiving this notice because your organization may not have SPIE eBooks access.*
*Shibboleth/Open Athens users─please
sign in
to access your institution's subscriptions.
To obtain this item, you may purchase the complete book in print or electronic format on
SPIE.org.
A no reference video quality assessment metric based on the region of interest (ROI) was proposed in this paper. In the
metric, objective video quality was evaluated by integrating the quality of two compressed artifacts, i.e. blurring
distortion and blocking distortion. The Gaussian kernel function was used to extract the human density maps of the
H.264 coding videos from the subjective eye tracking data. An objective bottom-up ROI extraction model based on
magnitude discrepancy of discrete wavelet transform between two consecutive frames, center weighted color opponent
model, luminance contrast model and frequency saliency model based on spectral residual was built. Then only the
objective saliency maps were used to compute the objective blurring and blocking quality. The results indicate that the
objective ROI extraction metric has a higher the area under the curve (AUC) value. Comparing with the conventional
video quality assessment metrics which measured all the video quality frames, the metric proposed in this paper not only
decreased the computation complexity, but improved the correlation between subjective mean opinion score (MOS) and
objective scores.
Access to the requested content is limited to institutions that have purchased or subscribe to SPIE eBooks.
You are receiving this notice because your organization may not have SPIE eBooks access.*
*Shibboleth/Open Athens users─please
sign in
to access your institution's subscriptions.
To obtain this item, you may purchase the complete book in print or electronic format on
SPIE.org.
No-reference image quality metrics are of fundamental interest as they can be embedded in practical applications.
The main goal of this paper is to perform a comparative study of seven well known no-reference learning-based
image quality algorithms. To test the performance of these algorithms, three public databases are used. As a
first step, the trial algorithms are compared when no new learning is performed. The second step investigates
how the training set influences the results. The Spearman Rank Ordered Correlation Coefficient (SROCC) is
utilized to measure and compare the performance. In addition, an hypothesis test is conducted to evaluate the
statistical significance of performance of each tested algorithm.
Access to the requested content is limited to institutions that have purchased or subscribe to SPIE eBooks.
You are receiving this notice because your organization may not have SPIE eBooks access.*
*Shibboleth/Open Athens users─please
sign in
to access your institution's subscriptions.
To obtain this item, you may purchase the complete book in print or electronic format on
SPIE.org.
Camera objective characterization methodologies are widely used in the digital camera industry. Most objective
characterization systems rely on a chart with specific patterns, a software algorithm measures a degradation or difference
between the captured image and the chart itself.
The Spatial Frequency Response (SFR) method, which is part of the ISO 122331 standard, is now very commonly used
in the imaging industry, it is a very convenient way to measure a camera Modulation transfer function (MTF). The SFR
algorithm can measure frequencies beyond the Nyquist frequency thanks to super-resolution, so it does provide useful
information on aliasing and can provide modulation for frequencies between half Nyquist and Nyquist on all color
channels of a color sensor with a Bayer pattern. The measurement process relies on a chart that is simple to manufacture:
a straight transition from a bright reflectance to a dark one (black and white for instance), while a sine chart requires
handling precisely shades of gray which can also create all sort of issues with printers that rely on half-toning. However,
no technology can create a perfect edge, so it is important to assess the quality of the chart and understand how it affects
the accuracy of the measurement.
In this article, I describe a protocol to characterize the MTF of a slanted edge chart, using a high-resolution flatbed
scanner. The main idea is to use the RAW output of the scanner as a high-resolution micro-densitometer, since the signal
is linear it is suitable to measure the chart MTF using the SFR algorithm. The scanner needs to be calibrated in
sharpness: the scanner MTF is measured with a calibrated sine chart and inverted to compensate for the modulation loss
from the scanner. Then the true chart MTF is computed. This article compares measured MTF from commercial charts
and charts printed on printers, and also compares how of the contrast of the edge (using different shades of gray) can
affect the chart MTF, then concludes on what distance range and camera resolution the chart can reliably measure the
camera MTF.
Access to the requested content is limited to institutions that have purchased or subscribe to SPIE eBooks.
You are receiving this notice because your organization may not have SPIE eBooks access.*
*Shibboleth/Open Athens users─please
sign in
to access your institution's subscriptions.
To obtain this item, you may purchase the complete book in print or electronic format on
SPIE.org.
In this study, we investigate the extent to which an image-difference metric based on structural similarity can
correlate with human judgment. We introduce a modified version of the recently published iCID metric and
present new results over two large image quality databases. It is particularly noteworthy that the proposed
metric yields a correlation of 0.861 with mean opinion scores on the 2013 version of the renowned Tampere
Image Database, without dedicated parameter optimization.
Access to the requested content is limited to institutions that have purchased or subscribe to SPIE eBooks.
You are receiving this notice because your organization may not have SPIE eBooks access.*
*Shibboleth/Open Athens users─please
sign in
to access your institution's subscriptions.
To obtain this item, you may purchase the complete book in print or electronic format on
SPIE.org.
Many display manufacturers have studied RGBW pixel structure adding a white sub-pixel to RGB LCD and recently
revealed UHD TVs based on novel RGBW LCD. The RGBW LCD has 50% higher white luminance and 25% lower
primary color luminance compared to RGB LCD.
In this paper, the image quality of RGBW and RGB LCD was dealt with. Before evaluating them, TV broadcast video
and IEC-62087 video were analyzed for test video clips. In order to analyze them, a TV reference video from TV
broadcast content in Korea was firstly collected. As a result of TV reference video analysis, RGBW LCD was expected
to improve image quality more because most of colors are distributed around white point and population ratio of
achromatic colors is higher.
RGB, RGBW and RGBW using wide color gamut (WCG) backlight unit (BLU) LCDs were prepared, and a series of
visual assessments were conducted. As a result, RGBW LCD obtained higher scores than RGB LCD about four
attributes (‘Brightness’, ‘Naturalness’, ‘Contrast’ and overall image quality) and ‘Colorfulness’ was not higher score
than RGB LCD in test still images. RGBW LCD’s overall image quality in the TV reference video clips also was
assessed higher than RGB LCD. Additionally, RGBW LCD using WCG BLU shows better performance about
especially ‘Colorfulness’ than RGBW LCD.
Access to the requested content is limited to institutions that have purchased or subscribe to SPIE eBooks.
You are receiving this notice because your organization may not have SPIE eBooks access.*
*Shibboleth/Open Athens users─please
sign in
to access your institution's subscriptions.
To obtain this item, you may purchase the complete book in print or electronic format on
SPIE.org.
The system gamma of the imaging pipeline, defined as the product of the encoding and decoding gammas,
is typically greater than one and is stronger for images viewed with a dark background (e.g. cinema) than
those viewed in lighter conditions (e.g. office displays).1-3 However, for high dynamic range (HDR) images
reproduced on a low dynamic range (LDR) monitor, subjects often prefer a system gamma of less than one,4
presumably reflecting the greater need for histogram equalization in HDR images. In this study we ask subjects
to rate the perceived quality of images presented on a LDR monitor using various levels of system gamma. We
reveal that the optimal system gamma is below one for images with a HDR and approaches or exceeds one for
images with a LDR. Additionally, the highest quality scores occur for images where a system gamma of one is
optimal, suggesting a preference for linearity (where possible). We find that subjective image quality scores can
be predicted by computing the degree of histogram equalization of the lightness distribution. Accordingly, an
optimal, image dependent system gamma can be computed that maximizes perceived image quality.
Access to the requested content is limited to institutions that have purchased or subscribe to SPIE eBooks.
You are receiving this notice because your organization may not have SPIE eBooks access.*
*Shibboleth/Open Athens users─please
sign in
to access your institution's subscriptions.
To obtain this item, you may purchase the complete book in print or electronic format on
SPIE.org.