Publications of the Department of Audiovisual Technology

The following list (automatically generated by the University Library) contains the publications from the year 2016. The publications up to the year 2015 can be found on an extra page.

Note: If you want to search through all the publications, select "Show All" and then you can use the browser search with Ctrl+F.

Results: 165
Created on: Thu, 02 May 2024 23:03:32 +0200 in 0.0779 sec


Ramachandra Rao, Rakesh Rao; Göring, Steve; Vogel, Patrick; Pachatz, Nicolas; Villamar Villarreal, Juan Jose; Robitza, Werner; List, Peter; Feiten, Bernhard; Raake, Alexander
Adaptive video streaming with current codecs and formats: extensions to parametric video quality model ITU-T P.1203. - In: Electronic imaging, ISSN 2470-1173, Bd. 31 (2019), 10, art00015, S. 314-1-314-6

Adaptive streaming is fast becoming the most widely used method for video delivery to the end users over the internet. The ITU-T P.1203 standard is the first standardized quality of experience model for audiovisual HTTP-based adaptive streaming. This recommendation has been trained and validated for H.264 and resolutions up to and including full-HD. The paper provides an extension for the existing standardized short-term video quality model mode 0 for new codecs i.e., H.265, VP9 and AV1 and resolutions larger than full-HD (e.g. UHD-1). The extension is based on two subjective video quality tests. In the tests, in total 13 different source contents of 10 seconds each were used. These sources were encoded with resolutions ranging from 360p to 2160p and various quality levels using the H.265, VP9 and AV1 codecs. The subjective results from the two tests were then used to derive a mapping/correction function for P.1203.1 to handle new codecs and resolutions. It should be noted that the standardized model was not re-trained with the new subjective data, instead only a mapping/correction function was derived from the two subjective test results so as to extend the existing standard to the new codecs and resolutions.



https://doi.org/10.2352/ISSN.2470-1173.2019.10.IQSP-314
Göring, Steve; Zebelein, Julian; Wedel, Simon; Keller, Dominik; Raake, Alexander
Analyze and predict the perceptibility of UHD video contents. - In: Electronic imaging, ISSN 2470-1173, Bd. 31 (2019), 12, art00009, S. 215-1-215-6

720p, Full-HD, 4K, 8K, ..., display resolutions are increasing heavily over the past time. However, many video streaming providers are currently streaming videos with a maximum of 4K/UHD-1 resolution. Considering that normal video viewers are enjoying their videos in typical living rooms, where viewing distances are quite large, the question arises if more resolution is even recognizable. In the following paper we will analyze the problem of UHD perceptibility in comparison with lower resolutions. As a first step, we conducted a subjective video test, that focuses on short uncompressed video sequences and compares two different testing methods for pairwise discrimination of two representations of the same source video in different resolutions.We selected an extended stripe method and a temporal switching method. We found that the temporal switching is more suitable to recognize UHD video content. Furthermore, we developed features, that can be used in a machine learning system to predict whether there is a benefit in showing a given video in UHD or not. Evaluating different models based on these features for predicting perceivable differences shows good performance on the available test data. Our implemented system can be used to verify UHD source video material or to optimize streaming applications.



https://doi.org/10.2352/ISSN.2470-1173.2019.12.HVEI-215
Keller, Dominik; Seybold, Tamara; Skowronek, Janto; Raake, Alexander
Assessing texture dimensions and video quality in motion pictures using sensory evaluation techniques. - In: 2019 Eleventh International Conference on Quality of Multimedia Experience (QoMEX), (2019), insges. 6 S.

The quality of images and videos is usually examined with well established subjective tests or instrumental models. These often target content transmitted over the internet, such as streaming or videoconferences and address the human preferential experience. In the area of high-quality motion pictures, however, other factors are relevant. These mostly are not error-related but aimed at the creative image design, which has gained comparatively little attention in image and video quality research. To determine the perceptual dimensions underlying movie-type video quality, we combine sensory evaluation techniques extensively used in food assessment - Degree of Difference test and Free Choice Profiling - with more classical video quality tests. The main goal of this research is to analyze the suitability of sensory evaluation methods for high-quality video assessment. To understand which features in motion pictures are recognizable and critical to quality, we address the example of image texture properties, measuring human perception and preferences with a panel of image-quality experts. To this aim, different capture settings were simulated applying sharpening filters as well as digital and analog noise to exemplary source sequences. The evaluation, involving Multidimensional Scaling, Generalized Procrustes Analysis as well as Internal and External Preference Mapping, identified two separate perceptual dimensions. We conclude that Free Choice Profiling connected with a quality test offers the highest level of insight relative to the needed effort. The combination enables a quantitative quality measurement including an analysis of the underlying perceptual reasons.



https://doi.org/10.1109/QoMEX.2019.8743189
Lebreton, Pierre; Hupont, Isabelle; Hirth, Matthias; Mäki, Toni; Skodras, Evangelos; Schubert, Anton; Raake, Alexander
CrowdWatcher: an open-source platform to catch the eye of the crowd. - In: Quality and user experience, ISSN 2366-0147, Volume 4 (2019), issue 1, article 1, Seite 1-17

https://doi.org/10.1007/s41233-019-0024-6
Fremerey, Stephan; Hofmeyer, Frank; Göring, Steve; Raake, Alexander
Impact of various motion interpolation algorithms on 360˚ video QoE. - In: 2019 Eleventh International Conference on Quality of Multimedia Experience (QoMEX), (2019), insges. 3 S.

https://doi.org/10.1109/QoMEX.2019.8743307
Göring, Steve; Ramachandra Rao, Rakesh Rao; Raake, Alexander
nofu - a lightweight no-reference pixel based video quality model for gaming content. - In: 2019 Eleventh International Conference on Quality of Multimedia Experience (QoMEX), (2019), insges. 6 S.

https://doi.org/10.1109/QoMEX.2019.8743262
Göring, Steve; Raake, Alexander
deimeq - a deep neural network based hybrid no-reference image quality model. - In: Proceedings of the 2018 7th European Workshop on Visual Information Processing (EUVIP), (2018), insges. 6 S.

https://doi.org/10.1109/EUVIP.2018.8611703
Lebreton, Pierre; Fremerey, Stephan; Raake, Alexander
V-BMS360: a video extention to the BMS360 image saliency model. - In: 2018 IEEE International Conference on Multimedia and Expo workshops (ICMEW), ISBN 978-1-5386-4195-8, (2018), insges. 4 S.

https://doi.org/10.1109/ICMEW.2018.8551523
Göring, Steve; Skowronek, Janto; Raake, Alexander
DeViQ - a deep no reference video quality model. - In: Electronic imaging, ISSN 2470-1173, Bd. 30 (2018), 14, art00017, S. 518-1-518-6

When enjoying video streaming services, users expect high video quality in various situations, including mobile phone connections with low bandwidths. Furthermore, the user's interest in consuming new large-size data content, such as high resolution/frame rate material or 360 degree videos, is gaining as well. To deal with such challenges, modern encoders adaptively reduce the size of the transmitted data. This in turn requires automated video quality monitoring solutions to ensure a sufficient quality of the material delivered. We present a no-reference video quality model; a model that does not require the original reference material, which is convenient for application in the field. Our approach uses a pretrained classification DNN in combination with hierarchical sub-image creation, some state-of-the-art features and a random forest model. Furthermore, the model can process UHD content and is trained on a large ground-truth data set, which is generated using a state-of-the-art full-reference model. The proposed model achieved a high quality prediction accuracy, comparable to a number of full-reference metrics. Thus our model is a proof-of-concept for a successful no-reference video quality estimation.



https://doi.org/10.2352/ISSN.2470-1173.2018.14.HVEI-518
Singla, Ashutosh; Robitza, Werner; Raake, Alexander
Comparison of subjective quality evaluation methods for omnidirectional videos with DSIS and modified ACR. - In: Electronic imaging, ISSN 2470-1173, Bd. 30 (2018), 14, art00025, S. 525-1-525-6

In this paper, we compare the Double-Stimulus Impairment Scale (DSIS) and a Modified Absolute Category Rating (M-ACR) subjective quality evaluation method for HEVC/H.265-encoded omnidirectional videos. These two methods differ in the type of rating scale and presentation of stimuli. Results of our test provide insight into the similarities and differences between these two subjective test methods. Also, we investigate whether the results obtained with these subjective test methods are content-dependent. We evaluated subjective quality on an Oculus Rift for two different resolutions (4K and FHD) and at five different bit-rates. Experimental results show that for 4K resolution, for the lower bit-rates at 1 and 2 MBit/s, M-ACR provides slightly higher MOS compared to DSIS. For 4, 8, 15 MBit/s, DSIS provides slightly higher MOS. While the correlation coefficient between these two methods is very high, M-ACR offers a higher statistical reliability than DSIS. We also compared simulator sickness scores and viewing behavior. Experimental results show that subjects are more prone to simulator sickness while evaluating 360˚ videos with the DSIS method.



https://doi.org/10.2352/ISSN.2470-1173.2018.14.HVEI-525