The University of Southampton
University of Southampton Institutional Repository

Joint source-channel decoding and unequal error protection aided video transmission

Joint source-channel decoding and unequal error protection aided video transmission
Joint source-channel decoding and unequal error protection aided video transmission
Shannon’s source and channel-coding separation theorem has been the basic philosophy followed by most state-of-the-art wireless transceivers, where all source signals are assumed to have the same importance and are protected equally by the channel encoder. However, due to its assumption of using idealized Gaussian channels, potentially infinite encoding/decoding delay as well as complexity, Shannon’s source and channel-coding theorem is not strictly applicable in practical wireless scenarios. This is because it is almost impossible to remove all the redundancy of the video source, therefore there exits a certain amount of residual correlation. Moreover, reallife applications are often required to transmit source signals associated with unequal importance, such as the base-layer and enhancement-layer of layered video codecs. Hence joint source-channel coding (JSCC) was proposed for wireless scenarios by jointly exploiting the residual redundancy of the video source and the intentionally imposed redundancy of unequal error protection (UEP) techniques.

A video codec suitable for lossless video compression and iterative source-channel decoding (ISCD) is proposed by further developing the Markov random field (MRF) soft source decoder for the transmission of video sequences, rather than still video pictures. More explicitly, we used low complexity frame-differencing (FD) for removing the temporal-domain inter-frame redundancy. The low-complexity FD technique does not rely on power-thirsty motion-vector-based motion-compensation and as its other benefit, it does not require the transmission of motion vectors. However, this results in more residual redundancy than the more complex motion-vector based motion-compensation. Furthermore, variable-length code (VrLC) is used for removing the residual spatial redundancy of the FD signal, while exploiting the correlations amongst the FD pixels within the current frame with the aid of our MRF model based soft-in-soft-out (SISO) module. By adopting the MRF for modelling of the video pixels, we can infer maximum a posteriori (MAP) image estimates from the 2-D spatial redundancy between the video pixels, while simpler models like the Markov chain can only explore the 1-D spatial redundancy. Although the estimation of the MRF parameters is a challenge compared to the 1-D Markov models, we conceived novel extrinsic information transfer (EXIT)-chart-based estimation methods, which are shown to be effective. Moreover, a novel three-stage ISCD structure is proposed, which outperforms the two-stage architecture. Furthermore, we examined the convergence of the three-stage iterative decoding process using 3D EXIT charts. The performance of our system operating both with and without FD is compared to our benchmarker schemes.

In support of inter-layer forward error correction (IL-FEC) coded layered video transmission, we conceived an adaptive truncated HARQ (ATHARQ) scheme for minimizing the video distortion under the constraint of a given total number of transmission time slots. More specifically, we investigated the merits of IL-FEC schemes in the context of truncated HARQ (THARQ) transmission schemes, where the packet scheduling arrangements were carefully designed for exploiting the specific characteristics of each IL-FEC coded packet. Furthermore, we developed a method of on-line optimization for our IL-ATHARQ transmission scheme, in order to find the most appropriate FEC code rate distribution among the video layers that reduced the video distortion. Type-I HARQ relying on Convolutional Codes (CC) was used for simplicity, because our focus was on the design of the scheduling schemes. The performance of our IL-ATHARQ scheme as well as of the rate-optimized IL-ATHARQ scheme using a RSC codec were compared to the benchmarkers using different video sequences, followed by characterizing both the effects of the delay as well as of the channel quality prediction errors on the attainable system performance.

Finally, we conceived an UEP scheme for layered video transmission in the downlink of a visible light communication system, explicitly, we proposed a hierarchical colour-shift keying (HCSK) modulation scheme based on the standardized colourshift keying (CSK), which is capable of conveying inter-dependent layers of video signals. More specifically, we proposed the Type I and Type II HCSK arrangements, which can be flexibly configured by according to the channel quality, video quality, etc., where the Type II constellation allows us to use a wider range of FERs for the higher layers upon varying the constellation-shaping parameter dl. Our simulation results show that Type II HCSK provides a high grade of flexibility in terms of both its system configuration and optimization. Furthermore, we provided a detailed design example for the employment of HCSK transmitting scalable video sources with the aid of a RSC code. An optimisation method was conceived in order to enhance the UEP and to improve the quality of the received video. The performance of our optimised M-HCSK-RSC video transmission system using different HCSK constellation sizes was compared to the relevant benchmarker schemes using different video sequences.
Zhu, Chuan
68faec7d-dfe8-42a6-ab1d-4ea51ce050b4
Zhu, Chuan
68faec7d-dfe8-42a6-ab1d-4ea51ce050b4
Hanzo, Lajos
66e7266f-3066-4fc0-8391-e000acce71a1

Zhu, Chuan (2016) Joint source-channel decoding and unequal error protection aided video transmission. University of Southampton, Faculty of Physical Sciences and Engineering, Doctoral Thesis, 197pp.

Record type: Thesis (Doctoral)

Abstract

Shannon’s source and channel-coding separation theorem has been the basic philosophy followed by most state-of-the-art wireless transceivers, where all source signals are assumed to have the same importance and are protected equally by the channel encoder. However, due to its assumption of using idealized Gaussian channels, potentially infinite encoding/decoding delay as well as complexity, Shannon’s source and channel-coding theorem is not strictly applicable in practical wireless scenarios. This is because it is almost impossible to remove all the redundancy of the video source, therefore there exits a certain amount of residual correlation. Moreover, reallife applications are often required to transmit source signals associated with unequal importance, such as the base-layer and enhancement-layer of layered video codecs. Hence joint source-channel coding (JSCC) was proposed for wireless scenarios by jointly exploiting the residual redundancy of the video source and the intentionally imposed redundancy of unequal error protection (UEP) techniques.

A video codec suitable for lossless video compression and iterative source-channel decoding (ISCD) is proposed by further developing the Markov random field (MRF) soft source decoder for the transmission of video sequences, rather than still video pictures. More explicitly, we used low complexity frame-differencing (FD) for removing the temporal-domain inter-frame redundancy. The low-complexity FD technique does not rely on power-thirsty motion-vector-based motion-compensation and as its other benefit, it does not require the transmission of motion vectors. However, this results in more residual redundancy than the more complex motion-vector based motion-compensation. Furthermore, variable-length code (VrLC) is used for removing the residual spatial redundancy of the FD signal, while exploiting the correlations amongst the FD pixels within the current frame with the aid of our MRF model based soft-in-soft-out (SISO) module. By adopting the MRF for modelling of the video pixels, we can infer maximum a posteriori (MAP) image estimates from the 2-D spatial redundancy between the video pixels, while simpler models like the Markov chain can only explore the 1-D spatial redundancy. Although the estimation of the MRF parameters is a challenge compared to the 1-D Markov models, we conceived novel extrinsic information transfer (EXIT)-chart-based estimation methods, which are shown to be effective. Moreover, a novel three-stage ISCD structure is proposed, which outperforms the two-stage architecture. Furthermore, we examined the convergence of the three-stage iterative decoding process using 3D EXIT charts. The performance of our system operating both with and without FD is compared to our benchmarker schemes.

In support of inter-layer forward error correction (IL-FEC) coded layered video transmission, we conceived an adaptive truncated HARQ (ATHARQ) scheme for minimizing the video distortion under the constraint of a given total number of transmission time slots. More specifically, we investigated the merits of IL-FEC schemes in the context of truncated HARQ (THARQ) transmission schemes, where the packet scheduling arrangements were carefully designed for exploiting the specific characteristics of each IL-FEC coded packet. Furthermore, we developed a method of on-line optimization for our IL-ATHARQ transmission scheme, in order to find the most appropriate FEC code rate distribution among the video layers that reduced the video distortion. Type-I HARQ relying on Convolutional Codes (CC) was used for simplicity, because our focus was on the design of the scheduling schemes. The performance of our IL-ATHARQ scheme as well as of the rate-optimized IL-ATHARQ scheme using a RSC codec were compared to the benchmarkers using different video sequences, followed by characterizing both the effects of the delay as well as of the channel quality prediction errors on the attainable system performance.

Finally, we conceived an UEP scheme for layered video transmission in the downlink of a visible light communication system, explicitly, we proposed a hierarchical colour-shift keying (HCSK) modulation scheme based on the standardized colourshift keying (CSK), which is capable of conveying inter-dependent layers of video signals. More specifically, we proposed the Type I and Type II HCSK arrangements, which can be flexibly configured by according to the channel quality, video quality, etc., where the Type II constellation allows us to use a wider range of FERs for the higher layers upon varying the constellation-shaping parameter dl. Our simulation results show that Type II HCSK provides a high grade of flexibility in terms of both its system configuration and optimization. Furthermore, we provided a detailed design example for the employment of HCSK transmitting scalable video sources with the aid of a RSC code. An optimisation method was conceived in order to enhance the UEP and to improve the quality of the received video. The performance of our optimised M-HCSK-RSC video transmission system using different HCSK constellation sizes was compared to the relevant benchmarker schemes using different video sequences.

Text
Final Thesis.pdf - Other
Available under License University of Southampton Thesis Licence.
Download (16MB)

More information

Published date: 1 July 2016
Organisations: University of Southampton, EEE

Identifiers

Local EPrints ID: 400650
URI: http://eprints.soton.ac.uk/id/eprint/400650
PURE UUID: 5ce40a46-fc9f-4c9d-84de-ec7afa1bfe4c
ORCID for Lajos Hanzo: ORCID iD orcid.org/0000-0002-2636-5214

Catalogue record

Date deposited: 29 Sep 2016 14:22
Last modified: 15 Mar 2024 05:54

Export record

Contributors

Author: Chuan Zhu
Thesis advisor: Lajos Hanzo ORCID iD

Download statistics

Downloads from ePrints over the past year. Other digital versions may also be available to download e.g. from the publisher's website.

View more statistics

Atom RSS 1.0 RSS 2.0

Contact ePrints Soton: eprints@soton.ac.uk

ePrints Soton supports OAI 2.0 with a base URL of http://eprints.soton.ac.uk/cgi/oai2

This repository has been built using EPrints software, developed at the University of Southampton, but available to everyone to use.

We use cookies to ensure that we give you the best experience on our website. If you continue without changing your settings, we will assume that you are happy to receive cookies on the University of Southampton website.

×