3 types of scalability in multimedia transmission

Upload: theinvinciblesun

Post on 04-Jun-2018

216 views

Category:

Documents


0 download

TRANSCRIPT

  • 8/13/2019 3 Types of Scalability in Multimedia Transmission

    1/8

  • 8/13/2019 3 Types of Scalability in Multimedia Transmission

    2/8

  • 8/13/2019 3 Types of Scalability in Multimedia Transmission

    3/8

    SVC uses multilayer coding technique in order to implement spatial scalable coding. Whilemultilayer coding is also used in older video standards such as mpeg2, coding efficiency ismuch better in h264/avc. This increased efficiency is gained by using several inter-layer

    prediction techniques. A dependency identifier d represents a single layer which is associated

    to a specific spatial resolution. Similar to temporal scalability the value of the identifier isincremented for every layer in the system starting from the base layer where the value is 0.motion-compensated AND INTRA-prediction are used for single layer coding for everylayer int he system.

    Fig. 4.

    Fig. 4. Multilayer structure with additional inter-layer prediction for enabling

    spatial scalable coding.

    As specified in the standard every layer uses the the same coding order which frees up andconserves memory and reduces complexity.

    As seen above where every layer is associated with a dependency identifier and a supportedassociated resolution which for a point in the time/space becomes an access unit. Theseaccess units are transmitted one after the other depending ont he value of D or the dependingindentifier in an increasing fashion.

    It can however be seen that the pictures in the lower layer can be omiited from the acess unitswhich allows us to use spatial and temporal scalability at once.

  • 8/13/2019 3 Types of Scalability in Multimedia Transmission

    4/8

    1) Inter-Layer Prediction: The main goal when designing inter-layer prediction tools is to

    enable the usage of as much lower layer information as possible for improving rate-distortionefciency of the enhancement layers. In H.262 MPEG -2 Video, H.263, and MPEG-4 Visual,the only supported inter- layer prediction methods employs the reconstructed samples of thelower layer signal. The prediction signal is either formed by motion-compensated predictioninside the enhancement layer, by upsampling the reconstructed lower layer signal, or byaveraging such an upsampled signal with a temporal prediction signal.

    Although the reconstructed lower layer samples represent the complete lower layerinformation, they are not necessarily the most suitable data that can be used for inter-layer

    prediction. Usually, the inter-layer predictor has to compete with the temporal predictor, and

    especially for sequences with slow motion and high spatial detail, the temporal predictionsignal mostly represents a better approximation of the original signal than the upsampledlower layer reconstruction. In order to improve the coding efciency for spatial scalablecoding, two additional inter-layer prediction concepts [33] have been added in SVC:

    prediction of macroblock modes and associated motion parameters and prediction of theresidual signal. When neglecting the minor syntax overhead for spatial enhancement layers,the coding efciency of spatial scalable coding should never become worse than that ofsimulcast, since in SVC, all inter-layer prediction mechanisms are switchable.

    An SVC conforming encoder can freely choose between intra and inter-layer prediction basedon the given local signal characteristics. Inter-layer prediction can only take place inside agiven access unit using a layer with a spatial layer identier D less than th e spatial layeridentier of the layer to be predicted. The layer that is employed for inter -layer prediction isalso referred to as reference layer, and it is signaled in the slice header of the enhancementlayer slices. Since the SVC inter-layer prediction concepts include techniques for motion aswell as residual prediction, an encoder should align the temporal prediction structures of allspatial layers. Although the SVC design supports spatial scalability with arbitrary resolutionratios [34], [35], for the sake of simplicity, we restrict our following description of the inter-layer prediction techniques to the case of dyadic spatial scalability, which is characterized by

    a doubling of the picture width and height from one layer to the next. Extensions of theseconcepts will be briey summarized in Section V -B.2.

    a) Inter-Layer Motion Prediction: For spatial enhancement layers, SVC includes anew macroblock type, which is signaled by a syntax element called base modeag.For this macroblock type, only a res idual signal but no additional sideinformation such as intra-prediction modes or motion parameters is transmitted. When

    base mode ag is equal to 1 and the corresponding 8 x 8 block in the reference layerlies inside an intra-coded macroblock, the macroblock is predicted by inter-layerintra-prediction as will be explained in Section V-B.1c. When the reference layermacroblock is inter-coded, the enhancement layer macroblock is also inter-coded. In

  • 8/13/2019 3 Types of Scalability in Multimedia Transmission

    5/8

    that case, the partitioning data of the enhancement layer macroblock together with theassociated reference indexes and motion vectors are derived from the correspondingdata of the co-located 8 x 8 block in the reference layer by so-called inter-layermotion prediction.

    b)

    Inter-Layer Residual Prediction: Inter-layer residual prediction can be employedfor all inter-coded macroblocks regardless whether they are coded using the newlyintroduced SVC macroblock type signaled by the base mode ag or by using any ofthe conventional macroblock types. A ag is added to t he macroblock syntax forspatial enhancement layers, which signals the usage of inter-layer residual prediction.When this residual prediction ag is equal to 1, the residual signal of thecorresponding 8 x 8 submacroblock in the reference layer is block-wise upsampledusing a bilinear lter and used as pre diction for the residual signal of theenhancement layer macroblock, so that only the corresponding difference signal needsto be coded in the enhancement layer. The upsampling of the reference layer residualis done on a transform block basis in order to ensure that no ltering is applied acrosstransformblock boundaries, by which disturbing signal components could begenerated [36]. Fig. 5 illustrates the visual impact of upsampling the residual byltering across block boundary and the block- based ltering in SVC.

    c) Inter-Layer Intra-Prediction: When an enhancement layer macroblock is codedwith base mode ag equal to 1 and the co -located 8 x 8 submacroblock in its referencelayer is intra-coded, the prediction signal of the enhancement layer macroblock is

    obtained by inter-layer intra-prediction, for which the corresponding reconstructedintra-signal of the reference layer is upsampled. For upsampling the luma component,one-dimensional 4- tap FIR lters are applied horizontally and vertically. The chromacomponents are upsampled by using a simple bilinear lter. Filtering is always

    performed across submacroblock boundaries using samples of neighboring intra- blocks. When the neighboring blocks are not intra-coded, the required samples aregenerated by specic border extension algorithms. In this way, it is avoided to re -construct inter-coded macroblocks in the reference layer and thus, so-called single-loop decoding is provided [37], [38], which will be further explained in Section V-

    B.3. To prevent disturbing signal components in the prediction signal, the H.264/AVCdeblocking lter is applied to the reconstructed intra -signal of the reference layer

    before upsampling.

  • 8/13/2019 3 Types of Scalability in Multimedia Transmission

    6/8

    Quality scalability

    Quality scalibilty is also refered to as grain quality scalability.Quality scalabilty is not sodifferent from apatial scalability. If we were to have completely same picture size for both thebase and enhacement layers in spatial scalability we tehn get quality scalability. Spatial

    scalability inherently provides support to implement this alteration to its configuration. Interlayer prediction techniques are used in spatial scalability without up sampling the

    picture. We can obtain a refined version of texture information in spatial acalability if we

    When utilizing inter-layer prediction, a refinement of texture information is typicallyachieved by re-quantizing the residual texture signal in the enhancement layer with a smallerquantization step size relative to that used for the preceding CGS layer. As a specific featureof this configuration, the deblocking of the reference layer intra signal for inter-layer intra

    prediction is omitted. Furthermore, inter-layer intra and residual prediction are directly performed in the transform coefficient domain in order to reduce the decoding complexity.

    The CGS concept only allows a few selected bit rates to be supported in a scalable bit stream.In general, the number of supported rate points is identical to the number of layers. Switching

    between different CGS layers can only be done at defined points in the bit stream.Furthermore, the CGS concept becomes less efficient, when the relative rate difference

    between successive CGS layers gets smaller. Especially for increasing the flexibility of bitstream adaptation and error robustness, but also for improving the coding efficiency for bitstreams that have to provide a variety of bit rates, a variation of the CGS approach, which isalso referred to as medium-grain quality scalability (MGS), is included in the SVC design.The differences to the CGS concept are a modified high-level signalling, which allows a

    switching between different MGS layers in any access unit, and the so-called key pictureconcept, which allows the adjustment of a suitable trade-off between drift and enhancement

  • 8/13/2019 3 Types of Scalability in Multimedia Transmission

    7/8

    layer coding efficiency for hierarchical prediction structures.

    Fig. 8. Various concepts for trading off enhancement layer coding efficiency and drift for packet-based quality scalable coding: (a) base layer only control, (b) enhancement layer onlycontrol, (c) two-loop control, (d) key picture concept of SVC for hierarchical prediction

    structures, where key pictures are marked by the black-framed boxes.

    Drift describes the effect that the motion-compensated prediction loops at encoder anddecoder are not synchronized, e.g., because quality refinement packets are discarded from a bit stream. Fig. 8 illustrates different concepts for trading off enhancement layer codingefficiency and drift for packet-based quality scalable coding.

    Base layer only control: In the mpeg4/ac

    For fine-grain quality scalable (FGS) coding in MPEG-4Visual, the prediction structure was chosen in a way thatdrift is completely omitted. As illustrated in Fig. 8a,motion compensation in MPEG-4 FGS is only performedusing the base layer reconstruction as reference, and thusany loss or modification of a quality refinement packetdoesnt have any impact on the motion compensationloop. The drawback of this approach, however, is that itsignificantly decreases enhancement layer codingefficiency in comparison to single-layer coding. Sinceonly base layer reconstruction signals are used formotion-compensated prediction, the portion of bit ratethat is spent for encoding MPEG-4 FGS enhancementlayers of a picture cannot be exploited for the coding of

    following pictures that use this picture as reference.

  • 8/13/2019 3 Types of Scalability in Multimedia Transmission

    8/8

    Enhancement layer only control: For quality scalablecoding in H.262/MPEG-2 Video, the other extreme caseof possible prediction structures was specified. Here, thereference with the highest available quality is alwaysemployed for motion-compensated prediction as depicted

    in Fig. 8b. This enables highly efficient enhancementlayer coding and ensures low complexity, since only asingle reference picture needs to be stored for each timeinstant. However, any loss of quality refinement packetsresults in a drift that can only be controlled by intraupdates. It should be noted that H.262/MPEG-2 Videodoes not allow partial discarding of quality refinement

    packets inside a video sequence, and thus the drift issuecan be completely avoided in conforming H.262/MPEG-2Video bit streams by controlling the reconstructionquality of both the base and the enhancement layer duringencoding.

    Two-loop control: As an alternative, a concept with twomotion compensation loops as illustrated in Fig. 8c could

    be employed. This concept is similar to spatial scalablecoding as specified in H.262/MPEG-2 Video, H.263, andMPEG-4 Visual. Although the base layer is notinfluenced by packet losses in the enhancement layer, anyloss of a quality refinement packet results in a drift for theenhancement layer reconstruction.

    SVC key picture concept: For MGS coding in SVC analternative approach using so-called key pictures (see Fig.8d) has been introduced. For each picture a flag istransmitted, which signals whether the base qualityreconstruction or the enhancement layer reconstruction ofthe reference pictures is employed for motion-compensated prediction. In order to limit the memoryrequirements, a second syntax element signals whetherthe base quality representation of a picture is additionallyreconstructed and stored in the decoded picture buffer. In

    order to limit the decoding overhead for such key pictures, SVC specifies that motion parameters must notchange between the base and enhancement layerrepresentations of key pictures, and thus also for key

    pictures, the decoding can be done with a single motion-compensation loop.