Jump to content

Media Delivery Index

From Wikipedia, the free encyclopedia

This is an old revision of this page, as edited by Jrib4354 (talk | contribs) at 08:44, 28 November 2008. The present address (URL) is a permanent link to this revision, which may differ significantly from the current revision.

The Media Delivery Index (MDI) is a set of measures that can be used to monitor both the quality of a delivered video stream as well as to show system margin for IPTV systems by providing an accurate measurement of jitter and delay at network level (Internet Protocol, IP), which are the main causes for quality loss. Identifying and quantizing such problems in this kind of networks is key to maintaining high quality video delivery and providing indications that warn system operators with enough advance notice to allow corrective action.[1]

The Media Delivery Index is typically displayed as two numbers separated by a colon: the Delay Factor (DF) and the Media Loss Rate (MLR).[2]

Context

The Media Delivery Index (MDI) may be able to identify problems caused by:

Time distortion

If packets are delayed by the network, some packets arrive in bursts with interpacket delays shorter than when they were transmitted, while others are delayed such that they arrive with greater delay between packets than when they were transmitted from the source (see figure below). This is defined as packet jitter or the time difference between when a packet actually arrives and the expected arrival time.[3]

Time distortion at Ethernet packet switched networks

A receiver displaying the video at its nominal rate must accommodate the varying input stream arrival times by buffering the data arriving early and assuring that there is enough already stored data to face the possible delays in the received data (because of this the buffer is filled before displaying).

Similarly, the network infrastructure (switches, routers,…) uses buffers at each node to avoid packet loss. These buffers must be sized appropriately to handle network congestion.

Packet delays can be caused by multiple facts, among which there are the way traffic is routed through the infrastructure and possible differences between link speeds in the infrastructure.

Moreover, some methods for delivering Quality of Service (QOS) using packet metering algorithms may intentionally hold back packets to meet the quality specifications in the transmission.[4][5] This is achieved by using buffer memories.

Packet loss

Packets may be lost due to buffer overflows or environmental electrical noise that creates corrupted packets. Even small packet loss rates result in a poor video display.

Description

Packet delay variation and packet loss have been shown to be the key characteristics in determining whether a network can transport good quality video. These features are represented as the Delay Factor (DF) and the Media Loss Rate (MLR), and they are combined to produce the Media Delivery Index (MDI), which is displayed as:

DF:MLR

Components

The different components of the Media Delivery Index (MDI) are explained in this section.

Delay Factor (DF)

The Delay Factor is a temporal value given in milliseconds that indicates how much time is required to drain the virtual buffer at the concrete network node and at a specific time. In other words, it is a time value indicating how many milliseconds’ worth of data the buffers must be able to contain in order to eliminate time distortions (jitter).[3]

The Delay Factor is computed as packets arrive at the node and is displayed/recorded at regular intervals (typically one second).

It is calculated as follows:

1. At every packet arrival, the difference between the bytes received and the bytes drained is calculated. This determines the MDI virtual buffer depth:

Δ = |received bytes - drained bytes|

2. Over a time interval, the difference between the minimum and maximum values of Δ is taken and then divided by the media rate:

DF = max(Δ)-min(Δ) / media rate

Maximum acceptable DF[5]: 9-50 ms

Media Loss Rate (MLR)

The Media Loss Rate is the number of media packets lost over a certain time interval (typically one second).[3]

It is computed by subtracting the number of media packets received during an interval from the number of media packets expected during that interval and scaling the value to the chosen time period (typically one second):

MLR = (Packets expected – Packets received) / Interval time in seconds

Maximum acceptable channel zapping MLR[5]: 0

Maximum acceptable average MLR:

  • SDTV: 0.004
  • VOD: 0.004
  • HDTV: 0.0005

Use

Generally, MDI can be used to install, modify or evaluate a video network following the next steps[3]:

  1. Identify, locate, and address any packet loss issues using MLR.
  2. Identify and measure jitter margins using DF.
  3. Establish an infrastructure monitor for both MDI components to analyze any possible scenarios of interest.

Given these results, measures must be taken to provide solutions to the problems found in the network. Some of them are: redefining system specifications, modifying the network components in order to meet the expected quality requirements (or number of users), etc.

Final comments

In summary, the Media Delivery Index is a lightweight and scalable metric for monitoring the effect a delivery network has on video and, ultimately, on end user Quality of Experience (QoE). Its two components predict the IPTV quality received, becoming measures of great importance to the success of IPTV deployments.

References

  1. ^ Florin Hodis (2008). "IPTV challenges and metrics Application Note". EXFO Electro-Optical Engineering Inc.
  2. ^ J. Welch, J. Clark (April 2006). "A Proposed Media Delivery Index (MDI)". Internet Engineering Task Force (IETF)
  3. ^ a b c d IneoQuest. "Media Delivery Index Application Note"
  4. ^ Francisco Palacios (2006). "IPTV testing over DSL". EXFO Electro-Optical Engineering Inc.
  5. ^ a b c Agilent Technologies (2008). "IPTV QoE: Understanding and interpreting MDI values"