First Category Second Category Third Category: The Files Are Compressed and Stored On A Server
First Category Second Category Third Category: The Files Are Compressed and Stored On A Server
First Category Second Category Third Category: The Files Are Compressed and Stored On A Server
1. INTRODUCTION
We can divide audio and video services into three broad categories: streaming stored
audio/video, streaming live audio/video, and interactive audio/video. Streaming means a user can
listen (or watch) the file after the downloading has started.
In the first category, streaming stored audio/video, the files are compressed and stored on a
server. A client downloads the files through the Internet. This is sometimes referred to as on-demand
audio/video. In the second category, streaming live audio/video refers to the broadcasting of radio
and TV programs through the Internet. In the third category, interactive audio/video refers to the use
of the Internet for interactive audio/video applications. A good example of this application is Internet
telephony and Internet teleconferencing.
1. The HTTP client accesses the Web server using a GET message.
2. The information about the metafile comes in the response.
3. The metafile is passed to the media player.
4. The media player uses the URL in the metafile to access the media server to download the file.
Downloading can take place by any protocol that uses UDP.
5. The media server responds.
1. The HTTP client accesses the Web server using a GET message.
2. The information about the metafile comes in the response.
3. The metafile is passed to the media player.
4. The media player sends a SETUP message to create a connection with the media server.
5. The media server responds.
6. The media player sends a PLAY message to start playing (downloading).
7. The audio/video file is downloaded using another protocol that runs over UDP.
8. The connection is broken using the TEARDOWN message.
9. The media server responds.
Time Relationship
Real-time data on a packet-switched network require the preservation of the time relationship
between packets of a session.
But what happens if the packets arrive with different delays? For example, the first packet arrives
at 00:00:01 (1-s delay), the second arrives at 00:00:15 (5-s delay), and the third arrives at 00:00:27 (7-s
delay). If the receiver starts playing the first packet at 00:00:01, it will finish at 00:00:11. However, the
next packet has not yet arrived; it arrives 4 s later. There is a gap between the first and second packets
and between the second and the third as the video is viewed at the remote site. This phenomenon is
called jitter. Jitter is introduced in real-time data by the delay between packets.
Timestamp
One solution to jitter is the use of a timestamp. If each packet has a timestamp that shows the
time it was produced relative to the first (or previous) packet, then the receiver can add this time
to the time at which it starts the playback. Imagine the first packet in the previous example has a
timestamp of 0, the second has a timestamp of 10, and the third a timestamp of 20. If the receiver starts
playing back the first packet at 00:00:08, the second will be played at 00:00:18, and the third at
00:00:28. There are no gaps between the packets.
Playback Buffer
To be able to separate the arrival time from the playback time, we need a buffer to store the data
until they are played back. The buffer is referred to as a playback buffer. In the previous example, the
first bit of the first packet arrives at 00:00:01; the threshold is 7 s, and the playback time is 00:00:08.
The threshold is measured in time units of data. The replay does not start until the time units of data are
equal to the threshold value.
Ordering
We need a sequence number for each packet. The timestamp alone cannot inform the receiver if a
packet is lost.
Multicasting
Multimedia play a primary role in audio and video conferencing. The traffic can be heavy, and the
data are distributed using multicasting methods. Conferencing requires two-way communication
between receivers and senders.
Mixing
If there is more than one source that can send data at the same time (as in a video or audio
conference), the traffic is made of multiple streams. Mixing means combining several streams of
traffic into one stream.
Sender Report:
The sender report is sent periodically by the active senders in a conference to report transmission and reception
statistics for all RTP packets sent during the interval.
Receiver Report:
The receiver report is for passive participants, those that do not send RTP packets. The report informs the sender
and other receivers about the quality of service.
Source Description Message:
The source periodically sends a source description message to give additional information about itself.
Bye Message:
A source sends a bye message to shut down a stream. It allows the source to announce that it is leaving the
conference.
Application-Specific Message
The application-specific message is a packet for an application that wants to use new applications (not defined in
the standard). It allows the definition of a new message type.