An Interactive Broadcasting Protocol for Video-on-Demand

Transcription

1 An Interactive Broadcasting Protocol for Video-on-Demand Jehan-François Pâris Department of Computer Science University of Houston Houston, TX Abstract Broadcasting protocols reduce the cost of video-ondemand services by distributing more efficiently videos that are likely to be simultaneously watched by several viewers. Unfortunately, they do not allow the customer to pause, move fast forward or backward while watching a video. We present an interactive pagoda broadcasting protocol that provides these functions at a very reasonable cost. Our protocol is based on the pagoda broadcasting protocol and requires a set-top box buffer large enough to keep in storage all video data until the customer has watched the entire video. As a result, rewind and pause interactions do not require any server intervention. To minimize the bandwidth requirements of fast forward interactions, the server only transmits the segments that are not available on any of the server broadcasting channels. We evaluate the overhead of these fast forward operations through a probabilistic model. Our data indicate that the most costly fast forward operations are those starting at the beginning of the video and jumping to the beginning of the second half of the video while most fast-forward operation taking place during the second half of the video require little or no additional data. 1. Introduction Despite all the attractiveness of the concept, video-ondemand (VOD) [16] has yet to succeed on the marketplace. The main reason for this lack of success is the high cost of providing video services. As a result, VOD cannot yet compete on a price basis with its two more entrenched rivals, namely pay-per-view and videocassette rentals. This situation has led to numerous proposals aiming at reducing the cost of providing video-on-demand (VOD) services. Many, if not most, of these proposals have focused on finding ways to distribute the top ten or twenty so-called hot videos more efficiently. Broadcasting protocols [3] distribute each video according to a fixed schedule that is not affected by the presence or the absence of requests for that video. Hence the number of viewers watching a given video does not affect their bandwidth requirements. With the sole exception of staggered broadcasting, all broadcasting protocols share the common limitation of not offering any interactive action capability. They require the viewers to watch each video in sequence as in a theater. Unlike VCRs, they do not provide controls allowing the viewers to pause the video and interrupt its viewing, to move fast forward or backward (rewind). While staggered broadcasting provides some interactive control capability, it only allows viewers to jump from one staggered stream to another [2]. The sole advantage of this solution is its simplicity. Its major disadvantages are its high bandwidth requirements and its lack of precision: given a video of duration D distributed over n broadcasting channels, staggered broadcasting only allows users to move forward or backward in increments of D/n times units. In addition, reducing the granularity of the jumps is very costly in terms of bandwidth. We present a more flexible, much less expensive solution. With the sole exception of staggered broadcasting, all broadcasting protocols require a set-top box (STB) capable of a) Simultaneously receiving video data from several channels and b) Storing these data in its local buffer until the customer watches them. The amount of data to be stored in the STB buffer varies with each distribution protocol but can be as high as 5 to 6 percent of the video duration. We propose to increase the STB buffer size in such a way that the STB will never have to discard any video data until the customer has watched the entire video. This would allow the STB to handle locally all pause and rewind commands. Fast forward interactions would still require the intervention of the video server and would be handled by a few contingency streams transmitting on demand the missing video data.

2 First Channel S 1 S 1 S 1 S 1 S 1 S 1 Second Channel S 2 S 4 S 2 S 5 S 2 S 4 Third Channel S 3 S 6 S 8 S 3 S 7 S 9 Fourth Channel repeats S 1 to S 14 and S 2 to S 29 Fifth Channel repeats S 15 to S 19 and S 3 to S 49 Sixth Channel repeats S 5 to S 99 Figure 1: How pagoda broadcasting maps 99 segments into six channels. To investigate the feasibility of our approach, we have evaluated the bandwidth overhead caused by adding interactive controls to an existing proactive video distribution protocol, namely, the pagoda broadcasting protocol [13]. We found that the most costly fast forward operations were those starting at the beginning of the video and jumping to the beginning of the second half of the video. Even in this case, transmitting only the missing video data reduces the cost of the operation by 5 percent. The remainder of our paper is organized as follows. Section 2 reviews previous work. Section 3 introduces our interactive pagoda broadcasting protocol and Section 4 contains our analysis of the protocol performance. Finally, Section 5 has our conclusions. 2. Previous Work The simplest video broadcasting protocol is staggered broadcasting [16]. A video broadcast under that protocol is continuously retransmitted over k distinct video channels at equal time intervals. The approach does not necessitate any significant modification to the set-top box (STB) but requires a fairly large number of channels per video to achieve a reasonable waiting time. Consider, for instance, a video that lasts two hours, which happens to be close to the average duration of a feature movie. Guaranteeing a maximum waiting time of five minutes would require starting a new instance of the video every five minutes for a total of 24 full-bandwidth streams. The past five years have seen the development of many more efficient broadcasting protocols [3]. Most of these protocols assume that the client set-top box has enough local storage to store at least one half of each video being watched. We can subdivide these protocols into two groups. The protocols in the first group are based on Viswanathan and Imielinski's pyramid broadcasting protocol [15]. They include Aggarwal, Wolf and Yu s permutation-based pyramid broadcasting protocol [1], Hua and Sheu s skyscraper broadcasting protocol [5] and Juhn and Tseng's fast broadcasting protocol [7]. While these protocols require less than half the bandwidth of staggered broadcasting to guarantee the same maximum waiting time, they cannot match the performance of the protocols based on the harmonic broadcasting (HB) protocol [6, 12]. These protocols divide videos into many equally sized segments and allocate a separate data stream to each segment. In addition, they require the STB to start receiving all these streams when the customer starts watching the first segment of a video. As a result, each segment can be broadcast using the minimum bandwidth required to ensure on time delivery of its data Even though the total bandwidth requirements for HB and its variants are quite small, the multitude of streams these protocols involve complicates the task of the STB's and the servers. Like HB, pagoda broadcasting (PB) [13] uses fixed-size segments. It assigns to each video m video channels whose bandwidths are all equal to the video consumption rate and partitions these m channels into slots of equal duration. The protocol then tries to find the segment mapping that maximizes the number n of segments that can be packed into these m channels while satisfying the condition that segment S k, for 1 k n, is repeated at least once every k slots. Figure 1 shows how the PB protocol maps 99 segments into six channels. Since the segment size is then equal to 1/99 of the duration of the video, the maximum customer waiting time for a two-hour video is 72/99 = 73 seconds. Most research on interactive video-on-demand has focussed on distribution protocols that allocate their video streams in a dynamic fashion. Li et al. proposed in 1996 to use contingent streams to handle interactive VOD operations [9]. More recent work has focused on minimizing the duration of these contingent streams by merging them as soon as possible with other streams [1, 11, 8, 4]. Poon et al. have proposed a single-rate multicast double-rate unicast protocol supporting full VCR functionality [14]. 3. The Interactive Pagoda Broadcasting Protocol The interactive pagoda broadcasting (IPB) protocol is based on the assumption that most customers of a videoon-demand service will watch videos that they had not watched before. Hence these customers will use the

3 pause and rewind controls much more frequently than the fast-forward control. Like the pagoda broadcasting protocol, the IPB protocol assigns to each video m video channels whose bandwidths are all equal to the video consumption rate. It then partitions these m channels into slots of equal duration and assigns to each slot one segment of the video in a way that guarantees that segment S k, for 1 k n, is repeated at least once every k slots. As we mentioned earlier, most broadcasting protocols require a customer STB that can store over one half of each video being watched on their disk drive. Assuming a video distributed in MPEG-2 format with an average bandwidth of 5Megabits/s, this means that 2.25 Gigabytes of storage are needed to store the first hour of a two-hour video. The cheapest way to provide this buffer capacity is to add a disk drive to the STB. Most new disk drives being sold today have capacities of at least seven Gigabytes. A disk drive of that size would allow the STB to keep any previously played segment of the video being watched until the end of that video. As a result, the STB could handle all pause and rewind requests locally without any server intervention. Implementing unrestricted fast forward interactions will require additional bandwidth unless we require each video to be completely downloaded by the customer STB before being viewed. This is not an acceptable solution as it would either result in unacceptable delays for the customer or require an inordinate amount of bandwidth. A better solution would be to allocate to each fast forward request a contingency video stream and use it to transmit to the customer STB the segments that it does not have and will not receive on time from the m broadcasting channels. One possible option would be to let these contingency streams use the extra server bandwidth that must be kept available to handle server disk failures. The next question is to decide whether the server or the STB will determine which segments are to be included in the contingency stream. Making this task the responsibility of the server would require the server to keep track of the state of the storage units of all STBs currently receiving the video. This could be a daunting task for very popular videos and would complicate server recovery. Conversely, giving this responsibility to the STB would prevent the sharing of segments among contingency streams. We propose instead to share this duty between the STB and the server. After each fast-forward interaction, the STB will send to the server a request specifying the video segments it does not have and does not expect to receive on time. When the server receives the request, it checks if some of the requested segments are not already included in one of the existing contingency streams. Having done that, it verifies that it has sufficient bandwidth to satisfy the customer request. If this is the case, the server replies to the STB with a message informing it of the scheduled time slots and channel allocations of all the missing segments. While our solution was strongly influenced by the work of Carter et al. [4], it differs in at least one major point. In order not to increase the size of the STB drive, these authors assume that the video server will handle all interactive commands. We believe our solution will require much less additional bandwidth without significantly increasing the cost of the STB. 4. Analytical Study As we mentioned in the previous section, our IPB protocol lets the STB handle all pause and rewind operations without server intervention. As a result, only fast forward operation can affect the server workload. We will thus focus our analysis on the cost of fast forward operations. Unable to find any reliable data on the frequency and the extents of fast-forward operations, we decided instead to measure directly the average costs of fast forward operations skipping a given amount of video data starting at a given position within the video. To ensure a steady flow of images, the IPB must ensure that the k th segment of video, say segment S k, will be broadcast at least once every k slots. As a first approximation, we can thus assume that there is at least a probability 1/k of finding segment S k in any arbitrary slot. Consider now a viewer watching segment S j and wanting to fast forward to some later segment S k. The probability p k,j that either segment S k is already in the STB buffer or can be received on time by the STB will satisfy the inequality j + 1 p k, j. k Similarly the probability p k+l,j+l that segment S k+l will either be already on the STB buffer or received on time by the STB will satisfy the inequality pk+ l, j+ l. k + l Not transmitting the segments that the client already has or will receive on time from the m broadcasting channels will reduce the average number of segments transmitted by the contingency stream by at least n k = k + l l where n is the number of segments in the video. Hence, the average cost of a fast forward from within segment j to within segment k will never exceed

4 n k (1 ) k + l = = n k n k k j 1 k + l k j 1 l = ( k j 1)( H( n) H( k 1)) where H (n) is the n-th harmonic number. In reality, mapping constraints force the protocol to broadcast some segments slightly more frequently than they should. As shown on Figure 1, segment S 5 is broadcast once every four slots even though it only has to be broadcast every five slots. Similarly, all segments in channel 6 are broadcast once every 3 slots. In general, segment S k, will be broadcast once every s(k) slots with s(k) k. The probability p k,j that either segment S k is already in the STB storage unit or it will be received on time by the STB is then given by j 1 p, j = min(,1). s( k) Similarly the probability p k+l,j+l that segment S k+l will either be already on the STB drive or received on time by the STB is given by pk+ l, j+ l = min(,1). s( k + l) Hence, the average cost of a fast forward from within segment j to within segment k will be equal to k + n k n k (1 min(,1)) = ( n k + 1) min(,1) s( k + l) s( k + l) where n is the number of segments in the video. Define θ as the index of the lowest-numbered segment that is repeated at the same periodicity as the last segment of the video S n : θ = min{ k s( k) = s( n)}. Since s(k) is normally a monotonic non-decreasing function of k, we will have s(k)=s(n) for all θ k n. Since s(θ ) θ, this means that s(k ) θ for all θ k n. Consider now a fast-forward interaction starting from a segment S j such that j θ. At that time, the customer STB would have already received all segments of the video since none of them would have been broadcast less frequently than once every θ slots. Therefore it would not require additional video data from the video server and would not result in any bandwidth overhead. Theorem 1: Any fast-forward interaction starting from a segment S j such that j θ would not cause any bandwidth overhead. Consider, for instance, the case of the IPB protocol with six channels. Segments S 5 to S 99 are all broadcast once every 5 slots, which means that θ = 5. Hence any fast-forward operation taking place during the second half of the video will be handled by the STB without any server intervention. Similarly IPB with five channels broadcasts segments S 3 to S 49 once every 3 slots. Any fast forward operation taking place during the last 4 percent of the video would also be handled without any server intervention. Figure 2 shows the influence of the number of skipped segments on the cost of a fast forward operation for the IPB protocol with 5 channels and 49 segments. Each segment thus represents 2 minutes and 27 seconds of playtime for a two-hour video. The X-axis represents the number of skipped segments during the fast forward operation while the Y-axis represents the number of additional video segments sent by the video server as a result of the fast-forward operation. Each curve corresponds to a different starting point. As we can see, the most expensive fast forward operations start while the customer is watching the first segment of the video and skip 23 segments thus bringing the viewer to the beginning of segment S 25, that is, just before the beginning of the second half of the video. The contingency stream that ensures on time delivery of segments S 25 to S 49 has an average duration of 13 segments, that is, 31 minutes and 5 seconds of data for a two-hour video. Even here, the protocol achieves significant bandwidth savings by excluding from the contingency stream the segments that the STB has already received or can receive on time from the five broadcast channels. A protocol that would always transmit segments S 25 to S 49 would have sent 26 segments, that is, exactly double of what our protocol requires. Fast forward operations taking place later in the video require considerably less bandwidth, mostly because the STB is more likely to have already received the segments or to be able to receive them on time from the five broadcasting channels. For instance, fast forward operations starting while customers watch the 16 th segment of the video never require more than an average of segments, that is, 7 minutes and 26 seconds of data for a twohour video. Fast forward operations starting while customers watch the 24 th segment of the video are almost free. All fast forward operations starting after that require no server intervention.

5 Contigency Stream Duration Starting at 1 Starting at 8 Starting at 16 Starting at Number of Skipped Segments Figure 2: Cost of a fast forward as a function of its starting point and the number of segments skipped for IPB with 5 channels and 49 segments. Contigency Stream Duration Starting from 1 Starting from 8 Starting from 16 Starting from 32 Starting from Number of Skipped Segments Figure 3: Cost of a fast forward as a function of its starting point and the number of segments skipped for IPB with 6 channels and 99 segments.

6 As Figure 3 indicates, similar conclusions hold for the IPB protocol with 6 channels and 99 segments. The most expensive fast forward operations start while the customer is watching the first segment of the video and skip 49 segments thus bringing the viewer to the beginning of segment S 51, that is, slightly after the beginning of the second half of the video. The contingency stream that ensures on time delivery of segments S 51 to S 99 has an average duration of segments, that is, 28 minutes and 31 seconds of data for a two-hour video. This is 48 percent of the 49 segments that a protocol that always transmits segments S 51 to S 99 would have required. Fast forward operations starting while customers watch the 48 th segment of the video are almost free. All fast forward operations starting after the 49 th segment of the video require no server intervention. We need however to point out that these results only hold for fast forward operations that are not preceded by any other interactive request. Fast forward operations taking place after one or more pause or rewind requests will require less bandwidth as the STB will have accumulated more segments in its buffer. On the other hand, cascading fast forward operations, especially when a customer performs incremental fast-forwards throughout an entire video, will require considerably more bandwidth. Our model did not either consider how segment sharing among contingency streams could reduce the average number of segments per contingency stream. We do not believe this impact to be significant as long as fast forward operations remain infrequent. 5. Conclusion Rather than answering individual requests, broadcasting protocols distribute each video according to a fixed schedule that is not affected by the presence or the absence of requests for that video. As a result, they do not provide controls allowing the viewers to pause the video and interrupt its viewing, to move fast forward or backward (rewind). The sole exception is staggered broadcasting, which allows viewers to jump from one staggered stream to another [2]. The biggest disadvantage of this solution is its high cost.. We have presented here an interactive broadcasting protocol that allows the customer greater control at a much lower cost. Our interactive pagoda broadcasting protocol (IPB) is based on the pagoda broadcasting protocol. Unlike the original pagoda broadcasting protocol, the IPB protocol requires a STB buffer large enough to keep in storage all video segments for the whole duration of the video. As a result, the STB can handle all rewind and pause interactions without any server intervention. To minimize the bandwidth requirements of fast forward interactions, the server only transmits the segments that are not already stored in the STB buffer and are not available on any of the VOD broadcasting channels. We have evaluated the actual cost of these fast forward operations through a probabilistic model. We found that the most costly fast forward operations were those starting at the beginning of the video and jumping to the beginning of the second half of the video while most fastforward operation taking place during the second half of the video required little or no additional data. Keeping in mind that rewind and pause operations do not require any server intervention, we can safely conclude that the bandwidth requirements of our IBP protocol will remain reasonable as long as fast forward operations will remain infrequent. Acknowledgments This research was supported by the Texas Advanced Research Program under grant and the National Science Foundation under grant CCR References [1] C. C. Aggarwal, J. L. Wolf, and P. S. Yu, "A permutation-based pyramid broadcasting scheme for video-on-demand systems," In Proceedings of International Conference on Multimedia Computing and Systems, pages , Hiroshima, Japan, June [2] K. C. Almeroth and M. H. Ammar, "The use of multicast delivery to provide a scalable and interactive video-on-demand service," IEEE Journal on Selected Areas in Communications, 14(5): , Aug [3] S. W. Carter, D. D. E. Long and J.-F. Pâris. "Video-on-Demand Broadcasting Protocols," In Multimedia Communications: Directions and Innovations (J. D. Gibson, Ed.), Academic Press, San Diego, 2, pages [4] S. W. Carter, D. D. E Long and J.-F. Pâris, "An efficient implementation of interactive video-ondemand," In Proceedings of the 8 th International Symposium on Modeling, Analysis and Simulation of Computer and Telecommunication Systems, pages , San Francisco, CA, Aug.-Sept. 2. [5] K. A. Hua and S. Sheu, "Skyscraper broadcasting: a new broadcasting scheme for metropolitan videoon-demand systems," In Proceedings of SIGCOMM 97Conference, pages 89 1, Cannes, France, Sep [6] L. Juhn and L. Tseng, "Harmonic broadcasting protocols for video-on-demand service," IEEE

Video-on-demand broadcasting protocols Jukka Leveelahti 17.4.2002 Tik-111.590 Multimedia Communications Motivation Watch any movie at home when ever you like MPEG-2 at least 4 MB per second Too expensive!

Video-on-Demand Nick Caggiano Walter Phillips Video-on-Demand What is Video-on-Demand? Storage, transmission, and display of archived video files in a networked environment Most popularly used to watch

Content storage architectures DAS: Directly Attached Store SAN: Storage Area Network allocates storage resources only to the computer it is attached to network storage provides a common pool of storage

Internet Protocol Television Table of Contents IPTV Definition History IPTV services in the World IPTV - in numbers What Is IPTV TV distribution methods Differences in technology IPTV - Technology IPTV

Current Status of ATSC 3.0 The Next Generation Broadcast Television System Jim Kutzner / PBS Skip Pizzi / NAB February 20, 2013 ATSC Advanced Television Systems Committee Established in the early 1980s

Metadata for Enhanced Electronic Program Guides by Gomer Thomas An increasingly popular feature for TV viewers is an on-screen, interactive, electronic program guide (EPG). The advent of digital television

INTRODUCING THE WORLD S FIRST HEVC H.265 METER & TV ANALYSER Digital terrestrial TV is at the dawn of a new transformation driven by the need to release yet further spectrum in the so called second dividend

Stream Conversion to Support Interactive Playout of Videos in a Client Station Ming-Syan Chen and Dilip D. Kandlur IBM Research Division Thomas J. Watson Research Center Yorktown Heights, New York 10598

Evaluation of SGI Vizserver James E. Fowler NSF Engineering Research Center Mississippi State University A Report Prepared for the High Performance Visualization Center Initiative (HPVCI) March 31, 2000

Development of Media Transport Protocol for 8K Super Hi Vision Satellite roadcasting System Using MMT ASTRACT An ultra-high definition display for 8K Super Hi-Vision is able to present much more information

Video Redundancy A Best-Effort Solution to Network Data Loss by Yanlin Liu A Thesis Submitted to the Faculty of the WORCESTER POLYTECHNIC INSTITUTE in partial fulfillment of the requirements of the Degree

6.UAP Project FunPlayer: A Real-Time Speed-Adjusting Music Accompaniment System Daryl Neubieser May 12, 2016 Abstract: This paper describes my implementation of a variable-speed accompaniment system that

March 7, 2012 # 7379 To media agency executives, media directors and all media committees. POV: Making Sense of Current Local TV Market Measurement This document is intended to raise awareness around the

156 1957 WESTERN COMPUTER PROCEEDINGS The Lincoln TX-2 Input-Output System*, JAMES w. FORGIEt INTRODUCTION THE input-output system of the Lincoln TX-2 computer contains a variety of input-output devices

Data Converters and DSPs Getting Closer to Sensors As the data converters used in military applications must operate faster and at greater resolution, the digital domain is moving closer to the antenna/sensor

Broadcast and media Transmitter systems TV transmitters: the best even better Thanks to their combined features, TV transmitters from Rohde & Schwarz already had a leading position worldwide, but now they

Course Presentation Multimedia Systems Video I (Basics of Analog and Digital Video) Mahdi Amiri April 2011 Sharif University of Technology Video Visual Effect of Motion The visual effect of motion is due

Asynchronous counters In the previous section, we saw a circuit using one J-K flip-flop that counted backward in a two-bit binary sequence, from 11 to 10 to 01 to 00. Since it would be desirable to have

Directional Gossip: Gossip in a Wide Area Network Meng-Jang Lin University of Texas at Austin Department of Electrical and Computer Engineering Austin, TX Keith Marzullo University of California, San Diego

This document is downloaded from DR-NTU, Nanyang Technological University Library, Singapore. Title Deregulation and commercialization of the broadcast media : implications for public service programmers

Schedules General operation of the DT80 data loggers centres on scheduling. Schedules determine when various processes are to occur, and can be triggered by the real time clock, by digital or counter events,