A Survey of Energy-Efficient Compression and Communication Techniques For Multimedia in Resource Constrained Systems

Download as pdf or txt
Download as pdf or txt
You are on page 1of 10

IEEE COMMUNICATIONS SURVEYS & TUTORIALS, VOL. 15, NO.

3, THIRD QUARTER 2013

963

A Survey of Energy-Efcient Compression and Communication Techniques for Multimedia in Resource Constrained Systems
Tao Ma, Michael Hempel, Dongming Peng, Member, IEEE, and Hamid Sharif, Senior Member, IEEE

AbstractAdvances in wireless multimedia communication technologies enable new types of pervasive and ubiquitous applications such as mobile health care, environmental monitoring, facility monitoring and trafc surveillance. Among different factors concerned, energy efciency is one of the most challenging issues in multimedia communication due to the resource constraints, and the requirements for high bandwidth and low transmission delay. In this survey, we provide a broad picture of the stateof-the-art energy efcient techniques that have been proposed in wireless multimedia communication for resource-constrained systems such as wireless sensor networks and mobile devices. Following the essential stages required for multimedia communication, we categorize these techniques into two groups: multimedia compression techniques and multimedia transmission techniques. In the rst group, we introduce the state-of-the-art compression algorithms and perform analyses and evaluations on energy efciency in applying these compression algorithms to resource-constrained multimedia transmission systems. In the second group, we will further categorize the energy efcient transmission techniques into two sub-categories according to their different communication architectures. We review both cross-layer communication, including Unequal Error Protection (UEP), and independent-layer communication, focusing on Routing, MAC, and Physical layer protocols. We present the basic problem statement and objectives of these techniques, and survey multiple potential approaches that have been reported in the literature. Our focus in this survey is to provide insight into different research directions to improve energy efciency in wireless multimedia communication protocols for future developments. Index TermsSurvey, Energy Efciency, Multimedia, Compression, Communication, Mobile, Resource Constrained, CrossLayer

I. I NTRODUCTION

IRELESS multimedia communication has found a variety of applications in numerous different elds such as: trafc surveillance, battle eld reconnaissance, security monitoring, health care, etc. Multimedia data, including audio, images and video is typically bandwidth intensive and delaysensitive. These characteristics result in high demands on the communication and computing aspects of these systems, and thus a very high demand on energy resources. For most typical

Manuscript received 30 September 2011; revised 4 May 2012. The authors are with the Department of Computer and Electronics Engineering, University of Nebraska-Lincoln, USA (email: [email protected]. edu, [email protected], [email protected], hsharif@unl notes.unl.edu). Digital Object Identier 10.1109/SURV.2012.060912.00149

applications, the system architecture exhibits severe resourceconstraints. Some of these constraints are a limited energy supply, low CPU speed, and limited memory for data storage. These constraints provide many challenges to provide desired application capabilities. In wireless sensor networks, for example, resource constraint communication is a great challenge. Many research studies on providing an energy-efcient multimedia communication platform have been reported over the years. The basic idea of these energy-efcient techniques is to design and develop new communication methods that provide optimal performance under constrained resources. In this survey, we summarize these techniques and present insights into the stateof-the-art in research activities in this area. A typical multimedia communication system consists of the following steps: 1) Multimedia Data Acquisition 2) Multimedia Compression and Processing 3) Multimedia Communication We focus on two of these categories: compression techniques and communication techniques, as shown in Fig.1. The rst major group of energy efcient techniques in this area is compression. In todays mobile communication system, especially in embedded sensor networks, energy consumption is dominated by the wireless radio. For an example, the eZ430RF sensors radio, the CC2500 chipset [1], consumes about 28mA current in active mode, while its microprocessor, the MSP430 [2], consumes only 270A current in active mode. That represents a 100-fold difference. For example, in an experiment did conducted by Kenneth Barr [3], text les were compressed using the LZW by compression algorithm LZW and then were transmitted with 2.48mMbps for a savreduction in consumed energy by es about 50% energy than compared to transmitting raw text les without compression. Compression of the amount of data to be transmitted and thus keeping the radio in active state, is therefore vital in reducing the energy demand and thus have become an integral component of the multimedia communication system. Existing compression algorithms were developed since the 1980s. In this survey, we will evaluate these compression algorithms from the perspective of suitability for use by energy-constrained multimedia communication systems. We also survey efforts to improve energy efciency of these

1553-877X/13/$31.00 c 2013 IEEE

964

IEEE COMMUNICATIONS SURVEYS & TUTORIALS, VOL. 15, NO. 3, THIRD QUARTER 2013

Fig. 1. A Summary of Energy Efcient Techniques for Multimedia Wireless Transmissions in Resource-Constrained Systems.

compression algorithms, provided for images and video, as since these are the most impactful for the energy efciency. In image compression, we introduce three major compression algorithms: JPEG, SPIHT and JPEG2000. An analysis of them in terms of computational demand, memory requirements, and compression efciency is conducted in order to provide insight into the question of: which one is the most suitable for energy-constrained systems? Additionally, related techniques on improving energy efciency of these compression algorithms are also discussed. In video compression, we introduce H.264 and High-Efciency Video Coding (HEVC). H.264 has been the state-of-the-art in video compression since its standardization in May 2003. HEVC is an emerging video coding standard designed for surpassing H.264 by providing two key approaches: 1) a low-complexity mode with signicant improvements in execution time with marginal compression rate improvements over H.264, and 2) a high-complexity mode with much higher improvement of compression rates over H.264. HEVC is still in its development stage. The rst draft of HEVC is expected to be complete in February 2012. Both H.264 and HEVC have the same core architecture: a complex encoder but simple decoder. This architecture makes them suitable for broadcastbased systems where the base station, serving many clients, is burdened by the complex encoder while each mobile client requires only a simple encoder. By contrast we also present another popular video coding algorithm: Distributed Video Coding (DVC). Its architecture is different from previous two video coding algorithms in that it has a simple encoder but complex decoder. That makes it more applicable in to upstream transmissions such as a camera surveillance network where energy-constrained mobile camera sensors require a simple encoder while the base station can handle the high computational requirements of the decoder. Related energy efcient techniques proposed for these three video algorithms are also summarized in this paper. The second major focus of the energy efcient techniques is communication. Multimedia transmission has several unique characteristics: unequal importance, error tolerance, and error propagation. These characteristics set multimedia transmission apart from traditional data transmission. In designing a cross-layer communication protocol suitable for multimedia, UEP was proposed. This strategy has been a

key research topic since the 1990s. Many articles have been published on UEP. In this survey, we provide the basic problem statement covering all these efforts, and briey summarize all solutions proposed in the literature. We provide insights on: 1) how well these solutions address this problem; and 2) the currently most effective solution. In addition, we survey one of the most important communication aspects in UEP, channel coding, and describe suitable channel codes proposed for multimedia. In addition to UEP, another cross-layer communication approach called Joint Compression and Communication Optimization is also reviewed. This approach considers both compression and communication together to optimize the energy consumption for a given transmission quality, instead of only communication-matching found in the UEP approach. The rest of this survey is organized as follows. Section II and Section III analyze energy efciency of compression techniques for images and video, respectively. Section IV reviews two cross-layer techniques: Unequal Error Protection and Joint Compression and Communication Optimization. Section V gives the open areas in the future work. Finally, our concluding remarks are given in Section VI. II. E NERGY-E FFICIENT I MAGE C OMPRESSION Digital images typically contain large redundancies. With todays compression techniques, a high-quality compressed image can usually occupy 1/8 the size of the original image. This is very favorable for energy savings in wireless communication since only a small portion of the image data needs to be transmitted. In this section, we compare existing image compression algorithms in terms of energy efciency and compression efciency, and give insights into their performance when these compression algorithms are applied to wireless communication systems. In addition, a survey of research efforts to improve energy efciency in image compression is also presented. Many image compression algorithms have been proposed in last 20 years. The basic idea of these algorithms is to reduce the high correlation in the image data so that any redundancy can be removed. Fig. 2 shows the basic concept of these algorithms. In general, most current image compression algorithms can be categorized into Discrete Cosine Transform (DCT)-based

MA et al.: ENERGY-EFFICIENT COMPRESSION AND COMMUNICATION TECHNIQUES FOR MULTIMEDIA IN RESOURCE CONSTRAINED SYSTEMS

965

TABLE I COMPUTATION COST COMPARISON OF JPEG IMPLEMENTATIONS. Methods Fig. 2. Basic Compression Step. Float Point (IJG) Slow INT (IJG) Adaptive INT(slow) Fast INT(IJG) Adaptive INT(fast) Executing Time 103.12 6.05 4.80 3.25 1.39 Energy 2268.60 133.15 105.50 71.40 30.67 PSNR (Goldhill) 27.0 23.8 26.1 27.0 27.0

compression and Discrete Wavelet Transform (DWT)-based compression. A. DCT-based compression JPEG compression is the most representative DCT-based compression algorithm. It is the most popular and widely used compression algorithm for images. In JPEG compression, the image source is rst partitioned into blocks of 88 pixels and each block is coded independently [4]. After the Discrete Cosine Transform (DCT), each block of 64 DCT coefcients is uniformly quantized by a quantization table. After that, all of the quantized coefcients are ordered into a zigzag sequence, such that low-frequency coefcients are to be processed prior to high-frequency coefcients. The DCT transform is very fast. Furthermore, the independent nature of the 88 block mechanism results in very low memory requirements for the JPEG algorithm. Therefore, the whole compression system is very fast and energy-efcient. The drawback of JPEG is the low compression quality. Despite that, it is still the most widely used compression in commercial products. JPEG consists of a large number of additions and multiplications involving real numbers. In processor architectures operations using oat data types are expensive and energy consuming by comparison. Most hardware platforms, instead of having dedicated oating point hardware units, rely on simple and low-cost xed-point arithmetic units to emulate oating-point operations. However, emulating oating point operations is slow and energy-inefcient. According to current studies, the precision of oating-point operations is often far greater than is necessary in light of quantization process in JPEG [5]. Some research efforts utilize this effect to improve computational energy efciency. Table 1 provides an overview of such efforts and their results. In Table 1, the oating point code and the fast INT (integer) code are from the Independent JPEG Group (IJG) library [6]. Adaptive INT codecs are from paper [5]. The quantization table is selected as JPEGs Table 50. As the results show, the oating point implementation has the highest computational load, but it has the same compression quality as the Slow INT codec. This fact demonstrates the oating point algorithm is unnecessarily precise due to subsequent quantization. The Fast INT and Slow INT codec from IJG use constant bit-width integers to emulate the oating point operation. They have capabilities to adapt to changing image quality requirements, however. In Adaptive INT [5], compression computation range and precision are analyzed in advance. First, the required integer bit-width to completely represent the signal is computed. Second, the precision analysis is conducted with the given quality requirement and the minimal required precision in terms of the fractional bit-width. Finally, according to calculated integer bit-width and fractional bit-width, the most

energy efcient compression code requiring minimal CPU cycles is selected. Thus, computational energy consumption is signicantly reduced compared to constant bit-width integer calculations. B. DWT-based compression Discrete Cosine Wavelet (DWT) is a very powerful transform that provides signicantly higher compression efciency than DCT at higher compression ratio. However, DWT is also more complex and computationally intensive than DCT. Thus, several research activities focus on reducing complexity and required memory. Among them, the lifting scheme [7-8] is one of the best algorithms to calculate the wavelet transform in an energy efcient manner. Compared to a convolution lter, the lifting scheme uses fewer lter coefcients; therefore it provides a faster implementation of DWT. Lifting wavelet also provides memory reduction through in-place computation of wavelet coefcients. In order to further reduce energy consumption in wavelet transform, papers such as [9-10] propose an approach of skipping the computation for the coefcients in high-pass lters in the wavelet transform. Their technique is inspired by the observation that high-pass lter output coefcients are of very low value, typically. According to their studies in image lena 512x512 80% of the high-pass coefcients for wavelet transform level 1 octave are less than 5, claiming that setting the high-pass coefcients to zero has minimal quality impact. By skipping the high pass lter coefcient computation, they can trade-off the reconstruction quality with computational energy consumption. Their experimental results show that 90% of energy savings can be achieved over adaptive wavelet image compression at a low PSNR of 25db. In DWT-based compression, the two best-known compression algorithms are Set Partitioning in Hierarchical Trees (SPIHT) and JPEG2000. In SPIHT, an image is rst transformed into the wavelet domain where the quad-tree structure is explored by a bitplane coding algorithm. The inter-band wavelet correlation is reduced by two pass lters: the sorting pass and the renement pass. The output of these two lters is a bit-stream with descending importance order. This compression is very fast and has high compression quality. In addition, many improvements have been made by many researchers over the traditional SPIHT image compression algorithm. In the SPIHT algorithm three lists are used to temporarily store the images quad-tree structure and signicance information. This requires a large amount of memory and causes hardware implementation difculties.

966

IEEE COMMUNICATIONS SURVEYS & TUTORIALS, VOL. 15, NO. 3, THIRD QUARTER 2013

Listless Zerotree Coding (LZC) [11] is proposed to overcome this drawback of SPIHT. LZC uses two ag maps to store coding information instead of three sets in SPIHT, which greatly reduces memory requirements. However, LZC uses the depth-rst searching strategy which lowers compression performance of the codec. Subsequently, this compression quality compromise problem was addressed in [12-13]. A recent paper [14] proposed an improved version of the SPIHT codec which surpassed the performance of traditional SPIHT without adding any overhead. In their codec, the PSNR is improved by 0.2 to 0.4dB over a broad compression rate range. That makes SPIHT without arithmetic coding competitive even with JPEG2000. Because of all these efforts, SPHIT has become one of the best choices for energy-efcient compression algorithms. JPEG2000 is another well-know DWT based compression algorithm. In JPEG2000, an image is divided into tiles and each tile is transformed independently. After that, the subbands of the tile are divided into packet partition locations. Finally, each packet partition location is divided into code-blocks. JPEG2000 adopts the Embedded Block Coding with Optimal Truncation (EBCOT) image compression technique [15]. In EBCOT, information coding and information ordering are separated. It is referred to as two-tier coding. In tier 1, each of these code blocks carries out a context-dependent, binary arithmetic coding of bit-planes [16]. An embedded bit stream is then generated for every code block. However, since the subbands are partitioned and are encoded independently, intrasubband redundancy is explored instead of inter-subbands in SPIHT. The tier-2 coder operates on the compressed information of every block to arrange their contributions in each quality layer, in a process referred to as packetization [17]. At the end of the tier-2 process, a compressed bit stream is obtained and ready for transmission. JPEG2000 includes comprehensive optimizations and features. The quantization step size can be adjusted by the rate-control mechanism. Also each code block bit stream is truncated in an optimal way to minimize distortion due to bit-rate constraints. The features of JPEG2000 include SNR scalability, resolution scalability as well as random access. JPEG2000 compression provides slightly higher compression quality than SPIHT. However, its multi-layer coding procedure is very complex and computationally intensive. Also, the optimization processes such as optimal truncation as well as resolution features greatly add to the computational requirements. The demand for multiple coding tables for arithmetic coding requires extra memory allocation which makes hardware implementations of the coder more complex and expensive. Based on our extensive studies on these three famous compression algorithms, we list their strengths in terms of memory requirement, compression quality, coding speed and computational load in Table II. These ndings are corroborated in [18]. Given the above information, we can conclude that SPIHT currently is the most energy-efcient and applicable compression algorithm for image-based wireless communication systems, since it provides a high compression ratio with relatively low complexity. JPEG2000 is typically too complex

TABLE II COMPUTATION COST COMPARISON OF COMPRESSION ALGORITHMS. Methods Forward Transform Computation Load Memory Reuquirement Coding Speed Compression Quality JPEG DCT low low high low SPIHT DWT low low high high JPEG2000 DWT high high average high

to apply in wireless sensor networks due to the sensor nodes limited computational ability. JPEG is not as energy efcient as SPIHT, but is a reasonable alternative when the image transmission quality requirement is not very high. JPEG2000 is a well-established image compression algorithm. Signicant research efforts have been conducted in improving its compression efciency. However, reducing the complexity and computational overhead of JPEG2000 is still an open research area.

III. E NERGY E FFICIENT V IDEO C OMPRESSION Video is another very important type of multimedia. It, however, generates a large amount of data for transmission and processing, and storage, as well as exhibits high bandwidth transmission requirements. Thus, it is a challenging task to transmit video in resource-constrained systems. Compression is obviously a major required step before transmission since it eliminates a large amount of data from having to be transmitted. In this section, the state-of-the-art in video coding algorithms is investigated. Energy efciency strengths of these video coding algorithms are analyzed for different application scenarios. A. H.264 For the last two decades, standardization efforts in MPEG and ITU-T have played a key role for traditional broadcasting paradigms in wireless video communication. Subsequently, the H.264/Advanced Video Coding (AVC) standard, completed in May 2003, represents the most efcient video codec available to date. In traditional broadcast architectures, the most computational complexity is concentrated at the encoder, while the decoder is signicantly less complex and thus exhibits far lower processing requirements. For typical implementations, the encoder is 5-10 times more complex than the decoder [19]. In H.264, the encoder includes processes of rate-distortion optimized motion estimation (ME) and mode decision. These processes are very complex and energy consuming. Thus, most research efforts in saving energy in H.264 compression are focused on these two processes. The methods can be categorized into two groups: encoding procedure simplication and complexity reduction in mode prediction. 1) Encoding procedure simplication: H.264 is a highly complex compression approach. For example, the intraprediction for 4x4 luminance blocks has to compute nine

MA et al.: ENERGY-EFFICIENT COMPRESSION AND COMMUNICATION TECHNIQUES FOR MULTIMEDIA IN RESOURCE CONSTRAINED SYSTEMS

967

different prediction modes to obtain the best prediction performance. To simplify this process, Wang et al. [20] developed a fast algorithm by using the property of linear transform and the xed-spatial relationship of predicted pixels in each intra mode. About 50% computational burden can be reduced by using the fast Sum of Absolute Transformed Differences (SATD) algorithm without a degradation in video quality. Furthermore, they propose a two-stage simplied SATD method in which the Sum of Absolute Transform (SAD) criterion is used to eliminate unwanted modes before fast SATD computation. Their experimental results demonstrate that about 70% computational cost can be saved with negligible PSNR loss for H.264 intra prediction. Andersen et al. [21] present a simple operational method that can trade-off between rate-distortion and complexity. The complexity is measured by the number of visited search positions weighted by the size of each block type. This operational method provides a general framework for realtime implementation of H.264 encoders in computationally constrained platforms. Results show a speed-up by a factor of 4 is achievable at only 1% increase in rate. 2) Reducing complexity in mode prediction: Hsia et al. [22] propose a fast intra prediction mode decision algorithm for 4x4 intra blocks. This algorithm reduces the computational complexity of the 4x4 block intra-prediction based on partial sampling prediction and symmetry of adjacent angle modes. Experimental results show that the proposed algorithm can reduce computation time of intra prediction by more than 60% while maintaining similar video quality and bit rate. Kannangara et al. [23] present a complexity reduction algorithm for H.264 encoders. The likely to be skipped macroblocks (MBs) are identied before motion estimation. Hence, the further computational processing of these MBs is also saved. Their new algorithm is based on a well-known rate-distortion optimization method and enables an H.264 encoder to decide whether to encode or to skip each MB. They claim that their algorithm can achieve computational savings of 19%-67% with no signicant loss of rate-distortion performance. Crecos et al. [24] show that the accurate detection of the skip mode type without performing the computationally intensive mode decision is highly desirable for computational speed-ups. By using a low-impact set of smoothness constraints and neighborhood information as well as a set of skip mode conditions, they propose an inter-mode decision scheme for P slices in the H.264 video coding standard. They achieve 35%-58% reduction in run time and a 33%-55% reduction in CPU cycles for both the rate-controlled and the non-ratecontrolled versions of H.264. Based on the analysis in paper [25], the entire computation for mode selection can be reduced signicantly if skipping calculation for less probable modes. In their work, an algorithm for fast coding mode selection in H.264/AVC encoders is proposed. Extensive experiments show that the proposed methods can reduce execution time of mode selection in H.264/AVC by 85% on average while the average PSNR loss is only 0.07dB. Kim et al. [26] propose a fast inter-mode determination algorithm on the basis of the rate-distortion cost of the tracked

MB for the current MB. They verify the performance of the proposed scheme through comparative analysis of experimental results using JM reference software. An improvement of 40%-70% with an average loss of 0.052dB and bit increment of 0.7%-0.8% is achieved. B. Distributed Video Coding Unfortunately, H.264 is not suitable for upstream video applications, such as low-power multimedia sensor networks and wireless video surveillance cameras, because these applications require simple encoders instead of simple decoders. In these upstream applications, lightweight encoding with high compression efciency is needed. Wyner and Ziv [27] established an information theory for lossy coding, which claims efcient compression can be achieved by leveraging knowledge of the source statistics at the decoder only. Based on this theory, a new coding concept called Distributed Video Coding (DVC) was presented. Different from traditional coding, DVC relies on a new statistical framework instead of the deterministic approach of conventional coding techniques. It changes the traditional codec to a simple encoder and complex decoder architecture. Following this direction, the seminal DVC schemes have been proposed in [28-29], followed by a large amount of related publications such [30-31]. In this survey, we only focus on the low complexity and energy saving aspects of DVC. To our knowledge, few publications are reported in this area. Areia et al. [32] propose a hybrid rate control solution to reduce the high WZ decoding complexity for Stanford architecture under constraints of 1) the added complexity at encoder is negligible; 2) the overall rate-distortion is negligible. Their results show that coding complexity reductions of up to 70 % may be possible for more complex sequences. Belief Propagation (BP) is one of the most signicant approximate algorithms available for decoders. It helps gather information from neighborhoods of each pixel in an image to nd the minimum matching cost of local points and their neighborhoods. Paper [33] examines the parallelisms in the BP method for stereo vision on multi-core processors. Their results show their BP method can achieve a 13.5-times speedup compared to single-processor implementations. Papers [34] and [35] propose a shufed belief propagation decoder of low-density parity-check codes (LDPC). Their results show that 10 iterations of the proposed approach with four sub-decoders achieve the same performance as 70 iterations of the standard BP decoder. C. Emerging HEVC standard In January 2010, ISO-IEC/MPEG and ITU-T/VCEG reached an agreement to form the joint collaborative team on video coding (JCT-VC) to develop a next-generation codec: High Efciency Video Coding (HEVC) [36]. HEVC aims to develop: 1) a low-complexity mode with signicant improvements in execution time with marginal compression rate improvements over H.264, and 2) a high-complexity mode with much higher improvement of compression rates over H.264. The initial draft is expected to be released in February 2012. Although HEVC is still under development,

968

IEEE COMMUNICATIONS SURVEYS & TUTORIALS, VOL. 15, NO. 3, THIRD QUARTER 2013

there are some subjective results [37, 35] already showing that it achieves similar visual quality to H.264/AVC with around 30% bit rate reduction for low-delay applications, and with around 20% bit rate reduction for random access experiments on average, but with lower complexity than the H.264/AVC baseline prole. Currently, HEVC is expected to surpass H.264 in terms of energy efciency at low complexity operation. In summary we can claim that: 1) The current H.264 standard is well-suited for downstream broadcast communication where decoders in mobile terminals are simple while the central encoders exhibit the highest complexity burden 2) DVC is highly preferred in multimedia sensor networks and surveillance camera system for its capability of reversing the complexity burden between encoder and decoder. 3) The currently under development HEVC standard is expected to provide substantial improvements in coding efciency over H.264.

IV. C ROSS -L AYER T ECHNIQUES FOR M ULTIMEDIA C OMMUNICATION In this section, we focus on two well-known techniques for wireless multimedia communication: Unequal Error Protection (UEP) and Joint Compression and Communication Optimization. The basic idea of UEP is to apply unequal communication protection to different parts of the compressed multimedia data in order to maximize the performance of the overall multimedia communication. In UEP technique, we provide the idea and theoretical problem statement of this technique. We then survey the state-of-the-art solutions for this problem. Thirdly, we also summarize the channel code design for this problem presented in the current literature. Also we describe the idea behind joint compression and communication optimization in this section. A. UEP Optimization Problem Due to the fact that multimedia information contains signicant amounts of redundancy, to efciently utilize limited resources data compression is a necessity. Compared with general data, however, compressed multimedia information has three distinct characteristics. 1) Unequal Importance: Different parts of the compressed data hold different importance for perceptually and structurally reconstructing the original data during decompression. 2) Error Tolerance: Multimedia data can be recovered with some tolerable degradation even though errors exist. 3) Error Propagation: If some bits are corrupted, the bits which depend on them become useless. Papers [39-40] have extensively looked into these unequal importance properties of compressed multimedia. The studies show that distortion reduction increases nonlinearly as the bitplane count increases. These studies suggest that a potential performance gain can be achieved when the UEP communication technique is applied accordingly.

This problem is often regarded as a UEP optimization problem. Over the past 10 years, many research activities have been reported in this area. Lu et al. [41] present an efcient video transmission approach which minimizes energy consumption under expected video quality constraints by jointly selecting scalable enhancement layers, channel transmission rate adaptation and forward error correction (FEC) error protection. Their experiment shows 5% to 30% energy can be saved while achieving the expected video distortion. Wei et al. [42] propose an energy-efcient JPEG2000 image transmission system over point-to-point wireless sensor networks. In their work, energy consumption is saved by jointly adjusting source coding schemes, channel coding rates, and transmitter power levels in an optimal way. An investigation of the use of an energy-constrained cross-layer approach for the transmission of progressively coded images over packetbased wireless channels is presented in [43]. An optimum power allocation algorithm is designed to enable unequal error protection of a pre-encoded image. Their experimental results demonstrate that it is possible to achieve a relevant performance enhancement with the proposed approach over uniform error protection. In addition, by jointly considering source coding, error concealment, and transmission power management at the physical layer, Wang et al. [44] propose an unequal error protection for the shape and texture information that minimizes the expected distortion at the receiver, given energy and delay constraints. Experimental results of their work indicate that the proposed unequal error protection scheme signicantly outperforms the equal error protection method. To sum up, the key idea of all these works is to match communication strategies to the characteristic of the compressed multimedia by the UEP technique in order to optimize overall performance of multimedia communication. The problem statement of this technique can thus be summarized as: 1) Given a distortion constraint, to minimize the energy consumption. 2) Given an energy consumption constraint, to minimize distortion. These two problems are actually the same problem in terms of the mathematical solution. For simplicitys sake, we do not distinguish them in this paper, and only formulate the rst problems solution since it equally applies to the second problem statement as well. The rst problem statement can be presented as follows: Assume that the compressed multimedia information has N layers. Let Si denote a particular communication strategy, with parameters such as coding algorithms, transmit powers, etc., applying in the layer i of the multimedia data. Let D0 denote the overall distortion when no layer is received by the decoder.GSi is the expected gain (distortion reduction) when the layer i is protected by strategy Si . Assume the distortion metric is additive, then the overall expected gain can be calculated as
N

GS =
i=1

G Si ,

(1)

The expected distortion is DS = D0 GS , (2)

MA et al.: ENERGY-EFFICIENT COMPRESSION AND COMMUNICATION TECHNIQUES FOR MULTIMEDIA IN RESOURCE CONSTRAINED SYSTEMS

969

Energy consumption is the sum of energy consumption of every layer.


N

ES =
i=1

E Si ,

(3)

The problem then can be formulated as a constrained discrete optimization problem arg min : E S
S

s.t. DS < Dmax

(4)

Finding an optimal solution to the above optimization problem is difcult because deriving analytical expressions for distortion and energy consumption as functions of communication strategies is very challenging, since these functions are nondeterministic (only worst case or average values can be determined) and nonlinear. B. Solutions Several optimization algorithms have been proposed to solve this problem. However, the dominant solution for this problem is dynamic programming. 1) Dynamic Programming: The rst and seminal dynamic programming solution is proposed by Chande et al [47]. In their work, two solution-space constraint optimization problems and their solutions were presented within the framework of dynamic programming: 1) the distortion-based optimal unequal protection, 2) the rate-based optimal unequal protection . Their dynamic solutions have the complexity of O(N 2 ) and O(N ) for these two problems, respectively. N is the number of transmitted packets. Later on, Stankovic et al. [48] found a fast linear-complexity algorithm to maximize the expected error-free source rate. Hamzaoui et al claim that the distortion-optimization problem can be solved with a local search starting from the rate-optimal solution [49, 50]. Thus, they combine the rate-optimal solution in [46, 47] with a local search algorithm to solve the distortionoptimization problem. This way, the complexity is reduced to a linear complexity. To the best of our knowledge, it is one of the fastest algorithms known so far for this problem. Another fast algorithm is reported in paper [51]. In their work, a method to nd an unequal error protection scheme for progressively encoded image data based on dichotomic search is presented. Their results have shown that this optimization method achieves an average quality that is virtually equivalent to that obtained by the linear search method. 2) Other Solution: a) Discrete Ergodic Search: Because the solution of the optimization problem is discrete and limited, some simple cases exhibit a small search space for possible strategies S. Therefore, discrete ergodic search would be very direct, accurate and simple. One solution proposed by paper [42] uses the convex concept to nd the optimal point in each layer. Only the optimal points whose distortion constraints are satised are collected. Complexity of the discrete ergodic search is directly proportional to O(N L), where L is the number of strategies and N is the number of layers. In paper [40], the number of strategies is only three. In this case, the discrete ergodic search is also adopted as their solution.

b) Theoretical Solution: Some researchers are trying to nd a closed form expression for this optimization problem. For instance, a closed mathematical expression for the optimal power allocation with various packets is formulated in paper [43]. However, their mathematical closed-form expression is heavily dependent on the theoretical model they assume. These theoretical models are usually too simplistic to be accurate in practice. In their solution, an erasure channel is assumed, and they only consider the slow fading channel with additive white Gaussian noise, which rarely is applicable. Still, their work represents an improvement of UEP performance and provides insight into potential future research in this direction. To sum up, the theoretical solution approach is typically very fast but is typically not applicable to practical use. Discrete ergodic search is simple to implement and is suitable for limited search spaces only. For the large search spaces in selecting a potential strategy S , especially for complex video transmissions, this problem quickly becomes very computationally intensive. Dynamic programming therefore is the best choice in this case. Despite all these research efforts, the optimization problem for UEP is still a challenging and difcult problem and will continue to attract the attention of researchers. C. Channel Codes for Cross-Layer Optimization The other important research issue in the cross-layer optimization problem is channel code design. The basic idea of this optimization problem is to match communication strategies to the multimedia datas structure. Channel codes which provide a wide selection of error protection capability thus become a key aspect of this matching problem. In past decades, many channel codes have been proposed in the literature. Among them are rate-compatible codes (RC), where the parity bits for higher-rate codes are embedded in those of lower-rate codes and all codes in the sequence can be encoded and decoded using a single encoder/decoder pair. This is a very attractive and widely used channel code. One of the more important RC codes is the Rate-Compatible Punctured Convolutional (RCPC) code which is a simple and easy to construct UEP code for different parts of an embedded code stream. It is among the most widely applied codes for this cross-layer optimization problem [52-55]. The other important RC code is the Rate-Compatible Punctured Turbo Code. The Punctured Turbo Code [56] provides stronger error protection than the RCPC code. However, the block size of turbo codes is typically larger than that of the RCPC code. Related reports can be found in paper [57]. Rate-Compatible Low-Density Parity Check (RCLDPC) code is another popular channel code used in this optimization problem [58-60]. It has some advantages over RCPC code, namely 1) it is parallelizable and can be realized at much faster speed than the turbo decoder and 2) almost all the errors are detectable. It also shows better performance than the RCTC code. However, the complexity is still the primary weakness of this code. To sum up, RCPC is the simplest and fastest channel code amongst them. Other channel codes such as RCTC and

970

IEEE COMMUNICATIONS SURVEYS & TUTORIALS, VOL. 15, NO. 3, THIRD QUARTER 2013

RCLDPC give better performance while suffering additional computational overhead and are thus not widely used for multimedia transmissions in resource-constrained environments. D. Joint Compression and Communication Optimization UEP matches the communication strategy to the source coding but does not consider the power consumption as a result of the source coding itself. As a result, researchers proposed a different approach to cross-layer optimization called Joint Compression and Communication Optimization. In this approach, compression power consumption and transmission power consumption are jointly considered in order to optimize the performance of the entire transmission system. This approach is mainly applied to video transmission systems, since the video coding itself also exhibits a high power consumption compared with the wireless transmission. This approach aims to nd the optimal trade-off between compression and transmission in order to achieve an overall better transmission quality. Below we reviewe some specic research publications related to this approach. Lu et al. in [61] formulated the power consumption of the H.264 codec as a function of INTRA rate, encoding rate, and transmission power consumption as a function of encoding rate, channel coding rate, and radiated energy per transmitted bit. They conclude that when channel conditions are good, source coding consumes signicantly more power than transmission. Source coding therefore is required to be simplied for less computational overhead at the expense of transmitting more data with more transmission overhead. On the other hand, when the channel conditions are deteriorating more complex source coding using the H.264 codec is required to achieve a better balance to the high transmission energy requirement. Lu et al. in [62] then proposed an approach for minimizing the total power consumption of a mobile transmitter given the xed end-to-end source distortion. In that approach, the H.264 encoder power consumption model and RS channel coding power consumption model are formulated. By using these power consumption models, they successfully optimized the total power consumption of the transmitter. In [63], He et al. extended the traditional rate-distortion (R-D) optimization approach to power-rate-distortion (P-R-D) optimization by exploring the encoding power consumption model. They analyzed the encoding mechanisms of typical video coding systems, and developed a parametric video encoding architecture. Also, by using dynamic voltage scaling the complexity scalability is able to be translated into energy consumption of the video encoder. By using this approach they successfully established an analytical P-R-D model and the power consumption is generically optimized using that method. Wang et al. in [64] proposed an optimization approach to minimize the sum of compression powers and transmission powers of all users, subject to the received video quality at each terminal in a multiple-terminal video transmitting environment. Thus, they extended previous research to multiple terminal users, globally optimizing the entire networks power consumption.

To summarize, the joint compression and communication optimizations strategy is to balance the resources allocated to compression and communication in order to optimize the energy consumption, and is driven by the desired end-to-end link quality. This approach is more generic and complex than the UEP approach. However, the solution of UEP can be integrated into this approach as well. Additional publications in this area can be found in [65-71].

V. F UTURE W ORK AND O PEN I SSUES Compression efciency of image compression algorithm design is a well-studied area. However, determining lowcomplexity efcient compression algorithms suitable for resource-constrained communication systems is still an open area. Especially the problem of how to reduce the complexity of JPEG2000 such that the compression energy can be saved in the communication system is challenging but a very rewarding open research area. One promising video codec, HEVC, is expected to provide substantial improvements in coding efciency over H.264. How to trade off the power consumption of HEVC against wireless transmission power consumption will very likely become the next cross-layer design research focus in wireless communication systems. Energy-efcient cross layer design is a very complex problem and requires to systematically and effectively research all the network layer optimizations jointly [72].

VI. C ONCLUSION In this paper, we conducted a comprehensive survey for existing energy-efcient techniques for multimedia communication in resource-constrained systems. We introduced two groups of energy-efcient techniques in two integral multimedia communication phases: energy-efcient multimedia compression and energy-efcient multimedia transmission. In the rst group, existing multimedia compression algorithms are evaluated and compared to present an insight into the potential for energy efciency for these compression algorithms. In the second group, important transmission techniques crosslayer communication technique is reviewed. The summary of these transmission techniques aims to provide a vision of the current stage of research in energy-efciency for transmission systems. This survey presents opportunities to identify future research directions and serve as a foundation for related research efforts. R EFERENCES
[1] CC2500 data sheet, Texas Instruments, USA. [2] MSP430F2274 data sheet, Texas Instruments, USA. [3] Kenneth Barr and Krste Asanovic, Energy Aware Lossless Data Compression, In First International Conference on Mobile Systems, Applications, and Services, May 2003. [4] G.K. Wallace, The JPEG still picture compression standard, IEEE Trans. Consum. Electron., vol.38, no.1, Feb 1992. [5] D. U. Lee, H. Kim, M. Rahimi, D. Estrin, J.D. Villasenor, EnergyEfcient Image Compression for Resource-Constrained Platforms, IEEE Trans. Image Process., vol.18, no.9, pp. 2100-2113, Sept 2009.

MA et al.: ENERGY-EFFICIENT COMPRESSION AND COMMUNICATION TECHNIQUES FOR MULTIMEDIA IN RESOURCE CONSTRAINED SYSTEMS

971

[6] T. Lane, P. Gladstone, L. Ortiz, J. Boucher, L. Crocker, J. Minguillon, G. Phillips, D. Rossi, and G. Weijers, The Independent JPEG Group JPEG Software Release 6b, 1998 [Online]. Available: http://www.ijg.org. [7] W. Sweldens, The lifting scheme: A custom-design construction of biorthogonal wavelets, Appl. Comput. Harmon. Anal., vol. 3, pp.186200, 1996. [8] W. Sweldens, The lifting scheme: a construction of second generation wavelets, SIAM J. Math. Anal., pp. 511-546, 1997. [9] D. G. Lee; S. Dey, Adaptive and energy efcient wavelet image compression for mobile multimedia data services, IEEE International Conference on Communication., vol.4, pp. 2484- 2490, 2002. [10] M. Nasri, A. Helali., H. Sghaier, H. Maaref, Energy-efcient wavelet image compression in Wireless Sensor Network, International Conference on Communication in Wireless Environments and Ubiquitous Systems: New Challenges (ICWUS), pp.1-7, Oct 2010. [11] K. L. Wen and N. Burgess, Listless Zerotree Coding for Color Images, Proc. 32nd Asilomar International Conferences on Signals, Systems and Computers, pp. 231-235, 1998. [12] J. Lian, K. Wang, J. Yang, Listless Zerotree Image Compression Algorithm, 8th International Conference on Signal Processing, vol.2, pp.16-20, 2006. [13] J. Lian, L. H. Li, T. Ma, L. D. Cao, Improved Listless Zerotree Image Compression Algorithm and Application, 2nd International Congress on Image and Signal Processing, pp.1-4, vol.17-19, Oct 2009. [14] T. Ma, P. L. Shrestha, M. Hempel, D. Peng, H. Sharif, LowComplexity Image Coder/Decoder with an Approaching-Entropy Quad-Tree Search Code for Embedded Computing Platforms, International Conference on Signal Processing, 2011. [15] D. Taubman, High performance scalable image compression with EBCOT, IEEE Trans. Image Process., vol. 9, no.7, pp. 1158-1170, Jul. 2000. [16] M. W. Marcellin, M. J. Gormish, A. Bilgin, M. P. Bolie,An overview of JPEG-2000, Data Compression Conference, pp.523-541, 2000. [17] H. Zhang and J. Fritts, EBCOT coprocessing architecture for JPEG2000, Proc. SPIE, vol.5308, pp. 1333-1340, Jan 2004. [18] L. W. Chew, L.M. Ang, K. P. Seng, Survey of image compression algorithms in wireless sensor networks, International Symposium on Information Technology, pp.1-9, Aug 2008. [19] B. Girod, A. Aaron, S. Rane, D. Rebollo-Monedero, Distributed video coding, Proc. IEEE, pp. 71-83, 2005. [20] H. M. Wang, C. H. Tseng and J. F. Yang, Computation reduction for intra 4x4 mode decision with SATD criterion in H.264/AVC, IET Signal Process., vol. 1, no. 3, pp. 121127, Aug 2007. [21] J. S. Andersen, S. F., and S. M. Aghito, Rate distortion-complexity optimization of fast motion estimation in H.264/MPEG-4 AVC, in Proc. IEEE Int. Conf. Image Process, Singapore, Oct 2004. [22] C. H. Hsia, J. S. Chiang, Y. H. Wang, T. Yuan, Fast Intra Prediction Mode Decision Algorithm for H.264/AVC Video Coding Standard, Third International Conference on Intelligent Information Hiding and Multimedia Signal Processing, vol.2, pp.535-538, Nov 2007. [23] C.S. Kannangara, I.E.G. Richardson, M. Bystrom, J.R., Solera, Y. F. Zhao, A. MacLennan, R. Cooney, Low-complexity skip prediction for H.264 through Lagrangian cost estimation, IEEE Trans. Circuits Syst. Video Technol., vol.16, no.2, pp. 202- 208, Feb 2006. [24] C. Crecos and M.Y. Yang, Fast inter mode prediction for P slices in the H.264 video coding standard, IEEE Trans. Broadcast., vol. 51, no. 2, pp. 256-263, June 2005. [25] H. Kim and Y. Altunbasak, Low-complexity macro block mode selection for H.264/AVC encoders, in Proc. Int. Conf. Image Process., Singapore, pp. 765768, 2004. [26] B. G. Kim and C. S. Cho, A fast inter-mode decision algorithm based on macro-Block tracking for P slices in the H.264/AVC video standard, in Proc. Int. Conf. Image Process., pp. 301-304, 2007. [27] A. Wyner and J. Ziv, The Rate-Distortion Function for Source Coding with Side Information at the Decoder, IEEE Trans. Inf. Theory, vol. 22, no. 1, pp. 1-10, January 1976. [28] R. Puri and K. Ramchandran, PRISM: A New Robust Video Coding Architecture Based on Distributed Compression Principles, in Proc. Allerton Conference on Communication, Control and Computing, Allerton, IL, USA, October 2002. [29] A. Aaron, R. Thang, and B. Girod, Wyner-Ziv Coding of Motion Video, in Proc. Asilomar Conference on Signals and Systems, Pacic Grove, CA, USA, November 2002. [30] C. Guillemot, F. Pereira, L. Torres, T. Ebrahimi, R. Leonardi and J. Ostermann, Distributed Monoview and Multiview Video Coding, IEEE Signal Processing Mag., vol. 24, no. 5, pp. 67-76, September 2007.

[31] P.L Dragotti and M. Gastpar, Distributed Source Coding: Theory, Algorithms and Applications, Academic Press, February 2009. [32] J. Areia, J. Ascenso, C. Brites and F. Pereira, Low Complexity Hybrid Rate Control for Lower Complexity Wyner-Ziv Video Decoding, in Proc. European Conference on Signal Processing (EUSIPCO 08), Lausanne, Switzerland, August 2008. [33] C. H. Lai, K.Y. Hsieh, S.H. Lai and J. K. Lee, Parallelization of Belief Propagation Method on Embedded Multicore Processors for Stereo Vision, ESTIMedia, Atlanta, GA, October 2008. [34] J. Zhang and M. Fossorier, Shufed Belief Propagation Decoding, Proc. of 36th Annual Asilomar Conference on Signals, Systems and Computers, Pacic Grove, CA, November 2002. [35] J. Zhang, Y. Wang, M. Fossorier, and J. S. Yedidia, Replica Shufed Belief Propagation Decoding of LDPC Codes, Proc. of 39th Conference on Information Sciences and Systems (CISS), The Johns Hopkins University, Baltimore, MD, March 2005. [36] T. Wiegand; J. R. Ohm; G. J. Sullivan; W. J. Han; R. Joshi; T. K. Tan; K. Ugur; Special Section on the Joint Call for Proposals on High Efciency Video Coding (HEVC) Standardization, IEEE Trans. Circuits Syst. Video Technol., vol.20, no.12, pp.1661-1666, Dec. 2010. [37] K. Ugur, K. Andersson, A. Fuldseth, G. Bjntegaard, L. P. Endresen,J. Lainema, A. Hallapuro, J. Ridge, D. Rusanovskyy, C. X. Zhang, A. Norkin, C. Priddle, T. Rusert, J. Samuelsson, R. Sjoberg, Z. F. Wu, Low complexity video coding and the emerging HEVC standard, Picture Coding Symposium (PCS), pp. 474-477, 8-10 Dec 2010. [38] K. Ugur, K. Andersson, A. Fuldseth, G. Bjntegaard, L.P. Endresen, J. Lainema, A. Hallapuro, J. Ridge, D. Rusanovskyy, C. X. Zhang, A. Norkin, C. Priddle, T. Rusert, J. Samuelsson, R. Sjoberg, Z. F. Wu, High Performance, Low Complexity Video Coding and the Emerging HEVC Standard, IEEE Trans. Circuits Syst. Video Technol., vol.20, no.12, pp. 1688-1697, Dec 2010. [39] W. Wang, D. Peng, H. Wang, H. Sharif, H. H. Chen, EnergyConstrained Distortion Reduction Optimization for Wavelet-based Coded Image Transmission in Wireless Sensor Networks, IEEE Trans. Multimedia (TMM), vol. 10, no. 6, pp. 1169-1180, Oct. 2008. [40] T. Ma, M. Hempel, D. Peng, H. Sharif, Rate-switching Unequal Error Protection for Wireless Electrocardiogram (ECG) Transmission, Military Communication Conference, 2010. [41] Y. Lu, W. S. Wang, H. J. Cui, K. Tang, Energy Efcient Wireless Scalable Video Streaming with Optimal Transmission Rate Adaptation and Unequal Error Protection, IMACS Multi-conference on Computational Engineering in Systems Applications, 2006. [42] Y. Wei, Z. Sahinoglu, A. Vetro, Energy efcient JPEG 2000 image transmission over wireless sensor networks, Global Telecommunications Conference, vol.5, pp. 2738-2743, Dec 2004. [43] C. Costa, F. Granelli, A. K. Katsaggelos, A cross-layer approach for energy efcient transmission of progressively coded images over wireless channels, IEEE International Conference on Image Processing, vol.1, pp.11-14 Sept 2005. [44] H. Wang, Y. Eisenherg, F. Zhai, A. K., Katsaggelos, Joint objectbased video encoding and power management for energy efcient wireless video communications, International Conference on Image Processing, vol.4, pp. 2543-2546 Oct 2004. [45] C. E. Luna, Y. Eisenberg, T. N. Pappas, R. Berry, A. K. Katsaggelos, Transmission energy minimization in wireless video streaming applications, Asilomar, 2001. [46] C. E. Luna, Y. Eisenberg, T. N. Pappas. R. Berry, A. K. Katsaggelos, Joint Source Coding and Data Rate Adaptation for Energy Efcient Wireless Video Streaming, Packet Video Workshop, April 2002. [47] V. Chande and N. Farvardin, Progressive transmission of images over memoryless noisy channels, IEEE J. Sel. Areas Commun., vol. 18, no.6, pp. 850-860, Jun 2000. [48] V. Stankovic, R. Hamzaoui, and D. Saupe, Fast algorithm for ratebased optimal error protection of embedded codes, IEEE Trans. Commun., vol. 51, no. 11, pp. 1788-1795, Nov 2003. [49] R. Hamzaoui, V. Stankovic, and Z. Xiong, Rate-based versus distortion based optimal joint source-channel coding, in Proc. DCC02, Snowbird, UT, Apr 2002. [50] R. Hamzaoui, V. Stankovic, Z. Xiong, Fast algorithm for distortionbased error protection of embedded image codes, IEEE Trans. Image Process., vol. 14, no. 10, pp.1417-1421, Oct 2005. [51] G. Baruffa, P. Micanti, F. Frescura, Error Protection and Interleaving for Wireless Transmission of JPEG 2000 Images and Video, IEEE Trans. Image Processing, vol.18, no.2, pp.346-356, Feb 2009. [52] A. Nosratinia, L. Jin, and B. Aazhang, Source-channel rate allocation for progressive transmission of images, IEEE Trans. Commun., vol. 51, no.2, pp. 186-196, Feb 2003.

972

IEEE COMMUNICATIONS SURVEYS & TUTORIALS, VOL. 15, NO. 3, THIRD QUARTER 2013

[53] J. Kim, R. M. Mersereau, and Y. Altunbasak, Error-resilient image and video transmission over the Internet using unequal error protection, IEEE Trans. Image Process., vol. 12, no. 2, pp. 121-131, Feb 2003. [54] P. G. Sherwood and K. Zeger, Progressive image coding on noisy channels, Proc. Data Compression Conf, pp. 7279, Mar 1997. [55] V. Sanchez and M. K. Mandal, Efcient channel protection for JPEG2000 bitstream, IEEE Trans. Circuits Syst. Video Technol., vol. 14, no. 4, pp. 554-558, Apr 2004. [56] B. A. Banister, B. Belzer, and T. R. Fischer, Robust image transmission using JPEG2000 and turbo-codes, IEEE Signal Process. Lett., vol.9, no. 4, pp. 117-119, Apr 2002. [57] T. L. Chu, Z. Liu, Z. Xiong, and X. Wu, Joint UEP and layered source coding with application to transmission of JPEG-2000 coded images, IEEE GLOBECOM, vol. 3, pp. 20362039, Nov 2001. [58] C. Lan, T. Chu, K. R. Narayanan, and Z. Xiong, Scalable image and video transmission using irregular repeat-accumulate codes with fast algorithm for optimal unequal error protection, IEEE Trans. Commun., vol.52, no. 7, pp. 1092-1101, Jul 2004. [59] X. Pan; A. Cuhadar, A. H. Banihashemi, Combined source and channel coding with JPEG2000 and rate-compatible low-density Paritycheck codes, IEEE Trans. Signal Process., vol.54, no.3, pp. 11601164, March 2006. [60] N. Thomos, N. V. Boulgouris, M. G. Strintzis,Product code optimization for determinate state LDPC decoding in robust image transmission, IEEE Trans. Image Process., vol.15, no.8, pp.21132119, Aug 2006. [61] X.Lu, Y.Wang, and E.Erkip,Power efcient H.263 video transmission over wireless channels, Proc. International Conference on Image Processing, Rochester, New York, September 2002. [62] X. Lu, E. Erkip, Y. Wang and D. Goodman. Power efcient multimedia communication over wireless channels, IEEE J. Sel. Areas Commun., Special Issue on Recent Advances in Wireless Multimedia, vol.21, no. 10, pp.1738-1751, December 2003. [63] Z. He, Y. Liang, L. Chen, I. Ahmad, and D. Wu, Power-RateDistortion Analysis for Wireless Video Communication under Energy Constraints, IEEE Trans. Circuits Syst. Video Technol., Special Issue on Integrated Multimedia Platforms, vol. 15, no. 5, pp. 645-658, May 2005. [64] X. Lu, Y. Wang, E. Erkip and D. Goodman, Total power minimization for multiuser video communications over CDMA networks, IEEE Trans. Circuits Syst. Video Technol., vol. 17, no. 6, pp. 674-685, June 2007. [65] Zhihai He, Wenjun Zeng, and Chang Wen Chen, Low-pass ltering of rate-distortion functions for quality smoothing and bandwidth control in real-time video coding, IEEE Trans. Circuits Syst. Video Technol., vol. 15, No. 8, pp. 973-981, 2005. [66] J. Cai, J. Wu, K. N. Ngan and Z. He, Joint mode selection and unequal error protection for bit plane coded video transmission over wireless channels, J. Visual Communications and Image Representation, special issue on visual communications in ubiquitous era, vol. 16, no. 4-5, pp. 412-431, Aug-Oct 2005. [67] Zhihai He and Sanjit K. Mitra, From rate-distortion analysis to resource-distortion analysis, IEEE Circuits Syst. Mag., vol.5, No. 3, pp. 6-18, 2005. [68] Zhihai He and Chang Wen Chen, Wireless video communication under resource constraints, (invited paper), Encyclopedia of Multimedia, December 2005. [69] Zhihai He, Power-rate-distortion analysis for wireless video, (invited paper), Encyclopedia of Multimedia, December 2005. [70] Zhihai He and Dapeng Wu, Resource Allocation and Cross-Layer Optimization for Real-Time Wireless Video Communication over Portable Devices, IEEE Trans. Circuits Syst. Video Technol., in Jan 2006. [71] Zhang, J., Wu, D., Ci, S., Wang, H., Katsaggelos, A., Power-Aware Mobile Multimedia: A Survey, J. Communications, 2009. [72] Miao, G., Himayat, N., Li, Y. and Swami, A. , Cross-layer optimization for energy-efcient wireless communications: A survey, Wireless Communications and Mobile Computing, pp. 529-542, 2009.

Tao Ma received the B.Sc. and M.Sc. degrees from Xian Jiaotong University, China, in 2005 and 2008 respectively, both in electrical engineering. He is currently pursuing the Ph.D. degree and in Department of Computer and Electronics Engineering, University of Nebraska-Lincoln. His research areas are Cross-layer design for QoS provisioning in wireless data networks, and multi-media distribution, ultralow power sensor network.

Michael Hempel received the Ph.D. degree in computer engineering from the University of NebraskaLincoln in 2007. He is currently a Research Assistant Professor in the Department of Computer and Electronics Engineering, University of NebraskaLincoln. His research interests include wireless communications networks and multimedia communications.

Dongming Peng received the B.A. and M.A. degrees in electrical engineering from Beijing University of Aeronautics and Astronautics, Beijing, China, in 1993 and 1996, respectively, and the Ph.D. degree in computer engineering from Texas A&M University, College Station, in 2003. From 1996 to 1997, he was a faculty member at Beijing University. In 2002, he joined the University of Nebraska-Lincoln, where he is currently an Associate Professor. His research interests include digital image processing, computer architectures, parallel and distributed computing, and sensor network. Dr. Peng is one of the recipients of the Best Paper Award in IEEE WCNC 2008. He has also served as a referee and program committee member for several conferences and journals.

Hamid Sharif is the Charles J. Vranek Professor in the Department of Computer and Electronics Engineering at the University of Nebraska-Lincoln. He received his BSEE from University of Iowa, MSEE from University of Missouri-Columbia, and PhD from University of Nebraska-Lincoln. He is the director of Advanced Telecommunications Engineering Laboratory (TEL) at University of Nebraska. Professor Sharif has published over 200 research articles in international journals and conferences and has been the recipient of a number of best papers. Dr. Sharif has been serving on many IEEE and other international journal editorial boards and currently is the co-editor-in-chief for the Wiley Journal of Security and Communication Networks. He has contributed to the IEEE in many roles including the elected Chair of the Nebraska Section, elected Chair of the Nebraska Computer Chapter, elected Chair of the Nebraska Communication Chapter, and the Chapter Coordinator for the IEEE Region 4 in US.

You might also like