Mise à jour du 13 mars 2010

Research Pages

Bibliography on Video Encoding - Table



JabRef References output

Global QuickSearch:   Number of matching entries: 0

Search Settings

    Author / Editor / Organization Title Year Journal / Proceedings / Book BibTeX type DOI/URL
    Moecke, M. & Seara, R. Sorting Rates in Video Encoding Process for Complexity Reduction 2010 Circuits and Systems for Video Technology, IEEE Transactions on
    Vol. 20 (1) , pp. 88 -101  
    article DOI    
    Abstract: The motion estimation process and coding mode selection are responsible for a large portion of the computational effort in H.264-based video encoding systems optimized for rate-distortion (RD). This paper presents a rate sorting and truncation strategy that incorporates the RD optimization criterion in the decision to evaluate distortion for both motion vectors and coding modes. Experimental results confirm the effectiveness of the proposed approach, yielding up to a 90% reduction in the computational complexity. An additional saving can also be obtained, with insignificant RD performance loss, by using a quality threshold.
    BibTeX:
    @article{2010_Moecke,
      author = {Moecke, M. and Seara, R.},
      title = {Sorting Rates in Video Encoding Process for Complexity Reduction},
      journal = {Circuits and Systems for Video Technology, IEEE Transactions on},
      year = {2010},
      volume = {20},
      number = {1},
      pages = {88 -101},
      doi = {http://dx.doi.org/10.1109/TCSVT.2009.2029022}
    }
    					
    Aghagolzadeh, M. & Oweiss, K. Compressed and Distributed Sensing of Neuronal Activity for Real Time Spike Train Decoding 2009 Neural Systems and Rehabilitation Engineering, IEEE Transactions on
    Vol. 17 (2) , pp. 116 -127  
    article DOI    
    Abstract: Multivariate point processes are increasingly being used to model neuronal response properties in the cortex. Estimating the conditional intensity functions underlying these processes is important to characterize and decode the firing patterns of cortical neurons. This paper proposes a new approach for estimating these intensity functions directly from a compressed representation of the neurons' extracellular recordings. The approach is based on exploiting a sparse representation of the extracellular spike waveforms, previously demonstrated to yield near-optimal denoising and compression properties. We show that by restricting this sparse representation to a subset of projections that simultaneously preserve features of the spike waveforms in addition to the temporal characteristics of the underlying intensity functions, we can reasonably approximate the instantaneous firing rates of the recorded neurons with variable tuning characteristics across a multitude of time scales. Such feature is highly desirable to detect subtle temporal differences in neuronal firing characteristics from single-trial data. An added advantage of this approach is that it eliminates multiple steps from the typical processing path of neural signals that are customarily performed for instantaneous neural decoding. We demonstrate the decoding performance of the approach using a stochastic cosine tuning model of motor cortical activity during a natural, nongoal-directed 2-D arm movement.
    BibTeX:
    @article{2009_Aghagolzadeh,
      author = {Aghagolzadeh, M. and Oweiss, K.},
      title = {Compressed and Distributed Sensing of Neuronal Activity for Real Time Spike Train Decoding},
      journal = {Neural Systems and Rehabilitation Engineering, IEEE Transactions on},
      year = {2009},
      volume = {17},
      number = {2},
      pages = {116 -127},
      doi = {http://dx.doi.org/10.1109/TNSRE.2009.2012711}
    }
    					
    Ben Jamaa, S.; Kieffer, M. & Duhamel, P. Improved sequential MAP estimation of CABAC encoded data with objective adjustment of the complexity/efficiency tradeoff 2009 Communications, IEEE Transactions on
    Vol. 57 (7) , pp. 2014 -2023  
    article DOI    
    Abstract: This paper presents an improved sequential MAP estimator to be used as a joint source-channel decoding technique for CABAC encoded data. The decoding process is compatible with realistic implementations of CABAC in standards like H.264, i.e, handling adaptive probabilities, context modeling and integer arithmetic coding. Soft-input decoding is obtained using an improved sequential decoding technique, which allows to obtain a tradeoff between complexity and efficiency. The algorithms are simulated in a context reminiscent of H264. Error detection is realized by exploiting on one side the properties of the binarization scheme and on the other side the redundancy left in the code string. As a result, the CABAC compression efficiency is preserved and no additional redundancy is introduced in the bit stream. Simulation results outline the efficiency of the proposed techniques for encoded data sent over AWGN and UMTS-OFDM channels.
    BibTeX:
    @article{2009_BenJamaa,
      author = {Ben Jamaa, S. and Kieffer, M. and Duhamel, P.},
      title = {Improved sequential MAP estimation of CABAC encoded data with objective adjustment of the complexity/efficiency tradeoff},
      journal = {Communications, IEEE Transactions on},
      year = {2009},
      volume = {57},
      number = {7},
      pages = {2014 -2023},
      doi = {http://dx.doi.org/10.1109/TCOMM.2009.07.070566}
    }
    					
    Bhattacharyya, D.; Dutta, J.; Das, P.; Bandyopadhyay, S. & hoon Kim, T. Discrete Cosine Transformation Based Image Authentication and Secret Message Transmission Scheme 2009 Computational Intelligence, Communication Systems and Networks, 2009. CICSYN '09. First International Conference on , pp. 374 -379   inproceedings DOI    
    Abstract: In this paper a novel ldquodiscrete cosine transformation based image authentication amp; secret message transmission schemerdquo (DCTIASMTT) has been proposed to authenticate a multimedia image and at the same time some ubiquitous secret message or image can be transmitted over the mobile network. Instead of direct embedding a message or image within the source image, choose a window of size 2x2 of the source image and then convert it from spatial domain to frequency domain using discrete cosine transformation (DCT). The bits of the authenticating message or image are then embedded at LSB of the transformed image (excluding first pixel). Inverse DCT has been performed for the transformation from frequency domain to spatial domain as final step of encoding. Decoding is done through reverse procedure. The experimental results have been discussed and compared with the existing steganographic algorithm S-Tools. Histogram analysis and chi-square test of source image with embedded image shows the better results in comparison with S-Tools.
    BibTeX:
    @inproceedings{2009_Bhattacharyya,
      author = {Bhattacharyya, D. and Dutta, J. and Das, P. and Bandyopadhyay, S.K. and Tai-hoon Kim},
      title = {Discrete Cosine Transformation Based Image Authentication and Secret Message Transmission Scheme},
      journal = {Computational Intelligence, Communication Systems and Networks, 2009. CICSYN '09. First International Conference on},
      year = {2009},
      pages = {374 -379},
      doi = {http://dx.doi.org/10.1109/CICSYN.2009.11}
    }
    					
    Chen, S.; Hu, R.; Park, J.H.; Xiong, N. & Yeo, S.-S. Analysis and Synthesis of Spatial Parameters Using MDCT 2009 Multimedia and Ubiquitous Engineering, 2009. MUE '09. Third International Conference on , pp. 18 -21   inproceedings DOI    
    Abstract: We present a method of analyzing and synthesizing spatial parameters using only modified discrete cosine transform (MDCT). Combing it with modified discrete sine transform (MDST) enables spatial parameters representation. And instead of direct transforming, MDST spectra can be perfectly built from neighboring MDCT spectra by a conversion matrix, which is highly diagonal and can be approximated by a small number of its sub-diagonals. With MDCT based core coders, like advanced audio coding (AAC), we need no separate transforming for spatial coding, cutting down significant computational complexity.
    BibTeX:
    @inproceedings{2009_Chen,
      author = {Shuixian Chen and Ruimin Hu and Jong Hyuk Park and Naixue Xiong and Sang-Soo Yeo},
      title = {Analysis and Synthesis of Spatial Parameters Using MDCT},
      journal = {Multimedia and Ubiquitous Engineering, 2009. MUE '09. Third International Conference on},
      year = {2009},
      pages = {18 -21},
      doi = {http://dx.doi.org/10.1109/MUE.2009.14}
    }
    					
    Chen, S.-H. & Perng, D.-B. Automatic Surface Inspection for Directional Textures Using Discrete Cosine Transform 2009 Pattern Recognition, 2009. CCPR 2009. Chinese Conference on , pp. 1 -5   inproceedings DOI    
    Abstract: A new global image restoration scheme using discrete cosine transform (DCT) is proposed in this paper. This DCT-based image restoration scheme can be used for inspecting the defects in directional texture surfaces automatically. The input spatial domain image is first transformed into DCT domain. The dominating directions of the textures in the input image will be compacted to orthogonal straight lines, respectively, throughout the direct current (DC) component in the spectrum. The linear primitives associated with high-energy frequency components in the DCT domain are eliminated by reducing them to zero and then transformed back to the spatial domain. This procedure will have all directional textures be blurred and will preserve only local defects if they are initially embedded in the input image. Experiments on a variety of product surfaces with directional texture such as straight, slant, orthogonal, and slant orthogonal linear primitives are given to demonstrate the effectiveness and robustness of the proposed method.
    BibTeX:
    @inproceedings{2009_Chena,
      author = {Ssu-Han Chen and Der-Baau Perng},
      title = {Automatic Surface Inspection for Directional Textures Using Discrete Cosine Transform},
      journal = {Pattern Recognition, 2009. CCPR 2009. Chinese Conference on},
      year = {2009},
      pages = {1 -5},
      doi = {http://dx.doi.org/10.1109/CCPR.2009.5344097}
    }
    					
    Dong, J.; Ngan, K.N.; Fong, C.-K. & Cham, W.-K. 2-D Order-16 Integer Transforms for HD Video Coding 2009 Circuits and Systems for Video Technology, IEEE Transactions on
    Vol. 19 (10) , pp. 1462 -1474  
    article DOI    
    Abstract: In this paper, the spatial properties of high-definition (HD) videos are investigated based on a large set of HD video sequences. Compared with lower resolution videos, the prediction errors of HD videos have higher correlation. Hence, we propose using 2-D order-16 transforms for HD video coding, which are expected to be more efficient to exploit this spatial property, and specifically propose two types of 2-D order-16 integer transforms, nonorthogonal integer cosine transform (ICT) and modified ICT. The former resembles the discrete cosine transform (DCT) and is approximately orthogonal, of which the transform error introduced by the nonorthogonality is proven to be negligible. The latter modifies the structure of the DCT matrix and is inherently orthogonal, no matter what the values of the matrix elements are. Both types allow selecting matrix elements more freely by releasing the orthogonality constraint and can provide comparable performance with that of the DCT. Each type is integrated into the audio and video coding standard (AVS) Enhanced Profile (EP) and the H.264 high profile (HP), respectively, and used adaptively as an alternative to the 2-D order-8 transform according to local activities. At the same time, many efforts have been devoted to further reducing the complexity of the 2-D order-16 transforms and specially for the modified ICT, a fast algorithm is developed and extended to a universal approach. Experimental results show that 2-D order-16 transforms provide significant performance improvement for both AVS enhanced profile and H.264 high profile, which means they can be efficient coding tools especially for HD video coding.
    BibTeX:
    @article{2009_Dong,
      author = {Jie Dong and King Ngi Ngan and Chi-Keung Fong and Wai-Kuen Cham},
      title = {2-D Order-16 Integer Transforms for HD Video Coding},
      journal = {Circuits and Systems for Video Technology, IEEE Transactions on},
      year = {2009},
      volume = {19},
      number = {10},
      pages = {1462 -1474},
      doi = {http://dx.doi.org/10.1109/TCSVT.2009.2026792}
    }
    					
    Hait, N. & Malah, D. Model-Based Transrating of H.264 Coded Video 2009 Circuits and Systems for Video Technology, IEEE Transactions on
    Vol. 19 (8) , pp. 1129 -1142  
    article DOI    
    Abstract: This paper presents a model-based transrating (bit-rate reduction) system for H.264 coded video via requantization. In works related to previous standards, optimal requantization step sizes were obtained via Lagrangian optimization that minimizes the distortion subject to a rate constraint. Due to H.264 advanced coding features, the choices of quantization step size and coding modes are dependent and the rate control becomes computationally expensive. Therefore, optimal requantization algorithms developed for previous standards cannot be applied as they are. Hence, previous works on transrating in H.264 focused on changing the input coding decisions rather than on rate control, while requantization was addressed by a simple one-pass algorithm. Here we propose new model-based optimal requantization algorithms for transrating of H.264 coded video. The optimal requantization goal is to achieve the target bit rate with minimal effect on video quality. Incorporation of the proposed models serves two goals. For intra-coded frames, a novel closed-loop statistical estimator that overcomes spatial neighbors dependencies is developed. For inter-coded frames, the proposed macroblock-level models reduce the computational burden of the optimization. Overall, as compared to re-encoding (cascaded decoder-encoder), the proposed system reduces the computational complexity by a factor of about four, at an average PSNR loss of only 0.4 dB for transrating CIF/SIF sequences from 2 Mb/s to 1 Mb/s. In comparison with a simple one-pass requantization, the proposed algorithm achieves better performance (an average PSNR gain of 0.45 dB), at the cost of just twice the complexity.
    BibTeX:
    @article{2009_Hait,
      author = {Hait, N. and Malah, D.},
      title = {Model-Based Transrating of H.264 Coded Video},
      journal = {Circuits and Systems for Video Technology, IEEE Transactions on},
      year = {2009},
      volume = {19},
      number = {8},
      pages = {1129 -1142},
      doi = {http://dx.doi.org/10.1109/TCSVT.2009.2020334}
    }
    					
    Iranpour, A. & Kuchcinski, K. Design space exploration for optimal memory mapping of data and instructions in multimedia applications to Scratch-Pad Memories 2009 Embedded Systems for Real-Time Multimedia, 2009. ESTIMedia 2009. IEEE/ACM/IFIP 7th Workshop on , pp. 89 -95   inproceedings DOI    
    Abstract: In this paper, we propose a new methodology for optimal memory mapping of data and instructions to Scratch-Pad Memories (SPM). In the mapping process, we optimize, as the main priority, the number of memory accesses to minimize power consumption. Minimization of external memory accesses lowers switching activity and therefore power consumption. The optimization is done by finding Pareto-points, using multi-objective optimization that combines different cost functions. Our methodology is intended to be used in real-life situations in industry where there is often a need for mapping third party applications to a specific architecture. For evaluating our methodology, we also use commercial video H.264 and audio eAAC+ applications. Our experiments show that SPM is well suited for these applications for reducing external accesses to reduce power consumption but has limited significance on overall performance improvements. The proposed methodology provides a way to combine SPMs with caches to optimally use this memory architecture. Our experiments indicate high accuracy of our methodology for predicting SPM and external memory accesses. We have obtained 90% accuracy between results of our methodology and results for executing applications on a given architecture.
    BibTeX:
    @inproceedings{2009_Iranpour,
      author = {Iranpour, A.R. and Kuchcinski, K.},
      title = {Design space exploration for optimal memory mapping of data and instructions in multimedia applications to Scratch-Pad Memories},
      journal = {Embedded Systems for Real-Time Multimedia, 2009. ESTIMedia 2009. IEEE/ACM/IFIP 7th Workshop on},
      year = {2009},
      pages = {89 -95},
      doi = {http://dx.doi.org/10.1109/ESTMED.2009.5336826}
    }
    					
    Jayanthi, V.; Selvalakshmi, V. & Rajamani, V. Digital watermarking robust to geometric distortions in biomedical images 2009 Control, Automation, Communication and Energy Conservation, 2009. INCACEC 2009. 2009 International Conference on , pp. 1 -6   inproceedings    
    Abstract: We propose a new normalization procedure, which is invariant to affine transform attacks .The resulting watermarking scheme is suitable for public watermarking applications, where the original image is not available for watermark extraction. , we employ watermarking scheme as a direct-sequence code division multiple access approach to embed multibit text information in DCT and DWT transform domain. In Biomedical images the watermark is embedded in the regions other than the area of interest so that the diagnosis is not affected. The area of interest of the biomedical image is found out using the K-means segmentation method. Performance of proposed watermarking scheme is robust against various types of attacks such as Affine transform, Gaussian noise, shearing, rotation, median filtering and JPEG compression. This proposed algorithm can be used in Hospital environment.
    BibTeX:
    @inproceedings{2009_Jayanthi,
      author = {Jayanthi, V.E. and Selvalakshmi, V.M. and Rajamani, V.},
      title = {Digital watermarking robust to geometric distortions in biomedical images},
      journal = {Control, Automation, Communication and Energy Conservation, 2009. INCACEC 2009. 2009 International Conference on},
      year = {2009},
      pages = {1 -6}
    }
    					
    Lee, Y.-M. & Lin, Y. Zero-Block Mode Decision Algorithm for H.264/AVC 2009 Image Processing, IEEE Transactions on
    Vol. 18 (3) , pp. 524 -533  
    article DOI    
    Abstract: In the previous paper , we proposed a zero-block intermode decision algorithm for H.264 video coding based upon the number of zero-blocks of 4times4 DCT coefficients between the current macroblock and the co-located macroblock. The proposed algorithm can achieve significant improvement in computation, but the computation performance is limited for high bit-rate coding. To improve computation efficiency, in this paper, we suggest an enhanced zero-block decision algorithm, which uses an early zero-block detection method to compute the number of zero-blocks instead of direct DCT and quantization (DCT/Q) calculation and incorporates two adequate decision methods into semi-stationary and nonstationary regions of a video sequence. In addition, the zero-block decision algorithm is also applied to the intramode prediction in the P frame. The enhanced zero-block decision algorithm brings out a reduction of average 27% of total encoding time compared to the zero-block decision algorithm.
    BibTeX:
    @article{2009_Lee,
      author = {Yu-Ming Lee and Yinyi Lin},
      title = {Zero-Block Mode Decision Algorithm for H.264/AVC},
      journal = {Image Processing, IEEE Transactions on},
      year = {2009},
      volume = {18},
      number = {3},
      pages = {524 -533},
      doi = {http://dx.doi.org/10.1109/TIP.2008.2009858}
    }
    					
    Li, F. & Liu, G. Compressed-Domain-Based Transmission Distortion Modeling for Precoded H.264/AVC Video 2009 Circuits and Systems for Video Technology, IEEE Transactions on
    Vol. 19 (12) , pp. 1908 -1914  
    article DOI    
    Abstract: Transmission distortion analysis for video streams is a considerably challenging task. In this letter, a compressed-domain-based (CDB) transmission distortion model for precoded H.264/advanced video coding video streams is developed. Unlike the earlier schemes, which were based on pixel domain and required a complete decoding of the compressed video streams, the CDB model only requires some information on the video features, which can be directly extracted from the compressed video streams. Therefore, the complexity of the calculations is substantially reduced, which is well suited for real-time applications. More specifically, the model is applicable to the real-time transmission for precoded video streams, such as video on demand and mobile video. The experimental results demonstrate high accuracy of the model. Furthermore, an application example using the CDB model in resource allocation in real-time multiuser video communication reveals the applicability and effectiveness of the model.
    BibTeX:
    @article{2009_Li,
      author = {Fan Li and Guizhong Liu},
      title = {Compressed-Domain-Based Transmission Distortion Modeling for Precoded H.264/AVC Video},
      journal = {Circuits and Systems for Video Technology, IEEE Transactions on},
      year = {2009},
      volume = {19},
      number = {12},
      pages = {1908 -1914},
      doi = {http://dx.doi.org/10.1109/TCSVT.2009.2031457}
    }
    					
    Li, G.; Pei, S.; Chen, G.; Cao, W. & Wu, B. A self-embedded watermarking scheme based on relationship function of corresponding inter-blocks DCT coefficient 2009 Computer Supported Cooperative Work in Design, 2009. CSCWD 2009. 13th International Conference on , pp. 107 -112   inproceedings DOI    
    Abstract: In the realm of computer supported cooperative work in design (CSCWD), how to ensure the authenticity and integrity of an image plays important roles. This paper presents a novel semi-fragile image watermarking scheme for authenticating and recovering image content. The scheme also shows strong robustness on those images' content reserved operations. It can precisely detect and locate malicious operations and tampers, and recover main content of tampered image region. Firstly, it assigns the exclusive precursor block and successor block for every block to be a block circle link, and regards the function mapping result of DCT coefficients of neighboring blocks as watermark, and embeds it into middle-frequency domain of DCT coefficients of successor block to form a watermarked image. Secondly, it verifies the blocks, which have been maliciously operated and tampered, by detecting whether the DCT coefficients of neighboring blocks satisfy the relationship function after operations. Thirdly, it uses corresponding relationship function of neighboring blocks in the link to estimate DCT direct current coefficients (DC) and low-frequency coefficients of the tampered blocks and recovers their main content. The experimental results show that the scheme is effective and feasible.
    BibTeX:
    @inproceedings{2009_Lia,
      author = {Guobo Li and Songwen Pei and Gang Chen and Wenjun Cao and Baifeng Wu},
      title = {A self-embedded watermarking scheme based on relationship function of corresponding inter-blocks DCT coefficient},
      journal = {Computer Supported Cooperative Work in Design, 2009. CSCWD 2009. 13th International Conference on},
      year = {2009},
      pages = {107 -112},
      doi = {http://dx.doi.org/10.1109/CSCWD.2009.4968043}
    }
    					
    Liang, H. Research on the MPEG-2 Video Watermarking Scheme Based on Spread Spectrum Technology 2009 Computer Engineering and Technology, 2009. ICCET '08. International Conference on
    Vol. 2 , pp. 408 -411  
    inproceedings DOI    
    Abstract: Digital watermarking is an effective technology for digital products copyright protection and data security maintenance as well as an important branch of information hiding technology. A kind of MPEG-2 video watermarking scheme based on spread spectrum technology is introduced in this paper. First starting with MPEG-2 compressed video standard and then the algorithm principle of video watermarking is given, after video streams are partly decoded, the watermarking information modulated by m sequence is embedded into the direct current coefficients of chroma DCT of video I frame. Because of partly decoding, the computational complexity of complete decoding and recoding is reduced and watermark embedding speed is increased, finally, the concrete video watermarking implementation scheme is proposed from watermark embedding and extraction.
    BibTeX:
    @inproceedings{2009_Liang,
      author = {Huang Liang},
      title = {Research on the MPEG-2 Video Watermarking Scheme Based on Spread Spectrum Technology},
      journal = {Computer Engineering and Technology, 2009. ICCET '08. International Conference on},
      year = {2009},
      volume = {2},
      pages = {408 -411},
      doi = {http://dx.doi.org/10.1109/ICCET.2009.241}
    }
    					
    Li, J.; Gabbouj, M.; Takala, J. & Chen, H. Direct 3-D DCT-to-DCT resizing algorithm for video coding 2009 Image and Signal Processing and Analysis, 2009. ISPA 2009. Proceedings of 6th International Symposium on , pp. 105 -110   inproceedings    
    Abstract: This paper proposes a direct DCT-to-DCT resizing algorithm for 3-D DCT based video coding. An 8 times 8 times 8 cube is resized to three modes along temporal dimension: a single 8 times 8 block, a downsized 8 times 8 times 4 cube and two 8 times 8 times 4 cubes. The mode selection is based on the local motion activity and determined after 2-D DCT on each block. In addition, the proposed resizing scheme directly resizes the DCT blocks in the transform domain using the DCT-to-DCT algorithm. Compared to traditional resizing approaches the proposed algorithm does not require the inverse transform and the computations in the spatial domain, thus is superior to other methods in terms of complexity. The proposed model is evaluated with the baseline video codec and the reference codec in the literature. Experimental results show a promising performance in terms of both coding efficiency and computational complexity. Potential applications could be for portable digital devices with low-power processors and other areas with real-time requirement.
    BibTeX:
    @inproceedings{2009_Lib,
      author = {Jin Li and Gabbouj, M. and Takala, J. and Hexin Chen},
      title = {Direct 3-D DCT-to-DCT resizing algorithm for video coding},
      journal = {Image and Signal Processing and Analysis, 2009. ISPA 2009. Proceedings of 6th International Symposium on},
      year = {2009},
      pages = {105 -110}
    }
    					
    Lou, J.; Jagmohan, A.; He, D.; Lu, L. & Sun, M.-T. H.264 Deblocking Speedup 2009 Circuits and Systems for Video Technology, IEEE Transactions on
    Vol. 19 (8) , pp. 1178 -1182  
    article DOI    
    Abstract: This letter tackles the problem of reducing the complexity of H.264 decoding. Since deblocking accounts for a significant percentage of H.264 decoding time, our focus is on the H.264 in-loop deblocking filter. Observing that branch operations are costly and that in the deblocking process there are events with significantly high probability of occurrence, we regroup and simplify the branch operations. We apply the idea of Huffman tree optimization to speed up the boundary strength derivation and the true-edge detection by taking advantage of the biased statistical distribution. Our analyses and experiments show that the proposed techniques can reduce the deblocking computation time typically by a factor of more than seven times, while maintaining the bit-exact output.
    BibTeX:
    @article{2009_Lou,
      author = {Jian Lou and Jagmohan, A. and Dake He and Ligang Lu and Ming-Ting Sun},
      title = {H.264 Deblocking Speedup},
      journal = {Circuits and Systems for Video Technology, IEEE Transactions on},
      year = {2009},
      volume = {19},
      number = {8},
      pages = {1178 -1182},
      doi = {http://dx.doi.org/10.1109/TCSVT.2009.2020262}
    }
    					
    Martinez, J.; Kalva, H.; Fernando, W.; Cuenca, P. & Quiles, F. Effiecient WZ-to-H264 transcoding using motion vector information sharing 2009 Multimedia and Expo, 2009. ICME 2009. IEEE International Conference on , pp. 1394 -1397   inproceedings DOI    
    Abstract: In mobile-to-mobile video communications, both the sender and the receiver devices should not have higher complexity requirements to perform complex video compression tasks. The traditional video coding solutions are not suitable to support this communications due to its extremely complex encoding algorithm. On the other hand, the new Wyner-Ziv video coding paradigm reduces the complexity of the encoder at the expenses of a more complex decoder. In this paper, we propose an improved WZ/H.264 video transcoder to support this mobile-to-mobile communications, using the low complexity Wyner-Ziv encoding and the traditional H.264 decoding to be implemented in the end-user devices. The improved transcoder converts the video from the Wyner-Ziv to H.264 and reuses the motion vectors generated in the Wyner-Ziv decoding, in order to reduce the computational complexity of the motion estimation process in the H.264 encoding. Simulations results show a complexity reduction up to 55% with negligible rate-distortion drop.
    BibTeX:
    @inproceedings{2009_Martinez,
      author = {Martinez, J.L. and Kalva, H. and Fernando, W.A.C. and Cuenca, P. and Quiles, F.J.},
      title = {Effiecient WZ-to-H264 transcoding using motion vector information sharing},
      journal = {Multimedia and Expo, 2009. ICME 2009. IEEE International Conference on},
      year = {2009},
      pages = {1394 -1397},
      doi = {http://dx.doi.org/10.1109/ICME.2009.5202764}
    }
    					
    Martinez-Enriquez, E. & Diaz-de-Maria, F. A hierarchical classification-based approach to Inter Mode Decision in H.264/AVC 2009 Multimedia and Expo, 2009. ICME 2009. IEEE International Conference on , pp. 221 -224   inproceedings DOI    
    Abstract: The H.264/AVC standard achieves a high coding efficiency compared with previous standards. However, it does so at a very high computational cost, with motion estimation being one of the most demanding subsystems. In this paper a hierarchical classification-based approach to the inter mode decision (MD) problem is proposed. A first classifier detects SKIP/Direct modes while a second one is able to decide whether using a large (16times16, 16times8 and 8times16) or a small mode (8times8, 8times4, 4times8 and 4times4). The same procedure is applied at SubMacroblock level. The suggested classifiers are binary and linear. The input features that feed both classifiers have been carefully selected. A novel cost function that pays more attention to the most critical samples during the classifier training process has been designed. The results are very promising: a 64 % computational saving of the total encoding time with respect to JM10.2 is achieved with negligible degradation in the rate-distortion performance.
    BibTeX:
    @inproceedings{2009_Martinez-Enriquez,
      author = {Martinez-Enriquez, E. and Diaz-de-Maria, F.},
      title = {A hierarchical classification-based approach to Inter Mode Decision in H.264/AVC},
      journal = {Multimedia and Expo, 2009. ICME 2009. IEEE International Conference on},
      year = {2009},
      pages = {221 -224},
      doi = {http://dx.doi.org/10.1109/ICME.2009.5202476}
    }
    					
    Mohammadnia, M.; Taheri, H. & Motamedi, S. Fast H.264/AVC Intra Mode Decision Implementation on DM648 DSP 2009 Signal Acquisition and Processing, 2009. ICSAP 2009. International Conference on , pp. 53 -56   inproceedings DOI    
    Abstract: One of the advantages of H264/AVC video coding standard is its new algorithms in intra coding mode, but making decision among all these prediction modes needs some complex computation so it makes a bottleneck in implementation of a real-time video codec. In this paper a new and simple method for intra prediction mode decision has been proposed and also implemented on a DM648 TI DSP. Experimental results illustrate the proposed algorithm efficiency in coding procedure speed versus neglectable PSNR and bit rate degradation.
    BibTeX:
    @inproceedings{2009_Mohammadnia,
      author = {Mohammadnia, M.R. and Taheri, H. and Motamedi, S.A.},
      title = {Fast H.264/AVC Intra Mode Decision Implementation on DM648 DSP},
      journal = {Signal Acquisition and Processing, 2009. ICSAP 2009. International Conference on},
      year = {2009},
      pages = {53 -56},
      doi = {http://dx.doi.org/10.1109/ICSAP.2009.16}
    }
    					
    Narkhede, N. & Limaye, S. FPGA Implementation of Low Phase Noise Oscillator 2009 Advances in Computing, Control, Telecommunication Technologies, 2009. ACT '09. International Conference on , pp. 244 -247   inproceedings DOI    
    Abstract: Many DSP applications require a complex sinusoid to accomplish various signal rotation tasks. Examples include the discrete, fast Fourier transforms and digital up/down conversions. This article presents a complex oscillator based on the unfolded CORDIC algorithm and produces periodic sine and cosine samples for any specified angle increment. Low phase noise is achieved by residual angle correction unit and stability is achieved using AGC unit. The simulation and synthesis results in FPGA are presented. Fourier analysis of the output using MATLAB is also presented.
    BibTeX:
    @inproceedings{2009_Narkhede,
      author = {Narkhede, N.P. and Limaye, S.S.},
      title = {FPGA Implementation of Low Phase Noise Oscillator},
      journal = {Advances in Computing, Control, Telecommunication Technologies, 2009. ACT '09. International Conference on},
      year = {2009},
      pages = {244 -247},
      doi = {http://dx.doi.org/10.1109/ACT.2009.68}
    }
    					
    Pescador, F.; Maturana, G.; Garrido, M.; Juarez, E. & Sanz, C. An H.264 video decoder based on a latest generation DSP 2009 Consumer Electronics, IEEE Transactions on
    Vol. 55 (1) , pp. 205 -212  
    article DOI    
    Abstract: Latest generation DSPs are becoming more efficient, being able to improve their forerunners while reducing their internal memory size to lower the cost. In this paper, an H.264 video decoder based on a latest generation DSP is described. Both the EDMA and the memory architecture of the processor have been fully exploited to increase the execution speed. Profiling tests have been carried out by using digital TV streams and DVD transcoded sequences. The speed of the new DSP running the decoder is 16% better than that of a forerunner with 20% more internal memory running the same decoder.
    BibTeX:
    @article{2009_Pescador,
      author = {Pescador, F. and Maturana, G. and Garrido, M.J. and Juarez, E. and Sanz, C.},
      title = {An H.264 video decoder based on a latest generation DSP},
      journal = {Consumer Electronics, IEEE Transactions on},
      year = {2009},
      volume = {55},
      number = {1},
      pages = {205 -212},
      doi = {http://dx.doi.org/10.1109/TCE.2009.4814436}
    }
    					
    Seitner, F.; Bleyer, M.; Gelautz, M. & Beuschel, R. Development of a High-Level Simulation Approach and Its Application to Multicore Video Decoding 2009 Circuits and Systems for Video Technology, IEEE Transactions on
    Vol. 19 (11) , pp. 1667 -1679  
    article DOI    
    Abstract: In this paper, we introduce a high-level simulation methodology for the modeling of multicore video processing architectures. This method allows design space explorations of parallel video processing applications (VPAs). It is used to test the performance of running a VPA on arbitrary virtual hardware and software configurations. The method represents an alternative to performing a "complete" decoder implementation on a field-programmable gate array or an application-specific integrated circuit. The use of our method, therefore, yields the advantage of being considerably more time, labor, and cost efficient. As an application, we use our method for designing a parallel H.264 decoder targeting 720 p 25 resolution at bit-rates up to 50 Mb/s. Starting from a single-core decoder implementation, we use our simulator for estimating the performance gain when using a multicore architecture. We then detect the major performance bottlenecks in this multicore system and perform additional decoder splittings accordingly until we reach the targeted requirements. The use of functional splitting (i.e., pipelining) and data-parallel processing is demonstrated. The final H.264 decoder architecture is capable of fulfilling our performance requirements.
    BibTeX:
    @article{2009_Seitner,
      author = {Seitner, F.H. and Bleyer, M. and Gelautz, M. and Beuschel, R.M.},
      title = {Development of a High-Level Simulation Approach and Its Application to Multicore Video Decoding},
      journal = {Circuits and Systems for Video Technology, IEEE Transactions on},
      year = {2009},
      volume = {19},
      number = {11},
      pages = {1667 -1679},
      doi = {http://dx.doi.org/10.1109/TCSVT.2009.2031523}
    }
    					
    Venkatraman, D. & Makur, A. A compressive sensing approach to object-based surveillance video coding 2009 Acoustics, Speech and Signal Processing, 2009. ICASSP 2009. IEEE International Conference on , pp. 3513 -3516   inproceedings DOI    
    Abstract: This paper studies the feasibility and investigates various choices in the application of compressive sensing (CS) to object-based surveillance video coding. The residual object error of a video frame is a sparse signal and CS, which aims to represent information of a sparse signal by random measurements, is considered for coding of object error. This work proposes several techniques using two approaches- direct CS and transform-based CS. The techniques are studied and analyzed by varying the different trade-off parameters such as the measurement index, quantization levels etc. Finally we recommend an optimal scheme for a range of bitrates. Experimental results with comparative bitrates-vs-PSNR graphs for the different techniques are presented.
    BibTeX:
    @inproceedings{2009_Venkatraman,
      author = {Venkatraman, D. and Makur, A.},
      title = {A compressive sensing approach to object-based surveillance video coding},
      journal = {Acoustics, Speech and Signal Processing, 2009. ICASSP 2009. IEEE International Conference on},
      year = {2009},
      pages = {3513 -3516},
      doi = {http://dx.doi.org/10.1109/ICASSP.2009.4960383}
    }
    					
    Vukobratovic, D.; Stankovic, V.; Sejdinovic, D.; Stankovic, L. & Xiong, Z. Scalable Video Multicast Using Expanding Window Fountain Codes 2009 Multimedia, IEEE Transactions on
    Vol. 11 (6) , pp. 1094 -1104  
    article DOI    
    Abstract: Fountain codes were introduced as an efficient and universal forward error correction (FEC) solution for data multicast over lossy packet networks. They have recently been proposed for large scale multimedia content delivery in practical multimedia distribution systems. However, standard fountain codes, such as LT or Raptor codes, are not designed to meet unequal error protection (UEP) requirements typical in real-time scalable video multicast applications. In this paper, we propose recently introduced UEP expanding window fountain (EWF) codes as a flexible and efficient solution for real-time scalable video multicast. We demonstrate that the design flexibility and UEP performance make EWF codes ideally suited for this scenario, i.e., EWF codes offer a number of design parameters to be ldquotunedrdquo at the server side to meet the different reception criteria of heterogeneous receivers. The performance analysis using both analytical results and simulation experiments of H.264 scalable video coding (SVC) multicast to heterogeneous receiver classes confirms the flexibility and efficiency of the proposed EWF-based FEC solution.
    BibTeX:
    @article{2009_Vukobratovic,
      author = {Vukobratovic, D. and Stankovic, V. and Sejdinovic, D. and Stankovic, L. and Zixiang Xiong},
      title = {Scalable Video Multicast Using Expanding Window Fountain Codes},
      journal = {Multimedia, IEEE Transactions on},
      year = {2009},
      volume = {11},
      number = {6},
      pages = {1094 -1104},
      doi = {http://dx.doi.org/10.1109/TMM.2009.2026087}
    }
    					
    Wang, Y. & Ji, X. A new algorithm for watermarking based On DCT and chaotic scrambling 2009 Computer Science Education, 2009. ICCSE '09. 4th International Conference on , pp. 1957 -1961   inproceedings DOI    
    Abstract: This paper presents a robust image watermarking method in discrete cosine transform (DCT) domain based on chaotic sequences encryption. Exploiting some characteristics of human visual system (HVS) and DC(direct current) components having much larger perceptual capacity than any AC (alternating current)components, watermark is embedded into the DC components of the host image. First, we scramble the watermark image to avoid the block effect. Then we split the host image and the scrambled watermark image into 8times8block respectively, and the scrambled watermark is embedded into the DC components of the host image. The experimental results show that the embedded watermark is invisible and robust against noise and commonly used image processing methods such as Gaussian, JPEG compression, Median filtering etc.
    BibTeX:
    @inproceedings{2009_Wang,
      author = {Yanling Wang and Xiuhua Ji},
      title = {A new algorithm for watermarking based On DCT and chaotic scrambling},
      journal = {Computer Science Education, 2009. ICCSE '09. 4th International Conference on},
      year = {2009},
      pages = {1957 -1961},
      doi = {http://dx.doi.org/10.1109/ICCSE.2009.5228222}
    }
    					
    Wei, G. & Xiaoxu, Z. Subjective assessment of switching in H264 SVC video coding standard 2009 Computer Science Education, 2009. ICCSE '09. 4th International Conference on , pp. 1086 -1089   inproceedings DOI    
    Abstract: H.264 SVC is the scalable extension of H.264/AVC standard. This article provides an overview of the technical features of which enhanced compression performance and provision of a ldquonetwork-friendlyrdquo video representation addressing ldquoconversationalrdquo (video telephony) and ldquonon-conversationalrdquo (storage, broadcast, or streaming) applications. And we describe the subjective assessment methodologies, make a prediction about the switching on the decoder side, then give a recommendation for it.
    BibTeX:
    @inproceedings{2009_Wei,
      author = {Gong Wei and Zhang Xiaoxu},
      title = {Subjective assessment of switching in H264 SVC video coding standard},
      journal = {Computer Science Education, 2009. ICCSE '09. 4th International Conference on},
      year = {2009},
      pages = {1086 -1089},
      doi = {http://dx.doi.org/10.1109/ICCSE.2009.5228549}
    }
    					
    Xun, C.; Wen, M.; Wu, W. & Zhang, C. A Framework for Stream Programming on DSP 2009 Embedded and Multimedia Computing, 2009. EM-Com 2009. 4th International Conference on , pp. 1 -5   inproceedings DOI    
    Abstract: There has recently been much interest in stream processing, both in industry (Cell, Storm series, NVIDIA G80, AMD FIRESTREAM) and academia (IMAGINE). Some researchers have accelerated a lot of applications in media processing, scientific computing and signal processing with a special programming style called stream programming. This paper presents a framework to program DSP with this special programming style. Stream program can run on DSP without any architectural support. H264 encoding is selected to evaluate our technique. The result shows that significant speedup is achieved, ranging from 3.2x for cavlc up to 7.1x for analysis.
    BibTeX:
    @inproceedings{2009_Xun,
      author = {Changqing Xun and Mei Wen and Wei Wu and Chunyuan Zhang},
      title = {A Framework for Stream Programming on DSP},
      journal = {Embedded and Multimedia Computing, 2009. EM-Com 2009. 4th International Conference on},
      year = {2009},
      pages = {1 -5},
      doi = {http://dx.doi.org/10.1109/EM-COM.2009.5403001}
    }
    					
    Zhang, J. & Matolak, D. MLO spreading codes for CDMA systems using DCT/DST 2009 Radio and Wireless Symposium, 2009. RWS '09. IEEE , pp. 695 -698   inproceedings DOI    
    Abstract: We propose the use of the discrete cosine transform (DCT) and discrete sine transform (DST), as well as their inverses, for spreading in multi-carrier (MC) code division multiple access (CDMA) systems. The DCT matrix works as a multi-level orthogonal (MLO) spreading code set. More generally, the DCT can be used in any spread spectrum system including single-carrier CDMA systems, MC-CDMA and MC direct sequence (DS) CDMA. The use of the DCT in MC-CDMA is shown to yield better system performance than traditional binary spreading codes.
    BibTeX:
    @inproceedings{2009_Zhang,
      author = {Jingtao Zhang and Matolak, D.W.},
      title = {MLO spreading codes for CDMA systems using DCT/DST},
      journal = {Radio and Wireless Symposium, 2009. RWS '09. IEEE},
      year = {2009},
      pages = {695 -698},
      doi = {http://dx.doi.org/10.1109/RWS.2009.4957447}
    }
    					
    Zhu, M.; Jia, W.; Liu, L. & Wei, S. Data compression and the 8 #x00D7;8 integer transform 2009 ASIC, 2009. ASICON '09. IEEE 8th International Conference on , pp. 902 -905   inproceedings DOI    
    Abstract: Data compression is a kind of power concentrating process. Complex data could be transferred into eigenvalues and sparse data concentrated directly. DCT is a very efficient transform but hard to implement. 8×8 integer transform derived here is easily mapping into a parallel accelerating system, and proved to earn a better performance than existing 4×4 ones. This work implemented 8×8 integer algorithm on the REmus system, a novel reconfigurable system designed by Tsinghua, verified the architecture with several media applications. A 46.0% speedup comparing with 4×4 integer ones and a 60.1% average boost comparing with other processors are achieved. Different dividing of frames also has an impact on other algorithms, such as deblocking, the most significant process in h264 codec. It gets 62.5% boost, too. Comparing with 4×4 ones, when quantity level is smaller than 10, 8×8 restores better pictures and a 10% improvement is always found on compress ratios.
    BibTeX:
    @inproceedings{2009_Zhu,
      author = {Min Zhu and Wen Jia and Leibo Liu and Shaojun Wei},
      title = {Data compression and the 8 #x00D7;8 integer transform},
      journal = {ASIC, 2009. ASICON '09. IEEE 8th International Conference on},
      year = {2009},
      pages = {902 -905},
      doi = {http://dx.doi.org/10.1109/ASICON.2009.5351549}
    }
    					
    yong, M.; min, T.Y. & hui, Q.Y. Adaptive video watermarking algorithm based on MPEG-4 streams 2008 Control, Automation, Robotics and Vision, 2008. ICARCV 2008. 10th International Conference on , pp. 1084 -1088   inproceedings DOI    
    Abstract: Video digital watermarking is a topic of the current watermarking research. To better ensure the robustness and invisibility, an adaptive video watermarking algorithm is proposed based on MPEG-4 video compression principle and the human visual system model. The adaptive factor is designed according to the direct current coefficient and the number of low and intermediate frequency coefficients after discrete cosine transform (DCT) of I frame. For different image blocks, the different embedded intensity is used. The watermarking is embedded into the low frequency coefficients. Experiments show that the algorithm works well with the compatibility of human visual system, and the common attack is very robust as well.
    BibTeX:
    @inproceedings{2008_,
      author = {Ma yong and Tian Yu-min and Qu Yun-hui},
      title = {Adaptive video watermarking algorithm based on MPEG-4 streams},
      journal = {Control, Automation, Robotics and Vision, 2008. ICARCV 2008. 10th International Conference on},
      year = {2008},
      pages = {1084 -1088},
      doi = {http://dx.doi.org/10.1109/ICARCV.2008.4795670}
    }
    					
    An, S. & Wang, C. Recursive algorithm, architectures and FPGA implementation of the two-dimensional discrete cosine transform 2008 Image Processing, IET
    Vol. 2 (6) , pp. 286 -294  
    article DOI    
    Abstract: A new recursive algorithm and two types of circuit architectures are presented for the computation of the two-dimensional discrete cosine transform (2D DCT). The new algorithm permits to compute the 2D DCT by a simple procedure of the 1D recursive calculations involving only cosine coefficients. The recursive kernel for the proposed algorithm contains a small number of operations. Also, it requires a smaller number of pre-computed data compared with many of existing algorithms in the same category. The kernel can be easily implemented in a simple circuit block with a short critical delay path. In order to evaluate the performance improvement resulting from the new algorithm, an architecture for the 2D DCT designed by direct mapping from the computation structure of the proposed algorithm has been implemented in an FPGA board. The results show that the reduction of the hardware consumption can easily reach 25% and the clock frequency can increase 17% compared with a system implementing a recently reported 2D DCT recursive algorithm. For a further reduction of the hardware, another architecture has been proposed for the same 2D DCT computation. Using one recursive computation block to perform different functions, this architecture needs only approximately one-half of the hardware that is required in the first architecture, which has been confirmed by an FPGA implementation.
    BibTeX:
    @article{2008_An,
      author = {An, S. and Wang, C.},
      title = {Recursive algorithm, architectures and FPGA implementation of the two-dimensional discrete cosine transform},
      journal = {Image Processing, IET},
      year = {2008},
      volume = {2},
      number = {6},
      pages = {286 -294},
      doi = {http://dx.doi.org/10.1049/iet-ipr:20080057}
    }
    					
    Bianchi, T.; Piva, A. & Barni, M. Discrete cosine transform of encrypted images 2008 Image Processing, 2008. ICIP 2008. 15th IEEE International Conference on , pp. 1668 -1671   inproceedings DOI    
    Abstract: Processing a signal directly in the encrypted domain provides an elegant solution in application scenarios where valuable signals must be protected from a malicious processing device. In a previous paper we considered the implementation of the ID discrete fourier transform (DFT) in the encrypted domain, by using the homomorphic properties of the underlying cryptosystem. In this paper we extend our previous results by considering the application of the 2-dimensional DCT to encrypted images. The effect of the consecutive application of the DCT algorithm first by rows then by columns is considered, as well as the differences between the implementation of the direct DCT algorithm and its fast version. Particular attention is given to block-based DCT, with emphasis on the possibility of lowering the computational burden by parallel application of the encrypted domain DCT algorithm to different image blocks.
    BibTeX:
    @inproceedings{2008_Bianchi,
      author = {Bianchi, T. and Piva, A. and Barni, M.},
      title = {Discrete cosine transform of encrypted images},
      journal = {Image Processing, 2008. ICIP 2008. 15th IEEE International Conference on},
      year = {2008},
      pages = {1668 -1671},
      doi = {http://dx.doi.org/10.1109/ICIP.2008.4712093}
    }
    					
    Boutekkouk, F.; Bilavarn, S.; Auguin, M. & Benmohammed, M. UML profile for estimating application Worst Case Execution Time on System-on-Chip 2008 System-on-Chip, 2008. SOC 2008. International Symposium on , pp. 1 -6   inproceedings DOI    
    Abstract: Systems-on-chip (SOC) design is confronted with the problem of the so-called productivity gap. In order to cope with this problem, authors emphasize on using the unified modeling language (UML) as a system level language, so higher level of abstraction is achieved. In this context, we present a UML profile and a methodology for estimating application worst case execution time (WCET) on SOC. The proposed profile allows the designer to express hierarchy among application tasks, and most of parallelism forms that exist in typical embedded applications such as task parallelism, pipelining, and data parallelism, while making control and communication over tasks explicit. In order to estimate application WCET, the hardware platform on which application is mapped on, should be abstracted too. Consequently, each hardware component is parameterized by a set of parameters matching the abstraction level of the application. A particularity of our flow is that it starts by establishing a sequential object model using UML sequence diagram, from which a task-level model is extracted. We think that the sequential model is strongly preferred from the system designerpsilas perspective for two reasons. First, because it facilitates the modelling task relieving the designer of the burden of concurrency modelling. Secondly, starting from an existing sequential model (e.g. legacy C code) which is generally considered as the reference model, we can then parallelize it, and explore the design space. We show how we have used our profile for H264 decoder modeling.
    BibTeX:
    @inproceedings{2008_Boutekkouk,
      author = {Boutekkouk, F. and Bilavarn, S. and Auguin, M. and Benmohammed, M.},
      title = {UML profile for estimating application Worst Case Execution Time on System-on-Chip},
      journal = {System-on-Chip, 2008. SOC 2008. International Symposium on},
      year = {2008},
      pages = {1 -6},
      doi = {http://dx.doi.org/10.1109/ISSOC.2008.4694865}
    }
    					
    Bulakci, O.; Schuster, M.; Bunge, C. & Spinnler, B. Reduced complexity precoding based peak-to-average power ratio reduction applied to optical direct-detection OFDM 2008 Optical Communication, 2008. ECOC 2008. 34th European Conference on , pp. 1 -2   inproceedings DOI    
    Abstract: DCT based precoding is proposed for PAPR reduction to improve the performance of optical OFDM at limited precoding complexity. Effective PAPR reduction and the resultant OSNR performance are shown.
    BibTeX:
    @inproceedings{2008_Bulakci,
      author = {Bulakci, O. and Schuster, M. and Bunge, C.A. and Spinnler, B.},
      title = {Reduced complexity precoding based peak-to-average power ratio reduction applied to optical direct-detection OFDM},
      journal = {Optical Communication, 2008. ECOC 2008. 34th European Conference on},
      year = {2008},
      pages = {1 -2},
      doi = {http://dx.doi.org/10.1109/ECOC.2008.4729497}
    }
    					
    Chen, G.; Cheng, Y. & Wang, Y. A Robust Adaptive Video Watermarking Algorithm Based on HVS 2008 Young Computer Scientists, 2008. ICYCS 2008. The 9th International Conference for , pp. 1424 -1428   inproceedings DOI    
    Abstract: Digital watermarking can be used to protect the intellectual property for multimedia data. In this paper, a robust content-adaptive watermarking algorithm based on human visual system (HVS) is presented for video copyright protection. By comparing the direct current discrete cosine transform (DCT) coefficients of the intra frame and its neighbor frame of a given video, a fast moving image block set was chosen firstly. Then watermarking bits were adaptively embedded into the DCT domain of image blocks in the fast-moving block set, which were decided based on HVS and the characteristics of middle frequency DCT coefficients of these blocks. The experimental results indicate that the proposed algorithm is highly robust against MPEG-2 compression and frame's geometrical distortions, such as rotation, aspect ratio and bending.
    BibTeX:
    @inproceedings{2008_Chen,
      author = {Guangxi Chen and Yan Cheng and Yinghao Wang},
      title = {A Robust Adaptive Video Watermarking Algorithm Based on HVS},
      journal = {Young Computer Scientists, 2008. ICYCS 2008. The 9th International Conference for},
      year = {2008},
      pages = {1424 -1428},
      doi = {http://dx.doi.org/10.1109/ICYCS.2008.201}
    }
    					
    Dhahri, S.; Zitouni, A.; Elhaji, M. & Tourki, R. Adaptive motion estimator for the H264 coder 2008 Design and Technology of Integrated Systems in Nanoscale Era, 2008. DTIS 2008. 3rd International Conference on , pp. 1 -5   inproceedings DOI    
    Abstract: In this paper we propose an adaptive motion estimation circuit that can be dynamically reconfigured by the best algorithm based on the video nature of the application under running. This configuration is based on a recognition step that collects the motion information's of each block from its same block in the previous images and their neighbour's blocs in the current image. The proposed circuit has been designed at Register Transfer Level (RTL) level by using the VHDL language and can be configured by three algorithms (full search, spiral search with threshold and spiral search without threshold). After being integrated into a H264 codec, experimental results show that the adaptive motion estimator allows better results in term of peak signal to noise ratio (PSNR) and compression ratio relatively to the static based estimators.
    BibTeX:
    @inproceedings{2008_Dhahri,
      author = {Dhahri, S. and Zitouni, A. and Elhaji, M. and Tourki, R.},
      title = {Adaptive motion estimator for the H264 coder},
      journal = {Design and Technology of Integrated Systems in Nanoscale Era, 2008. DTIS 2008. 3rd International Conference on},
      year = {2008},
      pages = {1 -5},
      doi = {http://dx.doi.org/10.1109/DTIS.2008.4540220}
    }
    					
    Ding, J.-R.; Chen, J.-Y.; Yang, F.-C. & Yang, J.-F. Two-layer and adaptive entropy coding algorithms for H.264-based lossless image coding 2008 Acoustics, Speech and Signal Processing, 2008. ICASSP 2008. IEEE International Conference on , pp. 1369 -1372   inproceedings DOI    
    Abstract: In this paper, we propose two-layer coding algorithm to improve the performance of the H.264-based lossless (H.264-LS) image coding. From universal access point of view, the proposed method is based on the H.264 lossy image coding with other CABAC layer to compensate the lossy portion. Besides, the H.264-LS with DPCM (H264-LS_DPCM) and H.264-LS achieve different coding performance in use of CABAC and CAVLC entropy coders without the DCT and quantization. We further suggest an adaptive entropy coding (AEC) algorithm to determine the best entropy coder by using the image content variations, which is calculated from the sum of absolute difference of intra prediction. Simulation results show that the proposed AEC method have good correct detection rates and improvement of compression rate for H.264-LS and H.264-LS_DPCM coders. The two-layer H.264-LS almost have the same compression rate than the H.264-LS_DPCM.
    BibTeX:
    @inproceedings{2008_Ding,
      author = {Jun-Ren Ding and Jiun-Yu Chen and Fu-Chun Yang and Jar-Ferr Yang},
      title = {Two-layer and adaptive entropy coding algorithms for H.264-based lossless image coding},
      journal = {Acoustics, Speech and Signal Processing, 2008. ICASSP 2008. IEEE International Conference on},
      year = {2008},
      pages = {1369 -1372},
      doi = {http://dx.doi.org/10.1109/ICASSP.2008.4517873}
    }
    					
    Dziri, A.; Diallo, A.; Kieffer, M. & Duhamel, P. P-Picture Based H.264 AVC to H.264 SVC Temporal Transcoding 2008 Wireless Communications and Mobile Computing Conference, 2008. IWCMC '08. International , pp. 425 -430   inproceedings DOI    
    Abstract: This paper proposes a transcoding method from a standard (P-picture based) H264 advanced video coder (AVC) bitstream to a H264 scalable video coder (SVC) bitstream with temporal scalability. This method avoids full AVC decoding followed by full SVC re-encoding and is performed in two steps. The IPPP... AVC bit stream is first transcoded to an IpPpP AVC bitstream, where P denotes a reference image and p is a non reference one. Then, the IpPpP... AVC bitstream is transformed to an IpPpP SVC bitstream by syntax adaptation. The resulting SVC bitstream is thus fully compatible and can be decoded e.g., by the JSVM reference software tool. We present a theoretical analysis of the first step of the proposed transcoder with two results. First, we prove that the transcoding process does not introduce any error propagation. Second, we present an evaluation of the transcoding errors bounds compared to full AVC decoding followed by full AVC re-encoding process, denoted in the rest of the paper as the perfect transcoding. Simulation and experimental results are presented for "foreman" CIF format sequence encoded at a frame rate of 30 frames per second. Performances are discussed in terms of coding efficiency and algorithm complexity. Compared to IPPP AVC to IpPpP AVC perfect transcoding, the proposed transcoder presents a 47% complexity reduction with almost the same quality for the enhancement layer. The resulting bitstream thus benefits of the traditional properties of scalable videos. Moreover, the base layer obtained from the decoding of the AVC/SVC transcoded bitstream presents an equivalent performance to that obtained from full AVC-decoding followed by AVC re-encoding of the even frames.
    BibTeX:
    @inproceedings{2008_Dziri,
      author = {Dziri, A. and Diallo, A. and Kieffer, M. and Duhamel, P.},
      title = {P-Picture Based H.264 AVC to H.264 SVC Temporal Transcoding},
      journal = {Wireless Communications and Mobile Computing Conference, 2008. IWCMC '08. International},
      year = {2008},
      pages = {425 -430},
      doi = {http://dx.doi.org/10.1109/IWCMC.2008.74}
    }
    					
    Elrefaey, L.; Allam, M.; Kader, H. & Selim, M. Robust blind image-adaptive watermarking 2008 Radio Science Conference, 2008. NRSC 2008. National , pp. 1 -13   inproceedings DOI    
    Abstract: This paper proposes a new blind, robust image watermarking for copyright protection and ownership verification in the DCT domain. The owner information is in the form of a logo image. Two different watermarking systems are considered. The first one is based on the principle of spread spectrum communication, in which we propose an adaptive direct sequence spread spectrum DSSS embedding, and a correlation based decoding. The second one is based on Spread Transform (ST) which is a hybrid combination between spread spectrum and a quantization watermarking scheme called Scalar Costa Scheme (SCS). We describe how to incorporate a perceptual model, based on Watson's perceptual model, into the frameworks. The proposed systems are tested using different standard test images and the performance is evaluated under JPEG compression and compared to each other and to other watermarking schemes. Experimental Results show that the proposed systems have good imperceptibility and higher robustness to JPEG image compression.
    BibTeX:
    @inproceedings{2008_Elrefaey,
      author = {Elrefaey, L.A. and Allam, M.E. and Kader, H.A. and Selim, M.},
      title = {Robust blind image-adaptive watermarking},
      journal = {Radio Science Conference, 2008. NRSC 2008. National},
      year = {2008},
      pages = {1 -13},
      doi = {http://dx.doi.org/10.1109/NRSC.2008.4542331}
    }
    					
    Fan, M. & Wang, H. Centroid-based robust audio watermarking scheme 2008 Audio, Language and Image Processing, 2008. ICALIP 2008. International Conference on , pp. 476 -479   inproceedings DOI    
    Abstract: In this paper, we propose a novel robust audio watermarking scheme based on DCT (discrete cosine transform). The key point lies in the fact that centroid of each audio frame will not be changed much in order to maintain high auditory quality. Each audio segment carrying the centroid of one audio frame is performed with DCT. Then the direct-current component of each audio segment is embedded scrambled watermark bit by quantization method. The extraction process of the robust watermark doesnpsilat need original audio signal. Experimental results show good transparency and high robustness against various common audio signal processing attacks of this scheme, and the stable essence of audio centroid under these various attacks.
    BibTeX:
    @inproceedings{2008_Fan,
      author = {Mingquan Fan and Hongxia Wang},
      title = {Centroid-based robust audio watermarking scheme},
      journal = {Audio, Language and Image Processing, 2008. ICALIP 2008. International Conference on},
      year = {2008},
      pages = {476 -479},
      doi = {http://dx.doi.org/10.1109/ICALIP.2008.4590170}
    }
    					
    Hamel, M.; Fontaine, R. & Boissy, P. In-home telerehabilitation for geriatric patients 2008 Engineering in Medicine and Biology Magazine, IEEE
    Vol. 27 (4) , pp. 29 -37  
    article DOI    
    Abstract: We investigate in this study, following the positive results from a proof-of-concept study [7], the effectiveness of providing in-home telerehabilitation services as an alternative to home care visits for physical therapy in orthopedic conditions following discharge from an acute care hospital and rehabilitation unit [8]. Based on the results from the initial proof- of-concept study and a user-centered design approach, a telerehabilitation platform was developed consisting of two H264 videoconferencing codecs (Tandberg 500 MXP) with integrated wide-angle view cameras and remotely controlled pan tilt zoom (PTZ) functions, local and remote computers with Digital Object Identifier 10.1109/MEMB.200S.919491 dedicated modular software interfaces for user-friendly control of videoconferencing connections, PTZ camera function, and external devices (i.e., tablet PC and sensors).
    BibTeX:
    @article{2008_Hamel,
      author = {Hamel, M. and Fontaine, R. and Boissy, P.},
      title = {In-home telerehabilitation for geriatric patients},
      journal = {Engineering in Medicine and Biology Magazine, IEEE},
      year = {2008},
      volume = {27},
      number = {4},
      pages = {29 -37},
      doi = {http://dx.doi.org/10.1109/MEMB.2008.919491}
    }
    					
    Ilbery, P.; Taubman, D. & Bradley, A. Mixed content image compression by gradient field integration 2008 Image Processing, 2008. ICIP 2008. 15th IEEE International Conference on , pp. 1053 -1056   inproceedings DOI    
    Abstract: Sketch based image coding decomposes an input image into a piece-wise smooth approximation image and a residual image. Image compression by gradient field integration follows this model, but differs by generating the approximation image from gradient data along edge contours and regularly sampled low resolution image data. This allows direct and efficient calculation of an approximation image which is smooth between edges. In this paper, we describe the image compression by gradient field integration approach, together with a low complexity implementation intended for near visually lossless compression of mixed content images. The implementation uses gradient field integration by scanline convolution, simple edge data extraction, DCT residual image coding and is block-based; it is suited to compression of images containing sharp edges occurring along pixel borders. Compression results are provided indicating potential gains from this method.
    BibTeX:
    @inproceedings{2008_Ilbery,
      author = {Ilbery, P. and Taubman, D. and Bradley, A.},
      title = {Mixed content image compression by gradient field integration},
      journal = {Image Processing, 2008. ICIP 2008. 15th IEEE International Conference on},
      year = {2008},
      pages = {1053 -1056},
      doi = {http://dx.doi.org/10.1109/ICIP.2008.4711939}
    }
    					
    Jubran, M.; Bansal, M. & Kondi, L. Low-Delay Low-Complexity Bandwidth-Constrained Wireless Video Transmission Using SVC Over MIMO Systems 2008 Multimedia, IEEE Transactions on
    Vol. 10 (8) , pp. 1698 -1707  
    article DOI    
    Abstract: We propose an efficient strategy for the transmission of scalable video over multiple-input multiple-output (MIMO) wireless systems. In this paper, we use the latest scalable H.264 codec (SVC), which provides combined temporal, quality and spatial scalability. At the transmitter, we estimate the decoded video distortion for given channel conditions taking into account the effects of quantization, packet loss and error concealment. The proposed scalable decoder distortion algorithm offers low delay and low complexity. The performance of this method is validated using experimental results. In our proposed system, we use a MIMO system with orthogonal space-time block codes (O-STBC) that provides spatial diversity and guarantees independent transmission of different symbols within the block code. The bandwidth constrained allocation problem considered here is simplified and solved for one O-STBC symbol at a time. Furthermore, we take the advantage of the hierarchical structure of SVC to attain the optimal solution for each group of pictures (GOP) of the video sequence. We incorporate the estimated decoder distortion to optimally select the application layer parameter, i.e., quantization parameter (QP), and physical layer parameters, i.e., channel coding rate and modulation type for wireless video transmission.
    BibTeX:
    @article{2008_Jubran,
      author = {Jubran, M.K. and Bansal, M. and Kondi, L.P.},
      title = {Low-Delay Low-Complexity Bandwidth-Constrained Wireless Video Transmission Using SVC Over MIMO Systems},
      journal = {Multimedia, IEEE Transactions on},
      year = {2008},
      volume = {10},
      number = {8},
      pages = {1698 -1707},
      doi = {http://dx.doi.org/10.1109/TMM.2008.2007317}
    }
    					
    Kamel, M.; Salah, T. & El-Bey, B. Material Architecture Proposition for the Block Matching Method of Motion Estimate in H264 Standard 2008 Information and Communication Technologies: From Theory to Applications, 2008. ICTTA 2008. 3rd International Conference on , pp. 1 -6   inproceedings DOI    
    Abstract: In this paper, we proposed hardware architecture for the most complicated bloc in the standard of video coding H264, it is the block of motion estimate and compensation. The proposed architecture has the advantage of gaining in the computation time and memory space for recording images, and even the advantage of keeping the same memory addresses for the different blocks object of treatment in the different images, which is one of many views physical realization. The architecture proposition also serves as reading block images of a video in real time. Also this architecture can be useful in the overall chain of dynamic image processing in a hardware/software hybrid diagram.
    BibTeX:
    @inproceedings{2008_Kamel,
      author = {Kamel, M. and Salah, T. and El-Bey, B.},
      title = {Material Architecture Proposition for the Block Matching Method of Motion Estimate in H264 Standard},
      journal = {Information and Communication Technologies: From Theory to Applications, 2008. ICTTA 2008. 3rd International Conference on},
      year = {2008},
      pages = {1 -6},
      doi = {http://dx.doi.org/10.1109/ICTTA.2008.4530092}
    }
    					
    Wahid, K.; Ko, S.-B.; Teng, D. & Dimitrov, V. Low-area and low-power video compressor for endoscopic capsules 2008 Electrical and Computer Engineering, 2008. CCECE 2008. Canadian Conference on , pp. 000507 -000510   inproceedings DOI    
    Abstract: The paper presents a hardware implementation of a low-power and low-area video compressor for capsule endoscopy application. The architecture uses a direct mapping to compute the two-dimensional Discrete Cosine Transform which eliminates the need of transpose operation and results in reduced area and low processing time. Instead of using the conventional binary finite precision approach, an error-free algebraic integer encoding has been used that guarantees lossless computation. The corresponding quantization table that complies with the JPEG standard has also been developed. The processor costs only 3466 cells, runs at a maximum frequency of 151 MHz, and consumes 10.30 mW of power. The test results of several endoscopic images show that higher compression ratio (over 90%) can be achieved with high quality image reconstruction.
    BibTeX:
    @inproceedings{2008_KhanWahid,
      author = {Khan Wahid, and Seok-Bum Ko, and Teng, Daniel and Dimitrov, Vassil},
      title = {Low-area and low-power video compressor for endoscopic capsules},
      journal = {Electrical and Computer Engineering, 2008. CCECE 2008. Canadian Conference on},
      year = {2008},
      pages = {000507 -000510},
      doi = {http://dx.doi.org/10.1109/CCECE.2008.4564586}
    }
    					
    Khirallah, C.; Stankovic, L. & Stankovic, V. Low complexity multi-view video streaming over multipath fading channels 2008 Information and Communication Technologies: From Theory to Applications, 2008. ICTTA 2008. 3rd International Conference on , pp. 1 -5   inproceedings DOI    
    Abstract: This paper addresses the problem of wireless delivery of a captured scene from multiple cameras, which do not communicate among one other, to a joint decoder. Correlation among different camera views calls for distributed source coding for efficient compression. The fact that the cameras may be placed within a short communication range of each other results in a high level of interference, multipath fading, and noise effects, during transmission. We develop a new two-camera system, where one camera employs H.264 SVC, whereas the other, low-complexity camera codec exploits statistical dependencies between closely-located cameras' views to jointly perform compression and spreading using Complete Complementary (CC) sequences for protection against interference. This paper proposes a very low complexity scalable coding scheme using fixed integer Discrete Cosine Transform (DCT) and fixed quantizer matrices instead of the 2D DCT. The proposed system is targeted at wireless multi-view video streaming applications, in which the data acquired form different viewing cameras are highly correlated spatially and temporally. Our results indicate that our proposed system is competitive with H264 SVC coded streams.
    BibTeX:
    @inproceedings{2008_Khirallah,
      author = {Khirallah, C. and Stankovic, L. and Stankovic, V.},
      title = {Low complexity multi-view video streaming over multipath fading channels},
      journal = {Information and Communication Technologies: From Theory to Applications, 2008. ICTTA 2008. 3rd International Conference on},
      year = {2008},
      pages = {1 -5},
      doi = {http://dx.doi.org/10.1109/ICTTA.2008.4530252}
    }
    					
    Kim, Y.-E.; Cho, S.-H. & Chung, J.-G. Modified CSD group multiplier design for predetermined coefficient groups 2008 Circuits and Systems, 2008. ISCAS 2008. IEEE International Symposium on , pp. 3362 -3365   inproceedings DOI    
    Abstract: Some digital signal processing applications, such as FFT, request multiplications with a group (or, groups) of a few predetermined coefficients. In this paper, based on a grouping method of CSD coefficients, an efficient multiplier design method for predetermined coefficient groups is proposed. In the case of the multiplier design for sine-cosine generator used in direct digital frequency synthesizer(DDFS), it is shown that by the proposed method, area, power and delay time can be reduced by 53.1%, 45.6% and 22.6%, respectively, compared with conventional design. Also, in the case of multiplier design used in 128 point radix-24 FFT, the area, power and delay time can be reduced by 42.9%, 58.5% and 19.7%, respectively.
    BibTeX:
    @inproceedings{2008_Kim,
      author = {Yong-Eun Kim and Su-Hyun Cho and Jin-Gyun Chung},
      title = {Modified CSD group multiplier design for predetermined coefficient groups},
      journal = {Circuits and Systems, 2008. ISCAS 2008. IEEE International Symposium on},
      year = {2008},
      pages = {3362 -3365},
      doi = {http://dx.doi.org/10.1109/ISCAS.2008.4542179}
    }
    					
    Kuo, T.-Y. & Lu, H.-J. Efficient Reference Frame Selector for H.264 2008 Circuits and Systems for Video Technology, IEEE Transactions on
    Vol. 18 (3) , pp. 400 -405  
    article DOI    
    Abstract: This paper proposes a simple yet effective mechanism to select proper reference frames for H.264 motion estimation. Unlike traditional video codecs, H.264 permits more than one reference frame for increased precision in motion estimation. However, motion estimation is complicated by variable block-size motion estimation, which requires significant encoding complexity to identify the best inter-coding. Our smart selection mechanism selects suitable reference frames by means of a simple test, and only the selected frames will be searched further in the variable block size motion estimation. One major advantage of our mechanism is that it enables working with any existing motion search algorithms developed for the traditional single reference frame. Experimental results demonstrate the effectiveness of our proposed algorithm.
    BibTeX:
    @article{2008_Kuo,
      author = {Tien-Ying Kuo and Hsin-Ju Lu},
      title = {Efficient Reference Frame Selector for H.264},
      journal = {Circuits and Systems for Video Technology, IEEE Transactions on},
      year = {2008},
      volume = {18},
      number = {3},
      pages = {400 -405},
      doi = {http://dx.doi.org/10.1109/TCSVT.2008.918111}
    }
    					
    Prats, F.; Reguant, V.; de Pozuelo, R.; Margalef, F. & Ubiergo, G. Intelligent playout for the optimal IP distribution of H264 SVC/MDC services over broadband and broadcast platforms 2008 Broadband Multimedia Systems and Broadcasting, 2008 IEEE International Symposium on , pp. 1 -5   inproceedings DOI    
    Abstract: This paper introduces an approach for the optimal management of a H264 SVC (Scalable Video Coding)/MDC (Multiple Description Coding) playout. The solution presented uses optimization and control strategies depending on the different type of delivered services, the terminals that will consume these services, the load process of the video servers and the network conditions.
    BibTeX:
    @inproceedings{2008_Prats,
      author = {Prats, F.E. and Reguant, V.D. and de Pozuelo, R.M. and Margalef, F.P. and Ubiergo, G.F.},
      title = {Intelligent playout for the optimal IP distribution of H264 SVC/MDC services over broadband and broadcast platforms},
      journal = {Broadband Multimedia Systems and Broadcasting, 2008 IEEE International Symposium on},
      year = {2008},
      pages = {1 -5},
      doi = {http://dx.doi.org/10.1109/ISBMSB.2008.4536636}
    }
    					
    Reguant, V.; Prats, F.; de Pozuelo, R.; Margalef, F. & Ubiergo, G. Delivery of H264 SVC/MDC streams over wimax and DVB-T networks 2008 Consumer Electronics, 2008. ISCE 2008. IEEE International Symposium on , pp. 1 -4   inproceedings DOI    
    Abstract: This paper introduces an approach for the optimal delivery (encapsulation and signalling) of video streams coded using H.264 scalable video coding (SVC) combined with multiple description coding (MDC). The solution presented uses optimization and control strategies depending on the different type of delivered services, the terminals that will consume these services, the load process of the video servers and the network conditions.
    BibTeX:
    @inproceedings{2008_Reguant,
      author = {Reguant, V.D. and Prats, F.E. and de Pozuelo, R.M. and Margalef, F.P. and Ubiergo, G.F.},
      title = {Delivery of H264 SVC/MDC streams over wimax and DVB-T networks},
      journal = {Consumer Electronics, 2008. ISCE 2008. IEEE International Symposium on},
      year = {2008},
      pages = {1 -4},
      doi = {http://dx.doi.org/10.1109/ISCE.2008.4559440}
    }
    					
    Shin, I.; Yu, J. & Hong, J.W. A fast down-sampling method for arbitrary-ratio spatial scalability based on type-II DCT 2008 Acoustics, Speech and Signal Processing, 2008. ICASSP 2008. IEEE International Conference on , pp. 685 -688   inproceedings DOI    
    Abstract: In a scalable video coding (SVC), spatial down/up-sampling of video sequences is an essential process for spatial scalability. We propose an arbitrary-ratio spatial down-sampling method based on the type-II DCT, with which down-sampling of H.264 SVC can be efficient. Simple windowing operation in the DCT domain is efficiently embedded in the proposed down-sampling method for aliasing reduction. A fast algorithm for arbitrary- ratio down-sampling is introduced with matrix decomposition and fast DCT. It achieves 80% reduction of computation complexity in comparison with the direct matrix calculation.
    BibTeX:
    @inproceedings{2008_Shin,
      author = {Ilhong Shin and Jeongju Yu and Jin Woo Hong},
      title = {A fast down-sampling method for arbitrary-ratio spatial scalability based on type-II DCT},
      journal = {Acoustics, Speech and Signal Processing, 2008. ICASSP 2008. IEEE International Conference on},
      year = {2008},
      pages = {685 -688},
      doi = {http://dx.doi.org/10.1109/ICASSP.2008.4517702}
    }
    					
    Teixeira, L. & Corte-Real, L. A novel approach to joint video coding 2008 Visual Information Engineering, 2008. VIE 2008. 5th International Conference on , pp. 824 -828   inproceedings    
    Abstract: It is expected that future delivery of Digital TV signals will use H.264. This paper presents a novel coding scheme for multi-program video transmission in which the channel capacity is distributed among the programs according to the program complexities resulting in a more uniform overall image quality. A complexity bit rate control algorithm based on the Structural Similarity Index (SSIM) is proposed. SSIM metric is presented under the hypothesis that the Human Visual System (HSV) is very specialized in extracting structural information from a video sequence but not in extracting the errors. Thus, a measurement on structural distortion should give a better correlation to the subjective impression. Computer simulations have demonstrated very promising results showing joint coding is able to effectively control the complexity of the multi-program encoding process whilst improving overall subjective compared to independent coding and algorithms based on traditional distortion/quality metrics.
    BibTeX:
    @inproceedings{2008_Teixeira,
      author = {Teixeira, Luis and Corte-Real, Luis},
      title = {A novel approach to joint video coding},
      journal = {Visual Information Engineering, 2008. VIE 2008. 5th International Conference on},
      year = {2008},
      pages = {824 -828}
    }
    					
    Tiu, C.-M.; Jong, T.-L. & Hsieh, C.-W. Self organizing map neural network with fuzzy screening for micro-calcifications detection on mammograms 2008 Soft Computing in Industrial Applications, 2008. SMCia '08. IEEE Conference on , pp. 421 -425   inproceedings DOI    
    Abstract: Mammography remains the main screening tool for detecting breast cancer. Depicting micro-calcifications is one of the major roles of mammography. The purpose of the study is to utilize some image processing techniques to enhance the detection of micro-calcifications on mammograms. Discrete wavelet transform and difference of Gaussian filter were applied to enhance mammograms and hybrid of spatial and frequency features with self-organizing map of neural network were used to estimate the efficiency of locating micro-calcifications. In the enhancement process, the raw image was enhanced by gradient enhancement, mean contrast enhancement, and discrete wavelet transform and difference of Gaussian filter. Mean, variance, direct cosine transform coefficients, and entropy- these were extracted in the assessment stage. Finally, a self-organizing map neural network with fuzzy criterion classifier was adopted to classify the regions with similar characteristics. Twenty mammograms, with different mammographic patterns and densities, were evaluated with common agreement of the breast imaging reporting and data system categories by two radiologists in simulation. The survey revealed the micro-calcification regions had a good clustering property in self-organizing map neural network index.
    BibTeX:
    @inproceedings{2008_Tiu,
      author = {Chui-Mei Tiu and Tai-Lang Jong and Chi-Wen Hsieh},
      title = {Self organizing map neural network with fuzzy screening for micro-calcifications detection on mammograms},
      journal = {Soft Computing in Industrial Applications, 2008. SMCia '08. IEEE Conference on},
      year = {2008},
      pages = {421 -425},
      doi = {http://dx.doi.org/10.1109/SMCIA.2008.5046001}
    }
    					
    Wahid, K.; Ko, S.-B. & Teng, D. Efficient hardware implementation of an image compressor for wireless capsule endoscopy applications 2008 Neural Networks, 2008. IJCNN 2008. (IEEE World Congress on Computational Intelligence). IEEE International Joint Conference on , pp. 2761 -2765   inproceedings DOI    
    Abstract: The paper presents an area- and power-efficient implementation of an image compressor for wireless capsule endoscopy application. The architecture uses a direct mapping to compute the two-dimensional discrete cosine transform which eliminates the need of transpose operation and results in reduced area and low processing time. The algorithm has been modified to comply with the JPEG standard and the corresponding quantization tables have been developed and the architecture is implemented using the CMOS 0.18um technology. The processor costs less than 3.5k cells, runs at a maximum frequency of 150 MHz, and consumes 10 mW of power. The test results of several endoscopic colour images show that higher compression ratio (over 85%) can be achieved with high quality image reconstruction (over 30 dB).
    BibTeX:
    @inproceedings{2008_Wahid,
      author = {Wahid, K. and Seok-Bum Ko and Teng, D.},
      title = {Efficient hardware implementation of an image compressor for wireless capsule endoscopy applications},
      journal = {Neural Networks, 2008. IJCNN 2008. (IEEE World Congress on Computational Intelligence). IEEE International Joint Conference on},
      year = {2008},
      pages = {2761 -2765},
      doi = {http://dx.doi.org/10.1109/IJCNN.2008.4634186}
    }
    					
    Yi, Y. & Song, B.C. High-Speed CAVLC Encoder for 1080p 60-Hz H.264 Codec 2008 Signal Processing Letters, IEEE
    Vol. 15 , pp. 891 -894  
    article DOI    
    Abstract: In H.264/AVC and the variants, the coding of context-based adaptive variable length codes (CAVLC) requires demanding operations, particularly at high bitrates such as 100 Mbps. This letter presents two approaches to accelerate the coding operation substantially. Firstly, in the architectural aspect, we propose component-level parallelism and pipeline techniques capable of processing high-bitrate video data in a macroblock (MB)-level pipelined codec architecture. The second approach focuses on a specific part of the coding process, i.e., the residual block coding, in which the coefficient levels are coded without using look-up tables so we minimize the pertaining logic depth in the critical path, and we achieve higher operating clock frequencies. Additionally, two coefficient levels are processed in parallel by exploiting a look-ahead technique. The resulting architecture, merged in the MB-level pipelined codec system, is capable of coding up to 100 Mbps bitstreams in real-time, thus accommodating the real-time encoding of 1080p@60 Hz video.
    BibTeX:
    @article{2008_Yi,
      author = {Yongseok Yi and Byung Cheol Song},
      title = {High-Speed CAVLC Encoder for 1080p 60-Hz H.264 Codec},
      journal = {Signal Processing Letters, IEEE},
      year = {2008},
      volume = {15},
      pages = {891 -894},
      doi = {http://dx.doi.org/10.1109/LSP.2008.2001982}
    }
    					
    Zaggoulos, G.; Tran, M. & Nix, A. Mobile WiMAX system performance - simulated versus experimental results 2008 Personal, Indoor and Mobile Radio Communications, 2008. PIMRC 2008. IEEE 19th International Symposium on , pp. 1 -5   inproceedings DOI    
    Abstract: This paper addresses the downlink performance of mobile WiMAX operating at 2.3 GHz in an urban environment. The analysis includes a comparison of simulated and experimental results. Simulated packet error rate (PER) versus signal to noise ratio (SNR) graphs are generated on a per link-speed basis using a fully compliant 512 carrier mobile WiMAX simulator. Experimental data is gathered using a carrier-class basestation, a mobile-WiMAX enabled laptop, and a suite of application layer logging software. An H264 AVC encoder and IP packetisation unit is used to transmit video to a mobile client. Results show strong agreement in terms of simulated and captured PER. Using this data, the downlink operating range is evaluated as a function of the Effective Isotropic Radiated Power (EIRP) and path loss exponent. Results indicate that at low EIRP (32 dBm) the expected outdoor operating range is around 200-400 m. Applying the UK OFCOM regulations for licensed operation in the 2.5 GHz band, downlink operation in excess of 2 km can be achieved.
    BibTeX:
    @inproceedings{2008_Zaggoulos,
      author = {Zaggoulos, G. and Tran, M. and Nix, A.},
      title = {Mobile WiMAX system performance - simulated versus experimental results},
      journal = {Personal, Indoor and Mobile Radio Communications, 2008. PIMRC 2008. IEEE 19th International Symposium on},
      year = {2008},
      pages = {1 -5},
      doi = {http://dx.doi.org/10.1109/PIMRC.2008.4699670}
    }
    					
    Zeng, G. & Qiu, Z. Image Watermarking Based on DC Component in DCT 2008 Intelligent Information Technology Application Workshops, 2008. IITAW '08. International Symposium on , pp. 573 -576   inproceedings DOI    
    Abstract: The peak signal-to-noise ratio (PSNR) is a simple and widely used fidelity measure in the image watermarking. The position of watermarking plays an important role when the tradeoff between robustness and invisibility is considered. In the paper we analyze the mathematical relationship between the embeddable position and the PSNR in the DCT (discrete cosine transform) domain. We consider DC (direct current) coefficient to tradeoff robustness and invisibility. And a blind watermarking scheme is implemented through quantization index modulation (QIM) technology. Experimental results show watermarking is robust to common signal processing operation such as additional noise, filtering, resizing and JPEG compressing, and invisibility is satisfaction.
    BibTeX:
    @inproceedings{2008_Zeng,
      author = {Gaorong Zeng and Zhengding Qiu},
      title = {Image Watermarking Based on DC Component in DCT},
      journal = {Intelligent Information Technology Application Workshops, 2008. IITAW '08. International Symposium on},
      year = {2008},
      pages = {573 -576},
      doi = {http://dx.doi.org/10.1109/IITA.Workshops.2008.126}
    }
    					
    Zeng, G. & Qiu, Z. Audio watermarking in DCT: Embedding strategy and algorithm 2008 Signal Processing, 2008. ICSP 2008. 9th International Conference on , pp. 2193 -2196   inproceedings DOI    
    Abstract: The position of watermarking plays an important role when the tradeoff between robustness and inaudibility is considered. In the paper we analyze the mathematical relationship between the embeddable position and the noise of the audio watermarking in the DCT (discrete cosine transform) domain. We consider DC (direct current) coefficient to tradeoff robustness and inaudibility. And a blind watermarking scheme is implemented through quantization index modulation (QIM) technology. Experimental results show watermarking is robust to common signal processing operation and mp3 compressing, and inaudibility is satisfaction.
    BibTeX:
    @inproceedings{2008_Zenga,
      author = {Gaorong Zeng and Zhengding Qiu},
      title = {Audio watermarking in DCT: Embedding strategy and algorithm},
      journal = {Signal Processing, 2008. ICSP 2008. 9th International Conference on},
      year = {2008},
      pages = {2193 -2196},
      doi = {http://dx.doi.org/10.1109/ICOSP.2008.4697583}
    }
    					
    Zhuang, Y.; Ikenaga, T. & Goto, S. Rate Estimation of RD Optimization for Intra Mode Decision of H.264/AVC 2008 Image and Signal Processing, 2008. CISP '08. Congress on
    Vol. 2 , pp. 100 -104  
    inproceedings DOI    
    Abstract: H.264 is the latest international video coding standard. It is a joint work by the ITUT Video Coding Experts Group and the ISO/IEC Moving Picture Experts Group. H.264 contains a number of new features which make it compress video sequences much more effectively than previous standards and can provide more flexibility for application to a wide variety of network environments. But the H.264 standard also brings much more computations, especially when the rate-distortion optimization function is on. In order to reduce the complexity of rate-distortion cost of the rate part, we propose al rate estimation methods to avoid doing the real entropy coding method. The estimation method proposed is based on the notification of CAVLC tables, and is changed adaptively according to total_coeff. The final result shows that the proposed estimation method can reduce the encoding computation by 50 #x025; through the intra coding with ignorable degradation of coding performance.
    BibTeX:
    @inproceedings{2008_Zhuang,
      author = {Zhuang, Yan and Ikenaga, Takeshi and Goto, Satoshi},
      title = {Rate Estimation of RD Optimization for Intra Mode Decision of H.264/AVC},
      journal = {Image and Signal Processing, 2008. CISP '08. Congress on},
      year = {2008},
      volume = {2},
      pages = {100 -104},
      doi = {http://dx.doi.org/10.1109/CISP.2008.617}
    }
    					
    Zrida, H.; Abid, M.; Ammri, A. & Jemai, A. A YAPI-KPN parallel model of a H264/AVC video encoder 2008 Research in Microelectronics and Electronics, 2008. PRIME 2008. Ph.D. , pp. 109 -112   inproceedings DOI    
    Abstract: H264/AVC (advanced video codec) is a new video coding standard developed by a joint effort of the ITU-TVCEG and ISO/IEC MPEG. This standard provides higher coding efficiency relative to former standards at the expense of higher computational requirements. This paper presents first a high-level complexity analysis of a H264 video encoder allowing for complexity reduction at the high system level. The complexity of the obtained cost-efficient configuration outlines the potential of using multi processor platforms for the execution of a parallel model of the encoder. For this, a YAPI-level parallel Kahn process network (KPN) model is proposed, implemented and validated at high-level using the YAPI library Programming Interface.
    BibTeX:
    @inproceedings{2008_Zrida,
      author = {Zrida, H.K. and Abid, M. and Ammri, A.C. and Jemai, A.},
      title = {A YAPI-KPN parallel model of a H264/AVC video encoder},
      journal = {Research in Microelectronics and Electronics, 2008. PRIME 2008. Ph.D.},
      year = {2008},
      pages = {109 -112},
      doi = {http://dx.doi.org/10.1109/RME.2008.4595737}
    }
    					
    Chen, Y.; Cao, X.; Xie, Q. & Peng, C. An Area Efficient High Performance DCT Distributed Architecture for Video Compression 2007 Advanced Communication Technology, The 9th International Conference on
    Vol. 1 , pp. 238 -241  
    inproceedings DOI    
    Abstract: Discrete cosine transform (DCT), which is an important component of image and video compression, is adopted in various standardized coding schemes, such as JPEG, MPEGx and H.26x. But when compute a two-dimensional (2D) DCT, a large number of multiplications and additions are required in the direct approach. Multiplications, which are the most time-consuming operations in simple processor, can be completely avoided in the proposed architecture for real-time image compression. An area efficient high performance VLSI architecture for DCT based on the distributed arithmetic is proposed in this paper. Minimum number of additions is used to the DCT by exploiting the timing property of the DCT transform based on the distributed arithmetic. A case study of 8 times 8 DCT architecture based on the DA is analyzed. Savings exceeding 97% are achieved for the DCT implementation.
    BibTeX:
    @inproceedings{2007_Chen,
      author = {Yanling Chen and Xixin Cao and Qingqing Xie and Chungan Peng},
      title = {An Area Efficient High Performance DCT Distributed Architecture for Video Compression},
      journal = {Advanced Communication Technology, The 9th International Conference on},
      year = {2007},
      volume = {1},
      pages = {238 -241},
      doi = {http://dx.doi.org/10.1109/ICACT.2007.358346}
    }
    					
    Chungan, P.; Xixin, C.; Dunshan, Y. & Xing, Z. A 250MHz optimized distributed architecture of 2D 8x8 DCT 2007 ASIC, 2007. ASICON '07. 7th International Conference on , pp. 189 -192   inproceedings DOI    
    Abstract: Discrete cosine transform (DCT) plays an important role in image and video compression, but computing a two-dimensional (2D) DCT, a large number of multiplications and additions are required in a direct approach. Multiplications, which are the most time-consuming and expensive operations in simple processor, can be completely avoided in our proposed architecture for multiple channel real-time image compression. In this paper, a compressed distributed arithmetic architecture for 2D 8times8 DCT is presented, which offers high speed and small area. The basic architecture consists of a ID row DCT followed by a transpose register array and another ID column DCT, in which an 8-input ID DCT structure only requires 15 adders to build a compressed adder matrix and no ROM is needed. Compared with other architectures available, it has a great improvement on computing speed and reducing area.
    BibTeX:
    @inproceedings{2007_Chungan,
      author = {Peng Chungan and Cao Xixin and Yu Dunshan and Zhang Xing},
      title = {A 250MHz optimized distributed architecture of 2D 8x8 DCT},
      journal = {ASIC, 2007. ASICON '07. 7th International Conference on},
      year = {2007},
      pages = {189 -192},
      doi = {http://dx.doi.org/10.1109/ICASIC.2007.4415599}
    }
    					
    Colonnese, S.; Panci, G.; Rinauro, S. & Scarano, G. Optimal video coding for bit-rate switching applications: a game-theoretic approach 2007 World of Wireless, Mobile and Multimedia Networks, 2007. WoWMoM 2007. IEEE International Symposium on a , pp. 1 -4   inproceedings DOI    
    Abstract: In this work1 we discuss a game theoretic approach to bitstream switching in video coding. Fast and bit-saving video bitstream switching is an important issue in video communication system on time varying channels. The most recent video coding standard, namely H.264, support the seamless switching among bitstreams coded at different bitrates by means of suitably coded frames, named Switching Pictures. Since the rate-distortion characteristics of switching frames differ from those of I and P frames, their location affect both the bit-rate and the quality of the coded sequence. In this work, we address the optimization of the SP frames location under an assigned bitrate budget. At this aim we restor to a game theoretic approach and we show that the optimal solution is met when the SP frames are assigned to the frame with the smallest innovation. Experimental results show the advantage in terms of both rate and distortion achieved by the optimized Switching frame insertion with respect to basic H264 coding.
    BibTeX:
    @inproceedings{2007_Colonnese,
      author = {Colonnese, Stefania and Panci, Gianpiero and Rinauro, Stefano and Scarano, Gaetano},
      title = {Optimal video coding for bit-rate switching applications: a game-theoretic approach},
      journal = {World of Wireless, Mobile and Multimedia Networks, 2007. WoWMoM 2007. IEEE International Symposium on a},
      year = {2007},
      pages = {1 -4},
      doi = {http://dx.doi.org/10.1109/WOWMOM.2007.4351753}
    }
    					
    Duh, D.; Jeng, J. & Chen, S. Fractal Image Compression with Predicted Dihedral Transformation 2007 Computers and Communications, 2007. ISCC 2007. 12th IEEE Symposium on , pp. 661 -666   inproceedings DOI    
    Abstract: Fractal image compression exploits the self-similarity of an image to achieve image compression. The conventional algorithm allows the transformations on domain blocks to obtain eight orientations so as to increase the quality of retrieved image. On the other hand, if no transformation is performed in order to speedup the encoder, the image quality will decay. In this paper, a direct allocating method to predict the desired transformation for similarity measure is proposed. Simulations show that the encoding time is almost the same as that of the method without transformations while the image quality is close to that of the standard method.
    BibTeX:
    @inproceedings{2007_Duh,
      author = {Duh, D.J. and Jeng, J.H. and Chen, S.Y.},
      title = {Fractal Image Compression with Predicted Dihedral Transformation},
      journal = {Computers and Communications, 2007. ISCC 2007. 12th IEEE Symposium on},
      year = {2007},
      pages = {661 -666},
      doi = {http://dx.doi.org/10.1109/ISCC.2007.4381598}
    }
    					
    George, T.G. & Malmurugan, N. A New Fast Architecture for HD H.264 CAVLC Multi-syntax Decoder and its FPGA Implementation 2007 Conference on Computational Intelligence and Multimedia Applications, 2007. International Conference on
    Vol. 3 , pp. 118 -122  
    inproceedings DOI    
    Abstract: In this paper, we present a fast architecture of real-time CAVLC decoder (CAVLD) implemented in a FPGA. The real-time performance is achieved by exploring the pipelining possibilities between the sub-modules and multi syntax decoding. The implemented fast CAVLD architecture, when integrated with H264 decoder was capable of parsing at 30fps for 1080p streams for an encoded bit stream at a bit rate of 200 Mbps to achieve the real-time performance, while the clock is operated at 74.25 MHz. The result numbers of ALUs are 3266 and the critical path is within 10.5ns
    BibTeX:
    @inproceedings{2007_George,
      author = {George, Tony Gladvin and Malmurugan, N.},
      title = {A New Fast Architecture for HD H.264 CAVLC Multi-syntax Decoder and its FPGA Implementation},
      journal = {Conference on Computational Intelligence and Multimedia Applications, 2007. International Conference on},
      year = {2007},
      volume = {3},
      pages = {118 -122},
      doi = {http://dx.doi.org/10.1109/ICCIMA.2007.335}
    }
    					
    George, T. & Malmurugan, N. The Architecture of Fast H.264 CAVLC Decoder and its FPGA Implementation 2007 Intelligent Information Hiding and Multimedia Signal Processing, 2007. IIHMSP 2007. Third International Conference on
    Vol. 2 , pp. 389 -392  
    inproceedings DOI    
    Abstract: In this paper, we present a fast architecture of realtime CAVLC decoder (CAVLD) implemented in a FPGA. The real-time performance is achieved by exploring the pipelining possibilities between the sub- modules and multi sub-symbol decoding. The implemented fast CAVLD architecture, when integrated with H264 decoder was capable of parsing at 30 fps for 1080 p streams for an encoded bit stream at a bit rate of 200 Mbps to achieve the real-time performance, while the clock is operated at 74.25 MHz. The result numbers of ALUs are 3266 and the critical path is within 10.5 ns.
    BibTeX:
    @inproceedings{2007_Georgea,
      author = {George, T.G. and Malmurugan, N.},
      title = {The Architecture of Fast H.264 CAVLC Decoder and its FPGA Implementation},
      journal = {Intelligent Information Hiding and Multimedia Signal Processing, 2007. IIHMSP 2007. Third International Conference on},
      year = {2007},
      volume = {2},
      pages = {389 -392},
      doi = {http://dx.doi.org/10.1109/IIH-MSP.2007.291}
    }
    					
    Huang, C.-M. & Lin, C.-W. A Novel 4-D Perceptual Quantization Modeling for H.264 Bit-Rate Control 2007 Multimedia, IEEE Transactions on
    Vol. 9 (6) , pp. 1113 -1124  
    article DOI    
    Abstract: Bit-rate control plays a major role In video coding and multimedia streaming. A well-designed bit-rate control mechanism can achieve line visual qualities and avoid network congestion over a time-varying channel. This paper proposes an H.264 bit-rate control using a 4D perceptual quantization modeling (PQrc), including two major encoding modules: the perceptual frame-level bit-allocation using a 1D temporal pattern and the macroblock-level quantizer decision using a 3D rate pattern. The temporal pattern is used to predict frame complexity and determine proper budget bits further. The rate pattern is depicted as a bit-complexity-quantization (B.C.Q.) model, in which a tangent slope of a B.C.Q. curve is a piece of unique information to find a proper quantizer. For newly generated video clips, the B.C.Q. model is updated continuously using a weighted least-square estimation. In comparison with the latest H.264 JM10.2, our experiment results show that the proposed PQrc can: 1) keep stable buffer fullness and 2) improve the SNR quality and control accuracy effectively.
    BibTeX:
    @article{2007_Huang,
      author = {Chung-Ming Huang and Chung-Wei Lin},
      title = {A Novel 4-D Perceptual Quantization Modeling for H.264 Bit-Rate Control},
      journal = {Multimedia, IEEE Transactions on},
      year = {2007},
      volume = {9},
      number = {6},
      pages = {1113 -1124},
      doi = {http://dx.doi.org/10.1109/TMM.2007.902840}
    }
    					
    gon Kim, D.; jung Yoo, C.; bae Chang, O.; mi Kim, E. & Choi, J.-R. Improved Fast Mode Decision Algorithm for Variable Macro Block Motion Compensation in H.264 2007 Information Technology Convergence, 2007. ISITC 2007. International Symposium on , pp. 184 -187   inproceedings DOI    
    Abstract: H.264 is the latest international video coding standard and it can achieve considerably higher coding efficiency versus previous standards This is accomplished by an enhanced exploitation of the spatiotemporal correlation, such as various macro-block modes, variable block sizes for motion compensation, multiple reference frames, quarter-pixel motion accuracy, and various predictive direction modes for intra prediction. However, although such new encoding technology is the main factor in improving the encoding efficiency, it is simultaneously a highly complex factor. Therefore, the H264 standard is used widely in real applications where the speed elevation of these technologies is essential. The proposed fast motion vector search technique supports the fastest compensation of the variable size block motion that is suitable for the motion vector conclusions which comprise a significant proportion of the encoding complexity in this paper. In addition, a fast-mode decision algorithm capable of deciding various encoding optimal values of H.264 is proposed. Because the proposed fast mode decision algorithm is capable of deciding the mode early, the method enables the omission of the computational procedure of rate-distortion cost and the efficient operation of the motion vector search technique. The results presented here confirmed the achievement of a very high efficiency while decreasing the required number of calculations without incurring any significant encoding damage, thereby improving the existing algorithm by using the information of the surrounding block.
    BibTeX:
    @inproceedings{2007_Kim,
      author = {Dae-gon Kim and Cheol-jung Yoo and Ok-bae Chang and Eun-mi Kim and Jong-Ryeol Choi},
      title = {Improved Fast Mode Decision Algorithm for Variable Macro Block Motion Compensation in H.264},
      journal = {Information Technology Convergence, 2007. ISITC 2007. International Symposium on},
      year = {2007},
      pages = {184 -187},
      doi = {http://dx.doi.org/10.1109/ISITC.2007.66}
    }
    					
    Kuech, F. & Edler, B. Aliasing Reduction for Modified Discrete Cosine Transform Domain Filtering and its Application to Speech Enhancement 2007 Applications of Signal Processing to Audio and Acoustics, 2007 IEEE Workshop on , pp. 131 -134   inproceedings DOI    
    Abstract: Efficient combinations of coding and manipulation of audio signals in the spectral domain are often desirable in communication systems. The modified discrete cosine transform (MDCT) represents a popular spectral transform in audio coding as it leads to compact signal representations. However, as the MDCT corresponds to a critically sampled filter bank, it is in general not appropriate to directly apply it to filtering tasks. In this paper we present a method to compensate for aliasing terms that arise from such direct MDCT domain filtering. The discussion is thereby based on a rigorous matrix representation of critically sampled filter banks which also leads to corresponding efficient realizations. As an application showcase, noise reduction for MDCT based speech coding is considered in simulations.
    BibTeX:
    @inproceedings{2007_Kuech,
      author = {Kuech, Fabian and Edler, Bernd},
      title = {Aliasing Reduction for Modified Discrete Cosine Transform Domain Filtering and its Application to Speech Enhancement},
      journal = {Applications of Signal Processing to Audio and Acoustics, 2007 IEEE Workshop on},
      year = {2007},
      pages = {131 -134},
      doi = {http://dx.doi.org/10.1109/ASPAA.2007.4392984}
    }
    					
    Martinez-Enriquez, E.; de Frutos-Lopez, M.; Pujol-Alcolado, J. & Diaz-de-Maria, F. A Fast Motion-Cost Based Algorithm for H.264/AVC Inter Mode Decision 2007 Image Processing, 2007. ICIP 2007. IEEE International Conference on
    Vol. 5 , pp. V -325 -V -328  
    inproceedings DOI    
    Abstract: The H.264/AVC standard achieves a high coding efficiency compared to previous standards. However, the encoder complexity results in very high computational cost due to motion estimation and macroblock mode decisions. In this paper we propose a fast mode decision for low computational complexity applications for which the rate distortion optimization mode decision becomes unacceptable. The proposed pruned mode decision method consists in a motion-cost based early termination algorithm and saves about 50% encoding time with negligible quality loss.
    BibTeX:
    @inproceedings{2007_Martinez-Enriquez,
      author = {Martinez-Enriquez, E. and de-Frutos-Lopez, M. and Pujol-Alcolado, J.C. and Diaz-de-Maria, F.},
      title = {A Fast Motion-Cost Based Algorithm for H.264/AVC Inter Mode Decision},
      journal = {Image Processing, 2007. ICIP 2007. IEEE International Conference on},
      year = {2007},
      volume = {5},
      pages = {V -325 -V -328},
      doi = {http://dx.doi.org/10.1109/ICIP.2007.4379831}
    }
    					
    Ou, S.; Zhao, X. & Gao, Y. Speech Enhancement Employing Modified a Priori SNR Estimation 2007 Software Engineering, Artificial Intelligence, Networking, and Parallel/Distributed Computing, 2007. SNPD 2007. Eighth ACIS International Conference on
    Vol. 3 , pp. 827 -831  
    inproceedings DOI    
    Abstract: In order to improve the performance of a speech enhancement system, Plapous introduced a novel method called two-step noise reduction (TSNR) technique to refine the a priori SNR estimation of the decision-directed (DD) approach. However, the performance of this method depends on the choice of gain function. In this paper, we propose a modified approach for the a priori SNR estimation in DCT domain with two steps like the TSNR method. While in the second step, the proposed approach computes directly the square of clean speech component using the estimated a priori SNR of the DD approach, its result is not restricted on the gain function, and thus the drawback of the TSNR method is removed while the advantages are kept. A number of objective tests under various conditions are provided, and the results show the improved performance of our approach.
    BibTeX:
    @inproceedings{2007_Ou,
      author = {Shifeng Ou and Xiaohui Zhao and Ying Gao},
      title = {Speech Enhancement Employing Modified a Priori SNR Estimation},
      journal = {Software Engineering, Artificial Intelligence, Networking, and Parallel/Distributed Computing, 2007. SNPD 2007. Eighth ACIS International Conference on},
      year = {2007},
      volume = {3},
      pages = {827 -831},
      doi = {http://dx.doi.org/10.1109/SNPD.2007.197}
    }
    					
    Siu, W.-C.; Chan, Y.-L. & Fung, K.-T. On Transcoding a B-Frame to a P-Frame in the Compressed Domain 2007 Multimedia, IEEE Transactions on
    Vol. 9 (6) , pp. 1093 -1102  
    article DOI    
    Abstract: Only a limited number of methods have been proposed to realize heterogeneous transcoding, for example from MPEG-2 to H.263, or from H.264 to H.263. The major difficulties of transcoding a B-picture to a P-picture are that the incoming discrete cosine transform (DCT) coefficients of the B-frame are prediction errors arising from both forward and backward predictions, whilst the prediction errors in the DCT domain arising from the prediction using the previous frame alone are not available. The required new prediction errors need to be re-estimated in the pixel domain. This process involves highly complex computation and introduces re-encoding errors. We propose a new approach to convert a B-picture into a P-picture by making use of some properties of motion compensation in the DCT domain and the direct addition of DCT coefficients. We derive a set of equations and formulate the problem of how to obtain the DCT coefficients. One difficulty is that the last P-frame inside a GOP with an IBBP structure, for example, needs to be transcoded to become the last P-frame in the IPPP structure, and it has to be linked to the previous reconstructed P-frame instead of to the I-frame. We increased the speed of the transcoding process by making use of the motion activity which is expressed in terms of the correlation between pictures. The whole transcoding process is done in the transform domain, hence re-encoding errors are completely avoided. Results from our experimental work show that the proposed video transcoder not only achieves a speed-up of two to six times that of the conventional video transcoder, but it also substantially improves the quality of the video.
    BibTeX:
    @article{2007_Siu,
      author = {Wan-Chi Siu and Yui-Lam Chan and Kai-Tat Fung},
      title = {On Transcoding a B-Frame to a P-Frame in the Compressed Domain},
      journal = {Multimedia, IEEE Transactions on},
      year = {2007},
      volume = {9},
      number = {6},
      pages = {1093 -1102},
      doi = {http://dx.doi.org/10.1109/TMM.2007.902895}
    }
    					
    Suresh, K. & Sreenivas, T. Direct MDCT Domain Psychoacoustic Modeling 2007 Signal Processing and Information Technology, 2007 IEEE International Symposium on , pp. 742 -747   inproceedings DOI    
    Abstract: We extend the recently proposed spectral integration based psychoacoustic model for sinusoidal distortions to the MDCT domain. The estimated masking threshold additionally depends on the sub-band spectral flatness measure of the signal which accounts for the non- sinusoidal distortion introduced by masking. The expressions for masking threshold are derived and the validity of the proposed model is established through perceptual transparency test of audio clips. Test results indicate that we do achieve transparent quality reconstruction with the new model. Performance of the model is compared with MPEG psychoacoustic models with respect to the estimated perceptual entropy (PE). The results show that the proposed model predicts a lower PE than other models.
    BibTeX:
    @inproceedings{2007_Suresh,
      author = {Suresh, K. and Sreenivas, T.V.},
      title = {Direct MDCT Domain Psychoacoustic Modeling},
      journal = {Signal Processing and Information Technology, 2007 IEEE International Symposium on},
      year = {2007},
      pages = {742 -747},
      doi = {http://dx.doi.org/10.1109/ISSPIT.2007.4458108}
    }
    					
    Voronenko, Y. & Puschel, M. Mechanical Derivation of Fused Multiply #x2013;Add Algorithms for Linear Transforms 2007 Signal Processing, IEEE Transactions on
    Vol. 55 (9) , pp. 4458 -4473  
    article DOI    
    Abstract: Several computer architectures offer fused multiply-add (FMA), also called multiply-and-accumulate (MAC) instructions, that are as fast as a single addition or multiplication. For the efficient implementation of linear transforms, such as the discrete Fourier transform or discrete cosine transforms, this poses a challenge to algorithm developers as standard transform algorithms have to be manipulated into FMA algorithms that make optimal use of FMA instructions. We present a general method to convert any transform algorithm into an FMA algorithm. The method works with both algorithms given as directed acyclic graphs (DAGs) and algorithms given as structured matrix factorizations. We prove bounds on the efficiency of the method. In particular, we show that it removes all single multiplications except at most as many as the transform has outputs. We implemented the DAG-based version of the method and show that we can generate many of the best-known hand-derived FMA algorithms from the literature as well as a few novel FMA algorithms.
    BibTeX:
    @article{2007_Voronenko,
      author = {Voronenko, Y. and Puschel, M.},
      title = {Mechanical Derivation of Fused Multiply #x2013;Add Algorithms for Linear Transforms},
      journal = {Signal Processing, IEEE Transactions on},
      year = {2007},
      volume = {55},
      number = {9},
      pages = {4458 -4473},
      doi = {http://dx.doi.org/10.1109/TSP.2007.896116}
    }
    					
    Wang, J.-C.; Wang, J.-F.; Yang, J.-F. & Chen, J.-T. A Fast Mode Decision Algorithm and Its VLSI Design for H.264/AVC Intra-Prediction 2007 Circuits and Systems for Video Technology, IEEE Transactions on
    Vol. 17 (10) , pp. 1414 -1422  
    article DOI    
    Abstract: In this paper, we present a fast mode decision algorithm and design its VLSI architecture for H.264 intra-prediction. A regular spatial domain filtering technique is proposed to compute the dominant edge strength (DES) to reduce the possible predictive modes. Experimental results revealed that the proposed fast intra-algorithm reduces 40% computation with slight peak signal-to-noise ratio (PSNR) degradation. The designed DES VLSI engine comprises a zigzag converter, a DES finite-state machine (FSM), and a DES core. The former two units handle memory allocation and control flow while the last performs pseudoblock computation, edge filtering, and dominant edge strength extraction. With semicustom design fabricated by 0.18 mum CMOS single-poly-six-metal technology, the realized die size is roughly 0.15 times 0.15 mm2 and can be operated at 66 MHz.
    BibTeX:
    @article{2007_Wang,
      author = {Jia-Ching Wang and Jhing-Fa Wang and Jar-Ferr Yang and Jang-Ting Chen},
      title = {A Fast Mode Decision Algorithm and Its VLSI Design for H.264/AVC Intra-Prediction},
      journal = {Circuits and Systems for Video Technology, IEEE Transactions on},
      year = {2007},
      volume = {17},
      number = {10},
      pages = {1414 -1422},
      doi = {http://dx.doi.org/10.1109/TCSVT.2007.903786}
    }
    					
    Xiaoyun, G.; Weiming, S.; Qiong, W. & Xiaoping, H. Error-Resilient Video Compression via Forecast of Error Concealment Modes Based on Multiple State Streams and Even-Odd Filed 2007 Wireless Communications, Networking and Mobile Computing, 2007. WiCom 2007. International Conference on , pp. 2944 -2947   inproceedings DOI    
    Abstract: This paper introduces the forecast of error concealment modes (FOECM) using in the error resilient coding scheme based on multiple state streams and even-odd field. The forecast of error concealment modes makes it extremely easy and flexible to obtain the best performance of error concealment at the decoder, which is obtained by the pre-evaluation of the error concealment modes at the encoder. Multiple states coding (MSC) based on even-odd filed presents error concealment modes of many dimensions, which provide improved error concealment and fast state recovery for the lost stream. In addition, we present three novelty algorithms: approximate parallel field projection (APFP) and rank edge line average (RELA) and bidirectional best neighborhood matching (BBNM). Simulations show that this approach has excellent performance. Moreover, the method proposed can be implemented in many source codes (e.g., H263, MPEG4, H264).
    BibTeX:
    @inproceedings{2007_Xiaoyun,
      author = {Guo Xiaoyun and Shen Weiming and Wu Qiong and Huang Xiaoping},
      title = {Error-Resilient Video Compression via Forecast of Error Concealment Modes Based on Multiple State Streams and Even-Odd Filed},
      journal = {Wireless Communications, Networking and Mobile Computing, 2007. WiCom 2007. International Conference on},
      year = {2007},
      pages = {2944 -2947},
      doi = {http://dx.doi.org/10.1109/WICOM.2007.731}
    }
    					
    Xie, S.; Rahardja, S. & Gu, Z. Performance of DS-CDMA Downlink Systems With Orthogonal UCHT Complex Sequences 2007 Communications, IEEE Transactions on
    Vol. 55 (2) , pp. 251 -256  
    article DOI    
    Abstract: This letter investigates a transmitted signaling technique using orthogonal unified complex Hadamard transform (UCHT) spreading sequences and the coherent RAKE receiver in direct-sequence code-division multiple-access (DS-CDMA) downlinks to maintain the orthogonality between users and reduce the effect of multipath fading and interference from other users. A general multipath-fading channel model is assumed. System performance is evaluated by means of signal-to-interference-plus-noise ratio (SINR) at the RAKE receiver. It is shown that the SINR of the system employing UCHT complex sequences is independent of the phase offsets between different paths, while the SINR of the system using Walsh-Hadamard (WH) sequences is related to the squared cosine of path phase offsets. As a result, the bit-error ratio performance of the DS-CDMA downlink system employing UCHT complex sequences is better than that of the system with WH sequences at high SINRs
    BibTeX:
    @article{2007_Xie,
      author = {Shoulie Xie and Susanto Rahardja and Zhenghui Gu},
      title = {Performance of DS-CDMA Downlink Systems With Orthogonal UCHT Complex Sequences},
      journal = {Communications, IEEE Transactions on},
      year = {2007},
      volume = {55},
      number = {2},
      pages = {251 -256},
      doi = {http://dx.doi.org/10.1109/TCOMM.2006.888524}
    }
    					
    Xiuhua, J.; Caiming, Z. & Yanling, W. Fast Algorithm of the 2-D 4x4 Inverse Integer Transform for H.264/AVC 2007 Innovative Computing, Information and Control, 2007. ICICIC '07. Second International Conference on , pp. 144 -144   inproceedings DOI    
    Abstract: A new fast two-dimension 4times4 (2-D 4times4 ) inverse integer transform algorithm is presented in this paper. The new algorithm reduces the addition and shift operations for the inverse transform greatly by using the two points: one is the regularities of the basic images; another is the characteristic of the post- transform data of the practical videos. By taking examples for several standard video clips, the new algorithm is compared with the existed influential fast algorithms. The experimental results indicate that the new algorithm needs 12.7838 addition operations and 1.69536 shift operations averagely which are much less than the operations of the existed influential fast algorithms. Moreover, the proposed direct 2-D algorithm is fit for parallel computing process.
    BibTeX:
    @inproceedings{2007_Xiuhua,
      author = {Ji Xiuhua and Zhang Caiming and Wang Yanling},
      title = {Fast Algorithm of the 2-D 4x4 Inverse Integer Transform for H.264/AVC},
      journal = {Innovative Computing, Information and Control, 2007. ICICIC '07. Second International Conference on},
      year = {2007},
      pages = {144 -144},
      doi = {http://dx.doi.org/10.1109/ICICIC.2007.304}
    }
    					
    Yajnanarayana, V.; Subramaniyan, R. & Schuette, M. Techniques to improve motion compensation performance of H264 video decoder using a vector processor 2007 Communications and Information Technologies, 2007. ISCIT '07. International Symposium on , pp. 1082 -1087   inproceedings DOI    
    Abstract: Motion Compensation for video decoding in standards like H.264 requires significant amount of computation. This is primarily because of H.264 six-tap FIR filtering for sub-sample computation. These algorithms typically take more than 50% of the computational time on a RISC processor like ARM. Novel algorithms proposed through this paper can be employed for systems which use vector processors as video decode accelerators to accelerate this process. The proposed algorithms are implemented on H264 video decode system with ARM9 host-processor and RSVP vector processor as an accelerator for key decode algorithms. By employing the proposed algorithms we were able to accelerate the motion compensation module by more than 4 times as compared to plain RISC implementation. This is achieved by efficiently vectorizing data on which FIR-filtering and reconstruction algorithm is operated on, together with optimal representation of FIR-filtering and reconstruction algorithm itself on vector processor.
    BibTeX:
    @inproceedings{2007_Yajnanarayana,
      author = {Yajnanarayana, V. and Subramaniyan, R. and Schuette, M.},
      title = {Techniques to improve motion compensation performance of H264 video decoder using a vector processor},
      journal = {Communications and Information Technologies, 2007. ISCIT '07. International Symposium on},
      year = {2007},
      pages = {1082 -1087},
      doi = {http://dx.doi.org/10.1109/ISCIT.2007.4392177}
    }
    					
    Ying, C. & Dejian, Y. The Design and Implementation of a Scalable Wireless Video Streaming System Adopting TCP Transmission Mode 2007 Computer and Information Technology, 2007. CIT 2007. 7th IEEE International Conference on , pp. 534 -538   inproceedings DOI    
    Abstract: Streaming services on wireless network is being popular since the bandwidth and QoS on wireless network has been improved. However, the dynamic fluctuated bandwidth, higher error rate and sharp delay jitter on wireless channel cause trouble to the real-time demand of streaming services. The wireless streaming media system aims at mobile users must also consider the large scale number of the potential users and the fast growth in the future. To solve these problems, we propose a scalable wireless video streaming system adopting TCP transmission mode in this paper. Our proposed system is compatible with the RTSP/RTP-protocols. It adopts RTP-on-TCP transmission mode to improve the quality of service on wireless network. It takes a hybrid centralized-distributed architecture on the server side to support the growth of users. Streaming services such as VOD are supported by the system with a wide range of media bitrates and several high-performance compression standards such as H264 and AMR.
    BibTeX:
    @inproceedings{2007_Ying,
      author = {Chai Ying and Ye Dejian},
      title = {The Design and Implementation of a Scalable Wireless Video Streaming System Adopting TCP Transmission Mode},
      journal = {Computer and Information Technology, 2007. CIT 2007. 7th IEEE International Conference on},
      year = {2007},
      pages = {534 -538},
      doi = {http://dx.doi.org/10.1109/CIT.2007.180}
    }
    					
    Zhan, Y. & Sapatnekar, S. High-Efficiency Green Function-Based Thermal Simulation Algorithms 2007 Computer-Aided Design of Integrated Circuits and Systems, IEEE Transactions on
    Vol. 26 (9) , pp. 1661 -1675  
    article DOI    
    Abstract: Due to technology scaling trends, the accurate and efficient calculations of the temperature distribution corresponding to a specific circuit layout and power density distribution will become indispensable in the design of high-performance very large scale integrated circuits. In this paper, we present three highly efficient thermal simulation algorithms for calculating the on-chip temperature distribution in a multilayered substrate structure. All three algorithms are based on the concept of the Green function and utilize the technique of discrete cosine transform. However, the application areas of the algorithms are different. The first algorithm is suitable for localized analysis in thermal problems, whereas the second algorithm targets full-chip temperature profiling. The third algorithm, which combines the advantages of the first two algorithms, can be used to perform thermal simulations where the accuracy requirement differs from place to place over the same chip. Experimental results show that all three algorithms can achieve relative errors of around 1% compared with that of a commercial computational fluid dynamic software package for thermal analysis, whereas their efficiencies are orders of magnitude higher than that of the direct application of the Green function method.
    BibTeX:
    @article{2007_Zhan,
      author = {Yong Zhan and Sapatnekar, S.S.},
      title = {High-Efficiency Green Function-Based Thermal Simulation Algorithms},
      journal = {Computer-Aided Design of Integrated Circuits and Systems, IEEE Transactions on},
      year = {2007},
      volume = {26},
      number = {9},
      pages = {1661 -1675},
      doi = {http://dx.doi.org/10.1109/TCAD.2007.895754}
    }
    					
    Adamo, O.; Mohanty, S.; Kougianos, E. & Varanasi, M. VLSI Architecture for Encryption and Watermarking Units Towards the Making of a Secure Camera 2006 SOC Conference, 2006 IEEE International , pp. 141 -144   inproceedings DOI    
    Abstract: Considerable amount of research is directed at putting biometric data in conventional forms of identification such as passports. However, putting biometric data in passports makes the data vulnerable to theft, causing privacy related issues. To address such issues, we present a new approach and architecture in the framework of a digital camera, conceptualized as a "Secure Digital Camera (SDC)". The SDC uses watermarking and encryption processes for image security and authentication. The Rijndael AES algorithm and a DCT-based visible watermarking algorithm were chosen for implementation in our camera. The proposed architectures were modeled, simulated and synthesized in Xilinx ISE.
    BibTeX:
    @inproceedings{2006_Adamo,
      author = {Adamo, O.B. and Mohanty, S.P. and Kougianos, E. and Varanasi, M.},
      title = {VLSI Architecture for Encryption and Watermarking Units Towards the Making of a Secure Camera},
      journal = {SOC Conference, 2006 IEEE International},
      year = {2006},
      pages = {141 -144},
      doi = {http://dx.doi.org/10.1109/SOCC.2006.283868}
    }
    					
    Balter, R.; Gioia, P. & Morin, L. Scalable and Efficient Video Coding Using 3-D Modeling 2006 Multimedia, IEEE Transactions on
    Vol. 8 (6) , pp. 1147 -1155  
    article DOI    
    Abstract: In this paper, we present a three-dimensional (3D) model-based video coding scheme for streaming static scene video in a compact way but also enabling time and spatial scalability according to network or terminal capability and providing 3D functionalities. The proposed format is based on encoding the sequence of reconstructed models using second-generation wavelets, and efficiently multiplexing the resulting geometric, topological, texture, and camera motion binary representations. The wavelets decomposition can be adaptive in order to fit to images and scene contents. To ensure time scalability, this representation is based on a common connectivity for all 3D models, which also allows straightforward morphing between successive models ensuring visual continuity at no additional cost. The method proves to be better than previous methods for video encoding of static scenes, even better than state-of-the-art video coders such as H264 (also known as MPEG AVC). Another application of our approach are smoothing camera path for suppression of jitter from hand-held acquisition and the fast transmission and real-time visualization of virtual environments obtained by video capture, for virtual or augmented reality and interactive walk-through in photo-realistic 3D environments around the original camera path
    BibTeX:
    @article{2006_Balter,
      author = {Balter, R. and Gioia, P. and Morin, L.},
      title = {Scalable and Efficient Video Coding Using 3-D Modeling},
      journal = {Multimedia, IEEE Transactions on},
      year = {2006},
      volume = {8},
      number = {6},
      pages = {1147 -1155},
      doi = {http://dx.doi.org/10.1109/TMM.2006.879873}
    }
    					
    Fan, C.-P. Fast 2-dimensional 4 times; 4 forward integer transform implementation for H.264/AVC 2006 Circuits and Systems II: Express Briefs, IEEE Transactions on
    Vol. 53 (3) , pp. 174 - 177  
    article DOI    
    Abstract: In this paper, the novel two-dimensional (2-D) fast algorithm for realization of 4 times; 4 forward integer transform in H.264 is proposed. Based on matrix operations with Kronecker product and direct sum, the efficient fast 2-D 4 times; 4 forward integer transform can be derived from the proposed one-dimensional fast 4 times; 4 forward integer transform through matrix decompositions. The proposed fast 2-D 4 times; 4 forward integer transform design doesn't need transpose memory for direct parallel pipelined architecture. The fast 2-D 4 times; 4 forward integer transform requires fewer latency delays than the state-of-the-art methods. With regular modularity, the proposed fast algorithm is suitable for VLSI implementation to achieve real-time H.264/advanced video coding (AVC) signal processing.
    BibTeX:
    @article{2006_Fan,
      author = {Chih-Peng Fan},
      title = {Fast 2-dimensional 4 times; 4 forward integer transform implementation for H.264/AVC},
      journal = {Circuits and Systems II: Express Briefs, IEEE Transactions on},
      year = {2006},
      volume = {53},
      number = {3},
      pages = { 174 - 177},
      doi = {http://dx.doi.org/10.1109/TCSII.2005.858748}
    }
    					
    Fu, C.-H.; Chan, Y.-L. & Siu, W.-C. Efficient reverse-play algorithms for MPEG video with VCR support 2006 Circuits and Systems for Video Technology, IEEE Transactions on
    Vol. 16 (1) , pp. 19 - 30  
    article DOI    
    Abstract: Reverse playback is the most common video cassette recording (VCR) function in digital video players and it involves playing video frames in reverse order. However, the predictive processing techniques employed in MPEG severely complicate the reverse-play operation. For displaying single frame during reverse playback, all frames from the previous I-frame to the requested frame must be sent by the server and decoded by the client machine. It requires much higher bandwidth of the network and complexity of the decoder. In this paper, we propose a compressed-domain approach for an efficient implementation of the MPEG video streaming system to provide reverse playback over a network with the minimal requirements on the network bandwidth and the decoder complexity. In the proposed video streaming server, it classifies macroblocks in the requested frame into two categories-backward macroblocks (BMBs) and forward macroblock (FMBs). Two novel MB-based techniques are used to manipulate the necessary MBs in the compressed domain and the server then sends the processed MBs to the client machine. For BMBs, we propose a sign inversion technique, which is operated in the variable length coding (VLC) domain, to reduce the number of MBs that need to be decoded by the decoder and the number of bits that need to be sent over the network in the reverse-play operation. The server also identifies the previous related MBs of FMBs and those related maroblocks coded without motion compensation are then processed by a technique of direction addition of discrete cosine transform (DCT) coefficients to further reduce the computational complexity of the client decoder. With the sign inversion and direct addition of DCT coefficients, the proposed architecture only manipulates MBs either on the VLC domain or DCT domain to achieve the server with low complexity. Experimental results show that, as compared to the conventional system, the new streaming system reduces the required network bandwidth and the decoder complexity significantly.
    BibTeX:
    @article{2006_Fu,
      author = {Chang-Hong Fu and Yui-Lam Chan and Wan-Chi Siu},
      title = {Efficient reverse-play algorithms for MPEG video with VCR support},
      journal = {Circuits and Systems for Video Technology, IEEE Transactions on},
      year = {2006},
      volume = {16},
      number = {1},
      pages = { 19 - 30},
      doi = {http://dx.doi.org/10.1109/TCSVT.2005.856901}
    }
    					
    Guo, H.; Xia, X.; Sun, W.; Zhou, J. & Yu, S. An memory-efficient variable length decoding scheme for embedded MPEG-4 video decoders 2006 Signal Processing, 2006 8th International Conference on
    Vol. 3  
    inproceedings DOI    
    Abstract: Variable length decoding (VLD) is inherent bit-serial operation and is the first stage of the whole video decoding task, the overall decoding system performance is determined by its throughput and efficiency. In this paper, a scheme for decoding of discrete cosine transform (DCT) coefficients is presented The DCT VLD tables are split into two different tables: the one with 0-3 leading zeros and the other with 4 or more leading zeros. The proposed VLD tables need less than 6K bytes memory space, and the decoding procedures are completed by applying numerical properties to codewords identification and symbol indexing. Compared with the direct indexing method, not only the memory space is reduced for symbol information, but also the code efficiency is improved by enabling compiling code with near mode. Experimental results show that the proposed solution can improve the VLD speed by 17-43%. This VLD method has been integrated into the developed embedded MPEG-4 video decoder successfully
    BibTeX:
    @inproceedings{2006_Guo,
      author = {Hongxing Guo and Xiaojian Xia and Weiping Sun and Jingli Zhou and Shengsheng Yu},
      title = {An memory-efficient variable length decoding scheme for embedded MPEG-4 video decoders},
      journal = {Signal Processing, 2006 8th International Conference on},
      year = {2006},
      volume = {3},
      doi = {http://dx.doi.org/10.1109/ICOSP.2006.345794}
    }
    					
    Kun, Y.; Chun, Z.; Guoze, D.; Jiangxiang, X. & Zhihua, W. A Hardware-Software Co-design for H.264/AVG Decoder 2006 Solid-State Circuits Conference, 2006. ASSCC 2006. IEEE Asian , pp. 119 -122   inproceedings DOI    
    Abstract: A single chip decoder SOC for H.264 baseline profile, called OR264 (OR1K based H264 decoder), is presented in this paper. The chip has mixed hardware/software architecture to combine performance and flexibility. It is partitioned that the hardware is used to boost the performance and efficiency of key operations in H.264 decoder while the software is used to control the decoding flow and to synchronize the hardware modules. All hardware units operate in parallel. The hardware can decode a MB in 851 clock cycles under ideal condition. The chip is fabricated using UMC 0.18-mum 6-layers metal CMOS process. It contains 1.5 M transistors and 176k bits embedded SRAM. The die size is 4.8 mm x 4.8 mm and the critical path is 10 ns.
    BibTeX:
    @inproceedings{2006_Kun,
      author = {Yang Kun and Zhang Chun and Du Guoze and Xie Jiangxiang and Wang Zhihua},
      title = {A Hardware-Software Co-design for H.264/AVG Decoder},
      journal = {Solid-State Circuits Conference, 2006. ASSCC 2006. IEEE Asian},
      year = {2006},
      pages = {119 -122},
      doi = {http://dx.doi.org/10.1109/ASSCC.2006.357866}
    }
    					
    Mohammad-Khani, G.-R.; Lee, C.-M.; Kieffer, M. & Duhamel, P. Simplification of VLC tables with application to ML and MAP decoding algorithms 2006 Communications, IEEE Transactions on
    Vol. 54 (10) , pp. 1835 -1844  
    article DOI    
    Abstract: Many source coding standards (JPEG, H263+, H264), rely heavily on entropy coding with variable-length codes (VLC). However, bitstreams made of VLC-encoded data are particularly sensitive to transmission errors. Recent results tend to use knowledge of the VLC structure in order to perform an efficient decoding of the bitstream. These techniques use a trellis describing the structure of the VLC codebook and assume that some a priori information is available at decoder side. Significant improvements, compared with prefix decoding of bitstreams are achieved. However, the complexity of these techniques may become intractable when realistic VLC codebooks are considered. This paper presents an algorithm for compacting VLC tables. The codewords are grouped into a minimum number of classes. Decoding algorithms may then work on a reduced number of classes, instead of working on the whole set of codewords. A proof of optimality is provided for the VLC table-compaction algorithm. The algorithm is applied to the H263+ VLC codebook and merges the 204 codewords into 25 classes. The resulting compact tables are shown to be exactly equivalent to the initial ones when used with hard decoding algorithms. The properties of the associated soft decoding algorithms using these compact tables are also evaluated
    BibTeX:
    @article{2006_Mohammad-Khani,
      author = {Mohammad-Khani, G.-R. and Chang-Ming Lee and Kieffer, M. and Duhamel, P.},
      title = {Simplification of VLC tables with application to ML and MAP decoding algorithms},
      journal = {Communications, IEEE Transactions on},
      year = {2006},
      volume = {54},
      number = {10},
      pages = {1835 -1844},
      doi = {http://dx.doi.org/10.1109/TCOMM.2006.881372}
    }
    					
    Nadamudi, D.; Clancy, M.; Yang, A.; DeLa Cruz, J.; Chandra, R.; Santhanam, S.; Burke, S. & Ko, U. Entering the hot zone - can you handle the heat and be cool? 2006 Design Automation Conference, 2006 43rd ACM/IEEE , pp. 174 -175   inproceedings DOI    
    Abstract: With the latest gaming systems such as X-box, Playstation 3, PSP, IPTV, and H264 requiring more bandwidth, packaged in smaller devices, and providing higher quality, in return they evacuate more heat and consume more power, thus making power consumption and thermal dissipation a major issue for chip designers in the nanometer era. Thermal effect on the chip and package is the next wave. First, there was signal integrity. Then power integrity and now, there is a new first-order effect in the horizon for IC and system designers called thermal integrity. And the traditional design methodology of using uniform temperature across the chip is becoming woefully inadequate and provides insufficient data to the package and system designers, unable to meet the requirement and demands of the consumer. To better understand the impact of temperature variation on chip's performance, both power distribution and package characteristics must be considered in the design and analysis flow. Likewise, a realistic view of the heat distribution across the chip is becoming essential for package and system designers. What makes the analysis of signal, power, and thermal integrity effects challenging is that they must all be considered concurrently across the entire chip, which requires novel approaches to mitigate the issue. The panel discusses how urgent is the impact of thermal integrity on system designs and is this a real concern or are we making it up? When and under what condition it became urgent, and is it related to process nodes, low power applications, type of packaging used, etc.? Is there specific design techniques used for more specific multimedia processing? Can you separate the IC design from package and system designs with some assumptions or is co-design the only way? How accurate is the industry's understanding of the physics of the chip (device), interconnect, and package? Is thermal integrity a first order or second order effect? Given the other variations from nomina- - l, how important is this? The moderator concludes with a summary of the panelist comments and be able to make a forward-looking statement about the future of power and heat and their impact on system design
    BibTeX:
    @inproceedings{2006_Nadamudi,
      author = {Nadamudi, D. and Clancy, M. and Yang, A. and DeLa Cruz, J.A. and Chandra, R. and Santhanam, S. and Burke, S. and Uming Ko},
      title = {Entering the hot zone - can you handle the heat and be cool?},
      journal = {Design Automation Conference, 2006 43rd ACM/IEEE},
      year = {2006},
      pages = {174 -175},
      doi = {http://dx.doi.org/10.1109/DAC.2006.229202}
    }
    					
    Petjanski, B. & Kalva, H. DCT domain intra MB mode decision for MPEG-2 to H.264 transcoding 2006 Consumer Electronics, 2006. ICCE '06. 2006 Digest of Technical Papers. International Conference on , pp. 419 -420   inproceedings DOI    
    Abstract: This paper presents a macroblock (MB) mode estimation technique for MPEG-2 to H.264 intra frame video transcoding. This technique can be applied in number of consumer electronic applications such as DVRs and content delivery systems. The DCT coefficients gathered from the MPEG-2 process are used to estimate the MB coding mode. The estimated mode is used to constrain the MB prediction mode computation just to the estimated MB mode. Our approach uses the DCT coefficients obtained during the MPEG-2 decoding stage and classifies an MB as INTRA16times16 or INTRA4times4. The computational cost of the DCT domain operations is negligible but results in substantial savings in the H.264 mode estimation. The results show that the PSNR drop caused by this direct estimation is negligible
    BibTeX:
    @inproceedings{2006_Petjanski,
      author = {Petjanski, B. and Kalva, H.},
      title = {DCT domain intra MB mode decision for MPEG-2 to H.264 transcoding},
      journal = {Consumer Electronics, 2006. ICCE '06. 2006 Digest of Technical Papers. International Conference on},
      year = {2006},
      pages = {419 -420},
      doi = {http://dx.doi.org/10.1109/ICCE.2006.1598489}
    }
    					
    Puschel, M. & Moura, J. The Algebraic Structure in Signal Processing: Time and Space 2006 Acoustics, Speech and Signal Processing, 2006. ICASSP 2006 Proceedings. 2006 IEEE International Conference on
    Vol. 5 , pp. V -V  
    inproceedings DOI    
    Abstract: The assumptions underlying linear signal processing (SP) produce more structure than vector spaces. We capture this structure by describing the space of filters as an algebra and the space of signals as the associated module. We formulate an algebraic approach to SP that is axiomatically based on the concept of a signal model. Signal models for time are visualized as directed graphs. We construct corresponding models for undirected graphs, which we hence call space models, and show that, in particular, the 16 DCTs and DSTs are Fourier transforms for these finite space models. Finally, we discuss the extension of our theory to separable and nonseparable 2-DSP
    BibTeX:
    @inproceedings{2006_Puschel,
      author = {Puschel, M. and Moura, J.M.F.},
      title = {The Algebraic Structure in Signal Processing: Time and Space},
      journal = {Acoustics, Speech and Signal Processing, 2006. ICASSP 2006 Proceedings. 2006 IEEE International Conference on},
      year = {2006},
      volume = {5},
      pages = {V -V},
      doi = {http://dx.doi.org/10.1109/ICASSP.2006.1661446}
    }
    					
    Sayit, M. & Tunah, T. Video Streaming with H.264 Over the Internet 2006 Signal Processing and Communications Applications, 2006 IEEE 14th , pp. 1 -4   inproceedings DOI    
    Abstract: The new video coding standard H264 introduces two different kinds of switching frames, namely SP and SI frames. SP frames are proposed for switching among video streams while SI frames are proposed for error resilience. In this study, we stream H.264 coded video over the Internet in a quality adaptive fashion. Video is pre-encoded in two different rates with SP frames. Depending on network conditions, quality is adapted among two different streams. For error resilience, periodic insertion of I frames is preferred over SI frame retransmission since our path between client and server has large propagation delay. Two different GOP patterns are used to compare the use of SP frames as an alternative to I frames
    BibTeX:
    @inproceedings{2006_Sayit,
      author = {Sayit, M.F. and Tunah, T.},
      title = {Video Streaming with H.264 Over the Internet},
      journal = {Signal Processing and Communications Applications, 2006 IEEE 14th},
      year = {2006},
      pages = {1 -4},
      doi = {http://dx.doi.org/10.1109/SIU.2006.1659854}
    }
    					
    Wang, X.; Yu, W. & Wang, Z. Efficient Direct Boundary Element Method for Resistance Extraction of Substrate With Arbitrary Doping Profile 2006 Computer-Aided Design of Integrated Circuits and Systems, IEEE Transactions on
    Vol. 25 (12) , pp. 3035 -3042  
    article DOI    
    Abstract: It is important to model the substrate coupling for mixed-signal or RF circuit designs. In this paper, a direct boundary element method (DBEM) and related efficient techniques are presented to calculate the coupling resistances for three-dimensional substrate structure. First, a nonuniform meshing scheme is presented to reduce boundary elements while preserving accuracy. Then, the unknowns on top medium surface are removed from the discretized linear system of DBEM with a matrix reduction technique. The third technique is applying the quasi-multiple medium idea (W. Yu, Z. Wang, and J. Gu, "Fast capacitance extraction of actual 3-D VLSI interconnects using quasi-multiple medium accelerated BEM," IEEE Trans. Microwave Theory Tech., vol. 51, no. 1, pp. 109-199, Jan. 2003), which greatly reduces the expense of matrix reduction and makes the final coefficient matrix much sparser. With these proposed techniques, the linear equation system is largely condensed and sparsified and then solved with a preconditioned generalized minimum residual solver for multiple right-hand sides to get the whole resistance matrix. Numerical experiments on typical substrates with various doping profiles show the high accuracy of the DBEM-based method. The authors also compared the DBEM method with the Green's function methods accelerated by discrete cosine transform or eigendecomposition techniques. The results show that the DBEM-based method is several times or tens of times faster than the other two. At the same time, the DBEM method has no difficulty in handling substrates with more complex than stratified doping profiles, which is a large advantage over the existing methods
    BibTeX:
    @article{2006_Wang,
      author = {Xiren Wang and Wenjian Yu and Zeyi Wang},
      title = {Efficient Direct Boundary Element Method for Resistance Extraction of Substrate With Arbitrary Doping Profile},
      journal = {Computer-Aided Design of Integrated Circuits and Systems, IEEE Transactions on},
      year = {2006},
      volume = {25},
      number = {12},
      pages = {3035 -3042},
      doi = {http://dx.doi.org/10.1109/TCAD.2006.882487}
    }
    					
    Yang, M.; Grecos, C. & Chen, L. Reduced Computation Mode decision using Error Domain heuristics for the H264 standard 2006 Research in Microelectronics and Electronics 2006, Ph. D. , pp. 117 -120   inproceedings DOI    
    Abstract: A new algorithm for fast mode decision in the H264 video coding standard is presented in this paper. The algorithm exploits mode grouping through moving averages of error cost functions for providing significant computational savings with similar rate distortion performance as compared to accepted standard contributions (Lim et al., 2003)
    BibTeX:
    @inproceedings{2006_Yang,
      author = {Yang, M.Y. and Grecos, C. and Lihui Chen},
      title = {Reduced Computation Mode decision using Error Domain heuristics for the H264 standard},
      journal = {Research in Microelectronics and Electronics 2006, Ph. D.},
      year = {2006},
      pages = {117 -120},
      doi = {http://dx.doi.org/10.1109/RME.2006.1689910}
    }
    					
    Zhang, X.; Wang, Z. & Han, Z. Video transcoding for wireless gateway system 2006 Wireless, Mobile and Multimedia Networks, 2006 IET International Conference on , pp. 1 -4   inproceedings    
    Abstract: In this paper we study the problem of accessing video streams through handheld devices and wireless networks. We first review and discuss the major technologies of video transcoding. Then, we introduce a wireless video transcoding gateway system to handle the heterogeneous mobile devices and the smallness capability of the handheld devices. To gain higher video quality and lower complexity, we propose an efficient video transcoding structure based on close-loop model and a fast motion vector re-estimation algorithm. Based on this, we design a novel service scheme, which can provide better video content access service on current mobile communication systems. Moreover, we implement a test prototype of MPEG2-to-H264 transcoding system and simulate the video streams access service in a wireless LAN.
    BibTeX:
    @inproceedings{2006_Zhang,
      author = {Zhang, Xinchen and Zhongyuan Wang and Zhen Han},
      title = {Video transcoding for wireless gateway system},
      journal = {Wireless, Mobile and Multimedia Networks, 2006 IET International Conference on},
      year = {2006},
      pages = {1 -4}
    }
    					
    Alam, M.; Badawy, W. & Jullien, G. A new time distributed DCT architecture for MPEG-4 hardware reference model 2005 Circuits and Systems for Video Technology, IEEE Transactions on
    Vol. 15 (5) , pp. 726 - 730  
    article DOI    
    Abstract: This paper presents the design of a new time distributed architecture (TDA) which outlines the architecture (ISO/IEC JTC1/SC29/WG11 MPEG2002/M8565) submitted to MPEG4 Part9 committee and included in the ISO/IEC JTC1/SC29/WG11 MPEG2002/9115N document. The proposed TDA optimizes the two-dimensional discrete cosine transform (2-D-DCT) architecture performance. It uses a time distribution mechanism to exploit the computational redundancy within the inner product computation module. The application specific requirements of input, output and coefficients word length are met by scheduling the input data. The coefficient matrix uses linear mappings to assign necessary computation to processor elements in both space and time domains. The performance analysis shows performance savings in excess of 96% as compared to the direct implementation and more than 71% as compared to other optimized application specific architectures for DCT.
    BibTeX:
    @article{2005_Alam,
      author = {Alam, M. and Badawy, W. and Jullien, G.},
      title = {A new time distributed DCT architecture for MPEG-4 hardware reference model},
      journal = {Circuits and Systems for Video Technology, IEEE Transactions on},
      year = {2005},
      volume = {15},
      number = {5},
      pages = { 726 - 730},
      doi = {http://dx.doi.org/10.1109/TCSVT.2005.846429}
    }
    					
    Au, K.; Law, N. & Siu, W. Direct image retrieval in JPEG and JPEG2000 2005 Image Processing, 2005. ICIP 2005. IEEE International Conference on
    Vol. 1 , pp. I - 529-32  
    inproceedings DOI    
    Abstract: Images are often compressed using JPEG or JPEG2000. Many retrieval systems operated in either uncompressed or compressed domains have been proposed. However, retrieving in multiple domains typically involves full decompression for feature analysis in spatial domain. Common features in different domains are thus worth of investigation for direct image indexing. By employing a common subband filtering model, outputs from JPEG and JPEG2000 can be compared directly without having a full decompression. Despite of high compression, similar translation and rotation invariant features can be extracted from these two domains. Simulation results reveal that JPEG and JPEG2000 compressed images can be searched from one another irrespective of the compression ratio. Our experimental studies confirm that retrieval in multiple domains is possible without a full decompression.
    BibTeX:
    @inproceedings{2005_Au,
      author = {Au, K.M. and Law, N.F. and Siu, W.C.},
      title = {Direct image retrieval in JPEG and JPEG2000},
      journal = {Image Processing, 2005. ICIP 2005. IEEE International Conference on},
      year = {2005},
      volume = {1},
      pages = { I - 529-32},
      doi = {http://dx.doi.org/10.1109/ICIP.2005.1529804}
    }
    					
    Barreh, W.; Tlili, F. & Benazza-Benyahia, A. Fast coding of bidirectional frame for H264 standard 2005 Electronics, Circuits and Systems, 2005. ICECS 2005. 12th IEEE International Conference on , pp. 1 -4   inproceedings DOI    
    Abstract: The new H264 video coding standard can achieve considerably higher coding efficiency than previous standards by extending the notion of bidirectional prediction. Unfortunately, this improvement requires heavy computational loads to select the best mode in the sense of rate-distortion tradeoff. In this paper, we propose to apply an accurate estimation of the bit rate in order to accelerate the rate-distortion optimization. experimental results indicate that substantial gains in time processing are achieved.
    BibTeX:
    @inproceedings{2005_Barreh,
      author = {Barreh, W. and Tlili, F. and Benazza-Benyahia, A.},
      title = {Fast coding of bidirectional frame for H264 standard},
      journal = {Electronics, Circuits and Systems, 2005. ICECS 2005. 12th IEEE International Conference on},
      year = {2005},
      pages = {1 -4},
      doi = {http://dx.doi.org/10.1109/ICECS.2005.4633417}
    }
    					
    Bialkowski, J.; Barkowsky, M. & Kaup, A. On Requantization in Intra-Frame Video Transcoding with Different Transform Block Sizes 2005 Multimedia Signal Processing, 2005 IEEE 7th Workshop on , pp. 1 -4   inproceedings DOI    
    Abstract: Transcoding is a technique to convert one video bit-stream into another. While homogeneous transcoding is done at the same coding standard, inhomogeneous transcoding converts from one standard format to another standard. Inhomogeneous transcoding between MPEG-2, MPEG-4 or H.263 was performed using the same transform. With the standardisation of H.264 also a new transform basis and different block size was defined. For requantization from block size 8times8 to 4times4 this leads to the effect that the quantization error of one coefficient in a block of size 8times8 is distributed over multiple coefficients in blocks of size 4times4. In our work, we analyze the requantization process for inhomogeneous transcoding with different transforms. The deduced equations result in an expression for the correlation of the error contributions from the coefficients of block size 8times8 at each coefficient of block size 4times4. We then compare the mathematical analysis to simulations on real sequences. The reference to the requantization process is the direct quantization of the undistorted signal. It will be shown that the loss is as high as 3 dB PSNR at equivalent step size for input and output bitstream. Also an equation for the choice of the second quantization step size in dependency of the requantization loss is deduced. The model is then extended from the DCT to the integer-based transform as defined in H.264
    BibTeX:
    @inproceedings{2005_Bialkowski,
      author = {Bialkowski, J. and Barkowsky, M. and Kaup, A.},
      title = {On Requantization in Intra-Frame Video Transcoding with Different Transform Block Sizes},
      journal = {Multimedia Signal Processing, 2005 IEEE 7th Workshop on},
      year = {2005},
      pages = {1 -4},
      doi = {http://dx.doi.org/10.1109/MMSP.2005.248669}
    }
    					
    Chen, C.; Wu, P.-H. & Chen, H. Transform-domain intra prediction for H.264 2005 Circuits and Systems, 2005. ISCAS 2005. IEEE International Symposium on , pp. 1497 - 1500 Vol. 2   inproceedings DOI    
    Abstract: H.264/AVC is the newest video coding standard jointly developed by the ITU-T Video Coding Experts Group and the ISO/IEC Moving Picture Experts Group. In contrast to some previous coding standards such as H.263+ and MPEG-4 Part-2, where intra prediction is performed in the transform domain, the intra prediction of H.264 is completely defined in the pixel domain. This presents a challenge to multimedia systems in which transcoding is conducted in the transform domain for the purpose of computational efficiency. In this paper, we show how to obtain the transform domain predictions for various intra modes of H264. We begin by converting the intra prediction from the pixel domain to the transform domain through matrix manipulation. Then we show how the operations involved in the matrix manipulation can be simplified. A computational complexity analysis of each intra prediction mode of H.264 is provided.
    BibTeX:
    @inproceedings{2005_Chen,
      author = {Chen Chen and Ping-Hao Wu and Chen, H.},
      title = {Transform-domain intra prediction for H.264},
      journal = {Circuits and Systems, 2005. ISCAS 2005. IEEE International Symposium on},
      year = {2005},
      pages = { 1497 - 1500 Vol. 2},
      doi = {http://dx.doi.org/10.1109/ISCAS.2005.1464883}
    }
    					
    Chen, T.-H.; Liu, S.-H.; Yao, H.-X. & Gao, W. Robust video watermarking based on DC coefficients of selected blocks 2005 Machine Learning and Cybernetics, 2005. Proceedings of 2005 International Conference on
    Vol. 9 , pp. 5273 -5278 Vol. 9  
    inproceedings DOI    
    Abstract: This paper presents a novel blind video watermarking method based on the stability of the direct current (DC) coefficient values in blocks transformed by discrete cosine transform (DCT). The frame is randomly selected to embed watermark, and then its luminance is blocked by 8-by-8 and transformed by DCT. The watermark is embedded in the high frequency coefficient of some consecutive DC coefficients which have been transformed by DCT. During the detection process, the correlation between the watermark and the high frequency coefficients of the DC coefficients transformed by DCT is computed to judge whether the frame has been embedded watermark. The experimental results indicate that the correlation curve peaks emerge at the watermarked I frames and P frames which suffered MPEG-4 and H.264 compression. The proposed watermarking method has strong robustness against some attacks such as frame deleting, frame inserting, frame shifting, frame statistical average and collusion attack.
    BibTeX:
    @inproceedings{2005_Chena,
      author = {Tian-Hang Chen and Shao-Hui Liu and Hong-Xun Yao and Wen Gao},
      title = {Robust video watermarking based on DC coefficients of selected blocks},
      journal = {Machine Learning and Cybernetics, 2005. Proceedings of 2005 International Conference on},
      year = {2005},
      volume = {9},
      pages = {5273 -5278 Vol. 9},
      doi = {http://dx.doi.org/10.1109/ICMLC.2005.1527875}
    }
    					
    Dai, Q.; Chen, X. & Lin, C. Fast algorithms for multidimensional DCT-to-DCT computation between a block and its associated subblocks 2005 Signal Processing, IEEE Transactions on
    Vol. 53 (8) , pp. 3219 - 3225  
    article DOI    
    Abstract: In this paper, we first propose an efficient algorithm for computing one-dimensional (1-D) discrete cosine transform (DCT) for a signal block, given its two adjacent subblocks in the DCT domain and then introduce several algorithms for the fast computation of multidimensional (m-D) DCT with size N1 times;N2 times;... times;Nm given 2m subblocks of DCT coefficients with size N1/2 times;N2/2 times;... times;Nm/2, where Ni(i=1,2,...,m) are powers of 2. Obviously, the row-column method, which employs the most efficient algorithms along each dimension, reduces the computational complexity considerably, compared with the traditional method, which employs only the one-dimensional (1-D) fast DCT and inverse DCT (IDCT) algorithms. However, when m ge;2, the traditional method, which employs the most efficient multidimensional DCT/IDCT algorithms, has lower computational complexity than the row-column method. Besides, we propose a direct method by dividing the data into 2m parts for independent fast computation, in which only two steps of r-dimensional (r=1,2,...,m) IDCT and additional multiplications and additions are required. If all the dimensional sizes are the same, the number of multiplications required for the direct method is only (2m-1)/m2m-1 times of that required for the row-column method, and if N ge;22m-1, the computational efficiency of the direct method is surely superior to that of the traditional method, which employs the most efficient multidimensional DCT/IDCT algorithms.
    BibTeX:
    @article{2005_Dai,
      author = {Qionghai Dai and Xinjian Chen and Chuang Lin},
      title = {Fast algorithms for multidimensional DCT-to-DCT computation between a block and its associated subblocks},
      journal = {Signal Processing, IEEE Transactions on},
      year = {2005},
      volume = {53},
      number = {8},
      pages = { 3219 - 3225},
      doi = {http://dx.doi.org/10.1109/TSP.2005.851115}
    }
    					
    Dong, P.; Brankov, J.; Galatsanos, N.; Yang, Y. & Davoine, F. Digital watermarking robust to geometric distortions 2005 Image Processing, IEEE Transactions on
    Vol. 14 (12) , pp. 2140 -2150  
    article DOI    
    Abstract: In this paper, we present two watermarking approaches that are robust to geometric distortions. The first approach is based on image normalization, in which both watermark embedding and extraction are carried out with respect to an image normalized to meet a set of predefined moment criteria. We propose a new normalization procedure, which is invariant to affine transform attacks. The resulting watermarking scheme is suitable for public watermarking applications, where the original image is not available for watermark extraction. The second approach is based on a watermark resynchronization scheme aimed to alleviate the effects of random bending attacks. In this scheme, a deformable mesh is used to correct the distortion caused by the attack. The watermark is then extracted from the corrected image. In contrast to the first scheme, the latter is suitable for private watermarking applications, where the original image is necessary for watermark detection. In both schemes, we employ a direct-sequence code division multiple access approach to embed a multibit watermark in the discrete cosine transform domain of the image. Numerical experiments demonstrate that the proposed watermarking schemes are robust to a wide range of geometric attacks.
    BibTeX:
    @article{2005_Dong,
      author = {Ping Dong and Brankov, J.G. and Galatsanos, N.P. and Yongyi Yang and Davoine, F.},
      title = {Digital watermarking robust to geometric distortions},
      journal = {Image Processing, IEEE Transactions on},
      year = {2005},
      volume = {14},
      number = {12},
      pages = {2140 -2150},
      doi = {http://dx.doi.org/10.1109/TIP.2005.857263}
    }
    					
    Fang, B.; Shen, G.; Li, S. & Chen, H. Techniques for efficient DCT/IDCT implementation on generic GPU 2005 Circuits and Systems, 2005. ISCAS 2005. IEEE International Symposium on , pp. 1126 - 1129 Vol. 2   inproceedings DOI    
    Abstract: The emergence of programmable graphics processing units (GPU) has led to increasing interest in off-loading numerically intensive computations on to graphics hardware. DCT/IDCT is widely adopted in modern image/video compression standards and is usually one of the most computationally expensive parts. We present several techniques for efficient implementation of DCT/IDCT on generic programmable GPU, using direct matrix multiplication. Our experimental results demonstrate that the speed of IDCT on a GPU using the proposed techniques can well exceed that on a CPU with MMX optimization.
    BibTeX:
    @inproceedings{2005_Fang,
      author = {Bo Fang and Guobin Shen and Shipeng Li and Huifang Chen},
      title = {Techniques for efficient DCT/IDCT implementation on generic GPU},
      journal = {Circuits and Systems, 2005. ISCAS 2005. IEEE International Symposium on},
      year = {2005},
      pages = { 1126 - 1129 Vol. 2},
      doi = {http://dx.doi.org/10.1109/ISCAS.2005.1464791}
    }
    					
    Grecos, C. & Yang, M.Y. Fast inter mode prediction for P slices in the H264 video coding standard 2005 Broadcasting, IEEE Transactions on
    Vol. 51 (2) , pp. 256 - 263  
    article DOI    
    Abstract: We propose an inter mode decision scheme for P slices in the H264 video coding standard. Our scheme initially exploits neighborhood information jointly with a set of skip mode conditions for enhanced skip mode decision. It subsequently performs inter mode decision for the remaining macroblocks by using a gentle set of smoothness constraints. For RD performance very close to the standard, we achieve 35-58% reduction in run times and 33-55% reduction in CPU cycles for both the rate controlled and the non rate controlled versions of H264. Compared to other non gradient based work that has been proposed as input to the standard, gains of 9-23% in run times and 7-22% in CPU cycles are also reported. Finally, with respect to other gradient based approaches for mode decision, gains of 9% are reported in run times.
    BibTeX:
    @article{2005_Grecos,
      author = {Grecos, C. and Ming Yuan Yang},
      title = {Fast inter mode prediction for P slices in the H264 video coding standard},
      journal = {Broadcasting, IEEE Transactions on},
      year = {2005},
      volume = {51},
      number = {2},
      pages = { 256 - 263},
      doi = {http://dx.doi.org/10.1109/TBC.2005.846192}
    }
    					
    Harmanci, O. & Tekalp, A. A Zero Error Propagation Extension to H264 for Low Delay Video Communications Over Lossy Channels 2005 Acoustics, Speech, and Signal Processing, 2005. Proceedings. (ICASSP '05). IEEE International Conference on
    Vol. 2 , pp. 185 - 188  
    inproceedings    
    BibTeX:
    @inproceedings{2005_Harmanci,
      author = { Harmanci, O. and Tekalp, A.M.},
      title = {A Zero Error Propagation Extension to H264 for Low Delay Video Communications Over Lossy Channels},
      journal = {Acoustics, Speech, and Signal Processing, 2005. Proceedings. (ICASSP '05). IEEE International Conference on},
      year = {2005},
      volume = {2},
      pages = { 185 - 188}
    }
    					
    Liu, L.-S.; Li, R.-H. & Gao, Q. A robust video watermarking scheme based on DCT 2005 Machine Learning and Cybernetics, 2005. Proceedings of 2005 International Conference on
    Vol. 8 , pp. 5176 -5180 Vol. 8  
    inproceedings DOI    
    Abstract: Using the stability of the direct current (DC) coefficient values in blocked discrete cosine transformation (DCT), a blind video watermarking method is proposed in the paper. The frame to embed watermark is randomly selected, and its luminance is transformed by 16-by-16 DCT. The watermark is embedded in the low frequency coefficients of the DCT blocks, which are chosen by its DC coefficient values. The watermark is detected by computing the correlation between the watermark and the absolute values of the low frequency coefficients of the blocks selected by DC values. The experimental results demonstrate, whether I, B or P frames are watermarked, the correlation curve emerges peaks in the watermarked frames which suffered compression and decompression with different bit rate. The watermark scheme has strong robustness against unintentional or malicious attacks such as frames deleting, frames inserting, frame statistical average, and collusion attack.
    BibTeX:
    @inproceedings{2005_Liu,
      author = {Lian-Shan Liu and Ren-Hou Li and Qi Gao},
      title = {A robust video watermarking scheme based on DCT},
      journal = {Machine Learning and Cybernetics, 2005. Proceedings of 2005 International Conference on},
      year = {2005},
      volume = {8},
      pages = {5176 -5180 Vol. 8},
      doi = {http://dx.doi.org/10.1109/ICMLC.2005.1527856}
    }
    					
    Modarressi, M. & Sarbazi-Azad, H. Parallel 3-dimensional DCT computation on k-ary n-cubes 2005 High-Performance Computing in Asia-Pacific Region, 2005. Proceedings. Eighth International Conference on , pp. 7 pp. -97   inproceedings DOI    
    Abstract: The three dimensional discrete cosine transform (3D DCT) has been widely used in many applications such as video compression. On the other hand, the k-ary n-cube is one of the most popular interconnection networks used in many recent multicomputers. As direct calculation of 3D DCT is very time consuming, many researchers have been working on developing algorithms and special-purpose architectures for fast computation of 3D DCT. This paper proposes a parallel algorithm for efficient calculation of 3D DCT on the k-ary n-cube multicomputers. The time complexity of the proposed algorithm is of O(N) for an N times N times N input data cube while direct calculation of 3D DCT has a complexity of O(N6)
    BibTeX:
    @inproceedings{2005_Modarressi,
      author = {Modarressi, M. and Sarbazi-Azad, H.},
      title = {Parallel 3-dimensional DCT computation on k-ary n-cubes},
      journal = {High-Performance Computing in Asia-Pacific Region, 2005. Proceedings. Eighth International Conference on},
      year = {2005},
      pages = {7 pp. -97},
      doi = {http://dx.doi.org/10.1109/HPCASIA.2005.64}
    }
    					
    Nunez, J. & Chouliaras, V. High-performance arithmetic coding VLSI macro for the H264 video compression standard 2005 Consumer Electronics, IEEE Transactions on
    Vol. 51 (1) , pp. 144 - 151  
    article DOI    
    Abstract: This paper investigates the algorithmic complexity of arithmetic coding in the new H264 video coding standard and proposes a processor-coprocessor architecture to reduce it by more than an order of magnitude. The proposed coprocessor is based on an innovative algorithm known as the MZ-coder and maintains the original coding efficiency via a low-complexity, multiplication-free, non-stalling, fully pipelined architecture. The coprocessor achieves a constant throughput for both coding and decoding processes of 1 symbol per cycle and is designed to be attached to a controlling embedded RISC CPU whose instruction set has been extended with arithmetic coding instructions.
    BibTeX:
    @article{2005_Nunez,
      author = {Nunez, J.L. and Chouliaras, V.A.},
      title = {High-performance arithmetic coding VLSI macro for the H264 video compression standard},
      journal = {Consumer Electronics, IEEE Transactions on},
      year = {2005},
      volume = {51},
      number = {1},
      pages = { 144 - 151},
      doi = {http://dx.doi.org/10.1109/TCE.2005.1405712}
    }
    					
    Panainte, E.; Bertels, K. & Vassiliadis, S. Instruction scheduling for dynamic hardware configurations [M-JPEG encoder case study] 2005 Design, Automation and Test in Europe, 2005. Proceedings , pp. 100 - 105 Vol. 1   inproceedings DOI    
    Abstract: Although the huge reconfiguration latency of the available FPGA platforms is a well-known shortcoming of the current FCCMs, little research in instruction scheduling has been undertaken to eliminate or diminish its negative influence on performance. In this paper, we introduce an instruction scheduling algorithm that minimizes the number of executed hardware reconfiguration instructions, taking into account the "FPGA area placement conflicts" between the available configurations. The algorithm is based on compiler analyses and feedback-directed techniques and it can switch from hardware execution to software execution for an operation, when the reconfiguration latency could not be reduced. The algorithm has been tested for the M-JPEG encoder application and the real hardware implementations for DCT quantization and VLC operations. Based on simulation results, we determine that, while a simple scheduling produces a significant performance decrease, our proposed scheduling contributes up to 16 times; M-JPEG encoder speedup.
    BibTeX:
    @inproceedings{2005_Panainte,
      author = {Panainte, E.M. and Bertels, K. and Vassiliadis, S.},
      title = {Instruction scheduling for dynamic hardware configurations [M-JPEG encoder case study]},
      journal = {Design, Automation and Test in Europe, 2005. Proceedings},
      year = {2005},
      pages = { 100 - 105 Vol. 1},
      doi = {http://dx.doi.org/10.1109/DATE.2005.184}
    }
    					
    Park, C.Y. & Cho, N.I. A fast algorithm for the conversion of DCT coefficients to H.264 transform coefficients 2005 Image Processing, 2005. ICIP 2005. IEEE International Conference on
    Vol. 3 , pp. III - 664-7  
    inproceedings DOI    
    Abstract: This paper proposes a fast algorithm that converts DCT coefficients into integer transform coefficients, for the transform domain transcoding from MPEG-x to H.264. For the transcoding in the same resolution, the 8 times; 8 DCT coefficients are converted to four 4 times; 4 integer transform coefficients by decomposing the conversion matrix into sparse ones. For the reduction of resolution by half, we also propose an algorithm that converts DCT coefficients in the lower band into 4 times; 4 integer transform coefficients. The sparse matrices derived in this paper require fewer computations than the direct and conventional conversion matrices, and thus the overall transcoding using the proposed algorithm requires less computational complexity.
    BibTeX:
    @inproceedings{2005_Park,
      author = {Chan Yul Park and Nam Ik Cho},
      title = {A fast algorithm for the conversion of DCT coefficients to H.264 transform coefficients},
      journal = {Image Processing, 2005. ICIP 2005. IEEE International Conference on},
      year = {2005},
      volume = {3},
      pages = { III - 664-7},
      doi = {http://dx.doi.org/10.1109/ICIP.2005.1530479}
    }
    					
    Pei, S.-C. & Kao, M.-P. Direct N-point DCT computation from three adjacent N/3-point DCT coefficients 2005 Signal Processing Letters, IEEE
    Vol. 12 (2) , pp. 89 - 92  
    article DOI    
    Abstract: An efficient method for computing a length-N Discrete Cosine Transform (DCT) from three consecutive length-N/3 DCTs is proposed. This method differs from previous ones in that it reduces considerable 31-38% arithmetic operations and uses only length-N/3 DCTs instead of length-N DCTs. We also find its great applications in fractional scaling of a block DCT-based image by the factor of 1/2 alpha;3 beta;. This would be very useful in high-definition television (HDTV) standard, whose display size is usually 16:9. The comparison with conventional methods is provided.
    BibTeX:
    @article{2005_Pei,
      author = {Soo-Chang Pei and Meng-Ping Kao},
      title = {Direct N-point DCT computation from three adjacent N/3-point DCT coefficients},
      journal = {Signal Processing Letters, IEEE},
      year = {2005},
      volume = {12},
      number = {2},
      pages = { 89 - 92},
      doi = {http://dx.doi.org/10.1109/LSP.2004.840868(410) 12}
    }
    					
    Raghavan, V.; Prabhu, K. & Sommen, P. An analysis of real-Fourier domain-based adaptive algorithms implemented with the Hartley transform using cosine-sine symmetries 2005 Signal Processing, IEEE Transactions on
    Vol. 53 (2) , pp. 622 - 629  
    article DOI    
    Abstract: The least mean squared (LMS) algorithm and its variants have been the most often used algorithms in adaptive signal processing. However the LMS algorithm suffers from a high computational complexity, especially with large filter lengths. The Fourier transform-based block normalized LMS (FBNLMS) reduces the computation count by using the discrete Fourier transform (DFT) and exploiting the fast algorithms for implementing the DFT. Even though the savings achieved with the FBNLMS over the direct-LMS implementation are significant, the computational requirements of FBNLMS are still very high, rendering many real-time applications, like audio and video estimation, infeasible. The Hartley transform-based BNLMS (HBNLMS) is found to have a computational complexity much less than, and a memory requirement almost of the same order as, that of the FBNLMS. This paper is based on the cosine and sine symmetric implementation of the discrete Hartley transform (DHT), which is the key in reducing the computational complexity of the FBNLMS by 33% asymptotically (with respect to multiplications). The parallel implementation of the discrete cosine transform (DCT) in turn can lead to more efficient implementations of the HBNLMS.
    BibTeX:
    @article{2005_Raghavan,
      author = {Raghavan, V. and Prabhu, K.M.M. and Sommen, P.C.W.},
      title = {An analysis of real-Fourier domain-based adaptive algorithms implemented with the Hartley transform using cosine-sine symmetries},
      journal = {Signal Processing, IEEE Transactions on},
      year = {2005},
      volume = {53},
      number = {2},
      pages = { 622 - 629},
      doi = {http://dx.doi.org/10.1109/TSP.2004.838983}
    }
    					
    Starck, J.-L.; Elad, M. & Donoho, D. Image decomposition via the combination of sparse representations and a variational approach 2005 Image Processing, IEEE Transactions on
    Vol. 14 (10) , pp. 1570 -1582  
    article DOI    
    Abstract: The separation of image content into semantic parts plays a vital role in applications such as compression, enhancement, restoration, and more. In recent years, several pioneering works suggested such a separation be based on variational formulation and others using independent component analysis and sparsity. This paper presents a novel method for separating images into texture and piecewise smooth (cartoon) parts, exploiting both the variational and the sparsity mechanisms. The method combines the basis pursuit denoising (BPDN) algorithm and the total-variation (TV) regularization scheme. The basic idea presented in this paper is the use of two appropriate dictionaries, one for the representation of textures and the other for the natural scene parts assumed to be piecewise smooth. Both dictionaries are chosen such that they lead to sparse representations over one type of image-content (either texture or piecewise smooth). The use of the BPDN with the two amalgamed dictionaries leads to the desired separation, along with noise removal as a by-product. As the need to choose proper dictionaries is generally hard, a TV regularization is employed to better direct the separation process and reduce ringing artifacts. We present a highly efficient numerical scheme to solve the combined optimization problem posed by our model and to show several experimental results that validate the algorithm's performance.
    BibTeX:
    @article{2005_Starck,
      author = {Starck, J.-L. and Elad, M. and Donoho, D.L.},
      title = {Image decomposition via the combination of sparse representations and a variational approach},
      journal = {Image Processing, IEEE Transactions on},
      year = {2005},
      volume = {14},
      number = {10},
      pages = {1570 -1582},
      doi = {http://dx.doi.org/10.1109/TIP.2005.852206}
    }
    					
    Wang, H. & Lv, J. A novel error concealment scheme for intra frames of H.264 video 2005 VLSI Design and Video Technology, 2005. Proceedings of 2005 IEEE International Workshop on , pp. 300 - 303   inproceedings DOI    
    Abstract: Several methods had been proposed for error concealment of H.264 video either in spatial domain or temporal domain. For H.264 video the macroblock is the more basically block unit and it tends to be lost in whole in an error prone channel. Single concealment approach can not achieve acceptable performance for a block of size 16 times;16. The scheme proposed is a combination of spatial concealment and temporal concealment and it has been demonstrated that it can significantly improve the quality of the video. Fast-DCT based spatial domain interpolation approach (Z. Alkachouh and M.G. Bellanger, 2000] and mean interpolation method are employed here for spatial domain concealment. For temporal domain direct copy is used after a determination operation.
    BibTeX:
    @inproceedings{2005_Wang,
      author = {Haiying Wang and Jing Lv},
      title = {A novel error concealment scheme for intra frames of H.264 video},
      journal = {VLSI Design and Video Technology, 2005. Proceedings of 2005 IEEE International Workshop on},
      year = {2005},
      pages = { 300 - 303},
      doi = {http://dx.doi.org/10.1109/IWVDVT.2005.1504610}
    }
    					
    Zhan, Y. & Sapatnekar, S. Fast computation of the temperature distribution in VLSI chips using the discrete cosine transform and table look-up 2005 Design Automation Conference, 2005. Proceedings of the ASP-DAC 2005. Asia and South Pacific
    Vol. 1 , pp. 87 - 92 Vol. 1  
    inproceedings DOI    
    Abstract: Temperature-related effects are critical in determining both the performance and reliability of VLSI circuits. Accurate and efficient estimation of the temperature distribution corresponding to a specific circuit layout is indispensable in physical design automation tools. In this paper, we propose a highly accurate fast algorithm for computing the on-chip temperature distribution due to power sources located on the top surface of the chip. The method is a combination of several computational techniques including the Green function method, the discrete cosine transform (DCT), and the table look-up technique. The high accuracy of the algorithm comes from the fully analytical nature of the Green function method, and the high efficiency is due to the application of the fast Fourier transform (FFT) technique to compute the DCT and later obtaining the temperature field for any power source distribution using the pre-calculated look-up table. Experimental results have demonstrated that our method has a relative error of below 1% compared with commercial computational fluid dynamic (CFD) softwares for thermal analysis, while the efficiency of our method is orders of magnitude higher than the direct application of the Green function method.
    BibTeX:
    @inproceedings{2005_Zhan,
      author = {Yong Zhan and Sapatnekar, S.S.},
      title = {Fast computation of the temperature distribution in VLSI chips using the discrete cosine transform and table look-up},
      journal = {Design Automation Conference, 2005. Proceedings of the ASP-DAC 2005. Asia and South Pacific},
      year = {2005},
      volume = {1},
      pages = { 87 - 92 Vol. 1},
      doi = {http://dx.doi.org/10.1109/ASPDAC.2005.1466136}
    }
    					
    Balter, R.; Gioia, P.; Morin, L. & Galpin, F. Scalable and efficient coding of 3D model extracted from a video 2004 3D Data Processing, Visualization and Transmission, 2004. 3DPVT 2004. Proceedings. 2nd International Symposium on , pp. 836 - 843   inproceedings DOI    
    Abstract: This work presents an efficient and scalable coding scheme for transmitting a stream of 3D models extracted from a video. As in classical model-based video coding, the geometry, connectivity, and texture of the 3D models have to be transmitted, as well as the camera position for each frame in the original video. The proposed method is based on exploiting the interrelations existing between each type of information, instead of coding them independently, allowing a better prediction of the next 3D model in the stream. Scalability is achieved through the use of wavelet-based representations for both texture and geometry of the models. A consistent connectivity is built for all 3D models extracted from the video sequence, which allows a more compact representation and straightforward geometric morphing between successive models. Furthermore this leads to a consistent wavelet decomposition for 3D models in the stream. Quantitative and qualitative results for the proposed scheme are compared with the state of the art video coder H264, 3D model-based Galpin coder and independent MPEG4-based coding of the information. Targeted applications include distant visualization of the original video at very low bit-rate and interactive navigation in the extracted 3D scene on heterogeneous terminals.
    BibTeX:
    @inproceedings{2004_Balter,
      author = {Balter, R. and Gioia, P. and Morin, L. and Galpin, F.},
      title = {Scalable and efficient coding of 3D model extracted from a video},
      journal = {3D Data Processing, Visualization and Transmission, 2004. 3DPVT 2004. Proceedings. 2nd International Symposium on},
      year = {2004},
      pages = { 836 - 843},
      doi = {http://dx.doi.org/10.1109/TDPVT.2004.1335402}
    }
    					
    Bernardini, R.; Durigon, M.; Rinaldo, R.; Celetto, L. & Vitali, A. Polyphase spatial subsampling multiple description coding of video streams with H264 2004 Image Processing, 2004. ICIP '04. 2004 International Conference on
    Vol. 5 , pp. 3213 - 3216 Vol. 5  
    inproceedings DOI    
    Abstract: In this work, we propose a multiple description (MD) coding system for video streams. In particular, our scheme originates four descriptions from the spatially downsampled polyphase components of the original frames. Each description is compressed independently with the recent H264/AVC video coding standard, it is packetized and sent over an error prone network. In case of errors in one or more descriptions, appropriate concealing is applied at the receiver, before insertion of the corrected frames into the corresponding receiver frame buffers. We propose and compare different concealment solutions and a post processing stage to attenuate visual effects related to MD coding. We analyze the trade off between robustness to channel errors and coding efficiency, comparing the proposed technique with single description (SD) video coding with H264/AVC. Experimental results validate the effectiveness of the proposed scheme.
    BibTeX:
    @inproceedings{2004_Bernardini,
      author = {Bernardini, R. and Durigon, M. and Rinaldo, R. and Celetto, L. and Vitali, A.},
      title = {Polyphase spatial subsampling multiple description coding of video streams with H264},
      journal = {Image Processing, 2004. ICIP '04. 2004 International Conference on},
      year = {2004},
      volume = {5},
      pages = { 3213 - 3216 Vol. 5},
      doi = {http://dx.doi.org/10.1109/ICIP.2004.1421797}
    }
    					
    Chen, C.-H.; Liu, B.-D. & Yang, J.-F. Direct recursive structures for computing radix-r two-dimensional DCT/IDCT/DST/IDST 2004 Circuits and Systems I: Regular Papers, IEEE Transactions on
    Vol. 51 (10) , pp. 2017 - 2030  
    article DOI    
    Abstract: In this paper, new recursive structures for computing radix-r two-dimensional (2-D) discrete cosine transform (DCT) and 2-D inverse DCT (IDCT) are proposed. The 2-D DCT/IDCT are first decomposed into cosine-cosine and sine-sine transforms. Based on indexes of transform bases, the regular pre-addition preprocess is established and the recursive structures for 2-D DCT/IDCT, which can be realized in a second-order infinite-impulse response (IIR) filter, are derived without involving any transposition procedure. For computation of 2-D DCT/IDCT, the recursive loops of the proposed structures are less than that of one-dimensional DCT/IDCT recursive structures, which require data transposition to achieve the so-called row-column approach. With advantages of fewer recursive loops and no transposition, the proposed recursive structures achieve more accurate results and less power consumption than the existed methods. The regular and modular properties are suitable for very large-scale integration (VLSI) implementation. By using similar procedures, the recursive structures for 2-D DST and 2-D IDST are also proposed.
    BibTeX:
    @article{2004_Chen,
      author = {Che-Hong Chen and Bin-Da Liu and Jar-Ferr Yang},
      title = {Direct recursive structures for computing radix-r two-dimensional DCT/IDCT/DST/IDST},
      journal = {Circuits and Systems I: Regular Papers, IEEE Transactions on},
      year = {2004},
      volume = {51},
      number = {10},
      pages = { 2017 - 2030},
      doi = {http://dx.doi.org/10.1109/TCSI.2004.835685}
    }
    					
    Chuang, Y.-J.; Pan, T.-J. & Wu, J.-L. General splitting and merging of 2-D DCT in the DCT domain 2004 Circuits and Systems, 2004. MWSCAS '04. The 2004 47th Midwest Symposium on
    Vol. 1 , pp. I - 17-20 vol.1  
    inproceedings DOI    
    Abstract: An efficient method for splitting an N times;N 2-D DCT block into four (N/2) times;(N/2) or two N times;(N/2) (or (N/2) times;N) 2-D DCT blocks is presented and vice versa. The computational complexity of the proposed methods is lower than the direct approach and the same as the most efficient converting approach existed in the literature. Besides, the proposed DCT splitter/merger is suitable for implementation by specific multimedia instruction set available nowadays. When N=8, our method can be applied to realize the transcoding between the latest video coding standards AVC/H.264 and the older ones, such as MPEG-1, MPEG-2 and MPEG-4 part 2.
    BibTeX:
    @inproceedings{2004_Chuang,
      author = {Yuh-Jue Chuang and Ting-Jian Pan and Ja-Ling Wu},
      title = {General splitting and merging of 2-D DCT in the DCT domain},
      journal = {Circuits and Systems, 2004. MWSCAS '04. The 2004 47th Midwest Symposium on},
      year = {2004},
      volume = {1},
      pages = { I - 17-20 vol.1},
      doi = {http://dx.doi.org/10.1109/MWSCAS.2004.1353886}
    }
    					
    Dasu, A. & Panchanathan, S. A wavelet-based sprite codec 2004 Circuits and Systems for Video Technology, IEEE Transactions on
    Vol. 14 (2) , pp. 244 - 255  
    article DOI    
    Abstract: The International Standards Organization (ISO) has proposed a family of standards for compression of image and video sequences, including the JPEG, MPEG-1, and MPEG-2. The latest MPEG-4 standard has many new dimensions to coding and manipulation of visual content. A video sequence usually contains a background object and many foreground objects. Portions of this background may not be visible in certain frames due to the occlusion of the foreground objects or camera motion. MPEG-4 introduces the novel concepts of video object planes (VOPs) and Sprites. A VOP is a visual representation of real world objects with shapes that need not be rectangular. Sprite is a large image composed of pixels belonging to a video object visible throughout a video segment. Since a sprite contains all parts of the background that were at least visible once, it can be used for direct reconstruction of the background VOP. Sprite reconstruction is dependent on the mode in which it is transmitted. In the static sprite mode, the entire sprite is decoded as an Intra VOP before decoding the individual VOPs. Since sprites consist of the information needed to display multiple frames of a video sequence, they are typically much larger than a single frame of video. Therefore, a static sprite can be considered as a large static image. In this paper, a novel solution to address the problem of spatial scalability has been proposed, where the sprite is encoded in discrete wavelet transform (DWT). A lifting kernel method of DWT implementation has been used for encoding and decoding sprites. Modifying the existing lifting scheme while maintaining it to be shape-adaptive results in a reduced complexity. The proposed scheme has the advantages of: 1) avoiding the need for any extensions to image or tile border pixels and is hence superior to the discrete cosine transform-based low latency scheme (used in the current MPEG-4 verification model) and 2) mapping the in place computed wavelet coefficients into a zero-tree structure without actually rearranging them, thereby saving allocation of additional memory. The proposed solutions provide efficient implementation of the sprite decoder, making possible a VLSI realization with a reduced real estate.
    BibTeX:
    @article{2004_Dasu,
      author = {Dasu, A.R. and Panchanathan, S.},
      title = {A wavelet-based sprite codec},
      journal = {Circuits and Systems for Video Technology, IEEE Transactions on},
      year = {2004},
      volume = {14},
      number = {2},
      pages = { 244 - 255},
      doi = {http://dx.doi.org/10.1109/TCSVT.2003.819185}
    }
    					
    Fung, K.-T.; Chan, Y.-L. & Siu, W.-C. Low-complexity and high-quality frame-skipping transcoder for continuous presence multipoint video conferencing 2004 Multimedia, IEEE Transactions on
    Vol. 6 (1) , pp. 31 - 46  
    article DOI    
    Abstract: This paper presents a new frame-skipping transcoding approach for video combiners in multipoint video conferencing. Transcoding is regarded as a process of converting a previously compressed video bitstream into a lower bitrate bitstream. A high transcoding ratio may result in an unacceptable picture quality when the incoming video bitstream is transcoded with the full frame rate. Frame skipping is often used as an efficient scheme to allocate more bits to representative frames, so that an acceptable quality for each frame can be maintained. However, the skipped frame must be decompressed completely, and should act as the reference frame to the nonskipped frame for reconstruction. The newly quantized DCT coefficients of prediction error need to be recomputed for the nonskipped frame with reference to the previous nonskipped frame; this can create an undesirable complexity in the real time application as well as introduce re-encoding error. A new frame-skipping transcoding architecture for improved picture quality and reduced complexity is proposed. The proposed architecture is mainly performed on the discrete cosine transform (DCT) domain to achieve a low complexity transcoder. It is observed that the re-encoding error is avoided at the frame-skipping transcoder when the strategy of direct summation of DCT coefficients is employed. By using the proposed frame-skipping transcoder and dynamically allocating more frames to the active participants in video combining, we are able to make more uniform peak signal-to-noise ratio (PSNR) performance of the subsequences and the video qualities of the active subsequences can be improved significantly.
    BibTeX:
    @article{2004_Fung,
      author = {Kai-Tat Fung and Yui-Lam Chan and Wan-Chi Siu},
      title = {Low-complexity and high-quality frame-skipping transcoder for continuous presence multipoint video conferencing},
      journal = {Multimedia, IEEE Transactions on},
      year = {2004},
      volume = {6},
      number = {1},
      pages = { 31 - 46},
      doi = {http://dx.doi.org/10.1109/TMM.2003.819761}
    }
    					
    Gunturk, B.; Altunbasak, Y. & Mersereau, R. Super-resolution reconstruction of compressed video using transform-domain statistics 2004 Image Processing, IEEE Transactions on
    Vol. 13 (1) , pp. 33 -43  
    article DOI    
    Abstract: Considerable attention has been directed to the problem of producing high-resolution video and still images from multiple low-resolution images. This multiframe reconstruction, also known as super-resolution reconstruction, is beginning to be applied to compressed video. Super-resolution techniques that have been designed for raw (i.e., uncompressed) video may not be effective when applied to compressed video because they do not incorporate the compression process into their models. The compression process introduces quantization error, which is the dominant source of error in some cases. In this paper, we propose a stochastic framework where quantization information as well as other statistical information about additive noise and image prior can be utilized effectively.
    BibTeX:
    @article{2004_Gunturk,
      author = {Gunturk, B.K. and Altunbasak, Y. and Mersereau, R.M.},
      title = {Super-resolution reconstruction of compressed video using transform-domain statistics},
      journal = {Image Processing, IEEE Transactions on},
      year = {2004},
      volume = {13},
      number = {1},
      pages = {33 -43},
      doi = {http://dx.doi.org/10.1109/TIP.2003.819221}
    }
    					
    Harmanci, O. & Tekalp, A. Optimization of H264 for low delay video communications over lossy channels 2004 Image Processing, 2004. ICIP '04. 2004 International Conference on
    Vol. 5 , pp. 3209 - 3212 Vol. 5  
    inproceedings DOI    
    Abstract: In this paper, we study the data partitioning (DP) and its optimization for H264 video coding standard. H264 does not include DP in baseline profile, which is the most suitable profile for low delay, low complexity, and loss prone environments. To analyze the optimization of DP, we first introduce the concept of subchannels to abstract the physical layer. This allows us to move channel coding from application layer to physical layer. Then, we build the video encoder system around NEWPRED (1996) so that error propagation and its analysis is eliminated. Finally, we provide macroblock and slice level optimizations that result in optimal mode decisions and unequal error protection (LTEP) rates for data partitions. Experimental results show about 0.5 dB performance increase as compared to no data partitioning.
    BibTeX:
    @inproceedings{2004_Harmanci,
      author = {Harmanci, O. and Tekalp, A.M.},
      title = {Optimization of H264 for low delay video communications over lossy channels},
      journal = {Image Processing, 2004. ICIP '04. 2004 International Conference on},
      year = {2004},
      volume = {5},
      pages = { 3209 - 3212 Vol. 5},
      doi = {http://dx.doi.org/10.1109/ICIP.2004.1421796}
    }
    					
    Ito, H.; Magai, K.; Fujii, R. & Suzuki, M. Watermarking for JPEG image authentication surviving integer rounding in decompression 2004 Image Processing, 2004. ICIP '04. 2004 International Conference on
    Vol. 4 , pp. 2641 - 2644 Vol. 4  
    inproceedings DOI    
    Abstract: A watermarking scheme for JPEG image authentication is proposed. It is a direct extension of Wong's algorithm to JPEG-coded images where the signature is embedded at the end of scanned DCT coefficients. instead of LSB's of raw pixel values. We address a problem that the watermark disappears after the integer rounding in JPEG decompression and show that imposing a restriction on the quantization step sizes for DCT coefficients can solve this problem. To avoid the limitation on compressed picture quality set by this restriction, we introduce an embedding technique to use different quantization vectors for the watermarking and the JPEG compression. Simulation results are shown to verify the proposed scheme.
    BibTeX:
    @inproceedings{2004_Ito,
      author = {Ito, H. and Magai, K. and Fujii, R. and Suzuki, M.},
      title = {Watermarking for JPEG image authentication surviving integer rounding in decompression},
      journal = {Image Processing, 2004. ICIP '04. 2004 International Conference on},
      year = {2004},
      volume = {4},
      pages = { 2641 - 2644 Vol. 4},
      doi = {http://dx.doi.org/10.1109/ICIP.2004.1421646}
    }
    					
    Li, H.-J.; Hsu, C.-T. & Chen, M.-J. Fast multiple reference frame selection method for motion estimation in JVT/H.264 2004 Circuits and Systems, 2004. Proceedings. The 2004 IEEE Asia-Pacific Conference on
    Vol. 1 , pp. 605 - 608 vol.1  
    inproceedings DOI    
    Abstract: The three main reasons why the new H.264 (MPEG-4 AVC) video coding standard has a significant performance better than the other standards are the adoption of variable block sizes, multiple reference frames, and the consideration of rate distortion optimization within the codec. However, these features incur a considerable increase in encoder complexity. As for the multiple reference frames motion estimation, the increased computation is in proportion to the number of searched reference frames. In this paper, a fast multi-frame selection method is proposed for H.264 video coding. The proposed scheme can efficiently determine the best reference frame from the allowed five reference frames. Simulation results show that the speed of the proposed method is over two times faster than that of the original scheme adopted in JVT reference software JM73 while keeping the similar video quality and bit-rate.
    BibTeX:
    @inproceedings{2004_Li,
      author = {Hung-Ju Li and Ching-Ting Hsu and Mei-Juan Chen},
      title = {Fast multiple reference frame selection method for motion estimation in JVT/H.264},
      journal = {Circuits and Systems, 2004. Proceedings. The 2004 IEEE Asia-Pacific Conference on},
      year = {2004},
      volume = {1},
      pages = { 605 - 608 vol.1},
      doi = {http://dx.doi.org/10.1109/APCCAS.2004.1412835}
    }
    					
    Lian, S.; Sun, J. & Wang, Z. Perceptual cryptography on MPEG compressed videos 2004 Signal Processing, 2004. Proceedings. ICSP '04. 2004 7th International Conference on
    Vol. 3 , pp. 2371 - 2374 vol.3  
    inproceedings DOI    
    Abstract: In this paper, a perceptual cryptography on MPEG encoded video is presented. By confusing different number of DCT coefficients, encrypting different number of coefficients' signs, confusing DCT blocks in different color planes and encrypting motion vectors according to certain quality factor, the videos can be degraded to different degree. Experimental results show that its encryption strength can be adjusted according to certain quality factor, it supports direct bit-rate control and is of low cost, which makes it suitable for multimedia applications with real-time operation requirement, such as video conference, multimedia network, mobile multimedia and so on.
    BibTeX:
    @inproceedings{2004_Lian,
      author = {Shiguo Lian and Jinsheng Sun and Zhiquan Wang},
      title = {Perceptual cryptography on MPEG compressed videos},
      journal = {Signal Processing, 2004. Proceedings. ICSP '04. 2004 7th International Conference on},
      year = {2004},
      volume = {3},
      pages = { 2371 - 2374 vol.3},
      doi = {http://dx.doi.org/10.1109/ICOSP.2004.1442257}
    }
    					
    Lian, S.; Sun, J. & Wang, Z. A novel image encryption scheme based-on JPEG encoding 2004 Information Visualisation, 2004. IV 2004. Proceedings. Eighth International Conference on , pp. 217 - 220   inproceedings DOI    
    Abstract: Image encryption is a suitable method to protect image data. The encryption algorithms based on position confusion and pixel substitution change compression ratio greatly. In this paper, an image encryption algorithm combining with JPEG encoding is proposed. In luminance and chrominance plane, the DCT blocks are confused by pseudo-random SFCs (space filling curves). In each DCT block, DCT coefficients are confused according to different frequency bands and their signs are encrypted by a chaotic stream cipher. The security of the cryptosystem against brute-force attack and known-plaintext attack is also analyzed. Experimental results show that, the algorithm is of high security and low cost. What's more, it supports direct bit-rate control or recompression, which means that the encrypted image can still be decrypted correctly even if its compression ratio has been changed. These advantages make it suitable for image transmission over network.
    BibTeX:
    @inproceedings{2004_Liana,
      author = {Shiguo Lian and Jinsheng Sun and Zhiquan Wang},
      title = {A novel image encryption scheme based-on JPEG encoding},
      journal = {Information Visualisation, 2004. IV 2004. Proceedings. Eighth International Conference on},
      year = {2004},
      pages = { 217 - 220},
      doi = {http://dx.doi.org/10.1109/IV.2004.1320147}
    }
    					
    Lie, W.-N.; Tsai, M.-L. & Lin, T. Rate-distortion optimized DCT-domain video transcoder for bit-rate reduction of MPEG videos 2004 Acoustics, Speech, and Signal Processing, 2004. Proceedings. (ICASSP '04). IEEE International Conference on
    Vol. 5 , pp. V - 969-72 vol.5  
    inproceedings DOI    
    Abstract: In this paper, we propose a rate-distortion optimized video transcoder which converts MPEG videos into a similar form at lower bit-rates. Our transcoder design is characterized by two features: 1) transcoding is performed in the DCT domain and the motion vector information is re-used; and 2) the rate-distortion relationship is optimized in both the frame-level rate allocation and macroblock-level rate control, thus leading to performances even better than the direct encoding and re-encoding methods based on the well-known TM5. In the proposed algorithm, the Lagrangian multiplier plays not only its traditional role in macroblock-level optimization, but also a variable to be optimized in frame-level rate allocation. These two levels of optimization process are highly linked. Experiments show that the R-D optimization is effective in getting better video quality, even the drift errors are ignored. Several speedy schemes were developed to make our transcoder design suitable for real-time video transmission over heterogeneous networks.
    BibTeX:
    @inproceedings{2004_Lie,
      author = {Wen-Nung Lie and Ming-Lun Tsai and Lin, T.C.L.},
      title = {Rate-distortion optimized DCT-domain video transcoder for bit-rate reduction of MPEG videos},
      journal = {Acoustics, Speech, and Signal Processing, 2004. Proceedings. (ICASSP '04). IEEE International Conference on},
      year = {2004},
      volume = {5},
      pages = { V - 969-72 vol.5},
      doi = {http://dx.doi.org/10.1109/ICASSP.2004.1327274}
    }
    					
    Mansour, M. A memory and computation efficient structure for MPEG polyphase synthesis 2004 Acoustics, Speech, and Signal Processing, 2004. Proceedings. (ICASSP '04). IEEE International Conference on
    Vol. 5 , pp. V - 245-8 vol.5  
    inproceedings DOI    
    Abstract: We propose a new structure for polyphase synthesis in the MPEG-1 audio standard. The algorithm is based on factorizing the DCT matrix in a way similar to FFT factorization. The proposed algorithm reduces the memory requirement significantly even with direct implementation. With optimized implementation, the computational requirement can be reduced considerably as well.
    BibTeX:
    @inproceedings{2004_Mansour,
      author = {Mansour, M.F.},
      title = {A memory and computation efficient structure for MPEG polyphase synthesis},
      journal = {Acoustics, Speech, and Signal Processing, 2004. Proceedings. (ICASSP '04). IEEE International Conference on},
      year = {2004},
      volume = {5},
      pages = { V - 245-8 vol.5},
      doi = {http://dx.doi.org/10.1109/ICASSP.2004.1327093}
    }
    					
    Overbye, D. & Priemer, R. Incorporation of prior knowledge of chip shapes in blind multiuser detection for DS-CDMA communication 2004 Electro/Information Technology Conference, 2004. EIT 2004. IEEE , pp. 261 -271   inproceedings DOI    
    Abstract: In this paper we present a new method of compensating for complex multipath channel fading in blind multiuser detection for DS-CDMA communication. This compensation allows for the use of different chip pulse shapes. The compensation is based on the incorporation of priors derived from the FFT of the pulse shape. The system employs detectors that incorporate methods of independent component analysis (ICA), subspace estimation of channel noise, and Hopfield type neural networks. Several chip pulse shapes are tested, including the cosine, raised cosine, and the duobinary shapes, which are the pulse shapes that provide superior performance with more traditional detectors. They do not retain this superior performance when used with the uncompensated ICA detectors operating in a multipath fading channel environment. When compensated, the ICA neural network detectors displayed significantly better bit error rate performance than traditional detectors and the uncompensated ICA detectors.
    BibTeX:
    @inproceedings{2004_Overbye,
      author = {Overbye, D. and Priemer, R.},
      title = {Incorporation of prior knowledge of chip shapes in blind multiuser detection for DS-CDMA communication},
      journal = {Electro/Information Technology Conference, 2004. EIT 2004. IEEE},
      year = {2004},
      pages = {261 -271},
      doi = {http://dx.doi.org/10.1109/EIT.2004.4569391}
    }
    					
    Pei, S.-C. & Kao, M.-P. Direct N-point DCT computation from three adjacent N/3-point DCT coefficients 2004 Image Processing, 2004. ICIP '04. 2004 International Conference on
    Vol. 2 , pp. 1113 - 1116 Vol.2  
    inproceedings DOI    
    Abstract: An efficient method for computing a length-N DCT given three consecutive length-N/3 DCTs is proposed. This method differs from previous ones in that it reduces considerable arithmetic operations and uses only length-N/3 DCTs instead of length-N DCTs. We also find its great applications in fractional scaling of a DCT-based image by the factor of N/2 alpha;3 beta;. This would be very useful in HDTV standard, whose display size is usually 16:9. The comparison with conventional methods is provided in this paper.
    BibTeX:
    @inproceedings{2004_Pei,
      author = {Soo-Chang Pei and Meng-Ping Kao},
      title = {Direct N-point DCT computation from three adjacent N/3-point DCT coefficients},
      journal = {Image Processing, 2004. ICIP '04. 2004 International Conference on},
      year = {2004},
      volume = {2},
      pages = { 1113 - 1116 Vol.2},
      doi = {http://dx.doi.org/10.1109/ICIP.2004.1419498}
    }
    					
    Porto, R. & Agostini, L. Project space exploration on the 2-D DCT architecture of a JPEG compressor directed to FPGA implementation 2004 Design, Automation and Test in Europe Conference and Exhibition, 2004. Proceedings
    Vol. 3 , pp. 224 - 229 Vol.3  
    inproceedings DOI    
    Abstract: This paper presents a project space exploration on the baseline JPEG compressor proposed and implemented in previous works. This exploration took as basis the substitution of the operators used in the 2-D DCT calculation architecture of the compressor and the consequent evaluation of impact in terms of performance and resources utilization. This substitution was made with main focus in the carry lookahead, hierarchical carry lookahead and carry select architectures, with the objective to increase the JPEG compressor performance. As the compressor architecture was designed in an hierarchical mode the operators substitution was an activity quite simple, because it has not involved the other hierarchy levels. The operators were described in VHDL, synthesized and validated. They were inserted in the 2-D DCT architecture for synthesis in the whole module. The 2-D DCT was synthesized for an altera FPGA. With this project space exploration, the highest performance obtained for the 2-D DCT was 23% higher than the original, using 11% more logic cells.
    BibTeX:
    @inproceedings{2004_Porto,
      author = {Porto, R.E.C. and Agostini, L.V.},
      title = {Project space exploration on the 2-D DCT architecture of a JPEG compressor directed to FPGA implementation},
      journal = {Design, Automation and Test in Europe Conference and Exhibition, 2004. Proceedings},
      year = {2004},
      volume = {3},
      pages = { 224 - 229 Vol.3},
      doi = {http://dx.doi.org/10.1109/DATE.2004.1269234}
    }
    					
    Tai, S.-C.; Chen, Y.-R. & Li, S.-J. Low complexity variable-size block-matching motion estimation for adaptive motion compensation block size in H.264 2004 Circuits and Systems, 2004. Proceedings. The 2004 IEEE Asia-Pacific Conference on
    Vol. 1 , pp. 613 - 616 vol.1  
    inproceedings DOI    
    Abstract: An efficient merge-based algorithm for variable-size block matching ME (motion estimation) to reduce the computation of block-size decision process in H.264 video encoder is proposed. Fixed-size 4 times; 4 small blocks are used for the integer-pixel accuracy full search ME and several candidate MVs are reserved for each 4 times; 4 blocks to determined if the neighboring blocks could be merge into other predefined block types or not. Simulation result shows the computational load of our method is significantly reduced than H.264 JM 7.3.
    BibTeX:
    @inproceedings{2004_Tai,
      author = {Shen-Chuan Tai and Ying-Ru Chen and Sheng-Jia Li},
      title = {Low complexity variable-size block-matching motion estimation for adaptive motion compensation block size in H.264},
      journal = {Circuits and Systems, 2004. Proceedings. The 2004 IEEE Asia-Pacific Conference on},
      year = {2004},
      volume = {1},
      pages = { 613 - 616 vol.1},
      doi = {http://dx.doi.org/10.1109/APCCAS.2004.1412837}
    }
    					
    Tsai, T.-H. & Pan, Y.-N. A novel predict hexagon search algorithm for fast block motion estimation on H.264 video coding 2004 Circuits and Systems, 2004. Proceedings. The 2004 IEEE Asia-Pacific Conference on
    Vol. 1 , pp. 609 - 612 vol.1  
    inproceedings DOI    
    Abstract: The upcoming video coding standard, MPEG-4 AVC/JVT/H.264, motion estimation is allowed to use multiple references and multiple block sizes to improve the rate-distortion performance. However, full exhaustive search of all block sizes is computational intensive with complexity increasing linearly to the number of allowed reference frame and block size. A novel search algorithm, predict hexagon search (PHS), is proposed. The PHS pattern is applied by the comparison with the hexagon based search pattern. It can predict the hexagon search pattern in horizontal or vertical direction. Analysis shows that the speed improvement of the PHS over the diamond search (DS) and the hexagon based search (HEXBS) is about 58% and 53% respectively.
    BibTeX:
    @inproceedings{2004_Tsai,
      author = {Tsung-Han Tsai and Yu-Nan Pan},
      title = {A novel predict hexagon search algorithm for fast block motion estimation on H.264 video coding},
      journal = {Circuits and Systems, 2004. Proceedings. The 2004 IEEE Asia-Pacific Conference on},
      year = {2004},
      volume = {1},
      pages = { 609 - 612 vol.1},
      doi = {http://dx.doi.org/10.1109/APCCAS.2004.1412836}
    }
    					
    Wong, P.; Chang, A. & Au, O. On improving the iterative watermark embedding technique for JPEG-to-JPEG watermarking 2004 Circuits and Systems, 2004. ISCAS '04. Proceedings of the 2004 International Symposium on
    Vol. 2 , pp. II - 161-4 Vol.2  
    inproceedings    
    Abstract: We previously proposed an iterative technique called iterative watermark embedding (IWE) for watermark embedding in JPEG compressed domains. The proposed technique embeds the watermark successfully when the quality factor of the JPEG image is high, but when the quality factor is low, not all the bits of the watermark can be embedded. This is due to uneven distribution of capacities in different regions of the image. Moreover, the computation complexity of the iterative loop is high. In this paper we improve the technique using shuffling to equalize the capacities of different regions. Also, the iterative loop is replaced by a direct approach to reduce the computational complexity. We call this direct approach Direct JPEG Watermarking Embedding (DJWE). A Human Visual System (HVS) model is used to prioritize the DCT coefficients to achieve good visual quality. Experimental results show that the watermarks are robust against Gaussian noise attack and JPEG transcoding attacks.
    BibTeX:
    @inproceedings{2004_Wong,
      author = {Wong, P.H.W. and Andy Chang and Au, O.C.},
      title = {On improving the iterative watermark embedding technique for JPEG-to-JPEG watermarking},
      journal = {Circuits and Systems, 2004. ISCAS '04. Proceedings of the 2004 International Symposium on},
      year = {2004},
      volume = {2},
      pages = { II - 161-4 Vol.2}
    }
    					
    Yeh, Y.-H. & Chen, S.-G. DCT-based channel estimation for OFDM systems 2004 Communications, 2004 IEEE International Conference on
    Vol. 4 , pp. 2442 - 2446 Vol.4  
    inproceedings DOI    
    Abstract: In this paper, based on the property of channel frequency response and the concept of interpolation in transform domain, we propose two discrete cosine transform (DCT)-based pilot-symbol-aided channel estimators, which can mitigate the aliasing error and high-frequency distortion of the direct discrete Fourier transform (DFT)-based channel estimators when the multipath fading channels have non-sample-spaced path delays. Both proposed estimators outperform the conventional DFT-based channel estimators. Of these two DCT-based estimators, one has its performance close to MMSE estimator, while the other one has the advantage of easy implementation with a little performance degradation. Furthermore, in implementation, the DCT-based estimators have the advantages of utilizing mature fast DCT algorithms and architectures, which is favorable to matrix-based channel estimators.
    BibTeX:
    @inproceedings{2004_Yeh,
      author = {Yen-Hui Yeh and Sau-Gee Chen},
      title = {DCT-based channel estimation for OFDM systems},
      journal = {Communications, 2004 IEEE International Conference on},
      year = {2004},
      volume = {4},
      pages = { 2442 - 2446 Vol.4},
      doi = {http://dx.doi.org/10.1109/ICC.2004.1312957}
    }
    					
    Zhang, Z.; Park, J. & Kim, Y. A novel deblocking algorithm using edge flow-directed filter and curvelet transform 2004 Multimedia and Expo, 2004. ICME '04. 2004 IEEE International Conference on
    Vol. 1 , pp. 683 -686 Vol.1  
    inproceedings DOI    
    Abstract: A new post-processing approach based on "edge flow" and curvelet transform is proposed for the suppression blocking artifacts in block discrete cosine transform (BDCT) compressed images. Firstly, by exploiting the edge flow correlations, edge information in the compressed images is extracted and protected, while blocky noise in the smooth background regions is smoothed out by an edge flow-directed filter in the wavelet domain. Then, the curvelet transform coefficients in different subbands are filtered with adaptive thresholds that are obtained according to the edge flow boundary map. The advantage of the new method is that it retains sharp features in images and, compared with other wavelet-based methods, it is capable of achieving higher peak signal-to-noise ratio (PSNR) improvement as well as giving visually very pleasing images.
    BibTeX:
    @inproceedings{2004_Zhang,
      author = {ZhiMing Zhang and JeongHoon Park and YongJe Kim},
      title = {A novel deblocking algorithm using edge flow-directed filter and curvelet transform},
      journal = {Multimedia and Expo, 2004. ICME '04. 2004 IEEE International Conference on},
      year = {2004},
      volume = {1},
      pages = {683 -686 Vol.1},
      doi = {http://dx.doi.org/10.1109/ICME.2004.1394284}
    }
    					
    Bae, B.; Yang, S.W. & Ro, Y.M. Fast MPEG-7 visual descriptor extraction using DCT coefficient 2003 TENCON 2003. Conference on Convergent Technologies for Asia-Pacific Region
    Vol. 3 , pp. 1136 - 1139 Vol.3  
    inproceedings DOI    
    Abstract: TMPEG-7 descriptors are known as standard descriptors for the content-based indexing and retrieval. To extract visual features from a compressed image, the spatial domain information should be extracted from the compression domain information. In this paper, we propose the fast feature extraction method that can be applied to MPEG-7 visual descriptor from DCT compressed image. In the proposed method, fast extraction of descriptor is achieved by combination of down sampling and direct calculation of spatial information in DCT domain. Experiment is performed with MPEG-7 reference database. The results show that the proposed method can be applied to fast feature extraction with large size of database in Internet.
    BibTeX:
    @inproceedings{2003_Bae,
      author = {Bae, B. and Sun Woo Yang and Yong Man Ro},
      title = {Fast MPEG-7 visual descriptor extraction using DCT coefficient},
      journal = {TENCON 2003. Conference on Convergent Technologies for Asia-Pacific Region},
      year = {2003},
      volume = {3},
      pages = { 1136 - 1139 Vol.3},
      doi = {http://dx.doi.org/10.1109/TENCON.2003.1273424}
    }
    					
    Chen, C.-H.; Liu, B.-D. & Yang, J.-F. Direct recursive structures for computing radix-r two-dimensional DCT 2003 Circuits and Systems, 2003. ISCAS '03. Proceedings of the 2003 International Symposium on
    Vol. 4 , pp. IV-269 - IV-272 vol.4  
    inproceedings DOI    
    Abstract: In this paper, new recursive structures for computing radix-r two-dimensional discrete cosine transform (2-D DCT) are proposed. Based on the same indices of transform bases, the regular pre-add preprocess is established and the recursive structures for 2-D DCT, which can be realized in a second-order infinite-impulse response (IIR) filter, are derived without involving any transposition procedure. For computation of 2-D DCT, the recursive loops of the proposed structures are less than that of one-dimensional DCT recursive structures, which need data transposition to achieve the so-called row-column approach. With advantages of fewer recursive loops and no transposition, the proposed recursive structures achieve more accurate results than the existed methods. The regular and modular properties are suitable for VLSI implementation.
    BibTeX:
    @inproceedings{2003_Chen,
      author = {Che-Hong Chen and Bin-Da Liu and Jar-Ferr Yang},
      title = {Direct recursive structures for computing radix-r two-dimensional DCT},
      journal = {Circuits and Systems, 2003. ISCAS '03. Proceedings of the 2003 International Symposium on},
      year = {2003},
      volume = {4},
      pages = { IV-269 - IV-272 vol.4},
      doi = {http://dx.doi.org/10.1109/ISCAS.2003.1205825}
    }
    					
    Chung, Y.Y. & Wong, M.T. Implementation of digital watermarking system 2003 Consumer Electronics, 2003. ICCE. 2003 IEEE International Conference on , pp. 214 - 215   inproceedings DOI    
    Abstract: It is easy to make perfect unauthorised copies on a large scale in the music, film, book and software publishing industries. A digital watermark - a hidden copyright message with hidden serial numbers-can be used as a direct method of preventing unauthorised copying. This paper presents a novel digital watermarking system based on vector quantisation (VQ) and discrete cosine transform (DCT), which can embed a grey level image. The proposed new system can embed 16 times more information than other traditional DCT based watermarking system. The recovered images have high peak signal to noise ratio (PSNR) value and good visual quality. This system is robust in terms of both JPEG compression and other signal processing attacks.
    BibTeX:
    @inproceedings{2003_Chung,
      author = {Yuk Ying Chung and Man To Wong},
      title = {Implementation of digital watermarking system},
      journal = {Consumer Electronics, 2003. ICCE. 2003 IEEE International Conference on},
      year = {2003},
      pages = { 214 - 215},
      doi = {http://dx.doi.org/10.1109/ICCE.2003.1218891}
    }
    					
    Lai, Y.-K. & Hsu, H.-J. A cost-effective 2-D discrete cosine transform processor with reconfigurable datapath 2003 Circuits and Systems, 2003. ISCAS '03. Proceedings of the 2003 International Symposium on
    Vol. 2 , pp. II-492 - II-495 vol.2  
    inproceedings DOI    
    Abstract: In this paper, a cost-effective 2D discrete cosine transform processor using a reconfigurable datapath is described. The proposed architecture uses some multiplexers to reduce computational complexity. This processor operates on 8 times;8 blocks. Unlike other direct methods, the proposed architecture is regular for VLSI implementation. The proposed 2D DCT processor costs 38598 transistors, with an operating frequency of 100 MHz, using 0.35 mu;m CMOS technology.
    BibTeX:
    @inproceedings{2003_Lai,
      author = {Yeong-Kang Lai and Han-Jen Hsu},
      title = {A cost-effective 2-D discrete cosine transform processor with reconfigurable datapath},
      journal = {Circuits and Systems, 2003. ISCAS '03. Proceedings of the 2003 International Symposium on},
      year = {2003},
      volume = {2},
      pages = { II-492 - II-495 vol.2},
      doi = {http://dx.doi.org/10.1109/ISCAS.2003.1206018}
    }
    					
    Laurent, N.; Buisson, A.; Laurent, C. & Brangoulo, S. A hybrid mesh-H264 video coder 2003 Image Processing, 2003. ICIP 2003. Proceedings. 2003 International Conference on
    Vol. 3 , pp. III - 865-8 vol.2  
    inproceedings DOI    
    Abstract: This paper presents a video coding scheme combining H264 with a triangular mesh based representation. In this scheme, we developed a mesh-based motion estimator in which the motion vectors are computed at each node of the mesh. All other motion vectors are then interpolated using the nodes denoting the patches. Finally, this model is integrated in the RD optimized model of the H264 coder. To end, we propose some results, which compare H264 with our hybrid approach. This approach permits us to improve the quality of P and B images obtained by H264 for the same coding cost.
    BibTeX:
    @inproceedings{2003_Laurent,
      author = {Laurent, N. and Buisson, A. and Laurent, C. and Brangoulo, S.},
      title = {A hybrid mesh-H264 video coder},
      journal = {Image Processing, 2003. ICIP 2003. Proceedings. 2003 International Conference on},
      year = {2003},
      volume = {3},
      pages = { III - 865-8 vol.2},
      doi = {http://dx.doi.org/10.1109/ICIP.2003.1247382}
    }
    					
    Pelekanakis, C.; Stojanovic, M. & Freitag, L. High rate acoustic link for underwater video transmission 2003 OCEANS 2003. Proceedings
    Vol. 2 , pp. 1091 - 1097 Vol.2  
    inproceedings    
    Abstract: A high bit rate acoustic link for video transmission over an underwater channel is investigated. The key to achieving this objective lies in two approaches: use of efficient data compression algorithms and use of high-level bandwidth-efficient modulation methods. Currently available video encoding standards allow video transmission at bit rates as low as 64 kbps. While this rate is still above the limit of commercially available acoustic modems, prototype acoustic modems based on phase coherent modulation/detection have demonstrated successful transmission up to 30 kbps over a deep-water vertical path. To bridge the final gap and provide acoustic transmission capability needed for near real-time video, we focus on the use of high-level bandwidth-efficient modulation methods. An experimental system, based on discrete cosine transform and Huffman entropy coding for video compression, and variable rate M-ary QAM was implemented. Phase-coherent detection is accomplished by decision-directed synchronization and adaptive equalization. System performance is demonstrated experimentally, using 25000 symbols/sec at a carrier frequency of 75 kHz over a short vertical path. Excellent results were obtained using modulation methods of 16, 32 and 64-QAM, thus achieving bit rates as high as 150 kbps, which are sufficient for real-time transmission of compressed video.
    BibTeX:
    @inproceedings{2003_Pelekanakis,
      author = {Pelekanakis, C. and Stojanovic, M. and Freitag, L.},
      title = {High rate acoustic link for underwater video transmission},
      journal = {OCEANS 2003. Proceedings},
      year = {2003},
      volume = {2},
      pages = { 1091 - 1097 Vol.2}
    }
    					
    Reulet, P.; Nortershauser, D. & Millan, P. Inverse method using infrared thermography for surface temperature and heat flux measurements 2003 Instrumentation in Aerospace Simulation Facilities, 2003. ICIASF '03. 20th International Congress on , pp. 118 - 126   inproceedings DOI    
    Abstract: Temperature measurements, for the direct identification of the surface heat flux, are not always possible considering an aggressive environment or an inaccessible zone. That is why an inverse method has been developed: the direct problem with the unknown boundary condition (wall heat flux) is solved by adding an observation equation given by temperature measurements on the opposite face of the wall. In order to estimate spatiotemporal variations of the surface heat flux, an inverse non linear three-dimensional unsteady model has been developed. The resolution is based on the minimisation of a function representing the sum of the differences between the observations (measured surface temperatures) and the calculated temperature at the same position. A compression method using DCT is used to filter the temperature measurements and reduce the number of heat flux components to be estimated. First a numerical validation of the inverse model has been conducted: a flat plate submitted to a specified spatiotemporal heat flux evolution. Then two validation experiments have been developed: impact of a moving laser beam on a flat steel plate and flame-wall interaction. The estimation of high resolution unsteady heat flux cartographies have been demonstrated.
    BibTeX:
    @inproceedings{2003_Reulet,
      author = {Reulet, P. and Nortershauser, D. and Millan, P.},
      title = {Inverse method using infrared thermography for surface temperature and heat flux measurements},
      journal = {Instrumentation in Aerospace Simulation Facilities, 2003. ICIASF '03. 20th International Congress on},
      year = {2003},
      pages = { 118 - 126},
      doi = {http://dx.doi.org/10.1109/ICIASF.2003.1274861}
    }
    					
    Wong, P. & Au, O. A blind watermarking technique for multiple watermarks 2003 Circuits and Systems, 2003. ISCAS '03. Proceedings of the 2003 International Symposium on
    Vol. 2 , pp. II-936 - II-939 vol.2  
    inproceedings DOI    
    Abstract: We propose a blind watermarking technique to embed multiple watermarks simultaneously. It allows the use of correlated key to embed multiple watermarks. A dual-key system is used to reduce the chance of the removal of watermark. Each embedded watermark can be decoded/detected by its own key. One direct approach and two iterative approaches are proposed to reduce the correlation effects among the noise-like keys. Experimental results show that multiple watermarks can be embedded without significant loss in PSNR and the watermark can be detected when the watermarked image is JPEG compressed.
    BibTeX:
    @inproceedings{2003_Wong,
      author = {Wong, P.H.W. and Au, O.C.},
      title = {A blind watermarking technique for multiple watermarks},
      journal = {Circuits and Systems, 2003. ISCAS '03. Proceedings of the 2003 International Symposium on},
      year = {2003},
      volume = {2},
      pages = { II-936 - II-939 vol.2},
      doi = {http://dx.doi.org/10.1109/ISCAS.2003.1206129}
    }
    					
    Yaroslavsky, L. & Chernobrodov, Y. DFT and DCT based discrete sinc-interpolation methods for direct Fourier tomographic reconstruction 2003 Image and Signal Processing and Analysis, 2003. ISPA 2003. Proceedings of the 3rd International Symposium on
    Vol. 1 , pp. 405 - 410 Vol.1  
    inproceedings DOI    
    Abstract: The paper describes the implementation and use of recently developed DFT and DCT based discrete sinc-interpolation algorithms for the direct Fourier method of reconstructing images from projections. DFT based discrete sinc-interpolation is the only completely reversible discrete interpolation technique. DCT based discrete sinc-interpolation algorithm implements the same interpolation kernel that of the DFT based method and allows to avoid its boundary affect artifacts. It is also computationally more efficient. Two modifications of the DCT based interpolation for tomographic reconstruction are suggested: interpolation by global zooming of 1-D DFT spectra of projections and interpolation with variable zooming factor along the angle coordinate. The latter allows to substantially reduce the computational complexity without compromising interpolation accuracy. Results of comparative numerical simulation of suggested algorithms show a good image reconstruction quality with a reduced level of artifacts.
    BibTeX:
    @inproceedings{2003_Yaroslavsky,
      author = {Yaroslavsky, L.P. and Chernobrodov, Y.},
      title = {DFT and DCT based discrete sinc-interpolation methods for direct Fourier tomographic reconstruction},
      journal = {Image and Signal Processing and Analysis, 2003. ISPA 2003. Proceedings of the 3rd International Symposium on},
      year = {2003},
      volume = {1},
      pages = { 405 - 410 Vol.1},
      doi = {http://dx.doi.org/10.1109/ISPA.2003.1296931}
    }
    					
    Armstrong, A. & Jiang, J. Direct DCT indexing using genetic algorithm concepts 2002 Eurographics UK Conference, 2002. Proceedings. The 20th , pp. 61 - 66   inproceedings DOI    
    Abstract: It is highly desirable in terms of speed and computational costs to perform image indexing and retrieval in the compressed domain. The exponential growth of digital media on both the WWW (Reddy and Fletcher, 1998) and home imaging equipment has prompted the development of faster, more accurate indexing algorithms. Successful techniques have the ability to summarise the features of an image into a relatively small key. We present a technique that uses the features of a genetic algorithm to quantize the content of an image directly in the DCT domain. The benefits of working in this domain, apart from the substantial savings of avoiding image reconstruction, is that the coefficients provide spatial information in a form well suited for image indexing.
    BibTeX:
    @inproceedings{2002_Armstrong,
      author = {Armstrong, A. and Jiang, J.},
      title = {Direct DCT indexing using genetic algorithm concepts},
      journal = {Eurographics UK Conference, 2002. Proceedings. The 20th},
      year = {2002},
      pages = { 61 - 66},
      doi = {http://dx.doi.org/10.1109/EGUK.2002.1011273}
    }
    					
    Fung, K.-T.; Chan, Y.-L. & Siu, W.-C. New architecture for dynamic frame-skipping transcoder 2002 Image Processing, IEEE Transactions on
    Vol. 11 (8) , pp. 886 - 900  
    article DOI    
    Abstract: Transcoding is a key technique for reducing the bit rate of a previously compressed video signal. A high transcoding ratio may result in an unacceptable picture quality when the full frame rate of the incoming video bitstream is used. Frame skipping is often used as an efficient scheme to allocate more bits to the representative frames, so that an acceptable quality for each frame can be maintained. However, the skipped frame must be decompressed completely, which might act as a reference frame to nonskipped frames for reconstruction. The newly quantized discrete cosine transform (DCT) coefficients of the prediction errors need to be re-computed for the nonskipped frame with reference to the previous nonskipped frame; this can create undesirable complexity as well as introduce re-encoding errors. In this paper, we propose new algorithms and a novel architecture for frame-rate reduction to improve picture quality and to reduce complexity. The proposed architecture is mainly performed on the DCT domain to achieve a transcoder with low complexity. With the direct addition of DCT coefficients and an error compensation feedback loop, re-encoding errors are reduced significantly. Furthermore, we propose a frame-rate control scheme which can dynamically adjust the number of skipped frames according to the incoming motion vectors and re-encoding errors due to transcoding such that the decoded sequence can have a smooth motion as well as better transcoded pictures. Experimental results show that, as compared to the conventional transcoder, the new architecture for frame-skipping transcoder is more robust, produces fewer requantization errors, and has reduced computational complexity.
    BibTeX:
    @article{2002_Fung,
      author = {Kai-Tat Fung and Yui-Lam Chan and Wan-Chi Siu},
      title = {New architecture for dynamic frame-skipping transcoder},
      journal = {Image Processing, IEEE Transactions on},
      year = {2002},
      volume = {11},
      number = {8},
      pages = { 886 - 900},
      doi = {http://dx.doi.org/10.1109/TIP.2002.800890}
    }
    					
    Hsiao, S.-F. & Tseng, J.-M. New matrix formulation for two-dimensional DCT/IDCT computation and its distributed-memory VLSI implementation 2002 Vision, Image and Signal Processing, IEE Proceedings -
    Vol. 149 (2) , pp. 97 - 107  
    article    
    Abstract: A direct method for the computation of 2-D DCT/IDCT on a linear-array architecture is presented. The 2-D DCT/IDCT is first converted into its corresponding I-D DCT/IDCT problem through proper input/output index reordering. Then, a new coefficient matrix factorisation is derived, leading to a cascade of several basic computation blocks. Unlike other previously proposed high-speed 2-D N times; N DCT/IDCT processors that usually require intermediate transpose memory and have computation complexity O(N3), the proposed hardware-efficient architecture with distributed memory structure has computation complexity O(N2 log2 N) and requires only log2 N multipliers. The new pipelinable and scalable 2-D DCT/IDCT processor uses storage elements local to the processing elements and thus does not require any address generation hardware or global memory-to-array routing.
    BibTeX:
    @article{2002_Hsiao,
      author = {Hsiao, S.-F. and Tseng, J.-M.},
      title = {New matrix formulation for two-dimensional DCT/IDCT computation and its distributed-memory VLSI implementation},
      journal = {Vision, Image and Signal Processing, IEE Proceedings -},
      year = {2002},
      volume = {149},
      number = {2},
      pages = { 97 - 107}
    }
    					
    Li, H.; Liu, G. & Li, Y. An effective approach to edge classification from DCT domain 2002 Image Processing. 2002. Proceedings. 2002 International Conference on
    Vol. 1 , pp. I-940 - I-943 vol.1  
    inproceedings DOI    
    Abstract: In the field of content-based visual information analysis, the detection of visual features is a significant topic. In order to process video data efficiently, visual features extraction is required. Many advanced video applications require direct manipulation of compressed video data. An effective approach that detects edges in MPEG compressed images is proposed. First, DCT coefficients of 8 times;8 subblocks are analyzed on their meaning in determining boundaries. Then, we consider two types of ideal linear edges cutting through a block of size 8 times;8. Based on these edge models, we derive an edge detection approach from ten normalized DCT coefficients, obtaining the general rules of edge classification. Finally, we test the proposed algorithm on different images, and compare our method with other edge classification approaches. Simulations show that our approach can be used to estimate the edge information of images from their DCT coefficients more effectively than those proposed previously.
    BibTeX:
    @inproceedings{2002_Li,
      author = {Hongliang Li and Guizhong Liu and Yongli Li},
      title = {An effective approach to edge classification from DCT domain},
      journal = {Image Processing. 2002. Proceedings. 2002 International Conference on},
      year = {2002},
      volume = {1},
      pages = { I-940 - I-943 vol.1},
      doi = {http://dx.doi.org/10.1109/ICIP.2002.1038182}
    }
    					
    Wahyudi, A. & Omondi, A. Parallel multimedia processor using customised Infineon TriCores 2002 Digital System Design, 2002. Proceedings. Euromicro Symposium on , pp. 140 - 147   inproceedings DOI    
    Abstract: This paper reports on our experiments on using the Infineon TriCore as a building block for a multimedia processor. The experiments aim to obtain a high performance processor using two strategies: integrating multimedia units into the TriCore CPU and constructing the TriCore in multiprocessor configuration. The design and implementation of the multimedia units for video, audio, and text compressions are discussed. Two hardware architectures for IMA ADPCM audio compression multimedia unit were designed: direct architecture and sequential architecture. The multimedia unit for text compression is based on a modification from another design; our design uses a more efficient timing operation and has a better hardware utilization than the original design. Two algorithms for parallel motion-estimation were implemented on the multiple TriCore system. The results show that the TriCore is a good building block for a multiprocessor system.
    BibTeX:
    @inproceedings{2002_Wahyudi,
      author = {Wahyudi, A. and Omondi, A.},
      title = {Parallel multimedia processor using customised Infineon TriCores},
      journal = {Digital System Design, 2002. Proceedings. Euromicro Symposium on},
      year = {2002},
      pages = { 140 - 147},
      doi = {http://dx.doi.org/10.1109/DSD.2002.1115362}
    }
    					
    Calvagno, G.; Mian, G.; Rinaldo, R. & Trabucco, W. Two-dimensional separable filters for optimal reconstruction of JPEG-coded images 2001 Circuits and Systems for Video Technology, IEEE Transactions on
    Vol. 11 (7) , pp. 777 -787  
    article DOI    
    Abstract: Transform coding is a technique used worldwide for image coding, and JPEG has become the most common tool for image compression. In a JPEG decoder, the quantized transform coefficient blocks are usually processed using the inverse discrete cosine transform (DCT) in order to reconstruct an approximation of the original image. The direct and inverse DCT pair can be arranged in the form of a perfect reconstruction filter bank, and it can be shown that, in the presence of quantization of the transform coefficients, the perfect reconstruction synthesis is not the best choice. In this paper, we propose a procedure for the design of separable 2-D synthesis filters that minimize the reconstruction error power for transform coders. The procedure is used to design a family of filters which are used in the decoder instead of the inverse DCT. The appropriate reconstruction filters are selected on the basis of the standard quantization information provided in the JPEG bit stream. We show that the proposed decoding method gives some gain with respect to the usual decoder in most cases, Moreover, it only makes use of the standard information provided by a JPEG bit stream
    BibTeX:
    @article{2001_Calvagno,
      author = {Calvagno, G. and Mian, G.A. and Rinaldo, R. and Trabucco, W.},
      title = {Two-dimensional separable filters for optimal reconstruction of JPEG-coded images},
      journal = {Circuits and Systems for Video Technology, IEEE Transactions on},
      year = {2001},
      volume = {11},
      number = {7},
      pages = {777 -787},
      doi = {http://dx.doi.org/10.1109/76.931106}
    }
    					
    Fung, K.-T.; Chan, Y.-L. & Siu, W.-C. Dynamic frame skipping for high-performance transcoding 2001 Image Processing, 2001. Proceedings. 2001 International Conference on
    Vol. 1 , pp. 425 -428 vol.1  
    inproceedings DOI    
    Abstract: Transcoding is a process of converting a previously compressed video bitstream into a lower bit-rate bitstream. When some incoming frames are dropped for the frame-rate conversion in transcoding, the newly quantized DCT coefficients of prediction error need to be re-computed, which can create an undesirable complexity as well as introduce re-encoding error. We propose a new architecture for a frame-skipping transcoder to improve picture quality and to reduce complexity. It is observed that re-encoding error is reduced significantly when the strategy of direct summation of DCT coefficients and the error compensation feedback loop are employed. Furthermore, we propose a frame-rate control scheme which can dynamically adjust the number of skipped frames according to the incoming motion vectors and the re-encoding error due to transcoding such that the decoded sequence can have smooth motion as well as better transcoded pictures. Experimental results show that, as compared to the conventional transcoder, the new frame-skipping transcoder is more robust, produces smaller requantization errors, and has simple computational complexity
    BibTeX:
    @inproceedings{2001_Fung,
      author = {Kai-Tat Fung and Yui-Lam Chan and Wan-Chi Siu},
      title = {Dynamic frame skipping for high-performance transcoding},
      journal = {Image Processing, 2001. Proceedings. 2001 International Conference on},
      year = {2001},
      volume = {1},
      pages = {425 -428 vol.1},
      doi = {http://dx.doi.org/10.1109/ICIP.2001.959044}
    }
    					
    Jian, B.L.; Xuan, Z.; Rong, T.J. & Yue, L. An efficient VLSI architecture for 2D-DCT using direct method 2001 ASIC, 2001. Proceedings. 4th International Conference on , pp. 393 -396   inproceedings DOI    
    Abstract: An efficient VLSI architecture for 8 times;8 two-dimensional (2D) discrete cosine transform (DCT) is proposed in this paper. It is a folded architecture using direct method. It can compute 2D-DCT of a 12-b 8 times;8 block using one 1D-DCT unit without transpose memory. Taking advantage of the direct method, the total number of multiplications in the proposed architecture is only half of that required for row-column method. It, in turn, results in the doubled operating speed compared with those conventional implementations with row-column method. Under 0.6 mu;m CMOS and double metal technology, the proposed architecture presents a chip with core size 3.9 times;0.9 mm2, transistor count 114 K and clock rate 200 MHz
    BibTeX:
    @inproceedings{2001_Jian,
      author = {Bian Li Jian and Zeng Xuan and Tong Jia Rong and Liu Yue},
      title = {An efficient VLSI architecture for 2D-DCT using direct method},
      journal = {ASIC, 2001. Proceedings. 4th International Conference on},
      year = {2001},
      pages = {393 -396},
      doi = {http://dx.doi.org/10.1109/ICASIC.2001.982583}
    }
    					
    Kim, D.; Yoon, B. & Choe, Y. Direct converting DV into MPEG-2 intra coding 2001 Consumer Electronics, IEEE Transactions on
    Vol. 47 (4) , pp. 941 -945  
    article DOI    
    Abstract: Converting the digital video (DV) format for digital video cassette recorders into the MPEG-2 intra coding is performed in the DCT domain to reduce conversion steps. To convert the 4:1:1 to the 4:2:2 chroma. formats and the 2-4-8 DCT mode to the 8-8 DCT mode, a matrix is multiplied by transformed data. Such multiplication enables parallel processing, and the m lowbar;quant of the MPEG-2 TM5 rate control is computed in the DCT domain
    BibTeX:
    @article{2001_Kim,
      author = {Donyeon Kim and Bumsik Yoon and Yoonsik Choe},
      title = {Direct converting DV into MPEG-2 intra coding},
      journal = {Consumer Electronics, IEEE Transactions on},
      year = {2001},
      volume = {47},
      number = {4},
      pages = {941 -945},
      doi = {http://dx.doi.org/10.1109/30.982812}
    }
    					
    Robertson, M. & Stevenson, R. Reduced-complexity iterative post-filtering of video 2001 Circuits and Systems for Video Technology, IEEE Transactions on
    Vol. 11 (10) , pp. 1121 -1127  
    article DOI    
    Abstract: There are numerous methods of post-processing that make use of iterative techniques. Many of these schemes have been demonstrated to be very effective in removing artifacts from compressed video, producing better and better image estimates at each iteration. However, this artifact removal comes at the cost of a large computational burden. This paper introduces two methods for iterative post-processing of compressed video in an efficient manner. One of these methods is applicable to one particular maximum a posteriori scheme. The other method has direct application to other, more general, iterative post-processing schemes that make use of a convex constraint set, which is the set of all images that will recompress to yield the originally-received data. The combination of these two methods produces a post-processing algorithm that has the advantage of many iterative schemes (excellent visual results), while requiring a relatively low amount of computational effort to achieve these results
    BibTeX:
    @article{2001_Robertson,
      author = {Robertson, M.A. and Stevenson, R.L.},
      title = {Reduced-complexity iterative post-filtering of video},
      journal = {Circuits and Systems for Video Technology, IEEE Transactions on},
      year = {2001},
      volume = {11},
      number = {10},
      pages = {1121 -1127},
      doi = {http://dx.doi.org/10.1109/76.954498}
    }
    					
    Yang, X.; Shi, S. & Wong, A. Tradeoffs in modified discrete cosine transform implementations 2001 ASIC, 2001. Proceedings. 4th International Conference on , pp. 370 -373   inproceedings DOI    
    Abstract: The performance dependence of modified discrete cosine transform (MDCT) on hardware architecture is investigated. The oddly stacked architecture is found to be superior to direct computation in terms of accuracy, power consumption, and circuit area
    BibTeX:
    @inproceedings{2001_Yang,
      author = {Xin Yang and ShiChang Shi and Wong, A.K.},
      title = {Tradeoffs in modified discrete cosine transform implementations},
      journal = {ASIC, 2001. Proceedings. 4th International Conference on},
      year = {2001},
      pages = {370 -373},
      doi = {http://dx.doi.org/10.1109/ICASIC.2001.982577}
    }
    					
    Britanak, V. & Rao, R. Two-dimensional DCT/DST universal computational structure for 2m times;2n block sizes 2000 Signal Processing, IEEE Transactions on
    Vol. 48 (11) , pp. 3250 -3255  
    article DOI    
    Abstract: A refined generalized signal flow graph for the direct two-dimensional discrete cosine transform (2-D DCT) and discrete sine transform (2-D DST) computation (the so-called 2-D DCT/DST universal computational structure) is described. It represents a generalized unified approach to the fast 2-D DCT and 2-D DST computation for any 2 m times;2n block sizes, i.e., both square and rectangular blocks, including the one-dimensional (1-D) case. The regular structure, moderate arithmetic complexity, numerical stability, and multiple block size capability makes it suitable for VLSI or parallel implementation
    BibTeX:
    @article{2000_Britanak,
      author = {Britanak, V. and Rao, R.},
      title = {Two-dimensional DCT/DST universal computational structure for 2m times;2n block sizes},
      journal = {Signal Processing, IEEE Transactions on},
      year = {2000},
      volume = {48},
      number = {11},
      pages = {3250 -3255},
      doi = {http://dx.doi.org/10.1109/78.875483}
    }
    					
    Chang, T.-S.; Kung, C.-S. & Jen, C.-W. A simple processor core design for DCT/IDCT 2000 Circuits and Systems for Video Technology, IEEE Transactions on
    Vol. 10 (3) , pp. 439 -447  
    article DOI    
    Abstract: This paper presents a cost-effective processor core design that features the simplest hardware and is suitable for discrete cosine transform/indiscrete cosine transform (DCT/IDCT) operations in H.263 and digital camera. This design combines the techniques of fast direct two-dimensional DCT algorithm, the bit level adder-based distributed arithmetic, and common subexpression sharing to reduce the hardware cost and enhance the computing speed. The resulting architecture is very simple and regular such that it can be easily scaled for higher throughput rate requirements. The DCT design has been implemented by 0.6 mu;m SPDM CMOS technology and only costs 1493 gate count, or 0.78 mm 2. The proposed design can meet real-time DCT/IDCT requirements of the H.263 codec system for QCIF image frame size at 10 frames/s with 4:2:0 color format. Moreover, the proposed design still possesses additional computing power for other operations when operating at 33 MHz
    BibTeX:
    @article{2000_Chang,
      author = {Tian-Sheuan Chang and Chin-Sheng Kung and Chein-Wei Jen},
      title = {A simple processor core design for DCT/IDCT},
      journal = {Circuits and Systems for Video Technology, IEEE Transactions on},
      year = {2000},
      volume = {10},
      number = {3},
      pages = {439 -447},
      doi = {http://dx.doi.org/10.1109/76.836290}
    }
    					
    Eng, H.-L. & Ma, K.-K. Unsupervised image object segmentation over compressed domain 2000 Image Processing, 2000. Proceedings. 2000 International Conference on
    Vol. 3 , pp. 758 -761 vol.3  
    inproceedings DOI    
    Abstract: Direct processing of JPEG images based on DCT coefficients could avoid computationally intensive full decoding and large memory storage. In this paper, we exploit the inherent information extracted from DCT coefficients to achieve unsupervised segmentation of image objects. First, a maximum entropy fuzzy clustering (MEFC) algorithm is proposed to achieve a coarse segmentation based on DCT-DC coefficients. The DCT-AC coefficients are then utilized to refine the segmentation boundary by a maximum a posteriori (MAP) approach. The major challenge of the problem is to achieve satisfactory segmentation simply based on DCT coefficients, which are quantized and coarse information in essence. Experimental results show the promising potential of the proposed algorithm in overcoming these fundamental limitations
    BibTeX:
    @inproceedings{2000_Eng,
      author = {How-Lung Eng and Kai-Kuang Ma},
      title = {Unsupervised image object segmentation over compressed domain},
      journal = {Image Processing, 2000. Proceedings. 2000 International Conference on},
      year = {2000},
      volume = {3},
      pages = {758 -761 vol.3},
      doi = {http://dx.doi.org/10.1109/ICIP.2000.899565}
    }
    					
    Hanzo, L.; Wong, C. & Cherriman, P. Burst-by-burst adaptive wideband wireless video telephony 2000 Communications and Vehicular Technology, 2000. SCVT-200. Symposium on , pp. 215 -232   inproceedings DOI    
    Abstract: The design trade-offs of interactive wireless video systems are discussed and performance comparisons are provided both in the context of second- and third-generation wireless videophone systems. We commence our discussions by a comparative study of arbitrarily programmable, but fixed-rate, videophone codecs using quarter common intermediate format (QCIF) video sequences scanned at 10 frames/s. These proprietary codecs were designed to allow direct replacement of mobile radio voice codecs in second generation wireless systems, such as the Pan-European GSM, the American IS-54 and IS-95 as well as the Japanese systems, operating at 13, 8, 9.6 and 6.7 kbps, respectively, although better video quality is maintained over higher-rate, 32 kbps cordless systems, such as the Japanese PHS and the European DECT and CT2 systems. From the range of codecs investigated, best overall performance was achieved by our vector-quantised codecs, followed by the discrete cosine transformed and the quadtree-based schemes. The associated video peak signal-to-noise ratio (PSNR) was around 30 dB, while the subjective video quality can be assessed under http://www-mobile.ecs.soton.ac.uk. A range of multimode wireless transceivers is also proposed. The second part of the paper is dedicated to burst-by-burst (BbB) adaptive wireless video transceivers employing the standard H.263 codec. It is demonstrated that the proposed BbB adaptive transceivers provide an improved video performance in comparison to their statically reconfigured counterparts in the context of both wideband BbB adaptive quadrature amplitude modulation (AQAM) transceivers and the joint-detection based code division multiple access (CDMA) transceivers of the third generation systems
    BibTeX:
    @inproceedings{2000_Hanzo,
      author = {Hanzo, L. and Wong, C.H. and Cherriman, P.},
      title = {Burst-by-burst adaptive wideband wireless video telephony},
      journal = {Communications and Vehicular Technology, 2000. SCVT-200. Symposium on},
      year = {2000},
      pages = {215 -232},
      doi = {http://dx.doi.org/10.1109/SCVT.2000.923365}
    }
    					
    Henttu, P.; Pouttu, A. & Raustia, M. Performance of PIE interference suppressor using FFT, DCT and ELT transformations in FH/DS communications 2000 EUROCOMM 2000. Information Systems for Enhanced Public Safety and Security. IEEE/AFCEA , pp. 126 -130   inproceedings DOI    
    Abstract: We have studied the BER performance of a hybrid direct sequence frequency hopping (DS/FH) system with interference suppressor in an environment where a strong narrowband interference perturbs the desired DS signal. In this approach a transform domain, real time suppression method PIE (phase interference extractor) has been chosen. The original PIE utilized FFT as the transformation. In this study we compared the performance and computational complexity of the original FFT PIE to DCT PIE and ELT PIE
    BibTeX:
    @inproceedings{2000_Henttu,
      author = {Henttu, P. and Pouttu, A. and Raustia, M.},
      title = {Performance of PIE interference suppressor using FFT, DCT and ELT transformations in FH/DS communications},
      journal = {EUROCOMM 2000. Information Systems for Enhanced Public Safety and Security. IEEE/AFCEA},
      year = {2000},
      pages = {126 -130},
      doi = {http://dx.doi.org/10.1109/EURCOM.2000.874786}
    }
    					
    Hu, P.-C.; Kavesh, M. & Zhang, Z.-L. A wavelet to DCT progressive image transcoder 2000 Image Processing, 2000. Proceedings. 2000 International Conference on
    Vol. 1 , pp. 968 -971 vol.1  
    inproceedings DOI    
    Abstract: A transcoder design is proposed in which the transcoder loads the pre-encoded embedded wavelet coefficients and computes the DCT coefficients. The resulting DCT coefficients are quantized and sorted by multi-grid embedded coding to output a compressed DCT bitstream. This transcoder has two operational modes: open-loop and closed-loop. The open-loop scheme is designed for direct transcoding, targeted to applications in which receivers are connected to the transcoder through a fast network and demand the same rate. In contrast, the closed-loop scheme is designed for fully progressive transcoding, targeted to receivers which are connected to the transcoder through a slow network connection, and may demand heterogeneous rates. The performance of the transcoder is evaluated in both open-loop mode and close-loop mode
    BibTeX:
    @inproceedings{2000_Hu,
      author = {Po-Chin Hu and Kavesh, M. and Zhi-Li Zhang},
      title = {A wavelet to DCT progressive image transcoder},
      journal = {Image Processing, 2000. Proceedings. 2000 International Conference on},
      year = {2000},
      volume = {1},
      pages = {968 -971 vol.1},
      doi = {http://dx.doi.org/10.1109/ICIP.2000.901122}
    }
    					
    Kim, Y.-M.; Choi, S.W. & Lee, S.-W. Fast scene change detection using direct feature extraction from MPEG compressed videos 2000 Pattern Recognition, 2000. Proceedings. 15th International Conference on
    Vol. 3 , pp. 174 -177 vol.3  
    inproceedings DOI    
    Abstract: In order to process video data efficiently, a video segmentation technique through scene change detection must be employed. Many of advanced video applications require manipulations of compressed video signals. So, the scene change detection process is achieved by analyzing the video directly in the compressed domain, thereby avoiding the overhead of decompressing video into individual frames in the pixel domain. In this paper, we propose a fast scene change detection algorithm using direct feature extraction from MPEG compressed videos, and evaluate this technique using sample video data. This process was made possible by a new mathematical formulation for deriving the edge information directly from the discrete cosine transform coefficients
    BibTeX:
    @inproceedings{2000_Kim,
      author = {Young-Min Kim and Sung Woo Choi and Seong-Whan Lee},
      title = {Fast scene change detection using direct feature extraction from MPEG compressed videos},
      journal = {Pattern Recognition, 2000. Proceedings. 15th International Conference on},
      year = {2000},
      volume = {3},
      pages = {174 -177 vol.3},
      doi = {http://dx.doi.org/10.1109/ICPR.2000.903513}
    }
    					
    Knockaert, L. Fast Hankel transform by fast sine and cosine transforms: the Mellin connection 2000 Signal Processing, IEEE Transactions on
    Vol. 48 (6) , pp. 1695 -1701  
    article DOI    
    Abstract: The Hankel transform of a function by means of a direct Mellin approach requires sampling on an exponential grid, which has the disadvantage of coarsely undersampling the tail of the function. A novel modified Hankel transform procedure that does not require exponential sampling is presented. The algorithm proceeds via a three-step Mellin approach to yield a decomposition of the Hankel transform into a sine, a cosine, and an inversion transform, which can be implemented by means of fast sine and cosine transforms
    BibTeX:
    @article{2000_Knockaert,
      author = {Knockaert, L.},
      title = {Fast Hankel transform by fast sine and cosine transforms: the Mellin connection},
      journal = {Signal Processing, IEEE Transactions on},
      year = {2000},
      volume = {48},
      number = {6},
      pages = {1695 -1701},
      doi = {http://dx.doi.org/10.1109/78.845927}
    }
    					
    Lee, S.-W.; Kim, Y.-M. & Choi, S.W. Fast scene change detection using direct feature extraction from MPEG compressed videos 2000 Multimedia, IEEE Transactions on
    Vol. 2 (4) , pp. 240 -254  
    article DOI    
    Abstract: In order to process video data efficiently, a video segmentation technique through scene change detection must be required. This is a fundamental operation used in many digital video applications such as digital libraries, video on demand (VOD), etc. Many of these advanced video applications require manipulations of compressed video signals. So, the scene change detection process is achieved by analyzing the video directly in the compressed domain, thereby avoiding the overhead of decompressing video into individual frames in the pixel domain. In this paper, we propose a fast scene change detection algorithm using direct feature extraction from MPEG compressed videos, and evaluate this technique using sample video data, First, we derive binary edge maps from the AC coefficients in blocks which were discrete cosine transformed. Second, we measure edge orientation, strength and offset using correlation between the AC coefficients in the derived binary edge maps. Finally, we match two consecutive frames using these two features (edge orientation and strength). This process was made possible by a new mathematical formulation for deriving the edge information directly from the discrete cosine transform (DCT) coefficients. We have shown that the proposed algorithm is faster or more accurate than the previously known scene change detection algorithms
    BibTeX:
    @article{2000_Lee,
      author = {Seong-Whan Lee and Young-Min Kim and Sung Woo Choi},
      title = {Fast scene change detection using direct feature extraction from MPEG compressed videos},
      journal = {Multimedia, IEEE Transactions on},
      year = {2000},
      volume = {2},
      number = {4},
      pages = {240 -254},
      doi = {http://dx.doi.org/10.1109/6046.890059}
    }
    					
    Salgado, L.; Menendez, J.; Rendon, E.; Garcia, N. & Larrosa, R. Efficient prediction error regions determination for region-based video coding through shape adaptive DCT 2000 Image Processing, 2000. Proceedings. 2000 International Conference on
    Vol. 1 , pp. 996 -999 vol.1  
    inproceedings DOI    
    Abstract: An efficient strategy to determine the prediction error regions to be coded within a region-based prediction error coding scheme is presented. Prediction error coding is based on the segmentation of the displaced field difference (DFD) and coding the resulting arbitrary shaped DFD regions using shape adaptive DCT. Efficiency in the determination of the DFD regions to be coded is achieved by eliminating from the selection process the direct computation of the cost of region contours and textures coding. With this scheme, perceptual distortion of the decoded images is reduced while quality is locally improved on relevant image areas. Comparative results with the complete H.263 coder are shown
    BibTeX:
    @inproceedings{2000_Salgado,
      author = {Salgado, L. and Menendez, J.M. and Rendon, E. and Garcia, N. and Larrosa, R.},
      title = {Efficient prediction error regions determination for region-based video coding through shape adaptive DCT},
      journal = {Image Processing, 2000. Proceedings. 2000 International Conference on},
      year = {2000},
      volume = {1},
      pages = {996 -999 vol.1},
      doi = {http://dx.doi.org/10.1109/ICIP.2000.901129}
    }
    					
    Stylianou, Y. A simple and fast way of generating a harmonic signal 2000 Signal Processing Letters, IEEE
    Vol. 7 (5) , pp. 111 -113  
    article DOI    
    Abstract: Harmonic models are widely used for text-to-speech (TTS) systems based on a concatenation of acoustic units. The fast generation of a harmonic signal is an important issue in reducing the complexity of TTS systems based on these models. In this letter, we propose a novel method of generating a harmonic signal based on delayed multi-resampled cosine functions (DMRCs). The DMRC method is compared with the direct (straightforward) synthesis method, SF, the use of the inverse fast Fourier transform, and synthesis using recurrence relations for trigonometric functions. DMRC was shown to outperform all the other techniques, reducing the complexity of the SF method by 95%
    BibTeX:
    @article{2000_Stylianou,
      author = {Stylianou, Y.},
      title = {A simple and fast way of generating a harmonic signal},
      journal = {Signal Processing Letters, IEEE},
      year = {2000},
      volume = {7},
      number = {5},
      pages = {111 -113},
      doi = {http://dx.doi.org/10.1109/97.841155}
    }
    					
    Xi, J. & Chicharo, J. Computing running DCTs and DSTs based on their second-order shift properties 2000 Circuits and Systems I: Fundamental Theory and Applications, IEEE Transactions on
    Vol. 47 (5) , pp. 779 -783  
    article DOI    
    Abstract: This paper presents a set of second-order recursive equations which are referred to as the second-order shift (SOS) properties of the discrete cosine transform (DCT) and the discrete sine transform (DST). The proposed SOS properties enable independent updating of the respective DCT and DST coefficients. This is in direct contrast with existing methodology for computing the running DCT and DST where there is an inherent interdependency between the DCT and DST coefficients. The SOS properties provide more efficient algorithms in terms of computational burden and memory requirements when implementing running DCTs and DSTs
    BibTeX:
    @article{2000_Xi,
      author = {Jiangtao Xi and Chicharo, J.F.},
      title = {Computing running DCTs and DSTs based on their second-order shift properties},
      journal = {Circuits and Systems I: Fundamental Theory and Applications, IEEE Transactions on},
      year = {2000},
      volume = {47},
      number = {5},
      pages = {779 -783},
      doi = {http://dx.doi.org/10.1109/81.847888}
    }
    					
    Chen, T.-H. A cost-effective 8 times;8 2-D IDCT core processor with folded architecture 1999 Consumer Electronics, IEEE Transactions on
    Vol. 45 (2) , pp. 333 -339  
    article DOI    
    Abstract: A dedicated cost-effective core processor of the 8 times;8 two-dimensional (2-D) inverse discrete transform (IDCT) architecture based on the direct realization approach is proposed. The folding scheme is developed to obtain a low gate-count and high throughput. The experimental result shows that the chip's throughput is one pixel per clock cycle with a structure of 78 K transistors, which reveals that the low cost of VLSI implementation is more attractive than most of previously reported chips. With 0.6 mu;m CMOS, double metal technology, the chip is a standard-cell implementation and requires a core size of 4.4 times;2.8 mm2, and is able to operate at a clock rate of more than 100 MHz
    BibTeX:
    @article{1999_Chen,
      author = {Thou-Ho Chen},
      title = {A cost-effective 8 times;8 2-D IDCT core processor with folded architecture},
      journal = {Consumer Electronics, IEEE Transactions on},
      year = {1999},
      volume = {45},
      number = {2},
      pages = {333 -339},
      doi = {http://dx.doi.org/10.1109/30.793417}
    }
    					
    Hernandez, J. & Perez-Gonzalez, F. Statistical analysis of watermarking schemes for copyright protection of images 1999 Proceedings of the IEEE
    Vol. 87 (7) , pp. 1142 -1166  
    article DOI    
    Abstract: In this paper, we address the problem of the performance analysis of image watermarking systems that do not require the availability of the original image during ownership verification. We focus on a statistical approach to obtain models that can serve as a basis for the application of decision theory to the design of efficient detector structures. Special attention is paid to the possible nonexistence of a statistical description of the original image. Different modeling approaches are proposed for the cases when such a statistical characterization is known and when it is not. Watermarks may encode a message, and the performance of the watermarking system is evaluated using as a measure the probability of false alarm, the probability of detection when the presence of the watermark is tested, and the probability of error when the information that it carries is extracted. Finally, the modeling techniques studied are applied to the analysis of two watermarking schemes, one of them defined in the spatial domain, and the other in the direct cosine transform (DCT) domain. The theoretical results are contrasted with empirical data obtained through experimentation covering several cases of interest. We show how choosing an appropriate statistical model for the original image can lead to considerable improvements in performance
    BibTeX:
    @article{1999_Hernandez,
      author = {Hernandez, J.R. and Perez-Gonzalez, F.},
      title = {Statistical analysis of watermarking schemes for copyright protection of images},
      journal = {Proceedings of the IEEE},
      year = {1999},
      volume = {87},
      number = {7},
      pages = {1142 -1166},
      doi = {http://dx.doi.org/10.1109/5.771069}
    }
    					
    Hsiao, S.-F. & Tseng, J.-M. Direct implementation of 2-D DCT on a low-cost linear-array architecture without intermediate transpose memory 1999 Signal Processing Systems, 1999. SiPS 99. 1999 IEEE Workshop on , pp. 90 -99   inproceedings DOI    
    Abstract: A direct method for the computation of 2-D DCT on a linear-array architecture is presented. The original 2-D DCT is converted into 1-D problem with representation of matrix-vector product. Then, we propose a fast algorithm with low computation complexity, and exploit an efficient mapping technique to generate from the algorithm a hardware-efficient architecture. Unlike other 2-D DCT processors that usually require transpose memory, our new architecture is easily pipelined for purpose of high throughput rate and is easily scalable for the computation of longer-length DCT
    BibTeX:
    @inproceedings{1999_Hsiao,
      author = {Shen-Fu Hsiao and Jian-Ming Tseng},
      title = {Direct implementation of 2-D DCT on a low-cost linear-array architecture without intermediate transpose memory},
      journal = {Signal Processing Systems, 1999. SiPS 99. 1999 IEEE Workshop on},
      year = {1999},
      pages = {90 -99},
      doi = {http://dx.doi.org/10.1109/SIPS.1999.822314}
    }
    					
    Skodras, A. Direct transform to transform computation 1999 Signal Processing Letters, IEEE
    Vol. 6 (8) , pp. 202 -204  
    article DOI    
    Abstract: An efficient direct method for the computation of a length-N discrete cosine transform (DCT) given two adjacent length-(N/2) DCT coefficients, is presented. The computational complexity of the proposed method is lower than the traditional approach for lengths N gt;8. Savings of N memory locations and 2N data transfers are also achieved
    BibTeX:
    @article{1999_Skodras,
      author = {Skodras, A.N.},
      title = {Direct transform to transform computation},
      journal = {Signal Processing Letters, IEEE},
      year = {1999},
      volume = {6},
      number = {8},
      pages = {202 -204},
      doi = {http://dx.doi.org/10.1109/97.774865}
    }
    					
    Song, J. & Yeo, B.-L. Fast extraction of spatially reduced image sequences from MPEG-2 compressed video 1999 Circuits and Systems for Video Technology, IEEE Transactions on
    Vol. 9 (7) , pp. 1100 -1114  
    article DOI    
    Abstract: MPEG-2 video standards are targeted for high-quality video broadcast and distribution and are optimized for efficient storage and transmission. However, it is difficult to process MPEG-2 for video browsing and database applications without first decompressing the video. Yeo and Liu (1995) have proposed fast algorithms for the direct extraction of spatially reduced images from MPEG-1 video. Reduced images have been demonstrated to be effective for shot detection, shot browsing and editing, and temporal processing of video for video presentation and content annotation. In this paper, we develop new tools to handle the extra complexity in MPEG-2 video for extracting spatially reduced images. In particular, we propose new classes of discrete cosine transform (DCT) domain and DCT inverse motion compensation operations for handling the interlaced modes in the different frame types of MPEG-2, and we design new and efficient algorithms for generating spatially reduced images of an MPEG-2 video. The algorithms proposed in this paper are fundamental for efficient and effective processing of MPEG-2 video
    BibTeX:
    @article{1999_Song,
      author = {Junehwa Song and Boon-Lock Yeo},
      title = {Fast extraction of spatially reduced image sequences from MPEG-2 compressed video},
      journal = {Circuits and Systems for Video Technology, IEEE Transactions on},
      year = {1999},
      volume = {9},
      number = {7},
      pages = {1100 -1114},
      doi = {http://dx.doi.org/10.1109/76.795061}
    }
    					
    BenAyed, M.; Dulau, L.; Nouel, P.; Berthournieu, Y.; Masmoudi, N.; Kadionik, P. & Kamoun, L. New design using a VHDL description for DCT based circuits 1998 Microelectronics, 1998. ICM '98. Proceedings of the Tenth International Conference on , pp. 87 -90   inproceedings DOI    
    Abstract: This paper presents a new method to implement the DCT algorithm for JPEG or MPEG compression using VHDL description in ASIC or FPGA circuits. Discrete Cosine Transform (DCT) is one of the most popular lossy techniques used today in video compression schemes. Several algorithms have been proposed to implement the DCT. Loeffler (1989) has given a new class of 1D-DCT using just 11 multiplications and 29 additions. To implement such an algorithm, one or more multipliers have to be integrated. This requires a high silicon occupation area. Arithmetic distribution is widely used for such algorithms. However, its direct implementation requires an important silicon occupation area due to the ROM size needed. This paper presents a new technique based on the fusion of distributed arithmetic with Loeffler's algorithm to over come the above problems. We present a 1D-DCT design. This circuit has been implemented in Xilinx FPGAs
    BibTeX:
    @inproceedings{1998_BenAyed,
      author = {BenAyed, M.A. and Dulau, L. and Nouel, P. and Berthournieu, Y. and Masmoudi, N. and Kadionik, P. and Kamoun, L.},
      title = {New design using a VHDL description for DCT based circuits},
      journal = {Microelectronics, 1998. ICM '98. Proceedings of the Tenth International Conference on},
      year = {1998},
      pages = {87 -90},
      doi = {http://dx.doi.org/10.1109/ICM.1998.825575}
    }
    					
    Chen, L.-G.; Jiu, J.-Y.; Chang, H.-C.; Lee, Y.-P. & Ku, C.-W. A low power 2D DCT chip design using direct 2D algorithm 1998 Design Automation Conference 1998. Proceedings of the ASP-DAC '98. Asia and South Pacific , pp. 145 -150   inproceedings DOI    
    Abstract: In this paper, a low power 8 times;8 2D DCT architecture based on direct 2D approach is proposed. The direct 2D consideration reduces computational complexity. According to this algorithm, a parallel distributed arithmetic (DA) architecture at reduced supply voltage is derived. In the real circuit implementation of the chip, a hybrid-architecture adder of low power consumption is designed, as well as a power-saving ROM and a low voltage two-port SRAM with sequential access. The resultant 2D DCT chip is realized by 0.6 mu;m single-poly double-metal technology. Critical path simulation indicates a maximum input rate of 133 MHz, and it consumes 138 mW at 100 MHz
    BibTeX:
    @inproceedings{1998_Chen,
      author = {Liang-Gee Chen and Juing-Ying Jiu and Hao-Chieh Chang and Yung-Pin Lee and Chung-Wei Ku},
      title = {A low power 2D DCT chip design using direct 2D algorithm},
      journal = {Design Automation Conference 1998. Proceedings of the ASP-DAC '98. Asia and South Pacific},
      year = {1998},
      pages = {145 -150},
      doi = {http://dx.doi.org/10.1109/ASPDAC.1998.669434}
    }
    					
    Chen, L.-G.; Jiu, J.-Y.; Chang, H.-C.; Lee, Y.-P. & Ku, C.-W. Low power 2D DCT chip design for wireless multimedia terminals 1998 Circuits and Systems, 1998. ISCAS '98. Proceedings of the 1998 IEEE International Symposium on
    Vol. 4 , pp. 41 -44 vol.4  
    inproceedings DOI    
    Abstract: In this paper, a low power 2-D DCT architecture based on direct 2-D approach is proposed. The direct 2-D consideration reduces computational complexity. According to this algorithm, a parallel distributed arithmetic (DA) architecture at reduced supply voltage is derived. In the real circuit implementation of the chip, an adder of low power consumption is designed, as well as a power-saving ROM and a low voltage two-port SRAM with sequential access. The resultant 2-D DCT chip is realized by 0.6 mu;m single-poly double-metal technology. Critical path simulation indicates a maximum input rate of 133 MHz, and it consumes 138 mW at 100 MHz
    BibTeX:
    @inproceedings{1998_Chena,
      author = {Liang-Gee Chen and Juing-Ying Jiu and Hao-Chieh Chang and Yung-Pin Lee and Chung-Wei Ku},
      title = {Low power 2D DCT chip design for wireless multimedia terminals},
      journal = {Circuits and Systems, 1998. ISCAS '98. Proceedings of the 1998 IEEE International Symposium on},
      year = {1998},
      volume = {4},
      pages = {41 -44 vol.4},
      doi = {http://dx.doi.org/10.1109/ISCAS.1998.698747}
    }
    					
    Chen, L.-G.; Jiu, J.-Y. & Chang, H.-C. Design and implementation of low-power DCT chip for portable multimedia terminals 1998 Signal Processing Systems, 1998. SIPS 98. 1998 IEEE Workshop on , pp. 85 -93   inproceedings DOI    
    Abstract: This paper describes the design and implementation of a low power 2D DCT chip for portable multimedia terminals. The chip architecture based on direct 2D approach reduces computational complexity and the power dissipation can be reduced accordingly. In the implementation of the direct 2D algorithm, a parallel distributed arithmetic (DA) architecture at reduced supply voltage is adopted. In the real circuit implementation of the chip, an adder of low power consumption is designed, as well as a power-saving ROM and a low-voltage two-port SRAM with sequential access. The resultant 2D DCT chip is realized by 0.6 mu;m single-poly double-metal technology. Critical path simulation indicates a maximum input rate of 133 MHz, and it consumes 138 mW at 100 MHz. The measured chip speed is around 100 MHz
    BibTeX:
    @inproceedings{1998_Chenb,
      author = {Chen, L.-G. and Jiu, J.-Y. and Chang, H.-C.},
      title = {Design and implementation of low-power DCT chip for portable multimedia terminals},
      journal = {Signal Processing Systems, 1998. SIPS 98. 1998 IEEE Workshop on},
      year = {1998},
      pages = {85 -93},
      doi = {http://dx.doi.org/10.1109/SIPS.1998.715771}
    }
    					
    Guo, H.; Sitton, G. & Burrus, C. The quick Fourier transform: an FFT based on symmetries 1998 Signal Processing, IEEE Transactions on
    Vol. 46 (2) , pp. 335 -341  
    article DOI    
    Abstract: This paper looks at an approach that uses symmetric properties of the basis function to remove redundancies in the calculation of the discrete Fourier transform (DFT). We develop an algorithm called the quick Fourier transform (QFT) that reduces the number of floating-point operations necessary to compute the DFT by a factor of two or four over direct methods or Goertzel's method for prime lengths. By further application of the idea to the calculation of a DFT of length-2M , we construct a new O(NlogN) algorithm, with computational complexities comparable to the Cooley-Tukey algorithm. We show that the power-of-two QFT can be implemented in terms of discrete sine and cosine transforms. The algorithm can be easily modified to compute the DFT with only a subset of either input or output points and reduces by nearly half the number of operations when the data are real
    BibTeX:
    @article{1998_Guo,
      author = {Haitao Guo and Sitton, G.A. and Burrus, C.S.},
      title = {The quick Fourier transform: an FFT based on symmetries},
      journal = {Signal Processing, IEEE Transactions on},
      year = {1998},
      volume = {46},
      number = {2},
      pages = {335 -341},
      doi = {http://dx.doi.org/10.1109/78.655419}
    }
    					
    Hong, S.-H. & Kim, S.-D. Joint video coding of MPEG-2 video programs for digital broadcasting services 1998 Broadcasting, IEEE Transactions on
    Vol. 44 (2) , pp. 153 -164  
    article DOI    
    Abstract: In digital broadcasting services such as digital satellite TV, cable TV, and digital terrestrial TV, several video programs are compressed by MPEG-2 and then simultaneously transmitted over a conventional CBR (constant bit rate) broadcasting channel. In this environment, the picture quality of the aggregated video programs should be kept as equal as possible to provide fair video services for all the viewers. We propose a joint video coding scheme for ensuring that the picture quality of all the programs are nearly the same. This objective is achieved by simultaneously controlling the video encoders to generate the VBR (variable bit rate) compressed video streams. This paper contributes in two ways. First, we propose a rate-distortion estimation method for MPEG-2 video, which enables us to predict the amount of bits and the distortion generated from an encoded picture at a given quantization step size and vice versa. The most attractive features of the proposed estimation method are its accuracy and a computational complexity low enough to be applied to real-time video coding applications. Second, this paper presents an efficient and accurate joint video coding scheme using the rate-distortion estimation results. The experimental results show that our coding scheme gives a higher and more stable picture quality and a more efficient channel utilization than an independent coding scheme that encodes each program independently
    BibTeX:
    @article{1998_Hong,
      author = {Sung-Hoon Hong and Seong-Dae Kim},
      title = {Joint video coding of MPEG-2 video programs for digital broadcasting services},
      journal = {Broadcasting, IEEE Transactions on},
      year = {1998},
      volume = {44},
      number = {2},
      pages = {153 -164},
      doi = {http://dx.doi.org/10.1109/11.713067}
    }
    					
    Shen, B.; Sethi, I. & Bhaskaran, V. DCT domain alpha blending 1998 Image Processing, 1998. ICIP 98. Proceedings. 1998 International Conference on
    Vol. 1 , pp. 857 -861 vol.1  
    inproceedings DOI    
    Abstract: For direct manipulation in the compressed domain, we develop a DCT domain convolution theorem which besides exploiting the sparseness of the DCT domain representation also exploits the orthogonality and symmetry in the DCT domain representation. These properties lead to efficient compressed domain based processing methods unlike their spatial domain counterparts, where such properties are not available. This theorem can be used in a variety of image and video editing functions when the image and video data are available only as a JPEG or MPEG bitstream. We illustrate the use of these DCT domain convolution theorems in a typical video editing application such as video bluescreen editing
    BibTeX:
    @inproceedings{1998_Shen,
      author = {Bo Shen and Sethi, I.K. and Bhaskaran, V.},
      title = {DCT domain alpha blending},
      journal = {Image Processing, 1998. ICIP 98. Proceedings. 1998 International Conference on},
      year = {1998},
      volume = {1},
      pages = {857 -861 vol.1},
      doi = {http://dx.doi.org/10.1109/ICIP.1998.723653}
    }
    					
    Shen, B.; Sethi, I. & Bhaskaran, V. DCT convolution and its application in compressed domain 1998 Circuits and Systems for Video Technology, IEEE Transactions on
    Vol. 8 (8) , pp. 947 -952  
    article DOI    
    Abstract: Conventional processing of JPEG or MPEG compressed image or video data involves first decompressing the data and applying the desired processing function, and then the processed data are recompressed for the purposes of transmission or storage. We propose an alternate processing pipeline which involves direct manipulation of the JPEG or MPEG compressed domain representation to achieve the desired spatial domain processing. For direct manipulation in the compressed domain, we develop a discrete cosine transform (DCT)-domain convolution theorem which besides exploiting the sparseness of the DCT-domain representation also exploits the orthogonality and symmetry in the DCT-domain representation. These properties lead to efficient compressed domain-based processing methods unlike their spatial domain counterparts, where such properties are not available. This theorem can be used in a variety of image and video editing functions when the image and video data are available only as a JPEG or MPEG bitstream. We illustrate the use of the DCT-domain convolution theorem in a typical video editing application such as video bluescreen editing
    BibTeX:
    @article{1998_Shena,
      author = {Bo Shen and Sethi, I.K. and Bhaskaran, V.},
      title = {DCT convolution and its application in compressed domain},
      journal = {Circuits and Systems for Video Technology, IEEE Transactions on},
      year = {1998},
      volume = {8},
      number = {8},
      pages = {947 -952},
      doi = {http://dx.doi.org/10.1109/76.736723}
    }
    					
    Turcza, P. & Zielinski, T. Fast cosine Gabor transform and its application to image compression 1998 Time-Frequency and Time-Scale Analysis, 1998. Proceedings of the IEEE-SP International Symposium on , pp. 461 -464   inproceedings DOI    
    Abstract: Fast DCT-IV implementation of previously presented novel discrete critically-sampled real Gabor transform is proposed in the paper and some aspects of its application to image compression are discussed including the choice of window functions and quantization schemes. Direct links between the new cosine Gabor transform and cosine modulated filter banks as well as cosine modulated lapped transforms are given
    BibTeX:
    @inproceedings{1998_Turcza,
      author = {Turcza, P. and Zielinski, T.P.},
      title = {Fast cosine Gabor transform and its application to image compression},
      journal = {Time-Frequency and Time-Scale Analysis, 1998. Proceedings of the IEEE-SP International Symposium on},
      year = {1998},
      pages = {461 -464},
      doi = {http://dx.doi.org/10.1109/TFSA.1998.721461}
    }
    					
    Wang, J. & Yu, S. Dynamic rate scaling of coded digital video for IVOD applications 1998 Consumer Electronics, IEEE Transactions on
    Vol. 44 (3) , pp. 743 -749  
    article DOI    
    Abstract: In this paper, a memory scalable architecture for rate scaling of pre-coded digital video is proposed. In this architecture, the anchor frames are compressed and scaled to fit the size of the memory used as the frame store, and motion compensation is performed in frequency domain. Its advantages include memory saving and memory scalability which make it possible to allocated the memory resource according to the on-line payload of the system in the VOD environment. Therefore, both high efficiency and flexibility can be obtained in respect of the utilization of memory. Another problem addressed in this paper is the rate control of the rate scaling transcoder. As a matter of fact, the content creation encoding and rate scaling can be regarded as a two-pass encoding system. In this paper, a rate control algorithm with look-ahead bit allocation and direct AC DCT bit usage profile tracking is implemented in the rate scaling transcoder based on the results of the first pass encoding
    BibTeX:
    @article{1998_Wang,
      author = {Jiansong Wang and Sile Yu},
      title = {Dynamic rate scaling of coded digital video for IVOD applications },
      journal = {Consumer Electronics, IEEE Transactions on},
      year = {1998},
      volume = {44},
      number = {3},
      pages = {743 -749},
      doi = {http://dx.doi.org/10.1109/30.713190}
    }
    					
    Yonghong, Z. & Lilun, Z. Fast algorithms for running EDCT and EDST 1998 Signal Processing Proceedings, 1998. ICSP '98. 1998 Fourth International Conference on , pp. 15 -18 vol.1   inproceedings DOI    
    Abstract: Fast algorithms for running the even discrete cosine transform (EDCT) and the even discrete sine transform (EDST) are presented in this paper. The computational complexity and stability are analyzed. The algorithms given here gain considerable computational savings compared with the direct algorithms, thus can effectively find applications in real-time signal processing
    BibTeX:
    @inproceedings{1998_Yonghong,
      author = {Zeng Yonghong and Zhang Lilun},
      title = {Fast algorithms for running EDCT and EDST},
      journal = {Signal Processing Proceedings, 1998. ICSP '98. 1998 Fourth International Conference on},
      year = {1998},
      pages = {15 -18 vol.1},
      doi = {http://dx.doi.org/10.1109/ICOSP.1998.770139}
    }
    					
    Chang, C.-M. & Pao, T.-L. Tomogram reconstruction via direct cosine method 1997 TENCON '97. IEEE Region 10 Annual Conference. Speech and Image Technologies for Computing and Telecommunications., Proceedings of IEEE
    Vol. 2 , pp. 691 -694 vol.2  
    inproceedings DOI    
    Abstract: This paper proposes a new algorithm, the direct cosine method (DCM), to reconstruct the tomogram. This method uses the cosine transform in a similar form as the Fourier transform, in direct Fourier method (DFM). The cosine projection-slice (CPS) theorem needed in the DCM is also derived. The CPS theorem states that the summation of the cosine transform of two projections is a slice of the cosine transform of the projected object. The real computation and good energy compaction will reduce the processing time and errors from truncation and interpolation. By applying the DCM to the reconstruction tomogram, more advanced applications are possible
    BibTeX:
    @inproceedings{1997_Chang,
      author = {Chia-Ming Chang and Tseng-Long Pao},
      title = {Tomogram reconstruction via direct cosine method},
      journal = {TENCON '97. IEEE Region 10 Annual Conference. Speech and Image Technologies for Computing and Telecommunications., Proceedings of IEEE},
      year = {1997},
      volume = {2},
      pages = {691 -694 vol.2},
      doi = {http://dx.doi.org/10.1109/TENCON.1997.648515}
    }
    					
    Kawahito, S.; Yoshida, M.; Sasaki, M.; Umehara, K.; Miyazaki, D.; Tadokoro, Y.; Murata, K.; Doushou, S. & Matsuzawa, A. A CMOS image sensor with analog two-dimensional DCT-based compression circuits for one-chip cameras 1997 Solid-State Circuits, IEEE Journal of
    Vol. 32 (12) , pp. 2030 -2041  
    article DOI    
    Abstract: This paper presents a CMOS image sensor with on-chip compression using an analog two-dimensional discrete cosine transform (2-D DCT) processor and a variable quantization level analog-to-digital converter (ADC). The analog 2-D DCT processor is essentially suitable for the on-sensor image compression, since the analog image sensor signal can be directly processed. The small and low-power nature of the analog design allows us to achieve low-power, low-cost, one-chip digital video cameras. The 8 times;8-point analog 2-D DCT processor is designed with fully differential switched-capacitor circuits to obtain sufficient precision for video compression purposes. An imager array has a dedicated eight-channel parallel readout scheme for direct encoding with the analog 2-D DCT processor. The variable level quantization after the 2-D DCT can be performed by the ADC at the same time. A prototype CMOS image sensor integrating these core circuits for compression is implemented based on triple-metal double-polysilicon 0.35- mu;m CMOS technology. Image encoding using the implemented analog 2-D DCT processor to the image captured by the sensor is successfully performed. The maximum peak signal-to-noise ratio (PSNR) is 36.7 dB
    BibTeX:
    @article{1997_Kawahito,
      author = {Kawahito, S. and Yoshida, M. and Sasaki, M. and Umehara, K. and Miyazaki, D. and Tadokoro, Y. and Murata, K. and Doushou, S. and Matsuzawa, A.},
      title = {A CMOS image sensor with analog two-dimensional DCT-based compression circuits for one-chip cameras},
      journal = {Solid-State Circuits, IEEE Journal of},
      year = {1997},
      volume = {32},
      number = {12},
      pages = {2030 -2041},
      doi = {http://dx.doi.org/10.1109/4.643661}
    }
    					
    Kok, C. Fast algorithm for computing discrete cosine transform 1997 Signal Processing, IEEE Transactions on
    Vol. 45 (3) , pp. 757 -760  
    article DOI    
    Abstract: An efficient method for computing the discrete cosine transform (DCT) is proposed. Based on direct decomposition of the DCT, the recursive properties of the DCT for an even length input sequence is derived, which is a generalization of the radix 2 DCT algorithm. Based on the recursive property, a new DCT algorithm for an even length sequence is obtained. The proposed algorithm is very structural and requires fewer computations when compared with others. The regular structure of the proposed algorithm is suitable for fast parallel algorithm and VLSI implementation
    BibTeX:
    @article{1997_Kok,
      author = {Kok, C.W.},
      title = {Fast algorithm for computing discrete cosine transform},
      journal = {Signal Processing, IEEE Transactions on},
      year = {1997},
      volume = {45},
      number = {3},
      pages = {757 -760},
      doi = {http://dx.doi.org/10.1109/78.558495}
    }
    					
    Lee, Y.-P.; Chen, T.-H.; Chen, L.-G.; Chen, M.-J. & Ku, C.-W. A cost-effective architecture for 8 times;8 two-dimensional DCT/IDCT using direct method 1997 Circuits and Systems for Video Technology, IEEE Transactions on
    Vol. 7 (3) , pp. 459 -467  
    article DOI    
    Abstract: Among the various transform techniques for image compression, the discrete cosine transform (DCT) is the most popular and effective one in practical image and video coding applications, such as high-definition television (HDTV). We develop a novel 8 times;8 two-dimensional (2-D) discrete cosine transform/inverse discrete cosine transform (DCT/IDCT) architecture based on the direct 2-D approach and the rotation technique. The computational complexity is reduced by taking advantage of the special attribute of a complex number. Both the parallel and the folded architectures are proposed. Unlike other approaches, the proposed architecture is regular and economically allowable for VLSI implementation. Compared to the row-column method, less internal wordlength is needed in order to meet the error requirement of IDCT, and the throughput of the proposed architecture can achieve two times that of the row-column method with 30% hardware increased
    BibTeX:
    @article{1997_Lee,
      author = {Yung-Pin Lee and Thou-Ho Chen and Liang-Gee Chen and Mei-Juan Chen and Chung-Wei Ku},
      title = {A cost-effective architecture for 8 times;8 two-dimensional DCT/IDCT using direct method},
      journal = {Circuits and Systems for Video Technology, IEEE Transactions on},
      year = {1997},
      volume = {7},
      number = {3},
      pages = {459 -467},
      doi = {http://dx.doi.org/10.1109/76.585925}
    }
    					
    Sanchez, M.; Lopez, J.; Plata, O. & Zapata, E. An efficient architecture for the in place fast cosine transform 1997 Application-Specific Systems, Architectures and Processors, 1997. Proceedings., IEEE International Conference on , pp. 499 -508   inproceedings DOI    
    Abstract: The cosine transform (DCT) is in the core of image encoding and compression applications. We present a new architecture to efficiently compute the fast direct and inverse cosine transform which is based on reordering the butterflies after their computation. The designed architecture exploits locality, allowing pipelining between stages and saving memory (in place). The result is an efficient architecture for high speed computation of the DCT that reduces significantly the area required to VLSI implementation
    BibTeX:
    @inproceedings{1997_Sanchez,
      author = {Sanchez, M. and Lopez, J. and Plata, O. and Zapata, E.L.},
      title = {An efficient architecture for the in place fast cosine transform },
      journal = {Application-Specific Systems, Architectures and Processors, 1997. Proceedings., IEEE International Conference on},
      year = {1997},
      pages = {499 -508},
      doi = {http://dx.doi.org/10.1109/ASAP.1997.606855}
    }
    					
    Scargall, L. & Dlay, S. A mobile videophone image codec using wavelets and classified vector quantisation for mobile radio speech channels 1997 Information Visualization, 1997. Proceedings., 1997 IEEE Conference on , pp. 267 -271   inproceedings DOI    
    Abstract: A bandwidth efficient image codec is offered as a direct replacement for mobile radio speech codecs in second generation wireless systems, such as the Pan European GSM system. The image codec is contrived for Quarter Common Intermediate Format (QCIF) videophone sequences, and uses the 2D orthogonal wavelet transform to decompose the Displaced Frame Difference (DFD), into a four band structure using Quadrature Mirror Filters (QMF). The transformed coefficients are then compressed using classified vector quantisation (CVQ) and then multiplexed to a time division multiple access (TDMA) slot and modulated using Pilot Symbol Assisted Quadrature Amplitude Modulation (PSAQAM). The proposed image codec does not suffer from blocking effects that are visually disjointed and has therefore a major advantage over current discrete cosine transform (DCT) methods. A further advantage of using wavelet based transforms is that they require less hardware and are simpler to implement on a DSP chip than Fourier based methods. The method proposed is suitable for implementation in VLSI technology
    BibTeX:
    @inproceedings{1997_Scargall,
      author = {Scargall, L.D. and Dlay, S.S.},
      title = {A mobile videophone image codec using wavelets and classified vector quantisation for mobile radio speech channels},
      journal = {Information Visualization, 1997. Proceedings., 1997 IEEE Conference on},
      year = {1997},
      pages = {267 -271},
      doi = {http://dx.doi.org/10.1109/IV.1997.626529}
    }
    					
    Angelidis, E. Frequency sampling design of 2-D quadrantally symmetric FIR filters by 2-D block decomposition technique 1996 Circuits and Systems II: Analog and Digital Signal Processing, IEEE Transactions on
    Vol. 43 (5) , pp. 401 -405  
    article DOI    
    Abstract: In this brief, two approaches-a direct and a progressive-for the frequency-sampling design of 2-D quadrantally symmetric FIR filters using efficient and recursive algorithms are presented. Allowing the samples to be distributed on sampling curves with shapes similar to the isocontours of the desired filter response, the method is suitable for the design of circular, elliptical, rectangular, or directional FIR filters. The method is based on a 2-D block decomposition technique and it is well suited to parallel computation, since almost all the operations are matrix multiplications and/or matrix-vector products. The proposed approaches can be also applied for computing 2-D inverse cosine transforms and can be extended to m-D design problems
    BibTeX:
    @article{1996_Angelidis,
      author = {Angelidis, E.},
      title = {Frequency sampling design of 2-D quadrantally symmetric FIR filters by 2-D block decomposition technique},
      journal = {Circuits and Systems II: Analog and Digital Signal Processing, IEEE Transactions on},
      year = {1996},
      volume = {43},
      number = {5},
      pages = {401 -405},
      doi = {http://dx.doi.org/10.1109/82.494391}
    }
    					
    Chan, D.-Y.; Yang, J.-F. & Chen, S.-Y. Regular implementation algorithms of time domain aliasing cancellation 1996 Vision, Image and Signal Processing, IEE Proceedings -
    Vol. 143 (6) , pp. 387 -392  
    article    
    Abstract: The authors propose two highly regular algorithms for realising the time domain aliasing cancellation (TDAC) technique. The first TDAC implementation, which is based on the fast discrete cosine transform, effectively adopts analysis and synthesis window functions in the transform structure. This implementation algorithm achieves the least computational complexity in TDAC processes. The second TDAC implementation, which extends Goertzel's concept, uses a simple selectable-fixed-coefficient second-order infinite impulse response (IIR) filter to recursively achieve multichannel audio encoding and decoding processes. With a properly selected coefficient, this recursive implementation achieves a lower round-off-error than the current fast implementations and the direct implementation in finite wordlength. In recently developed high quality consumer products, the first algorithm is suitable to be realised in digital signal processing chips and the second one will be a better choice for VLSI implementation
    BibTeX:
    @article{1996_Chan,
      author = {Chan, D.-Y. and Yang, J.-F. and Chen, S.-Y.},
      title = {Regular implementation algorithms of time domain aliasing cancellation},
      journal = {Vision, Image and Signal Processing, IEE Proceedings -},
      year = {1996},
      volume = {143},
      number = {6},
      pages = {387 -392}
    }
    					
    Hamada, T. & Matsumoto, S. WHT-based composite motion compensated NTSC interframe direct coding 1996 Communications, IEEE Transactions on
    Vol. 44 (12) , pp. 1711 -1719  
    article DOI    
    Abstract: The motion compensated interframe differential pulse code modulation (DPCM) and discrete cosine transform (DCT) hybrid (MC DCT) coding was nominated as a standard scheme for component TV signals by ISO and ITU-R. However, in cases where an NTSC composite TV signal is used such as the United States and Japan, applying the MC DCT scheme with its luminance/chrominance separating and composing process causes unavoidable quality degradation. The reason for this additional process required for MC DCT is that a composite TV signal presents a ldquo;color subcarrier phase shift problem rdquo; in which the color subcarrier phase varies between a coding block and reference block according to the motion vector. In this paper, we propose a Walsh Hadamard transform (WHT)-based composite motion compensated NTSC interframe direct coding scheme. In this scheme, phase shifts of a color subcarrier and modulated chrominance components between a coding block and reference block can be effectively compensated by a simple process of coefficient permutation and polarity changes of several pairs of WHT coefficients to which 100% of the subcarrier energy and most of the modulated chrominance component's energy are packed. In the motion compensated DCT scheme, however, the energy of the color subcarrier and modulated chrominance components are spread over too many coefficients and a pair-based coefficient handling rule is not given to solve this problem. This paper demonstrates that the proposed scheme provides higher coding performance for a composite NTSC signal than does the motion compensated DCT scheme with its luminance/chrominance separating and composing process
    BibTeX:
    @article{1996_Hamada,
      author = {Hamada, T. and Matsumoto, S.},
      title = {WHT-based composite motion compensated NTSC interframe direct coding},
      journal = {Communications, IEEE Transactions on},
      year = {1996},
      volume = {44},
      number = {12},
      pages = {1711 -1719},
      doi = {http://dx.doi.org/10.1109/26.545901}
    }
    					
    Lee, Y.-P.; Chen, L.-G.; Chen, M.-J. & Ku, C.-W. A new design and implementation of 8 times;8 2-D DCT/IDCT 1996 VLSI Signal Processing, IX, 1996., [Workshop on] , pp. 408 -417   inproceedings DOI    
    Abstract: Among various transform techniques for image compression, the discrete cosine transform (DCT) is the most popular and effective one in practical applications because it gives an almost optimal performance and can be implemented at an acceptable cost. We describe a novel 8 times;8 2-D DCT/IDCT architecture based on the direct 2-D approach and the rotation technique. The computational complexity is reduced by taking advantage of the special attribute of complex numbers. Unlike other direct approach, the proposed architecture is regular, hence, it is suitable for VLSI implementation
    BibTeX:
    @inproceedings{1996_Lee,
      author = {Yung-Pin Lee and Liang-Gee Chen and Mei-Juan Chen and Chung-Wei Ku},
      title = {A new design and implementation of 8 times;8 2-D DCT/IDCT},
      journal = {VLSI Signal Processing, IX, 1996., [Workshop on]},
      year = {1996},
      pages = {408 -417},
      doi = {http://dx.doi.org/10.1109/VLSISP.1996.558373}
    }
    					
    Nicholls, J. & Monro, D. Scalable video with background segmentation 1996 Image Processing, 1996. Proceedings., International Conference on
    Vol. 1 , pp. 529 -532 vol.2  
    inproceedings DOI    
    Abstract: Intelligent identification of background and foreground in video scenes is used for gain compression. This is implemented as a two layer object model in a software-only video compressor with a bit rate range from less than 10 kbps up to to 1.2 Mbps. Quadtree decomposition on an error metric between the input and transmitted images directs the coder towards a foreground layer of active image fragments. A rate buffering system limits the bandwidth by transmitting only the foreground blocks which most improve the image and are above some error threshold. A high fidelity background layer is identified and communicated to the decoder, which can be used to redraw background fragments as foreground objects move across them. Blocks can be coded by various methods, such as fractal transforms or truncated DCTs. This system can be implemented in RISC processors without the need for dedicated hardware. It is suitable for low bit rate applications with slowly varying backgrounds, such as personal video communications over packet networks, or closed circuit TV surveillance using fixed or wireless links
    BibTeX:
    @inproceedings{1996_Nicholls,
      author = {Nicholls, J.A. and Monro, D.M.},
      title = {Scalable video with background segmentation},
      journal = {Image Processing, 1996. Proceedings., International Conference on},
      year = {1996},
      volume = {1},
      pages = {529 -532 vol.2},
      doi = {http://dx.doi.org/10.1109/ICIP.1996.560908}
    }
    					
    Wei, X.; Shaw, M. & Varley, M. Efficient implementation of Koilpillai-Vaidyanathan pseudo quadrature mirror filter (PQMF) banks 1996 Signal Processing, IEEE Transactions on
    Vol. 44 (12) , pp. 3135 -3138  
    article DOI    
    Abstract: An efficient implementation algorithm for the Koilpillai-Vaidyanathan (see ibid., vol.41, no.1, p.82-92, 1993) pseudo quadrature mirror filter (KVPQMF) bank, which is useful in audio compression schemes, is presented. The implementation employs a polyphase system with discrete cosine transforms (DCTs). Theoretical and practical results show a typical saving in computational load of 82% over the direct implementation
    BibTeX:
    @article{1996_Wei,
      author = {Xiang Wei and Shaw, M.J. and Varley, M.R.},
      title = {Efficient implementation of Koilpillai-Vaidyanathan pseudo quadrature mirror filter (PQMF) banks},
      journal = {Signal Processing, IEEE Transactions on},
      year = {1996},
      volume = {44},
      number = {12},
      pages = {3135 -3138},
      doi = {http://dx.doi.org/10.1109/78.553487}
    }
    					
    Angelidis, E. & Diamessis, J. A fast algorithm for computing inverse cosine transforms for designing zero-phase FIR filters in frequency domain 1995 Signal Processing Letters, IEEE
    Vol. 2 (1) , pp. 13 -16  
    article DOI    
    Abstract: A new algorithm for computing inverse cosine transforms or for designing zero-phase FIR filters from nonuniform frequency samples is presented. The algorithm is simple, fast, recursive and can be used in 1-D or 2-D applications. Based on the three-term recursive relation of the Chebyshev polynomials, the cosine matrix is decomposed into LU products using parallel computations. Two alternative approaches-a direct and a progressive-suitable for serial computations are also derived. Given N samples, the direct version requires 2.5N2+O(N) flops for computing the inverse cosine transforms or for calculating the filter coefficients, whereas the progressive version needs only O(5N) flops when the next N+1th sample appears. The algorithm guarantees real results and produces accurate solutions even in cases of designing high-order 1-D or 2-D FIR filters or when the interpolation matrix is ill conditioned. It can be also used in LU-factorization and can be extended to m-D filter design
    BibTeX:
    @article{1995_Angelidis,
      author = {Angelidis, E. and Diamessis, J.E.},
      title = {A fast algorithm for computing inverse cosine transforms for designing zero-phase FIR filters in frequency domain},
      journal = {Signal Processing Letters, IEEE},
      year = {1995},
      volume = {2},
      number = {1},
      pages = {13 -16},
      doi = {http://dx.doi.org/10.1109/97.365520}
    }
    					
    Chau, L.-P. & Siu, W.-C. Direct formulation for the realization of discrete cosine transform using recursive structure 1995 Circuits and Systems II: Analog and Digital Signal Processing, IEEE Transactions on
    Vol. 42 (1) , pp. 50 -52  
    article DOI    
    Abstract: Effective formulations for the conversion of the Discrete Fourier Transform (DFT) into recursive structure are available and have been found very effective for realization using software, hardware, and VLSI techniques. Little research work has been reported on an effective way to convert the Discrete Cosine Transform (DCT) into recursive form and the related realization. In this paper, we propose a new method to convert a prime length DCT into a recursive structure. A trivial approach is to use a conventional approach to convert the DCT into DFT and to apply Goertzel's algorithm for the rest of the realization. However, this method is inefficient and requires the realization of long length DFT's. In our approach, we suggest using some suitable mappings to convert a prime length DCT into two suitable transforms with approximately half of the original length to effect fast realization. The number of operations is greatly reduced and the structure is extremely regular
    BibTeX:
    @article{1995_Chau,
      author = {Lap-Pui Chau and Wan-Chi Siu},
      title = {Direct formulation for the realization of discrete cosine transform using recursive structure},
      journal = {Circuits and Systems II: Analog and Digital Signal Processing, IEEE Transactions on},
      year = {1995},
      volume = {42},
      number = {1},
      pages = {50 -52},
      doi = {http://dx.doi.org/10.1109/82.363541}
    }
    					
    Deng, G. & Cahill, L. Isotropic quadratic filter design using the discrete cosine transform 1995 Circuits and Systems, 1995. ISCAS '95., 1995 IEEE International Symposium on
    Vol. 2 , pp. 873 -876 vol.2  
    inproceedings DOI    
    Abstract: The problems associated with applying the discrete cosine transform (DCT) to the design and implementation of the 1-D isotropic quadratic filter are investigated in this paper. A DCT implementation of the filter is proposed and its computational complexity is analysed. The advantage of the DCT implementation is illustrated by a nonlinear system modelling problem. Results show that the DCT implementation converges faster than the direct implementation
    BibTeX:
    @inproceedings{1995_Deng,
      author = {Deng, G. and Cahill, L.W.},
      title = {Isotropic quadratic filter design using the discrete cosine transform},
      journal = {Circuits and Systems, 1995. ISCAS '95., 1995 IEEE International Symposium on},
      year = {1995},
      volume = {2},
      pages = {873 -876 vol.2},
      doi = {http://dx.doi.org/10.1109/ISCAS.1995.519903}
    }
    					
    Neogi, R. Embedded real-time video decompression algorithm and architecture for HDTV applications 1995 Algorithms and Architectures for Parallel Processing, 1995. ICAPP 95. IEEE First ICA/sup 3/PP., IEEE First International Conference on
    Vol. 1 , pp. 414 -421 vol.1  
    inproceedings DOI    
    Abstract: DCT/IDCT bared source coding and decoding techniques are widely accepted in HDTV systems and other MPEG based applications. In this paper, we propose a new direct 2-D IDCT algorithm bared on the parallel divide-and-conquer approach. The algorithm distributes computation by considering one transformed coefficient at a time and doing partial computation and updating as every coefficient arrives. A novel parallel and fully pipelined architecture with an effective processing time of one cycle per pixel for an N times;N size block is designed to implement the algorithm. An unique feature of this architecture is that it integrates inverse-shuffling, inverse-quantization, inverse-source-coding, and motion-compensation into a single compact data-path. We avoid the insertion of a FIFO between the bit-stream decoder and decompression engine. The entire block of pixel values are sampled in a single cycle for post-processing after de-compression. Also, we use only (N/2(N/2+1))/2 multipliers and N2 adders
    BibTeX:
    @inproceedings{1995_Neogi,
      author = {Neogi, R.},
      title = {Embedded real-time video decompression algorithm and architecture for HDTV applications},
      journal = {Algorithms and Architectures for Parallel Processing, 1995. ICAPP 95. IEEE First ICA/sup 3/PP., IEEE First International Conference on},
      year = {1995},
      volume = {1},
      pages = {414 -421 vol.1},
      doi = {http://dx.doi.org/10.1109/ICAPP.1995.472212}
    }
    					
    Neogi, R. & Saha, A. Embedded parallel divide-and-conquer video decompression algorithm and architecture for HDTV applications 1995 Consumer Electronics, IEEE Transactions on
    Vol. 41 (1) , pp. 160 -171  
    article DOI    
    Abstract: DCT/IDCT based source coding and decoding techniques are widely accepted in HDTV systems and other MPEG based applications. We propose a new direct 2-D IDCT algorithm based on the parallel divide-and-conquer approach. The algorithm distributes computation by considering one transformed coefficient at a time and doing partial computation and updating as every coefficient arrives. A novel parallel and fully pipelined architecture with an effective processing time of one cycle per pixel for an N times;N size block is designed to implement the algorithm. An unique feature of the architecture is that it integrates inverse-shuffling, inverse-quantization, inverse-source-coding and motion-compensation into a single compact data-path. The entire block of pixel values are sampled in a single cycle for post processing after decompression. We use only (N/2(N/2+1))/2 multipliers and N2 adders. The configuration of the adders is such that motion compensation is realized in a single cycle following decompression
    BibTeX:
    @article{1995_Neogia,
      author = {Neogi, R. and Saha, A.},
      title = {Embedded parallel divide-and-conquer video decompression algorithm and architecture for HDTV applications},
      journal = {Consumer Electronics, IEEE Transactions on},
      year = {1995},
      volume = {41},
      number = {1},
      pages = {160 -171},
      doi = {http://dx.doi.org/10.1109/30.370323}
    }
    					
    Neogi, R. Real-time integrated video-compression architecture for broadcasting HDTV and multimedia applications 1995 ASIC Conference and Exhibit, 1995., Proceedings of the Eighth Annual IEEE International , pp. 79 -82   inproceedings DOI    
    Abstract: DCT/IDCT based source coding and decoding techniques are widely accepted in HDTV systems and other MPEG based multimedia applications. In this paper we propose a new direct 2-D DCT algorithm based on parallel divide and conquer approach for computation in real-time. The algorithm distributes computation by considering one time-domain coefficient at a time and doing partial computation and updating as every coefficient arrives. A novel parallel and fully pipelined architecture with an effective processing time of one cycle per pixel for an N times;N size block is designed to implement the algorithm. An unique feature of this architecture is that it integrates shuffling and source-coding into a single compact data-path. We avoid the insertion of a FIFO between the motion-estimator and compression engine. The entire block of frequency coefficients are sampled in a single cycle for statistical encoding after compression. Also, we use only N2 multipliers and N2 adders
    BibTeX:
    @inproceedings{1995_Neogib,
      author = {Neogi, R.},
      title = {Real-time integrated video-compression architecture for broadcasting HDTV and multimedia applications},
      journal = {ASIC Conference and Exhibit, 1995., Proceedings of the Eighth Annual IEEE International},
      year = {1995},
      pages = {79 -82},
      doi = {http://dx.doi.org/10.1109/ASIC.1995.580686}
    }
    					
    Otomo, G.; Hara, H.; Oto, T.; Seta, K.; Kitagaki, K.; Ishiwata, S.; Michinaka, S.; Shimazawa, T.; Matsui, M.; Demura, T.; Koyama, M.; Watanabe, Y.; Sano, F.; Chiba, A.; Matsuda, K. & Sakurai, T. Special memory and embedded memory macros in MPEG environment 1995 Custom Integrated Circuits Conference, 1995., Proceedings of the IEEE 1995 , pp. 139 -142   inproceedings DOI    
    Abstract: Special memory and embedded memories used in a newly designed MPEG2 decoder LSI are described. Orthogonal memory is employed in a IDCT (Inverse Discrete Cosine Transform) block for small area and power. FIFOs and other dual-port memories are designed by using a single-port RAM operated twice in one clock cycle to reduce cost. As for testability, direct test mode is implemented for small area. An instruction RAM is placed outside the pad area in parallel to a normal instruction ROM and activated by Al-masterslice for extensive debugging and an early sampling. Other memory related techniques and the key features of the decoder are also described
    BibTeX:
    @inproceedings{1995_Otomo,
      author = {Otomo, G. and Hara, H. and Oto, T. and Seta, K. and Kitagaki, K. and Ishiwata, S. and Michinaka, S. and Shimazawa, T. and Matsui, M. and Demura, T. and Koyama, M. and Watanabe, Y. and Sano, F. and Chiba, A. and Matsuda, K. and Sakurai, T.},
      title = {Special memory and embedded memory macros in MPEG environment},
      journal = {Custom Integrated Circuits Conference, 1995., Proceedings of the IEEE 1995},
      year = {1995},
      pages = {139 -142},
      doi = {http://dx.doi.org/10.1109/CICC.1995.518153}
    }
    					
    Potkonjak, M. & Chandrakasan, A. Synthesis and selection of DCT algorithms using behavioral synthesis-based algorithm space exploration 1995 Image Processing, 1995. Proceedings., International Conference on
    Vol. 1 , pp. 65 -68 vol.1  
    inproceedings DOI    
    Abstract: Numerous fast algorithms for the discrete cosine transform (DCT) have been proposed in image and video processing literature. Until recently, it has been difficult to compare different DCT algorithms and select one which is best suited for implementation under a given set of design goals and constraints. In this paper, we propose an approach for design space exploration at the algorithm and behavioral levels using high level synthesis tools. In particular, we study and compare the following nine DCT algorithms: Lee's, Wang's, DIT, DFT, QR, Givens, Arai, MCM, and direct algorithm. The main conclusion of this study is that the best choice among fast DCT algorithms depends on a particular set of design goals and constraints. Another important conclusion is that for almost all sets of implementation goals and constraints more than an order of magnitude improvement can be achieved using algorithm and behavioral design space exploration
    BibTeX:
    @inproceedings{1995_Potkonjak,
      author = {Potkonjak, M. and Chandrakasan, A.},
      title = {Synthesis and selection of DCT algorithms using behavioral synthesis-based algorithm space exploration},
      journal = {Image Processing, 1995. Proceedings., International Conference on},
      year = {1995},
      volume = {1},
      pages = {65 -68 vol.1},
      doi = {http://dx.doi.org/10.1109/ICIP.1995.529040}
    }
    					
    Tazebay, M. & Akansu, A. A comparative performance study of excisers in spread spectrum communications 1995 Global Telecommunications Conference, 1995. GLOBECOM '95., IEEE
    Vol. 2 , pp. 1131 -1135 vol.2  
    inproceedings DOI    
    Abstract: The comparative performance of a direct-sequence spread-spectrum (DSSS) receiver for different excisers is presented. The performance of the adaptive time-frequency (ATF) exciser is ranked along with fixed and optimal transform domain-based excisers. The smart ATF exciser is able to decide on the domain of processing. For the narrowband interference case, the ATF exciser employs an adaptive subband filter bank with the best transform basis. The bit error rate performance of DSSS receiver is simulated for ATF, DFT, DCT, optimal KLT, and fixed subband filter bank-based interference cancellers. It is shown that the smart ATF exciser significantly outperforms the others for the scenarios considered. The performance of ATF is very robust to the variations of the interference
    BibTeX:
    @inproceedings{1995_Tazebay,
      author = {Tazebay, M.V. and Akansu, A.N.},
      title = {A comparative performance study of excisers in spread spectrum communications},
      journal = {Global Telecommunications Conference, 1995. GLOBECOM '95., IEEE},
      year = {1995},
      volume = {2},
      pages = {1131 -1135 vol.2},
      doi = {http://dx.doi.org/10.1109/GLOCOM.1995.502579}
    }
    					
    Tsai, T.-H.; Chen, T.-H. & Chen, L.-G. An MPEG audio decoder chip 1995 Consumer Electronics, IEEE Transactions on
    Vol. 41 (1) , pp. 89 -96  
    article DOI    
    Abstract: A single chip MPEG audio decoder which performs the decoding scheme with a direct hardware implementation approach, without programming, is presented. Based on the computation analysis of the decoder, we develop a novel design which only requires half the computational capacity of previous decoders and half the storage memory too. The chip is designed with cost-effective and simple circuitry features, and thus it provides a consumer-economical solution, especially in design time and complexity
    BibTeX:
    @article{1995_Tsai,
      author = {Tsung-Han Tsai and Thou-Ho Chen and Liang-Gee Chen},
      title = {An MPEG audio decoder chip},
      journal = {Consumer Electronics, IEEE Transactions on},
      year = {1995},
      volume = {41},
      number = {1},
      pages = {89 -96},
      doi = {http://dx.doi.org/10.1109/30.370314}
    }
    					
    Yeo, B.-L. & Liu, B. On the extraction of DC sequence from MPEG compressed video 1995 Image Processing, 1995. Proceedings., International Conference on
    Vol. 2 , pp. 260 -263 vol.2  
    inproceedings DOI    
    Abstract: Reduced images formed from the collection of DC coefficients in intra-coded DCT compressed video retain a ldquo;global rdquo; feature which is useful for processing purposes. This paper examines the direct reconstruction of such DC images from motion-compensated P-frames and B-frames of MPEG compressed video. Exact expressions are derived and approximations are given for the computational speedup. The analysis and experimental results show that the approximations yield images that are very close to the true ones
    BibTeX:
    @inproceedings{1995_Yeo,
      author = {Yeo, B.-L. and Liu, B.},
      title = {On the extraction of DC sequence from MPEG compressed video},
      journal = {Image Processing, 1995. Proceedings., International Conference on},
      year = {1995},
      volume = {2},
      pages = {260 -263 vol.2},
      doi = {http://dx.doi.org/10.1109/ICIP.1995.537464}
    }
    					
    Yeo, B.-L. & Liu, B. Volume rendering of DCT-based compressed 3D scalar data 1995 Visualization and Computer Graphics, IEEE Transactions on
    Vol. 1 (1) , pp. 29 -43  
    article DOI    
    Abstract: The paper proposes a scheme to perform volume rendering from compressed scalar data. Instead of decompressing the entire data set before rendering, blocks of data are decompressed as needed. Discrete cosine transform based compression technique is used to illustrate the method. The data is partitioned into overlapping blocks to permit local rendering and allow easy parallelization. Compression by factor of 20 to 30 produces rendering virtually indistinguishable from rendering using the original uncompressed data. Speedup is obtained by making use of spatial homogeneity detected in the transform domain. Rendering time using the proposed approach is less than that of direct rendering from the entire uncompressed data. The proposed method thus offers an attractive option to reduce storage, computation, and transmission overhead of otherwise huge data sets
    BibTeX:
    @article{1995_Yeoa,
      author = {Boon-Lock Yeo and Bede Liu},
      title = {Volume rendering of DCT-based compressed 3D scalar data},
      journal = {Visualization and Computer Graphics, IEEE Transactions on},
      year = {1995},
      volume = {1},
      number = {1},
      pages = {29 -43},
      doi = {http://dx.doi.org/10.1109/2945.468390}
    }
    					
    Yun, I.D. & Lee, S.U. On the fixed-point error analysis of several fast IDCT algorithms 1995 Circuits and Systems II: Analog and Digital Signal Processing, IEEE Transactions on
    Vol. 42 (11) , pp. 685 -693  
    article DOI    
    Abstract: In this paper, a fixed-point error analysis for well-known fast l-D IDCT algorithms, such as Lee, Hou, and Vetterli, are presented. For a comparison purpose, a direct-form method is also included in our investigation. Based on the l-D analysis, the fixed-point error analysis of the row-column method and the Cho-Lee algorithm are also investigated for 2-D IDCT. Closed-form expressions for the rounding error variances are derived and compared with the experimental results. There is a close agreement between the theory and experiment, demonstrating that the analysis presented in this paper is valid. In addition, we also discuss the minimum word length to satisfy requirements for the implementation of 8 times;8 IDCT
    BibTeX:
    @article{1995_Yun,
      author = {Il Dong Yun and Sang Uk Lee},
      title = {On the fixed-point error analysis of several fast IDCT algorithms },
      journal = {Circuits and Systems II: Analog and Digital Signal Processing, IEEE Transactions on},
      year = {1995},
      volume = {42},
      number = {11},
      pages = {685 -693},
      doi = {http://dx.doi.org/10.1109/82.475243}
    }
    					
    Akopian, D.; Egiazarian, I.; Agaian, S. & Astola, J. Pipeline processors for fast trigonometric transforms 1994 Signals, Systems and Computers, 1994. 1994 Conference Record of the Twenty-Eighth Asilomar Conference on
    Vol. 2 , pp. 1031 -1035 vol.2  
    inproceedings DOI    
    Abstract: Pipeline processors are proposed for computation of fast discrete trigonometric transforms (DTT), particularly for cosine (DCT) and sine (DST) transforms. The transform time is O(N), where N=2n is the length of the input data vector. Because of the using the internal properties of the considered transforms, the numbers of adders, multipliers and registers are reduced when compared with the pipeline structures for fast Fourier transform and direct DCT
    BibTeX:
    @inproceedings{1994_Akopian,
      author = {Akopian, D. and Egiazarian, I. and Agaian, S. and Astola, J.},
      title = {Pipeline processors for fast trigonometric transforms},
      journal = {Signals, Systems and Computers, 1994. 1994 Conference Record of the Twenty-Eighth Asilomar Conference on},
      year = {1994},
      volume = {2},
      pages = {1031 -1035 vol.2},
      doi = {http://dx.doi.org/10.1109/ACSSC.1994.471616}
    }
    					
    Bhaskaran, R. & Kwatra, S. Compression of HDTV signals for low bit-rate transmission using motion compensated subband transform coding and a self-organization neural network 1994 Data Compression Conference, 1994. DCC '94. Proceedings , pp. 195 -204   inproceedings DOI    
    Abstract: The quality of visual communications has greatly improved with HDTV. Due to the advantages of digital communications, and the trend towards it, it is desirable to develop a cost effective digital codec for HDTV signals. One of the criteria in the transmission of digital signals via satellites is the transmission bandwidth which dictates the transmission costs. Apart from development of good compression schemes, due consideration should also be given to the ease of hardware implementation, which is crucial in deciding factors like processing delay and the level of encoder/decoder complexity. With the above mentioned factors in mind a cost-effective HDTV codec is proposed. Few US organizations have addressed the problem of HDTV distribution via communication satellites. The majority of efforts in the development of an HDTV codec have been aimed at the 100 to 140 Mbits/sec data rate. To facilitate cost effective satellite transmission, the DS3 rate of 45 Mbits/sec, is desirable. The purpose of the present research is to compress the HDTV signals to around 20 Mbits/sec so that 2 HDTV channels can be transmitted through the DS3 rate channel. To achieve this degree of compression a new scheme is proposed. First subband coding is used to decompose the image into different frequency bands. Next motion compensation is applied to the low band, and the resulting interframe difference is discrete cosine transform (DCT) coded. The high bands are also DCT coded. The DCT coefficients of the difference signal of the low band and the DCT coefficients of the high band are then vector quantized
    BibTeX:
    @inproceedings{1994_Bhaskaran,
      author = {Bhaskaran, R. and Kwatra, S.C.},
      title = {Compression of HDTV signals for low bit-rate transmission using motion compensated subband transform coding and a self-organization neural network},
      journal = {Data Compression Conference, 1994. DCC '94. Proceedings},
      year = {1994},
      pages = {195 -204},
      doi = {http://dx.doi.org/10.1109/DCC.1994.305927}
    }
    					
    Bhattacharya, A. & Haider, S. A VLSI architecture of an inverse discrete cosine transform 1994 VLSI Design, 1994., Proceedings of the Seventh International Conference on , pp. 87 -90   inproceedings DOI    
    Abstract: The inverse discrete cosine transform (IDCT) is an important function in HDTV and multimedia systems complying with JPEG or MPEG standards for video compression. However, the IDCT is computationally intensive and therefore very expensive to implement in VLSI using direct matrix multiplication. By properly arranging the sequence of input coefficients and the output data, the rows and columns of the transform matrix can be reordered to build modular regularity which is suitable for VLSI implementation. Based on this technique, an architecture using only seven constant multipliers and only one 1 dimensional IDCT processor is presented
    BibTeX:
    @inproceedings{1994_Bhattacharya,
      author = {Bhattacharya, A.K. and Haider, S.S.},
      title = {A VLSI architecture of an inverse discrete cosine transform},
      journal = {VLSI Design, 1994., Proceedings of the Seventh International Conference on},
      year = {1994},
      pages = {87 -90},
      doi = {http://dx.doi.org/10.1109/ICVD.1994.282662}
    }
    					
    Gangal, A. & Kayikcioglu, T. A new bit rate reduction technique in video coding based on motion compensation using adaptive block matching 1994 Electrotechnical Conference, 1994. Proceedings., 7th Mediterranean , pp. 1113 -1116 vol.3   inproceedings DOI    
    Abstract: A new block matching algorithm is examined for motion compensated coding of videophone signals. In this algorithm, firstly, a motion vector field is obtained by using small size of blocks. Secondly, the motion vector filed is postprocessed and appropriate block dimensions are chosen for each region. Estimated displacement vectors are used to direct the motion compensated interframe predictor. Finally, the DCT of the motion compensated difference signal between current and previous frames, displacement vectors and block size information are coded. Simulation results of this algorithm shows less image degradation compared to the fixed size block matching algorithm for typical videophone images at the bit rate of 64 kbits/s
    BibTeX:
    @inproceedings{1994_Gangal,
      author = {Gangal, A. and Kayikcioglu, T.},
      title = {A new bit rate reduction technique in video coding based on motion compensation using adaptive block matching},
      journal = {Electrotechnical Conference, 1994. Proceedings., 7th Mediterranean},
      year = {1994},
      pages = {1113 -1116 vol.3},
      doi = {http://dx.doi.org/10.1109/MELCON.1994.380875}
    }
    					
    Matsui, M.; Hara, H.; Uetani, Y.; Kim, L.-S.; Nagamatsu, T.; Watanabe, Y.; Chiba, A.; Matsuda, K. & Sakurai, T. A 200 MHz 13 mm2 2-D DCT macrocell using sense-amplifying pipeline flip-flop scheme 1994 Solid-State Circuits, IEEE Journal of
    Vol. 29 (12) , pp. 1482 -1490  
    article DOI    
    Abstract: The two-dimensional discrete cosine transform (2D DCT) has been widely recognized as a key processing unit for image data compression/decompression. In this paper, the implementation of a 200 MHz 13.3 mm2 8 times;8 2-D DCT macrocell capable of HDTV rates, based on a direct realization of the DCT, and using distributed arithmetic is presented. The macrocell, fabricated using 0.8 mu;m base-rule CMOS technology and 0.5 mu;m MOSFET's, performs the DCT processing with 1 sample-(pixel)-per-clock throughput. The fast speed and small area are achieved by a novel sense-amplifying pipeline flip-flop (SA-F/F) circuit technique in combination with nMOS differential logic. The SA-F/F, a class of delay flip-flops, can be used as a differential synchronous sense-amplifier, and can amplify dual-rail inputs with swings lower than 100 mV. A 1.6 ns 20 bit carry skip adder used in the DCT macrocell, which was designed by the same scheme, is also described. The adder is 50% faster and 30% smaller than a conventional CMOS carry look ahead adder, which reduces the macrocell size by 15% compared to a conventional CMOS implementation
    BibTeX:
    @article{1994_Matsui,
      author = {Matsui, M. and Hara, H. and Uetani, Y. and Lee-Sup Kim and Nagamatsu, T. and Watanabe, Y. and Chiba, A. and Matsuda, K. and Sakurai, T.},
      title = {A 200 MHz 13 mm2 2-D DCT macrocell using sense-amplifying pipeline flip-flop scheme},
      journal = {Solid-State Circuits, IEEE Journal of},
      year = {1994},
      volume = {29},
      number = {12},
      pages = {1482 -1490},
      doi = {http://dx.doi.org/10.1109/4.340421}
    }
    					
    Campbell, E.J. Inversion of time domain signals from a Balle #x2013;Flygare type microwave spectrometer 1993 Review of Scientific Instruments
    Vol. 64 (8) , pp. 2166 -2172  
    article DOI    
    Abstract: The method of sine and cosine Fourier transforms with a phase correction, well known in NMR and traveling wave Fourier transform microwave spectroscopy, is examined for the pulsed beam Fabry #x2013;Perot cavity. The validity of this approach is closely related to the result, noted previously elsewhere [J. #x2010;L. Le Gou #xeb;t and P. R. Berman, Phys. Rev. A 20, 1105 (1979), and F. Rohart and B. Macke, J. Phys. 41, 837 (1980)] that only the polarization spatial harmonic that projects onto the driven fundamental mode of the cavity contributes appreciably to the free induction decay signal. When the full signal is recorded after a short polarization pulse, an absorption line shape function in the low power limit for a single nondegenerate transition can be recovered as an angular distribution weighted integral over the coordinate #x3b8; between the nozzle axis and a point on the cavity axis of Doppler #x2010;shifted Voight profiles. Passage of the beam through the transverse Gaussian profile of a TEM00q mode contributes an effective line broadening temperature. These results are a direct extension of a semiclassical derivation given earlier [E. J. Campbell, L. W. Buxton, T. J. Balle, and W. H. Flygare, J. Chem. Phys. 74, 813 (1981), and E. J. Campbell, L. W. Buxton, T. J. Balle, M. R. Keenan, and W. H. Flygare, J. Chem. Phys. 74, 829 (1981)]. Analytical results are derived using simplified cavity mode functional forms. Calculations using the exact mode expressions are reported. The relation between this approach and the direct absorption method is discussed.
    BibTeX:
    @article{1993_Campbell,
      author = {Campbell, Edward J.},
      title = {Inversion of time domain signals from a Balle #x2013;Flygare type microwave spectrometer},
      journal = {Review of Scientific Instruments},
      year = {1993},
      volume = {64},
      number = {8},
      pages = {2166 -2172},
      doi = {http://dx.doi.org/10.1063/1.1143955}
    }
    					
    Chang, S. & Messerschmitt, D. A new approach to decoding and compositing motion-compensated DCT-based images 1993 Acoustics, Speech, and Signal Processing, 1993. ICASSP-93., 1993 IEEE International Conference on
    Vol. 5 , pp. 421 -424 vol.5  
    inproceedings DOI    
    Abstract: A novel decoding algorithm for the MC-DCT (motion-compensation discrete-cosine-transform)-based video, which performs inverse MC before inverse DCT, is designed. This algorithm can be applied in compositing compressed video within the network, which may take multiple compressed video sources and combine them into a single compressed output stream. The proposed algorithm convers all MC-DCT compressed video into the DCT domain and performs compositing in the DCT domain. This DCT-domain approach can reduce the required computations with a speedup factor depending on the compression ratio and the nonzero motion vector percentage. However, dropping some least-significant DCT coefficients may be necessary for the worst case of high-motion video in real-time implementations. Some issues of networked video compositing are also discussed. Another direct application of the proposed decoding algorithm is converting MC-DCT compressed video to the DCT compressed format directly in the DCT domain
    BibTeX:
    @inproceedings{1993_Chang,
      author = {Chang, S.F. and Messerschmitt, D.G.},
      title = {A new approach to decoding and compositing motion-compensated DCT-based images},
      journal = {Acoustics, Speech, and Signal Processing, 1993. ICASSP-93., 1993 IEEE International Conference on},
      year = {1993},
      volume = {5},
      pages = {421 -424 vol.5},
      doi = {http://dx.doi.org/10.1109/ICASSP.1993.319837}
    }
    					
    Cho, N.; Yun, I. & Lee, S. On the regular structure for the fast 2-D DCT algorithm 1993 Circuits and Systems II: Analog and Digital Signal Processing, IEEE Transactions on
    Vol. 40 (4) , pp. 259 -266  
    article DOI    
    Abstract: N. I. Cho and S. U. Lee (1991) proposed a fast algorithm for 2-D N times;N DCT, where N=2m. It requires only half the number of multiplications of the conventional row-column approach. However, the signal flow graph for the postaddition stage seems very complicated and the order of the output index is seemingly irregular, because the postaddition stage was not based on the mathematical expressions. Consequently, derivation of the signal flow graph becomes complicated as the transform size increases. Systematic expressions for the postaddition stage of the algorithm that enable any N times;N DCT to be implemented in a straightforward manner are provided here. The results show that the signal flow graph from input to output has a recursive structure in which the structure for smaller N reappears for larger N. However, the number of additions increases in the new signal flow graph at the expense of improving the regularity in the structure
    BibTeX:
    @article{1993_Cho,
      author = {Cho, N.I. and Yun, I.D. and Lee, S.U.},
      title = {On the regular structure for the fast 2-D DCT algorithm},
      journal = {Circuits and Systems II: Analog and Digital Signal Processing, IEEE Transactions on},
      year = {1993},
      volume = {40},
      number = {4},
      pages = {259 -266},
      doi = {http://dx.doi.org/10.1109/82.224317}
    }
    					
    Gongli, Z. Split-radix DET algorithms and the computation of DCT and DST 1993 TENCON '93. Proceedings. Computer, Communication, Control and Power Engineering.1993 IEEE Region 10 Conference on (0) , pp. 550 -553 vol.3   inproceedings DOI    
    Abstract: Split-radix direct algorithms for four types of discrete exponent transforms (DETs) are introduced. The DCT and DST may be directly computed using the split-radix DET algorithms. It is shown that the direct computations of DCT and DST have simple structure and in-place implementation. Their output is in bit-reserved order, rather than Hadamard order or other complex order. The direct computations can be implemented using the software or hardware of split-radix DFT algorithms
    BibTeX:
    @inproceedings{1993_Gongli,
      author = {Zhang Gongli},
      title = {Split-radix DET algorithms and the computation of DCT and DST},
      journal = {TENCON '93. Proceedings. Computer, Communication, Control and Power Engineering.1993 IEEE Region 10 Conference on},
      year = {1993},
      number = {0},
      pages = {550 -553 vol.3},
      doi = {http://dx.doi.org/10.1109/TENCON.1993.328046}
    }
    					
    Lin, Z.; Chen, J. & McCallum, R. Transform domain adaptive filtering using recursive running DCT and its application in surface recording of small intestine 1993 Southeastcon '93, Proceedings., IEEE , pp. 7 p.   inproceedings DOI    
    Abstract: An adaptive system is proposed for the enhancement of the small intestine signal. To obtain better performance, adaptive signal enhancement is performed in the transform domain using the discrete cosine transform (DCT). A fast recursive algorithm is developed for the calculation of running DCT. The computational complexity of the proposed recursive algorithm is only 2/N (where N is the length of the adaptive filter) of the direct calculation of the running DCT. A series of simulations are conducted to investigate the performance of the proposed transform domain adaptive filtering using DCT in comparison with time-domain adaptive filter and with transform-domain adaptive filtering using the discrete Fourier transform. The parameters of the proposed adaptive system are optimized, and their effects on system performance are investigated
    BibTeX:
    @inproceedings{1993_Lin,
      author = {Lin, Z. and Chen, J. and McCallum, R.W.},
      title = {Transform domain adaptive filtering using recursive running DCT and its application in surface recording of small intestine},
      journal = {Southeastcon '93, Proceedings., IEEE},
      year = {1993},
      pages = {7 p.},
      doi = {http://dx.doi.org/10.1109/SECON.1993.465666}
    }
    					
    Tun, I.D. & Lee, S.U. On the fixed-point-error analysis of several fast DCT algorithms 1993 Circuits and Systems for Video Technology, IEEE Transactions on
    Vol. 3 (1) , pp. 27 -41  
    article DOI    
    Abstract: A fixed-point-error analysis for several 1D fast DCT algorithms is presented. For comparison, a direct-form approach is also included in the investigation. A statistical model is used as the basis for predicting the fixed-point error in implementing the algorithms, and a suitable scaling scheme is selected to avoid overflow. Closed-form expressions for both the mean and variance for fixed-point error are derived and compared with experimental results. Simulation results show close agreement between theory and experiment, validating the analysis. The results show that one of the algorithms is better than others in terms of average SNR performance. Based on the 1D analysis, attempts are made to investigate the fixed-point-error analysis of the two-column approach for 2D DCT. It is found that the fixed-point-error characteristics of the row-column approach for 2D DCTs is very similar to that of their 1D counterparts
    BibTeX:
    @article{1993_Tun,
      author = {Il Dong Tun and Sang Uk Lee},
      title = {On the fixed-point-error analysis of several fast DCT algorithms },
      journal = {Circuits and Systems for Video Technology, IEEE Transactions on},
      year = {1993},
      volume = {3},
      number = {1},
      pages = {27 -41},
      doi = {http://dx.doi.org/10.1109/76.180688}
    }
    					
    Yang, Y.; Galatsanos, N. & Katsaggelos, A. Iterative projection algorithms for removing the blocking artifacts of block-DCT compressed images 1993 Acoustics, Speech, and Signal Processing, 1993. ICASSP-93., 1993 IEEE International Conference on
    Vol. 5 , pp. 405 -408 vol.5  
    inproceedings DOI    
    Abstract: Iterative projection algorithms for reconstructing visually pleasing images from block discrete cosine transform (BDCT) compressed image data are presented. Two algorithms are proposed. The first is based on the theory of projections onto convex sets (POCS). The second is motivated by the theory of POCS. In order for the POCS approach to be effective, the convex sets must be as pertinent as possible to the required objective. For the deblocking of the DCT compressed images two types of constraint sets capture most of the information pertinent to this problem: the set that contains the information of the transmitted DCT coefficients and the sets that express the between-block smoothness properties of the desired image. The proposed iterative reconstruction algorithms have been tested thoroughly on different DCT-based compression approaches for a number of 256 times;256 and 512 times;512 images. The experimental results demonstrate that the proposed algorithms yield superior images to those obtained by direct reconstruction from the compressed data only
    BibTeX:
    @inproceedings{1993_Yang,
      author = {Yang, Y. and Galatsanos, N.P. and Katsaggelos, A.K.},
      title = {Iterative projection algorithms for removing the blocking artifacts of block-DCT compressed images},
      journal = {Acoustics, Speech, and Signal Processing, 1993. ICASSP-93., 1993 IEEE International Conference on},
      year = {1993},
      volume = {5},
      pages = {405 -408 vol.5},
      doi = {http://dx.doi.org/10.1109/ICASSP.1993.319833}
    }
    					
    Hsieh, C.-H. DCT-based codebook design for vector quantization of images 1992 Circuits and Systems for Video Technology, IEEE Transactions on
    Vol. 2 (4) , pp. 401 -409  
    article DOI    
    Abstract: A codebook design algorithm based on a two-dimensional discrete cosine transform (2-D DCT) is presented for vector quantization (VQ) of images. The significant features of training images are extracted by using the 2-D DCT. A codebook is generated by partitioning the training set into a binary tree. Each training vector at a nonterminal node of the binary tree is directed to one of the two descendants by comparing a single feature associated with that node to a threshold. Compared with the pairwise nearest neighbor (PNN) algorithm, the algorithm results in a considerable reduction in computation time and shows better picture quality
    BibTeX:
    @article{1992_Hsieh,
      author = {Chaur-Heh Hsieh},
      title = {DCT-based codebook design for vector quantization of images},
      journal = {Circuits and Systems for Video Technology, IEEE Transactions on},
      year = {1992},
      volume = {2},
      number = {4},
      pages = {401 -409},
      doi = {http://dx.doi.org/10.1109/76.168905}
    }
    					
    Kondo, H.; Maeda, E. & Hirai, H. Efficient coding method of chest X-ray images 1992 Singapore ICCS/ISITA '92. 'Communications on the Move' , pp. 340 -344 vol.1   inproceedings DOI    
    Abstract: An efficient discrete cosine transform (DCT) coding of chest X-ray images is presented. This coding method has a great ability to reduce the redundancy in the coding of chest X-ray images. Excellent performance is demonstrated in terms of signal to noise ratio (SNR), and in also by direct comparison of original and reconstructed images
    BibTeX:
    @inproceedings{1992_Kondo,
      author = {Kondo, H. and Maeda, E. and Hirai, H.},
      title = {Efficient coding method of chest X-ray images},
      journal = {Singapore ICCS/ISITA '92. 'Communications on the Move'},
      year = {1992},
      pages = {340 -344 vol.1},
      doi = {http://dx.doi.org/10.1109/ICCS.1992.255011}
    }
    					
    Wolter, S.; Birreck, D.; Heine, M. & Laur, R. Parallel architectures for 8 times;8 discrete cosine transforms 1992 Circuits and Systems, 1992. ISCAS '92. Proceedings., 1992 IEEE International Symposium on
    Vol. 1 , pp. 149 -152 vol.1  
    inproceedings DOI    
    Abstract: The design of multiplier-free parallel architectures for computing the 8 times;8 discrete cosine transform (DCT) is addressed. The focus is on direct methods, which avoid a row-column decomposition. Two architectures are proposed and compared. One uses polynomial transforms; the other computes the DCT via the Walsh-Hadamard transform (WHT). Both architectures achieve a high degree of parallelism and regularity. The architectures are designed for HDTV sampling rates and can be efficiently realized in CMOS technology
    BibTeX:
    @inproceedings{1992_Wolter,
      author = {Wolter, S. and Birreck, D. and Heine, M. and Laur, R.},
      title = {Parallel architectures for 8 times;8 discrete cosine transforms},
      journal = {Circuits and Systems, 1992. ISCAS '92. Proceedings., 1992 IEEE International Symposium on},
      year = {1992},
      volume = {1},
      pages = {149 -152 vol.1},
      doi = {http://dx.doi.org/10.1109/ISCAS.1992.229992}
    }
    					
    Chau, K.; Wang, I.-F. & Eldridge, C. VLSI implementation of a 2-D DCT in a compiler 1991 Acoustics, Speech, and Signal Processing, 1991. ICASSP-91., 1991 International Conference on , pp. 1233 -1236 vol.2   inproceedings DOI    
    Abstract: A 100 MHz realization of a 2D DCT (discrete cosine transform) in a compiler is presented. An optimal nibble-serial distributed arithmetic architecture is used for an efficient direct implementation of the DCT. The compiler, which is based on BiCMOS gate array technology and functional module generation technology, is used to support a rapid prototyping environment. An 8 times;8 inverse 2D DCT example in a 0.8 mu;m BiCMOS gate array is given to illustrate the performance and flexibility of this approach. The design flow and a proposed development system, the Signal Analysis Workstation (SAW), are shown for customized system integrations
    BibTeX:
    @inproceedings{1991_Chau,
      author = {Chau, K.K. and Wang, I.-F. and Eldridge, C.L.},
      title = {VLSI implementation of a 2-D DCT in a compiler},
      journal = {Acoustics, Speech, and Signal Processing, 1991. ICASSP-91., 1991 International Conference on},
      year = {1991},
      pages = {1233 -1236 vol.2},
      doi = {http://dx.doi.org/10.1109/ICASSP.1991.150616}
    }
    					
    Gluth, R. Regular FFT-related transform kernels for DCT/DST-based polyphase filter banks 1991 Acoustics, Speech, and Signal Processing, 1991. ICASSP-91., 1991 International Conference on , pp. 2205 -2208 vol.3   inproceedings DOI    
    Abstract: A realization for the discrete cosine and sine transforms (DCT and DST) is proposed which is based on an FFT (fast Fourier transform) optimized for DSP (digital signal processor) use. Besides its regular structure, even the theoretically interesting total number of required operations is about the same as needed by the most regular direct approach and only slightly higher than the minimum number yet published. Compared to a similar approach published recently by N. Ramamurthy and M.N.S. Swamy (1990), it is even more efficient
    BibTeX:
    @inproceedings{1991_Gluth,
      author = {Gluth, R.},
      title = {Regular FFT-related transform kernels for DCT/DST-based polyphase filter banks},
      journal = {Acoustics, Speech, and Signal Processing, 1991. ICASSP-91., 1991 International Conference on},
      year = {1991},
      pages = {2205 -2208 vol.3},
      doi = {http://dx.doi.org/10.1109/ICASSP.1991.150852}
    }
    					
    Kou, W. & Fjallbrant, T. A direct computation of DCT coefficients for a signal block taken from two adjacent blocks 1991 Signal Processing, IEEE Transactions on
    Vol. 39 (7) , pp. 1692 -1695  
    article DOI    
    Abstract: A direct computational algorithm for obtaining the DCT (discrete cosine transform) coefficients of a signal block taken from two adjacent blocks is proposed. This algorithm reduces the number of both multiplications and additions/subtractions compared to the traditional method, which requires inverse transforms of two received coefficient blocks followed by a forward transform. The proposed algorithm is efficient and useful for speech coding and image processing, especially for real-time applications
    BibTeX:
    @article{1991_Kou,
      author = {Kou, W. and Fjallbrant, T.},
      title = {A direct computation of DCT coefficients for a signal block taken from two adjacent blocks},
      journal = {Signal Processing, IEEE Transactions on},
      year = {1991},
      volume = {39},
      number = {7},
      pages = {1692 -1695},
      doi = {http://dx.doi.org/10.1109/78.134412}
    }
    					
    Cominetti, M. & Molo, F. A codec for HDTV signal transmission through terrestrial and satellite digital links 1990 Global Telecommunications Conference, 1990, and Exhibition. 'Communications: Connecting the Future', GLOBECOM '90., IEEE , pp. 987 -992 vol.2   inproceedings DOI    
    Abstract: A codec for digital transmission of HDTV signals designed to operate with the 1125/60 and 1250/50 interlaced studio systems is described. The codec is based on the application of discrete cosine transforms (DCT), combined with temporal differential pulse-code modulation (DPCM) on blocks of 8 times;8 samples for both luminance and color differences. The line bit rate can range from a minimum of about 60 Mb/s up to more than 140 Mb/s. This range includes the standard digital CCITT hierarchies of 2 times;34 Mb/s, 2 times;45 Mb/s, and 140 Mb/s. For higher bit rates the full studio quality (1920 pixels/line) can be transmitted, while for lower bit rates, a horizontal resolution of about 1440 pixels/line is a better compromise, in order not to stress the compression algorithm. Applications of the codec to various transmission links, including satellite transmission, are discussed
    BibTeX:
    @inproceedings{1990_Cominetti,
      author = {Cominetti, M. and Molo, F.},
      title = {A codec for HDTV signal transmission through terrestrial and satellite digital links},
      journal = {Global Telecommunications Conference, 1990, and Exhibition. 'Communications: Connecting the Future', GLOBECOM '90., IEEE},
      year = {1990},
      pages = {987 -992 vol.2},
      doi = {http://dx.doi.org/10.1109/GLOCOM.1990.116650}
    }
    					
    Duhamel, P.; Guillemot, C. & Carlach, J. A DCT chip based on a new structured and computationally efficient DCT algorithm 1990 Circuits and Systems, 1990., IEEE International Symposium on , pp. 77 -80 vol.1   inproceedings DOI    
    Abstract: A discrete cosine transform (DCT) algorithm and architecture that minimize both software and hardware costs are presented. The proposed approaches are either direct or indirect and are based on the decomposition of the DCT in three operations: permutation, fast Fourier transform, and rotation. The main characteristics of the VLSI implementation chosen for this DCT and inverse DCT algorithm are show. Its data path coupled with a twin-pages memory and its controller, which contains the microprograms of the DCT algorithm are described. The results in terms of data processing rate and silicon area are given
    BibTeX:
    @inproceedings{1990_Duhamel,
      author = {Duhamel, P. and Guillemot, C. and Carlach, J.C.},
      title = {A DCT chip based on a new structured and computationally efficient DCT algorithm},
      journal = {Circuits and Systems, 1990., IEEE International Symposium on},
      year = {1990},
      pages = {77 -80 vol.1},
      doi = {http://dx.doi.org/10.1109/ISCAS.1990.111917}
    }
    					
    Duhamel, P. & Guillemot, C. Polynomial transform computation of the 2-D DCT 1990 Acoustics, Speech, and Signal Processing, 1990. ICASSP-90., 1990 International Conference on , pp. 1515 -1518 vol.3   inproceedings DOI    
    Abstract: A 2-D DCT (discrete cosine transform) algorithm based on a direct polynomial approach is presented. The resulting algorithm reduces the number of both multiplications and additions compared to previous algorithms. It is shown that, although being mathematically involved, it possesses a clean, butterfly-based structure. Tables comparing the number of operations are provided, as well as flowgraphs
    BibTeX:
    @inproceedings{1990_Duhamela,
      author = {Duhamel, P. and Guillemot, C.},
      title = {Polynomial transform computation of the 2-D DCT},
      journal = {Acoustics, Speech, and Signal Processing, 1990. ICASSP-90., 1990 International Conference on},
      year = {1990},
      pages = {1515 -1518 vol.3},
      doi = {http://dx.doi.org/10.1109/ICASSP.1990.115696}
    }
    					
    Hughes, R. & Heron, M. Physical Science, Measurement and Instrumentation, Management and Education, IEE Proceedings A 1989 Physical Science, Measurement and Instrumentation, Management and Education, IEE Proceedings A
    Vol. 136 (4) , pp. 223 -228  
    article    
    Abstract: The technique of spectral analysis, by truncated approximations to the sine and cosine functions, is evaluated using a three-level approximation (-1, 0, +1), where the -1 and +1 sections are of equal length, and the length of the zero level is variable. An optimum ratio is found where the zero level together cover 120 degrees of the cycle length, to give the least amount of ringing, or leakage. The transform that converts the square-wave spectral estimates to the Fourier coefficients is evaluated. In applications where spectral estimates are required only for limited frequency bands, the technique is flexible and efficient. This analysis is also well suited to irregularly spaced samples, and for direct application to analogue signals.<>
    BibTeX:
    @article{1989_Hughes,
      author = {Hughes, R.D. and Heron, M.L.},
      title = {Physical Science, Measurement and Instrumentation, Management and Education, IEE Proceedings A},
      journal = {Physical Science, Measurement and Instrumentation, Management and Education, IEE Proceedings A},
      year = {1989},
      volume = {136},
      number = {4},
      pages = {223 -228}
    }
    					
    Lee, B.G. Input and output index mappings for a prime-factor-decomposed computation of discrete cosine transform 1989 Acoustics, Speech and Signal Processing, IEEE Transactions on
    Vol. 37 (2) , pp. 237 -244  
    article DOI    
    Abstract: A formal direct derivation of the prime-factor-decomposed computation algorithm is presented. The derivation is direct in the sense that it is based on the real cosine function without resort to the discrete Fourier transform expressions or the complex functions. Based on the equations obtained from the derivation, input and output index mappings are introduced in the form of tables. This tabulation enables any prime-factor-decomposable discrete cosine transform (DCT) to be implemented in a straight-forward manner. The use of the index mapping tables is demonstrated for the 12-point DCT
    BibTeX:
    @article{1989_Lee,
      author = {Byeong G Lee},
      title = {Input and output index mappings for a prime-factor-decomposed computation of discrete cosine transform},
      journal = {Acoustics, Speech and Signal Processing, IEEE Transactions on},
      year = {1989},
      volume = {37},
      number = {2},
      pages = {237 -244},
      doi = {http://dx.doi.org/10.1109/29.21686}
    }
    					
    Artieri, A.; Kritter, S.; Jutand, F. & Demassieux, N. A one chip VLSI for real time two-dimensional discrete cosine transform 1988 Circuits and Systems, 1988., IEEE International Symposium on , pp. 701 -704 vol.1   inproceedings DOI    
    Abstract: A single-chip two-dimensional discrete cosine transform processor is presented. This chip meets the challenge of high throughput rate (13.5 MHz) and versatility (block size from 4 times;4 to 16 times;16, and direct and inverse DCT) with a die area as small as 40 mm2. An efficient optimized architecture providing high computation power is described. The chip is found to exhibit excellent precision performances. A full-custom approach was chosen because of the required speed and economic reasons (implementation of a low-cost real-time video coder/decoder). The chip uses a 1.25- mu;m CMOS technology and contains 114000 transistors on 5.4 times;7.5 mm2
    BibTeX:
    @inproceedings{1988_Artieri,
      author = {Artieri, A. and Kritter, S. and Jutand, F. and Demassieux, N.},
      title = {A one chip VLSI for real time two-dimensional discrete cosine transform},
      journal = {Circuits and Systems, 1988., IEEE International Symposium on},
      year = {1988},
      pages = {701 -704 vol.1},
      doi = {http://dx.doi.org/10.1109/ISCAS.1988.15022}
    }
    					
    Hsu, C.-Y. & Wu, J.-L. Block-diagonal structure of Walsh-Hadamard/discrete cosine transform 1987 Electronics Letters
    Vol. 23 (21) , pp. 1123 -1124  
    article DOI    
    Abstract: The A-matrix, the conversion matrix for Walsh-Hadamard/discrete cosine transform, is known for its efficient block-diagonal structure. This associates with the even/odd structure of the transform kernels. In the letter we present a direct matrix derivation by using the intrinsic properties of the discrete cosine transform and the Walsh-Hadamard transform.
    BibTeX:
    @article{1987_Chau-YunHsu,
      author = {Chau-Yun Hsu, and Ja-Ling Wu,},
      title = {Block-diagonal structure of Walsh-Hadamard/discrete cosine transform},
      journal = {Electronics Letters},
      year = {1987},
      volume = {23},
      number = {21},
      pages = {1123 -1124},
      doi = {http://dx.doi.org/10.1049/el:19870783}
    }
    					
    Guglielmo, M. An Analysis of Error Behavior in the Implementation of 2-D Orthogonal Transformations 1986 Communications, IEEE Transactions on
    Vol. 34 (9) , pp. 973 - 975  
    article    
    Abstract: The paper analyzes the effect of finite-length arithmetic in the calculation of 2-D linear transformations employed in some picture coding algorithms. Since the condition of zero-error in general direct and reverse transformations leads to results of little practical importance, an analysis is carried out on the statistical properties of error in 2-D linear transformation with given length of arithmetics. Then the important case of discrete cosine transform (DCT) applied to real images is considered in detail. The results of the paper allow a circuit designer to determine the representation accuracy of the one- and two-dimensional coefficients required to satisfy a preassigned reconstruction error on the image.
    BibTeX:
    @article{1986_Guglielmo,
      author = { Guglielmo, M.},
      title = {An Analysis of Error Behavior in the Implementation of 2-D Orthogonal Transformations},
      journal = {Communications, IEEE Transactions on},
      year = {1986},
      volume = {34},
      number = {9},
      pages = { 973 - 975}
    }
    					
    Ersoy, O. On relating discrete Fourier, sine, and symmetric cosine transforms 1985 Acoustics, Speech and Signal Processing, IEEE Transactions on
    Vol. 33 (1) , pp. 219 - 222  
    article    
    Abstract: The relationship among a real formalism of the discrete Fourier transform, discrete sine transform, and discrete symmetric cosine transform is discussed. It is shown that the real formalism of the discrete Fourier transform is basically equivalent to the direct sum of the other two transforms, with modifications in the pre- and post-computations with the data vector.
    BibTeX:
    @article{1985_Ersoy,
      author = {Ersoy, O.},
      title = {On relating discrete Fourier, sine, and symmetric cosine transforms},
      journal = {Acoustics, Speech and Signal Processing, IEEE Transactions on},
      year = {1985},
      volume = {33},
      number = {1},
      pages = { 219 - 222}
    }
    					
    Haque, M. A two-dimensional fast cosine transform 1985 Acoustics, Speech and Signal Processing, IEEE Transactions on
    Vol. 33 (6) , pp. 1532 - 1539  
    article    
    Abstract: A two-dimensional fast cosine transform algorithm (2-D FCT) is developed for 2m #215; 2ndata points. This algorithm is an extended version of the 1-D FCT algorithm introduced in a recent paper, but with significantly reduced computations for a 2-D field. The rationale for this 2-D FCT is a 2-D decomposition of data sequences into 2- D subblocks with reduced dimension (halves), rather than serial, one-dimensional, separable treatment for the columns and rows of the data sets. Computer simulation for the 2-D FCT algorithms, using a smaller block of data and finite word precision, proves to be excellent in comparison with the direct 2-D discrete cosine transform (2-D DCT). An example of a 4 #215; 4 2-D inverse fast cosine transform (2-D IFCT) algorithm development is presented in this paper, together with a signal flow graph.
    BibTeX:
    @article{1985_Haque,
      author = {Haque, M.},
      title = {A two-dimensional fast cosine transform},
      journal = {Acoustics, Speech and Signal Processing, IEEE Transactions on},
      year = {1985},
      volume = {33},
      number = {6},
      pages = { 1532 - 1539}
    }
    					
    Arnould, E. & Dugre, J. Real time discrete cosine transform an original architecture 1984 Acoustics, Speech, and Signal Processing, IEEE International Conference on ICASSP '84.
    Vol. 9 , pp. 557 - 560  
    inproceedings    
    Abstract: This paper presents an original hardware architecture for a two-dimensional real time Discrete Cosine Transform (DCT) processor for television signals. The work has been focused on low power consumption, low cost and minimum size because of the necessity (at least for the expander) to insert the equipment in the subscriber's home. The basic element of the transformer consists of a single chip 2D-DCT processor working on (8 #215; 8) blocks of 8 bits with an objective cycle time of 100 ns. The highly pipelined structure of the processor is optimized for the Fast Discrete Cosine Transform (FDCT) algorithm proposed by W.H. Chen and al [1]. In order to satisfy speed and resolution constraints for real time compression of television signals, four of these basic DCT processors need to work in parallel, controlled by two Direct Memory Access Controllers (DMAC). This architecture compares favourably with other works published in the literature [2] [3].
    BibTeX:
    @inproceedings{1984_Arnould,
      author = {Arnould, E. and Dugre, J.},
      title = {Real time discrete cosine transform an original architecture},
      journal = {Acoustics, Speech, and Signal Processing, IEEE International Conference on ICASSP '84.},
      year = {1984},
      volume = {9},
      pages = { 557 - 560}
    }
    					
    Hein, D. & Ahmed, N. On a Real-Time Walsh-Hadamard/Cosine Transform Image Processor 1978 Electromagnetic Compatibility, IEEE Transactions on
    Vol. EMC-20 (3) , pp. 453 -457  
    article DOI    
    Abstract: A real-time image processor which is capable of video compression using either the sequency-ordered Walsh-Hadamard transform (WHT)W, or the discrete cosine transform (DCT), is considered. The processing is done on an intraframe basis in (8 X 8) data blocks. The (WHT)W coefficients are computed directly, and then used to obtain the DCT coefficients. This is achieved via an (8 X 8) transformation matrix which is orthonormal, and has a block-diagonal structure. As such, it results in substantial savings in the number of multiplications and additions required to obtain the DCT, relative to its direct computation. Some aspects of a hardware implementation of the processor are also included.
    BibTeX:
    @article{1978_Hein,
      author = {Hein, D. and Ahmed, N.},
      title = {On a Real-Time Walsh-Hadamard/Cosine Transform Image Processor},
      journal = {Electromagnetic Compatibility, IEEE Transactions on},
      year = {1978},
      volume = {EMC-20},
      number = {3},
      pages = {453 -457},
      doi = {http://dx.doi.org/10.1109/TEMC.1978.303679}
    }
    					
    Roberts, J. & Gaster, M. Rapid estimation of spectra from irregularly sampled records 1978 Electrical Engineers, Proceedings of the Institution of
    Vol. 125 (2) , pp. 92 -96  
    article DOI    
    Abstract: Records of physical quantities often arise as continuous electrical signals. Spectral estimates may be formed either by analogue means or from digitised samples that are then processed on a computer, When the samples are provided at regularly spaced time instants, this can be achieved very quickly with the aid of the f.f.t (fast Fourier transform) algorithm. There are situations, however, where the data is known only at random time instants, and the paper is concerned with the computation of spectral estimates from such data. When the sample times are Poisson distributed, it has been shown, in previous papers, that unbiased alias-free estimates can be formed, either through the correlation function or by a direct Fourier transform of short blocks of data. Random sampling introduces additional variability in these spectral estimates, and it is consequently necessary to process a large amount of data in order to achieve stable results. Unfortunately, this is very time consuming, most of the computer effort being spent evaluating sine and cosine functions which are then multiplied by the data samples. Here, two methods that can be used to simplify this operation are discussed. It is shown that when the sine and cosine functions are replaced by their equivalent rectangular waveforms, the resulting estimates can be related to spectral estimates through the Fourier expansion for the rectangular waves. A second way of speeding up the processing of Gaussian signals can be achieved by quantising the data to a sign bit and using the `arc-sine¿ rule to transform the autocorrelation function to that of the full signal. It is shown that when both techniques are used together, and the processing reduced to 1-bit logical operations, valid spectral estimates can indeed be formed. These ideas are tested on various simulated sets of data.
    BibTeX:
    @article{1978_Roberts,
      author = {Roberts, J.B. and Gaster, M.},
      title = {Rapid estimation of spectra from irregularly sampled records},
      journal = {Electrical Engineers, Proceedings of the Institution of},
      year = {1978},
      volume = {125},
      number = {2},
      pages = {92 -96},
      doi = {http://dx.doi.org/10.1049/piee.1978.0026}
    }
    					
    Chen, W.-H. & Smith, C. Adaptive Coding of Monochrome and Color Images 1977 Communications, IEEE Transactions on
    Vol. 25 (11) , pp. 1285 - 1292  
    article    
    Abstract: An efficient adaptive encoding technique using a new implementation of the Fast Discrete Cosine Transform (FDCT) for bandwidth compression of monochrome and color images is described. Practical system application is attained by maintaining a balance between complexity of implementation and performance. FDCT sub-blocks are sorted into four classes according to level of image activity, measured by the total ac energy within each sub-block. Adaptivity is provided by distributing bits between classes, favoring higher levels of activity over lower levels. Excellent performance is demonstrated in terms of mean square error and direct comparison of original and reconstructed images. Results are presented for both noiseless and noisy transmission at a total rate of 1 bit and 0.5 bit per pixel for a monochrome image and for a total rate of 2 bits and 1 bit per pixel for a color image. In every case the total bit rate includes all overhead required for image reconstruction and bit protection.
    BibTeX:
    @article{1977_Chen,
      author = { Wen-Hsiung Chen and Smith, C.},
      title = {Adaptive Coding of Monochrome and Color Images},
      journal = {Communications, IEEE Transactions on},
      year = {1977},
      volume = {25},
      number = {11},
      pages = { 1285 - 1292}
    }
    					
    Imai, S.; Kitamura, T. & Takeya, H. A direct approximation technique of log magnitude response for digital filters 1977 Acoustics, Speech and Signal Processing, IEEE Transactions on
    Vol. 25 (2) , pp. 127 - 133  
    article    
    Abstract: A new direct approximation technique of log magnitude response for digital filters is presented in this paper. The facts that the log magnitude response of digital filters can be expanded into Fourier series and a fairly accurate cosine type log magnitude response can be realized by the elemental digital filter presented in this paper are used in the present technique. The system functions obtained by this method provide the best mean-square approximation to an arbitrarily prescribed log magnitude response. The resulting digital filters are realized in the cascade form of the elemental digital filters, and they give relatively low coefficient sensitivity. The elemental filter is recursive but its form is very simple. Its coefficients are easily obtained by the cepstrum of the impulse response which is the Fourier transform of the desired log magnitude response. This method is very powerful in the realization of digital filters for speech synthesis filters with complicated log magnitude responses.
    BibTeX:
    @article{1977_Imai,
      author = {Imai, S. and Kitamura, T. and Takeya, H.},
      title = {A direct approximation technique of log magnitude response for digital filters},
      journal = {Acoustics, Speech and Signal Processing, IEEE Transactions on},
      year = {1977},
      volume = {25},
      number = {2},
      pages = { 127 - 133}
    }
    					
    Warren, B.E. & Averbach, B.L. The Effect of Cold #x2010;Work Distortion on X #x2010;Ray Patterns 1950 Journal of Applied Physics
    Vol. 21 (6) , pp. 595 -599  
    article DOI    
    Abstract: With modern experimental technique, it is possible to measure a peak shape with sufficient accuracy to justify an interpretation based on the precise shape of the reflection. The corrected shape is represented by a cosine Fourier series and a set of An coefficients determined. A plot of the An coefficients vs. n will distinguish between distortion and particle size broadening. From the An coefficients, root mean square values of strain averaged over lengths na3 are obtained. The decrease in these values for increasing length na3 is a direct indication of the non #x2010;uniform nature of the strains in cold #x2010;worked metal. By measuring several orders of a given plane, it is theoretically possible to obtain a distribution function of the strains directly from a Fourier transform of the An coefficients.
    BibTeX:
    @article{1950_Warren,
      author = {Warren, B. E. and Averbach, B. L.},
      title = {The Effect of Cold #x2010;Work Distortion on X #x2010;Ray Patterns},
      journal = {Journal of Applied Physics},
      year = {1950},
      volume = {21},
      number = {6},
      pages = {595 -599},
      doi = {http://dx.doi.org/10.1063/1.1699713}
    }
    					

    Created by JabRef on 15/03/2010.

    Valid HTML 4.01 Transitional      Valid CSS!           Valid HTML 4.01 Transitional      Valid CSS!

    © Stéphane BURIGNAT