Efficient bitstream compression using historical-based grouping and size-based coding rearrangement

Published date
Resource type
Call no.
Other identifier(s)
Copyrighted date
File type
10 pages
Other title(s)
Other Contributor(s)
IEEE Celebrating 125 Years of Engineering the Future (November 2009)
Degree name
Degree level
Degree discipline
Degree department
Degree grantor
Data storage becomes one of the important factors for most of large enterprises in increasing the cost, chip area, and dissipation. Large enterprises usually create redundancy of repeated digital messages through communication and documentary. In order to minimize these kinds of redundant bitstream, this paper presents a novel algorithm for lossless data compression by developing Historical-Based grouping and Sized-Based Coding Rearrangement approaches to reduce the code size used in compress redundant data flows. The Historical-Based grouping technique improves efficiency of data compression by implementing knowledge-based dictionary that is adaptive and can analyze the existing text coding with the incoming text by considering frequency of same text that has been compresses. In this case, if a compressed text besides that particular compressed text is the same, then this algorithm will group them into a "phase" by using just a particular compression code or symbol to identify that phase. This technique can also efficiently reduce compressed data size by extending a particular compression text length from "word" to "phase", from "phase" to "sentence", from "sentence" to "paragraph", indefinitely depended on repetition of the same binary patterns and amount of available memory allocated to store it. This technique is also able to group various groups of duplicated texts into a single code if those text groups are located in the same text area, same length of space between groups of texts, and set special symbols to link among each group of texts if space lengths are not the same. The objective of this technique is to analyze the historical coded data to minimize the coded size either to group or not to group the same binary patterns based on their frequencies. However, coding of the same binary patterns by using only the knowledge-based dictionary approaches might not be able to efficiently minimize the compressed patterns since frequencies of specific patterns are different. The data that being encoded should be frequency-based analyzed and grouped as long as possible by using historical-Based Grouping approach, then the compressed output should be rearrange based on their coding sizes and text lengths in which the result of total bit size should be concerned and minimized. Normally, longest or mostly-use coded text should be matched with the shortest code or symbol to ensure that every codes are efficiently utilized and total memory space requirement for that particular file is minimized.
Table of contents
Spatial Coverage
Access rights
Rights holder(s)
View External Resources