A Comparison between English and. Arabic Text Compression

Similar documents
EE-575 INFORMATION THEORY - SEM 092

Engineering Mathematics II Lecture 16 Compression

A Novel Image Compression Technique using Simple Arithmetic Addition

Data Compression. Media Signal Processing, Presentation 2. Presented By: Jahanzeb Farooq Michael Osadebey

So, what is data compression, and why do we need it?

CS 335 Graphics and Multimedia. Image Compression

Repetition 1st lecture

15 Data Compression 2014/9/21. Objectives After studying this chapter, the student should be able to: 15-1 LOSSLESS COMPRESSION

Image Compression for Mobile Devices using Prediction and Direct Coding Approach

Image coding and compression

Fundamentals of Multimedia. Lecture 5 Lossless Data Compression Variable Length Coding

IMAGE COMPRESSION TECHNIQUES

Data compression with Huffman and LZW

A New Compression Method Strictly for English Textual Data

Compression. storage medium/ communications network. For the purpose of this lecture, we observe the following constraints:

Encoding. A thesis submitted to the Graduate School of University of Cincinnati in

Chapter 1. Digital Data Representation and Communication. Part 2

AN ANALYTICAL STUDY OF LOSSY COMPRESSION TECHINIQUES ON CONTINUOUS TONE GRAPHICAL IMAGES

Noise Reduction in Data Communication Using Compression Technique

Digital Image Processing

7: Image Compression

EE67I Multimedia Communication Systems Lecture 4

A Comprehensive Review of Data Compression Techniques

Analysis of Parallelization Effects on Textual Data Compression

Department of electronics and telecommunication, J.D.I.E.T.Yavatmal, India 2

Multimedia Systems. Part 20. Mahdi Vasighi

Multimedia Networking ECE 599

REVIEW ON IMAGE COMPRESSION TECHNIQUES AND ADVANTAGES OF IMAGE COMPRESSION

A Compression Technique Based On Optimality Of LZW Code (OLZW)

Compression; Error detection & correction

Ch. 2: Compression Basics Multimedia Systems

DEFLATE COMPRESSION ALGORITHM

Compression; Error detection & correction

IMAGE COMPRESSION. Image Compression. Why? Reducing transportation times Reducing file size. A two way event - compression and decompression

CIS 121 Data Structures and Algorithms with Java Spring 2018

THE RELATIVE EFFICIENCY OF DATA COMPRESSION BY LZW AND LZSS

Comparative Study between Various Algorithms of Data Compression Techniques

Lossless Compression Algorithms

OPTIMIZATION OF LZW (LEMPEL-ZIV-WELCH) ALGORITHM TO REDUCE TIME COMPLEXITY FOR DICTIONARY CREATION IN ENCODING AND DECODING

JPEG. Table of Contents. Page 1 of 4

Introduction to Data Compression

Keywords Data compression, Lossless data compression technique, Huffman Coding, Arithmetic coding etc.

HARDWARE IMPLEMENTATION OF LOSSLESS LZMA DATA COMPRESSION ALGORITHM

LZW Compression. Ramana Kumar Kundella. Indiana State University December 13, 2014

Keywords : FAX (Fascimile), GIF, CCITT, ITU, MH (Modified haffman, MR (Modified read), MMR (Modified Modified read).

An Efficient Technique for Text Compression

Research Article Does an Arithmetic Coding Followed by Run-length Coding Enhance the Compression Ratio?

Image Coding and Compression

Image compression. Stefano Ferrari. Università degli Studi di Milano Methods for Image Processing. academic year

WIRE/WIRELESS SENSOR NETWORKS USING K-RLE ALGORITHM FOR A LOW POWER DATA COMPRESSION

Text Data Compression and Decompression Using Modified Deflate Algorithm

Data Compression. An overview of Compression. Multimedia Systems and Applications. Binary Image Compression. Binary Image Compression

Study of LZ77 and LZ78 Data Compression Techniques

S 1. Evaluation of Fast-LZ Compressors for Compacting High-Bandwidth but Redundant Streams from FPGA Data Sources

Journal of Computer Engineering and Technology (IJCET), ISSN (Print), International Journal of Computer Engineering

Entropy Coding. - to shorten the average code length by assigning shorter codes to more probable symbols => Morse-, Huffman-, Arithmetic Code

IMAGE PROCESSING (RRY025) LECTURE 13 IMAGE COMPRESSION - I

International Journal of Advanced Research in Computer Science and Software Engineering

A Research Paper on Lossless Data Compression Techniques

Dynamic with Dictionary Technique for Arabic Text Compression

Image Compression. CS 6640 School of Computing University of Utah

A Methodology to Detect Most Effective Compression Technique Based on Time Complexity Cloud Migration for High Image Data Load

Lecture 5: Compression I. This Week s Schedule

VC 12/13 T16 Video Compression

CS 493: Algorithms for Massive Data Sets Dictionary-based compression February 14, 2002 Scribe: Tony Wirth LZ77

Enhancing the Compression Ratio of the HCDC Text Compression Algorithm

Data Representation. Types of data: Numbers Text Audio Images & Graphics Video

Data and information. Image Codning and Compression. Image compression and decompression. Definitions. Images can contain three types of redundancy

Design and Implementation of FPGA- based Systolic Array for LZ Data Compression

Some Algebraic (n, n)-secret Image Sharing Schemes

Data Storage. Slides derived from those available on the web site of the book: Computer Science: An Overview, 11 th Edition, by J.

A QUAD-TREE DECOMPOSITION APPROACH TO CARTOON IMAGE COMPRESSION. Yi-Chen Tsai, Ming-Sui Lee, Meiyin Shen and C.-C. Jay Kuo

IMAGE COMPRESSION- I. Week VIII Feb /25/2003 Image Compression-I 1

Optimized Compression and Decompression Software

Comparative Study of Dictionary based Compression Algorithms on Text Data

FPGA based Data Compression using Dictionary based LZW Algorithm

Overview. Last Lecture. This Lecture. Next Lecture. Data Transmission. Data Compression Source: Lecture notes

Comparative data compression techniques and multi-compression results

Topic 5 Image Compression

Improving LZW Image Compression

Image Compression Algorithm and JPEG Standard

1.6 Graphics Packages

Designing and Building an Automatic Information Retrieval System for Handling the Arabic Data

CS/COE 1501

David Rappaport School of Computing Queen s University CANADA. Copyright, 1996 Dale Carnegie & Associates, Inc.

A Image Comparative Study using DCT, Fast Fourier, Wavelet Transforms and Huffman Algorithm

Category: Informational May DEFLATE Compressed Data Format Specification version 1.3

FRACTAL IMAGE COMPRESSION OF GRAYSCALE AND RGB IMAGES USING DCT WITH QUADTREE DECOMPOSITION AND HUFFMAN CODING. Moheb R. Girgis and Mohammed M.

STUDY OF VARIOUS DATA COMPRESSION TOOLS

Volume 2, Issue 9, September 2014 ISSN

7.5 Dictionary-based Coding

ECE 533 Digital Image Processing- Fall Group Project Embedded Image coding using zero-trees of Wavelet Transform

An Implementation on Pattern Creation and Fixed Length Huffman s Compression Techniques for Medical Images

An Advanced Text Encryption & Compression System Based on ASCII Values & Arithmetic Encoding to Improve Data Security

Medical Image Compression using DCT and DWT Techniques

Enhancing Text Compression Method Using Information Source Indexing

COMPRESSION OF SMALL TEXT FILES

Information Technology Department, PCCOE-Pimpri Chinchwad, College of Engineering, Pune, Maharashtra, India 2

Implementation of Robust Compression Technique using LZ77 Algorithm on Tensilica s Xtensa Processor

G64PMM - Lecture 3.2. Analogue vs Digital. Analogue Media. Graphics & Still Image Representation

Transcription:

Contemporary Engineering Sciences, Vol. 6, 2013, no. 3, 111-119 HIKARI Ltd, www.m-hikari.com A Comparison between English and Arabic Text Compression Ziad M. Alasmer, Bilal M. Zahran, Belal A. Ayyoub, Monther A. Kanan Department of Computer Science Al-Balqa Applied University, Amman, Jordan ziad_alasmer@yahoo.com, zahranb@ bau.edu.jo, belal_ayyoub@hotmail.com, kananmonther@yahoo.com Abdelaziz I. Hammouri Department of Computer Information System Al-Balqa Applied University, Al-Salt, Jordan aziz@bau.edu.jo Jafar Ababneh Department of Computer Network System The world Islamic Sciences and Education University, Amman, Jordan jafar.ababneh@wise.edu.jo Copyright 2013 Ziad M. Alasmer et al. This is an open access article distributed under the Creative Commons Attribution License, which permits unrestricted use, distribution, and reproduction in any medium, provided the original work is properly cited. Abstract A Comparison between applying two Techniques that compress document data in both languages Arabic and English is introduced. In order to compress the data document, two or more constituent's data documents in both languages are identified. The comparison takes to its consideration, for the first time to the best of our knowledge, the Arabic data compressing. The problem is solved using an efficient language that uses Borland C++ builder to ensure compression for any documents. Our numerical experiments show that Huffman technique can be better used for Arabic Documents. LZW algorithm is better to use for TIFF, GIF and English textual files. Keywords: Data compression, Huffman compression, LZW Compression

112 Ziad M. Alasmer et al 1 Introduction Data compression is the removal of redundant data. This, therefore, reduces the number of binary bits necessary to represent the information contained within that data<. Thus a compressor is made of at least two different tasks: predicting the probabilities of the input and generating codes from those probabilities, which is done with a model and a coder respectively [3]. A compressor can be either lossy or lossless. A lossless compressor makes files smaller by finding redundant patterns of data and then replacing them with tokens or other symbols that take up less space. With a lossless compressor and decompressor, the original and decompressed files are identical bit per bit. If an image is compressed using lossless compression, after decompressing, it will be an identical image. No data is lost or changed in any way. It s like a sponge you can squeeze it down, and when you let go it reverts to its original form. A Lossy compressor makes files smaller by removing ostensibly less important data from a file. This type, actually removes information in the process of squeezing the data. Individual lossy compression methods work only on specific kinds of images, and typically yield much smaller file sizes than lossless compression methods. Good lossy schemes drop out data in a very intelligent manner to minimize the noticeable effect of lost pixels. To do so, they start with assumptions about what kinds of data are most important. For instance, JPEG thinks that coarse tonal details are most important and fine color details have the least value [2, 6, 14, 15, 17]. We may want to compress different kinds of data such as text, data bases, binary programs, sound, image and video. In practice text compression and signal compression are distinguish about. This separation is done because data bases and binary programs have the same characteristic as text. Likewise sound, image and video are signals and thus share properties. In the other hand text and image data have nothing in common, and that's what they don't belong to the same group. Text documents can be introduced with different languages, in this research we focused on Arabic language, where Arabic is the mother language of millions of people all over the world. It is a highly inflected language, it has much richer morphology than English [19]. Among several sources that discussed the difficulty of Arabic text classification, the following are some of the challenges in Arabic text classification [5]: Arabic language differs syntactically, morphologically and semantically from other Indo-European languages. Compared to English, Arabic language is sparser, which means that English words repeated more often than Arabic words for the same text length. In written Arabic, most letters take many forms of writing. Moreover, there is a punctuation associated with some letters that may change the meaning of two identical words. The omission of diacritics (vowels) in written Arabic altashkiil. Comparing to English roots, Arabic roots are more complex.

A comparison between English and Arabic text compression 113 2 Huffman Compression Also known as Huffman encoding was invented by David Huffman back in 1952. It is one of many compression techniques in use today and is used as part of a number of other compression schemes, like CCITT and JPEG. One of the main benefits of Huffman Compression is how easy it is to understand and implement yet still gets a decent compression ratio on average files. The Huffman encoding assumed data files consist of some byte values that occur more frequently than other byte values in the same file. This is very true for text files and most raw gfx images, as well as EXE and COM file code segments. Huffman encoding is a technique that takes a set of symbols, like the letters in a text file, and analyzes them to determine the frequency of each symbol. It then uses the fewest possible bits to represent the most frequently occurring symbols. For instance, is the most common letter in Standard English text[8]. Huffman encoding might represent it in as few as 2 bits (1 followed by 0) instead of the 8 bits needed to signal in ASCII, which is used to store and transmit virtually all text on and between computers. On the other hand, a little used letter like x or y might require 11 or 12 bits to represent[11, 16]. 2.1 Huffman Compression Algorithm The algorithm steps are: 1. For each byte value within the file, calculate the number of occurrences, and build a Frequency Table. 2. Build a binary tree which represents the bytes of the file. 3. In the binary tree, the highest occurrence must be in the most left, the lowest occurrence must be in the most right 4. To scan the tree: for each byte when you go left fill zero, for each right fill one; so the byte can be represented by one bit or 2 bits or instead of 8 bits. By analyzing the algorithm, it can be noticed that Huffman encoding builds a "Frequency Table" for each byte value within a file. With the frequency table the algorithm can then build the "Huffman Tree" from the frequency table. The purpose of the tree is to associate each byte value with a bit string of variable length. The more frequently used characters get shorter bit strings, while the less frequent characters get longer bit strings. Thusly the data file may be compressed. To compress the file, the Huffman algorithm reads the file a second time, converting each byte value into the bit string assigned to it by the Huffman Tree and then writing the bit string to a new file[1, 3, 11, 13, 15, 16]. 3 LZW Compression (Abraham Lempel, Jakob Ziv and Terry Welch) LZW is named after Abraham Lempel, Jakob Ziv and Terry Welch, the scientists

114 Ziad M. Alasmer et al who developed this compression algorithm. It is a lossless 'dictionary based' compression algorithm. Dictionary based algorithms scan a file for sequences of data that occur more than once. These sequences are then stored in a dictionary and within the compressed file, references are put where ever repetitive data occurred. Their first algorithm was published in 1977, hence its name: LZ77. This compression algorithm maintains its dictionary within the data themselves[21]. Suppose the following string of text to be compressed: the quick brown fox jumps over the lazy dog. The word 'the' occurs twice in the file so the data can be compressed like this: the quick brown fox jumps over << lazy dog. in which << is a pointer to the first 4 characters in the string. In 1978, Lempel and Ziv published a second paper outlining a similar algorithm that is now referred to as LZ78. This algorithm maintains a separate dictionary. Suppose the following string of text to be compressed again: the quick brown fox jumps over the lazy dog. The word 'the' occurs twice in the file so this string is put in an index that is added to the compressed file and this entry is referred to as *. The data then look like this: * quick brown fox jumps over * lazy dog. In 1984, Terry Welch was working on a compression algorithm for high-performance disk controllers. He developed a rather simple algorithm that was based on the LZ78 algorithm and that is now called LZW[18, 20, 21]. 3.1 LZW Compression Algorithm LZW compression replaces strings of characters with single codes. It does not do any analysis of the incoming text. Instead, it just adds every new string of characters it sees to a table of strings. Compression occurs when a single code is output instead of a string of characters. The code that the LZW algorithm outputs can be of any arbitrary length, but it must have more bits in it than a single character. The first 256 codes (when using eight bit characters) are by default assigned to the standard character set. The remaining codes are assigned to strings as the algorithm proceeds. The algorithm below uses 12 bit codes for output codes. This means codes 0-255 refer to individual bytes, while codes 256-4095 refers to substrings [4, 9, 18]. 4 Compression ratio Compression ratio is used to determine how much a file has been compressed after applying a specific compression algorithm on it. Compression ratio is measured in several ways that will be discussed next, and compression ratios for examples in previous parts: (Huffman Algorithm) & (LZW Algorithm) parts will be calculated in this part. Compression ratio is also used to compare between different compressions algorithms when applied on the same file. A study to compare between Huffman and LZW algorithms using compression ratio is performed in part. Measuring the Compression Ratio:

A comparison between English and Arabic text compression 115 1. Bits Per Byte (bpb): bpb is the most used way of measuring the compression achieved by a program. It is computed as: (compressed length / original length)*8 (1) If a 400 bytes file is compressed down to 100 bytes, then bpb ratio is: (100/400)*8 = 2bpb, which means that only 2 bits are needed to represent one byte. This measure is accurate enough, for example: (47/134)*8 = 2.805970149254 bpb, though usually three digits are used, like 2.806 bpb (no rules concerning rounding). Note that when the expected compression of a given algorithm is known then the supposed length of the output can be known: (input length / 8) * bpb = output length. This kind of measurement is recommended to be used. 2. Percentage Compression Ratio (%): Compression ratio can also be measured using % as: (Compressed length / original length)*100 (2) For example if a file with a size of 400 bytes is compressed down to 100 bytes, the ratio will be (100/400)*100 = 25% so the output compressed file is only 25% of the original. However there's other method: (1-(compressed length/original length))*100 (3) In this case the ratio is 75% meaning that 75% of the original file is subtracted. In both cases the compression is the same, but the ratios are different. Here, one need to determine which form of the percentage compression ratio is used ((2) or (3)). Form (2) will be used in calculations during this part [7, 10, 12]. 5 Results We tested our algorithm on Arabic dataset, which has been in-house collected corpus from online Arabic newspapers archives, including Al-Jazeera, Al-Hayat, Al-Ahram and Addostour as well as a few other specialized web sites. In this Arabic dataset, each document was saved in a separate file within the directory for the corresponding category, i.e., the documents in this dataset are single-labeled. The code was written using C++ builder, this language uses C++ codes to write program, added to that it is a GUI (Graphical User Interface) programming language. C++ was the choice because of the facilities and data structures it offers for writing programs such as programs for Huffman and LZW algorithms Comparison between LZW and Huffman: Under the title (Group 1 Results) both LZW and Huffman will be used to compress and decompress different types of files, tries and results will be represented in a table, then figured in a chart to compare the efficiency of both programs in compressing and decompressing different types of files, conclusions and discussions are given at the end. Study the following table and chart to see the results.

116 Ziad M. Alasmer et al Table 1: Comparison between LZW and Huffman File Name Input File Size Output File Size/LZW Output File Size/Huffman Compress Ratio/LZW Compress Ratio/Huffman Example1. doc 68096 30580 29433 55% 57% Example2. doc 58880 23814 23640 60% 66% Example3. doc 83968 48984 46876 42% 45% Example4. Doc 20480 2530 4836 88% 76% Example5. Doc 27648 8222 10921 70% 60% Example6. Doc 57856 30993 27163 46% 53% Example7. Doc 87552 54229 47101 38% 46% Example8. Doc 48128 23631 21600 51% 55% Example9. Doc 79360 30363 32416 62% 59% Example10.Doc 68096 30581 29433 55% 57% Pict3.bmp 1440054 193888 276506 87% 81% Pict4.bmp 1440054 100338 282824 93% 80% Pict5.bmp 1440054 461637 318178 68% 78% Pict6.bmp 1365318 371601 366830 73% 73% Inprise.gif 4654 6634 5073-43% -9% Baby.jpg 26183 35367 26487-35% -1% Cake.jpg 23036 32457 23479-41% -2% Candels.jpg 17639 23230 17885-32% -1% Class.jpg 5851 6764 6035-16% -3% Earth.jpg 9370 12955 9811-38% -5% Figure 1 shows the results of using the program in compressing different types of files. In the chart, the dark line curve represents the input files sizes, the grey curve represents the output files sizes when compressed using LZW and the white curve represents the output file sizes when compressed using Huffman. Figure 1:Comparison between LZW and Huffman compression ratios. From the table and the chart above, the following discussions can be listed: LZW and Huffman give nearly results when used for compressing document or text files, as appears in the table and in the chart. The difference in the

A comparison between English and Arabic text compression 117 compression ratio is related to the different mechanisms of both in the compression process; which depends in LZW on replacing strings of characters with single codes, where in Huffman depends on representing individual characters with bit sequences. When LZW and Huffman are used to compress a binary file (all of its contents either 1 or 0), LZW gives a better compression ratio than Huffman. If you tried for example to compress one line of binary (00100101010111001101001010101110101.) using LZW, you will arrive to a stage in which 5 or 6 consecutive binary digits are represented by a single new code (9 bits), while in Huffman you will represent every individual binary digit with a bit sequence of 2 bits, so in Huffman the 5 or 6 binary digits which were represented in LZW by 9 bits are represented now with 10 or 12 bits; this decreases the compression ratio in the case of Huffman. LZW and Huffman are used in compressing bmp files; bmp files contain images, in which each dot in the image is represented by a byte, as appears in the chart for compressing bmp files, the results are somehow different. LZW seems to be better in compressing bmp files than Huffman; since it replaces sets of dots (instead of strings of characters in text files) with single codes; resulting in new codes that are useful when the dots that consists the image are repeated, while in Huffman, individual dots in the image are represented by bit sequences of a length depending on its probabilities. Because of the large different dots representing the image, the binary tree to be built is large, so the length of bit sequences which represents the individual dots increases, resulting in a less compression ratio compared to LZW compression ratio. When LZW or Huffman is used to compress a file of type gif or type jpg, you will notice as in the table and in the chart that the compressed file size is larger than the original file size; this is due to being the images of these files are already compressed, so when compressed using LZW the number of the new output codes will increase, resulting in a file size larger than the original, while in Huffman the size of the binary tree built increases because of the less of probabilities, resulting in longer bit sequences that represent the individual dots of the image, so the compressed file size will be larger than the original. But because of being the new output code in LZW represented by 9 bits, while in Huffman the individual dot is represented with bits less than 9, this makes the resulting file size after compression in LZW larger than that in Huffman. Decompression operation is the opposite operation for compression; so the results will be the same as in compression.

118 Ziad M. Alasmer et al Table 2:Arabic compression File Name Input File Size Output File Size/LZW Output File Size /Huffman Compress Ratio /LZW Compress Ratio /Huffman DOC1 1007B 832B 655B 18% 35% DOC2 965B 807B 619B 16% 36% DOC38 1181B 931B 744B 22% 38% DOC71 762B 670B 513B 13% 33% DOC20 892B 745B 578B 17% 36% DOC66 705B 631B 479B 11% 33% As we see in the table 2 and when applying the two methods with same Document sizes in Arabic language, we find that Huffman better than LZW in Arabic Language so we should enhance LZW method to give better results for Arabic Documents this can be done by change the dictionary method to be suitable with Arabic. 6 Conclusion A comparison in between Huffman and LZW Techniques on has been applied on Arabic and English documents with identical sizes in both Arabic and English, we found that Huffman get better and efficient results in Arabic document and LZW techniques is better in English documents specially when it is converted to a binary file. It seems that LZW need to be improved, because it is based on English language, also Huffman need to be developed so it can suite other language. References [1] O. C. L. Au and J. Zhou, "System and method for encoding data based on a compression technique with security features," ed: Google Patents, 2011. [2] M. Deering, "Geometry compression," in Proceedings of the 22nd annual conference on Computer graphics and interactive techniques, 1995, pp. 13-20. [3] C. Delfs, et al., "Dictionary-based compression and decompression," ed: Google Patents, 2007. [4] J. Dvorský, et al., "Word-based compression methods and indexing for text retrieval systems," in Advances in Databases and Information Systems, 1999, pp. 76-84. [5] A. Farghaly and K. Shaalan, "Arabic natural language processing: Challenges and solutions," ACM Transactions on Asian Language Information Processing (TALIP), vol. 8, p. 14, 2009.

A comparison between English and Arabic text compression 119 [6] C. J. Goosmann, "Data Compression In A Mainframe World (Less Is More)," in CMG-CONFERENCE-, 1995, pp. 74-84. [7] E. Y. Hamid and Z. I. Kawasaki, "Wavelet-based data compression of power system disturbances using the minimum description length criterion," Power Delivery, IEEE Transactions on, vol. 17, pp. 460-466, 2002. [8] D. A. Huffman, "A method for the construction of minimum-redundancy codes," Proceedings of the IRE, vol. 40, pp. 1098-1101, 1952. [9] Z. Li and S. Hauck, "Configuration compression for virtex FPGAs," in Field-Programmable Custom Computing Machines, 2001. FCCM'01. The 9th Annual IEEE Symposium on, 2001, pp. 147-159. [10] C. H. Lin, et al., "LZW-based code compression for VLIW embedded systems," in Design, Automation and Test in Europe Conference and Exhibition, 2004. Proceedings, 2004, pp. 76-81. [11] M. Nelson and J. L. Gailly, "The data compression book 2nd edition," M & T Books, New York, NY, 1995. [12] M. Nourani and M. H. Tehranipour, "RL-Huffman encoding for test compression and power reduction in scan applications," ACM Transactions on Design Automation of Electronic Systems (TODAES), vol. 10, pp. 91-115, 2005. [13] B. E. Ross, "Method and system for compressing publication documents in a computer system by selectively eliminating redundancy from a hierarchy of constituent data structures," ed: Google Patents, 1999. [14] D. Salomon, A concise introduction to data compression: Springer, 2008. [15] D. Salomon, "Data compression," Handbook of massive data sets, pp. 245-309, 2002. [16] D. Salomon, A guide to data compression methods vol. 1: Springer, 2002. [17] E. L. Schwartz and A. Zandi, "Reversible DCT for lossless-lossy compression," ed: Google Patents, 2000. [18] D. Sculley and C. E. Brodley, "Compression and machine learning: A new perspective on feature space vectors," in Data Compression Conference, 2006. DCC 2006. Proceedings, 2006, pp. 332-341. [19] M. M. Syiam, et al., "An intelligent system for Arabic text categorization," International Journal of Intelligent Computing and Information Sciences, vol. 6, pp. 1-19, 2006. [20] F. G. Wolff and C. Papachristou, "Multiscan-based test compression and hardware decompression using LZ77," in Test Conference, 2002. Proceedings. International, 2002, pp. 331-339. [21] S. Yadav and V. Gupta, "A 4-D Sequential Multispectral Lossless Images Compression Over Changed Data Using LZW Techniques," International Journal of Engineering Research and Applications, vol. 2, 2012. Received: February 9, 2013