| Title: | Mathematics at DEC |
| Moderator: | RUSURE::EDP |
| Created: | Mon Feb 03 1986 |
| Last Modified: | Fri Jun 06 1997 |
| Last Successful Update: | Fri Jun 06 1997 |
| Number of topics: | 2083 |
| Total number of notes: | 14613 |
Is the Lempel-Ziv data compression algorithm noted in 336 suitable for
compressing binary images? I ask this becasue 334.last mentions that
Lempel-Ziv was derived from a more complicated algorithm for binary
images. Would Lempel-Ziv it fare well in a system code environment with
a limited stack space? If not, are there simpler algorithms to
implement that might not have the same reduction capabilitites but
would peform useful reduction.
Our application is an image for the EEPROM on our CPU board. The
EEPROM image and its blasting code are piggy-backed onto the console
code in one ROM. Space is tight, and we need a way to reduce the
EEPROM image significantly, without making a great increase in the
blasting code, which is about 200 (hex) bytes. Currently the entire
EEPROM image is blasted into the ROM.
The image is divided into several sections. Each section starts out
with a normal pattern of data, followed by fill characters (0) until
the beginning of the next section.
I could hard code the data reduction to blast the varying portion of
each section, loop to blast the zeroes to the end of that section and
repeat this for each section. That information is static and
reasonably easy to code into the blasting routine. Before doing the
brute force method, however, I wanted to inquire if a data reduction
algorithm would make more sense.
Thanks.
Charles Norton
Rigel System Integration/Console Development
Low-End Mid-Range VAX Systems Business | BXB2-2/F02 | DTN: 293-5487
| T.R | Title | User | Personal Name | Date | Lines |
|---|---|---|---|---|---|
| 924.1 | Try it and see. | AKQJ10::YARBROUGH | I prefer Pi | Fri Sep 02 1988 08:40 | 22 |
> Is the Lempel-Ziv data compression algorithm noted in 336 suitable for > compressing binary images? Yes and no. It depends on what's in the binary image. I have seen 40% reduction of large .exe files and BACKUP save sets, using the L-Z algorithm. I have also seen binary files grow LARGER when a compression algorithm is applied. This behavior is to be expected: if you start with a file that has already been compressed, or that has, for other reasons, very low redundancy to begin with, you cannot expect further compression, and in fact slight growth in size can be predicted. This effect is independent of the compression algorithm chosen. So what you need to do is to run the algorithm on your image and see how much it compresses. If you find that it compresses satisfactorily, then measure how long it takes to decompress and see if you can live with that startup time in your application. LZCOMP in the ToolShed has the L-Z algorithm; give it a try. Lynn Yarbrough | |||||
| 924.2 | Reduction and Stack Space Only Constraints/Tnx | WONDER::NORTON | Charles McKinley Norton | MSB | Fri Sep 02 1988 11:26 | 6 |
Thanks.
Fortunately, time is not an issue. It's only if the algorithm reduces
that counts. Thanks for the pointers.
CMN
| |||||