Multi-threaded encoding and decoding.
This crate provides a near drop in replacement for Write
that has will compress chunks of data in parallel and write
to an underlying writer in the same order that the bytes were handed to the writer. This allows for much faster
compression of data.
Additionally, this provides multi-threaded decompressors for Mgzip and BGZF formats.
By default gzp
has the deflate_default
and libdeflate
features enabled which brings in the best performing zlib
implementation as the backend for flate2
as well as libdeflater
for the block gzip formats.
toml
[dependencies]
gzp = { version = "*" }
Zlib
format will not be available.toml
[dependencies]
gzp = { version = "*", default-features = false, features = ["deflate_rust"] }
toml
[dependencies]
gzp = { version = "*", default-features = false, features = ["snap_default"] }
Simple example
```rust use std::{env, fs::File, io::Write};
use gzp::{deflate::Gzip, ZBuilder, ZWriter};
fn main() {
let mut writer = vec![];
// ZBuilder will return a trait object that transparent over ParZ
or SyncZ
let mut parz = ZBuilder::
An updated version of pgz.
```rust use gzp::{ ZWriter, deflate::Mgzip, par::{compress::{ParCompress, ParCompressBuilder}} }; use std::io::{Read, Write};
fn main() { let chunksize = 64 * (1 << 10) * 2;
let stdout = std::io::stdout();
let mut writer: ParCompress<Mgzip> = ParCompressBuilder::new().from_writer(stdout);
let stdin = std::io::stdin();
let mut stdin = stdin.lock();
let mut buffer = Vec::with_capacity(chunksize);
loop {
let mut limit = (&mut stdin).take(chunksize as u64);
limit.read_to_end(&mut buffer).unwrap();
if buffer.is_empty() {
break;
}
writer.write_all(&buffer).unwrap();
buffer.clear();
}
writer.finish().unwrap();
} ```
Same thing but using Snappy instead.
```rust use gzp::{parz::{ParZ, ParZBuilder}, snap::Snap}; use std::io::{Read, Write};
fn main() { let chunksize = 64 * (1 << 10) * 2;
let stdout = std::io::stdout();
let mut writer: ParZ<Snap> = ParZBuilder::new().from_writer(stdout);
let stdin = std::io::stdin();
let mut stdin = stdin.lock();
let mut buffer = Vec::with_capacity(chunksize);
loop {
let mut limit = (&mut stdin).take(chunksize as u64);
limit.read_to_end(&mut buffer).unwrap();
if buffer.is_empty() {
break;
}
writer.write_all(&buffer).unwrap();
buffer.clear();
}
writer.finish().unwrap();
} ```
pigz
, including
implementation details for some functions.PRs are very welcome! Please run tests locally and ensure they are passing. May tests are ignored in CI because the CI instances don't have enough threads to test them / are too slow.
bash
cargo test --all-features && cargo test --all-features -- --ignored
Note that tests will take 30-60s.
All benchmarks were run on the file in ./bench-data/shakespeare.txt
catted together 100 times which creates a rough
550Mb file.
The primary benchmark takeaway is that compression time decreases proportionately to the number of threads used.