我们正在开发一个程序,我们需要刷新(强制压缩和发送数据)GZIPOutputStream。问题是,GZIPOutputStream的flush方法不能按预期工作(强制压缩和发送数据),而是Stream等待更多数据进行有效的数据压缩。
当你调用完成后,数据被压缩并通过输出流发送,但GZIPOutputStream(不是底层流)将被关闭,所以我们不能写更多的数据,直到我们创建一个新的GZIPOutputStream,这需要时间和性能。
希望任何人都可以提供帮助。
最好的问候。
答案 0 :(得分:10)
我还没有尝试过这个,在我们掌握Java 7之前,这个建议没有用,但是GZIPOutputStream
的{{3}}方法的文档继承自{{1}依赖于在构造时指定的刷新模式 flush()
(与DeflaterOutputStream
相关)来决定是否刷新要压缩的待处理数据。 Deflater#SYNC_FLUSH
在施工时也接受syncFlush
参数。
听起来你想要使用the syncFlush
argument或甚至Deflator#SYNC_FLUSH
,但是,在深入挖掘这一点之前,首先尝试使用Deflater#FULL_FLUSH
或the two-argument并通过GZIPOutputStream
参数的true
。这将激活你想要的冲洗行为。
答案 1 :(得分:9)
我没有找到另一个工作的答案。它仍然拒绝刷新,因为GZIPOutputStream正在使用的本机代码保留在数据上。
值得庆幸的是,我发现有人在Apache Tomcat项目中实现了FlushableGZIPOutputStream。这是神奇的部分:
@Override
public synchronized void flush() throws IOException {
if (hasLastByte) {
// - do not allow the gzip header to be flushed on its own
// - do not do anything if there is no data to send
// trick the deflater to flush
/**
* Now this is tricky: We force the Deflater to flush its data by
* switching compression level. As yet, a perplexingly simple workaround
* for
* http://developer.java.sun.com/developer/bugParade/bugs/4255743.html
*/
if (!def.finished()) {
def.setLevel(Deflater.NO_COMPRESSION);
flushLastByte();
flagReenableCompression = true;
}
}
out.flush();
}
你可以在这个jar中找到整个类(如果你使用Maven):
<dependency>
<groupId>org.apache.tomcat</groupId>
<artifactId>tomcat-coyote</artifactId>
<version>7.0.8</version>
</dependency>
或者只是去抓取源代码FlushableGZIPOutputStream.java
它是在Apache-2.0许可下发布的。
答案 2 :(得分:1)
Bug ID 4813885处理此问题。 2006年9月9日提交的“DamonHD”评论(大约是bug报告的一半)包含FlushableGZIPOutputStream
的一个例子,他在Jazzlib's net.sf.jazzlib.DeflaterOutputStream
之上建立了/**
* Substitute for GZIPOutputStream that maximises compression and has a usable
* flush(). This is also more careful about its output writes for efficiency,
* and indeed buffers them to minimise the number of write()s downstream which
* is especially useful where each write() has a cost such as an OS call, a disc
* write, or a network packet.
*/
public class FlushableGZIPOutputStream extends net.sf.jazzlib.DeflaterOutputStream {
private final CRC32 crc = new CRC32();
private final static int GZIP_MAGIC = 0x8b1f;
private final OutputStream os;
/** Set when input has arrived and not yet been compressed and flushed downstream. */
private boolean somethingWritten;
public FlushableGZIPOutputStream(final OutputStream os) throws IOException {
this(os, 8192);
}
public FlushableGZIPOutputStream(final OutputStream os, final int bufsize) throws IOException {
super(new FilterOutputStream(new BufferedOutputStream(os, bufsize)) {
/** Suppress inappropriate/inefficient flush()es by DeflaterOutputStream. */
@Override
public void flush() {
}
}, new net.sf.jazzlib.Deflater(net.sf.jazzlib.Deflater.BEST_COMPRESSION, true));
this.os = os;
writeHeader();
crc.reset();
}
public synchronized void write(byte[] buf, int off, int len) throws IOException {
somethingWritten = true;
super.write(buf, off, len);
crc.update(buf, off, len);
}
/**
* Flush any accumulated input downstream in compressed form. We overcome
* some bugs/misfeatures here so that:
* <ul>
* <li>We won't allow the GZIP header to be flushed on its own without real compressed
* data in the same write downstream.
* <li>We ensure that any accumulated uncompressed data really is forced through the
* compressor.
* <li>We prevent spurious empty compressed blocks being produced from successive
* flush()es with no intervening new data.
* </ul>
*/
@Override
public synchronized void flush() throws IOException {
if (!somethingWritten) { return; }
// We call this to get def.flush() called,
// but suppress the (usually premature) out.flush() called internally.
super.flush();
// Since super.flush() seems to fail to reliably force output,
// possibly due to over-cautious def.needsInput() guard following def.flush(),
// we try to force the issue here by bypassing the guard.
int len;
while((len = def.deflate(buf, 0, buf.length)) > 0) {
out.write(buf, 0, len);
}
// Really flush the stream below us...
os.flush();
// Further flush()es ignored until more input data data written.
somethingWritten = false;
}
public synchronized void close() throws IOException {
if (!def.finished()) {
def.finish();
do {
int len = def.deflate(buf, 0, buf.length);
if (len <= 0) {
break;
}
out.write(buf, 0, len);
} while (!def.finished());
}
// Write trailer
out.write(generateTrailer());
out.close();
}
// ...
}
。
供参考,这是一个(重新格式化)提取物:
{{1}}
您可能会发现它很有用。
答案 3 :(得分:1)
Android
也有同样的问题。接受者答案不起作用,因为def.setLevel(Deflater.NO_COMPRESSION);
抛出异常。根据{{1}}方法,它会更改flush
的压缩级别。所以我想在编写数据之前应该调用压缩,但我不确定。
还有其他2个选项:
答案 4 :(得分:1)
此代码在我的应用程序中对我很有用。
public class StreamingGZIPOutputStream extends GZIPOutputStream {
public StreamingGZIPOutputStream(OutputStream out) throws IOException {
super(out);
}
@Override
protected void deflate() throws IOException {
// SYNC_FLUSH is the key here, because it causes writing to the output
// stream in a streaming manner instead of waiting until the entire
// contents of the response are known. for a large 1 MB json example
// this took the size from around 48k to around 50k, so the benefits
// of sending data to the client sooner seem to far outweigh the
// added data sent due to less efficient compression
int len = def.deflate(buf, 0, buf.length, Deflater.SYNC_FLUSH);
if (len > 0) {
out.write(buf, 0, len);
}
}
}
答案 5 :(得分:0)
正如@seh所说,这很好用:
ByteArrayOutputStream stream = new ByteArrayOutputStream();
// the second param need to be true
GZIPOutputStream gzip = new GZIPOutputStream(stream, true);
gzip.write( .. );
gzip.flush();
...
gzip.close()