mirror of https://github.com/python/cpython.git
2124 lines
51 KiB
C
2124 lines
51 KiB
C
/*
|
|
|
|
python-bz2 - python bz2 library interface
|
|
|
|
Copyright (c) 2002 Gustavo Niemeyer <niemeyer@conectiva.com>
|
|
Copyright (c) 2002 Python Software Foundation; All Rights Reserved
|
|
|
|
*/
|
|
|
|
#include "Python.h"
|
|
#include <stdio.h>
|
|
#include <bzlib.h>
|
|
#include "structmember.h"
|
|
|
|
#ifdef WITH_THREAD
|
|
#include "pythread.h"
|
|
#endif
|
|
|
|
static char __author__[] =
|
|
"The bz2 python module was written by:\n\
|
|
\n\
|
|
Gustavo Niemeyer <niemeyer@conectiva.com>\n\
|
|
";
|
|
|
|
#define BUF(v) PyString_AS_STRING((PyStringObject *)v)
|
|
|
|
#define MODE_CLOSED 0
|
|
#define MODE_READ 1
|
|
#define MODE_READ_EOF 2
|
|
#define MODE_WRITE 3
|
|
|
|
#define BZ2FileObject_Check(v) ((v)->ob_type == &BZ2File_Type)
|
|
|
|
#if SIZEOF_LONG >= 8
|
|
#define BZS_TOTAL_OUT(bzs) \
|
|
(((long)bzs->total_out_hi32 << 32) + bzs->total_out_lo32)
|
|
#elif SIZEOF_LONG_LONG >= 8
|
|
#define BZS_TOTAL_OUT(bzs) \
|
|
(((LONG_LONG)bzs->total_out_hi32 << 32) + bzs->total_out_lo32)
|
|
#else
|
|
#define BZS_TOTAL_OUT(bzs) \
|
|
bzs->total_out_lo32;
|
|
#endif
|
|
|
|
#ifdef WITH_THREAD
|
|
#define ACQUIRE_LOCK(obj) PyThread_acquire_lock(obj->lock, 1)
|
|
#define RELEASE_LOCK(obj) PyThread_release_lock(obj->lock)
|
|
#else
|
|
#define ACQUIRE_LOCK(obj)
|
|
#define RELEASE_LOCK(obj)
|
|
#endif
|
|
|
|
#ifdef WITH_UNIVERSAL_NEWLINES
|
|
/* Bits in f_newlinetypes */
|
|
#define NEWLINE_UNKNOWN 0 /* No newline seen, yet */
|
|
#define NEWLINE_CR 1 /* \r newline seen */
|
|
#define NEWLINE_LF 2 /* \n newline seen */
|
|
#define NEWLINE_CRLF 4 /* \r\n newline seen */
|
|
#endif
|
|
|
|
/* ===================================================================== */
|
|
/* Structure definitions. */
|
|
|
|
typedef struct {
|
|
PyFileObject file;
|
|
BZFILE *fp;
|
|
int mode;
|
|
long pos;
|
|
long size;
|
|
#ifdef WITH_THREAD
|
|
PyThread_type_lock lock;
|
|
#endif
|
|
} BZ2FileObject;
|
|
|
|
typedef struct {
|
|
PyObject_HEAD
|
|
bz_stream bzs;
|
|
int running;
|
|
#ifdef WITH_THREAD
|
|
PyThread_type_lock lock;
|
|
#endif
|
|
} BZ2CompObject;
|
|
|
|
typedef struct {
|
|
PyObject_HEAD
|
|
bz_stream bzs;
|
|
int running;
|
|
PyObject *unused_data;
|
|
#ifdef WITH_THREAD
|
|
PyThread_type_lock lock;
|
|
#endif
|
|
} BZ2DecompObject;
|
|
|
|
/* ===================================================================== */
|
|
/* Utility functions. */
|
|
|
|
static int
|
|
Util_CatchBZ2Error(int bzerror)
|
|
{
|
|
int ret = 0;
|
|
switch(bzerror) {
|
|
case BZ_OK:
|
|
case BZ_STREAM_END:
|
|
break;
|
|
|
|
case BZ_CONFIG_ERROR:
|
|
PyErr_SetString(PyExc_SystemError,
|
|
"the bz2 library was not compiled "
|
|
"correctly");
|
|
ret = 1;
|
|
break;
|
|
|
|
case BZ_PARAM_ERROR:
|
|
PyErr_SetString(PyExc_ValueError,
|
|
"the bz2 library has received wrong "
|
|
"parameters");
|
|
ret = 1;
|
|
break;
|
|
|
|
case BZ_MEM_ERROR:
|
|
PyErr_NoMemory();
|
|
ret = 1;
|
|
break;
|
|
|
|
case BZ_DATA_ERROR:
|
|
case BZ_DATA_ERROR_MAGIC:
|
|
PyErr_SetString(PyExc_IOError, "invalid data stream");
|
|
ret = 1;
|
|
break;
|
|
|
|
case BZ_IO_ERROR:
|
|
PyErr_SetString(PyExc_IOError, "unknown IO error");
|
|
ret = 1;
|
|
break;
|
|
|
|
case BZ_UNEXPECTED_EOF:
|
|
PyErr_SetString(PyExc_EOFError,
|
|
"compressed file ended before the "
|
|
"logical end-of-stream was detected");
|
|
ret = 1;
|
|
break;
|
|
|
|
case BZ_SEQUENCE_ERROR:
|
|
PyErr_SetString(PyExc_RuntimeError,
|
|
"wrong sequence of bz2 library "
|
|
"commands used");
|
|
ret = 1;
|
|
break;
|
|
}
|
|
return ret;
|
|
}
|
|
|
|
#if BUFSIZ < 8192
|
|
#define SMALLCHUNK 8192
|
|
#else
|
|
#define SMALLCHUNK BUFSIZ
|
|
#endif
|
|
|
|
#if SIZEOF_INT < 4
|
|
#define BIGCHUNK (512 * 32)
|
|
#else
|
|
#define BIGCHUNK (512 * 1024)
|
|
#endif
|
|
|
|
/* This is a hacked version of Python's fileobject.c:new_buffersize(). */
|
|
static size_t
|
|
Util_NewBufferSize(size_t currentsize)
|
|
{
|
|
if (currentsize > SMALLCHUNK) {
|
|
/* Keep doubling until we reach BIGCHUNK;
|
|
then keep adding BIGCHUNK. */
|
|
if (currentsize <= BIGCHUNK)
|
|
return currentsize + currentsize;
|
|
else
|
|
return currentsize + BIGCHUNK;
|
|
}
|
|
return currentsize + SMALLCHUNK;
|
|
}
|
|
|
|
/* This is a hacked version of Python's fileobject.c:get_line(). */
|
|
static PyObject *
|
|
Util_GetLine(BZ2FileObject *self, int n)
|
|
{
|
|
char c;
|
|
char *buf, *end;
|
|
size_t total_v_size; /* total # of slots in buffer */
|
|
size_t used_v_size; /* # used slots in buffer */
|
|
size_t increment; /* amount to increment the buffer */
|
|
PyObject *v;
|
|
int bzerror;
|
|
#ifdef WITH_UNIVERSAL_NEWLINES
|
|
int newlinetypes = ((PyFileObject*)self)->f_newlinetypes;
|
|
int skipnextlf = ((PyFileObject*)self)->f_skipnextlf;
|
|
int univ_newline = ((PyFileObject*)self)->f_univ_newline;
|
|
#endif
|
|
|
|
total_v_size = n > 0 ? n : 100;
|
|
v = PyString_FromStringAndSize((char *)NULL, total_v_size);
|
|
if (v == NULL)
|
|
return NULL;
|
|
|
|
buf = BUF(v);
|
|
end = buf + total_v_size;
|
|
|
|
for (;;) {
|
|
Py_BEGIN_ALLOW_THREADS
|
|
#ifdef WITH_UNIVERSAL_NEWLINES
|
|
if (univ_newline) {
|
|
while (1) {
|
|
BZ2_bzRead(&bzerror, self->fp, &c, 1);
|
|
self->pos++;
|
|
if (bzerror != BZ_OK || buf == end)
|
|
break;
|
|
if (skipnextlf) {
|
|
skipnextlf = 0;
|
|
if (c == '\n') {
|
|
/* Seeing a \n here with
|
|
* skipnextlf true means we
|
|
* saw a \r before.
|
|
*/
|
|
newlinetypes |= NEWLINE_CRLF;
|
|
BZ2_bzRead(&bzerror, self->fp,
|
|
&c, 1);
|
|
if (bzerror != BZ_OK)
|
|
break;
|
|
} else {
|
|
newlinetypes |= NEWLINE_CR;
|
|
}
|
|
}
|
|
if (c == '\r') {
|
|
skipnextlf = 1;
|
|
c = '\n';
|
|
} else if ( c == '\n')
|
|
newlinetypes |= NEWLINE_LF;
|
|
*buf++ = c;
|
|
if (c == '\n') break;
|
|
}
|
|
if (bzerror == BZ_STREAM_END && skipnextlf)
|
|
newlinetypes |= NEWLINE_CR;
|
|
} else /* If not universal newlines use the normal loop */
|
|
#endif
|
|
do {
|
|
BZ2_bzRead(&bzerror, self->fp, &c, 1);
|
|
self->pos++;
|
|
*buf++ = c;
|
|
} while (bzerror == BZ_OK && c != '\n' && buf != end);
|
|
Py_END_ALLOW_THREADS
|
|
#ifdef WITH_UNIVERSAL_NEWLINES
|
|
((PyFileObject*)self)->f_newlinetypes = newlinetypes;
|
|
((PyFileObject*)self)->f_skipnextlf = skipnextlf;
|
|
#endif
|
|
if (bzerror == BZ_STREAM_END) {
|
|
self->size = self->pos;
|
|
self->mode = MODE_READ_EOF;
|
|
break;
|
|
} else if (bzerror != BZ_OK) {
|
|
Util_CatchBZ2Error(bzerror);
|
|
Py_DECREF(v);
|
|
return NULL;
|
|
}
|
|
if (c == '\n')
|
|
break;
|
|
/* Must be because buf == end */
|
|
if (n > 0)
|
|
break;
|
|
used_v_size = total_v_size;
|
|
increment = total_v_size >> 2; /* mild exponential growth */
|
|
total_v_size += increment;
|
|
if (total_v_size > INT_MAX) {
|
|
PyErr_SetString(PyExc_OverflowError,
|
|
"line is longer than a Python string can hold");
|
|
Py_DECREF(v);
|
|
return NULL;
|
|
}
|
|
if (_PyString_Resize(&v, total_v_size) < 0)
|
|
return NULL;
|
|
buf = BUF(v) + used_v_size;
|
|
end = BUF(v) + total_v_size;
|
|
}
|
|
|
|
used_v_size = buf - BUF(v);
|
|
if (used_v_size != total_v_size)
|
|
_PyString_Resize(&v, used_v_size);
|
|
return v;
|
|
}
|
|
|
|
#ifndef WITH_UNIVERSAL_NEWLINES
|
|
#define Util_UnivNewlineRead(a,b,c,d,e) BZ2_bzRead(a,b,c,d)
|
|
#else
|
|
/* This is a hacked version of Python's
|
|
* fileobject.c:Py_UniversalNewlineFread(). */
|
|
size_t
|
|
Util_UnivNewlineRead(int *bzerror, BZFILE *stream,
|
|
char* buf, size_t n, BZ2FileObject *fobj)
|
|
{
|
|
char *dst = buf;
|
|
PyFileObject *f = (PyFileObject *)fobj;
|
|
int newlinetypes, skipnextlf;
|
|
|
|
assert(buf != NULL);
|
|
assert(stream != NULL);
|
|
|
|
if (!f->f_univ_newline)
|
|
return BZ2_bzRead(bzerror, stream, buf, n);
|
|
|
|
newlinetypes = f->f_newlinetypes;
|
|
skipnextlf = f->f_skipnextlf;
|
|
|
|
/* Invariant: n is the number of bytes remaining to be filled
|
|
* in the buffer.
|
|
*/
|
|
while (n) {
|
|
size_t nread;
|
|
int shortread;
|
|
char *src = dst;
|
|
|
|
nread = BZ2_bzRead(bzerror, stream, dst, n);
|
|
assert(nread <= n);
|
|
n -= nread; /* assuming 1 byte out for each in; will adjust */
|
|
shortread = n != 0; /* true iff EOF or error */
|
|
while (nread--) {
|
|
char c = *src++;
|
|
if (c == '\r') {
|
|
/* Save as LF and set flag to skip next LF. */
|
|
*dst++ = '\n';
|
|
skipnextlf = 1;
|
|
}
|
|
else if (skipnextlf && c == '\n') {
|
|
/* Skip LF, and remember we saw CR LF. */
|
|
skipnextlf = 0;
|
|
newlinetypes |= NEWLINE_CRLF;
|
|
++n;
|
|
}
|
|
else {
|
|
/* Normal char to be stored in buffer. Also
|
|
* update the newlinetypes flag if either this
|
|
* is an LF or the previous char was a CR.
|
|
*/
|
|
if (c == '\n')
|
|
newlinetypes |= NEWLINE_LF;
|
|
else if (skipnextlf)
|
|
newlinetypes |= NEWLINE_CR;
|
|
*dst++ = c;
|
|
skipnextlf = 0;
|
|
}
|
|
}
|
|
if (shortread) {
|
|
/* If this is EOF, update type flags. */
|
|
if (skipnextlf && *bzerror == BZ_STREAM_END)
|
|
newlinetypes |= NEWLINE_CR;
|
|
break;
|
|
}
|
|
}
|
|
f->f_newlinetypes = newlinetypes;
|
|
f->f_skipnextlf = skipnextlf;
|
|
return dst - buf;
|
|
}
|
|
#endif
|
|
|
|
/* This is a hacked version of Python's fileobject.c:drop_readahead(). */
|
|
static void
|
|
Util_DropReadAhead(BZ2FileObject *self)
|
|
{
|
|
PyFileObject *f = (PyFileObject*)self;
|
|
if (f->f_buf != NULL) {
|
|
PyMem_Free(f->f_buf);
|
|
f->f_buf = NULL;
|
|
}
|
|
}
|
|
|
|
/* This is a hacked version of Python's fileobject.c:readahead(). */
|
|
static int
|
|
Util_ReadAhead(BZ2FileObject *self, int bufsize)
|
|
{
|
|
int chunksize;
|
|
int bzerror;
|
|
PyFileObject *f = (PyFileObject*)self;
|
|
|
|
if (f->f_buf != NULL) {
|
|
if((f->f_bufend - f->f_bufptr) >= 1)
|
|
return 0;
|
|
else
|
|
Util_DropReadAhead(self);
|
|
}
|
|
if (self->mode == MODE_READ_EOF) {
|
|
return -1;
|
|
}
|
|
if ((f->f_buf = PyMem_Malloc(bufsize)) == NULL) {
|
|
return -1;
|
|
}
|
|
Py_BEGIN_ALLOW_THREADS
|
|
chunksize = Util_UnivNewlineRead(&bzerror, self->fp, f->f_buf,
|
|
bufsize, self);
|
|
Py_END_ALLOW_THREADS
|
|
self->pos += chunksize;
|
|
if (bzerror == BZ_STREAM_END) {
|
|
self->size = self->pos;
|
|
self->mode = MODE_READ_EOF;
|
|
} else if (bzerror != BZ_OK) {
|
|
Util_CatchBZ2Error(bzerror);
|
|
Util_DropReadAhead(self);
|
|
return -1;
|
|
}
|
|
f->f_bufptr = f->f_buf;
|
|
f->f_bufend = f->f_buf + chunksize;
|
|
return 0;
|
|
}
|
|
|
|
/* This is a hacked version of Python's
|
|
* fileobject.c:readahead_get_line_skip(). */
|
|
static PyStringObject *
|
|
Util_ReadAheadGetLineSkip(BZ2FileObject *bf, int skip, int bufsize)
|
|
{
|
|
PyFileObject *f = (PyFileObject*)bf;
|
|
PyStringObject* s;
|
|
char *bufptr;
|
|
char *buf;
|
|
int len;
|
|
|
|
if (f->f_buf == NULL)
|
|
if (Util_ReadAhead(bf, bufsize) < 0)
|
|
return NULL;
|
|
|
|
len = f->f_bufend - f->f_bufptr;
|
|
if (len == 0)
|
|
return (PyStringObject *)
|
|
PyString_FromStringAndSize(NULL, skip);
|
|
bufptr = memchr(f->f_bufptr, '\n', len);
|
|
if (bufptr != NULL) {
|
|
bufptr++; /* Count the '\n' */
|
|
len = bufptr - f->f_bufptr;
|
|
s = (PyStringObject *)
|
|
PyString_FromStringAndSize(NULL, skip+len);
|
|
if (s == NULL)
|
|
return NULL;
|
|
memcpy(PyString_AS_STRING(s)+skip, f->f_bufptr, len);
|
|
f->f_bufptr = bufptr;
|
|
if (bufptr == f->f_bufend)
|
|
Util_DropReadAhead(bf);
|
|
} else {
|
|
bufptr = f->f_bufptr;
|
|
buf = f->f_buf;
|
|
f->f_buf = NULL; /* Force new readahead buffer */
|
|
s = Util_ReadAheadGetLineSkip(
|
|
bf, skip+len, bufsize + (bufsize>>2) );
|
|
if (s == NULL) {
|
|
PyMem_Free(buf);
|
|
return NULL;
|
|
}
|
|
memcpy(PyString_AS_STRING(s)+skip, bufptr, len);
|
|
PyMem_Free(buf);
|
|
}
|
|
return s;
|
|
}
|
|
|
|
/* ===================================================================== */
|
|
/* Methods of BZ2File. */
|
|
|
|
PyDoc_STRVAR(BZ2File_read__doc__,
|
|
"read([size]) -> string\n\
|
|
\n\
|
|
Read at most size uncompressed bytes, returned as a string. If the size\n\
|
|
argument is negative or omitted, read until EOF is reached.\n\
|
|
");
|
|
|
|
/* This is a hacked version of Python's fileobject.c:file_read(). */
|
|
static PyObject *
|
|
BZ2File_read(BZ2FileObject *self, PyObject *args)
|
|
{
|
|
long bytesrequested = -1;
|
|
size_t bytesread, buffersize, chunksize;
|
|
int bzerror;
|
|
PyObject *ret = NULL;
|
|
|
|
if (!PyArg_ParseTuple(args, "|l:read", &bytesrequested))
|
|
return NULL;
|
|
|
|
ACQUIRE_LOCK(self);
|
|
switch (self->mode) {
|
|
case MODE_READ:
|
|
break;
|
|
case MODE_READ_EOF:
|
|
ret = PyString_FromString("");
|
|
goto cleanup;
|
|
case MODE_CLOSED:
|
|
PyErr_SetString(PyExc_ValueError,
|
|
"I/O operation on closed file");
|
|
goto cleanup;
|
|
default:
|
|
PyErr_SetString(PyExc_IOError,
|
|
"file is not ready for reading");
|
|
goto cleanup;
|
|
}
|
|
|
|
if (bytesrequested < 0)
|
|
buffersize = Util_NewBufferSize((size_t)0);
|
|
else
|
|
buffersize = bytesrequested;
|
|
if (buffersize > INT_MAX) {
|
|
PyErr_SetString(PyExc_OverflowError,
|
|
"requested number of bytes is "
|
|
"more than a Python string can hold");
|
|
goto cleanup;
|
|
}
|
|
ret = PyString_FromStringAndSize((char *)NULL, buffersize);
|
|
if (ret == NULL)
|
|
goto cleanup;
|
|
bytesread = 0;
|
|
|
|
for (;;) {
|
|
Py_BEGIN_ALLOW_THREADS
|
|
chunksize = Util_UnivNewlineRead(&bzerror, self->fp,
|
|
BUF(ret)+bytesread,
|
|
buffersize-bytesread,
|
|
self);
|
|
self->pos += chunksize;
|
|
Py_END_ALLOW_THREADS
|
|
bytesread += chunksize;
|
|
if (bzerror == BZ_STREAM_END) {
|
|
self->size = self->pos;
|
|
self->mode = MODE_READ_EOF;
|
|
break;
|
|
} else if (bzerror != BZ_OK) {
|
|
Util_CatchBZ2Error(bzerror);
|
|
Py_DECREF(ret);
|
|
ret = NULL;
|
|
goto cleanup;
|
|
}
|
|
if (bytesrequested < 0) {
|
|
buffersize = Util_NewBufferSize(buffersize);
|
|
if (_PyString_Resize(&ret, buffersize) < 0)
|
|
goto cleanup;
|
|
} else {
|
|
break;
|
|
}
|
|
}
|
|
if (bytesread != buffersize)
|
|
_PyString_Resize(&ret, bytesread);
|
|
|
|
cleanup:
|
|
RELEASE_LOCK(self);
|
|
return ret;
|
|
}
|
|
|
|
PyDoc_STRVAR(BZ2File_readline__doc__,
|
|
"readline([size]) -> string\n\
|
|
\n\
|
|
Return the next line from the file, as a string, retaining newline.\n\
|
|
A non-negative size argument will limit the maximum number of bytes to\n\
|
|
return (an incomplete line may be returned then). Return an empty\n\
|
|
string at EOF.\n\
|
|
");
|
|
|
|
static PyObject *
|
|
BZ2File_readline(BZ2FileObject *self, PyObject *args)
|
|
{
|
|
PyObject *ret = NULL;
|
|
int sizehint = -1;
|
|
|
|
if (!PyArg_ParseTuple(args, "|i:readline", &sizehint))
|
|
return NULL;
|
|
|
|
ACQUIRE_LOCK(self);
|
|
switch (self->mode) {
|
|
case MODE_READ:
|
|
break;
|
|
case MODE_READ_EOF:
|
|
ret = PyString_FromString("");
|
|
goto cleanup;
|
|
case MODE_CLOSED:
|
|
PyErr_SetString(PyExc_ValueError,
|
|
"I/O operation on closed file");
|
|
goto cleanup;
|
|
default:
|
|
PyErr_SetString(PyExc_IOError,
|
|
"file is not ready for reading");
|
|
goto cleanup;
|
|
}
|
|
|
|
if (sizehint == 0)
|
|
ret = PyString_FromString("");
|
|
else
|
|
ret = Util_GetLine(self, (sizehint < 0) ? 0 : sizehint);
|
|
|
|
cleanup:
|
|
RELEASE_LOCK(self);
|
|
return ret;
|
|
}
|
|
|
|
PyDoc_STRVAR(BZ2File_readlines__doc__,
|
|
"readlines([size]) -> list\n\
|
|
\n\
|
|
Call readline() repeatedly and return a list of lines read.\n\
|
|
The optional size argument, if given, is an approximate bound on the\n\
|
|
total number of bytes in the lines returned.\n\
|
|
");
|
|
|
|
/* This is a hacked version of Python's fileobject.c:file_readlines(). */
|
|
static PyObject *
|
|
BZ2File_readlines(BZ2FileObject *self, PyObject *args)
|
|
{
|
|
long sizehint = 0;
|
|
PyObject *list = NULL;
|
|
PyObject *line;
|
|
char small_buffer[SMALLCHUNK];
|
|
char *buffer = small_buffer;
|
|
size_t buffersize = SMALLCHUNK;
|
|
PyObject *big_buffer = NULL;
|
|
size_t nfilled = 0;
|
|
size_t nread;
|
|
size_t totalread = 0;
|
|
char *p, *q, *end;
|
|
int err;
|
|
int shortread = 0;
|
|
int bzerror;
|
|
|
|
if (!PyArg_ParseTuple(args, "|l:readlines", &sizehint))
|
|
return NULL;
|
|
|
|
ACQUIRE_LOCK(self);
|
|
switch (self->mode) {
|
|
case MODE_READ:
|
|
break;
|
|
case MODE_READ_EOF:
|
|
list = PyList_New(0);
|
|
goto cleanup;
|
|
case MODE_CLOSED:
|
|
PyErr_SetString(PyExc_ValueError,
|
|
"I/O operation on closed file");
|
|
goto cleanup;
|
|
default:
|
|
PyErr_SetString(PyExc_IOError,
|
|
"file is not ready for reading");
|
|
goto cleanup;
|
|
}
|
|
|
|
if ((list = PyList_New(0)) == NULL)
|
|
goto cleanup;
|
|
|
|
for (;;) {
|
|
Py_BEGIN_ALLOW_THREADS
|
|
nread = Util_UnivNewlineRead(&bzerror, self->fp,
|
|
buffer+nfilled,
|
|
buffersize-nfilled, self);
|
|
self->pos += nread;
|
|
Py_END_ALLOW_THREADS
|
|
if (bzerror == BZ_STREAM_END) {
|
|
self->size = self->pos;
|
|
self->mode = MODE_READ_EOF;
|
|
if (nread == 0) {
|
|
sizehint = 0;
|
|
break;
|
|
}
|
|
shortread = 1;
|
|
} else if (bzerror != BZ_OK) {
|
|
Util_CatchBZ2Error(bzerror);
|
|
error:
|
|
Py_DECREF(list);
|
|
list = NULL;
|
|
goto cleanup;
|
|
}
|
|
totalread += nread;
|
|
p = memchr(buffer+nfilled, '\n', nread);
|
|
if (p == NULL) {
|
|
/* Need a larger buffer to fit this line */
|
|
nfilled += nread;
|
|
buffersize *= 2;
|
|
if (buffersize > INT_MAX) {
|
|
PyErr_SetString(PyExc_OverflowError,
|
|
"line is longer than a Python string can hold");
|
|
goto error;
|
|
}
|
|
if (big_buffer == NULL) {
|
|
/* Create the big buffer */
|
|
big_buffer = PyString_FromStringAndSize(
|
|
NULL, buffersize);
|
|
if (big_buffer == NULL)
|
|
goto error;
|
|
buffer = PyString_AS_STRING(big_buffer);
|
|
memcpy(buffer, small_buffer, nfilled);
|
|
}
|
|
else {
|
|
/* Grow the big buffer */
|
|
_PyString_Resize(&big_buffer, buffersize);
|
|
buffer = PyString_AS_STRING(big_buffer);
|
|
}
|
|
continue;
|
|
}
|
|
end = buffer+nfilled+nread;
|
|
q = buffer;
|
|
do {
|
|
/* Process complete lines */
|
|
p++;
|
|
line = PyString_FromStringAndSize(q, p-q);
|
|
if (line == NULL)
|
|
goto error;
|
|
err = PyList_Append(list, line);
|
|
Py_DECREF(line);
|
|
if (err != 0)
|
|
goto error;
|
|
q = p;
|
|
p = memchr(q, '\n', end-q);
|
|
} while (p != NULL);
|
|
/* Move the remaining incomplete line to the start */
|
|
nfilled = end-q;
|
|
memmove(buffer, q, nfilled);
|
|
if (sizehint > 0)
|
|
if (totalread >= (size_t)sizehint)
|
|
break;
|
|
if (shortread) {
|
|
sizehint = 0;
|
|
break;
|
|
}
|
|
}
|
|
if (nfilled != 0) {
|
|
/* Partial last line */
|
|
line = PyString_FromStringAndSize(buffer, nfilled);
|
|
if (line == NULL)
|
|
goto error;
|
|
if (sizehint > 0) {
|
|
/* Need to complete the last line */
|
|
PyObject *rest = Util_GetLine(self, 0);
|
|
if (rest == NULL) {
|
|
Py_DECREF(line);
|
|
goto error;
|
|
}
|
|
PyString_Concat(&line, rest);
|
|
Py_DECREF(rest);
|
|
if (line == NULL)
|
|
goto error;
|
|
}
|
|
err = PyList_Append(list, line);
|
|
Py_DECREF(line);
|
|
if (err != 0)
|
|
goto error;
|
|
}
|
|
|
|
cleanup:
|
|
RELEASE_LOCK(self);
|
|
if (big_buffer) {
|
|
Py_DECREF(big_buffer);
|
|
}
|
|
return list;
|
|
}
|
|
|
|
PyDoc_STRVAR(BZ2File_write__doc__,
|
|
"write(data) -> None\n\
|
|
\n\
|
|
Write the 'data' string to file. Note that due to buffering, close() may\n\
|
|
be needed before the file on disk reflects the data written.\n\
|
|
");
|
|
|
|
/* This is a hacked version of Python's fileobject.c:file_write(). */
|
|
static PyObject *
|
|
BZ2File_write(BZ2FileObject *self, PyObject *args)
|
|
{
|
|
PyObject *ret = NULL;
|
|
char *buf;
|
|
int len;
|
|
int bzerror;
|
|
|
|
if (!PyArg_ParseTuple(args, "s#", &buf, &len))
|
|
return NULL;
|
|
|
|
ACQUIRE_LOCK(self);
|
|
switch (self->mode) {
|
|
case MODE_WRITE:
|
|
break;
|
|
|
|
case MODE_CLOSED:
|
|
PyErr_SetString(PyExc_ValueError,
|
|
"I/O operation on closed file");
|
|
goto cleanup;;
|
|
|
|
default:
|
|
PyErr_SetString(PyExc_IOError,
|
|
"file is not ready for writing");
|
|
goto cleanup;;
|
|
}
|
|
|
|
PyFile_SoftSpace((PyObject*)self, 0);
|
|
|
|
Py_BEGIN_ALLOW_THREADS
|
|
BZ2_bzWrite (&bzerror, self->fp, buf, len);
|
|
self->pos += len;
|
|
Py_END_ALLOW_THREADS
|
|
|
|
if (bzerror != BZ_OK) {
|
|
Util_CatchBZ2Error(bzerror);
|
|
goto cleanup;
|
|
}
|
|
|
|
Py_INCREF(Py_None);
|
|
ret = Py_None;
|
|
|
|
cleanup:
|
|
RELEASE_LOCK(self);
|
|
return ret;
|
|
}
|
|
|
|
PyDoc_STRVAR(BZ2File_writelines__doc__,
|
|
"writelines(sequence_of_strings) -> None\n\
|
|
\n\
|
|
Write the sequence of strings to the file. Note that newlines are not\n\
|
|
added. The sequence can be any iterable object producing strings. This is\n\
|
|
equivalent to calling write() for each string.\n\
|
|
");
|
|
|
|
/* This is a hacked version of Python's fileobject.c:file_writelines(). */
|
|
static PyObject *
|
|
BZ2File_writelines(BZ2FileObject *self, PyObject *seq)
|
|
{
|
|
#define CHUNKSIZE 1000
|
|
PyObject *list = NULL;
|
|
PyObject *iter = NULL;
|
|
PyObject *ret = NULL;
|
|
PyObject *line;
|
|
int i, j, index, len, islist;
|
|
int bzerror;
|
|
|
|
ACQUIRE_LOCK(self);
|
|
islist = PyList_Check(seq);
|
|
if (!islist) {
|
|
iter = PyObject_GetIter(seq);
|
|
if (iter == NULL) {
|
|
PyErr_SetString(PyExc_TypeError,
|
|
"writelines() requires an iterable argument");
|
|
goto error;
|
|
}
|
|
list = PyList_New(CHUNKSIZE);
|
|
if (list == NULL)
|
|
goto error;
|
|
}
|
|
|
|
/* Strategy: slurp CHUNKSIZE lines into a private list,
|
|
checking that they are all strings, then write that list
|
|
without holding the interpreter lock, then come back for more. */
|
|
for (index = 0; ; index += CHUNKSIZE) {
|
|
if (islist) {
|
|
Py_XDECREF(list);
|
|
list = PyList_GetSlice(seq, index, index+CHUNKSIZE);
|
|
if (list == NULL)
|
|
goto error;
|
|
j = PyList_GET_SIZE(list);
|
|
}
|
|
else {
|
|
for (j = 0; j < CHUNKSIZE; j++) {
|
|
line = PyIter_Next(iter);
|
|
if (line == NULL) {
|
|
if (PyErr_Occurred())
|
|
goto error;
|
|
break;
|
|
}
|
|
PyList_SetItem(list, j, line);
|
|
}
|
|
}
|
|
if (j == 0)
|
|
break;
|
|
|
|
/* Check that all entries are indeed strings. If not,
|
|
apply the same rules as for file.write() and
|
|
convert the rets to strings. This is slow, but
|
|
seems to be the only way since all conversion APIs
|
|
could potentially execute Python code. */
|
|
for (i = 0; i < j; i++) {
|
|
PyObject *v = PyList_GET_ITEM(list, i);
|
|
if (!PyString_Check(v)) {
|
|
const char *buffer;
|
|
int len;
|
|
if (PyObject_AsCharBuffer(v, &buffer, &len)) {
|
|
PyErr_SetString(PyExc_TypeError,
|
|
"writelines() "
|
|
"argument must be "
|
|
"a sequence of "
|
|
"strings");
|
|
goto error;
|
|
}
|
|
line = PyString_FromStringAndSize(buffer,
|
|
len);
|
|
if (line == NULL)
|
|
goto error;
|
|
Py_DECREF(v);
|
|
PyList_SET_ITEM(list, i, line);
|
|
}
|
|
}
|
|
|
|
PyFile_SoftSpace((PyObject*)self, 0);
|
|
|
|
/* Since we are releasing the global lock, the
|
|
following code may *not* execute Python code. */
|
|
Py_BEGIN_ALLOW_THREADS
|
|
for (i = 0; i < j; i++) {
|
|
line = PyList_GET_ITEM(list, i);
|
|
len = PyString_GET_SIZE(line);
|
|
BZ2_bzWrite (&bzerror, self->fp,
|
|
PyString_AS_STRING(line), len);
|
|
if (bzerror != BZ_OK) {
|
|
Py_BLOCK_THREADS
|
|
Util_CatchBZ2Error(bzerror);
|
|
goto error;
|
|
}
|
|
}
|
|
Py_END_ALLOW_THREADS
|
|
|
|
if (j < CHUNKSIZE)
|
|
break;
|
|
}
|
|
|
|
Py_INCREF(Py_None);
|
|
ret = Py_None;
|
|
|
|
error:
|
|
RELEASE_LOCK(self);
|
|
Py_XDECREF(list);
|
|
Py_XDECREF(iter);
|
|
return ret;
|
|
#undef CHUNKSIZE
|
|
}
|
|
|
|
PyDoc_STRVAR(BZ2File_seek__doc__,
|
|
"seek(offset [, whence]) -> None\n\
|
|
\n\
|
|
Move to new file position. Argument offset is a byte count. Optional\n\
|
|
argument whence defaults to 0 (offset from start of file, offset\n\
|
|
should be >= 0); other values are 1 (move relative to current position,\n\
|
|
positive or negative), and 2 (move relative to end of file, usually\n\
|
|
negative, although many platforms allow seeking beyond the end of a file).\n\
|
|
\n\
|
|
Note that seeking of bz2 files is emulated, and depending on the parameters\n\
|
|
the operation may be extremely slow.\n\
|
|
");
|
|
|
|
static PyObject *
|
|
BZ2File_seek(BZ2FileObject *self, PyObject *args)
|
|
{
|
|
int where = 0;
|
|
long offset;
|
|
char small_buffer[SMALLCHUNK];
|
|
char *buffer = small_buffer;
|
|
size_t buffersize = SMALLCHUNK;
|
|
int bytesread = 0;
|
|
int readsize;
|
|
int chunksize;
|
|
int bzerror;
|
|
int rewind = 0;
|
|
PyObject *func;
|
|
PyObject *ret = NULL;
|
|
|
|
if (!PyArg_ParseTuple(args, "l|i:seek", &offset, &where))
|
|
return NULL;
|
|
|
|
ACQUIRE_LOCK(self);
|
|
Util_DropReadAhead(self);
|
|
switch (self->mode) {
|
|
case MODE_READ:
|
|
case MODE_READ_EOF:
|
|
break;
|
|
|
|
case MODE_CLOSED:
|
|
PyErr_SetString(PyExc_ValueError,
|
|
"I/O operation on closed file");
|
|
goto cleanup;;
|
|
|
|
default:
|
|
PyErr_SetString(PyExc_IOError,
|
|
"seek works only while reading");
|
|
goto cleanup;;
|
|
}
|
|
|
|
if (offset < 0) {
|
|
if (where == 1) {
|
|
offset = self->pos + offset;
|
|
rewind = 1;
|
|
} else if (where == 2) {
|
|
if (self->size == -1) {
|
|
assert(self->mode != MODE_READ_EOF);
|
|
for (;;) {
|
|
Py_BEGIN_ALLOW_THREADS
|
|
chunksize = Util_UnivNewlineRead(
|
|
&bzerror, self->fp,
|
|
buffer, buffersize,
|
|
self);
|
|
self->pos += chunksize;
|
|
Py_END_ALLOW_THREADS
|
|
|
|
bytesread += chunksize;
|
|
if (bzerror == BZ_STREAM_END) {
|
|
break;
|
|
} else if (bzerror != BZ_OK) {
|
|
Util_CatchBZ2Error(bzerror);
|
|
goto cleanup;
|
|
}
|
|
}
|
|
self->mode = MODE_READ_EOF;
|
|
self->size = self->pos;
|
|
bytesread = 0;
|
|
}
|
|
offset = self->size + offset;
|
|
if (offset >= self->pos)
|
|
offset -= self->pos;
|
|
else
|
|
rewind = 1;
|
|
}
|
|
if (offset < 0)
|
|
offset = 0;
|
|
} else if (where == 0) {
|
|
if (offset >= self->pos)
|
|
offset -= self->pos;
|
|
else
|
|
rewind = 1;
|
|
}
|
|
|
|
if (rewind) {
|
|
BZ2_bzReadClose(&bzerror, self->fp);
|
|
func = Py_FindMethod(PyFile_Type.tp_methods, (PyObject*)self,
|
|
"seek");
|
|
if (bzerror != BZ_OK) {
|
|
Util_CatchBZ2Error(bzerror);
|
|
goto cleanup;
|
|
}
|
|
if (!func) {
|
|
PyErr_SetString(PyExc_RuntimeError,
|
|
"can't find file.seek method");
|
|
goto cleanup;
|
|
}
|
|
ret = PyObject_CallFunction(func, "(i)", 0);
|
|
if (!ret)
|
|
goto cleanup;
|
|
Py_DECREF(ret);
|
|
ret = NULL;
|
|
self->pos = 0;
|
|
self->fp = BZ2_bzReadOpen(&bzerror,
|
|
PyFile_AsFile((PyObject*)self),
|
|
0, 0, NULL, 0);
|
|
if (bzerror != BZ_OK) {
|
|
Util_CatchBZ2Error(bzerror);
|
|
goto cleanup;
|
|
}
|
|
self->mode = MODE_READ;
|
|
} else if (self->mode == MODE_READ_EOF) {
|
|
goto exit;
|
|
}
|
|
|
|
if (offset == 0)
|
|
goto exit;
|
|
|
|
/* Before getting here, offset must be set to the number of bytes
|
|
* to walk forward. */
|
|
for (;;) {
|
|
if ((size_t)offset-bytesread > buffersize)
|
|
readsize = buffersize;
|
|
else
|
|
readsize = offset-bytesread;
|
|
Py_BEGIN_ALLOW_THREADS
|
|
chunksize = Util_UnivNewlineRead(&bzerror, self->fp,
|
|
buffer, readsize, self);
|
|
self->pos += chunksize;
|
|
Py_END_ALLOW_THREADS
|
|
bytesread += chunksize;
|
|
if (bzerror == BZ_STREAM_END) {
|
|
self->size = self->pos;
|
|
self->mode = MODE_READ_EOF;
|
|
break;
|
|
} else if (bzerror != BZ_OK) {
|
|
Util_CatchBZ2Error(bzerror);
|
|
goto cleanup;
|
|
}
|
|
if (bytesread == offset)
|
|
break;
|
|
}
|
|
|
|
exit:
|
|
Py_INCREF(Py_None);
|
|
ret = Py_None;
|
|
|
|
cleanup:
|
|
RELEASE_LOCK(self);
|
|
return ret;
|
|
}
|
|
|
|
PyDoc_STRVAR(BZ2File_tell__doc__,
|
|
"tell() -> int\n\
|
|
\n\
|
|
Return the current file position, an integer (may be a long integer).\n\
|
|
");
|
|
|
|
static PyObject *
|
|
BZ2File_tell(BZ2FileObject *self, PyObject *args)
|
|
{
|
|
PyObject *ret = NULL;
|
|
|
|
if (self->mode == MODE_CLOSED) {
|
|
PyErr_SetString(PyExc_ValueError,
|
|
"I/O operation on closed file");
|
|
goto cleanup;
|
|
}
|
|
|
|
ret = PyInt_FromLong(self->pos);
|
|
|
|
cleanup:
|
|
return ret;
|
|
}
|
|
|
|
PyDoc_STRVAR(BZ2File_notsup__doc__,
|
|
"Operation not supported.\n\
|
|
");
|
|
|
|
static PyObject *
|
|
BZ2File_notsup(BZ2FileObject *self, PyObject *args)
|
|
{
|
|
PyErr_SetString(PyExc_IOError, "operation not supported");
|
|
return NULL;
|
|
}
|
|
|
|
PyDoc_STRVAR(BZ2File_close__doc__,
|
|
"close() -> None or (perhaps) an integer\n\
|
|
\n\
|
|
Close the file. Sets data attribute .closed to true. A closed file\n\
|
|
cannot be used for further I/O operations. close() may be called more\n\
|
|
than once without error.\n\
|
|
");
|
|
|
|
static PyObject *
|
|
BZ2File_close(BZ2FileObject *self)
|
|
{
|
|
PyObject *file_close;
|
|
PyObject *ret = NULL;
|
|
int bzerror = BZ_OK;
|
|
|
|
ACQUIRE_LOCK(self);
|
|
switch (self->mode) {
|
|
case MODE_READ:
|
|
case MODE_READ_EOF:
|
|
BZ2_bzReadClose(&bzerror, self->fp);
|
|
break;
|
|
case MODE_WRITE:
|
|
BZ2_bzWriteClose(&bzerror, self->fp,
|
|
0, NULL, NULL);
|
|
break;
|
|
}
|
|
self->mode = MODE_CLOSED;
|
|
file_close = Py_FindMethod(PyFile_Type.tp_methods, (PyObject*)self,
|
|
"close");
|
|
if (!file_close) {
|
|
PyErr_SetString(PyExc_RuntimeError,
|
|
"can't find file.close method");
|
|
goto cleanup;
|
|
}
|
|
ret = PyObject_CallObject(file_close, NULL);
|
|
if (bzerror != BZ_OK) {
|
|
Util_CatchBZ2Error(bzerror);
|
|
Py_XDECREF(ret);
|
|
ret = NULL;
|
|
goto cleanup;
|
|
}
|
|
|
|
cleanup:
|
|
RELEASE_LOCK(self);
|
|
return ret;
|
|
}
|
|
|
|
static PyMethodDef BZ2File_methods[] = {
|
|
{"read", (PyCFunction)BZ2File_read, METH_VARARGS, BZ2File_read__doc__},
|
|
{"readline", (PyCFunction)BZ2File_readline, METH_VARARGS, BZ2File_readline__doc__},
|
|
{"readlines", (PyCFunction)BZ2File_readlines, METH_VARARGS, BZ2File_readlines__doc__},
|
|
{"write", (PyCFunction)BZ2File_write, METH_VARARGS, BZ2File_write__doc__},
|
|
{"writelines", (PyCFunction)BZ2File_writelines, METH_O, BZ2File_writelines__doc__},
|
|
{"seek", (PyCFunction)BZ2File_seek, METH_VARARGS, BZ2File_seek__doc__},
|
|
{"tell", (PyCFunction)BZ2File_tell, METH_NOARGS, BZ2File_tell__doc__},
|
|
{"truncate", (PyCFunction)BZ2File_notsup, METH_VARARGS, BZ2File_notsup__doc__},
|
|
{"readinto", (PyCFunction)BZ2File_notsup, METH_VARARGS, BZ2File_notsup__doc__},
|
|
{"close", (PyCFunction)BZ2File_close, METH_NOARGS, BZ2File_close__doc__},
|
|
{NULL, NULL} /* sentinel */
|
|
};
|
|
|
|
|
|
/* ===================================================================== */
|
|
/* Slot definitions for BZ2File_Type. */
|
|
|
|
static int
|
|
BZ2File_init(BZ2FileObject *self, PyObject *args, PyObject *kwargs)
|
|
{
|
|
PyObject *file_args = NULL;
|
|
static char *kwlist[] = {"filename", "mode", "buffering",
|
|
"compresslevel", 0};
|
|
char *name = NULL;
|
|
char *mode = "r";
|
|
int buffering = -1;
|
|
int compresslevel = 9;
|
|
int bzerror;
|
|
int mode_char = 0;
|
|
int univ_newline = 0;
|
|
|
|
self->size = -1;
|
|
|
|
if (!PyArg_ParseTupleAndKeywords(args, kwargs, "et|sii:BZ2File",
|
|
kwlist, Py_FileSystemDefaultEncoding,
|
|
&name, &mode, &buffering,
|
|
&compresslevel))
|
|
return -1;
|
|
|
|
if (compresslevel < 1 || compresslevel > 9) {
|
|
PyErr_SetString(PyExc_ValueError,
|
|
"compresslevel must be between 1 and 9");
|
|
return -1;
|
|
}
|
|
|
|
for (;;) {
|
|
int error = 0;
|
|
switch (*mode) {
|
|
case 'r':
|
|
case 'w':
|
|
if (mode_char)
|
|
error = 1;
|
|
mode_char = *mode;
|
|
break;
|
|
|
|
case 'b':
|
|
break;
|
|
|
|
case 'U':
|
|
univ_newline = 1;
|
|
break;
|
|
|
|
default:
|
|
error = 1;
|
|
break;
|
|
}
|
|
if (error) {
|
|
PyErr_Format(PyExc_ValueError,
|
|
"invalid mode char %c", *mode);
|
|
return -1;
|
|
}
|
|
mode++;
|
|
if (*mode == '\0')
|
|
break;
|
|
}
|
|
|
|
if (mode_char == 'r')
|
|
mode = univ_newline ? "rbU" : "rb";
|
|
else
|
|
mode = univ_newline ? "wbU" : "wb";
|
|
|
|
file_args = Py_BuildValue("(ssi)", name, mode, buffering);
|
|
if (!file_args)
|
|
return -1;
|
|
|
|
/* From now on, we have stuff to dealloc, so jump to error label
|
|
* instead of returning */
|
|
|
|
if (PyFile_Type.tp_init((PyObject *)self, file_args, NULL) < 0)
|
|
goto error;
|
|
|
|
#ifdef WITH_THREAD
|
|
self->lock = PyThread_allocate_lock();
|
|
if (!self->lock)
|
|
goto error;
|
|
#endif
|
|
|
|
if (mode_char == 'r')
|
|
self->fp = BZ2_bzReadOpen(&bzerror,
|
|
PyFile_AsFile((PyObject*)self),
|
|
0, 0, NULL, 0);
|
|
else
|
|
self->fp = BZ2_bzWriteOpen(&bzerror,
|
|
PyFile_AsFile((PyObject*)self),
|
|
compresslevel, 0, 0);
|
|
|
|
if (bzerror != BZ_OK) {
|
|
Util_CatchBZ2Error(bzerror);
|
|
goto error;
|
|
}
|
|
|
|
self->mode = (mode_char == 'r') ? MODE_READ : MODE_WRITE;
|
|
|
|
Py_XDECREF(file_args);
|
|
PyMem_Free(name);
|
|
return 0;
|
|
|
|
error:
|
|
#ifdef WITH_THREAD
|
|
if (self->lock)
|
|
PyThread_free_lock(self->lock);
|
|
#endif
|
|
Py_XDECREF(file_args);
|
|
PyMem_Free(name);
|
|
return -1;
|
|
}
|
|
|
|
static void
|
|
BZ2File_dealloc(BZ2FileObject *self)
|
|
{
|
|
int bzerror;
|
|
#ifdef WITH_THREAD
|
|
if (self->lock)
|
|
PyThread_free_lock(self->lock);
|
|
#endif
|
|
switch (self->mode) {
|
|
case MODE_READ:
|
|
case MODE_READ_EOF:
|
|
BZ2_bzReadClose(&bzerror, self->fp);
|
|
break;
|
|
case MODE_WRITE:
|
|
BZ2_bzWriteClose(&bzerror, self->fp,
|
|
0, NULL, NULL);
|
|
break;
|
|
}
|
|
Util_DropReadAhead(self);
|
|
((PyObject*)self)->ob_type->tp_free((PyObject *)self);
|
|
}
|
|
|
|
/* This is a hacked version of Python's fileobject.c:file_getiter(). */
|
|
static PyObject *
|
|
BZ2File_getiter(BZ2FileObject *self)
|
|
{
|
|
if (self->mode == MODE_CLOSED) {
|
|
PyErr_SetString(PyExc_ValueError,
|
|
"I/O operation on closed file");
|
|
return NULL;
|
|
}
|
|
Py_INCREF((PyObject*)self);
|
|
return (PyObject *)self;
|
|
}
|
|
|
|
/* This is a hacked version of Python's fileobject.c:file_iternext(). */
|
|
#define READAHEAD_BUFSIZE 8192
|
|
static PyObject *
|
|
BZ2File_iternext(BZ2FileObject *self)
|
|
{
|
|
PyStringObject* ret;
|
|
ACQUIRE_LOCK(self);
|
|
if (self->mode == MODE_CLOSED) {
|
|
PyErr_SetString(PyExc_ValueError,
|
|
"I/O operation on closed file");
|
|
return NULL;
|
|
}
|
|
ret = Util_ReadAheadGetLineSkip(self, 0, READAHEAD_BUFSIZE);
|
|
RELEASE_LOCK(self);
|
|
if (ret == NULL || PyString_GET_SIZE(ret) == 0) {
|
|
Py_XDECREF(ret);
|
|
return NULL;
|
|
}
|
|
return (PyObject *)ret;
|
|
}
|
|
|
|
/* ===================================================================== */
|
|
/* BZ2File_Type definition. */
|
|
|
|
PyDoc_VAR(BZ2File__doc__) =
|
|
PyDoc_STR(
|
|
"BZ2File(name [, mode='r', buffering=0, compresslevel=9]) -> file object\n\
|
|
\n\
|
|
Open a bz2 file. The mode can be 'r' or 'w', for reading (default) or\n\
|
|
writing. When opened for writing, the file will be created if it doesn't\n\
|
|
exist, and truncated otherwise. If the buffering argument is given, 0 means\n\
|
|
unbuffered, and larger numbers specify the buffer size. If compresslevel\n\
|
|
is given, must be a number between 1 and 9.\n\
|
|
")
|
|
#ifdef WITH_UNIVERSAL_NEWLINES
|
|
PyDoc_STR(
|
|
"\n\
|
|
Add a 'U' to mode to open the file for input with universal newline\n\
|
|
support. Any line ending in the input file will be seen as a '\\n' in\n\
|
|
Python. Also, a file so opened gains the attribute 'newlines'; the value\n\
|
|
for this attribute is one of None (no newline read yet), '\\r', '\\n',\n\
|
|
'\\r\\n' or a tuple containing all the newline types seen. Universal\n\
|
|
newlines are available only when reading.\n\
|
|
")
|
|
#endif
|
|
;
|
|
|
|
static PyTypeObject BZ2File_Type = {
|
|
PyObject_HEAD_INIT(NULL)
|
|
0, /*ob_size*/
|
|
"bz2.BZ2File", /*tp_name*/
|
|
sizeof(BZ2FileObject), /*tp_basicsize*/
|
|
0, /*tp_itemsize*/
|
|
(destructor)BZ2File_dealloc, /*tp_dealloc*/
|
|
0, /*tp_print*/
|
|
0, /*tp_getattr*/
|
|
0, /*tp_setattr*/
|
|
0, /*tp_compare*/
|
|
0, /*tp_repr*/
|
|
0, /*tp_as_number*/
|
|
0, /*tp_as_sequence*/
|
|
0, /*tp_as_mapping*/
|
|
0, /*tp_hash*/
|
|
0, /*tp_call*/
|
|
0, /*tp_str*/
|
|
0, /*tp_getattro*/
|
|
0, /*tp_setattro*/
|
|
0, /*tp_as_buffer*/
|
|
Py_TPFLAGS_DEFAULT|Py_TPFLAGS_BASETYPE, /*tp_flags*/
|
|
BZ2File__doc__, /*tp_doc*/
|
|
0, /*tp_traverse*/
|
|
0, /*tp_clear*/
|
|
0, /*tp_richcompare*/
|
|
0, /*tp_weaklistoffset*/
|
|
(getiterfunc)BZ2File_getiter, /*tp_iter*/
|
|
(iternextfunc)BZ2File_iternext, /*tp_iternext*/
|
|
BZ2File_methods, /*tp_methods*/
|
|
0, /*tp_members*/
|
|
0, /*tp_getset*/
|
|
0, /*tp_base*/
|
|
0, /*tp_dict*/
|
|
0, /*tp_descr_get*/
|
|
0, /*tp_descr_set*/
|
|
0, /*tp_dictoffset*/
|
|
(initproc)BZ2File_init, /*tp_init*/
|
|
0, /*tp_alloc*/
|
|
0, /*tp_new*/
|
|
0, /*tp_free*/
|
|
0, /*tp_is_gc*/
|
|
};
|
|
|
|
|
|
/* ===================================================================== */
|
|
/* Methods of BZ2Comp. */
|
|
|
|
PyDoc_STRVAR(BZ2Comp_compress__doc__,
|
|
"compress(data) -> string\n\
|
|
\n\
|
|
Provide more data to the compressor object. It will return chunks of\n\
|
|
compressed data whenever possible. When you've finished providing data\n\
|
|
to compress, call the flush() method to finish the compression process,\n\
|
|
and return what is left in the internal buffers.\n\
|
|
");
|
|
|
|
static PyObject *
|
|
BZ2Comp_compress(BZ2CompObject *self, PyObject *args)
|
|
{
|
|
char *data;
|
|
int datasize;
|
|
int bufsize = SMALLCHUNK;
|
|
LONG_LONG totalout;
|
|
PyObject *ret = NULL;
|
|
bz_stream *bzs = &self->bzs;
|
|
int bzerror;
|
|
|
|
if (!PyArg_ParseTuple(args, "s#", &data, &datasize))
|
|
return NULL;
|
|
|
|
ACQUIRE_LOCK(self);
|
|
if (!self->running) {
|
|
PyErr_SetString(PyExc_ValueError,
|
|
"this object was already flushed");
|
|
goto error;
|
|
}
|
|
|
|
ret = PyString_FromStringAndSize(NULL, bufsize);
|
|
if (!ret)
|
|
goto error;
|
|
|
|
bzs->next_in = data;
|
|
bzs->avail_in = datasize;
|
|
bzs->next_out = BUF(ret);
|
|
bzs->avail_out = bufsize;
|
|
|
|
totalout = BZS_TOTAL_OUT(bzs);
|
|
|
|
for (;;) {
|
|
Py_BEGIN_ALLOW_THREADS
|
|
bzerror = BZ2_bzCompress(bzs, BZ_RUN);
|
|
Py_END_ALLOW_THREADS
|
|
if (bzerror != BZ_RUN_OK) {
|
|
Util_CatchBZ2Error(bzerror);
|
|
goto error;
|
|
}
|
|
if (bzs->avail_out == 0) {
|
|
bufsize = Util_NewBufferSize(bufsize);
|
|
if (_PyString_Resize(&ret, bufsize) < 0) {
|
|
BZ2_bzCompressEnd(bzs);
|
|
goto error;
|
|
}
|
|
bzs->next_out = BUF(ret) + (BZS_TOTAL_OUT(bzs)
|
|
- totalout);
|
|
bzs->avail_out = bufsize - (bzs->next_out - BUF(ret));
|
|
} else if (bzs->avail_in == 0) {
|
|
break;
|
|
}
|
|
}
|
|
|
|
_PyString_Resize(&ret, (int)(BZS_TOTAL_OUT(bzs) - totalout));
|
|
|
|
RELEASE_LOCK(self);
|
|
return ret;
|
|
|
|
error:
|
|
RELEASE_LOCK(self);
|
|
Py_XDECREF(ret);
|
|
return NULL;
|
|
}
|
|
|
|
PyDoc_STRVAR(BZ2Comp_flush__doc__,
|
|
"flush() -> string\n\
|
|
\n\
|
|
Finish the compression process and return what is left in internal buffers.\n\
|
|
You must not use the compressor object after calling this method.\n\
|
|
");
|
|
|
|
static PyObject *
|
|
BZ2Comp_flush(BZ2CompObject *self)
|
|
{
|
|
int bufsize = SMALLCHUNK;
|
|
PyObject *ret = NULL;
|
|
bz_stream *bzs = &self->bzs;
|
|
LONG_LONG totalout;
|
|
int bzerror;
|
|
|
|
ACQUIRE_LOCK(self);
|
|
if (!self->running) {
|
|
PyErr_SetString(PyExc_ValueError, "object was already "
|
|
"flushed");
|
|
goto error;
|
|
}
|
|
self->running = 0;
|
|
|
|
ret = PyString_FromStringAndSize(NULL, bufsize);
|
|
if (!ret)
|
|
goto error;
|
|
|
|
bzs->next_out = BUF(ret);
|
|
bzs->avail_out = bufsize;
|
|
|
|
totalout = BZS_TOTAL_OUT(bzs);
|
|
|
|
for (;;) {
|
|
Py_BEGIN_ALLOW_THREADS
|
|
bzerror = BZ2_bzCompress(bzs, BZ_FINISH);
|
|
Py_END_ALLOW_THREADS
|
|
if (bzerror == BZ_STREAM_END) {
|
|
break;
|
|
} else if (bzerror != BZ_FINISH_OK) {
|
|
Util_CatchBZ2Error(bzerror);
|
|
goto error;
|
|
}
|
|
if (bzs->avail_out == 0) {
|
|
bufsize = Util_NewBufferSize(bufsize);
|
|
if (_PyString_Resize(&ret, bufsize) < 0)
|
|
goto error;
|
|
bzs->next_out = BUF(ret);
|
|
bzs->next_out = BUF(ret) + (BZS_TOTAL_OUT(bzs)
|
|
- totalout);
|
|
bzs->avail_out = bufsize - (bzs->next_out - BUF(ret));
|
|
}
|
|
}
|
|
|
|
if (bzs->avail_out != 0)
|
|
_PyString_Resize(&ret, (int)(BZS_TOTAL_OUT(bzs) - totalout));
|
|
|
|
RELEASE_LOCK(self);
|
|
return ret;
|
|
|
|
error:
|
|
RELEASE_LOCK(self);
|
|
Py_XDECREF(ret);
|
|
return NULL;
|
|
}
|
|
|
|
static PyMethodDef BZ2Comp_methods[] = {
|
|
{"compress", (PyCFunction)BZ2Comp_compress, METH_VARARGS,
|
|
BZ2Comp_compress__doc__},
|
|
{"flush", (PyCFunction)BZ2Comp_flush, METH_NOARGS,
|
|
BZ2Comp_flush__doc__},
|
|
{NULL, NULL} /* sentinel */
|
|
};
|
|
|
|
|
|
/* ===================================================================== */
|
|
/* Slot definitions for BZ2Comp_Type. */
|
|
|
|
static int
|
|
BZ2Comp_init(BZ2CompObject *self, PyObject *args, PyObject *kwargs)
|
|
{
|
|
int compresslevel = 9;
|
|
int bzerror;
|
|
static char *kwlist[] = {"compresslevel", 0};
|
|
|
|
if (!PyArg_ParseTupleAndKeywords(args, kwargs, "|i:BZ2Compressor",
|
|
kwlist, &compresslevel))
|
|
return -1;
|
|
|
|
if (compresslevel < 1 || compresslevel > 9) {
|
|
PyErr_SetString(PyExc_ValueError,
|
|
"compresslevel must be between 1 and 9");
|
|
goto error;
|
|
}
|
|
|
|
#ifdef WITH_THREAD
|
|
self->lock = PyThread_allocate_lock();
|
|
if (!self->lock)
|
|
goto error;
|
|
#endif
|
|
|
|
memset(&self->bzs, 0, sizeof(bz_stream));
|
|
bzerror = BZ2_bzCompressInit(&self->bzs, compresslevel, 0, 0);
|
|
if (bzerror != BZ_OK) {
|
|
Util_CatchBZ2Error(bzerror);
|
|
goto error;
|
|
}
|
|
|
|
self->running = 1;
|
|
|
|
return 0;
|
|
error:
|
|
#ifdef WITH_THREAD
|
|
if (self->lock)
|
|
PyThread_free_lock(self->lock);
|
|
#endif
|
|
return -1;
|
|
}
|
|
|
|
static void
|
|
BZ2Comp_dealloc(BZ2CompObject *self)
|
|
{
|
|
#ifdef WITH_THREAD
|
|
if (self->lock)
|
|
PyThread_free_lock(self->lock);
|
|
#endif
|
|
BZ2_bzCompressEnd(&self->bzs);
|
|
((PyObject*)self)->ob_type->tp_free((PyObject *)self);
|
|
}
|
|
|
|
|
|
/* ===================================================================== */
|
|
/* BZ2Comp_Type definition. */
|
|
|
|
PyDoc_STRVAR(BZ2Comp__doc__,
|
|
"BZ2Compressor([compresslevel=9]) -> compressor object\n\
|
|
\n\
|
|
Create a new compressor object. This object may be used to compress\n\
|
|
data sequentially. If you want to compress data in one shot, use the\n\
|
|
compress() function instead. The compresslevel parameter, if given,\n\
|
|
must be a number between 1 and 9.\n\
|
|
");
|
|
|
|
static PyTypeObject BZ2Comp_Type = {
|
|
PyObject_HEAD_INIT(NULL)
|
|
0, /*ob_size*/
|
|
"bz2.BZ2Compressor", /*tp_name*/
|
|
sizeof(BZ2CompObject), /*tp_basicsize*/
|
|
0, /*tp_itemsize*/
|
|
(destructor)BZ2Comp_dealloc, /*tp_dealloc*/
|
|
0, /*tp_print*/
|
|
0, /*tp_getattr*/
|
|
0, /*tp_setattr*/
|
|
0, /*tp_compare*/
|
|
0, /*tp_repr*/
|
|
0, /*tp_as_number*/
|
|
0, /*tp_as_sequence*/
|
|
0, /*tp_as_mapping*/
|
|
0, /*tp_hash*/
|
|
0, /*tp_call*/
|
|
0, /*tp_str*/
|
|
0, /*tp_getattro*/
|
|
0, /*tp_setattro*/
|
|
0, /*tp_as_buffer*/
|
|
Py_TPFLAGS_DEFAULT|Py_TPFLAGS_BASETYPE, /*tp_flags*/
|
|
BZ2Comp__doc__, /*tp_doc*/
|
|
0, /*tp_traverse*/
|
|
0, /*tp_clear*/
|
|
0, /*tp_richcompare*/
|
|
0, /*tp_weaklistoffset*/
|
|
0, /*tp_iter*/
|
|
0, /*tp_iternext*/
|
|
BZ2Comp_methods, /*tp_methods*/
|
|
0, /*tp_members*/
|
|
0, /*tp_getset*/
|
|
0, /*tp_base*/
|
|
0, /*tp_dict*/
|
|
0, /*tp_descr_get*/
|
|
0, /*tp_descr_set*/
|
|
0, /*tp_dictoffset*/
|
|
(initproc)BZ2Comp_init, /*tp_init*/
|
|
0, /*tp_alloc*/
|
|
0, /*tp_new*/
|
|
0, /*tp_free*/
|
|
0, /*tp_is_gc*/
|
|
};
|
|
|
|
|
|
/* ===================================================================== */
|
|
/* Members of BZ2Decomp. */
|
|
|
|
#define OFF(x) offsetof(BZ2DecompObject, x)
|
|
|
|
static PyMemberDef BZ2Decomp_members[] = {
|
|
{"unused_data", T_OBJECT, OFF(unused_data), RO},
|
|
{NULL} /* Sentinel */
|
|
};
|
|
|
|
|
|
/* ===================================================================== */
|
|
/* Methods of BZ2Decomp. */
|
|
|
|
PyDoc_STRVAR(BZ2Decomp_decompress__doc__,
|
|
"decompress(data) -> string\n\
|
|
\n\
|
|
Provide more data to the decompressor object. It will return chunks\n\
|
|
of decompressed data whenever possible. If you try to decompress data\n\
|
|
after the end of stream is found, EOFError will be raised. If any data\n\
|
|
was found after the end of stream, it'll be ignored and saved in\n\
|
|
unused_data attribute.\n\
|
|
");
|
|
|
|
static PyObject *
|
|
BZ2Decomp_decompress(BZ2DecompObject *self, PyObject *args)
|
|
{
|
|
char *data;
|
|
int datasize;
|
|
int bufsize = SMALLCHUNK;
|
|
LONG_LONG totalout;
|
|
PyObject *ret = NULL;
|
|
bz_stream *bzs = &self->bzs;
|
|
int bzerror;
|
|
|
|
if (!PyArg_ParseTuple(args, "s#", &data, &datasize))
|
|
return NULL;
|
|
|
|
ACQUIRE_LOCK(self);
|
|
if (!self->running) {
|
|
PyErr_SetString(PyExc_EOFError, "end of stream was "
|
|
"already found");
|
|
goto error;
|
|
}
|
|
|
|
ret = PyString_FromStringAndSize(NULL, bufsize);
|
|
if (!ret)
|
|
goto error;
|
|
|
|
bzs->next_in = data;
|
|
bzs->avail_in = datasize;
|
|
bzs->next_out = BUF(ret);
|
|
bzs->avail_out = bufsize;
|
|
|
|
totalout = BZS_TOTAL_OUT(bzs);
|
|
|
|
for (;;) {
|
|
Py_BEGIN_ALLOW_THREADS
|
|
bzerror = BZ2_bzDecompress(bzs);
|
|
Py_END_ALLOW_THREADS
|
|
if (bzerror == BZ_STREAM_END) {
|
|
if (bzs->avail_in != 0) {
|
|
Py_DECREF(self->unused_data);
|
|
self->unused_data =
|
|
PyString_FromStringAndSize(bzs->next_in,
|
|
bzs->avail_in);
|
|
}
|
|
self->running = 0;
|
|
break;
|
|
}
|
|
if (bzerror != BZ_OK) {
|
|
Util_CatchBZ2Error(bzerror);
|
|
goto error;
|
|
}
|
|
if (bzs->avail_out == 0) {
|
|
bufsize = Util_NewBufferSize(bufsize);
|
|
if (_PyString_Resize(&ret, bufsize) < 0) {
|
|
BZ2_bzDecompressEnd(bzs);
|
|
goto error;
|
|
}
|
|
bzs->next_out = BUF(ret);
|
|
bzs->next_out = BUF(ret) + (BZS_TOTAL_OUT(bzs)
|
|
- totalout);
|
|
bzs->avail_out = bufsize - (bzs->next_out - BUF(ret));
|
|
} else if (bzs->avail_in == 0) {
|
|
break;
|
|
}
|
|
}
|
|
|
|
if (bzs->avail_out != 0)
|
|
_PyString_Resize(&ret, (int)(BZS_TOTAL_OUT(bzs) - totalout));
|
|
|
|
RELEASE_LOCK(self);
|
|
return ret;
|
|
|
|
error:
|
|
RELEASE_LOCK(self);
|
|
Py_XDECREF(ret);
|
|
return NULL;
|
|
}
|
|
|
|
static PyMethodDef BZ2Decomp_methods[] = {
|
|
{"decompress", (PyCFunction)BZ2Decomp_decompress, METH_VARARGS, BZ2Decomp_decompress__doc__},
|
|
{NULL, NULL} /* sentinel */
|
|
};
|
|
|
|
|
|
/* ===================================================================== */
|
|
/* Slot definitions for BZ2Decomp_Type. */
|
|
|
|
static int
|
|
BZ2Decomp_init(BZ2DecompObject *self, PyObject *args, PyObject *kwargs)
|
|
{
|
|
int bzerror;
|
|
|
|
if (!PyArg_ParseTuple(args, ":BZ2Decompressor"))
|
|
return -1;
|
|
|
|
#ifdef WITH_THREAD
|
|
self->lock = PyThread_allocate_lock();
|
|
if (!self->lock)
|
|
goto error;
|
|
#endif
|
|
|
|
self->unused_data = PyString_FromString("");
|
|
if (!self->unused_data)
|
|
goto error;
|
|
|
|
memset(&self->bzs, 0, sizeof(bz_stream));
|
|
bzerror = BZ2_bzDecompressInit(&self->bzs, 0, 0);
|
|
if (bzerror != BZ_OK) {
|
|
Util_CatchBZ2Error(bzerror);
|
|
goto error;
|
|
}
|
|
|
|
self->running = 1;
|
|
|
|
return 0;
|
|
|
|
error:
|
|
#ifdef WITH_THREAD
|
|
if (self->lock)
|
|
PyThread_free_lock(self->lock);
|
|
#endif
|
|
Py_XDECREF(self->unused_data);
|
|
return -1;
|
|
}
|
|
|
|
static void
|
|
BZ2Decomp_dealloc(BZ2DecompObject *self)
|
|
{
|
|
#ifdef WITH_THREAD
|
|
if (self->lock)
|
|
PyThread_free_lock(self->lock);
|
|
#endif
|
|
Py_XDECREF(self->unused_data);
|
|
BZ2_bzDecompressEnd(&self->bzs);
|
|
((PyObject*)self)->ob_type->tp_free((PyObject *)self);
|
|
}
|
|
|
|
|
|
/* ===================================================================== */
|
|
/* BZ2Decomp_Type definition. */
|
|
|
|
PyDoc_STRVAR(BZ2Decomp__doc__,
|
|
"BZ2Decompressor() -> decompressor object\n\
|
|
\n\
|
|
Create a new decompressor object. This object may be used to decompress\n\
|
|
data sequentially. If you want to decompress data in one shot, use the\n\
|
|
decompress() function instead.\n\
|
|
");
|
|
|
|
static PyTypeObject BZ2Decomp_Type = {
|
|
PyObject_HEAD_INIT(NULL)
|
|
0, /*ob_size*/
|
|
"bz2.BZ2Decompressor", /*tp_name*/
|
|
sizeof(BZ2DecompObject), /*tp_basicsize*/
|
|
0, /*tp_itemsize*/
|
|
(destructor)BZ2Decomp_dealloc, /*tp_dealloc*/
|
|
0, /*tp_print*/
|
|
0, /*tp_getattr*/
|
|
0, /*tp_setattr*/
|
|
0, /*tp_compare*/
|
|
0, /*tp_repr*/
|
|
0, /*tp_as_number*/
|
|
0, /*tp_as_sequence*/
|
|
0, /*tp_as_mapping*/
|
|
0, /*tp_hash*/
|
|
0, /*tp_call*/
|
|
0, /*tp_str*/
|
|
0, /*tp_getattro*/
|
|
0, /*tp_setattro*/
|
|
0, /*tp_as_buffer*/
|
|
Py_TPFLAGS_DEFAULT|Py_TPFLAGS_BASETYPE, /*tp_flags*/
|
|
BZ2Decomp__doc__, /*tp_doc*/
|
|
0, /*tp_traverse*/
|
|
0, /*tp_clear*/
|
|
0, /*tp_richcompare*/
|
|
0, /*tp_weaklistoffset*/
|
|
0, /*tp_iter*/
|
|
0, /*tp_iternext*/
|
|
BZ2Decomp_methods, /*tp_methods*/
|
|
BZ2Decomp_members, /*tp_members*/
|
|
0, /*tp_getset*/
|
|
0, /*tp_base*/
|
|
0, /*tp_dict*/
|
|
0, /*tp_descr_get*/
|
|
0, /*tp_descr_set*/
|
|
0, /*tp_dictoffset*/
|
|
(initproc)BZ2Decomp_init, /*tp_init*/
|
|
0, /*tp_alloc*/
|
|
0, /*tp_new*/
|
|
0, /*tp_free*/
|
|
0, /*tp_is_gc*/
|
|
};
|
|
|
|
|
|
/* ===================================================================== */
|
|
/* Module functions. */
|
|
|
|
PyDoc_STRVAR(bz2_compress__doc__,
|
|
"compress(data [, compresslevel=9]) -> string\n\
|
|
\n\
|
|
Compress data in one shot. If you want to compress data sequentially,\n\
|
|
use an instance of BZ2Compressor instead. The compresslevel parameter, if\n\
|
|
given, must be a number between 1 and 9.\n\
|
|
");
|
|
|
|
static PyObject *
|
|
bz2_compress(PyObject *self, PyObject *args, PyObject *kwargs)
|
|
{
|
|
int compresslevel=9;
|
|
char *data;
|
|
int datasize;
|
|
int bufsize;
|
|
PyObject *ret = NULL;
|
|
bz_stream _bzs;
|
|
bz_stream *bzs = &_bzs;
|
|
int bzerror;
|
|
static char *kwlist[] = {"data", "compresslevel", 0};
|
|
|
|
if (!PyArg_ParseTupleAndKeywords(args, kwargs, "s#|i",
|
|
kwlist, &data, &datasize,
|
|
&compresslevel))
|
|
return NULL;
|
|
|
|
if (compresslevel < 1 || compresslevel > 9) {
|
|
PyErr_SetString(PyExc_ValueError,
|
|
"compresslevel must be between 1 and 9");
|
|
return NULL;
|
|
}
|
|
|
|
/* Conforming to bz2 manual, this is large enough to fit compressed
|
|
* data in one shot. We will check it later anyway. */
|
|
bufsize = datasize + (datasize/100+1) + 600;
|
|
|
|
ret = PyString_FromStringAndSize(NULL, bufsize);
|
|
if (!ret)
|
|
return NULL;
|
|
|
|
memset(bzs, 0, sizeof(bz_stream));
|
|
|
|
bzs->next_in = data;
|
|
bzs->avail_in = datasize;
|
|
bzs->next_out = BUF(ret);
|
|
bzs->avail_out = bufsize;
|
|
|
|
bzerror = BZ2_bzCompressInit(bzs, compresslevel, 0, 0);
|
|
if (bzerror != BZ_OK) {
|
|
Util_CatchBZ2Error(bzerror);
|
|
Py_DECREF(ret);
|
|
return NULL;
|
|
}
|
|
|
|
for (;;) {
|
|
Py_BEGIN_ALLOW_THREADS
|
|
bzerror = BZ2_bzCompress(bzs, BZ_FINISH);
|
|
Py_END_ALLOW_THREADS
|
|
if (bzerror == BZ_STREAM_END) {
|
|
break;
|
|
} else if (bzerror != BZ_FINISH_OK) {
|
|
BZ2_bzCompressEnd(bzs);
|
|
Util_CatchBZ2Error(bzerror);
|
|
Py_DECREF(ret);
|
|
return NULL;
|
|
}
|
|
if (bzs->avail_out == 0) {
|
|
bufsize = Util_NewBufferSize(bufsize);
|
|
if (_PyString_Resize(&ret, bufsize) < 0) {
|
|
BZ2_bzCompressEnd(bzs);
|
|
Py_DECREF(ret);
|
|
return NULL;
|
|
}
|
|
bzs->next_out = BUF(ret) + BZS_TOTAL_OUT(bzs);
|
|
bzs->avail_out = bufsize - (bzs->next_out - BUF(ret));
|
|
}
|
|
}
|
|
|
|
if (bzs->avail_out != 0)
|
|
_PyString_Resize(&ret, (int)BZS_TOTAL_OUT(bzs));
|
|
BZ2_bzCompressEnd(bzs);
|
|
|
|
return ret;
|
|
}
|
|
|
|
PyDoc_STRVAR(bz2_decompress__doc__,
|
|
"decompress(data) -> decompressed data\n\
|
|
\n\
|
|
Decompress data in one shot. If you want to decompress data sequentially,\n\
|
|
use an instance of BZ2Decompressor instead.\n\
|
|
");
|
|
|
|
static PyObject *
|
|
bz2_decompress(PyObject *self, PyObject *args)
|
|
{
|
|
char *data;
|
|
int datasize;
|
|
int bufsize = SMALLCHUNK;
|
|
PyObject *ret;
|
|
bz_stream _bzs;
|
|
bz_stream *bzs = &_bzs;
|
|
int bzerror;
|
|
|
|
if (!PyArg_ParseTuple(args, "s#", &data, &datasize))
|
|
return NULL;
|
|
|
|
if (datasize == 0)
|
|
return PyString_FromString("");
|
|
|
|
ret = PyString_FromStringAndSize(NULL, bufsize);
|
|
if (!ret)
|
|
return NULL;
|
|
|
|
memset(bzs, 0, sizeof(bz_stream));
|
|
|
|
bzs->next_in = data;
|
|
bzs->avail_in = datasize;
|
|
bzs->next_out = BUF(ret);
|
|
bzs->avail_out = bufsize;
|
|
|
|
bzerror = BZ2_bzDecompressInit(bzs, 0, 0);
|
|
if (bzerror != BZ_OK) {
|
|
Util_CatchBZ2Error(bzerror);
|
|
Py_DECREF(ret);
|
|
return NULL;
|
|
}
|
|
|
|
for (;;) {
|
|
Py_BEGIN_ALLOW_THREADS
|
|
bzerror = BZ2_bzDecompress(bzs);
|
|
Py_END_ALLOW_THREADS
|
|
if (bzerror == BZ_STREAM_END) {
|
|
break;
|
|
} else if (bzerror != BZ_OK) {
|
|
BZ2_bzDecompressEnd(bzs);
|
|
Util_CatchBZ2Error(bzerror);
|
|
Py_DECREF(ret);
|
|
return NULL;
|
|
}
|
|
if (bzs->avail_out == 0) {
|
|
bufsize = Util_NewBufferSize(bufsize);
|
|
if (_PyString_Resize(&ret, bufsize) < 0) {
|
|
BZ2_bzDecompressEnd(bzs);
|
|
Py_DECREF(ret);
|
|
return NULL;
|
|
}
|
|
bzs->next_out = BUF(ret) + BZS_TOTAL_OUT(bzs);
|
|
bzs->avail_out = bufsize - (bzs->next_out - BUF(ret));
|
|
} else if (bzs->avail_in == 0) {
|
|
BZ2_bzDecompressEnd(bzs);
|
|
PyErr_SetString(PyExc_ValueError,
|
|
"couldn't find end of stream");
|
|
Py_DECREF(ret);
|
|
return NULL;
|
|
}
|
|
}
|
|
|
|
if (bzs->avail_out != 0)
|
|
_PyString_Resize(&ret, (int)BZS_TOTAL_OUT(bzs));
|
|
BZ2_bzDecompressEnd(bzs);
|
|
|
|
return ret;
|
|
}
|
|
|
|
static PyMethodDef bz2_methods[] = {
|
|
{"compress", (PyCFunction) bz2_compress, METH_VARARGS|METH_KEYWORDS,
|
|
bz2_compress__doc__},
|
|
{"decompress", (PyCFunction) bz2_decompress, METH_VARARGS,
|
|
bz2_decompress__doc__},
|
|
{NULL, NULL} /* sentinel */
|
|
};
|
|
|
|
/* ===================================================================== */
|
|
/* Initialization function. */
|
|
|
|
PyDoc_STRVAR(bz2__doc__,
|
|
"The python bz2 module provides a comprehensive interface for\n\
|
|
the bz2 compression library. It implements a complete file\n\
|
|
interface, one shot (de)compression functions, and types for\n\
|
|
sequential (de)compression.\n\
|
|
");
|
|
|
|
DL_EXPORT(void)
|
|
initbz2(void)
|
|
{
|
|
PyObject *m;
|
|
|
|
BZ2File_Type.ob_type = &PyType_Type;
|
|
BZ2File_Type.tp_base = &PyFile_Type;
|
|
BZ2File_Type.tp_new = PyFile_Type.tp_new;
|
|
BZ2File_Type.tp_getattro = PyObject_GenericGetAttr;
|
|
BZ2File_Type.tp_setattro = PyObject_GenericSetAttr;
|
|
BZ2File_Type.tp_alloc = PyType_GenericAlloc;
|
|
BZ2File_Type.tp_free = _PyObject_Del;
|
|
|
|
BZ2Comp_Type.ob_type = &PyType_Type;
|
|
BZ2Comp_Type.tp_getattro = PyObject_GenericGetAttr;
|
|
BZ2Comp_Type.tp_setattro = PyObject_GenericSetAttr;
|
|
BZ2Comp_Type.tp_alloc = PyType_GenericAlloc;
|
|
BZ2Comp_Type.tp_new = PyType_GenericNew;
|
|
BZ2Comp_Type.tp_free = _PyObject_Del;
|
|
|
|
BZ2Decomp_Type.ob_type = &PyType_Type;
|
|
BZ2Decomp_Type.tp_getattro = PyObject_GenericGetAttr;
|
|
BZ2Decomp_Type.tp_setattro = PyObject_GenericSetAttr;
|
|
BZ2Decomp_Type.tp_alloc = PyType_GenericAlloc;
|
|
BZ2Decomp_Type.tp_new = PyType_GenericNew;
|
|
BZ2Decomp_Type.tp_free = _PyObject_Del;
|
|
|
|
m = Py_InitModule3("bz2", bz2_methods, bz2__doc__);
|
|
|
|
PyModule_AddObject(m, "__author__", PyString_FromString(__author__));
|
|
|
|
Py_INCREF(&BZ2File_Type);
|
|
PyModule_AddObject(m, "BZ2File", (PyObject *)&BZ2File_Type);
|
|
|
|
Py_INCREF(&BZ2Comp_Type);
|
|
PyModule_AddObject(m, "BZ2Compressor", (PyObject *)&BZ2Comp_Type);
|
|
|
|
Py_INCREF(&BZ2Decomp_Type);
|
|
PyModule_AddObject(m, "BZ2Decompressor", (PyObject *)&BZ2Decomp_Type);
|
|
}
|