Streaming, Serialization, and IPC¶
Writing and Reading Streams¶
Arrow defines two types of binary formats for serializing record batches:
Streaming format: for sending an arbitrary length sequence of record batches. The format must be processed from start to end, and does not support random access
File or Random Access format: for serializing a fixed number of record batches. Supports random access, and thus is very useful when used with memory maps
To follow this section, make sure to first read the section on Memory and IO.
Using streams¶
First, let’s create a small record batch:
In [1]: import pyarrow as pa
In [2]: data = [
...: pa.array([1, 2, 3, 4]),
...: pa.array(['foo', 'bar', 'baz', None]),
...: pa.array([True, None, False, True])
...: ]
...:
In [3]: batch = pa.record_batch(data, names=['f0', 'f1', 'f2'])
In [4]: batch.num_rows
Out[4]: 4
In [5]: batch.num_columns
Out[5]: 3
Now, we can begin writing a stream containing some number of these batches. For
this we use RecordBatchStreamWriter
, which can write to a
writeable NativeFile
object or a writeable Python object. For convenience,
this one can be created with new_stream()
:
In [6]: sink = pa.BufferOutputStream()
In [7]: with pa.ipc.new_stream(sink, batch.schema) as writer:
...: for i in range(5):
...: writer.write_batch(batch)
...:
Here we used an in-memory Arrow buffer stream (sink
),
but this could have been a socket or some other IO sink.
When creating the StreamWriter
, we pass the schema, since the schema
(column names and types) must be the same for all of the batches sent in this
particular stream. Now we can do:
In [8]: buf = sink.getvalue()
In [9]: buf.size
Out[9]: 1984
Now buf
contains the complete stream as an in-memory byte buffer. We can
read such a stream with RecordBatchStreamReader
or the
convenience function pyarrow.ipc.open_stream
:
In [10]: with pa.ipc.open_stream(buf) as reader:
....: schema = reader.schema
....: batches = [b for b in reader]
....:
In [11]: schema
Out[11]:
f0: int64
f1: string
f2: bool
In [12]: len(batches)
Out[12]: 5
We can check the returned batches are the same as the original input:
In [13]: batches[0].equals(batch)
Out[13]: True
An important point is that if the input source supports zero-copy reads
(e.g. like a memory map, or pyarrow.BufferReader
), then the returned
batches are also zero-copy and do not allocate any new memory on read.
Writing and Reading Random Access Files¶
The RecordBatchFileWriter
has the same API as
RecordBatchStreamWriter
. You can create one with
new_file()
:
In [14]: sink = pa.BufferOutputStream()
In [15]: with pa.ipc.new_file(sink, batch.schema) as writer:
....: for i in range(10):
....: writer.write_batch(batch)
....:
In [16]: buf = sink.getvalue()
In [17]: buf.size
Out[17]: 4226
The difference between RecordBatchFileReader
and
RecordBatchStreamReader
is that the input source must have a
seek
method for random access. The stream reader only requires read
operations. We can also use the open_file()
method to open a file:
In [18]: with pa.ipc.open_file(buf) as reader:
....: num_record_batches = reader.num_record_batches
....:
In [19]: b = reader.get_batch(3)
Because we have access to the entire payload, we know the number of record batches in the file, and can read any at random.
In [20]: num_record_batches
Out[20]: 10
In [21]: b.equals(batch)
Out[21]: True
Reading from Stream and File Format for pandas¶
The stream and file reader classes have a special read_pandas
method to
simplify reading multiple record batches and converting them to a single
DataFrame output:
In [22]: with pa.ipc.open_file(buf) as reader:
....: df = reader.read_pandas()
....:
In [23]: df[:5]
Out[23]:
f0 f1 f2
0 1 foo True
1 2 bar None
2 3 baz False
3 4 None True
4 1 foo True
Efficiently Writing and Reading Arrow Data¶
Being optimized for zero copy and memory mapped data, Arrow allows to easily read and write arrays consuming the minimum amount of resident memory.
When writing and reading raw Arrow data, we can use the Arrow File Format or the Arrow Streaming Format.
To dump an array to file, you can use the new_file()
which will provide a new RecordBatchFileWriter
instance
that can be used to write batches of data to that file.
For example to write an array of 10M integers, we could write it in 1000 chunks of 10000 entries:
In [24]: BATCH_SIZE = 10000
In [25]: NUM_BATCHES = 1000
In [26]: schema = pa.schema([pa.field('nums', pa.int32())])
In [27]: with pa.OSFile('bigfile.arrow', 'wb') as sink:
....: with pa.ipc.new_file(sink, schema) as writer:
....: for row in range(NUM_BATCHES):
....: batch = pa.record_batch([pa.array(range(BATCH_SIZE), type=pa.int32())], schema)
....: writer.write(batch)
....:
record batches support multiple columns, so in practice we always write the
equivalent of a Table
.
Writing in batches is effective because we in theory need to keep in memory only the current batch we are writing. But when reading back, we can be even more effective by directly mapping the data from disk and avoid allocating any new memory on read.
Under normal conditions, reading back our file will consume a few hundred megabytes of memory:
In [28]: with pa.OSFile('bigfile.arrow', 'rb') as source:
....: loaded_array = pa.ipc.open_file(source).read_all()
....:
In [29]: print("LEN:", len(loaded_array))
LEN: 10000000
In [30]: print("RSS: {}MB".format(pa.total_allocated_bytes() >> 20))
RSS: 38MB
To more efficiently read big data from disk, we can memory map the file, so that Arrow can directly reference the data mapped from disk and avoid having to allocate its own memory. In such case the operating system will be able to page in the mapped memory lazily and page it out without any write back cost when under pressure, allowing to more easily read arrays bigger than the total memory.
In [31]: with pa.memory_map('bigfile.arrow', 'rb') as source:
....: loaded_array = pa.ipc.open_file(source).read_all()
....:
In [32]: print("LEN:", len(loaded_array))
LEN: 10000000
In [33]: print("RSS: {}MB".format(pa.total_allocated_bytes() >> 20))
RSS: 0MB
Note
Other high level APIs like read_table()
also provide a
memory_map
option. But in those cases, the memory mapping can’t help with
reducing resident memory consumption. See Reading Parquet and Memory Mapping for details.
Arbitrary Object Serialization¶
Warning
The custom serialization functionality is deprecated in pyarrow 2.0, and will be removed in a future version.
While the serialization functions in this section utilize the Arrow stream
protocol internally, they do not produce data that is compatible with the
above ipc.open_file
and ipc.open_stream
functions.
For arbitrary objects, you can use the standard library pickle
functionality instead. For pyarrow objects, you can use the IPC
serialization format through the pyarrow.ipc
module, as explained
above.
PyArrow serialization was originally meant to provide a higher-performance
alternative to pickle
thanks to zero-copy semantics. However,
pickle
protocol 5 gained support for zero-copy using out-of-band
buffers, and can be used instead for similar benefits.
In pyarrow
we are able to serialize and deserialize many kinds of Python
objects. As an example, consider a dictionary containing NumPy arrays:
In [34]: import numpy as np
In [35]: data = {
....: i: np.random.randn(500, 500)
....: for i in range(100)
....: }
....:
We use the pyarrow.serialize
function to convert this data to a byte
buffer:
In [36]: buf = pa.serialize(data).to_buffer()
In [37]: type(buf)
Out[37]: pyarrow.lib.Buffer
In [38]: buf.size
Out[38]: 200028928
pyarrow.serialize
creates an intermediate object which can be converted to
a buffer (the to_buffer
method) or written directly to an output stream.
pyarrow.deserialize
converts a buffer-like object back to the original
Python object:
In [39]: restored_data = pa.deserialize(buf)
In [40]: restored_data[0]
Out[40]:
array([[ 0.97956465, -0.43815265, -1.18364808, ..., 0.35111594,
-0.71647183, -0.66108611],
[ 0.24558811, -0.52441559, -0.44386121, ..., 0.13860664,
-1.58523707, -0.87087192],
[ 0.91167483, 2.37571145, 1.05785704, ..., -0.28283493,
-0.46741056, 1.01571402],
...,
[ 1.28581923, 0.7127101 , 0.4776438 , ..., -0.61280796,
-1.16608896, -0.29708614],
[ 1.51817581, -2.06175867, 0.02850169, ..., 0.80922794,
0.05205157, -0.7339787 ],
[-0.1629145 , -0.66528386, -0.03009694, ..., -0.16359135,
0.72314738, 0.43702372]])
Serializing Custom Data Types¶
If an unrecognized data type is encountered when serializing an object,
pyarrow
will fall back on using pickle
for converting that type to a
byte string. There may be a more efficient way, though.
Consider a class with two members, one of which is a NumPy array:
class MyData:
def __init__(self, name, data):
self.name = name
self.data = data
We write functions to convert this to and from a dictionary with simpler types:
def _serialize_MyData(val):
return {'name': val.name, 'data': val.data}
def _deserialize_MyData(data):
return MyData(data['name'], data['data']
then, we must register these functions in a SerializationContext
so that
MyData
can be recognized:
context = pa.SerializationContext()
context.register_type(MyData, 'MyData',
custom_serializer=_serialize_MyData,
custom_deserializer=_deserialize_MyData)
Lastly, we use this context as an additional argument to pyarrow.serialize
:
buf = pa.serialize(val, context=context).to_buffer()
restored_val = pa.deserialize(buf, context=context)
The SerializationContext
also has convenience methods serialize
and
deserialize
, so these are equivalent statements:
buf = context.serialize(val).to_buffer()
restored_val = context.deserialize(buf)
Component-based Serialization¶
For serializing Python objects containing some number of NumPy arrays, Arrow
buffers, or other data types, it may be desirable to transport their serialized
representation without having to produce an intermediate copy using the
to_buffer
method. To motivate this, suppose we have a list of NumPy arrays:
In [41]: import numpy as np
In [42]: data = [np.random.randn(10, 10) for i in range(5)]
The call pa.serialize(data)
does not copy the memory inside each of these
NumPy arrays. This serialized representation can be then decomposed into a
dictionary containing a sequence of pyarrow.Buffer
objects containing
metadata for each array and references to the memory inside the arrays. To do
this, use the to_components
method:
In [43]: serialized = pa.serialize(data)
In [44]: components = serialized.to_components()
The particular details of the output of to_components
are not too
important. The objects in the 'data'
field are pyarrow.Buffer
objects,
which are zero-copy convertible to Python memoryview
objects:
In [45]: memoryview(components['data'][0])
Out[45]: <memory at 0x7f554e27fa00>
A memoryview can be converted back to a Arrow Buffer
with
pyarrow.py_buffer
:
In [46]: mv = memoryview(components['data'][0])
In [47]: buf = pa.py_buffer(mv)
An object can be reconstructed from its component-based representation using
deserialize_components
:
In [48]: restored_data = pa.deserialize_components(components)
In [49]: restored_data[0]
Out[49]:
array([[ 0.15405192, -1.89691648, 1.91123496, 1.20678999, -0.72872127,
1.13618062, -0.0109724 , -0.66596511, -1.57619739, 0.24086739],
[-1.12051462, -1.7785293 , 1.61868304, -1.98418646, 0.9279108 ,
-0.75383802, -0.37204145, 0.75505408, -1.41052382, 1.69605963],
[ 0.25575821, 0.78383125, -0.13768737, -1.48913567, -0.18957406,
-1.9003326 , -1.87557182, 1.39887766, 0.30055991, 0.06515482],
[-0.52767355, 0.39300569, 0.24267334, 0.04720361, -0.87532373,
-0.6400917 , -0.16335227, -0.19619457, 0.96717332, -0.96515108],
[-0.03407856, 1.07316338, -2.4538942 , 0.83285584, 0.88774567,
-0.74674442, -1.98508259, -0.01434152, -0.11095245, -0.14543038],
[-0.81546323, 0.33614943, -0.63915333, -0.13789921, -0.7438201 ,
0.15560259, 0.79711476, -0.76994378, 1.44792185, 1.40927863],
[ 0.34514766, -0.94452552, -1.64195498, -1.35263511, 1.1553221 ,
-0.7131641 , -1.52515459, 0.32283589, -0.50853271, -0.59899314],
[ 0.78379451, -0.4686877 , -0.12630493, -0.77553473, -0.72181552,
-0.67757499, -0.35683496, -0.40838114, 0.21381165, -2.30721251],
[ 0.89797548, 1.3460108 , -1.48832741, 0.74504223, 0.69061888,
0.30011252, 0.51167582, -0.32839633, 0.72025667, -2.20418585],
[ 0.33067039, -0.02858438, 0.12896881, -0.26793734, 1.22449541,
-0.04449203, -0.22875291, 1.28817077, -0.8902981 , -0.80530942]])
deserialize_components
is also available as a method on
SerializationContext
objects.