Commit 5a8423e3 authored by Yuxin Wu's avatar Yuxin Wu

add strict mode in ThreadedMapData (#139)

parent 545fc709
...@@ -364,7 +364,8 @@ class RandomChooseData(RNGDataFlow): ...@@ -364,7 +364,8 @@ class RandomChooseData(RNGDataFlow):
class RandomMixData(RNGDataFlow): class RandomMixData(RNGDataFlow):
""" """
Perfectly mix datapoints from several DataFlow. Will stop when all DataFlow exhausted. Perfectly mix datapoints from several DataFlow using their :meth:`size()`.
Will stop when all DataFlow exhausted.
""" """
def __init__(self, df_lists): def __init__(self, df_lists):
...@@ -395,6 +396,7 @@ class RandomMixData(RNGDataFlow): ...@@ -395,6 +396,7 @@ class RandomMixData(RNGDataFlow):
assert idxs.max() == len(itrs) - 1, "{}!={}".format(idxs.max(), len(itrs) - 1) assert idxs.max() == len(itrs) - 1, "{}!={}".format(idxs.max(), len(itrs) - 1)
for k in idxs: for k in idxs:
yield next(itrs[k]) yield next(itrs[k])
# TODO run till exception
class ConcatData(DataFlow): class ConcatData(DataFlow):
......
...@@ -12,7 +12,6 @@ import os ...@@ -12,7 +12,6 @@ import os
import zmq import zmq
from .base import ProxyDataFlow, DataFlowTerminated from .base import ProxyDataFlow, DataFlowTerminated
from .common import RepeatedData
from ..utils.concurrency import (ensure_proc_terminate, from ..utils.concurrency import (ensure_proc_terminate,
mask_sigint, start_proc_mask_signal, mask_sigint, start_proc_mask_signal,
StoppableThread) StoppableThread)
...@@ -231,19 +230,30 @@ class ThreadedMapData(ProxyDataFlow): ...@@ -231,19 +230,30 @@ class ThreadedMapData(ProxyDataFlow):
This is useful when the mapping function is the bottleneck, but you don't This is useful when the mapping function is the bottleneck, but you don't
want to start processes for the entire dataflow pipeline. want to start processes for the entire dataflow pipeline.
With threads, there are tiny communication overhead, but due to GIL, you Notes:
should avoid starting the threads in your main process. 1. There is tiny communication overhead with threads, but you
Note that the threads will only start in the process which calls should avoid starting many threads in your main process to avoid GIL.
:meth:`reset_state()`.
So you can use ``PrefetchDataZMQ(ThreadedMapData(...), 1)`` to avoid GIL. The threads will only start in the process which calls :meth:`reset_state()`.
Therefore you can use ``PrefetchDataZMQ(ThreadedMapData(...), 1)`` to avoid GIL.
2. Threads run in parallel and can take different time to run the
mapping function. Therefore the order of datapoints won't be
preserved, and datapoints from one pass of `df.get_data()` might get
mixed with datapoints from the next pass.
You can use **strict mode**, where `ThreadedMapData.get_data()`
is guranteed to produce the exact set which `df.get_data()`
produces. Although the order of data still isn't preserved.
""" """
class _WorkerThread(StoppableThread): class _WorkerThread(StoppableThread):
def __init__(self, inq, outq, map_func): def __init__(self, inq, outq, map_func, strict):
super(ThreadedMapData._WorkerThread, self).__init__() super(ThreadedMapData._WorkerThread, self).__init__()
self.inq = inq self.inq = inq
self.outq = outq self.outq = outq
self.func = map_func self.func = map_func
self.daemon = True self.daemon = True
self._strict = strict
def run(self): def run(self):
while not self.stopped(): while not self.stopped():
...@@ -251,18 +261,23 @@ class ThreadedMapData(ProxyDataFlow): ...@@ -251,18 +261,23 @@ class ThreadedMapData(ProxyDataFlow):
dp = self.func(dp) dp = self.func(dp)
if dp is not None: if dp is not None:
self.queue_put_stoppable(self.outq, dp) self.queue_put_stoppable(self.outq, dp)
else:
assert not self._strict, \
"[ThreadedMapData] Map function cannot return None when strict mode is used."
def __init__(self, ds, nr_thread, map_func, buffer_size=200): def __init__(self, ds, nr_thread, map_func, buffer_size=200, strict=False):
""" """
Args: Args:
ds (DataFlow): the dataflow to map ds (DataFlow): the dataflow to map
nr_thread (int): number of threads to use nr_thread (int): number of threads to use
map_func (callable): datapoint -> datapoint | None map_func (callable): datapoint -> datapoint | None
buffer_size (int): number of datapoints in the buffer buffer_size (int): number of datapoints in the buffer
strict (bool): use "strict mode", see notes above.
""" """
super(ThreadedMapData, self).__init__(ds) super(ThreadedMapData, self).__init__(ds)
self._iter_ds = RepeatedData(ds, -1) self._iter_ds = ds
self._strict = strict
self.nr_thread = nr_thread self.nr_thread = nr_thread
self.buffer_size = buffer_size self.buffer_size = buffer_size
self.map_func = map_func self.map_func = map_func
...@@ -276,29 +291,42 @@ class ThreadedMapData(ProxyDataFlow): ...@@ -276,29 +291,42 @@ class ThreadedMapData(ProxyDataFlow):
self._in_queue = queue.Queue() self._in_queue = queue.Queue()
self._out_queue = queue.Queue() self._out_queue = queue.Queue()
self._threads = [ThreadedMapData._WorkerThread( self._threads = [ThreadedMapData._WorkerThread(
self._in_queue, self._out_queue, self.map_func) self._in_queue, self._out_queue, self.map_func, self._strict)
for _ in range(self.nr_thread)] for _ in range(self.nr_thread)]
for t in self._threads: for t in self._threads:
t.start() t.start()
# fill the buffer self._iter = self._iter_ds.get_data()
self._itr = self._iter_ds.get_data()
# only call once, to ensure inq+outq has a total of buffer_size elements
self._fill_buffer() self._fill_buffer()
def _fill_buffer(self): def _fill_buffer(self):
n = self.buffer_size - self._in_queue.qsize() - self._out_queue.qsize() n = self.buffer_size - self._in_queue.qsize() - self._out_queue.qsize()
if n <= 0: assert n >= 0, n
if n == 0:
return return
try: try:
for _ in range(n): for _ in range(n):
self._in_queue.put(next(self._itr)) self._in_queue.put(next(self._iter))
except StopIteration: except StopIteration:
logger.error("[ThreadedMapData] buffer_size cannot be larger than the size of the DataFlow!") logger.error("[ThreadedMapData] buffer_size cannot be larger than the size of the DataFlow!")
raise raise
def get_data(self): def get_data(self):
self._fill_buffer() for dp in self._iter:
sz = self.size() self._in_queue.put(dp)
for _ in range(sz):
self._in_queue.put(next(self._itr))
yield self._out_queue.get() yield self._out_queue.get()
self._iter = self._iter_ds.get_data()
if self._strict:
# first call get() to clear the queues, then fill
for k in range(self.buffer_size):
dp = self._out_queue.get()
if k == self.buffer_size - 1:
self._fill_buffer()
yield dp
else:
for _ in range(self.buffer_size):
self._in_queue.put(next(self._iter))
yield self._out_queue.get()
Markdown is supported
0% or
You are about to add 0 people to the discussion. Proceed with caution.
Finish editing this message first!
Please register or to comment