有没有一种优雅的方式来分块处理流? [英] Is there an elegant way to process a stream in chunks?

查看:25
本文介绍了有没有一种优雅的方式来分块处理流?的处理方法,对大家解决问题具有一定的参考价值,需要的朋友们下面随着小编来一起学习吧!

问题描述

我的确切场景是批量向数据库插入数据,所以我想累积 DOM 对象,然后每 1000 个,刷新它们.

My exact scenario is inserting data to database in batches, so I want to accumulate DOM objects then every 1000, flush them.

我通过将代码放入累加器中以检测填充度然后刷新来实现它,但这似乎是错误的 - 刷新控制应该来自调用者.

I implemented it by putting code in the accumulator to detect fullness then flush, but that seems wrong - the flush control should come from the caller.

我可以将流转换为 List,然后以迭代方式使用 subList,但这似乎也很笨重.

I could convert the stream to a List then use subList in an iterative fashion, but that too seems clunky.

是否有一种巧妙的方法可以对每 n 个元素采取行动,然后继续处理流,同时只处理一次流?

It there a neat way to take action every n elements then continue with the stream while only processing the stream once?

推荐答案

优雅在情人眼中.如果你不介意在 groupingBy 中使用有状态函数,你可以这样做:

Elegance is in the eye of the beholder. If you don't mind using a stateful function in groupingBy, you can do this:

AtomicInteger counter = new AtomicInteger();

stream.collect(groupingBy(x->counter.getAndIncrement()/chunkSize))
    .values()
    .forEach(database::flushChunk);

这不会比您的原始解决方案赢得任何性能或内存使用点,因为它仍然会在执行任何操作之前实现整个流.

This doesn't win any performance or memory usage points over your original solution because it will still materialize the entire stream before doing anything.

如果您想避免具体化列表,流 API 将无济于事.您将必须获取流的迭代器或拆分器并执行以下操作:

If you want to avoid materializing the list, stream API will not help you. You will have to get the stream's iterator or spliterator and do something like this:

Spliterator<Integer> split = stream.spliterator();
int chunkSize = 1000;

while(true) {
    List<Integer> chunk = new ArrayList<>(size);
    for (int i = 0; i < chunkSize && split.tryAdvance(chunk::add); i++){};
    if (chunk.isEmpty()) break;
    database.flushChunk(chunk);
}

这篇关于有没有一种优雅的方式来分块处理流?的文章就介绍到这了,希望我们推荐的答案对大家有所帮助,也希望大家多多支持IT屋!

查看全文
相关文章
登录 关闭
扫码关注1秒登录
发送“验证码”获取 | 15天全站免登陆