Aggregator Limit
Moderators: chulett, rschirm, roy
Aggregator Limit
I receive the following error during aggregation;
Aggregator_37,0: Failure during execution of operator logic.
Aggregator_37,0: Input 0 consumed 9614492 records.
Aggregator_37,0: Output 0 produced 9614492 records.
Aggregator_37,0: Fatal Error: pipe write failed: Broken pipe
Aggregator_37,0: Failure during execution of operator logic.
Aggregator_37,0: Input 0 consumed 9614046 records.
Aggregator_37,0: Output 0 produced 6683230 records.
Aggregator_37,0: Fatal Error: sendReadAck(): write failed on node SASHQOKWSDA Broken pipe
node_node2: Player 18 terminated unexpectedly.
I assume it's a space issue as the program worked fine with a 5 million row output. I ran it multiple time and it keeps aborting around the same amount of records. Is there a limit that I should change.
Thanks,
Aggregator_37,0: Failure during execution of operator logic.
Aggregator_37,0: Input 0 consumed 9614492 records.
Aggregator_37,0: Output 0 produced 9614492 records.
Aggregator_37,0: Fatal Error: pipe write failed: Broken pipe
Aggregator_37,0: Failure during execution of operator logic.
Aggregator_37,0: Input 0 consumed 9614046 records.
Aggregator_37,0: Output 0 produced 6683230 records.
Aggregator_37,0: Fatal Error: sendReadAck(): write failed on node SASHQOKWSDA Broken pipe
node_node2: Player 18 terminated unexpectedly.
I assume it's a space issue as the program worked fine with a 5 million row output. I ran it multiple time and it keeps aborting around the same amount of records. Is there a limit that I should change.
Thanks,
Jim Stewart
Have you monitored your scratch space while the job was running? Perhaps the broken pipe was caused by space issues.
<a href=http://www.worldcommunitygrid.org/team/ ... TZ9H4CGVP1 target="WCGWin">
</a>
</a>
Re: Aggregator Limit
Are you doing some pre-sorting before aggregation ?
Definitely its an space/memory issue. Temporary fix can be to run the job as sequential or make aggregator stage as sequential.
Definitely its an space/memory issue. Temporary fix can be to run the job as sequential or make aggregator stage as sequential.
Re: Aggregator Limit
I think in the aggregate method you are using hash, use "sort" instead.
It took me fifteen years to discover I had no talent for ETL, but I couldn't give it up because by that time I was too famous.
Follow-up:
We have two servers. DEV and PROD If I run the same job on each server, I get two different results. On the DEV server, the job completes normally. On the PROD server, the job aborts. The only DS difference is a hot patch IBM offered to fix another problem. I have a suspicion that the patch has caused the problem. I will keep you posted.
I will also try some of your suggestions.
Thanks
We have two servers. DEV and PROD If I run the same job on each server, I get two different results. On the DEV server, the job completes normally. On the PROD server, the job aborts. The only DS difference is a hot patch IBM offered to fix another problem. I have a suspicion that the patch has caused the problem. I will keep you posted.
I will also try some of your suggestions.
Thanks
Jim Stewart