How to calculate the optimal size of all the lookup data set

Post questions here relative to DataStage Enterprise/PX Edition for such areas as Parallel job design, Parallel datasets, BuildOps, Wrappers, etc.

Moderators: chulett, rschirm, roy

Post Reply
venkat_nk
Participant
Posts: 2
Joined: Wed Jul 26, 2006 8:11 am

How to calculate the optimal size of all the lookup data set

Post by venkat_nk »

used at the same time?
Based on
1. Number of CPU
2. CPU GHz,
3. Scratch size
4 Number of nodes and
5. RAM GB


Thanks,
Venkat
thanks,
Venkat
ray.wurlod
Participant
Posts: 54595
Joined: Wed Oct 23, 2002 10:52 pm
Location: Sydney, Australia
Contact:

Post by ray.wurlod »

Based, much more importantly, on the volume of data you propose to load into the reference Data Sets, and whether you propose to use virtual Data Set or Lookup File Set for your reference source.

The optimal size is one with the capacity to store that total volume.

You can not succeed without all the rows and columns you need in the lookups. The trick is to make sure that that's all you have.

Then you need to make sure that you have sufficient resources to cope with that volume, not the reverse (which is where you seem to be coming from).
IBM Software Services Group
Any contribution to this forum is my own opinion and does not necessarily reflect any position that IBM may hold.
Post Reply