I am trying to read 2 million row table (60 Columns of which 4 Columns are keys) into a hash file.its been taking around 2 hrs to load the data into the file with the default options that are provided with the Hash file.I tried to optimize the hash file using HFC.exe(Hash File Calculator tool),it said either to use a File Type of 14,18 or Modulo of 367097,still the performance is no different.Could someone provide me any options in tuning the hash file.
pre-sizing the modulo on a dynamic file and/or using a static hashed file of appropriate size works well. Also, do you have row buffering enabled? That will increase your speed. Are you sure that your hashed file write is the bottleneck? If you do short test and change your hashed file stage into a sequential file writing to /dev/null do you get a much better speed?
Ok... that would have been a nice fact to mention up front. Next question, why are you doing that? You are obviously doing a wee bit more than simply reading 2 million rows into a hashed file...
-craig
"You can never have too many knives" -- Logan Nine Fingers