Hi
I have Data stage Job which process the file has 30GB(14m Rows). Which will Remove the Duplicates and writes the valid data into Dataset.
Which is creating 230 GB datafiles in Resource Disk and 200GB temp files in Scratch Disk. Does Dataset Data files become these many time on actual raw files? As I don't have any transformation or extra fields in it.
Could someone suggest on this?
Thanks
Scratch Disk & Resource Disk Space Issue
Moderators: chulett, rschirm, roy
Scratch Disk & Resource Disk Space Issue
Last edited by Nagac on Tue Oct 28, 2014 12:06 pm, edited 1 time in total.
I am Sorry.... if it means in that way.
But I would like to know How much space Resource Disk We need to process 1 GB Flat File(CSV), Because I am processing one file which is 30GB which is creating 230GB Files in Resource Disk Area where I have no transformations Just removing the Duplications if there are any.
Thanks
But I would like to know How much space Resource Disk We need to process 1 GB Flat File(CSV), Because I am processing one file which is 30GB which is creating 230GB Files in Resource Disk Area where I have no transformations Just removing the Duplications if there are any.
Thanks
Those environment variables will do nothing for your space issue.
Change your nvarchar(255) to simply nvarchar with no maximum length specified.
An nvarchar(255) physically requires 510 bytes of storage in a dataset. An nvarchar with no length specified physically requires 2 bytes per character of data plus a couple of bytes to store the length.
Dataset storage changed at version 7.0.1 to favor processing speed over storage space.
Mike
Change your nvarchar(255) to simply nvarchar with no maximum length specified.
An nvarchar(255) physically requires 510 bytes of storage in a dataset. An nvarchar with no length specified physically requires 2 bytes per character of data plus a couple of bytes to store the length.
Dataset storage changed at version 7.0.1 to favor processing speed over storage space.
Mike