Challenge with RM Server - Running out of memory
Hi there!
I am a newbie and this is my first post in the community. We have got a RM server installation on top of a MS SQL server box. We have a job container with 64GB RAM. I have built some workflows using sample data on studio environment and trying to run those processes after necessary changes in server environment connecting to original SQL data tables. These workflows mainly involve some basic data joins and summarization after application of few domain specific business rules.
When I am trying to run a flow, I am quickly running into the issue of Running Out of Memory. The challenge I have is, even the first part of my flow which involves reading few variables from a 40GB dataset is not getting completed. Due to the nature of data and business knowledge involved, I am not in a position to share the XML flow or log files here.
I have few specific questions for the community:
1. How does RM Server handles memory internally? Will the whole source data file be read and kept in memory while processing?
2. What is the maximum database size at source that can be handled by a 64GB single container?
3. Will you recommend RM server for huge data processing operations (i.e. data running closer to a TB in size).
Thanks,
Ramesh
Answers
David
Another nice tool to use for cases like this is the caching operators from Old World Computing in their Jackhammer extension. They have just published some new blog articles showing how this is done - you can find part 1 here: https://oldworldcomputing.com/tutorial-introduction-to-caching-functions-of-the-jackhammer-extension-by-old-world-computing/ It is designed almost exactly for your use case. I'm cc'ing @land in case he has something more to add.
Scott
@ David_A and @sgenzer:
Thank you for taking time to respond and for your inputs.
I am using a Read Database operator with a query to pull few variables. Will look into InDatabase Processing and JackHammer extension. Will keep this thread updated.
Thanks,
Ramesh