|
[Sponsors] |
January 23, 2009, 11:30 |
Fluent Gunzip
|
#1 |
Guest
Posts: n/a
|
I'm trying to read large data files (> 4 GB) into fluent and get an error message.
These files are older and were created 4 yrs ago. The person who ran them had similar problems but does not remember the fix. If I gunzip (using /bin/gunzip) on the command line and read that file in it works since Fluent is not doing the gunzip. So it seems that Fluent is pointing at a different out of date gunzip. How can I check? Can I change this setting? Thanks Scott |
|
January 27, 2009, 16:12 |
Re: Fluent Gunzip
|
#2 |
Guest
Posts: n/a
|
You are probably just running out of memory. Given the size of the data files, I assume the cas file is fairly large, and is already loaded into fluent, so your memory usage is already high. When you read a gzipped data file, fluent creates a new process via fork/exec to run gzip and read the output. If you don't have enough swap space (virtual memory), the fork will fail.
|
|
|
|
Similar Threads | ||||
Thread | Thread Starter | Forum | Replies | Last Post |
Fluent in Linux vs. Fluent in Windows | Melih | FLUENT | 6 | November 16, 2014 09:39 |
Fluent Gunzip | Scott Nordsen | Main CFD Forum | 0 | January 26, 2009 11:47 |
how to save files with gunzip?? | Kim | FLUENT | 3 | October 12, 2004 09:20 |