[yt-users] Memory overloading on time series

Sam Geen samgeen at astro.ox.ac.uk
Tue May 20 05:28:52 PDT 2014


Hmm, nope, memory usage still slowly creeping up (I also tried reloading 
yt.mods). I have a couple more ideas to try - one is to run the function 
in a separate process, the other is to check for existing images and 
skip that output if they exist, then just re-run every time the memory 
fills up. Is there any way of profiling what's in the memory with 
Python? I hope/assume it's not numpy or something leaking.

On 20/05/14 13:48, Matthew Turk wrote:
> Hi Sam,
>
> I've looked over your script and I'm not entirely sure where the
> memory usage is not being freed.  My one thought is to try this:
>
> import yt
> yt.frontends.ramses.RAMSESDataset._skip_cache=True
>
> at the top of your script and seeing if it helps.
>
> -Matt
>
> On Mon, May 19, 2014 at 9:10 AM, Sam Geen <samgeen at astro.ox.ac.uk> wrote:
>> Hi,
>>
>> The memory usage in yt is filling up the machine for a time series with ~100
>> outputs. It seems like the data from previous datasets isn't being cleared
>> from memory, despite calling "del pf" and "gc.collect()". I've tried various
>> suggestions made previously in the mailing list to no effect. I'm using the
>> Ramses frontend with the script attached. Any ideas?
>>
>> Thanks,
>>
>> Sam
>>
>> _______________________________________________
>> yt-users mailing list
>> yt-users at lists.spacepope.org
>> http://lists.spacepope.org/listinfo.cgi/yt-users-spacepope.org
>>
> _______________________________________________
> yt-users mailing list
> yt-users at lists.spacepope.org
> http://lists.spacepope.org/listinfo.cgi/yt-users-spacepope.org




More information about the yt-users mailing list