Services & Resources / Wolfram Forums / MathGroup Archive
-----

MathGroup Archive 2012

[Date Index] [Thread Index] [Author Index]

Search the Archive

Re: Memory Blowup Issues

  • To: mathgroup at smc.vnet.net
  • Subject: [mg126721] Re: Memory Blowup Issues
  • From: David Reiss <dbreiss at gmail.com>
  • Date: Sat, 2 Jun 2012 05:44:14 -0400 (EDT)
  • Delivered-to: l-mathgroup@mail-archive0.wolfram.com
  • References: <jqa1h7$bj7$1@smc.vnet.net>

One thing to try is to set

$HistoryLength = 0.

This way Mathematica does not keep a record of the In[n]  and Out[n]
values.


On Jun 1, 5:20 am, "Kevin J. McCann" <k... at KevinMcCann.com> wrote:
> I am dealing with some good sized data sets from satellite data. These
> data are stored in hdf format. The problem is that I process a sequence
> of these data sets by Importing and then processing; however, it seems
> that the memory usage grows with each Import, even if I Remove the data.
>
> Here is a somewhat simplified example:
>
> MemoryInUse[] (* some data already read in *)
> 2781549920
>
> Data=Import[fileName,"Data"]; (* import a new data set *)
> MemoryInUse[]
> 3302168936
>
> Data=Import[fileName,"Data"];(* import  it again *)
> MemoryInUse[]
> 3822781920
>
> Remove[Data] (* this has no effect on the memory usage *)
> MemoryInUse[]
> 3822787840
>
> As you can see, I will shortly run out of memory. Any ideas?
>
> Thanks,
>
> Kevin




  • Prev by Date: Re: Memory Blowup Issues
  • Next by Date: Re: Flattening a hypermatrix into an ordinary matrix
  • Previous by thread: Re: Memory Blowup Issues
  • Next by thread: Re: Memory Blowup Issues