MathGroup Archive 2010

[Date Index] [Thread Index] [Author Index]

Search the Archive

Need speed partitioning WeatherData

  • To: mathgroup at
  • Subject: [mg112488] Need speed partitioning WeatherData
  • From: "P. Fonseca" <public at>
  • Date: Fri, 17 Sep 2010 06:41:17 -0400 (EDT)


Meanwhile I figured out that the problem comes from the time  
DateDifference takes.

One process of DateDifference costs 0.0045 seconds on my computer (ex.  
DateDifference[{1990, 10, 1, 0, 0, 0}, {1999, 10, 1, 0, 0, 0}]).

In all the different algorithms I implemented, each one of the 500 000  
samples, is "DateDifferenced" against around 3 of its neighbors. This  

3 * 500 000 * 0.0045 = 6750 seconds !!!!!

I can now imagine a way of partitioning the data without applying so  
many times the DateDifference function: one first run to register the  
differences between every consecutive pair, and then I just work with  
these difference values.
Nevertheless, this still means 500 000*0.0045=2250 seconds!

Should I create my own DateDifference function (less options -> more  
fast)? Does someone has a simpler solution for a 100x speed up? (I  
already dropped the illusion of the couple of seconds...)

Thank you in advance,

  • Prev by Date: Range of data with BezierFunction
  • Next by Date: Re: ANOVA question
  • Previous by thread: Range of data with BezierFunction
  • Next by thread: Re: Need speed partitioning WeatherData