-
Notifications
You must be signed in to change notification settings - Fork 7
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
Optimize codebase #25
Comments
As the codebase increases, more things are merged in that are well below optimal. E.g., cc @JanJereczek the new segmented window code is poorly optimized. Even the most basic thing, getting the segmented window, violates many rules outlined in Julia's "performance tips": TransitionsInTimeseries.jl/src/analysis/segmented_window.jl Lines 100 to 103 in daa2bbd
The function does broadcasting that allocates many new vectors in place by doing |
Hi @Datseris, I tried to optimize the segmented analysis. Improving things like changing function testsgen(sgen)
for _ in 1:100
sgen()
end
end
x = collect(1.0:100)
sgen = surrogenerator(x, signif.surrogate, Random.Xoshiro(1995))
@btime testsgen($sgen)
# 143.765 μs (200 allocations: 135.94 KiB) Now imagine we have Either we accept this or perform, for instance, following change: I have implemented this for the segment analysis - in a somewhat dirty way so far, but which seems to confirm that we reduce the memory allocation on following code: using TransitionsInTimeseries
n = 1001
t = collect(1.0:n)
x = copy(t)
t1 = [23.6, 300.7, 777.4] # some segmentation lower bounds
t2 = [245.5, 684.7, 952] # some segmentation upper bounds
indicators = (mean, var, skewness, kurtosis)
change_metrics = RidgeRegressionSlope()
config = SegmentedWindowConfig(indicators, change_metrics, t1, t2;
width_ind = 20, min_width_cha = 10, whichtime = last)
res = estimate_indicator_changes(config, x, t);
signif = SurrogatesSignificance(n = 10_000, tail = :right)
pvals = significant_transitions(res, signif)
@btime significant_transitions($res, $signif)
# Former version: 494.023 ms (241554 allocations: 391.31 MiB)
# Current version: 311.283 ms (180845 allocations: 101.18 MiB) Please tell me what you think. |
This sounds like a bug. The surrogate generation shouldn't allocate for random Fourier surrogates. I'll take a look at it. Option 2 seems good, but what is the impact on performance? Reducing allocations by itself doesn't matter. it i s the run time we care about. |
I am confused about what you do in teh segmented analysis. I thought yuou generate the surrogates once and then you take the same segmented windows as the original data. If this is not what you do, then what you do is incorrect and we have to change the source code. Only 100 surrogates must be generated, not more, if the user asked for 100 surrogates. The surrogates must be generated for the whole timeseries, not the window. |
I am speaking about our general design, which includes both the sliding and the segmented window analysis. Some time ago I had programmed the following (here in pseudo-code for the sake of illustration): for n in 1:n_surrogates
s = sgen()
for i in 1:n_indicator
perform_analysis(s)
end
end You changed it to: for i in 1:n_indicator
for n in 1:n_surrogates
s = sgen()
perform_analysis(s)
end
end Which, to me, does not make sense, since it implies generating more surrogates than necessary. See for instance |
I don't agree with that. Since you have transitions in the time series, your dynamic regime might be quite different depending on the segment. For instance, segment 1 displays much larger variance than segment 2. Making a surrogate for the whole timeseries would give you an intermediate value of variance. You "average" the statistics of different segments in a certain way. I think we rather, quite clearly want a surrogate of the segment that is being analysed. |
Yes, this doesn't make sense to me either. That was obviously a mistake that we need to fix! |
It isn't as clear cut as you say though. Surrogates aren't magic: they want to conserve various statistical properties for their generation to be meaningful and the statistical hypothesis to be validly stated. If you have small windows with not enough statistics to estimate the quantity that surrogates coserve, the whole hypothesis testing infrastructure breaks down. So, the larger the window, the more accurate the statement of hte hypothesis testing. |
Sure. Our whole approach of estimating statistics needs timeseries (or segments of it) that are long enough - that's not even specific to the surrogates. That's also why we require minimum lengths of segments in the current version of the code. That should also ensure that the surrogates are meaningful. |
We've spent a lot of time devising the new interface, but practically 0 seconds optimizing it. There are memory "leakages" all over the place. Running:
This amount of allocations is of course nonsense. There are probably a lot of incorrect allocations done in
windowmap!
which shouldnt happen because we use the in-place version with pre-allocated dummies.Generally speaking, almost all parts of the code base can be optimized. We can start by detecting the bottlenecks. Using the profiling functionality of VSCode we can get a flame graph of what parts of the code are the slowest. Then, we can benchmark the slow functions in separation and see what goes wrong. Do we have type instabilities? Do we allocate unecessary vectors?
Another point to improve is that our functions are not written in a way that allows for easy optimizations. Normally you would want to have a setup function that initializes all containers, and then a function that does the compitations without initializing anything.
The text was updated successfully, but these errors were encountered: