What sort of fit does Beeminder perform when you enable the “Turquoise Swath”?
I’m wondering because I noticed wild variations sometimes appearing before or after the existing data points on various graphs that have it enabled. Have the bees considered using Gaussian processes to perform the fitting a la the Automatic Statistician? Depending on the kernel selected, you can provide a much stronger prior on what curves should look like, reducing those wild swings outside the existing data.
There has even been some work on compositional kernel construction (e.g. this paper, which might be useful for capturing structure like a predictable flurry of activity at the end of the month or breaks on the weekend.
I’m not entirely sure what the computational demands are for fitting something like a Gaussian process compared to what Beeminder does now but just wanted to throw the idea out there in case it sparked ideas for features down the road.
Thanks for the pointer to something potentially much better! I have a bunch of notes collected on that question as well…
[1] That’s right, all the code that generates the graphs is open source now! Well, as soon as we finish the migration to the new Beebrain: [oops, I originally linked to an internal thread here]
Perhaps I’m using it incorrectly, but my primary use case for the swath is to less to smooth out existing data and more to make sane predictions. Whatever improves that is a win in my book .
One other option to consider that’s somewhat similar to Kriging, is some sort of Bayesian regression in which you perform inference over both the type of curve (sinusoidal, linear, parabolic, cubic,…) and the specific parameters of the curve. Depending on your setup, you can tradeoff between the complexity of the curve and fit in a principled way.