, 12 tweets, 2 min read
in state estimation problems you get to pick from one of two approaches, filtering or optimization. if x is your state and z are your measurements,

filtering states the problem as :

find p(x | z_{1:t}) i.e. a distribution over states given all observations so far
optimization states it as;

argmin_x (E(x, z_{1:t}) i.e. a state such that some energy E is minimized
filtering has the advantage that the distribution you get also encodes uncertainty about your estimate, which is useful

you can do something analogous by perturbing your argmin_x (gonna call it x* now) which tells you how sensitive your minima is in various dimensions
filtering also works mainly with updates of a single measurement on your state, z_t. every bayes filter is a recursive state estimator that updates some belief about the state given new information

someone on here called it a running average, which is not too far from truth
contrast optimization requires many measurements for your estimate to be meaningful, some set {z_1 .. z_n}
optimization usually comes out on top in accuracy though.

what i am curious about is whether it is meaningful to filter over intermediate optimization results (this has come up in two problems I've worked on now)
i.e. if x*_k is the minima for some window of measurements {k_t... k_{t-n}}, is it meaningful to construct a filter p(x | x*_{1:k}), and what are the optimality characteristics of this filter.
the closest thing to this that exists in literature I think is marginalizing over factors in a windowed optimization scheme, very common in e.g. visual-inertial odometry b/c the number of factors in the energy grows very quickly over time
the converse idea is also interesting, rather than marginalizing out factors, just run filters to maintain sparse factors and optimize later; keyframe depth approaches _kinda_ do this by maintaining 1-D kalman filters for each pixel but full optimization has proven better
however the way those filters regard uncertainty may not be the best way (weight each factor by the inverse of its variance at the time it goes into the problem), cerainly seems room to improve here
the best way to analyze this might be to set up some toy convex problem and get at least some closed form results for the optimization, compare with the filter.
hmm there is something dualistic here, another way to look at this might be to write down an energy that instead moves measurements around such that the filter variance is minimized at each timestep
Missing some Tweet in this thread? You can try to force a refresh.

Enjoying this thread?

Keep Current with anton

Profile picture

Stay in touch and get notified when new unrolls are available from this author!

Read all threads

This Thread may be Removed Anytime!

Twitter may remove this content at anytime, convert it as a PDF, save and print for later use!

Try unrolling a thread yourself!

how to unroll video

1) Follow Thread Reader App on Twitter so you can easily mention us!

2) Go to a Twitter thread (series of Tweets by the same owner) and mention us with a keyword "unroll" @threadreaderapp unroll

You can practice here first or read more on our help page!

Follow Us on Twitter!

Did Thread Reader help you today?

Support us! We are indie developers!


This site is made by just three indie developers on a laptop doing marketing, support and development! Read more about the story.

Become a Premium Member ($3.00/month or $30.00/year) and get exclusive features!

Become Premium

Too expensive? Make a small donation by buying us coffee ($5) or help with server cost ($10)

Donate via Paypal Become our Patreon

Thank you for your support!