For those wondering if they are, I'll give a few comments:
T-SNE is always something I want to apply, but can never quite figure out the right way to do it. There is certainly some benefits to be had from having a basic understanding of what this all means so you can get a better chance at visualizing your features in 2D...
Much like stochastic methods, as much as I would never make them the center of a model, there is always use as a feature or ensemble. Ensemble is truly the free lunch of alpha...
Ricci curvature, etc. There are plenty of papers that try to predict crashes with these sorts of things but it is very tricky to pick out one structure in particular and not find some false flags. Differential, topological, and algebraic structures are all key components...
Back to ensemble, for those who aren't aware an ensemble of all GARCH family models beats any single EGARCH, TGARCH, HDMI-FLATSCREEN-TV-GARCH model you can whip up..
I think it has a little bit of use for visualizing data but always put it alongside something like PCA so you can see all the elements of structure, it is particularly nice when you are creating synthetic data (if you haven't heard of this already, try MAD-GAN, it's sick)...
Dimensionality reduction is definitely an art. PCA smashes most autoencoders unless you have loads of data so use a mix. Keep in mind PCA is unsupervised and simply orders your data in the form of key components, you then pick how many of the components you want...
You'll find some great results just from applying PCA and leaving it the same dimensionality and then applying a supervised autoencoder. Same with manifolds. You wouldn't use an NN without Kalman, HHT, FFT, TA, etc as features so why use an AE without features...
Even if it is supervised it can still do with some help. I've seen MLP and LSTM supervised AEs so far, both work great. Think of PCA and Manifold learning for dimensionality reduction as your features.
Just remember if I take in 50 features, use PCA, I still get 50 components. All it's done is split it into the shared (linearly) information and the residuals. I.e. X and Y perform a linear regression. I take the linear regression as my new first variable, and the residual...
as my second variable. Still the same data, I've just transformed it into something with more meaning. It's just that people then take only the first variable and chuck away the residuals. A good AE will find assistance and guidance from that structure and then...
pick up on non-linear effects. Don't waste data for it to learn the easy linear shit, use established models -> introduce that rigid basis -> and then let it do it's magic.
There is more to this thread, I just replied to the tweet above this on (the one ending with "without features...") instead so twitter had fucked it all up
• • •
Missing some Tweet in this thread? You can try to
force a refresh
Large quantity of unique features
Really good dimensionality reduction
Ensemble everywhere!
A word on each...
When it comes to modeling everyone always goes to their favorite NNs like LSTMs etc or LGBMs and those are great, but everyone has them, and frankly, they aren't that hard to implement! Just look at Kaggle if you want an example of DS students using them everywhere...
For real alpha, you need to focus on the three most ignored areas (there is a fourth, speed, but that's not really modeling, and a fifth which I'm not telling you because I like my alpha unleaked). That sounded super guru-like, but I promise these work and I use them.
A key concept for MMs is how you manage inventory. Avellaneda and Stoikov is basically the model everyone uses for this. Then there comes the offset, basically how wide your spreads are. That's your basic model of liquidity provision...
From there we get to have some fun! If you can create multiple forecasts for different timeframes (and at a super-advanced level compute speeds) you can make spreads asymmetric and intentionally hold inventory...
Entirely unprompted here, but please check out @FadingRallies. Also @choffstein's Liquidity Cascades paper (link below). The flow between MMs, passive funds, ELS, and generally the effects of reflexive dealer hedging are key to understanding this regime!
Even if you aren't a trader (I certainly am not, although I try to keep up with it all) it is still super important to understand the regime and how it all fits in from a risk perspective. You CANNOT just take the models as your risk! Eigenportfolios decay, I would know I work
with them all the time so that isn't even the perfect metric (although I do love them). Statistical models will capture some risk but at the end of the day, you choose the parameters and the distribution you feed in is key. Knowing fat tails exist is incredibly important for this
Tweeting a question I was asked/ response regarding MM:
(me adding bonus resources):
A great example of C++ HFT MM algorithms. An improvement idea I have suggested to the author but can also be attempted by interested algotraders is that a fast model like XGBOOST (there is a C++
library) is used alongside some alphas to make spreads asymmetric before traders can trade against you and you get negative edge in those trades. A large part of market making is cheaply executing alphas by trying to get inventory on the side of your predictions and also by
getting out the way of adverse conditions by making your spreads asymmetrically wide (traders with alpha against you). github.com/hello2all/gamm…
I think you can probably classify modeling and feature engineering into a few areas: ML, Statistics, Time Series, Microstructural, some fun extras like entropy which are super weird to work with TA, data-mined alphas, and signal processing.
1/who knows lol
I'll probably speak on each of these eventually, but today I think it'd be good to get some publicity on signal processing. It's underhyped compared to ML and just as deserving.
2/
A lot of the literature is exclusive to electrical engineering and CS, but I can tell you there is lots of alpha in the area. As the story usually goes NN models like LSTMs get a bad rap performance-wise bc of their terrible application
3/