Temporary TC Bragging Rights

I don’t think I’ve ever had a run like this so I have to brag. This is under the model id Bridgeface.
I have the top rating in a few rounds (100 percentile).
Unfortunately I have my largest stake (200 NMR) on a different model, lol.

mmm, beats me.
How do you create the model?
What is the sharp ratio?

Thanks for the reply. The Sharpe is not the best 0.84, max drawdown 0.2. However, none of my high scoring diagnostic models did well in the tournament (including Sharpe 1.0+). In fact, the performance was crap, very low Corr and TCs, I used all 1500 variables for those models. For my latest models I do feature engineering and use a smaller set of variables (~50), also use parameters that minimize overfitting (regularization, early stop, etc).


nice! what type of feature engineering you’re doing?

classic - that’s how staking works :smile:

1 Like

Wow that’s a small subset of features. I used to think it was good to use all the features because Numerai does so much to test our features. However, I think with TC creative feature selection makes a lot of sense especially to find the best features which the current Stake Weighted Meta Model has no exposure too.

One of my more consistent models TC-wise uses only the “small” feature set – 32 features, no engineering.


BTW, how small features were selected?

This is the small feature set picked out by Numerai in the features.json file. I believe they were the features that were giving the most from an information standpoint, something like that. (So informative and non-redundant.) It’s been the same since v3 data – doesn’t use any of the newer features. (Originally it was 36 features, but 4 of them were in the 10 “bad” features they got rid of, so it’s down to 32 now.)

The reason I was using them wasn’t for a small simple model – it’s because I was playing around with this combinatorial explosion thing and it can’t handle many features at all without becoming untenable to calculate. So the 32 they picked out seemed like a good choice, and it turned out the resulting models were pretty good. So you don’t necessarily need a gazillion features. And there is a ton of redundancy in the features anyway so we don’t actually have as many as it looks like on a fundamental level.

1 Like

The small feature set was created using BorutaShap: Feature Selection with BorutaShap

1 Like

Btw your model get worse.

Why does bragging feel so good?

My last drawdown on May 30th submission an then green all the way up:

Because it is temporary? Anyway… CARPE DIEM

1 Like

I want t brag a little bit here as well . never been in top top 15 in both corr and tc. I know nothing last forever so enjoying it :slight_smile:


Actually, if your models start with ML_ you are doing very well:

  • 7 models in Top 100 TC leaderboard
  • 15 models in CORRV2 TOP 100


Almost got it. ml_yn* are mine. The fact is that they are the same approach (one of the 60D targets with weekly retraining on very short period) but with different flavor of feature engineering and feature neutralization.

So in summary it is one model with different parameters. Could be just a lucky constellation of random parameters that works today. Will see