Quantopian's community platform is shutting down. Please read this post for more information and download your code.
Back to Community
What is preventing challenge submissions from getting tailored on the past?

Example https://www.quantopian.com/posts/new-challenge-build-smart-beta-factors

I probably don't understand something fundamental on how challenge submissions are evaluated. Naively I'd think if all data is known about a past time period, most algorithms can be backfitted to look good for that period, for just about any metric.

How can the jury differentiate between truly valuable factors, and backfitted ones?

1 response

Quantopian evaluates winners on both in-sample (i.e. the period they list in the challenge) as well as out-of-sample (the ~2 years of data they hold out from the community) performance. They often also use true out-of-sample data by running the algorithms after every trading day. Looking at all of these things combined, with some different weighting metrics, etc., they can arrive at a "score" that weeds out most of overfit strategies that you're describing.