This is decent for what it is. Some of the problems are pretty open ended which has pros and cons, but that is very different from leetcode, which has very specific data and test cases.
For example, implement linear regression but the example solution uses a random number generator without a fixed seed. It’s fine, reproducibility isn’t the point, but leetcode problems are more structured.
In leetcode they usually don’t tell you exactly what data structure you must use, only that it must pass certain test cases. By analogy this might not tell you which architecture to use but require that it passes certain eval metrics.
Most ML problems in real life don’t constrain you to use linear regression or a CNN either. But there will be some metric you need to optimize.
What would take this repo to the next level is to have a reproducible data generation function for each exercise as well as a reasonable metric which must be passed. I don’t see anything that requires my classification auc to be over 0.5 which would be a basic criteria of bug-free code.
For example, implement linear regression but the example solution uses a random number generator without a fixed seed. It’s fine, reproducibility isn’t the point, but leetcode problems are more structured.
In leetcode they usually don’t tell you exactly what data structure you must use, only that it must pass certain test cases. By analogy this might not tell you which architecture to use but require that it passes certain eval metrics.