Best Practices for Imbalanced Data and Partitioning
In this two-part learning session we discuss best practices around data partitioning and working with imbalanced datasets.
Five-fold cross-validation is often the silver bullet for partitioning your validation dataset, but there are some dangerous caveats you have to be aware of to make sure that you’re building robust models. In part 1 of this learning session, we talk about those pitfalls and outline strategies for handling them.
Binary target variables are very common in data science use cases, many of which are severely imbalanced. When you’re building models for infrequent events, such as predicting fraud or identifying product failures, it’s important to watch out for imbalance in your data. In part 2 of this learning session, we discuss strategies for working with imbalanced datasets and provide some rules-of-thumb for these types of use cases.
- Matt Marzillo (DataRobot, Customer Facing Data Scientist)
- Mitch Carmen (DataRobot, Customer Facing Data Scientist)
- Jack Jablonski (DataRobot, AI Success Manager)
After watching this two-part learning session, you should check out these resources for more information.
DataRobot Platform Documentation: