DataRobot PartnersUnify your AI stack with our open platform, extend your cloud investments, and connect with service providers to help you build, deploy, or migrate to DataRobot.
This post was originally part of the DataRobot Community. Visit now to browse discussions and ask questions about DataRobot, AI Platform, data science, and more.
In this two-part learning session we discuss best practices around data partitioning and working with imbalanced datasets.
Five-fold cross-validation is often the silver bullet for partitioning your validation dataset, but there are some dangerous caveats you have to be aware of to make sure that you’re building robust models. In part 1 of this learning session, we talk about those pitfalls and outline strategies for handling them.
Binary target variables are very common in data science use cases, many of which are severely imbalanced. When you’re building models for infrequent events, such as predicting fraud or identifying product failures, it’s important to watch out for imbalance in your data. In part 2 of this learning session, we discuss strategies for working with imbalanced datasets and provide some rules-of-thumb for these types of use cases.
Hosts
Matt Marzillo (DataRobot, Customer Facing Data Scientist)
Mitch Carmen (DataRobot, Customer Facing Data Scientist)
Jack Jablonski (DataRobot, AI Success Manager)
Now what?
After watching this two-part learning session, you should check out these resources for more information.