Blog blue BKG
  • Blog
  • AI Partners
  • New DataRobot and Snowflake Integrations: Seamless Data Prep, Model Deployment, and Monitoring

New DataRobot and Snowflake Integrations: Seamless Data Prep, Model Deployment, and Monitoring

March 16, 2023
· 5 min read

Data scientists run experiments. They iterate. They experiment again. They generate insights that drive business decisions. They work with partners in IT to harden ML use cases into production systems. To work effectively, data scientists need agility in the form of access to enterprise data, streamlined tooling, and infrastructure that just works. Agility and enterprise security, compliance, and governance are often at odds. This tension results in more friction for data scientists, more headaches for IT, and missed opportunities for businesses to maximize their investments in data and AI platforms. 

Resolving this tension and helping you make the most of your current ecosystem investments is core to the DataRobot AI Platform. The DataRobot team has been working hard on new integrations that make data scientists more agile and meet the needs of enterprise IT, starting with Snowflake. In our 9.0 release, we’ve made it easy for you to rapidly prepare data, engineer new features and subsequently automate model deployment and monitoring into your Snowflake data landscape, all with limited data movement. We’ve tightened the loop between ML data prep, experimentation and testing all the way through to putting models into production. Now data scientists can be agile across the machine learning life cycle with the benefit of Snowflake’s scale, security, and governance. 

Data Science with Snowflake and DataRobot

Why are we focusing on this? Because the current ML lifecycle process is broken. On average, 54% of AI projects make it from pilot to production. Hence, nearly half of AI projects fail. There are a couple of reasons for this. 

First, being able to experiment long enough to identify meaningful patterns and drivers of change is difficult. The prototyping loop, particularly the ML data prep for each new experiment, is tedious at best. It’s difficult for data scientists to securely connect to, browse and preview, and prepare data for ML models particularly when data is spread across multiple tables. From there, every time you run a new experiment, you’re back to prepping the data again. And when you do find a signal and have built a great model, it’s difficult to put those ML models into production. 

Models that do make it into production require time-consuming management through monitoring and replacement to maintain prediction quality. A lack of integrated tooling along the entire process not only slows down data scientist productivity, but it increases the total cost of ownership as teams have to stitch together tooling to get through this process. The DataRobot AI Platform has been focused on making the entire ML lifecycle seamless, and today we’re doing even more with our new Snowflake integration. 

Secure, Seamless, and Scalable ML Data Preparation and Experimentation

Now DataRobot and Snowflake customers can maximize their return on investment in AI and their cloud data platform. You can seamlessly and securely connect to Snowflake with support for External OAuth authentication in addition to basic authentication. DataRobot secure OAuth configuration sharing allows IT administrators to configure and manage access to Snowflake.

DataRobot will automatically inherit access controls, so you can focus on creating value-driven AI, and IT can streamline their backlog. 

With our new integration, you can quickly browse and preview data across the Snowflake landscape to identify the data you need for your machine learning use case. Automated data preparation and well-defined APIs allow you to quickly frame business problems as training datasets. The push-down integration minimizes data movement and allows you to leverage Snowflake for secure and scalable data preparation, and as a feature engineering engine so you don’t have to worry about compute resources, or wait on processes to complete. Now you can take full advantage of the scale and elasticity of your Snowflake instance.  

Secure, Seamless, and Scalable ML Data Preparation and Experimentation - DataRobot and Snowflake

With our DataRobot hosted notebooks, you can leverage Snowpark for Python alongside the DataRobot Python Client to quickly connect to Snowflake, explore, prepare, and create machine learning experiments with your Snowflake data. You can leverage the two platforms in the way that make the most sense for you – leveraging Snowpark and the DataRobot developer framework that has native support for Python, Java, and Scala. Because this integration is native to the DataRobot AI Platform, you get your time back with one frictionless experience. 

One-Click Model Deployment and Monitoring in Snowflake

Once trained models are ready to be deployed, you can operationalize them in Snowflake with a single click. Supported models can be deployed directly into Snowflake as a Java UDF by DataRobot. This functionality includes being able to deploy models, built outside of DataRobot, in Snowflake. This means you can bring a model directly into the governed runtime of Snowflake, allowing businesses to make accurate predictions in-database on sensitive data at scale, and without the fuss of configuration. One-click model deployment also gives ML practitioners the flexibility to use normal queries or more advanced features like Stored Procedures from within Snowflake to read scoring data, score data, and write predictions.

One-Click Model Deployment and Monitoring in Snowflake - DataRobot

Along with one-click model deployment come more robust monitoring capabilities, allowing for ongoing monitoring of not just deployment service health, but also drift and accuracy. Model replacement is made easy with retraining and deployment workflows to ensure enterprise-grade reliability of production machine learning on Snowflake. 

Snowflake and DataRobot: Combining Data and AI for Business Results

The new Snowflake and DataRobot integration provides organizations a unique and scalable enterprise platform for data and AI driven business results. We shrunk the ML cycle time, and made it easy for you to experiment more, prepare datasets and build ML models fast, and then get those models out into production to drive value even faster. 

Torsten Grabs, Director of Product Management at Snowflake, and Venky Veeraraghavan, CPO DataRobot

Try out the new integration and let us know what you like. Learn more from Torsten Grabs, Director of Product Management at Snowflake, who will share more about these new innovative capabilities at the DataRobot virtual on-demand event: From Vision to Value: Creating Impact with AI. Join us on March 16 and see more of the DataRobot and Snowflake integration first hand! 

DataRobot Launch Event
From Vision to Value. Creating Impact with AI
Watch Now

1 Gartner®, Gartner Survey Analysis: The Most Successful AI Implementations Require Discipline, not Ph.D.s, Erick Brethenoux, Anthony Mullen, Published 26 August 2022

About the author
Kian Kamyab
Kian Kamyab

Senior Product Manager, DataRobot

Kian Kamyab is a Senior Product Manager at DataRobot. He honed his customer empathy and analytical edge as an Executive Director at SAP’s New Ventures and Technologies group, a Senior Data Scientist at an enterprise software venture studio, and a founding team member of a James Beard award-nominated cocktail bar. When he’s not crafting AI/ML products that solve real world problems, he’s handcrafting furniture and exploring the woods in and around San Francisco.

Meet Kian Kamyab
  • Listen to the blog
  • Share this post
    Subscribe to DataRobot Blog
    Newsletter Subscription
    Subscribe to our Blog