Snowflake sponsored this post.
Everywhere you look, companies are identifying ways to use data to improve or resolve business problems. Automated machine learning (AutoML) tools enable organizations to quickly, accurately and consistently develop and apply machine learning models at scale across their operations to address real-world issues rather than relying solely on data scientists to hand-code models.
There are a plethora of real-world use cases for applying AI models. It could be helping to improve demand forecasting for raw materials, for finished goods, or for the number of employees needed to staff a facility for the year ahead. Alternatively, perhaps an enterprise’s corporate finance department wants to enhance its ability to predict fiscal results, an airline is seeking insight into dynamic pricing, or a midsize company needs to increase the number of resolved customer support tickets within its call centers.
Traditionally, whenever a new AI model is required, a data scientist would reach for a new Jupyter Notebook and start programming. While this approach works well for pure sandbox experimentation, these notebooks are not the ideal medium in which to build and then attempt to share and scale production code. Hand coding can be a real obstacle to realizing AI success, since no one is going to deploy 100 AI models where each model has been developed in its own separate notebook.
AutoML aims to automate all the tasks involved in developing and deploying machine learning models, from cleaning the raw data all the way through to parameter optimization.
As well as making data scientists much more productive, AutoML tools can be used by non-experts to run the same dataset against combinations of hundreds or even thousands of different algorithms simultaneously to arrive at the optimal AI model. Contrast that scenario with a data scientist hand coding to create a data pipeline, build the model, do all the testing and then run the dataset against a handful of algorithms.
One way to think of AutoML is that it’s acting as the “cruise control” for machine learning as the user is driving down an AI model highway. They don’t need to be in the manual mode of hand coding at that stage; programming is only required when the user is looking to turn off to a specific exit or design element for that model.
The four major benefits of using AutoML instead of hand coding are:
- Speedier time to value. This is the No. 1 benefit for organizations that are laser focused on what an AI model is worth, including how long it took to create it in order to determine return on investment and the projected operating efficiency or net revenue uplift. As data scientists hand code models, they may be overly focused on the journey to evolve a model and the experiments they made, while a business manager might only be interested in results and the actionable insights generated by the AI model. Automation may also provide a way to clearly define when an AI model is good enough to be deployed, versus that same assessment made by a human programmer.
- Greater acceleration at scale. This benefit goes hand in hand with time to value as a way to generate accurate models as quickly as possible. Automation enables a retailer to scale from two hand-coded models to 10,000 AutoML models, or a model for every SKU in the store. By relying on automation, an organization can avoid human-based errors that may occur during manual data modeling.
- Realizing operational excellence. AutoML tools allow an organization to inject more standardization and consistency into the generation of AI models. For instance, if a data scientist were to leave the company which has adopted AutoML, a replacement would not have to reinterpret their predecessor’s hand-coded models. All the new hire would have to do is apply or be retrained on AutoML. Again, an AutoML tool functions as a great system of record, a central repository where managers and employees can collaborate on models and interpret results.
- Democratizing AI. AutoML tools can empower business analysts and subject matter experts to generate models in a guardrail-safe manner that embeds standard data science best practices into the models. These roles no longer have to depend on their data scientist peers to create models for them. AutoML tools can help an organization to open up data science to a broader community of users and make AI more mainstream. While Jupyter Notebooks function as outliers to an organization’s overall software engineering process, AutoML-generated models can be embedded within those existing operations and then be easily shared with software and ML engineers.
As organizations become fully convinced of these benefits, there will be widespread adoption of AutoML, in part driven by people’s eagerness, regardless of their roles, for tools to help them become more data literate.
There is already good consensus in the market about the functionality of AutoML products, regardless of whether the tools originate from market pioneer DataRobot, other early entrants such as Dataiku or H2O, or market validators like the large cloud providers Amazon, Microsoft or Google.
The New Stack is a wholly owned subsidiary of Insight Partners, an investor in the following companies mentioned in this article: Dataiku.
Photo by Tara Winstead from Pexels.