Case Study

Learn how to optimize machine learning model training using concurrent.futures for parallel hyperparameter tuning and grid search evaluation strategy.

One of the problems that often plagues data scientists working on machine learning applications is the amount of time it takes to train a model. In our specific example of the kk-nearest neighbors implementation, training means performing the hyperparameter tuning to find an optimal value of kk and the right distance algorithm. In the previous chapters of our case study, we’ve tacitly assumed there will be an optimal set of hyperparameters. We’ll look at one way to locate the optimal parameters.

In more complex and less well-defined problems, the time spent training the model can be quite long. If the volume of data is immense, then very expensive compute and storage ...