Scaling scikit-learn: introducing new sets of computational routines
- Room:
- Liffey Hall 2
- Start (Dublin time):
- Start (your time):
- Duration:
- 30 minutes
Abstract
For more than 10 years, scikit-learn has been bringing machine learning and data science methods to the world. Since then, the library always aimed to deliver quality implementations, focusing on a clear and accessible code-base built on top of the PyData ecosystem.
This talk aims at explaining the recent on-going work of the scikit-learn developers to boost the native performances of the library.
TalkPyData: Software Packages & Jupyter
Description
scikit-learn is an open-source scientific library for machine learning in Python.
Since its first release in 2010, the library gained a lot of traction in education, research and the wider society, and has set several standards for API designs in ML software. Nowadays scikit-learn is of one the most used scientific library in the world for data analysis. It provides reference implementations of many methods and algorithms to a userbase of millions.
With the renewed interest in machine-learning based methods in the last years, other libraries providing efficient and highly optimised methods (such as for instance LightGBM and XGBoost for Gradient-Boosting-based methods) have emerged. Those libraries have encountered a similar success, and have put performance and computational efficiency as top priorities.
In this talk, we will present the recent work carried over by the scikit-learn core-developers team to improve its native performance.
This talk will cover elements of the PyData ecosystem and the CPython interpreter with an emphasis on their impact on performances. Computationally expensive patterns will then be covered before presenting the technical choices associated with the new routines implementations, keeping the project requirements in mind. At the end, we will take a quick look at the future work and collaborations on hardware-specialised computational routines.