Automated Machine Learning (AutoML) provides methods and processes to automate the end-to-end machine learning workflow towards solving real-world business problems.

In traditional machine learning world, data scientist used to spend a considerable amount of time in Data wrangling, model selection and tuning, now with the advances of AutoML it provides methods and processes to make Machine Learning available for non-Machine Learning experts, to improve efficiency of Machine Learning and to accelerate research on Machine Learning.

Automating the end-to-end process of applying machine learning offers the advantages of producing simpler solutions, faster creation of those solutions, and models that often outperform models that were designed by hand.

1 favorite thumb_down thumb_up 0 comments visibility_off  Remove from Watchlist visibility  Add to Watchlist

Outline/Structure of the Talk

1. Machine and Deep Learning Introduction

2. Why Automated Machine learning

3. Auto ML libraries (H2O, Auto-Keras, T-POT) and algorithms in R and Python

4. Industry usage and integration with Cloud (GCP)

5. Auto ML demo using Banking industry case study in R

Learning Outcome

Participants will take back good fundamental concepts of Deep Learning and Automated Machine Learning using different libraries in R and Python.

Beginners to Machine learning will learn how to develop model in 15 mins and experts in ML will learn about how to leverage Auto ML to develop model for complex use cases.

Cloud integration using GCP will also demoed.

Target Audience

Machine Learning enthusiast, Beginners and Data scientist

Prerequisites for Attendees

1. Basics of Machine Learning and Deep Learning

2. Supervised and Unsupervised Learning

3. Cloud (GCP, AWS) knowledge - Good to have (not must to have)

schedule Submitted 2 weeks ago

Public Feedback

comment Suggest improvements to the Speaker

  • Liked Dr. Saptarsi Goswami

    Dr. Saptarsi Goswami - Meta features and clustering based approaches for feature selection

    45 Mins

    Feature selection is one of the most important processes for pattern recognition, machine learning and data mining problems. A successful feature selection method facilitates improvement of learning model performance and interpretability as well as reduces computational cost of the classifier by dimensionality reduction of the data. Feature selection refers to the retention of discriminatory features while discarding the redundant and irrelevant features. In this process, a subset of D features are selected from a set of N features (D<N). There is another way of achieving dimensionality reduction by projecting higher dimensional data to lower dimension, normally referred to feature extraction. This thesis refers to the former one i.e. feature subset selection. Optimal feature subset selection method comprises of developing 1) an evaluation function for measuring the goodness of a feature or a feature subset and 2) a search algorithm to find out the best subset of features from all possible subsets of the whole feature set. Based on the nature of the objective function used in the search algorithms, feature subset selection algorithms are broadly classified into filter approach and wrapper approach. Classifier dependent wrapper approaches use classifier accuracy as the objective function while filter approaches use any evaluation function representing the intrinsic characteristics of the data set and the resulting feature subset works equally well for any classifier. This work focusses on filter based feature subset selection approach. In this work, initially a study has been done with currently available search based filter type feature selection algorithms for supervised as well as unsupervised classification with both the single objective and multi-objective evaluation functions. Some improvements over the current algorithms have been proposed and their efficiency has been examined by simulation experiments with bench mark data sets. In the second step, an inexpensive feature evaluation measure based on feature relevance to be used with a filter type feature selection for unsupervised classification has been proposed. It has been noticed during literature study that the concept of feature relevance in case of unsupervised classification is difficult to form and current methods are complex and time consuming. The proposed measure which considers individual variability as well as overall variability of the dataset,is found to be effective compared to the current methods by simulation experiments with bench mark data sets. Thirdly, it seems that the most of the current feature selection algorithms are based on search strategies to find out the best feature subset from the available feature set. For a large number of features, exhaustive search is computationally prohibitive which leads to combinatorial optimization problem and some sort of heuristic is used for the solution. With the increase of the number of features, the computational time for optimal feature subset selection increases.An alternative solution to this problem is to use clustering of the features to find out the best feature subset which is not yet explored sufficiently. In this work, an efficient clustering based feature selection algorithm has been proposed and simulation experiments have been done with bench mark data sets. The main contributions of the proposed algorithm are introduction of a novel method to determine the optimal number of clusters, a way of interpretation of the importance of the feature clusters and a method of selection of the final subset of features from the feature clusters. Finally, it is found that though lots of feature selection algorithms are available, it is very difficult to decide which algorithm is suitable for a particular real world application. Here a study has been done to establish the relation between the feature selection algorithm and the characteristics of the data set. A technique has been proposed to define a data set according to its intrinsic characteristics represented by some meta-features. Finally a feature selection strategy is recommended based on the characteristics of the data set and has been implemented with bench mark data sets to judge its effectiveness.

  • Liked Rahee Walambe

    Rahee Walambe / Vishal Gokhale - Processing Sequential Data using RNNs

    480 Mins

    Data that forms the basis of many of our daily activities like speech, text, videos has sequential/temporal dependencies. Traditional deep learning models, being inadequate to model this connectivity needed to be made recurrent before they brought technologies such as voice assistants (Alexa, Siri) or video based speech translation (Google Translate) to a practically usable form by reducing the Word Error Rate (WER) significantly. RNNs solve this problem by adding internal memory. The capacities of traditional neural networks are bolstered with this addition and the results outperform the conventional ML techniques wherever the temporal dynamics are more important.
    In this full-day immersive workshop, participants will develop an intuition for sequence models through hands-on learning along with the mathematical premise of RNNs.

  • Liked Parul pandey

    Parul pandey - Jupyter Ascending : The journey from Jupyter Notebook to Jupyter Lab

    Parul pandey
    Parul pandey
    Data Science Communicator
    schedule 3 weeks ago
    Sold Out!
    45 Mins

    For many of the researchers and data scientists, Jupyter Notebooks are the de-facto platform when it comes to quick prototyping and exploratory analysis. Right from Paul Romer- the Ex-World bank chief Economist and also the co-winner 2018 Nobel prize in Economics to Netflix, Jupyter Notebooks are used almost everywhere. The browser-based computing environment, coupled with a reproducible document format has made them the choice of tool for millions of data scientists and researchers around the globe. But have we fully exploited the benefits of Jupyter Notebooks and do we know all about the best practises of using it? if not, then this talk is just for you.

    Through this talk/demo, I'll like to discuss three main points:

    1. Best Practises for Jupyter Notebooks since a lot of Jupyter functionalities sometimes lies under the hood and is not adequately explored. We will try and explore Jupyter Notebooks’ features which can enhance our productivity while working with them.
    2. In this part, we get acquainted with Jupyter Lab, the next-generation UI developed by the Project Jupyter team, and its emerging ecosystem of extensions. JupyterLab differs from Jupyter Notebook in the fact that it provides a set of core building blocks for interactive computing (e.g. notebook, terminal, file browser, console) and well-designed interfaces for them that allow users to combine them in novel ways. The new interface enables users to do new things in their interactive computing environment, like tiled layouts for their activities, dragging cells between notebooks, and executing markdown code blocks in a console and many more cool things.
    3. Every tool/features come with their set of pros and cons and so does Jupyter Notebooks/Lab and it is equally important to discuss the pain areas along with the good ones.

    JAYA SUSAN MATHEW - Breaking the language barrier: how do we quickly add multilanguage support in our AI application?

    Sr. Data Scientist
    schedule 3 months ago
    Sold Out!
    20 Mins

    With the need to cater to a global audience, there is a growing demand for applications to support speech identification/translation/transliteration from one language to another. This session aims at introducing the audience to the topic and how to quickly use some of the readily available APIs to identify, translate or even transliterate speech/text within their application.