Behavioral Cloning for Self-Driving Cars
We all know that with companies like Tesla and Waymo leading the Self Driving Car space, it is one of the hottest areas of research and technology. What seemed like science fiction a few years ago is now a reality, especially with Tesla's recent software update of Full Self Driving (FSD) for its fleet. Having said that, having the skill set to build self-driving cars as an AI Developer surely gives you the edge. Although India might seem behind in this space, there are actually a lot of Indian startups emerging now.
As you participate, you will get an overview of the basics of self-driving cars and then we will go through a particularly interesting concept of Behavioral Cloning which is nothing but using supervised learning for self-driving cars. In Behavioral Cloning, a human driver drives the car for some time while all that driving data is captured by cameras in front of the car as the input data and the respective steering angle and acceleration as the target data. Then a CNN is trained on that data to self-drive a car with just visual input without any additional sensors like lidar etc which are costly. This technique is used in the real world by the self-driving car startup, comma.ai as well as Tesla to some extent.
Outline/Structure of the Demonstration
- Introduction (1 min)
- What goes into making a Self Driving Car? (5 min)
- Behavioral Cloning (12 min)
- Data
- Model Architecture
- Code Demo with a simulator
- Conclusion (2 min)
Learning Outcome
- Attendees will get an overview of the systems used to make a self-driving car
- They will particularly understand the usage of supervised learning for self-driving cars
Target Audience
AI/Deep Learning Enthusiasts, Researchers and anyone with practical experience in ML and DL
Prerequisites for Attendees
A Practical Understanding of
- Deep Learning
- Convolutional Neural Networks
- Python
- Tensorflow or Keras
Video
Links
React Frontends for Data Scientists: Slides
Deploying a Pytorch Model to AWS Sagemaker: Slides
Deploying Tensorflow Model on AWS
Project Presentation at CellStrat AI Conclave: My part of the presentation starts at 8:35
I also have an upcoming webinar on this topic on April 14, it can be accessed here on meetup
schedule Submitted 2 years ago
People who liked this proposal, also liked:
-
keyboard_arrow_down
Darshan Ganji - On-Demand Accelerating Deep Neural Network Inference via Edge Computing
20 Mins
Demonstration
Advanced
Deep Neural networks are both computationally intensive and memory intensive, making them difficult to deploy on mobile phones and embedded systems with limited hardware resources and taking more time for Inference and Training. For many mobile-first companies such as Baidu and Facebook, various apps are updated via different app stores, and they are very sensitive to the size of the binary files. For example, App Store has the restriction “apps above 100 MB will not download until you connect to Wi-Fi”. As a result, a feature that increases the binary size by 100MB will receive much more scrutiny than one that increases it by 10MB. It is challenging to run computation-intensive DNN-based tasks on mobile devices due to the limited computation resources.
This talk introduces the Algorithms and Hardware that can be used to accelerate the Inferencing or reduce the latency of deep learning workloads. We will discuss how to compress the Deep Neural Networks and techniques like Graph Fusion, Kernel Auto-Tuning for accelerating inference, as well as Data and model parallelization, automatic mixed precision, and other techniques for accelerating training. We will also discuss specialized hardware for deep learning such as GPUs, FPGAs, and ASICs, including the Tensor Cores in NVIDIA’s Volta GPUs as well as Google’s Tensor Processing Units (TPUs). We will also discuss the Deployment of the Large Size Deep Learning Models on the Edge devices.
Keywords: Graph Optimization, Tensor Fusion, Kernel Auto Tuning, Pruning, Weight sharing, quantization, low-rank approximations, binary networks, ternary networks, Winograd transformations, data parallelism, model parallelism, mixed precision, FP16, FP32, model distillation, Dense-Sparse-Dense training, NVIDIA Volta, Tensor Core, Google TPU.
-
keyboard_arrow_down
Saakshi Bhargava / Vrishank Gupta - Generative Adversarial Networks (GANs)
20 Mins
Demonstration
Beginner
Deep learning has accomplished pronounced triumph in the field of artificial intelligence, there are many deep learning models that have been developed in the recent time. Generative Models (GAN) are one of the deep learning models, that was given based on the game theory called zero-sum and now has been treated as the hot area for research. Generative Models are modern techniques used in computer vision. Unlike other neural networks that are used for predictions from images, generative models can generate new images as well for specific objectives. They can be used for generating huge datasets. This session will review several applications of generative modeling such as image generation and image translation, video frame prediction using CNNs and GANs.
-
keyboard_arrow_down
Abdul Azeez - Photo-Realistic Single Image Super-Resolution using SRGAN
20 Mins
Demonstration
Advanced
As we know, a basic GAN has two neural Networks – the Discriminator (D) and the Generator (G). The Generator attempts to generate images that look like real images. The Discriminator tries to distinguish the generated images from real images. By combined loss minimization of these two neural networks, the entire model trains and eventually reach a state of equilibrium, where the Discriminator no longer can distinguish the fake images, generated by the Generator, from real images.
In this talk, I'll be discussing how GANs can be used to achieve Super Resolution. Super Resolution is the process of upscaling and or improving the details within an image. Often a Low Resolution image is taken as an input and the same image is upscaled to a higher resolution, which is the output. The details in the High Resolution output are filled in where the details are essentially unknown.The traditional way of upscaling an image was to perform interpolation over the pixels such as Bicubic Interpolation. Drawbacks of this method are images that get smoothened or details and definitions are lost.
The need for high resolution is extensive right from a portrait we capture from our phone to image recognition, forensic science, etc.
-
keyboard_arrow_down
Vivek Singhal - How to start and grow an AI Lab
20 Mins
Experience Report
Beginner
I will discuss how I started CellStrat AI Lab and what goes to make it successful as one of the leading AI research groups in India.
I will cover facets such as origination, growth, researcher motivation, content portfolio, project activity, research areas, marketing techniques and talent development pipeline.
This presentation will give an idea to corporate and academic institutions as to how they can create and nurture a world-class AI Lab within their organizations.
-
keyboard_arrow_down
Vishal Singhal - How to Fulfill Management’s Enterprise AI Initiative
20 Mins
Talk
Executive
Artificial Intelligence fulfillment is a hot topic these days across continents and boardrooms of all small to mid-range to large companies alike. A common question for everybody is to understand as to how they can utilise the technology, deploy it in their business and what will it take to achive this business objective. “How to Fulfill Management’s Enterprise AI Initiative” topic addresses most of such queries and attempts to answer them taking a well-rounded perspective in the Indian Enterprise context.
-
keyboard_arrow_down
Vivek Singhal / Niraj Kale - Scalable and Efficient Object Detection with EfficientDet
Vivek SinghalCo-Founder & Chief Data ScientistCellStratNiraj KaleAI ResearcherCellStratschedule 2 years ago
20 Mins
Talk
Intermediate
There has been much research in efficient and scalable model approaches required for network design and object detection. Feature Pyramid Networks (FPN) enabled feature fusion at different scales. Path Aggregation Networks (PANet) conduct bottom-up path augmentation in an FPN, which shortens information transfer from bottom layers to topmost features. A more recent scheme, NAS-FPN (Neural Architecture Search-Feature Pyramid Network), combines RL-based NAS with an FPN to enable combination of top-down and bottom-up connections to fuse features across scales.
A recent state-of-the art model EfficientDet uses a weighted bi-directional BiFPN with multi-scale feature fusion across layers to account for learnable feature importance while applying bottom-up and top-down feature fusion. EfficientDet combines BiFPN with efficient model scaling techniques proposed in EfficientNet model, such that the baseline network, the feature network and the bounding box/class prediction networks are all scaled uniformly and efficiently using compound scaling technique across resolution/width/depth dimensions. The EfficientNet-D6 achieves top of the line accuracy (mAP) on COCO dataset with 4x lesser parameters and 13x fewer FLOPS then the recent comparable NAS-FPN model.
-
keyboard_arrow_down
Indrajit Singh - Modern QA System using BERT and DistilBERT
20 Mins
Experience Report
Advanced
A QA system finds the answers out of the help with the information retrieval, computational linguistics and knowledge representation of the data for finding those answers for the questions given. I have used the #google #BERT pretrained model for building a QA system. The QA system is extracting the data from Web/Database using IR and feeding the paragraph as input to the system which retrieves the answer based on the answer given as input to the system. BERT is one such pre-trained model developed by Google which can be fine-tuned on new data which can be used to create NLP systems like question answering, text generation, text classification, text summarization and sentiment analysis.
In this talk, I will be discussing the end to end architecture of modern Question Answering Systems and how they are different from chat-bots using SQUAD 2.0. How QA Based systems can help modernize multiple industries such as pharmaceutical, Automobile and Retail etc,domains in solving various complex challenges. I will be discussing in detail about masked language models and sentence prediction and how that's helping BERT/DistilBERT to be top production-ready algorithms.
We will see how various BERT variants like SciBERT and BioBERT are helping in focusing on a specific domain or area. A clear understanding of Token Embedding - Segmanet Embedding - Position Embedding as the primary base behind these algorithms.
-
keyboard_arrow_down
Shubha Manikarnike / Vivek Singhal - RL in Finance with ActorCritic
Shubha ManikarnikeDeep Reinforcement ResearcherCellstrat AI LabVivek SinghalCo-Founder & Chief Data ScientistCellStratschedule 2 years ago
20 Mins
Demonstration
Intermediate
Reinforcement Learning broadly involves Value-based methods and Policy-based Methods. Actor-based methods are suitable for discrete action spaces, whereas Policy-based methods are suitable for continuous action spaces.
A simple Policy Gradient algorithm (a policy-based method) predicts the probability of Actions of an Agent. Since, the Policy is updated using Monte Carlo approach (i.e. taking random samples), the variance of log probabilities produced by the PG algorithm is very high, causing unstable learning. This variance can be reduced by using a Baseline to remove outliers. A good choice to be used as a baseline is the Q-Value of the State, Action pair.
In this session, we will talk about Actor Critic algorithms, which uses two networks - an Actor and a Critic. The actor outputs the Policy, while the Critic gives out the Q-value which can be used to evaluate the Policy. In this zero-sum game, an equilibrium is reached when an ideal policy is discovered.
We will demonstrate with a Code demo, how this algorithm can be used to predict Actions in a Stock market scenario.
-
keyboard_arrow_down
Shubha Manikarnike - Collaboration and Competition in Multi-Agent RL
20 Mins
Demonstration
Intermediate
In Deep Reinforcement Learning, we know that policy Gradient methods work efficiently in predicting the probability of the Actions, in case of discrete Action Spaces and the exact action itself incase of continous Actions. The Predicted Action is evaluated by another Neural Network. We have used this to train a Robotic Arm to perform specific Tasks.
But what if we have more One Agent in an environment to be trained?.You can have an autonomous group of agents sharing a common environment. These Agents can work Collaboratively towards maximizing their total goal, or they could be interacting Competitively against each other to get the Reward, or they can be interacting in a mixed mode.
In this session, I will talk about Multi Agent Reinforcement Algorithms (MARL) which can be used in either collaborative mode, or competitive mode or a mixed mode between Agents. I will present a brief Code walk through of using the above algorithm in Open AI's Multi Agent Particle Environment.
-
keyboard_arrow_down
Niraj Kale - Quantum Neural Network
20 Mins
Talk
Beginner
Quantum computing has surged in the news and the new breakthroughs put it in a spotlight. Quantum computing can greatly enhance the ML/DL landscape. The new innovation quantum neural networks can greatly accelerate the training and inference of neural networks.
In this presentation we will see different architectures of QNN. QNN can be build using variational circuits.
The quantum variational circuits can be used as neural networks. The different architectures in quantum neural network are explored in this presentation. What are the advantages of QNN over traditional ANN. A simple image classification problem using QNN is demonstrated.
-
keyboard_arrow_down
Niraj Kale - Identifying COVID-19 patients using Image Classification of X-Rays
20 Mins
Demonstration
Intermediate
With the cases of COVID-19 exponentially rising, it is essential to identify the disease at an early stage. Also, given the volume of the patients, it is impossible to do testing on each and every person who feels sick. Also, when testing kits are scarce, we need to find an alternative method to classify between the COVID-19 and other cases. To solve this problem we can use the chest X-ray images of the patients and identify the patients with COVID-19 patients as the first level of screening.
This talk focuses on creating the dataset of normal Xrays and COVID-19 patient Xrays. What are the hurdles in using various pre-trained models? Identifying the correct pre-trained model for this application. This talk further focuses on training and creating a viable model to identify COVID-19 patients.
Furthermore, it talks about how to deploy this model quickly on the production server to make it available to the healthcare facilities so that they can do the pre-screening of the patients.