schedule Aug 8th 09:00 - 09:45 AM place Grand Ball Room people 13 Interested

Since we originally proposed the need for a first-class language, compiler and ecosystem for machine learning (ML) - a view that is increasingly shared by many, there have been plenty of interesting developments in the field. Not only have the tradeoffs in existing systems, such as TensorFlow and PyTorch, not been resolved, but they are clearer than ever now that both frameworks contain distinct "static graph" and "eager execution" interfaces. Meanwhile, the idea of ML models fundamentally being differentiable algorithms – often called differentiable programming – has caught on.

Where current frameworks fall short, several exciting new projects have sprung up that dispense with graphs entirely, to bring differentiable programming to the mainstream. Myia, by the Theano team, differentiates and compiles a subset of Python to high-performance GPU code. Swift for TensorFlow extends Swift so that compatible functions can be compiled to TensorFlow graphs. And finally, the Flux ecosystem is extending Julia’s compiler with a number of ML-focused tools, including first-class gradients, just-in-time CUDA kernel compilation, automatic batching and support for new hardware such as TPUs.

This talk will demonstrate how Julia is increasingly becoming a natural language for machine learning, the kind of libraries and applications the Julia community is building, the contributions from India (there are many!), and our plans going forward.

 
7 favorite thumb_down thumb_up 1 comment visibility_off  Remove from Watchlist visibility  Add to Watchlist
 

Target Audience

All

schedule Submitted 3 months ago

Public Feedback

comment Suggest improvements to the Speaker
  • Anoop Kulkarni
    By Anoop Kulkarni  ~  3 months ago
    reply Reply

    Viral, thanks for your submission. I have been using Julia for some time now as part of quantum computing.. moving only recently to Julia for machine learning.. looking foward to this talk from you given your pedigree in the area


    ~anoop


  • Liked Anuj Gupta
    keyboard_arrow_down

    Anuj Gupta - Continuous Learning Systems: Building ML systems that keep learning from their mistakes

    Anuj Gupta
    Anuj Gupta
    Scientist
    Intuit
    schedule 1 month ago
    Sold Out!
    45 Mins
    Talk
    Beginner

    Won't it be great to have ML models that can update their “learning” as and when they make mistake and correction is provided in real time? In this talk we look at a concrete business use case which warrants such a system. We will take a deep dive to understand the use case and how we went about building a continuously learning system for text classification. The approaches we took, the results we got.

    For most machine learning systems, “train once, just predict thereafter” paradigm works well. However, there are scenarios when this paradigm does not suffice. The model needs to be updated often enough. Two of the most common cases are:

    1. When the distribution is non-stationary i.e. the distribution of the data changes. This implies that with time the test data will have very different distribution from the training data.
    2. The model needs to learn from its mistakes.

    While (1) is often addressed by retraining the model, (2) is often addressed using batch update. Batch updation requires collecting a sizeable number of feedback points. What if you have much fewer feedback points? You need model that can learn continuously - as and when model makes a mistake and feedback is provided. To best of our knowledge there is a very limited literature on this.

  • Liked Dr. Vikas Agrawal
    keyboard_arrow_down

    Dr. Vikas Agrawal - Non-Stationary Time Series: Finding Relationships Between Changing Processes for Enterprise Prescriptive Systems

    45 Mins
    Talk
    Intermediate

    It is too tedious to keep on asking questions, seek explanations or set thresholds for trends or anomalies. Why not find problems before they happen, find explanations for the glitches and suggest shortest paths to fixing them? Businesses are always changing along with their competitive environment and processes. No static model can handle that. Using dynamic models that find time-delayed interactions between multiple time series, we need to make proactive forecasts of anomalous trends of risks and opportunities in operations, sales, revenue and personnel, based on multiple factors influencing each other over time. We need to know how to set what is “normal” and determine when the business processes from six months ago do not apply any more, or only applies to 35% of the cases today, while explaining the causes of risk and sources of opportunity, their relative directions and magnitude, in the context of the decision-making and transactional applications, using state-of-the-art techniques.

    Real world processes and businesses keeps changing, with one moving part changing another over time. Can we capture these changing relationships? Can we use multiple variables to find risks on key interesting ones? We will take a fun journey culminating in the most recent developments in the field. What methods work well and which break? What can we use in practice?

    For instance, we can show a CEO that they would miss their revenue target by over 6% for the quarter, and tell us why i.e. in what ways has their business changed over the last year. Then we provide the prioritized ordered lists of quickest, cheapest and least risky paths to help turn them over the tide, with estimates of relative costs and expected probability of success.

  • Liked Favio Vázquez
    keyboard_arrow_down

    Favio Vázquez - Complete Data Science Workflows with Open Source Tools

    90 Mins
    Tutorial
    Beginner

    Cleaning, preparing , transforming, exploring data and modeling it's what we hear all the time about data science, and these steps maybe the most important ones. But that's not the only thing about data science, in this talk you will learn how the combination of Apache Spark, Optimus, the Python ecosystem and Data Operations can form a whole framework for data science that will allow you and your company to go further, and beyond common sense and intuition to solve complex business problems.

  • Liked Paolo Tamagnini
    keyboard_arrow_down

    Paolo Tamagnini / Kathrin Melcher - Guided Analytics - Building Applications for Automated Machine Learning

    90 Mins
    Tutorial
    Beginner

    In recent years, a wealth of tools has appeared that automate the machine learning cycle inside a black box. We take a different stance. Automation should not result in black boxes, hiding the interesting pieces from everyone. Modern data science should allow automation and interaction to be combined flexibly into a more transparent solution.

    In some specific cases, if the analysis scenario is well defined, then full automation might make sense. However, more often than not, these scenarios are not that well defined and not that easy to control. In these cases, a certain amount of interaction with the user is highly desirable.

    By mixing and matching interaction with automation, we can use Guided Analytics to develop predictive models on the fly. More interestingly, by leveraging automated machine learning and interactive dashboard components, custom Guided Analytics Applications, tailored to your business needs, can be created in a few minutes.

    We'll build an application for automated machine learning using KNIME Software. It will have an input user interface to control the settings for data preparation, model training (e.g. using deep learning, random forest, etc.), hyperparameter optimization, and feature engineering. We'll also create an interactive dashboard to visualize the results with model interpretability techniques. At the conclusion of the workshop, the application will be deployed and run from a web browser.

  • Liked Dr. C.S.Jyothirmayee
    keyboard_arrow_down

    Dr. C.S.Jyothirmayee / Usha Rengaraju / Vijayalakshmi Mahadevan - Deep learning powered Genomic Research

    90 Mins
    Workshop
    Advanced

    The event disease happens when there is a slip in the finely orchestrated dance between physiology, environment and genes. Treatment with chemicals (natural, synthetic or combination) solved some diseases but others persisted and got propagated along the generations. Molecular basis of disease became prime center of studies to understand and to analyze root cause. Cancer also showed a way that origin of disease, detection, prognosis and treatment along with cure was not so uncomplicated process. Treatment of diseases had to be done case by case basis (no one size fits).

    With the advent of next generation sequencing, high through put analysis, enhanced computing power and new aspirations with neural network to address this conundrum of complicated genetic elements (structure and function of various genes in our systems). This requires the genomic material extraction, their sequencing (automated system) and analysis to map the strings of As, Ts, Gs, and Cs which yields genomic dataset. These datasets are too large for traditional and applied statistical techniques. Consequently, the important signals are often incredibly small along with blaring technical noise. This further requires far more sophisticated analysis techniques. Artificial intelligence and deep learning gives us the power to draw clinically useful information from the genetic datasets obtained by sequencing.

    Precision of these analyses have become vital and way forward for disease detection, its predisposition, empowers medical authorities to make fair and situationally decision about patient treatment strategies. This kind of genomic profiling, prediction and mode of disease management is useful to tailoring FDA approved treatment strategies based on these molecular disease drivers and patient’s molecular makeup.

    Now, the present scenario encourages designing, developing, testing of medicine based on existing genetic insights and models. Deep learning models are helping to analyze and interpreting tiny genetic variations ( like SNPs – Single Nucleotide Polymorphisms) which result in unraveling of crucial cellular process like metabolism, DNA wear and tear. These models are also responsible in identifying disease like cancer risk signatures from various body fluids. They have the immense potential to revolutionize healthcare ecosystem. Clinical data collection is not streamlined and done in a haphazard manner and the requirement of data to be amenable to a uniform fetchable and possibility to be combined with genetic information would power the value, interpretation and decisive patient treatment modalities and their outcomes.

    There is hugh inflow of medical data from emerging human wearable technologies, along with other health data integrated with ability to do quickly carry out complex analyses on rich genomic databases over the cloud technologies … would revitalize disease fighting capability of humans. Last but still upcoming area of application in direct to consumer genomics (success of 23andMe).

    This road map promises an end-to-end system to face disease in its all forms and nature. Medical research, and its applications like gene therapies, gene editing technologies like CRISPR, molecular diagnostics and precision medicine could be revolutionized by tailoring a high-throughput computing method and its application to enhanced genomic datasets.

  • Liked Badri Narayanan Gopalakrishnan
    keyboard_arrow_down

    Badri Narayanan Gopalakrishnan / Shalini Sinha / Usha Rengaraju - Lifting Up: Deep Learning for implementing anti-hunger and anti-poverty programs

    45 Mins
    Case Study
    Intermediate

    Ending poverty and zero hunger are top two goals United Nations aims to achieve by 2030 under its sustainable development program. Hunger and poverty are byproducts of multiple factors and fighting them require multi-fold effort from all stakeholders. Artificial Intelligence and Machine learning has transformed the way we live, work and interact. However economics of business has limited its application to few segments of the society. A much conscious effort is needed to bring the power of AI to the benefits of the ones who actually need it the most – people below the poverty line. Here we present our thoughts on how deep learning and big data analytics can be combined to enable effective implementation of anti-poverty programs. The advancements in deep learning , micro diagnostics combined with effective technology policy is the right recipe for a progressive growth of a nation. Deep learning can help identify poverty zones across the globe based on night time images where the level of light correlates to higher economic growth. Once the areas of lower economic growth are identified, geographic and demographic data can be combined to establish micro level diagnostics of these underdeveloped area. The insights from the data can help plan an effective intervention program. Machine Learning can be further used to identify potential donors, investors and contributors across the globe based on their skill-set, interest, history, ethnicity, purchasing power and their native connect to the location of the proposed program. Adequate resource allocation and efficient design of the program will also not guarantee success of a program unless the project execution is supervised at grass-root level. Data Analytics can be used to monitor project progress, effectiveness and detect anomaly in case of any fraud or mismanagement of funds.

  • Liked Govind Chada
    keyboard_arrow_down

    Govind Chada - Using 3D Convolutional Neural Networks with Visual Insights for Classification of Lung Nodules and Early Detection of Lung Cancer

    Govind Chada
    Govind Chada
    Student
    Cy Woods
    schedule 1 month ago
    Sold Out!
    45 Mins
    Case Study
    Intermediate

    Lung cancer is the leading cause of cancer death among both men and women in the U.S., with more than a hundred thousand deaths every year. The five-year survival rate is only 17%; however, early detection of malignant lung nodules significantly improves the chances of survival and prognosis.

    This study aims to show that 3D Convolutional Neural Networks (CNNs) which use the full 3D nature of the input data perform better in classifying lung nodules compared to previously used 2D CNNs. It also demonstrates an approach to develop an optimized 3D CNN that performs with state of art classification accuracies. CNNs, like other deep neural networks, have been black boxes giving users no understanding of why they predict what they predict. This study, for the first time, demonstrates that Gradient-weighted Class Activation Mapping (Grad-CAM) techniques can provide visual explanations for model decisions in lung nodule classification by highlighting discriminative regions. Several CNN architectures using Keras and TensorFlow were implemented as part of this study. The publicly available LUNA16 dataset, comprising 888 CT scans with candidate nodules manually annotated by radiologists, was used to train and test the models. The models were optimized by varying the hyperparameters, to reach accuracies exceeding 90%. Grad-CAM techniques were applied to the optimized 3D CNN to generate images that provide quality visual insights into the model decision making. The results demonstrate the promise of 3D CNNs as highly accurate and trustworthy classifiers for early lung cancer detection, leading to improved chances of survival and prognosis.

  • Liked Gaurav Godhwani
    keyboard_arrow_down

    Gaurav Godhwani / Swati Jaiswal - Fantastic Indian Open Datasets and Where to Find Them

    45 Mins
    Case Study
    Beginner

    With the big boom in Data Science and Analytics Industry in India, a lot of data scientists are keen on learning a variety of learning algorithms and data manipulation techniques. At the same time, there is this growing interest among data scientists to give back to the society, harness their acquired skills and help fix some of the major burning problems in the nation. But how does one go about finding meaningful datasets connecting to societal problems and plan data-for-good projects? This session will summarize our experience of working in Data-for-Good sector in last 5 years, sharing few interesting datasets and associated use-cases of employing machine learning and artificial intelligence in social sector. Indian social sector is replete with good volume of open data on attributes like annotated images, geospatial information, time-series, Indic languages, Satellite Imagery, etc. We will dive into understanding journey of a Data-for-Good project, getting essential open datasets and understand insights from certain data projects in development sector. Lastly, we will explore how we can work with various communities and scale our algorithmic experiments in meaningful contributions.

  • Liked Aditya Singh Tomar
    keyboard_arrow_down

    Aditya Singh Tomar - Building Your Own Data Visualization Platform

    Aditya Singh Tomar
    Aditya Singh Tomar
    Data Consultant
    ACT Insights
    schedule 4 weeks ago
    Sold Out!
    45 Mins
    Demonstration
    Beginner

    Ever thought about having a mini interactive visualization tool that caters to your specific requirements. That is the product I created when I started independent consulting. 2 years since, and I have now decided to make it public – even the source code.

    This session will give you an overview about creating a custom, personalized version of a visualization platform built on R and Shiny. We will focus on a mix of structure and flexibility to address the varying requirements. We will look at the code itself and the various components involved while exploring the customization options available to ensure that the outcome is truly a personal product.

  • Liked Deepak Mukunthu
    keyboard_arrow_down

    Deepak Mukunthu - Democratizing & Accelerating AI through Automated Machine Learning

    90 Mins
    Workshop
    Beginner

    Intelligent experiences powered by AI can seem like magic to users. Developing them, however, is pretty cumbersome involving a series of sequential and interconnected decisions along the way that is pretty time-consuming. What if there was an automated service that identifies the best machine learning pipelines for a given problem/data? Automated Machine Learning does exactly that!

    Automated ML is based on a breakthrough from our Microsoft Research division. The approach combines ideas from collaborative filtering and Bayesian optimization to search an enormous space of possible machine learning pipelines intelligently and efficiently. It's essentially a recommender system for machine learning pipelines. Similar to how streaming services recommend movies for users, Automated ML recommends machine learning pipelines for data sets.

    Just as important, Automated ML accomplishes all this without having to see the customer’s data, preserving privacy. Automated ML is designed to not look at the customer’s data. Customer data and execution of the machine learning pipeline both live in the customer’s cloud subscription (or their local machine), which they have complete control of. Only the results of each pipeline run are sent back to the Automated ML service, which then makes an intelligent, probabilistic choice of which pipelines should be tried next.

    By making Automated ML available through the Azure Machine Learning service (Python-based SDK), we're empowering data scientists with a powerful productivity tool. We also have Automated ML available through PowerBI so that business analysts and BI professionals can also take advantage of machine learning. For developers familiar with Visual Studio and C#, we now have Automated ML available in C#.Net. If you are a SQL data engineer, we have a solution for you as well. And stay tuned as we continue to incorporate it into other product channels to bring the power of Automated ML to everyone!

    This session will provide an overview of Automated machine learning, how it works and how you can get started! We will walk through real-world use cases, build ML models using Automated ML and go through the E2E ML process of training, deployment, inferencing and operationalization of models.

  • Liked Dipanjan Sarkar
    keyboard_arrow_down

    Dipanjan Sarkar / Anuj Gupta - Natural Language Processing Bootcamp - Zero to Hero

    Dipanjan Sarkar
    Dipanjan Sarkar
    Data Scientist
    Red Hat
    Anuj Gupta
    Anuj Gupta
    Scientist
    Intuit
    schedule 5 months ago
    Sold Out!
    480 Mins
    Workshop
    Intermediate

    Data is the new oil and unstructured data, especially text, images and videos contain a wealth of information. However, due to the inherent complexity in processing and analyzing this data, people often refrain from spending extra time and effort in venturing out from structured datasets to analyze these unstructured sources of data, which can be a potential gold mine. Natural Language Processing (NLP) is all about leveraging tools, techniques and algorithms to process and understand natural language based unstructured data - text, speech and so on.

    Being specialized in domains like computer vision and natural language processing is no longer a luxury but a necessity which is expected of any data scientist in today’s fast-paced world! With a hands-on and interactive approach, we will understand essential concepts in NLP along with extensive case- studies and hands-on examples to master state-of-the-art tools, techniques and frameworks for actually applying NLP to solve real- world problems. We leverage Python 3 and the latest and best state-of- the-art frameworks including NLTK, Gensim, SpaCy, Scikit-Learn, TextBlob, Keras and TensorFlow to showcase our examples. You will be able to learn a fair bit of machine learning as well as deep learning in the context of NLP during this bootcamp.

    In our journey in this field, we have struggled with various problems, faced many challenges, and learned various lessons over time. This workshop is our way of giving back a major chunk of the knowledge we’ve gained in the world of text analytics and natural language processing, where building a fancy word cloud from a bunch of text documents is not enough anymore. You might have had questions like ‘What is the right technique to solve a problem?’, ‘How does text summarization really work?’ and ‘Which are the best frameworks to solve multi-class text categorization?’ among many other questions! Based on our prior knowledge and learnings from publishing a couple of books in this domain, this workshop should help readers avoid some of the pressing issues in NLP and learn effective strategies to master NLP.

    The intent of this workshop is to make you a hero in NLP so that you can start applying NLP to solve real-world problems. We start from zero and follow a comprehensive and structured approach to make you learn all the essentials in NLP. We will be covering the following aspects during the course of this workshop with hands-on examples and projects!

    • Basics of Natural Language and Python for NLP tasks
    • Text Processing and Wrangling
    • Text Understanding - POS, NER, Parsing
    • Text Representation - BOW, Embeddings, Contextual Embeddings
    • Text Similarity and Content Recommenders
    • Text Clustering
    • Topic Modeling
    • Text Summarization
    • Sentiment Analysis - Unsupervised & Supervised
    • Text Classification with Machine Learning and Deep Learning
    • Multi-class & Multi-Label Text Classification
    • Deep Transfer Learning and it's promise
    • Applying Deep Transfer Learning - Universal Sentence Encoders, ELMo and BERT for NLP tasks
    • Generative Deep Learning for NLP
    • Next Steps

    With over 10 hands-on projects, the bootcamp will be packed with plenty of hands-on examples for you to go through, try out and practice and we will try to keep theory to a minimum considering the limited time we have and the amount of ground we want to cover. We hope at the end of this workshop you can takeaway some useful methodologies to apply for solving NLP problems in the future. We will be using Python to showcase all our examples.

  • Liked Amit Doshi
    keyboard_arrow_down

    Amit Doshi - Integrating Digital Twin and AI for Smarter Engineering Decisions

    45 Mins
    Talk
    Intermediate

    With the increasing popularity of AI, new frontiers are emerging in predictive maintenance and manufacturing decision science. However, there are many complexities associated with modeling plant assets, training predictive models for them, and deploying these models at scale for near real-time decision support. This talk will discuss these complexities in the context of building an example system.

    First, you must have failure data to train a good model, but equipment failures can be expensive to introduce for the sake of building a data set! Instead, physical simulations can be used to create large, synthetic data sets to train a model with a variety of failure conditions.

    These systems also involve high-frequency data from many sensors, reporting at different times. The data must be time-aligned to apply calculations, which makes it difficult to design a streaming architecture. These challenges can be addressed through a stream processing framework that incorporates time-windowing and manages out-of-order data with Apache Kafka. The sensor data must then be synchronized for further signal processing before being passed to a machine learning model.

    As these architectures and software stacks mature in areas like manufacturing, it is increasingly important to enable engineers and domain experts in this workflow to build and deploy the machine learning models and work with system architects on the system integration. This talk also highlights the benefit of using apps and exposing the functionality through API layers to help make these systems more accessible and extensible across the workflow.

    This session will focus on building a system to address these challenges using MATLAB, Simulink. We will start with a physical model of an engineering asset and walk through the process of developing and deploying a machine learning model for that asset as a scalable and reliable cloud service.

  • Liked Anuj Gupta
    keyboard_arrow_down

    Anuj Gupta - NLP Bootcamp

    Anuj Gupta
    Anuj Gupta
    Scientist
    Intuit
    schedule 1 month ago
    Sold Out!
    480 Mins
    Workshop
    Beginner

    Recent advances in machine learning have rekindled the quest to build machines that can interact with outside environment like we human do - using visual clues, voice and text. An important piece of this trilogy are systems that can process and understand text in order to automate various workflows such as chat bots, named entity recognition, machine translation, information extraction, summarization, FAQ system, etc.

    A key step towards achieving any of the above task is - using the right set of techniques to represent text in a form that machine can understand easily. Unlike images, where directly using the intensity of pixels is a natural way to represent the image; in case of text there is no such natural representation. No matter how good is your ML algorithm, it can do only so much unless there is a richer way to represent underlying text data. Thus, whatever NLP application you are building, it’s imperative to find a good representation for your text data.

    In this bootcamp, we will understand key concepts, maths, and code behind the state-of-the-art techniques for text representation. We will cover mathematical explanations as well as implementation details of these techniques. This bootcamp aims to demystify, both - Theory (key concepts, maths) and Practice (code) that goes into building these techniques. At the end of this bootcamp participants would have gained a fundamental understanding of these schemes with an ability to implement them on datasets of their interest.

    This would be a 1-day instructor-led hands-on training session to learn and implement an end-to-end deep learning model for natural language processing.

  • 480 Mins
    Workshop
    Intermediate

    In this session, data scientists from CellStrat AI Lab will present demos and presentations on cutting-edge AI solutions in :-

    • Computer Vision - Image Segmentation with FCN/UNets/DeepLab/ESPNet, Image Processing, Pose Estimation with DensePose
    • Natural Language Processing (NLP) - Latest NLP and Text Analytics with BERT, NER, Neural Language Translation etc to solve problems such as text summarization, QnA systems, video captioning etc.
    • Reinforcement Learning (RL) - Train Atari Video Games with RL, Augmented Random Search, Deep Q Learning etc. Apply RL techniques for gaming, financial portfolios, driverless cars etc. Train Robots with MuJoCo simulator.
    • Driverless Cars - Demo on multi-class roads datasets, path planning and navigation control for cars etc.
    • Neural Network Architectures - Faster and Smaller Neural Networks with MorphNet
  • Liked Antrixsh Gupta
    keyboard_arrow_down

    Antrixsh Gupta - Creating Custom Interactive Data Visualization Dashboards with Bokeh

    90 Mins
    Workshop
    Beginner

    This will be a hands-on workshop how to build a custom interactive dashboard application on your local machine or on any cloud service provider. You will also learn how to deploy this application with both security and scalability in mind.

    Powerful Data visualization software solutions are extremely useful when building interactive data visualization dashboards. However, these types of solutions might not provide sufficient customization options. For those scenarios, you can use open source libraries like D3.js, Chart.js, or Bokeh to create custom dashboards. While these libraries offer a lot of flexibility for building dashboards with tailored features and visualizations.

  • 90 Mins
    Workshop
    Intermediate

    Machine learning and deep learning have been rapidly adopted in various spheres of medicine such as discovery of drug, disease diagnosis, Genomics, medical imaging and bioinformatics for translating biomedical data into improved human healthcare. Machine learning/deep learning based healthcare applications assist physicians to make faster, cheaper and more accurate diagnosis.

    We have successfully developed three deep learning based healthcare applications and are currently working on three more healthcare related projects. In this workshop, we will discuss one healthcare application titled "Deep Learning based Craniofacial Distance Measurement for Facial Reconstructive Surgery" which is developed by us using TensorFlow. Craniofacial Distances play important role in providing information related to facial structure. They include measurements of head and face that are to be measured from image.They are used in facial reconstructive surgeries such as cephalometry, treatment planning of various malocclusions, and craniofacial anomalies, where reliable and accurate data are very important and cannot be compromised.

    Our discussion on healthcare application will include precise problem statement, the major steps involved in the solution (deep learning based face detection, deep learning based facial landmarking and craniofacial distance measurement), data set collection, experimental analysis and challenges faced & overcame to achieve this success. Subsequently, we will provide hands-on exposure to implement this healthcare solution using TensorFlow. Finally, we will briefly discuss the possible extensions of our work and the future scope of research in healthcare sector.

  • 45 Mins
    Talk
    Intermediate

    This session will discuss Reinforcement Learning (RL) algorithms such as Policy Gradients, TD Learning and Deep-Q Learning. We will discuss how emerging RL algorithms can be used to train games, driverless cars, financial decision models and home automation systems.

  • Liked Indranil Basu
    keyboard_arrow_down

    Indranil Basu - Machine Generation of Recommended Image from Human Speech

    45 Mins
    Talk
    Advanced

    Introduction:

    Synthesizing audio for specific domains has many practical applications in creative sound design for music and film. But the application is not restricted to entertainment industry. We propose an architecture that will convert audio (human voice) to the voice owner’s preferred image – for the time being we restrict the intended images to two domains – Object Design and Human body. Many times, human beings are unable to describe a design (may be power-point presentation or interior decoration of a house) or a known person by verbally described attributes as they are able to visualise the same design or the person. But the other person, the listener may be unable to interpret the object or human descriptions from the speaker’s verbal descriptions as he/she is not visualising the same. Complete communication thus needs much of a trial and error and overall hazardous and time consuming. Examples of such situations are 1) While making presentation, an executive or manager can visualise something and an express to his/her employee to make the same. But, making the best slides from manger’s description may not be proper. Another relevant example is that a house owner or office owner wants his/her premises to have certain design which he/she can visualise and express to the concerned vendor. But the vendor may not be able to produce the same. Also, trial and error in this case is highly expensive. Having an automated Image, recommended to him/her can address this problem. 2) Verbal description of a terrorist or criminal suspect (facial description and/or attribute) may not be always available to all the security people every time, in Airports or Railway Stations or sensitive areas. Presence of a software system having Machine Generated Image with Ranked Recommendation for such suspect can immediately point to one or very few people in a crowded Airport or even Railway Station or any such sensitive place. Security agencies can then frisk only those people or match their attributes with existing database. This can avoid hazardous manual checking of every people in the same process and can help the security agencies to do adequate checking for those recommended individuals.

    We can use a Sequential Architecture consisting of simple NLP and more complex Deep Learning algorithms primarily based on Generative Adversarial Network (GAN) and Neural Personalised Ranking (NPR) to help the object designers and security personnel for serving their specific purposes.

    The idea to combat the problem:

    I propose a combination of Deep Learning and Recommender System approach to tackle this problem. Architecture of the Solution model consists of 4 major Components – 1) Speech to Text

    2) Text Classification into Person or Design; 3) Text to Image Formation; 4) Recommender System

    We are trying to address these four steps in consecutive applications of effective Machine Learning and Deep Learning Algorithms. Deep Learning community has already been able to make significant progress in terms of Text to Image generation and also in Ranking based Recommender System

    Brief Details about the four major pillars of this problem:

    Deep Learning based Speech Recognition – Primary technique for Speech to text could be Baidu’s DeepSpeech for which a Tensorflow implementation is readily available. Also, Google Cloud Speech-to-Text enables the develop to convert Voice to text. Voice of the user needs to be converted in .wav file. Our steps for Deep-Speech-2 are like this – Fixing GPU memory, Adding Batch normalization to RNN, implement row Convolution layer and generate text.

    Nowadays, we have quite a few free Speech to Text software, e.g. Google Docs Voice typing, windows Speech Recognition, Speech-notes etc.

    Text Classification of Content – This is needed to classify the converted text into two classes – a) Design Description or b) Human Attribute Description because these two applications and therefore image types are different. This may be Statistically easier part, but its importance is immense. A Dictionary of words related to Designs and Personal Attributes can be built using online available resources. Then, a supervised algorithm using tf-idf and Latent Semantic Analysis (LSA) should be able to classify the text into two classes – Object and Person. These are very much traditional and proven techniques in many NLP research

    Text to Image Formation – This is our main component for this proposal. Today, one of the most challenging problems in the world of Computer Vision is synthesizing high-quality images from text descriptions. In recent years, GANs have been found to generate good results. Samples generated by existing text-to-image approaches can roughly reflect the meaning of the given descriptions, but they fail to contain necessary details and vivid object parts. There have been a few approaches to address this problem, all using GAN. One of those is given as Stacked Generative Adversarial Networks (StackGAN). Heart of such approaches is Conditional GAN which is an extension of GAN where both generator and discriminator receive additional conditioning variables c, yielding G(z, c) and D(x, c). This formulation allows G to generate images conditioned on variables c.

    In our case, we train deep convolutional generative adversarial network (DC-GAN) conditioned on text features. These text features are encoded by a hybrid character-level convolutional-recurrent neural network. Overall, DC-GAN uses text embeddings where the context of a word is of prime importance. Class label determined in the earlier step will be of help in this case. This will simply help DC-GAN to generate more relevant images than irrelevant ones. Details will be discussed during the talk

    The most straightforward way to train a conditional GAN is to view (text, image) pairs as joint observations and train the discriminator to judge pairs as real or fake. The discriminator has no explicit notion of whether real training images match the text embedding context. To account for this, in GAN-CLS, in addition to the real/fake inputs to the discriminator during training, a third type of input consisting of real images with mismatched text is added, which the discriminator must learn to score as fake. By learning to optimize image/text matching in addition to the image realism, the discriminator can provide an additional signal to the generator. (details are in talk)

    Image Recommender System – In the last step, we propose personalised image recommendation for the user from the set of images generated by GAN-CLS architecture. Image Recommendation brings down the number of choice of images to a top N (N=3, 5, 10 ideally) with a rank given to each of those and therefore user finds it easier to choose. In this case, we propose Neural Personalized Ranking (NPR) – a personalized pairwise ranking model over implicit feedback datasets – that is inspired by Bayesian Personalized Ranking (BPR) and recent advances in neural networks. We like to mention that, now NPR is improved to contextual enhanced NPR. This enhanced Model depends on implicit feedbacks from the users, its contexts and incorporates the idea of generalized matrix factorization. Contextual NPR significantly outperforms its competitors

    In the presentation, we shall describe the complete sequence in detail

  • Liked Pankaj Kumar
    keyboard_arrow_down

    Pankaj Kumar / Abinash Panda / Usha Rengaraju - Quantitative Finance :Global macro trading strategy using Probabilistic Graphical Models

    90 Mins
    Workshop
    Advanced

    { This is a handson workshop in pgmpy package. The creator of pgmpy package Abinash Panda will do the code demonstration }

    Crude oil plays an important role in the macroeconomic stability and it heavily influences the performance of the global financial markets. Unexpected fluctuations in the real price of crude oil are detrimental to the welfare of both oil-importing and oil-exporting economies.Global macro hedge-funds view forecast the price of oil as one of the key variables in generating macroeconomic projections and it also plays an important role for policy makers in predicting recessions.

    Probabilistic Graphical Models can help in improving the accuracy of existing quantitative models for crude oil price prediction as it takes in to account many different macroeconomic and geopolitical variables .

    Hidden Markov Models are used to detect underlying regimes of the time-series data by discretising the continuous time-series data. In this workshop we use Baum-Welch algorithm for learning the HMMs, and Viterbi Algorithm to find the sequence of hidden states (i.e. the regimes) given the observed states (i.e. monthly differences) of the time-series.

    Belief Networks are used to analyse the probability of a regime in the Crude Oil given the evidence as a set of different regimes in the macroeconomic factors . Greedy Hill Climbing algorithm is used to learn the Belief Network, and the parameters are then learned using Bayesian Estimation using a K2 prior. Inference is then performed on the Belief Networks to obtain a forecast of the crude oil markets, and the forecast is tested on real data.

  • Liked Saikat Sarkar
    keyboard_arrow_down

    Saikat Sarkar / Dhanya Parameshwaran / Dr Sweta Choudhary / Raunak Bhandari / Srikanth Ramaswamy / Usha Rengaraju - AI meets Neuroscience

    480 Mins
    Workshop
    Advanced

    This is a mixer workshop with lot of clinicians , medical experts , Neuroimaging experts ,Neuroscientists, data scientists and statisticians will come under one roof to bring together this revolutionary workshop.

    The theme will be updated soon .

    Our celebrity and distinguished presenter Srikanth Ramaswamy who is an advisor at Mysuru Consulting Group and also works Blue Brain Project at the EPFL will be delivering an expert talk in the workshop.

    https://www.linkedin.com/in/ramaswamysrikanth/

    { This workshop will be a combination of panel discussions , expert talk and neuroimaging data science workshop ( applying machine learning and deep learning algorithms to Neuroimaging data sets}

    { We are currently onboarding several experts from Neuroscience domain --Neurosurgeons , Neuroscientists and Computational Neuroscientists .Details of the speakers will be released soon }

    Abstract for the Neuroimaging Data Science Part of the workshop:

    The study of the human brain with neuroimaging technologies is at the cusp of an exciting era of Big Data. Many data collection projects, such as the NIH-funded Human Connectome Project, have made large, high- quality datasets of human neuroimaging data freely available to researchers. These large data sets promise to provide important new insights about human brain structure and function, and to provide us the clues needed to address a variety of neurological and psychiatric disorders. However, neuroscience researchers still face substantial challenges in capitalizing on these data, because these Big Data require a different set of technical and theoretical tools than those that are required for analyzing traditional experimental data. These skills and ideas, collectively referred to as Data Science, include knowledge in computer science and software engineering, databases, machine learning and statistics, and data visualization.

    The workshop covers Data analysis, statistics and data visualization and applying cutting-edge analytics to complex and multimodal neuroimaging datasets . Topics which will be covered in this workshop are statistics, associative techniques, graph theoretical analysis, causal models, nonparametric inference, and meta-analytical synthesis.