SQUAD application through Knowledge Graph for COVID-19 Literature
There are numerous documents and research papers being published for COVID-19 and doctors are not able to absorb the content of all the literature. It has become a real challenge to extract relevant information in a short span of time.
Knowledge Graph along with SQUAD application can help process multiple documents and extract precise information from a set of documents quickly. This will be a very handy application for healthcare professional to extract relevant information without going in detail with each application.
The session will demonstrate the following:
a) Text Processing of COVID-19 literature
b) Named Entity Extraction from the documents using BERT/Spacy
c) Building a Knowledge Graph of the documents
d) Building question-answer application
Outline/Structure of the Demonstration
1. Emerging landscape of AI in 2019-20 - 2 Mins
2. Semantic Information Extraction from text 5 Mins
3. Google BERT VS Knowledge Graph - 5 Mins
4. Building Knowledge Graph and Information Retreival - 5 Mins
5. Q&A - 3 Mins
Learning Outcome
1. Understand Text Processing for Natural Language Processing
2. Learn Named Entity Relationship (NER)
3. Understand Creation of Knowledge Graph
4. Information extraction from Knowledge Graph
Target Audience
Healthcare Professionals, Doctors, AI Researchers
Video
Links
https://www.youtube.com/watch?v=x5SlgaW2WDA&t=3s
https://youtu.be/gz0e82YSCTU
schedule Submitted 3 years ago
People who liked this proposal, also liked:
-
keyboard_arrow_down
Sharmistha Chatterjee - Machine Learning and Data Governance in Telecom Industry
20 Mins
Demonstration
Intermediate
The key aspect in solving ML problems in telecom industry lies in continuous data collection and evaluation from different categories of customers and networks so as to track and dive into varying performance metrics. The KPIs form the basis of network monitoring helping network/telecom operators to automatically add and scale network resources. Such smart automated systems are built with the objective of increasing customer engagement through enhanced customer experience and tracking customer behavior anomaly with timely detection and correction. Further the system is designed to scale and serve current LTE, 4G and upcoming 5G networks with minimal non-effective cell site visits and quick identification of Root Cause Analysis (RCA).
Network congestion has remained an ever-increasing problem. Operators have attempted a variety of strategies to match the network demand capacity with existing infrastructure, as the cost of deploying additional network capacities is expensive. To keep the cost under control, operators apply control measures to attempt to allocate bandwidth fairly among users and throttle the bandwidth of users that consume excessive bandwidth. This approach had limited success. Alternatively, techniques that utilize extra bandwidth for quality of experience (QOE) efficiency by over-provisioning the network has proved to be ineffective and inefficient due to lack of proper estimation.
The evolution of 5G networks, would lead manufacturers and telecom operators to use high-data transfer rates, wide network coverage, low latency to build smart factories using automation, artificial intelligence and Internet of Things (IoT). The application of advanced data science and AI can provide better predictive insights to improve network capacity-planning accuracy. Better network provisioning would yield better network utilization for both next-generation networks based on 5G technology and current LTE and 4G networks. Further AI models can be designed to link application throughput with network performance, prompting users to plan their daily usage based on their current location and total monthly budget.
In this talk, we will understand the current challenges in the telecom industry, the need for an AIOPS platform, and the mission held by telecom operators, communication service providers across the world for designing such AI frameworks, platforms, and best practices. We will see how increasing operator collaborations are helping to create, deploy and produce AI platforms for different AI use-cases. We will study one industrial use-case (with code) based on real-time field research to predict network capacity. In this respect, we will investigate how deep learning networks can be used to train large volumes of data at scale (millions of network cells), and how its use can help the upcoming 5G networks. We will also examine an end-to-end pipeline of hosting the scalable framework on Google Cloud. As data volume is huge and data needs to be stored in highly secured systems, we build our high-performing system with extra security features that can process millions of request in an order of few mili-secs. As the session highlights parameters and metrics in creating a LSTM based neural network, it also discusses the challenges and some of the key aspects involved in designing and scaling the system.
-
keyboard_arrow_down
Darshan Ganji - On-Demand Accelerating Deep Neural Network Inference via Edge Computing
20 Mins
Demonstration
Advanced
Deep Neural networks are both computationally intensive and memory intensive, making them difficult to deploy on mobile phones and embedded systems with limited hardware resources and taking more time for Inference and Training. For many mobile-first companies such as Baidu and Facebook, various apps are updated via different app stores, and they are very sensitive to the size of the binary files. For example, App Store has the restriction “apps above 100 MB will not download until you connect to Wi-Fi”. As a result, a feature that increases the binary size by 100MB will receive much more scrutiny than one that increases it by 10MB. It is challenging to run computation-intensive DNN-based tasks on mobile devices due to the limited computation resources.
This talk introduces the Algorithms and Hardware that can be used to accelerate the Inferencing or reduce the latency of deep learning workloads. We will discuss how to compress the Deep Neural Networks and techniques like Graph Fusion, Kernel Auto-Tuning for accelerating inference, as well as Data and model parallelization, automatic mixed precision, and other techniques for accelerating training. We will also discuss specialized hardware for deep learning such as GPUs, FPGAs, and ASICs, including the Tensor Cores in NVIDIA’s Volta GPUs as well as Google’s Tensor Processing Units (TPUs). We will also discuss the Deployment of the Large Size Deep Learning Models on the Edge devices.
Keywords: Graph Optimization, Tensor Fusion, Kernel Auto Tuning, Pruning, Weight sharing, quantization, low-rank approximations, binary networks, ternary networks, Winograd transformations, data parallelism, model parallelism, mixed precision, FP16, FP32, model distillation, Dense-Sparse-Dense training, NVIDIA Volta, Tensor Core, Google TPU.
-
keyboard_arrow_down
Vrishank Gupta / Saakshi Bhargava - Assembling a Perfect Personal Computer using Genetic Algorithms
20 Mins
Demonstration
Beginner
Assembling a perfect personal computer, that meets various varying requirements of a family such as gaming, regular usage, programming, etc., in such a huge market of features is quite a challenge nowadays.
Despite the efforts put by consumers to customize their computers to meet the different requirements, the percentage of satisfied consumers is very less. This session aims to propose and demonstrate a genetic algorithm approach to find the optimum set of features, given that each feature adds to the cost of the computer but provides some benefit to the consumer, the selected features must be fulfilled within a given budget. The experimental result yields the average fitness convergence at value 5524 which is a marked improvement over 23% over a recently published paper that used the Group Selection Technique along with single-point crossover for hardware selection.
-
keyboard_arrow_down
Saakshi Bhargava / Vrishank Gupta - Generative Adversarial Networks (GANs)
20 Mins
Demonstration
Beginner
Deep learning has accomplished pronounced triumph in the field of artificial intelligence, there are many deep learning models that have been developed in the recent time. Generative Models (GAN) are one of the deep learning models, that was given based on the game theory called zero-sum and now has been treated as the hot area for research. Generative Models are modern techniques used in computer vision. Unlike other neural networks that are used for predictions from images, generative models can generate new images as well for specific objectives. They can be used for generating huge datasets. This session will review several applications of generative modeling such as image generation and image translation, video frame prediction using CNNs and GANs.
-
keyboard_arrow_down
Gouthaman Asokan - Real Time Multi Person Pose Estimation
20 Mins
Demonstration
Intermediate
Openpose is a library written in C++ with python wrapper available for real time multi person key point detection and multithreading. This model predicts the location of various human keypoints such as chest, hips, shoulder, neck, elbows, knees. This model uses part affinity fields and greedy inference to connect these localized keypoints.
In this talk, I'll be discussing how Openpose helps in the real time multi person detection system to jointly detect human body,hand,facial and foot keypoints detection and the part affinity field.
Also,discuss the model architecture,comparing with other models like Mask RCNN and AlphaPose. Finally show how pose estimation can be done on single as well as multiple person images using pretrained models
-
keyboard_arrow_down
Abdul Azeez - Photo-Realistic Single Image Super-Resolution using SRGAN
20 Mins
Demonstration
Advanced
As we know, a basic GAN has two neural Networks – the Discriminator (D) and the Generator (G). The Generator attempts to generate images that look like real images. The Discriminator tries to distinguish the generated images from real images. By combined loss minimization of these two neural networks, the entire model trains and eventually reach a state of equilibrium, where the Discriminator no longer can distinguish the fake images, generated by the Generator, from real images.
In this talk, I'll be discussing how GANs can be used to achieve Super Resolution. Super Resolution is the process of upscaling and or improving the details within an image. Often a Low Resolution image is taken as an input and the same image is upscaled to a higher resolution, which is the output. The details in the High Resolution output are filled in where the details are essentially unknown.The traditional way of upscaling an image was to perform interpolation over the pixels such as Bicubic Interpolation. Drawbacks of this method are images that get smoothened or details and definitions are lost.
The need for high resolution is extensive right from a portrait we capture from our phone to image recognition, forensic science, etc.
-
keyboard_arrow_down
Vivek Singhal - How to start and grow an AI Lab
20 Mins
Experience Report
Beginner
I will discuss how I started CellStrat AI Lab and what goes to make it successful as one of the leading AI research groups in India.
I will cover facets such as origination, growth, researcher motivation, content portfolio, project activity, research areas, marketing techniques and talent development pipeline.
This presentation will give an idea to corporate and academic institutions as to how they can create and nurture a world-class AI Lab within their organizations.
-
keyboard_arrow_down
Vishal Singhal - How to Fulfill Management’s Enterprise AI Initiative
20 Mins
Talk
Executive
Artificial Intelligence fulfillment is a hot topic these days across continents and boardrooms of all small to mid-range to large companies alike. A common question for everybody is to understand as to how they can utilise the technology, deploy it in their business and what will it take to achive this business objective. “How to Fulfill Management’s Enterprise AI Initiative” topic addresses most of such queries and attempts to answer them taking a well-rounded perspective in the Indian Enterprise context.
-
keyboard_arrow_down
Vivek Singhal / Niraj Kale - Scalable and Efficient Object Detection with EfficientDet
Vivek SinghalCo-Founder & Chief Data ScientistCellStratNiraj KaleAI ResearcherCellStratschedule 3 years ago
20 Mins
Talk
Intermediate
There has been much research in efficient and scalable model approaches required for network design and object detection. Feature Pyramid Networks (FPN) enabled feature fusion at different scales. Path Aggregation Networks (PANet) conduct bottom-up path augmentation in an FPN, which shortens information transfer from bottom layers to topmost features. A more recent scheme, NAS-FPN (Neural Architecture Search-Feature Pyramid Network), combines RL-based NAS with an FPN to enable combination of top-down and bottom-up connections to fuse features across scales.
A recent state-of-the art model EfficientDet uses a weighted bi-directional BiFPN with multi-scale feature fusion across layers to account for learnable feature importance while applying bottom-up and top-down feature fusion. EfficientDet combines BiFPN with efficient model scaling techniques proposed in EfficientNet model, such that the baseline network, the feature network and the bounding box/class prediction networks are all scaled uniformly and efficiently using compound scaling technique across resolution/width/depth dimensions. The EfficientNet-D6 achieves top of the line accuracy (mAP) on COCO dataset with 4x lesser parameters and 13x fewer FLOPS then the recent comparable NAS-FPN model.
-
keyboard_arrow_down
Aditya Bhattacharya - Application of Masked RCNN for segmentation of brain haemorrhage from Computed Tomography Images
20 Mins
Demonstration
Intermediate
Automated analysis of CT scan images using AI solutions to diagnose abnormalities will help in overcoming the costly, time consuming and prone to error from manual analysis. Deep Learning has proved to be quite efficient to mimic human cognitive abilities (and even exceed that in many cases), especially with unstructured data.
DL algorithms can detect, localize and quantify a growing list of brain pathologies including intra-cerebral bleeds and their subtypes, infarcts, mass effect, midline shift, and cranial fractures. So, with advanced DL algorithms, analysis of radiographic data can be easily achieved and this can accelerate early detection of certain critical medical conditions, powered by AI.
As mentioned, Deep Learning algorithms for computer vision use cases has been extremely successful for classification and localization related problems. With the availability of annotated dataset, object of interest or region of interest segmentation using Deep Learning has been plausible.
Algorithms like Regional Convolutional Neural Network (RCNN) and it’s evolved forms, Faster RCNN and Masked RCNN is being widely used in the field of advanced radiology to auto detect medical conditions through radio-graphic images.
For this session, I am particularly going to talk about application of Masked RCNN for detection of regions of brain haemorrhage from CT scan images of the brain.
-
keyboard_arrow_down
Aditya Bhattacharya - Using Deep Learning to identify medical conditions related to Thorax Region from Radiographic X-Ray Images
20 Mins
Talk
Intermediate
Automated analysis of Chest X-ray images to diagnose various pathologies will help in overcoming the costly, time consuming and prone to error from manual analysis of them, especially using deep learning based approaches. One of such recent efforts in this direction is Classification of Common Thorax which combines the advantages of CNN based feature extraction and problem transformation methods in multi-label classification task.
So this is one of the key areas where deep learning based solution has already made an impact and has the potential to come up with even a better and well improved performance.
For this session, I am going to discuss about the problem at hand, the data-set, several approaches that has been explored and that worked quite well so far in this research. Also I am going to mention about the potential use case and the real world impact of such a real world healthcare application that can save millions of lives by early and effective detection.
Also I am going to mention about some of the key challenges faced during this research and how it can be scaled to build an end to end software solution! -
keyboard_arrow_down
Dr Purnendu Sekhar Das - Screening for COVID-19 Patients using Deep Convolutional Neural Networks on Chest X-Ray Images
20 Mins
Talk
Intermediate
As the COVID-19 pandemic rages on across the world as one of the most catastrophic healthcare crises in recent memory, the medical world is grappling with several challenges at once to cope with this crisis in the best possible manner. Following problem areas need urgent attention:
- Timely screening and identification of infected Covid-19 patients to help in triaging of scarce healthcare resources
- Monitoring the response to clinical interventions to decide best treatment possible.
- Rapid scale-up of capacities in both testing of new patients and monitoring treatment response
The current gold standard for confirmatory diagnosis of SARS-CoV2 infection is the RT-PCR lab test which is an expensive and time-consuming test relying on identification of viral genetic material – it is not possible to quickly scale up testing capacity using this technique in countries like India due to lack of the required equipment. In this scenario, we need to come up with an accurate and rapidly scalable method as an alternative testing strategy.
The Chest X-Ray is a basic test that is widely available in healthcare facilities across India and the developing world and can be done and analyzed within a short time frame.
Deep Convolutional Neural Networks offer great promise in automating the classification of Chest X-ray images to accurately identify COVID-19 patients from patients with normal lungs and those with other lung pathology related conditions like viral and bacterial pneumonias. This technique can be used as a reliable screening methodology to identify COVID-19 patients and recommend the appropriate therapy before it is too late.
Various CNN architectures have been explored recently to perform efficient image classification and segmentation on lung radiology images originating from X-Ray and CT Scans. These Deep Learning based techniques leverage the power of CNNs to extract hidden features from radiology images that can further be exploited for accurate identification of the disease process affecting the patient.
This technique promises to be a game changer in COVID-19 screening that can both be scaled up rapidly and can also be used as a Clinical Decision Support Tool by radiologists inside hospitals, critical care units and isolation wards. It also has the potential to reduce the time required by experienced radiologists to screen through huge volumes of Chest X-Ray images, while maintaining optimal levels of accuracy in diagnosis.
-
keyboard_arrow_down
Indrajit Singh - Modern QA System using BERT and DistilBERT
20 Mins
Experience Report
Advanced
A QA system finds the answers out of the help with the information retrieval, computational linguistics and knowledge representation of the data for finding those answers for the questions given. I have used the #google #BERT pretrained model for building a QA system. The QA system is extracting the data from Web/Database using IR and feeding the paragraph as input to the system which retrieves the answer based on the answer given as input to the system. BERT is one such pre-trained model developed by Google which can be fine-tuned on new data which can be used to create NLP systems like question answering, text generation, text classification, text summarization and sentiment analysis.
In this talk, I will be discussing the end to end architecture of modern Question Answering Systems and how they are different from chat-bots using SQUAD 2.0. How QA Based systems can help modernize multiple industries such as pharmaceutical, Automobile and Retail etc,domains in solving various complex challenges. I will be discussing in detail about masked language models and sentence prediction and how that's helping BERT/DistilBERT to be top production-ready algorithms.
We will see how various BERT variants like SciBERT and BioBERT are helping in focusing on a specific domain or area. A clear understanding of Token Embedding - Segmanet Embedding - Position Embedding as the primary base behind these algorithms.
-
keyboard_arrow_down
Shubha Manikarnike / Vivek Singhal - RL in Finance with ActorCritic
Shubha ManikarnikeDeep Reinforcement ResearcherCellstrat AI LabVivek SinghalCo-Founder & Chief Data ScientistCellStratschedule 3 years ago
20 Mins
Demonstration
Intermediate
Reinforcement Learning broadly involves Value-based methods and Policy-based Methods. Actor-based methods are suitable for discrete action spaces, whereas Policy-based methods are suitable for continuous action spaces.
A simple Policy Gradient algorithm (a policy-based method) predicts the probability of Actions of an Agent. Since, the Policy is updated using Monte Carlo approach (i.e. taking random samples), the variance of log probabilities produced by the PG algorithm is very high, causing unstable learning. This variance can be reduced by using a Baseline to remove outliers. A good choice to be used as a baseline is the Q-Value of the State, Action pair.
In this session, we will talk about Actor Critic algorithms, which uses two networks - an Actor and a Critic. The actor outputs the Policy, while the Critic gives out the Q-value which can be used to evaluate the Policy. In this zero-sum game, an equilibrium is reached when an ideal policy is discovered.
We will demonstrate with a Code demo, how this algorithm can be used to predict Actions in a Stock market scenario.
-
keyboard_arrow_down
Shubha Manikarnike - Collaboration and Competition in Multi-Agent RL
20 Mins
Demonstration
Intermediate
In Deep Reinforcement Learning, we know that policy Gradient methods work efficiently in predicting the probability of the Actions, in case of discrete Action Spaces and the exact action itself incase of continous Actions. The Predicted Action is evaluated by another Neural Network. We have used this to train a Robotic Arm to perform specific Tasks.
But what if we have more One Agent in an environment to be trained?.You can have an autonomous group of agents sharing a common environment. These Agents can work Collaboratively towards maximizing their total goal, or they could be interacting Competitively against each other to get the Reward, or they can be interacting in a mixed mode.
In this session, I will talk about Multi Agent Reinforcement Algorithms (MARL) which can be used in either collaborative mode, or competitive mode or a mixed mode between Agents. I will present a brief Code walk through of using the above algorithm in Open AI's Multi Agent Particle Environment.
-
keyboard_arrow_down
Bhavesh Laddagiri - Getting closer to Artificial General Intelligence with Neuro-Symbolic AI
20 Mins
Demonstration
Advanced
The recent advances in AI over the last decade has been heavily powered by Deep Learning with a new SOTA algorithm being released every other week by developers and researchers all across the globe. The concept of neural networks came into existence in the 1960s but made a come-back when fueled by the massive datasets and GPU farms. But the field of AI is much richer than just this one type of algorithm. Symbolic reasoning based algorithms also pioneered in the 1960s and were more practical for the time as they were less computationally intensive.
Deep Learning is great at understanding and finding patterns in massive streams of data but fails when it comes to using logic and reasoning to understand a concept with minimal data. Symbolic AI, on the other hand, is good at reasoning and creating abstractions from pre-defined concepts but fails to scale to real-world problems. However, there is something interesting here, the weaknesses of neural nets are actually the strengths of symbolic AI (and vice-versa), for the most part. So why not, Hybridize AI with Neural Nets and Symbolic AI to pave out a path for human-level AGI?
In this session, we are going to learn how this radical approach to AI can help solve real-world problems by demonstrating my proof-of-concept which can navigate a simple environment using natural language commands in a few shots of training. This demo will also allow the attendees to understand the usage of pretrained models for generalizing to new tasks using a multi-modal approach. Finally, we will look at the industry use cases of Neuro-Symbolic AI and the potential advantages of this approach over vanilla deep learning in that domain.
-
keyboard_arrow_down
Niraj Kale - Quantum Neural Network
20 Mins
Talk
Beginner
Quantum computing has surged in the news and the new breakthroughs put it in a spotlight. Quantum computing can greatly enhance the ML/DL landscape. The new innovation quantum neural networks can greatly accelerate the training and inference of neural networks.
In this presentation we will see different architectures of QNN. QNN can be build using variational circuits.
The quantum variational circuits can be used as neural networks. The different architectures in quantum neural network are explored in this presentation. What are the advantages of QNN over traditional ANN. A simple image classification problem using QNN is demonstrated.
-
keyboard_arrow_down
Rajesh Sampathkumar - Defence Against the Dark Arts: Taking Deep Neural Networks from Confusion to Clarity
20 Mins
Talk
Advanced
Deep Neural Networks are being used in production scale systems for computer vision, audio/video classification and processing, and in many other rich-data centric situations. However, at their core, neural networks are found to be susceptible to noisy, so-called adversarial input, which can potentially enable them to be hacked, or otherwise attacked. How do we understand and quantify the robustness of neural networks to such input? Are there best practices and methods that enable us to understand and accommodate these kinds of attacks on neural networks?
This talk addresses this specific topic, starting from some of the fundamentals, while exploring adversarial examples, their generation and how they can impact neural network efficacy. We then see how real world attack patterns are related to these adversarial examples, and evaluate some of the scenarios that may arise, specifically in a telecommunications network context. We then discuss some general ideas around defence against such adversarial examples. Additionally, we may also discuss elements of explainability and interpretability as relevant to adversarial attacks on neural networks, and practical considerations with respect to implementation of these techniques and best practices. Relevant research or resources on adversarial attacks and how they can be addressed for deep learning models may also be discussed if time permits.
-
keyboard_arrow_down
Mohammad Salim Ansari / Dr Purnendu Sekhar Das - Using AI to predict ICU treatment cycle
Mohammad Salim AnsariDirectorAppstract Mobile SolutionsDr Purnendu Sekhar DasPrincipal Data ScientistAccentureschedule 3 years ago
20 Mins
Talk
Intermediate
While handling the a patient in ICU for treatment there are two main chains of actions that need to be optimized
- Forecasting the progression of the treatment in the ICU for each patient from the time he/she enters the ICU to the time he exits
- Optimizing the treatment and testing for the complete cycle leading to the reduction of the overall cost and equipment cost
Using AI we can address both of the above using Deep learning and Deep RL .By addressing these jointly and create a overall solution using the following Algorithmic solution:
- Running Policy – Using LSTM
- Generate Patient Experience – Using treatment history – LSTM
- Training Sequential DQN – Create treatment schedule - Deep RL
- Per time Off policy Evaluation – evaluate and optimize the Treatment/testing Schedule- Deep RL
The idea is to create the forecasting model using LSTM and optimize the treatment/testing cycle using Deep RL and give a consolidated solution to the Healthcare establishment, which can give them a holistic view for each ICU patient from the time he/she is tested to the discharge time both in terms of time and the Cost of equipment .
I plan to present a brief Code walk through of using the above algorithms
-
keyboard_arrow_down
Niraj Kale - Identifying COVID-19 patients using Image Classification of X-Rays
20 Mins
Demonstration
Intermediate
With the cases of COVID-19 exponentially rising, it is essential to identify the disease at an early stage. Also, given the volume of the patients, it is impossible to do testing on each and every person who feels sick. Also, when testing kits are scarce, we need to find an alternative method to classify between the COVID-19 and other cases. To solve this problem we can use the chest X-ray images of the patients and identify the patients with COVID-19 patients as the first level of screening.
This talk focuses on creating the dataset of normal Xrays and COVID-19 patient Xrays. What are the hurdles in using various pre-trained models? Identifying the correct pre-trained model for this application. This talk further focuses on training and creating a viable model to identify COVID-19 patients.
Furthermore, it talks about how to deploy this model quickly on the production server to make it available to the healthcare facilities so that they can do the pre-screening of the patients.