Introduction :

"Special children" includes children who are affected with a complex neuro-behavioral conditions like autism, which includes impairments in social interaction, language development and communication skills, combined with rigid, repetitive behaviors. Children with autism particularly face a very difficult childhood as they have extreme difficulty in communication. They have trouble in understanding what other people think and feel. This makes it very hard for them to express themselves either with words or through gestures.

Such special children need “special” care for the development of their cognitive abilities. The amount of learning resources required for teaching such children are extremely hard to find and less accessible to many.

So, can artificial intelligence with the help of modern deep learning algorithms generate animated videos for developing or improving cognitive abilities of such a special group?

The idea to combat the problem:

Well, I feel it can be done!

An animated video consists of 3 main components:

1. Graphical video (sequence of images put together to tell a story),

2. A background story and

3. A relevant background audio or music.

Now if we have to come up with a system that produces machine generated animated video, we would have to think about these three components:

  1. Machine generated sequence of images with a spatial coherence
  2. Machine generated text, or the story
  3. Machine generated audio or music, that highlights the mood or the theme of the video

If these three discrete components are put together in a cohesive flow, our purpose can be achieved. And the Deep Learning community has already been able to make significant progress in terms of machine generated images and audio and machine generated text.

Details about the three pillars of this problem:

Machine generated sequence of images with a spatial coherence

Generative Adversarial Networks (GANs) has been quite successful till date to come up with generated images and audio. Also, for our use case, to maintain a coherency in spatial features, Variational Auto Encoders (VAEs) have been even better.

If we start with a popular use case of a very popular cartoon series, Tom & Jerry, specially modified for autistic children, let’s consider a simple scene where tom is chasing jerry. On an image level, for the entire scene, the posture of tom and jerry will remain constant, only their location will vary in every subsequent image frame in the entire scene. Which means, only their spatial location with respect to the entire image background will vary and hence VAEs will have the potential to implement such a use case as VAEs helps to provide probabilistic descriptions of features or observations in latent spaces.

Machine generated text, or the story

Coming to text generation or story generation, recurrent neural networks like Long/Short Term Memory (LSTM) has been quite successful. Already, LSTM has been used to artificially generate chapters from popular novels or stories like Harry Potter and Cinderella. So, for a simple animated video story specially structured for autistic children, LSTM can be effective. Although Gradient Recurrent Units (GRU) can be the other alternative, but till date LSTM has been more successful, so the first preference will always be LSTM.

Machine generated audio or music

For music generation, GANs have been proved effective till date. For our use case, Natural Language Processing or NLP can used to determine the type of scene from the generated story, e.g. for the Tom & Jerry scene, it will be a chase scene. Based on this classification, Deep Convolution Generative Adversarial Networks (DCGAN) can be used to generate music which is relevant to such a chase scene and at the same time be soothing and enjoyable to such children!

Assembling everything together

Now if we can put all these discrete pieces of the puzzle together, we can come up with a completely machine generated animated video tailor-made for developing and improving cognitive abilities of children with autism. This will be a new progress in the field of Artificial Intelligence!

These machine generated videos can be trained on Neural Network in such a way that it can be a source of fun and enjoyment for this special group and at the same time reward their good behavior and educate them in a sensitive way without any human dependency.

Future scope and extension

As a future scope, if this approach is successful, the gaming industry can adopt usage of such a technology and with the help of reinforcement learning, can come up with machine generated video games and educational games specially designed for such children that can disrupt the entire gaming industry and can be a source of happiness for such children!

2 favorite thumb_down thumb_up 4 comments visibility_off  Remove from Watchlist visibility  Add to Watchlist

Outline/Structure of the Talk

Introduction : Discussion on the problem statement

Objectives : Discussion on the target that can be achieved

Technical discussion on the three components of the problem:

  1. Image Generation using GAN and VAE
  2. Text Generation using LSTM and GRU
  3. Music Generation using GAN


Social Impact

Sustainability and Future Scope

AI for ALL

Brief Demonstration

Learning Outcome

The audience is expected to receive concrete knowledge on the following topic:

1. Computer Vision with Deep Learning

2. Image generation with VAEs

3. Use of deep learning in sequential data

4. Music generation with GAN

Also, some of the applications of this solution can be extended to other use cases as well which will be discussed during the talk

Target Audience

Researchers, Developers, AI Enthusiasts , Government Organizations and NGO members interested to know the potential of AI and anyone who wants to join our journey to improve the childhood experience of autistic children.

Prerequisites for Attendees

1. Basics of Machine Learning

2. Basics of Neural Network

3. Basics of Computer Vision with Deep Learning

4. Basics of Natural Language Processing

5. High level idea about modern deep learning algorithms

6. Passion for solving human life problems

schedule Submitted 2 weeks ago

Public Feedback

comment Suggest improvements to the Speaker
  • Dipanjan Sarkar
    By Dipanjan Sarkar  ~  1 week ago
    reply Reply

    Hi Aditya thanks for this submission. The topic definitely looks to be innovative and having a good scope.

    Few questions here: 

    1. Would you be showcasing any examples around how these are done?
    2. What I'm having trouble understanding is, in what way would the AI generated content be specially geared for children with autism vs normal children i.e how would it benefit them
    3. Would you have enough time to tie the three major components in your proposal and showcase how them connect together?



    - DJ

    • Aditya Bhattacharya
      By Aditya Bhattacharya  ~  1 week ago
      reply Reply

      Thanks DJ for reading my proposal. 

      Coming to your queries:

      1. Yes, I have plans to show how the three main components that I have talked about can be done.

      2. Generating content typically designed for autism requires extensive research and domain knowledge and a lot of training data. I might not be able to come up with the entire solution before ODSC, given the time constraint. But from what I was able to find out, autism requires different type of music like more extensive use of binaural beats , image patterns and textual structure as compared to that of normal content and hence I thought that it can be an interesting point to start with. I would try to demo my progress till then and would try to talk about challenges that I will face so that I can reach out for help from the community. 

      3. Time will be a major constraint, but definitely I would try to talk about how these three components can be tied together. But yet I don't want to make the content too lengthy with too much details as the audience might lose the focus!

      I hope I was able to answer most of your queries. If not, let's connect offline. Feel free to reach me through email or LinkedIn :)

      • Dipanjan Sarkar
        By Dipanjan Sarkar  ~  1 week ago
        reply Reply

        That is good. It's alright you don't need to build a full-fledged product or something but even if you can showcase some of these aspects it would be great, particularly, " But from what I was able to find out, autism requires different type of music like more extensive use of binaural beats , image patterns and textual structure as compared to that of normal content and hence I thought that it can be an interesting point to start with"

        I think if you can showcase some of these distinctions and how they would be of use to children with special needs, it would be amazing.

        • Aditya Bhattacharya
          By Aditya Bhattacharya  ~  1 week ago
          reply Reply

          Thanks DJ, definitely would try to work on and showcase some of these distinct features.

  • Liked Suvro Shankar Ghosh

    Suvro Shankar Ghosh - Learning Entity embedding’s form Knowledge Graph

    45 Mins
    Case Study
    • Over a period of time, a lot of Knowledge bases have evolved. A knowledge base is a structured way of storing information, typically in the following form Subject, Predicate, Object
    • Such Knowledge bases are an important resource for question answering and other tasks. But they often suffer from their incompleteness to resemble all the data in the world, and thereby lack of ability to reason over their discrete Entities and their unknown relationships. Here we can introduce an expressive neural tensor network that is suitable for reasoning over known relationships between two entities.
    • With such a model in place, we can ask questions, the model will try to predict the missing data links within the trained model and answer the questions, related to finding similar entities, reasoning over them and predicting various relationship types between two entities, not connected in the Knowledge Graph.
    • Knowledge Graph infoboxes were added to Google's search engine in May 2012

    What is the knowledge graph?

    ▶Knowledge in graph form!

    ▶Captures entities, attributes, and relationships

    More specifically, the “knowledge graph” is a database that collects millions of pieces of data about keywords people frequently search for on the World wide web and the intent behind those keywords, based on the already available content

    ▶In most cases, KGs is based on Semantic Web standards and have been generated by a mixture of automatic extraction from text or structured data, and manual curation work.

    ▶Structured Search & Exploration
    e.g. Google Knowledge Graph, Amazon Product Graph

    ▶Graph Mining & Network Analysis
    e.g. Facebook Entity Graph

    ▶Big Data Integration
    e.g. IBM Watson

    ▶Diffbot, GraphIQ, Maana, ParseHub, Reactor Labs, SpazioDati

  • 480 Mins


    Computer vision is an interesting area of Machine Learning and Deep Learning that still needs to evolve a lot to match human-level accuracy. So, I believe to inorder to increase the community strength for computer vision, we need more people equipped to work in this field and hence such a session on computer vision will be important. Hence this session will be a hands-on workshop on the introductory concepts of computer vision using deep learning algorithms and concepts.


    The objective of this workshop will be to teach the audience on the following concepts of Computer Vision:

    1. Image Manipulations like applying filters to an image
    2. Convolutional Neural Network
    3. Different CNN Architectures
    4. Image Classification
    5. Data Augmentation and other regularization techniques
    6. Transfer Learning
    7. Object Detection
    8. Neural Style Transfer
    9. Image Generation

    The target will be to teach the audience with introductory to intermediate level of knowledge on the above topics. Based on the audience background and interest, the level of details can be modified.

    Programming language will be Python. Preferred framework can be keras, tensorflow and open cv. Preferred medium of training will be Anaconda Spyder IDE or Jupyter notebook.

    The idea is to teach the fundamental concepts in each area, show sample code-snippet for each applications and provide equivalent assignments to the audience, so that the audience can learn by hands-on coding and gain practical experience on this area.

  • Liked Aditya Bhattacharya

    Aditya Bhattacharya - Impact of Data Science and Artificial Intelligence in Digital Transformation

    45 Mins

    Introduction :

    Most organizations, from Manufacturing to Healthcare to Financial sectors have realized the importance of modern technology to scale up their operations. So all massive organizations are going through it's Digital Transformation journey post 2010 and make most of the latest technology. This involves transformation of broadly two key components:

    1. The internal experience of the organization

    2. The external experience of the organization

    1. The internal experience talks about empowering the internal employees of the organization through modern tools and technologies so as to increase the overall business productivity.

    2. The external experience focuses mainly on the reducing the overall time to market for the external products yet keeping quality as the top priority and improving the quality of the client services provided by the organization.

    On a technology aspect, both internal and external experience depends on use of modern hardware, software and IT infrastructure. But since such a major breakthrough can not be always smooth for all organizations, the main question is, what can possibility go wrong in such a journey, considering time as a constraint since all organizations would want to reach there as quickly as possible?

    The major issue is related to the decisions that the organization takes. Decision about the overall budget to invest, decision about moving certain set of operations to the cloud, decisions about the security, decisions about the infrastructure cost, decisions about the ideal team size and etc.

    But all these decisions can be made more impactful when these are backed by data and supported by artificial intelligence.


    In this session, I would talk about how data science and AI can help such organizations take such critical decisions to improve user experience and scale their operations.

    In the session I would talk about certain use cases in which most companies end up taking wrong decisions while in the process of digital transformation and eventually end up taking a step back.

    Use Case 1: Move operations to the cloud for scaling up existing operations

    When we talk about digital transformation, moving some of the existing infrastructure to cloud or automating existing operations and managing operations remotely with the help of cloud platforms is a very common strategy. But the question is what can possibly go wrong in such cases?

    Although most cloud platforms provide pay as you use subscription models, but what the users really fail to estimate "what to use" ? The most probable decisions that go wrong is the "SKU" of the cloud resources, like what is the most optimal specifications of the cloud based resources. For example, if I am working on a machine learning algorithm on a cloud based VM, I won't be using a GPU and instead a CPU core will be sufficient and hence saving a lot of cost. Hence I propose for an AI system that will help the organization take such decisions and plan their budget accordingly.

    To start with, the user interface can be a simple chat-bot, which will ask the user about his requirements. Using Natural Language Processing (NLP) from the user requirements and using certain ML algorithms ( it is based on the complexity of the problem, and can be as simple as a decision tree ) the user will get the exact specifications and the suggestion about which cloud platform provider to go for, considering the cost in the long run. The system will also help to estimate, that if they go for such a cloud based approach, will it actually help the organization to improve their productivity and operations or not.

    Use Case 2: Planning the hiring strategy and finding the optimal team size

    Hiring strategy is something, which I have felt from my experience, that the companies often go wrong. If the organizations hire more people than required, in the long run, it can be a huge problem, both for the employees as well as the employers. In such a case, for the employees , they will get lesser opportunities to learn and grow and work on new and challenging products and services, which indirectly impacts their job satisfaction. If the hiring strength is lesser than what is required, it results in slower throughput, lower output and more work pressure for the employees.

    Now, how can the proposed AI system help organizations for such a use case?

    I believe the recruiting strategy has alot to do with the vision and mission of the company, along with capability of the organization to invest. So, the proposed system can have a similar chat-bot like UI to make these requirements, like to what scale the company wants to increase it's operations and how many end-point users that the company would like to reach.

    Now, employee churn rate, average employment period, geo-location and average employee salary depends on the specific job role. So, based on this parameter, the system can come up with a predictive model to tell what should be the number of people hired for a particular job role for the next two years based on the organization's demand and recruitment.

    Use Case 3: Securing the existing technology

    Every organization from Manufacturing, to supply chain, to typical software companies, invest a lot on security, may it be cyber security, may it be employee safety and security or it can be related to securing the infrastructures. Like the other two use cases, the proposed system can help to come up with better options in choosing the right specifications and investment budget for improving the overall security and detect and report anomaly whenever possible.

    So, these three use cases can be three key areas in which artificial intelligence backed by relevant data can help organizations to complete their journey of digital transformation very easily. There are plenty of other use cases as well, but from my experience the above three use cases are the major ones in which organizations struggle the most.

  • Liked Joy Mustafi

    Joy Mustafi - Human-Machine Interaction through Multi-Modal Interface with Combination of Speech, Text, Image and Sensor Data

    45 Mins


    In the context of human–computer interaction, a modality is the classification of a single independent channel of sensory input / output between a computer and a human. A system is designated uni-modal if it has only one modality implemented, and multi-modal if it has more than one. When multiple modalities are available for some tasks or aspects of a task, the system is said to have overlapping modalities. If multiple modalities are available for a task, the system is said to have redundant modalities. Multiple modalities can be used in combination to provide complementary methods that may be redundant but convey information more effectively. Modalities can be generally defined in two forms: human-computer and computer-human modalities.

    With the increasing popularity of smartphones, the general public are becoming more comfortable with the more complex modalities. Speech recognition was a major selling point of the iPhone and following Apple products, with the introduction of Siri. This technology gives users an alternative way to communicate with computers when typing is less desirable. However, in a loud environment, the audition modality is not quite effective. This exemplifies how certain modalities have varying strengths depending on the situation. Other complex modalities such as computer vision in the form of Microsoft's Kinect or other similar technologies can make sophisticated tasks easier to communicate to a computer especially in the form of three dimensional movement.

    This talk is based on a physical robot (a personalized humanoid built in MUST Research), equipped with various types of input devices and sensors to allow them to receive information from humans, which are interchangeable and a standardized method of communication with the computer, affording practical adjustments to the user, providing a richer interaction depending on the context, and implementing robust system with features like; keyboard; pointing device; touchscreen; computer vision; speech recognition; motion, orientation etc.

    Cognitive computing makes a new class of problems computable. 

To respond to the fluid nature of users understanding of their problems, the cognitive computing system offers a synthesis not just of information sources but of influences, contexts, and insights. 

These systems differ from current computing applications in that they move beyond tabulating and calculating based on pre-configured rules and programs. 

They can infer and even reason based on broad objectives. In this sense, cognitive computing is a new type of computing with the goal of more accurate models of how the human brain or mind senses, reasons, and responds to stimulus. 

It is a field of study which studies how to create computers and computer software that are capable of intelligent behavior. This field is interdisciplinary, in which a number of sciences and professions converge, including computer science, electronics, mathematics, statistics, psychology, linguistics, philosophy, neuroscience and biology.

    Computer–Human Modalities

    Computers utilize a wide range of technologies to communicate and send information to humans:

    • Vision – computer graphics typically through a screen
    • Audition – various audio outputs
    • Tactition – vibrations or other movement
    • Gustation (taste)
    • Olfaction (smell)
    • Thermoception (heat)
    • Nociception (pain)
    • Equilibrioception (balance)

    Human–computer Modalities

    Computers can be equipped with various types of input devices and sensors to allow them to receive information from humans. Common input devices are often interchangeable if they have a standardized method of communication with the computer and afford practical adjustments to the user. Certain modalities can provide a richer interaction depending on the context, and having options for implementation allows for more robust systems.

    • Keyboard
    • Pointing device
    • Touchscreen
    • Computer vision
    • Speech recognition
    • Motion
    • Orientation

    Project Features

    Adaptive: They MUST learn as information changes, and as goals and requirements evolve. They MUST resolve ambiguity and tolerate unpredictability. They MUST be engineered to feed on dynamic data in real time.

    Interactive: They MUST interact easily with users so that those users can define their needs comfortably. They MUST interact with other processors, devices, services, as well as with people.

    Iterative and Stateful: They MUST aid in defining a problem by asking questions or finding additional source input if a problem statement is ambiguous or incomplete. They MUST remember previous interactions in a process and return information that is suitable for the specific application at that point in time.

    Contextual: They MUST understand, identify, and extract contextual elements such as meaning, syntax, time, location, appropriate domain, regulation, user profile, process, task and goal. They may draw on multiple sources of information, including both structured and unstructured digital information, as well as sensory inputs (visual, gestural, auditory, or sensor-provided).

  • Liked Dr. Saptarsi Goswami

    Dr. Saptarsi Goswami - Meta features and clustering based approaches for feature selection

    45 Mins

    Feature selection is one of the most important processes for pattern recognition, machine learning and data mining problems. A successful feature selection method facilitates improvement of learning model performance and interpretability as well as reduces computational cost of the classifier by dimensionality reduction of the data. Feature selection refers to the retention of discriminatory features while discarding the redundant and irrelevant features. In this process, a subset of D features are selected from a set of N features (D<N). There is another way of achieving dimensionality reduction by projecting higher dimensional data to lower dimension, normally referred to feature extraction. This thesis refers to the former one i.e. feature subset selection. Optimal feature subset selection method comprises of developing 1) an evaluation function for measuring the goodness of a feature or a feature subset and 2) a search algorithm to find out the best subset of features from all possible subsets of the whole feature set. Based on the nature of the objective function used in the search algorithms, feature subset selection algorithms are broadly classified into filter approach and wrapper approach. Classifier dependent wrapper approaches use classifier accuracy as the objective function while filter approaches use any evaluation function representing the intrinsic characteristics of the data set and the resulting feature subset works equally well for any classifier. This work focusses on filter based feature subset selection approach. In this work, initially a study has been done with currently available search based filter type feature selection algorithms for supervised as well as unsupervised classification with both the single objective and multi-objective evaluation functions. Some improvements over the current algorithms have been proposed and their efficiency has been examined by simulation experiments with bench mark data sets. In the second step, an inexpensive feature evaluation measure based on feature relevance to be used with a filter type feature selection for unsupervised classification has been proposed. It has been noticed during literature study that the concept of feature relevance in case of unsupervised classification is difficult to form and current methods are complex and time consuming. The proposed measure which considers individual variability as well as overall variability of the dataset,is found to be effective compared to the current methods by simulation experiments with bench mark data sets. Thirdly, it seems that the most of the current feature selection algorithms are based on search strategies to find out the best feature subset from the available feature set. For a large number of features, exhaustive search is computationally prohibitive which leads to combinatorial optimization problem and some sort of heuristic is used for the solution. With the increase of the number of features, the computational time for optimal feature subset selection increases.An alternative solution to this problem is to use clustering of the features to find out the best feature subset which is not yet explored sufficiently. In this work, an efficient clustering based feature selection algorithm has been proposed and simulation experiments have been done with bench mark data sets. The main contributions of the proposed algorithm are introduction of a novel method to determine the optimal number of clusters, a way of interpretation of the importance of the feature clusters and a method of selection of the final subset of features from the feature clusters. Finally, it is found that though lots of feature selection algorithms are available, it is very difficult to decide which algorithm is suitable for a particular real world application. Here a study has been done to establish the relation between the feature selection algorithm and the characteristics of the data set. A technique has been proposed to define a data set according to its intrinsic characteristics represented by some meta-features. Finally a feature selection strategy is recommended based on the characteristics of the data set and has been implemented with bench mark data sets to judge its effectiveness.

  • Lakshya
    Applied Researcher-2
    schedule 1 week ago
    Sold Out!
    45 Mins

    Deep learning has significantly improved state-of-the-art performance for natural language processing (NLP) tasks, but each one is typically studied in isolation. The Natural Language Decathlon (decaNLP) is a new benchmark for studying general NLP models that can perform a variety of complex, natural language tasks. By requiring a single system to perform ten disparate natural language tasks, decaNLP offers a unique setting for multitask, transfer, and continual learning. decaNLP is maintained by salesforce and is publicly available on github in order to use for tasks like Question Answering, Machine Translation, Summarization, Sentiment Analysis etc.

  • Liked Siboli mukherjee

    Siboli mukherjee - Real time Anomaly Detection in Network KPI using Time Series

    20 Mins
    Experience Report


    How to accurately detect Key Performance Indicator (KPI) anomalies is a critical issue in cellular network management. In this talk I shall introduce CNR(Cellular Network Regression) a unified performance anomaly detection framework for KPI time-series data. CNR realizes simple statistical modelling and machine-learning-based regression for anomaly detection; in particular, it specifically takes into account seasonality and trend components as well as supports automated prediction model retraining based on prior detection results. I demonstrate here how CNR detects two types of anomalies of practical interest, namely sudden drops and correlation changes, based on a large-scale real-world KPI dataset collected from a metropolitan LTE network. I explore various prediction algorithms and feature selection strategies, and provide insights into how regression analysis can make automated and accurate KPI anomaly detection viable.

    Index Terms—anomaly detection, NPAR (Network Performance Analysis)


    The continuing advances of cellular network technologies make high-speed mobile Internet access a norm. However, cellular networks are large and complex by nature, and hence production cellular networks often suffer from performance degradations or failures due to various reasons, such as back- ground interference, power outages, malfunctions of network elements, and cable disconnections. It is thus critical for network administrators to detect and respond to performance anomalies of cellular networks in real time, so as to maintain network dependability and improve subscriber service quality. To pinpoint performance issues in cellular networks, a common practice adopted by network administrators is to monitor a diverse set of Key Performance Indicators (KPIs), which provide time-series data measurements that quantify specific performance aspects of network elements and resource usage. The main task of network administrators is to identify any KPI anomalies, which refer to unexpected patterns that occur at a single time instant or over a prolonged time period.

    Today’s network diagnosis still mostly relies on domain experts to manually configure anomaly detection rules such a practice is error-prone, labour intensive, and inflexible. Recent studies propose to use (supervised) machine learning for anomaly detection in cellular networks . ellular networks, a common practice adopted by network administrators is to monitor a diverse set of Key Performance Indicators (KPIs), which provide time-series data measurements that quantify specific performance aspects of network elements and resource usage. The main task of network administrators is to identify any KPI anomalies, which refer to unexpected patterns that occur at a single time instant or over a prolonged time period.

    Today’s network diagnosis still mostly relies on domain experts to manually configure anomaly detection rules such a practice is error-prone, labour intensive, and inflexible. Recent studies propose to use (supervised) machine learning for anomaly detection in cellular networks .

  • Liked Siboli mukherjee

    Siboli mukherjee - AI in Telecommunication -An Obstacle or Opportunity

    45 Mins


    “Alexa, launch Netflix!”

    No longer limited to providing basic phone and Internet service, the telecom industry is at the epicentre of technological growth, led by its mobile and broadband services in the Internet of Things (IoT) era.This growth is expected to continue,The driver for this growth? Artificial intelligence (AI).

    Artificial Intelligent applications are revolutionizing the way telecoms operate, optimize and provide service to their customers

    Today’s communications service providers (CSPs) face increasing customer demands for higher quality services and better customer experiences (CX). Telecoms are addressing these opportunities by leveraging the vast amounts of data collected over the years from their massive customer base. This data is culled from devices, networks, mobile applications, geolocations, detailed customer profiles, services usage and billing data.

    Telecoms are harnessing the power of AI to process and analyse these huge volumes of Big Data in order to extract actionable insights to provide better customer experiences, improve operations, and increase revenue through new products and services.

    With Gartner forecasting that 20.4 billion connected devices will be in use worldwide by 2020, more and more CSPs are jumping on the bandwagon, recognizing the value of artificial intelligence applications in the telecommunications industry.

    Forward-thinking CSPs have focused their efforts on four main areas where AI has already made significant inroads in delivering tangible business results: Network optimization, preventive maintenance, Virtual Assistants, and robotic process automation (RPA)

    Network optimisation

    AI is essential for helping CSPs build self-optimizing networks (SONs), where operators have the ability to automatically optimize network quality based on traffic information by region and time zone. Artificial intelligence applications in the telecommunications industry use advanced algorithms to look for patterns within the data, enabling telecoms to both detect and predict network anomalies, and allowing operators to proactively fix problems before customers are negatively impacted.

    Some popular AI solutions for telecoms are ZeroStack’s ZBrain Cloud Management, which analyses private cloud telemetry storage and use for improved capacity planning, upgrades and general management; Aria Networks, an AI-based network optimization solution that counts a growing number of Tier-1 telecom companies as customers, and Sedona Systems’ NetFusion, which optimizes the routing of traffic and speed delivery of 5G-enabled services like AR/VR. Nokia launched its own machine learning-based AVA platform, a cloud-based network management solution to better manage capacity planning, and to predict service degradations on cell sites up to seven days in advance.

    Predictive maintenance

    AI-driven predictive analytics are helping telecoms provide better services by utilizing data, sophisticated algorithms and machine learning techniques to predict future results based on historical data. This means telecoms can use data-driven insights to can monitor the state of equipment, predict failure based on patterns, and proactively fix problems with communications hardware, such as cell towers, power lines, data centre servers, and even set-top boxes in customers’ homes.

    In the short-term, network automation and intelligence will enable better root cause analysis and prediction of issues. Long term, these technologies will underpin more strategic goals, such as creating new customer experiences and dealing efficiently with business demands. An innovative solution by AT&Tis using AI to support its maintenance procedures: the company is testing a drone to expand its LTE network coverage and to utilize the analysis of video data captured by drones for tech support and infrastructure maintenance of its cell towers.Preventive maintenance is not only effective on the network side, but on the customer’s side as well.Dutch telecom KPN analyses the notes generated by its call centre agents, and uses the insights generated to make changes to the interactive voice response (IVR) system.

    Virtual Assistants

    Conversational AI platforms — known as virtual assistants — have learned to automate and scale one-on-one conversations so efficiently that they are projected to cut business expenses by as much as $8 billion in the next five years. Telecoms have turned to virtual assistants to help contend with the massive number of support requests for installation, set up, troubleshooting and maintenance, which often overwhelm customer support centre Using AI, telecoms can implement self-service capabilities that instruct customers how to install and operate their own devices.

    Vodafone introduced its new chatbot — TOBi to handle a range of customer service-type questions. The chatbotscales responses to simple customer queries, thereby delivering the speed that customers demand. Nokia’s virtual assistant MIKA suggests solutions for network issues, leading to a 20% to 40% improvement in first-time resolution.

    Robotic process automation (RPA)

    CSPs all have vast numbers of customers and an endless volume of daily transactions, each susceptible to human error. Robotic Process Automation (RPA) is a form of business process automation technology based on AI. RPA can bring greater efficiency to telecommunications functions by allowing telecoms to more easily manage their back office operations and the large volumes of repetitive and rules-based processes. By streamlining execution of once complex, labor-intensive and time-consuming processes such as billing, data entry, workforce management and order fulfillment, RPA frees CSP staff for higher value-add work.

    According to a survey by Deloitte, 40% of Telecom, Media and Tech executives say they have garnered “substantial” benefits from cognitive technologies, with 25% having invested $10 million or more. More than three-quarters expect cognitive computing to “substantially transform” their companies within the next three years.


    Artificial intelligence applications in the telecommunications industry is increasingly helping CSPs manage, optimize and maintain not only their infrastructure, but their customer support operations as well. Network optimization, predictive maintenance, virtual assistants and RPA are examples of use cases where AI has impacted the telecom industry, delivering an enhanced CX and added value for the enterprise overall.

  • Liked Debjyoti Paul

    Debjyoti Paul - Transfer Learning in Unsupervised text processing

    45 Mins

    Today we are facing enormous amount of unstructured textual data. Given a text processing problem, how to start? What models to build language model with? Can models trained in similar domains be exploited. These are some trailing questions.

    1. When and how to use Transfer Learning- new vocabulary? 2. Challenges in Text processing and Transfer Learning 3. Effectively method selection for transfer learning 4. Applications 5. How to validate your model?

    Presentation on Aspect detection in unsupervised domain using Transfer Learning from structure prediction.

  • Liked Kaushik Dey

    Kaushik Dey - Algorithms at Edge leveraging decentralized learning

    45 Mins

    The problem of network behavior prediction has been an ongoing study by researchers for quite a while now. Network behavior typically exhibits a complex sequential pattern and is often difficult to predict. Nowadays there are several techniques to predict the degradation in Network KPIs like throughput, latency etc., using various machine learning techniques like Deep Neural Networks, where the initial layers have learnt to map the raw features like performance counter measurements, weather, system configuration details etc into a feature space where classification by the final layers can be performed.

    Given the initial number of counters( which constitutes the dimensions) is substantial (more than 2000 in number) the problem requires huge amount of data to train the Deep Neural Networks. Often this needs resources and time and more importantly this requires provisioning of huge amount of data for every trial. Given each node generates huge amount of data ( data on every 2000 counters generated at 15 minutes interval for each of 6 cells in an eNodeB) and the data needs to be transported across several hundred of eNodeBs to one central data center, it requires a very fat data pipe and consequently huge investment to enable a predictive fault prediction apparatus across the network.

    The alernative is to have a compute infrastructure at the node and take the intelligence at the edge. However the challenge is given the huge amount of data generated in a single node having a compute at each node was proving to be expensive. Nowadays this compute requirement at node could be reduced through use of transfer learning. However the other challenge is on sharing the intelligence and developing a system which is collectively intelligent across nodes.

    Network topology, climate features and user patterns vary across regions and service providers and hence an unique model is often necesarry to serve the node. However in order to deal with unseen patterns intelligence from other nodes can be useful which leads us to building an global model which again leads to the challenge of fat data pipeline requirement which makes it commercially less attractive.

    In order to get around this challenge, an combination of federated learning is used in combination with transfer learning.

    This presentation details such deep learning architectures which combines federated learning with transfer learning to enable construction and updation of Global models which imbibes intelligence from nodes but can be constructed by a consensus mechanism whereby weights and changes to weights of local models are shared to global. Also the local models are periodically updated once global model update iteration is complete. Further updation of local models is only done in final layers and initial layers are freezed. This reduces the compute requirement at node also...

    The above principles are being implemented as First of a kind implementation and has prooved to be a success across multiple customers in delivering a compelling ML enabled fault prediction and self-healing mechanism but keeping the investments in infrastructure lower than would have been required in traditional Deep Learning architectures

    This talk will specifically detail the leverage of above principles of federated and transfer learning on LSTMs..