Breaking the language barrier: how do we quickly add multilanguage support in our AI application?

location_city Bengaluru schedule Aug 8th 02:45 - 03:05 PM place Grand Ball Room 2 people 81 Interested

With the need to cater to a global audience, there is a growing demand for applications to support speech identification/translation/transliteration from one language to another. This session aims at introducing the audience to the topic, learn the inner working of the AI/ML models and eventually how to quickly use some of the readily available APIs to identify, translate or even transliterate speech/text within their application.

 
 

Outline/Structure of the Talk

Over the last few years, multitude of customers routinely come with various machine translation use cases. Here are a few examples:

  • Identify which language is being spoken/typed by the user.
  • How do I translate one language to another for the user?
  • Is there a way to transliterate text from one language to another?

Hard problems such as these and many more like them are now readily solved using advanced API’s that are readily available without having to reinvent the wheel.

This talk starts off with a brief introduction to the topic of Machine Translation, then it moves on to introduce some of the typical customer cases and finally ends with how to embed such functionality in your application. The talk will end with a Python based solution demo and introduce the audience to some resources and tools that could help them as they continue to explore the domain.

Learning Outcome

The goal of this talk is to provide a robust means for the audience to discover opportunities and to learn how to quickly apply Machine Translation in their applications.

We will provide a framework for attendees to accomplish the following goals:

  • Improve their understanding of the domain of Machine Translation and learn the inner working of the underlying AI/ML models.
  • Qualify new opportunities and assess their fit for Machine Translation.
  • Smoothly embed Machine Translation using APIs in their applications.

Target Audience

Data Scientist, Data Engineer, Program manager

Prerequisites for Attendees

Some familiarity with Python, ML, AI concept.

Anyone aspiring to learn about machine translation and how to identify languages being spoken/typed or how to perform translation/transliteration of text.

People should attend this session if they are interested in any of the following 3 questions:

  • What are some of the AI powered machine translation methods?
  • What is the underlying AI/ML techniques used for these models?
  • What are some of the common applications of machine translation?
  • How can these AI methods be quickly embedded in my application?
schedule Submitted 1 year ago

Public Feedback


    • Dr. Saptarsi Goswami
      keyboard_arrow_down

      Dr. Saptarsi Goswami - Mastering feature selection: basics for developing your own algorithm

      45 Mins
      Tutorial
      Beginner

      Feature selection is one of the most important processes for pattern recognition, machine learning and data mining problems. A successful feature selection method facilitates improvement of learning model performance and interpretability as well as reduces computational cost of the classifier by dimensionality reduction of the data. Feature selection is computationally expensive and becomes intractable even for few 100 features. This is a relevant problem because text, image and next generation sequence data all are inherently high dimensional. In this talk, I will discuss about few algorithms we have developed in last 5/6 years. Firstly, we will set the context of feature selection ,with some open issues , followed by definition and taxonomy. Which will take about 20 odd minutes. Then in next 20 minutes we will discuss couple of research efforts where we have improved feature selection for textual data and proposed a graph based mechanism to view the feature interaction. After the talk, participants will be appreciate the need of feature selection, the basic principles of feature selection algorithm and finally how they can start developing their own models

    • Dr. Rahee Walambe
      keyboard_arrow_down

      Dr. Rahee Walambe / Vishal Gokhale - Processing Sequential Data using RNNs

      480 Mins
      Workshop
      Beginner

      Data that forms the basis of many of our daily activities like speech, text, videos has sequential/temporal dependencies. Traditional deep learning models, being inadequate to model this connectivity needed to be made recurrent before they brought technologies such as voice assistants (Alexa, Siri) or video based speech translation (Google Translate) to a practically usable form by reducing the Word Error Rate (WER) significantly. RNNs solve this problem by adding internal memory. The capacities of traditional neural networks are bolstered with this addition and the results outperform the conventional ML techniques wherever the temporal dynamics are more important.
      In this full-day immersive workshop, participants will develop an intuition for sequence models through hands-on learning along with the mathematical premise of RNNs.

    • Varun Sharma
      keyboard_arrow_down

      Varun Sharma - Blockchain with Machine Learning - The ultimate industry disruptor

      Varun Sharma
      Varun Sharma
      Sr. Manager
      SAP Labs
      schedule 1 year ago
      Sold Out!
      45 Mins
      Talk
      Beginner

      The fusion of blockchain and machine learning is an ultimate game changer. Machine learning relies on high volume of data to build models for accurate prediction. A lot of the challenges incurred in getting this data lies in collecting, organizing and auditing the data for accuracy. This is an area that can significantly be improved by using blockchain technology. By using smart contracts, data can be directly and reliably transferred straight from its place of origin. Smart contracts could, however, improve the whole process significantly by using digital signatures.

      Blockchain is a good candidate to store sensitive information that should not be modified in any way. Machine learning works on the principle of “Garbage In, Garbage Out,” which means that if the data that was used to build a prediction model was corrupted in any way, the resultant model would not be of much use either. Combining both these technologies creates an industry disruptor which leverages the power of both Blockchain and Machine learning.

    • Varun Sharma
      keyboard_arrow_down

      Varun Sharma - Automated Machine Leaning - Transforming the AI Landscape

      Varun Sharma
      Varun Sharma
      Sr. Manager
      SAP Labs
      schedule 1 year ago
      Sold Out!
      45 Mins
      Talk
      Intermediate

      Automated Machine Learning (AutoML) provides methods and processes to automate the end-to-end machine learning workflow towards solving real-world business problems.

      In traditional machine learning world, data scientist used to spend a considerable amount of time in Data wrangling, model selection and tuning, now with the advances of AutoML it provides methods and processes to make Machine Learning available for non-Machine Learning experts, to improve efficiency of Machine Learning and to accelerate research on Machine Learning.

      Automating the end-to-end process of applying machine learning offers the advantages of producing simpler solutions, faster creation of those solutions, and models that often outperform models that were designed by hand.

    • Parul pandey
      keyboard_arrow_down

      Parul pandey - Jupyter Ascending : The journey from Jupyter Notebook to Jupyter Lab

      Parul pandey
      Parul pandey
      Data Science Communicator
      -
      schedule 1 year ago
      Sold Out!
      45 Mins
      Demonstration
      Beginner

      For many of the researchers and data scientists, Jupyter Notebooks are the de-facto platform when it comes to quick prototyping and exploratory analysis. Right from Paul Romer- the Ex-World bank chief Economist and also the co-winner 2018 Nobel prize in Economics to Netflix, Jupyter Notebooks are used almost everywhere. The browser-based computing environment, coupled with a reproducible document format has made them the choice of tool for millions of data scientists and researchers around the globe. But have we fully exploited the benefits of Jupyter Notebooks and do we know all about the best practises of using it? if not, then this talk is just for you.

      Through this talk/demo, I'll like to discuss three main points:

      1. Best Practises for Jupyter Notebooks since a lot of Jupyter functionalities sometimes lies under the hood and is not adequately explored. We will try and explore Jupyter Notebooks’ features which can enhance our productivity while working with them.
      2. In this part, we get acquainted with Jupyter Lab, the next-generation UI developed by the Project Jupyter team, and its emerging ecosystem of extensions. JupyterLab differs from Jupyter Notebook in the fact that it provides a set of core building blocks for interactive computing (e.g. notebook, terminal, file browser, console) and well-designed interfaces for them that allow users to combine them in novel ways. The new interface enables users to do new things in their interactive computing environment, like tiled layouts for their activities, dragging cells between notebooks, and executing markdown code blocks in a console and many more cool things.
      3. Every tool/features come with their set of pros and cons and so does Jupyter Notebooks/Lab and it is equally important to discuss the pain areas along with the good ones.