Breaking the language barrier: how do we quickly add multilanguage support in our AI application?

schedule Aug 8th 02:45 - 03:05 PM place Grand Ball Room 2 people 81 Interested

With the need to cater to a global audience, there is a growing demand for applications to support speech identification/translation/transliteration from one language to another. This session aims at introducing the audience to the topic, learn the inner working of the AI/ML models and eventually how to quickly use some of the readily available APIs to identify, translate or even transliterate speech/text within their application.

 
3 favorite thumb_down thumb_up 13 comments visibility_off  Remove from Watchlist visibility  Add to Watchlist
 

Outline/Structure of the Talk

Over the last few years, multitude of customers routinely come with various machine translation use cases. Here are a few examples:

  • Identify which language is being spoken/typed by the user.
  • How do I translate one language to another for the user?
  • Is there a way to transliterate text from one language to another?

Hard problems such as these and many more like them are now readily solved using advanced API’s that are readily available without having to reinvent the wheel.

This talk starts off with a brief introduction to the topic of Machine Translation, then it moves on to introduce some of the typical customer cases and finally ends with how to embed such functionality in your application. The talk will end with a Python based solution demo and introduce the audience to some resources and tools that could help them as they continue to explore the domain.

Learning Outcome

The goal of this talk is to provide a robust means for the audience to discover opportunities and to learn how to quickly apply Machine Translation in their applications.

We will provide a framework for attendees to accomplish the following goals:

  • Improve their understanding of the domain of Machine Translation and learn the inner working of the underlying AI/ML models.
  • Qualify new opportunities and assess their fit for Machine Translation.
  • Smoothly embed Machine Translation using APIs in their applications.

Target Audience

Data Scientist, Data Engineer, Program manager

Prerequisites for Attendees

Some familiarity with Python, ML, AI concept.

Anyone aspiring to learn about machine translation and how to identify languages being spoken/typed or how to perform translation/transliteration of text.

People should attend this session if they are interested in any of the following 3 questions:

  • What are some of the AI powered machine translation methods?
  • What is the underlying AI/ML techniques used for these models?
  • What are some of the common applications of machine translation?
  • How can these AI methods be quickly embedded in my application?
schedule Submitted 7 months ago

Public Feedback

comment Suggest improvements to the Speaker
  • Naresh Jain
    By Naresh Jain  ~  3 months ago
    reply Reply

    Hi Jaya,

    Thanks for your response to all the comments below.

    • You did reply to Ashay saying you would include a couple of slides on the core ML/DL models to ensure the audience understands how the underlying technology works. Adding to Vikas' comment, going deeper into the technical subject focusing on Data Scientists would be really helpful. Can we please request you to focus more on the core ML/DL topic and less on the product/API demo?
    • Also, I think this is a very interesting topic and instead of 20 mins, if we can extend it to 45 mins it would be great.
    • JAYA SUSAN MATHEW
      By JAYA SUSAN MATHEW  ~  2 months ago
      reply Reply

      Yes Naresh, this talk can be a 45 min technical deep dive. I had replied to an email from the organizers and did not check this portal, hence the delay in response. 

      • Naresh Jain
        By Naresh Jain  ~  2 months ago
        reply Reply

        Thank you. Can you please update the proposal?

        • JAYA SUSAN MATHEW
          By JAYA SUSAN MATHEW  ~  1 month ago
          reply Reply

          Sure, I just updated it to include the technical pieces as well. 

          • Naresh Jain
            By Naresh Jain  ~  1 month ago
            reply Reply

            Will 20 mins be sufficient?

            • JAYA SUSAN MATHEW
              By JAYA SUSAN MATHEW  ~  1 month ago
              reply Reply

              Yes Naresh, I will be able to do the talk in the allotted 20 min. 

  • Dr. Vikas Agrawal
    By Dr. Vikas Agrawal  ~  4 months ago
    reply Reply

    Dear Jaya: Are you planning to share details of known issues/challenges with machine translation and technical deep dive on one or more of those? Given your experience and expertise we would love to see that. Your talk is currently aimed at Data Scientists, Data Engineers and Program Managers. Can we make it significantly more technical and deep dive please?

    The usage of Microsoft Azure APIs is available online showing your good work. 

    Warm Regards,

    Vikas

    • JAYA SUSAN MATHEW
      By JAYA SUSAN MATHEW  ~  2 months ago
      reply Reply

      Yes Vikas, I believe I had sent an email to the organizers that I was planning to share technical details about the model as well as issues/challenges. The content can be tapered to more technical/deep dive session. 

    • JAYA SUSAN MATHEW
      By JAYA SUSAN MATHEW  ~  3 months ago
      reply Reply

      Thanks Vikas for your feedback. Since I have worked with some customers in this domain, I can outline some of the issues/challenges in this area as well. 

  • Ashay Tamhane
    By Ashay Tamhane  ~  4 months ago
    reply Reply

    Hi Jaya, thanks for the proposal. Would like to understand if you will be spending time on the core ML/DL models that drive the APIs, or will this be more of a tutorial on the usage of APIs?

    • JAYA SUSAN MATHEW
      By JAYA SUSAN MATHEW  ~  3 months ago
      reply Reply

      Thanks Ashay for the feedback. There will be a couple of slides on the core ML/DL models to ensure the audience understands how the underlying technology works. 

  • Anoop Kulkarni
    By Anoop Kulkarni  ~  5 months ago
    reply Reply

    Jaya, thanks for your proposal. The scope of talk is explained very well and exhibits clarity of thought. Was just curious if you could explain briefly what the python demo would be like? E.g which language to which language , is it text to speech or speech to text or speech to speech but from one language to another?

    Also, how long you intend to spend on the demo?

    Thanks

    ~anoop

    • JAYA SUSAN MATHEW
      By JAYA SUSAN MATHEW  ~  3 months ago
      reply Reply

      Thanks Anoop for your feedback. I was planning on doing a short 5 minute demo.


  • Liked Dr. Saptarsi Goswami
    keyboard_arrow_down

    Dr. Saptarsi Goswami - Mastering feature selection: basics for developing your own algorithm

    45 Mins
    Tutorial
    Beginner

    Feature selection is one of the most important processes for pattern recognition, machine learning and data mining problems. A successful feature selection method facilitates improvement of learning model performance and interpretability as well as reduces computational cost of the classifier by dimensionality reduction of the data. Feature selection is computationally expensive and becomes intractable even for few 100 features. This is a relevant problem because text, image and next generation sequence data all are inherently high dimensional. In this talk, I will discuss about few algorithms we have developed in last 5/6 years. Firstly, we will set the context of feature selection ,with some open issues , followed by definition and taxonomy. Which will take about 20 odd minutes. Then in next 20 minutes we will discuss couple of research efforts where we have improved feature selection for textual data and proposed a graph based mechanism to view the feature interaction. After the talk, participants will be appreciate the need of feature selection, the basic principles of feature selection algorithm and finally how they can start developing their own models

  • Liked Rahee Walambe
    keyboard_arrow_down

    Rahee Walambe / Vishal Gokhale - Processing Sequential Data using RNNs

    Rahee Walambe
    Rahee Walambe
    Research and Teaching Faculty
    Symbiosis Institute of Technology
    Vishal Gokhale
    Vishal Gokhale
    Sr. Consultant
    XNSIO
    schedule 4 months ago
    Sold Out!
    480 Mins
    Workshop
    Beginner

    Data that forms the basis of many of our daily activities like speech, text, videos has sequential/temporal dependencies. Traditional deep learning models, being inadequate to model this connectivity needed to be made recurrent before they brought technologies such as voice assistants (Alexa, Siri) or video based speech translation (Google Translate) to a practically usable form by reducing the Word Error Rate (WER) significantly. RNNs solve this problem by adding internal memory. The capacities of traditional neural networks are bolstered with this addition and the results outperform the conventional ML techniques wherever the temporal dynamics are more important.
    In this full-day immersive workshop, participants will develop an intuition for sequence models through hands-on learning along with the mathematical premise of RNNs.

  • Liked Varun Sharma
    keyboard_arrow_down

    Varun Sharma - Blockchain with Machine Learning - The ultimate industry disruptor

    Varun Sharma
    Varun Sharma
    Sr. Manager
    SAP Labs
    schedule 4 months ago
    Sold Out!
    45 Mins
    Talk
    Beginner

    The fusion of blockchain and machine learning is an ultimate game changer. Machine learning relies on high volume of data to build models for accurate prediction. A lot of the challenges incurred in getting this data lies in collecting, organizing and auditing the data for accuracy. This is an area that can significantly be improved by using blockchain technology. By using smart contracts, data can be directly and reliably transferred straight from its place of origin. Smart contracts could, however, improve the whole process significantly by using digital signatures.

    Blockchain is a good candidate to store sensitive information that should not be modified in any way. Machine learning works on the principle of “Garbage In, Garbage Out,” which means that if the data that was used to build a prediction model was corrupted in any way, the resultant model would not be of much use either. Combining both these technologies creates an industry disruptor which leverages the power of both Blockchain and Machine learning.

  • Liked Varun Sharma
    keyboard_arrow_down

    Varun Sharma - Automated Machine Leaning - Transforming the AI Landscape

    Varun Sharma
    Varun Sharma
    Sr. Manager
    SAP Labs
    schedule 4 months ago
    Sold Out!
    45 Mins
    Talk
    Intermediate

    Automated Machine Learning (AutoML) provides methods and processes to automate the end-to-end machine learning workflow towards solving real-world business problems.

    In traditional machine learning world, data scientist used to spend a considerable amount of time in Data wrangling, model selection and tuning, now with the advances of AutoML it provides methods and processes to make Machine Learning available for non-Machine Learning experts, to improve efficiency of Machine Learning and to accelerate research on Machine Learning.

    Automating the end-to-end process of applying machine learning offers the advantages of producing simpler solutions, faster creation of those solutions, and models that often outperform models that were designed by hand.

  • Liked Parul pandey
    keyboard_arrow_down

    Parul pandey - Jupyter Ascending : The journey from Jupyter Notebook to Jupyter Lab

    Parul pandey
    Parul pandey
    Data Science Communicator
    -
    schedule 4 months ago
    Sold Out!
    45 Mins
    Demonstration
    Beginner

    For many of the researchers and data scientists, Jupyter Notebooks are the de-facto platform when it comes to quick prototyping and exploratory analysis. Right from Paul Romer- the Ex-World bank chief Economist and also the co-winner 2018 Nobel prize in Economics to Netflix, Jupyter Notebooks are used almost everywhere. The browser-based computing environment, coupled with a reproducible document format has made them the choice of tool for millions of data scientists and researchers around the globe. But have we fully exploited the benefits of Jupyter Notebooks and do we know all about the best practises of using it? if not, then this talk is just for you.

    Through this talk/demo, I'll like to discuss three main points:

    1. Best Practises for Jupyter Notebooks since a lot of Jupyter functionalities sometimes lies under the hood and is not adequately explored. We will try and explore Jupyter Notebooks’ features which can enhance our productivity while working with them.
    2. In this part, we get acquainted with Jupyter Lab, the next-generation UI developed by the Project Jupyter team, and its emerging ecosystem of extensions. JupyterLab differs from Jupyter Notebook in the fact that it provides a set of core building blocks for interactive computing (e.g. notebook, terminal, file browser, console) and well-designed interfaces for them that allow users to combine them in novel ways. The new interface enables users to do new things in their interactive computing environment, like tiled layouts for their activities, dragging cells between notebooks, and executing markdown code blocks in a console and many more cool things.
    3. Every tool/features come with their set of pros and cons and so does Jupyter Notebooks/Lab and it is equally important to discuss the pain areas along with the good ones.