"Today we’re excited to share DeepMind’s first significant milestone in demonstrating how artificial intelligence research can drive and accelerate new scientific discoveries. With a strongly interdisciplinary approach to our work, DeepMind has brought together experts from the fields of structural biology, physics, and machine learning to apply cutting-edge techniques to predict the 3D structure of a protein based solely on its genetic sequence." source: https://deepmind.com/blog/alphafold/

Over the past five decades, scientists have been able to determine shapes of proteins in labs using experimental techniques like cryo-electron microscopy, nuclear magnetic resonance or X-ray crystallography, but each method depends on a lot of trial and error, which can take years and cost tens of thousands of dollars per structure. This is why biologists are turning to AI methods as an alternative to this long and laborious process for difficult proteins.

Recently released by Deepmind, Alpha fold, beat top pharmaceutical companies with 100K+ employees like Pfizer, Novartis, etc. at predicting protein structures in the CASP13 challenge. It outperformed all the other competitors and emerged first with a huge difference of correctly predicting 25 proteins correctly whereas the second place winner only predicted 9 of them correctly and that too with only 29K of the 129K present data about different proteins

This research is the greatest breakthrough in this field which will be able to predict how proteins fold for the formation of different types of proteins for different functions. This is important because this could lead to a better understanding and possibly a cure for diseases like Alzheimer's, mad cow's disease etc. because these diseases are believed to be caused due to malfunction in the folding of the proteins in the body.

The architecture for the network was simple, on a high level it constituted of residual convolutional neural network and gradient descent to optimize full protein features in the end.

The audience from this talk will be able to learn about how to reproduce the architecture of the Alpha Fold and also some basics about how different proteins strands affect the body and function of the proteins. This talk will be mostly on the technical side of the Alpha Fold.


Outline/Structure of the Talk

On a very high level the presentation will be as follows:

  • Introduction to some concepts about different proteins (a very basic introduction)
  • What is the problem and what all this can solve.

  • The Problem
    • Different diseases caused by the problems

  • Previous implementation to solve these problems
    • Simple Deep Neural Network
      • An overview
      • The limitations
    • DeepSF
      • An overview
      • The limitations

  • the architecture of Alpha Fold
    • Some basic knowledge about the different algorithms used in its architecture
    • The complete architecture
    • Deep analysis on the different components

  • A basic implementation and Demo

Learning Outcome

The audience from this talk will be able to learn about how to reproduce the architecture of the Alpha Fold and also some basics about how different proteins strands affect the body and function of the proteins. This talk will be mostly on the technical side of the Alpha Fold.

Target Audience

Data Scientists, ML Enthusiasts and anyone who wants to learn about the Alpha Fold.

Prerequisites for Attendees

Some basic knowledge of Neural Networks and also some Knowledge of Machine Learning.

schedule Submitted 1 year ago

Public Feedback

    • Liked Dr. Vikas Agrawal

      Dr. Vikas Agrawal - Non-Stationary Time Series: Finding Relationships Between Changing Processes for Enterprise Prescriptive Systems

      45 Mins

      It is too tedious to keep on asking questions, seek explanations or set thresholds for trends or anomalies. Why not find problems before they happen, find explanations for the glitches and suggest shortest paths to fixing them? Businesses are always changing along with their competitive environment and processes. No static model can handle that. Using dynamic models that find time-delayed interactions between multiple time series, we need to make proactive forecasts of anomalous trends of risks and opportunities in operations, sales, revenue and personnel, based on multiple factors influencing each other over time. We need to know how to set what is “normal” and determine when the business processes from six months ago do not apply any more, or only applies to 35% of the cases today, while explaining the causes of risk and sources of opportunity, their relative directions and magnitude, in the context of the decision-making and transactional applications, using state-of-the-art techniques.

      Real world processes and businesses keeps changing, with one moving part changing another over time. Can we capture these changing relationships? Can we use multiple variables to find risks on key interesting ones? We will take a fun journey culminating in the most recent developments in the field. What methods work well and which break? What can we use in practice?

      For instance, we can show a CEO that they would miss their revenue target by over 6% for the quarter, and tell us why i.e. in what ways has their business changed over the last year. Then we provide the prioritized ordered lists of quickest, cheapest and least risky paths to help turn them over the tide, with estimates of relative costs and expected probability of success.

    • Liked Dipanjan Sarkar

      Dipanjan Sarkar - Explainable Artificial Intelligence - Demystifying the Hype

      45 Mins

      The field of Artificial Intelligence powered by Machine Learning and Deep Learning has gone through some phenomenal changes over the last decade. Starting off as just a pure academic and research-oriented domain, we have seen widespread industry adoption across diverse domains including retail, technology, healthcare, science and many more. More than often, the standard toolbox of machine learning, statistical or deep learning models remain the same. New models do come into existence like Capsule Networks, but industry adoption of the same usually takes several years. Hence, in the industry, the main focus of data science or machine learning is more ‘applied’ rather than theoretical and effective application of these models on the right data to solve complex real-world problems is of paramount importance.

      A machine learning or deep learning model by itself consists of an algorithm which tries to learn latent patterns and relationships from data without hard-coding fixed rules. Hence, explaining how a model works to the business always poses its own set of challenges. There are some domains in the industry especially in the world of finance like insurance or banking where data scientists often end up having to use more traditional machine learning models (linear or tree-based). The reason being that model interpretability is very important for the business to explain each and every decision being taken by the model.However, this often leads to a sacrifice in performance. This is where complex models like ensembles and neural networks typically give us better and more accurate performance (since true relationships are rarely linear in nature).We, however, end up being unable to have proper interpretations for model decisions.

      To address and talk about these gaps, I will take a conceptual yet hands-on approach where we will explore some of these challenges in-depth about explainable artificial intelligence (XAI) and human interpretable machine learning and even showcase with some examples using state-of-the-art model interpretation frameworks in Python!

    • Liked Akshay Bahadur

      Akshay Bahadur - Minimizing CPU utilization for deep networks

      Akshay Bahadur
      Akshay Bahadur
      Symantec Softwares
      schedule 1 year ago
      Sold Out!
      45 Mins

      The advent of machine learning along with its integration with computer vision has enabled users to efficiently to develop image-based solutions for innumerable use cases. A machine learning model consists of an algorithm which draws some meaningful correlation between the data without being tightly coupled to a specific set of rules. It's crucial to explain the subtle nuances of the network along with the use-case we are trying to solve. With the advent of technology, the quality of the images has increased which in turn has increased the need for resources to process the images for building a model. The main question, however, is to discuss the need to develop lightweight models keeping the performance of the system intact.
      To connect the dots, we will talk about the development of these applications specifically aimed to provide equally accurate results without using much of the resources. This is achieved by using image processing techniques along with optimizing the network architecture.
      These applications will range from recognizing digits, alphabets which the user can 'draw' at runtime; developing state of the art facial recognition system; predicting hand emojis, developing a self-driving system, detecting Malaria and brain tumor, along with Google's project of 'Quick, Draw' of hand doodles.
      In this presentation, we will discuss the development of such applications with minimization of CPU usage.

    • Liked Maryam Jahanshahi

      Maryam Jahanshahi - Applying Dynamic Embeddings in Natural Language Processing to Analyze Text over Time

      45 Mins
      Case Study

      Many data scientists are familiar with word embedding models such as word2vec, which capture semantic similarity of words in a large corpus. However, word embeddings are limited in their ability to interrogate a corpus alongside other context or over time. Moreover, word embedding models either need significant amounts of data, or tuning through transfer learning of a domain-specific vocabulary that is unique to most commercial applications.

      In this talk, I will introduce exponential family embeddings. Developed by Rudolph and Blei, these methods extend the idea of word embeddings to other types of high-dimensional data. I will demonstrate how they can be used to conduct advanced topic modeling on datasets that are medium-sized, which are specialized enough to require significant modifications of a word2vec model and contain more general data types (including categorical, count, continuous). I will discuss how my team implemented a dynamic embedding model using Tensor Flow and our proprietary corpus of job descriptions. Using both categorical and natural language data associated with jobs, we charted the development of different skill sets over the last 3 years. I will specifically focus the description of results on how tech and data science skill sets have developed, grown and pollinated other types of jobs over time.