With all the focus on Test practices and tools getting better for the present, it is easy to forget about the Future of Testing, especially in terms of Tools and Infrastructure? A question that always comes up in my mind - “Are we so caught up in the past and present, that we will not be as effective in the future?”

In this talk, we will go on a journey to figure out what new challenges are coming up in the future, and more importantly, what do we need to do next to prepare for it. Also, just preparing for the future is not sufficient. We have an opportunity to stretch beyond our current set of skills, capabilities and boundaries to influence out future! The question is - will we make use of this opportunity? Will Selenium help us take the step in the right direction? Or, will it hold us back?

 
2 favorite thumb_down thumb_up 0 comments visibility_off  Remove from Watchlist visibility  Add to Watchlist
 

Outline/structure of the Session

  • Skills & Capabilities required for being a good QA
  • Evolution and trends of hardware & software
  • Spatial operating systems / Sixth Sense - future or reality?
  • Which life are we living - Past, Present or Future?
  • Tips and Techniques to be innovative

Learning Outcome

  • Importance to be in-sync with what is coming up next
  • What does the future look like from Testing perspective? How can we influence it?
  • Tips and Techniques for being leaders, instead of followers

Target Audience

Testers, Developers, Managers, Innovators

schedule Submitted 1 year ago

Comments Subscribe to Comments

comment Comment on this Proposal

  • Liked Anand Bagmar
    keyboard_arrow_down

    Anand Bagmar - To Deploy or Not-to-Deploy - decide using TTA's Trend & Failure Analysis

    Anand Bagmar
    Anand Bagmar
    Director - Quality
    Vuclip Inc.
    schedule 1 year ago
    Sold Out!
    60 mins
    Demonstration
    Intermediate

    The key objectives of organizations is to provide / derive value from the products / services they offer. To achieve this, they need to be able to deliver their offerings in the quickest time possible, and of good quality!

    In order for these organizations to to understand the quality / health of their products at a quick glance, typically a team of people scramble to collate and collect the information manually needed to get a sense of quality about the products they support. All this is done manually.

    So in the fast moving environment, where CI (Continuous Integration) and CD (Continuous Delivery) are now a necessity and not a luxury, how can teams take decisions if the product is ready to be deployed to the next environment or not?

    Test Automation across all layers of the Test Pyramid is one of the first building blocks to ensure the team gets quick feedback into the health of the product-under-test.

    The next set of questions are:
        •    How can you collate this information in a meaningful fashion to determine - yes, my code is ready to be promoted from one environment to the next?
        •    How can you know if the product is ready to go 'live'?
        •    What is the health of you product portfolio at any point in time?
        •    Can you identify patterns and do quick analysis of the test results to help in root-cause-analysis for issues that have happened over a period of time in making better decisions to better the quality of your product(s)?

    The current set of tools are limited and fail to give the holistic picture of quality and health, across the life-cycle of the products.

    The solution - TTA - Test Trend Analyzer

    TTA is an open source product that becomes the source of information to give you real-time and visual insights into the health of the product portfolio using the Test Automation results, in form of Trends, Comparative Analysis, Failure Analysis and Functional Performance Benchmarking. This allows teams to take decisions on the product deployment to the next level using actual data points, instead of 'gut-feel' based decisions.

  • Liked Sridharan Vembu
    keyboard_arrow_down

    Sridharan Vembu - Over-selling the "Enterprise Agile Frameworks and Certifications"

    45 mins
    Talk
    Intermediate
    Agile is only for smaller projects and/or startup organisations - Not Anymore. Taking my own and my organisation's experience, Agile is a proven methodology that is well suited for delivering complex, distributed, multi-year enterprise programs, for many years now.
     
    While this is really a great thing for agile enthusiasts and practitioners, it’s a bit of worrying sign for me the increased recognition and popularity the ‘Agile Certifications’ and ‘Agile Frameworks’ are receiving among individuals and organisations who would like to adopt Agile to stay relevant in current world.
     
    I would like to share my views on the adoption of these frameworks and certifications, why I feel they are not-so-agile and how am I and my organisation are solving similar problems without the need for any of these frameworks and certifications.
     
    I am planning to walk through the complete life cycle of the most recent program that I’m part of (from Inception to Initiation to on-going Execution to Post-Production Support) and bring out the relevant agile principles that we adopted, context based customizations we did and the best practices that we have come up with.
    • For instance, one should know the clear difference between hygienic practices vs context based practices - the first ones are not to be compromised at any cost, whereas the latter ones are to be applied based on the need, not because some framework prescribes it.
    The typical life cycle stages that we follow in any program / project delivery is normally: Discovery - Inception - Initiation - Execution - Transition, whereas the actual set of practices within any of these stages and how they are being implemented could be very different from project to project, team to team. 
    • For example, in the Execution Phase, doing pair programming and following TDD are hygienic practices for us. Having said that, it’s perfectly okay for a pair to split and work on a specific task on a case-to-case basis (we call this Pragmatic Programming) and the pair decides when and how long they would split and when to re-join.
    To give an idea on the complexity, enterprise and distributed nature of the program, some key data points:
    • Started almost 3 years ago, on-going
    • 10 quarterly planning workshops done so far
    • 10+ teams, 7 timezones
    • peak program size: 250+
    • peak team size from my org.: 50+
    • total no of systems: 10+
    • geographical spread: plan: 100 countries, 132 locales launched so far: 53 countries, 56 locales
    • 140 page-views / sec
    • Av. response time: 1.3s
    • Handling 100+K products in the catalog, 15+ K pages , 300+ K responsive images
    • Blue-Green production deployment (zero downtime over 1.5 years)
    • 3 weeks cycle of production releases