Testing & Quality Engineering

A Journey Through Testing Talks Sydney 2024


AWS Advanced Tier Partner

Insights from the Testing Talks Conference - Sydney

In the heart of Sydney, the Testing Talks Conference convened, bringing together some of the brightest minds in the testing world. As I navigated through the sessions in the Green Track, I found myself not just learning, but also rethinking the very fabric of software testing. Each presentation offered a unique lens into the future, where digital twins, AI, and continuous testing redefine the landscape.


Take on Digital Twins

Tarek El Merachli

Imagine a world where every critical system in an airplane is mirrored in a virtual environment, a digital twin. In this twin, engineers can predict failures before they happen, simulate scenarios, and ensure safety without risking a single life. The presentation transported us into this world, reflecting on past aviation disasters that might have been averted with this technology.

As Tarek outlined, the digital twin market is booming, poised to revolutionize industries beyond aviation. But with this innovation comes challenges—unique to each phase of the Software Development Life Cycle (SDLC). Tarek didn't shy away from these hurdles; instead, they dissected them, offering a roadmap for integrating digital twins into testing frameworks. The session ended on a forward-looking note, painting a future where digital twins are not just tools, but essential companions in every testing strategy.


Vision of the Future

Manoj kumar kumar

Next, we looked into the world of AI with Manoj. Here, the focus shifted from physical twins to digital brains—Large Language Models (LLMs) that can understand, generate, and even dream up text. Manoj's presentation wasn't just about the capabilities of these models; it was about their potential to transform test automation.

In a world increasingly powered by AI and machine learning, testing isn't just about finding bugs; it's about ensuring that the algorithms driving our systems are trustworthy. Manoj took us on a journey through the complexities of testing AI itself. They explored the concept of Human-in-the-Loop (HITL), where human judgment is critical in guiding AI, and discussed the dangers of LLM hallucinations—instances where AI generates plausible but false information.

One of the most intriguing concepts presented was Retrieval Augmented Generation (RAG), likened to Docker containers for AI. Just as containers isolate and manage environments for software, RAG isolates and manages information, ensuring AI models generate accurate and relevant outputs. This analogy resonated with the audience, as it bridged the familiar with the cutting-edge.


API-Centric Systems: Blazemeter's Continuous Testing Journey

Srdjan Nalis

In a world where systems are increasingly interconnected through APIs, the importance of robust API testing cannot be overstated. Srdjan's presentation pulled back the curtain on the hidden complexities of today's digital ecosystems. APIs are the glue that holds our modern applications together, but they also represent a critical point of failure if not properly tested.

Srdjan introduced us to their approach, which goes beyond traditional testing. They highlighted the challenges of testing in environments where APIs are constantly evolving, and how Blazemeter's suite of tools—rooted in open source—can navigate these challenges. The session showcased Blazemeter's Virtual Services, which allow teams to simulate complex environments and interactions without the need for full-scale deployments. This capability is a game-changer, especially in a world where time-to-market is everything.

But what stood out most was their comparison between traditional API monitoring tools and Blazemeter's approach. In a side-by-side analysis, it became clear that Blazemeter offers a more integrated and comprehensive solution, particularly for organizations operating in both cloud and on-prem environments.


AI in Testing: A Panel of Perspectives

Lisa Pfitzner

The conference's panel discussion on AI in testing was a conversation among thought leaders who weren't just predicting the future—they were shaping it. The debate centered on the role of AI in automating tests, with a consensus that automation should be purposeful, not endless.

One key takeaway was the importance of focusing on what truly matters in testing. It's easy to get caught up in automating everything, but as the panelists pointed out, not everything needs to be automated. The discussion also touched on the concept of “hard gates”—non-negotiable checkpoints in the testing process that ensure quality isn't compromised, even as automation takes on more tasks.


The Quest for Meaningful Automation: Applitools' Perspective

Anand Bagmar

Anand' session began with a provocative question: Why do we automate tests? It was a simple question, but one that prompted deep reflection. The answer, as the speaker explained, isn't just about efficiency; it's about achieving outcomes that matter.

To drive this point home, Anand introduced Testwiz, a tool designed to optimize test automation efforts. But more than the tool itself, it was the underlying philosophy that resonated. Automation should not be an end in itself; it should serve the broader goal of delivering quality software. The importance of setting “hard gates” was revisited, emphasizing that these gates are crucial in maintaining control over automated processes.


Quality at Every Phase: The REA Group's Blueprint

Deepika Nagaraja

Quality isn't something that can be tacked on at the end of a project; it must be woven into every phase. Deepika's presentation was a masterclass in how to achieve this. They shared a framework for conducting health checks at every stage of development, ensuring that quality is maintained from start to finish.

Their approach was holistic, considering not just the technical aspects of quality, but also the human factors. It was a reminder that quality engineering is as much about people as it is about processes.


The Metrics of Quality: Inflectra's Approach

Adam Sandman

Measuring quality can be elusive, but Adam's session offered practical insights into how it can be done effectively. They presented tools and methodologies designed to quantify quality, providing teams with the data they need to make informed decisions.

What was particularly interesting was their focus on continuous measurement. In today's fast-paced development environments, quality isn't static—it's dynamic, and Inflectra's tools are built to reflect that reality.


Decoding Quality Engineering in the Age of AI: DevOps1's Vision

Alejandro Sanchez-Giraldo

The conference ended with a thought-provoking session from DevOps1, which explored the evolving role of human insight in an age increasingly dominated by AI. As AI takes on more responsibilities, the challenge will be to maintain the human element that has always been at the heart of quality engineering.

DevOps1's vision is one where AI and human expertise work hand-in-hand, each enhancing the other. It's a future where AI handles the heavy lifting, but human judgment ensures that the outcomes align with our values and expectations.

Conclusion

The Testing Talks Conference in Sydney wasn't just a showcase of the latest tools and technologies; it was a glimpse into the future of testing. From digital twins to AI-driven automation, the sessions on the Green Track offered a rich tapestry of ideas, challenges, and solutions. As we move forward, these insights will be invaluable in shaping how we think about testing—not just as a technical discipline, but as a critical component of delivering quality in the digital age.

Alejandro Sanchez-Giraldo
Alejandro Sanchez Giraldo
Head of Quality Engineering and Observability

Alejandro is a seasoned professional with over 15 years of experience in the tech industry, specialising in quality and observability within both enterprise settings and start-ups. With a strong focus on quality engineering, he is dedicated to helping companies enhance their overall quality posture while actively engaging with the community.

Alejandro actively collaborates with cross-functional teams to cultivate a culture of continuous improvement, ensuring that organisations develop the necessary capabilities to elevate their quality standards. By fostering collaboration and building strong relationships with internal and external stakeholders, Alejandro effectively aligns teams towards a shared goal of delivering exceptional quality while empowering individuals to expand their skill sets.

With Alejandro's extensive experience and unwavering dedication, he consistently strives to elevate the quality engineering landscape, both within organisations and across the wider community.

Related Posts

3 May, 2023

DevOps1 Approach to Enterprise Data Masking

Introduction Data is an asset as well as a liability. In a complex enterprise environment, we have large IT systems involving multiple enhancements and change requests all the time.

1 May, 2023

Delphix Partners with DevOps1 in Australia

Our technologists are dedicated to using their collective experience and knowledge of modern engineering best practices to deliver effective, secure, and forward-thinking solutions.

13 Mar, 2023

Why is Certificate Management important to businesses?

Certificate management refers to the process of managing digital certificates in a DevOps environment.