6 expensive ways to fail in an automated testing project

Here are 6 ways that we failed in a very big Selenium based testing project. These automated testing mistakes cost us a lot of money, due to the maintainability, performance and visibility problems they created in our tests.

Want to know more about how to make your Selenium tests faster? follow this one.

Rephrasing Tolstoy, each automation project has the same green-yellow-red “traffic light”, but each project fails in a unique way. We failed on all the points that will soon be mentioned 😊 Yeay!
But worry not! We now have a state of the art Test Automation project, running tests 10s of times a day in our CI\CD pipeline.

The test automation project I will refer to here is in high focus inside the company (thanks TimeToKnow for letting me share this experience here), used for “end to end” sanity and regression tests. Due to the nature of the product (a web based SASS platform), the tests are aimed to run through browsers or mobile devices.

Enjoy our automated testing mistakes😊

1. “Here are the test results! What? You can’t understand stderr output on your own?”

Once we wanted to make the entire R&D organization an integral part of the testing analysis mechanism, we started getting questions about “how do I read this test output?”.
For us as the ones who wrote the tests, it was fairly easy to read a report that looks like this:

6 expensive ways to fail in an automated testing project - the standard error used as the test report. Not readable for anyone, but the ones who wrote it. hopefully.
6 expensive ways to fail in an automated testing project – the standard error used as the test report. Readable for the ones who wrote it, not for our customers.

Readable indeed. Not really.

After some research we started using a product called ExtentReports (yes, it has a great free version).
And now it looks like this:

6 expensive ways to fail in an automated testing project - first improvement of the test result's report, using basic configuration of ExtentReports
6 expensive ways to fail in an automated testing project – first improvement of the test result’s report, using basic configuration of ExtentReports

This is better, and answers the “what failed” question that developers are eager to know.

But what about bug reproduction? Video to the rescue!
Our latest improvement related to the reports mechanism was to add video capabilities, based on this predefined docker container, and – it’s free.

6 expensive ways to fail in an automated testing project - a much better report with statistics from previous runs, and an embedded video of the failed test, using ExtentReports

2. More git repositories than developers

The system under test is made up of microservices both on the client and server sides (that’s a subject for a different post). When we started the automation project we had a designated test repository for each of the microservices, and a common infra repository. With a total of 7 repositories with dependencies between them.
During that time, the team underwent significant changes – maternity leaves, some personal shifts to other places in the organization etc. At the end, we ended up with 2 developers, working on 7 different repositories.
This doesn’t make sense 😊.
There’s no mathematical correctness here, but as a rule of thumb – if you spend too much time moving between repositories and updating repositories – it’s time to unite.

3. Too much boiler plate code

When we first started the automation project, it was very simple. Start a new project, add dependencies for Selenium, start building the project’s structure based on the Page Object pattern, and the skies were blue, rainbows everywhere.
After a while, we needed our own customizations:

  • Customized reports that have a very specific look, with a nice dashboard, aggregating test results from many test runs.
  • Pluggable “browser” object to support many types of configurations such as browser type, running on mobile or tablet, etc.

A year into the project – we saw that almost 20% of our code handles things that looked (to us) like they are not a part of our business logic.
20% means lots of time to develop, lots of time to test, and lots of focus put on these places.
While sometimes it is a must to do these things “in our own way”, most of the time it’s better to invest in RESEARCH rather than on reinventing stuff.
Need a customized report? Google it.
Need a way to capture videos? Google it.

It took time, but now we google it.

4. Hard coded configurations

Some of the tests need to have information about their surroundings, for instance: login URL, server URL (for API usage), bucket definitions, specific timeouts for different environments, a different failure threshold for specific tests on specific environments.
It starts small, but the number of configurations can easily get larger when tests become more complex. And when it grows – you will definitely have more environment-related logic in the test code.
How can you tell if you suffer from the same disease?

Does this look familiar?

If (environment == "my-staging-ip") {
…
}

How to solve it?
A common solution is to pass configurations in runtime, using parameterized builds in Jenkins.
The code becomes more “stupid” and the control over the tests becomes way simpler.
A new environment to test? 30 seconds and it’s configured!

5. Timeouts. Lots of timeouts.

In tests – timing is the king. And when the timing fails (clicking on an item before it’s there, not waiting for a loader to disappear, starting the tests a millisecond before the environment is up) – all hell breaks loose.
A common mistake, that obviously we made too, is to base our waiting mechanism on absolute time.

“When will the loader disappear? Wait for 5 seconds, it will definitely disappear”. This approach leads to the following “bad code smell”:

// enter a page
Some code here..

/// wait 5 seconds for the loader to disappear
Thread.sleep(5000);

// now we continue to work on the loaded page, clicking elements, etc
Some more code here..

What’s wrong with it? Well, everything. One day there’s a glitch in your office’s network and 5 seconds isn’t enough, and it fails. The hard truth is that you can almost never know when it’s enough.

To overcome this issue, we switched to a “wait for an expected condition” approach, where we wait for specific conditions, for example waiting at most 10 seconds for that loader element not to appear in the DOM and only then proceed. If it disappeared after 2 seconds – the code continues, no time is wasted.

Bonus: A good approach here is to have a default timeout used in all places, and to have the ability to control it by a configuration.

6. Not realizing who the customers are

An automated testing project is a software product in every single way, with an allocated budget, organizational needs, and… it has customers.
Who are the customers? It really depends on the goals of the project. It can be the test team, managers, developers, devops team, etc.
In any case, a project cannot be successful unless you know WHO your customers are, and WHAT they want to do with it.
In our case there are 2 types of customers (forgive me for oversimplifying):

  1. Developers: “I committed new code, does it break anything in the system?”
  2. Managers: “The release is tomorrow – is the version ready for production?”

When tests pass – everyone is happy. But when something breaks – there are many things to discover: what failed, is it a regression, how to reproduce it, what’s its severity?
To handle these, we put lots of effort into improving test reports (see the paragraphs above), and lots of time TALKING TO THE DEVELOPERS so we’ll be in the same boat, both hearing the needs and explaining what the failures mean.

Is the version ready for production? To answer this one, a deeper understanding of the real users’ usage of the system is required.

Meaning: a RED test may be a regression that has a low priority for customers, so it’s worth releasing a new version to production and only then fixing the regression. This is our approach anyway, and when we started to talk in this language, the managers became more satisfied and felt a real contribution to the automation project.