The New York Times had a quiz a while back with a simple task: figure out the mathematical rule behind a sequence of numbers.
As a reader, you wou...
For further actions, you may consider blocking this person and/or reporting abuse
Great post!
One thing I find easy to forget about TDD is that you do not merely start with a test. You start by writing a FAILING test. Witnessing a test fail is essential so that you see the test doing its job. If you never saw your tests fail, how do you know they work?
Excellent post, and quite timely! Just made the same point in a comment elsewhere, and then came across this in my feed.
I mentioned in an article earlier, I recommend "testing blind". Even if one writes tests first, rewriting them later without looking at your code helps overcome that little subconscious bias we have of tiptoeing around error traps.
Also, as Jeff Atwood says, Do Terrible Things To Your Code.
To a certain extent, I love unhappy paths, but I see too often when someone who does that manually incorrectly applies it to code.
I know some QAs that would very happily write that the sky isn't green and the sky isn't purple and the sky isn't red. When they could just write that the sky is blue and be done with it.
While that may seem silly with that example, consider an api integration tests that does like
expect(Object.keys(getResponse)).not.toContain('links');
Lovely, we no longer use HATEOAS links. Good job, team. But also, that getResponse doesn't contain
'puppies'
or'kittens'
properties either. It might have a'link'
property because someone got cheeky and changed the name on us.expect(Object.keys(getResponse)).toEqual(['thing', 'stuff', 'bunnies']);
gives use better info. The negative would be testing that being unauthorized gives us['error', 'message']
but too many people seem to think it's looking for puppies and kittens and verifying it's not there.Somewhat related, but using De Morgan's Laws in tests is fun, too, especially during code review.
.not.toBeLessThan
is the same as.toBeGreaterThanOrEqual
but that second function is only in more recent versions of Jasmine.Some very good points! I think this would make a great follow up post, going through different scenarios and how to ensure you've covered the major scenarios, but not gone too far in testing unneeded scenarios.
I think it stems from more tests being assumed to always mean better. And having lots of tests obviously means we'll never have bugs ever.
Bugs will always happen, but good tests mean they get caught earlier and hopefully before things get merged to the main codebase, so yay! But that doesn't mean there are no bugs. And that doesn't mean that there are no missing requirements. It just means tests passed :P
And there will always be someone writing a test to verify the sky isn't red, and someone looking at the metrics and being happy test count and coverage increased.
A good test is like a scientific experiment. You are not trying to prove your code works but you are trying to falsify assumptions that need to be true for it to not work. Then when it breaks later, it will be because of something you did not think of. It will be an assumption you made that turned out to be wrong that you did not test.
This post is a realization of how I feel about testing. Don't just do happy path! Life doesn't follow a happy path always, and neither should your testing. Great write-up!
People have come up with an interesting approach for this, mutation testing. For example check stryker. stryker-mutator.io/
Yes, this is an incredibly important point. Part concepts we should pay attention to, part basic self-awareness and bias attention.
Pretty cool article and yes, sometimes we should break a head before we find a negative test case to pass. That helps you better to understand a code
Great article, Kevin! Brilliant!