Test-Driven Development with Python

My notes and reflections from reading Test-Driven Development with Python by Harry J.W. Percival and coding along (code found here).

TLDR: I highly recommend the book as an introduction to TDD in Python, more specifically using the Django framework. Pushing the reader to actually setup a staging server together with deployment scripts and a Jenkins server for continuous testing ensures that you get an end-to-end hands-on basic understanding of the process.

Part 1: The fundamental workflow of TDD and a gentle introduction to Django

Setup according to instructions went really smoothly, which was really nice. The instructions were also clear, so very easy to code along. Quite soon I got into the habit to first try to think about what the next step should be after finishing a piece of code, or after updating a test and getting a new failure. This practice helped keeping me more engaged.

I really appreciate how unit testing and functional testing are handled in parallel, as that keeps the focus both on the low level internal consistency, but also tying back to the user story level to ensure the code actually is achieving something of use.

Part 2: Real Developers Ship

8. Prettification: Layout and Styling, and What to Test About It

Quick intro to bootstrap and django templating.

9. Testing Deployment Using a Staging Site

Pushing you to setup the site on an actual server and get a proper domain. Really appreciated the push, and it went smoothly to set it up in AWS, though I expect that for a beginner with no previous experience of servers, this might have been a very steep learning curve.

10. Getting to a Production-Ready Deployment

Installing and configuring nginx and gunicorn in order to replace the Django built-in server. Further, updating settings.py to look for environment variables so that debug and allowed_hosts have more secure settings in production mode.

Overall straight forward, but a good reminder of how much time you can end up spending on troubleshooting the smallest typo when it comes to server configurations.

11. Automating Deployment with Fabric

Excellent chapter that shows both how easy deployment becomes with Fabric, but also some really neat CL commands to work with template files. No matter how much else from this book I actually end up using, I’m sure this will come in handy on many occasions.

Example; cat ./deploy_tools/gunicorn-systemd.template.service \ | sed "s/DOMAIN/prod.randomtechramblings.com/g" \ | sudo tee /etc/systemd/system/gunicorn-prod.randomtechramblings.com.service

12. Splitting Our Tests into Multiple Files, and a Generic Wait Helper

Quick chapter just to refactor the tests into a more maintainable structure.

13. Validation at the Database Layer

Starting to add a bit more Django magic to help out with routing and form validation.

14. A Simple Form

Long chapter introducing Django forms, and a bit intentionally tedious refactoring to really push for TDD discipline. As my interest lies more in learning more Python and TDD than Django, some of the content felt less interesting than many of the previous chapters.

One thing to remember; self.fail(object) prints the object to the CL, which can help a lot when debugging failing tests…

15. More Advanced Forms

Guess I want to be a completionist in regards to this book, because I went through this chapter too though it was clearly marked as optional. Hopefully, getting a bit more familiarity with the Django forms will come in handy at some point in the future.

16. Dipping Our Toes, Very Tentatively, into JavaScript

Short chapter with just the bare minimums, which suits me quite nice. I would though have appreciated a bit of discussion on the duplication of html code, when the unit test doesn’t actually test the django template it seems like it’s bound to happen that the markup there at some point will change, without the javascript test being properly updated.

17. Deploying Our New Code

Another short chapter, since the release process was semi automated already.

Part 3: More Advanced Topics in Testing

18. User Authentication, Spiking, and De-Spiking

Example of doing a quick and dirty proof of concept first to assure that a solution is viable, and once proven to scrap it and rewrite the code using TDD.

19. Using Mocks to Test External Dependencies or Reduce Duplication

Finishing up auth functionality along with tests that mocks the e-mail send out. Back to focus on the core python unit test packages, which I appreciate.

20. Test Fixtures and a Decorator for Explicit Waits

Adding fixtures to help test isolated parts more efficiently without having to go through the whole setup process (eg. logging in) every time in the browser.

Further, providing a neat example of creating a decorator in order to avoid duplicate code.

21. Server-Side Debugging

Setting up FT so that we check that e-mails actually gets sent out on the staging server. Here there were some inaccuracies in regards as to how the POP protocol works, and I had to update the code a bit to get it working. Further, the Yahoo mail server needed an app password (and for that, 2F auth enabled) to allow POP connection.

Further, getting into some neat Python functionality with a good example of using decorators. Lastly, revisiting Fabric, and how to allow the functional tests to interact with the staging server when run from localhost. Great chapter all in all.

22. Finishing “My Lists”: Outside-In TDD

Clarifying the concept of working from outside and in, i.e. start with a failing functional test mirroring a user story, and then gradually work inwards as needed in order to get the full test to pass.

Among the benefits of this approach is that your APIs gets built based on what makes sense on higher abstraction levels rather than vice versa. Additionally, keeping the development tied to solving the user story reduces the risk of ending up writing too much low level code trying to support features that you probably are never going to need.

Lastly, working from the outside-in with the user flows should help you identify UX problems, or at least be able to provide early prototypes to stakeholders for testing as needed.

23. Test Isolation, and “Listening to Your Tests”

At the models layer, we no longer need to write isolated tests—​the whole point of the models layer is to integrate with the database, so it’s appropriate to write integrated tests”

I don’t agree fully with this. It can make sense to keep models as framework agnostic as possible as well, and instead rely on a repository pattern for the storage. The main purpose of this would then be to more easily be able to switch out the persistence backend without affecting the logic of the model.

The chapter was overall informative and gave some good insights into pros and cons with more integrated vs. isolated tests. I think the resetting of master to a tag wasn’t the nicest git workflow, meaning in the end I had to use git push --force to overwrite the previous master state in the origin repo.

24. Continuous Integration (CI)

Great chapter on settings up and configuring Jenkins to do continuous testing. Involved some troubleshooting to get the tests working, and in the end I had some functional test that inconsistently kept failing on the Jenkins server, but runs fine on localhost and staging server.

It seems that there’s something weird in the current Linux Firefox version, as far as I can tell, that means that the javascript doesn’t always get executed correctly. In the end, I paused the test in question (ItemValidationTest.test_error_messages_are_cleared_on_input) in order to be able to move on.

25. The Token Social Bit, the Page Pattern, and an Exercise for the Reader

The chapter gives you a broken functional test, and then points you through initial steps that will be needed to solve.

After two weeks of vacation, coming back into this and now having to apply the methodology myself without training wheels, it was a good experience and helped drive home some of the points from earlier chapters.

It’s becoming clear to me the importance of using ‘spikes’ to prototype quickly and test things, because the disciplined approach of always writing tests first really hinges on that you have a clear idea of where you are going.

26. Fast Tests, Slow Tests, and Hot Lava

The final chapter discusses unit tests, integrated tests, system tests and functional tests and the pros and cons of each.

The main benefit of pure unit tests is that they stay fast to run when the number of tests increases, which means they are also more likely to be run often since they are fast iterations. Further, since they rely on identifying and mocking dependencies, they also help pushing for a decoupled design which, depending on context, can save a lot of headache later on in the software lifecycle.

On the other hand, using mocks can tie the test closely to the implementation, if it relies on certain methods on objects being called, which will mean the tests will have to be refactored if the implementation changes, which goes against the basic idea of unit testing as a ‘black box’ test of the contract.

A strategy to design the architecture to lessen these challenges is the “Functional Core, Imperative Shell” pushed by Gary Bernhardt. This means that the core of the application is kept ‘pure’ so it can be tested in isolation, and the shell is where code more coupled to implementation/framework/external dependencies is kept. This sounds similar to using a repository pattern for models in order to keep the models agnostic, and keep the storage implementation for them separate.

Though, this gives me a new understanding of the quote from chapter 23, that the models main purpose is to interface with the database, which is the case for the Django models. If the actual data contained in the model is implemented as a separate data structure owned by the model, then the django model just becomes the shell for db interaction.

Comments