Classes that use other classes (as members, or as arguments to methods) need instances that behave properly for unit test. If you have these classes available and they introduce no additional dependencies, isn’t it better to use the real thing instead of a mock?
Classes that use other classes (as members, or as arguments to methods) need instances
Share
I say use real classes whenever you can.
I’m a big believer in expanding the boundaries of ‘unit’ tests as much as possible. At this point they aren’t really unit tests in the traditional sense, but rather just an automated regression suite for your application. I still practice TDD and write all my tests first, but my tests are a little bigger than most people’s and my green-red-green cycles take a little longer. But now that I’ve been doing this for a little while I’m completely convinced that unit tests in the traditional sense aren’t all they’re cracked up to be.
In my experience writing a bunch of tiny unit tests ends up being an impediment to refactoring in the future. If I have a class A that uses B and I unit test it by mocking out B, when I decide to move some functionality from A to B or vice versa all of my tests and mocks have to change. Now if I have tests that verify that the end to end flow through the system works as expected then my tests actually help me to identify places where my refactorings might have caused a change in the external behavior of the system.
The bottom line is that mocks codify the contract of a particular class and often end up actually specifying some of the implementation details too. If you use mocks extensively throughout your test suite your code base ends up with a lot of extra inertia that will resist any future refactoring efforts.