Python returns MagicMock object instead of return_value
I have a python file a.py
which contains two classes A
and B
.
class A(object):
def method_a(self):
return "Class A method a"
class B(object):
def method_b(self):
a = A()
print a.method_a()
I would like to unittest method_b
in class B
by mocking A
. Here is the content of the file testa.py
for this purpose:
import unittest
import mock
import a
class TestB(unittest.TestCase):
@mock.patch('a.A')
def test_method_b(self, mock_a):
mock_a.method_a.return_value = 'Mocked A'
b = a.B()
b.method_b()
if __name__ == '__main__':
unittest.main()
I expect to get Mocked A
in the output. But what I get is:
<MagicMock name='A().method_a()' id='4326621392'>
Where am I doing wrong?
When you @mock.patch('a.A')
, you are replacing the class A
in the code under test with mock_a
.
In B.method_b
you then set a = A()
, which is now a = mock_a()
- i.e. a
is the return_value
of mock_a
. As you haven't specified this value, it's a regular MagicMock
; this isn't configured either, so you get the default response (yet another MagicMock
) when calling methods on it.
Instead, you want to configure the return_value
of mock_a
to have the appropriate method, which you can do as either:
mock_a().method_a.return_value = 'Mocked A'
# ^ note parentheses
or, perhaps more explicitly:
mock_a.return_value.method_a.return_value = 'Mocked A'
Your code would have worked in the case a = A
(assigning the class, not creating an instance), as then a.method_a()
would have triggered your mock method.