When declaring a model in pytorch, having a model class member variable declared and initialized mysteriously prevents it from being populated in the constructor. Is this expected? If so, why?
Testing code below, with example models with a component member variable.
The initialization value of the component (e.g. None, a number or a Tensor) does not change the behaviour.
import torch
class Lin1(torch.nn.Module):
def __init__(self):
super(Lin1, self).__init__()
self.component = torch.nn.Linear(100,200)
class Lin2(torch.nn.Module):
component = None
def __init__(self):
super(Lin2, self).__init__()
self.component = torch.nn.Linear(100,200)
# instantiate and check member component
for cl in [Lin1, Lin2]:
model = cl()
print("nModel:", cl)
print("Component:", model.component)
print("It's None?: ", model.component is None)
Model: <class '__main__.Lin1'> Component: Linear(in_features=100, out_features=200, bias=True) It's None?: False Model: <class '__main__.Lin2'> Component: None It's None?: True
Advertisement
Answer
This happens because nn.Module overwrites __getattr__, and it would only work as you expect if component was not in Lin2.__dict__ (nor in Lin2().__dict__). Since component is a class attribute, it is in Lin2.__dict__ and will be returned as it should.
When you write self.x = nn.Linear(...) or any other nn.Module (or even nn.Buffer or nn.Parameter), x is actually registered in a dictionary called _modules (or _buffers, etc.) In this way, when you ask for self.component, if component is already in the __dict__ of the class or the instance, Python will not call the custom nn.Module‘s __getattr__().
You can check the source-code of __getattr__ from nn.Module here. A similar discussion was done here. There was also a discussion about changing from __getattr__ to __getattribute__ in PyTorch, but as of now, this is a wontfix issue.