-
Notifications
You must be signed in to change notification settings - Fork 0
/
hoist_conv_packed_params.cpp
141 lines (121 loc) · 4.75 KB
/
hoist_conv_packed_params.cpp
1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
20
21
22
23
24
25
26
27
28
29
30
31
32
33
34
35
36
37
38
39
40
41
42
43
44
45
46
47
48
49
50
51
52
53
54
55
56
57
58
59
60
61
62
63
64
65
66
67
68
69
70
71
72
73
74
75
76
77
78
79
80
81
82
83
84
85
86
87
88
89
90
91
92
93
94
95
96
97
98
99
100
101
102
103
104
105
106
107
108
109
110
111
112
113
114
115
116
117
118
119
120
121
122
123
124
125
126
127
128
129
130
131
132
133
134
135
136
137
138
139
140
141
#include <stack>
#include <torch/csrc/jit/api/module.h>
#include <torch/csrc/jit/jit_log.h>
#include <torch/csrc/jit/passes/constant_pooling.h>
#include <torch/csrc/jit/passes/constant_propagation.h>
#include <torch/csrc/jit/passes/hoist_conv_packed_params.h>
#include <torch/csrc/jit/passes/quantization/helper.h>
namespace torch {
namespace jit {
// Hoists packed params from a conv module to the parent module.
// The benefit is that after this hoisting, the conv module
// no longer holds anything and can be deleted, reducing model
// size.
//
// Before (easy case):
//
// %1 = prim::GetAttr[name="conv1"][%self]
// %2 = prim::GetAttr[name="_packed_params][%1]
//
// After (easy case):
//
// %2 = prim::GetAttr[name="{prefix}.conv1._packed_params"][%self]
//
// Before (generic case):
//
// %1 = prim::GetAttr[name="name1"][%self]
// %2 = prim::GetAttr[name="name2"][%1]
// ...
// %n = prim::GetAttr[name="_packed_params][%n-1]
//
// After (generic case):
//
// %n =
// prim::GetAttr[name="{prefix}.name1{...}.name(n-1)._packed_params"][%self]
//
static void hoistConvPackedParams(
Module& rootModule,
Node* getConvPackedParamsNode,
const std::string& prefix,
int& nameUniqueCounter) {
auto method = rootModule.get_method("forward");
auto graph = method.graph();
Value* rootModuleAsValue = graph->inputs()[0];
// get a path from root module to conv module
Value* convModuleAsValue = getConvPackedParamsNode->inputs()[0];
std::vector<std::string> rootToConvPath =
getModuleAccessPath(convModuleAsValue, rootModuleAsValue);
// get a module object representing the conv
Module convModule = findChildModule(rootModule, rootToConvPath);
// get the packed params value
c10::IValue packedParams = convModule.attr("_packed_params");
// create the new name
std::string suffix = "";
for (const auto& attrName : rootToConvPath) {
suffix += attrName + ".";
}
std::string newNameBase = prefix + "." + suffix + "_packed_params";
nameUniqueCounter++;
std::string newName = newNameBase + "." + c10::to_string(nameUniqueCounter);
while (rootModule.hasattr(newName)) {
nameUniqueCounter++;
newName = newNameBase + "." + c10::to_string(nameUniqueCounter);
}
// copy the packed params
rootModule.register_attribute(newName, packedParams.type(), packedParams);
// change target module to rootModule
getConvPackedParamsNode->replaceInput(0, rootModuleAsValue);
// change attribute name to new name
getConvPackedParamsNode->s_(Symbol::attr("name"), newName);
}
void HoistConvPackedParams(script::Module& m) {
auto method = m.get_method("forward");
auto graph = method.graph();
std::stack<Block*> blocks_to_visit;
blocks_to_visit.push(graph->block());
std::string attr_name_base = "_jit_pass_hoist_conv_packed_params";
// counter to ensure new attribute names are unique
int nameUniqueCounter = 0;
while (!blocks_to_visit.empty()) {
Block* b = blocks_to_visit.top();
blocks_to_visit.pop();
for (Node* n : b->nodes()) {
// make sure this node is fetching {foo}.{_packed_params}
bool isGetPackedParamsNode =
n->kind() == prim::GetAttr && n->s(attr::name) == "_packed_params";
if (isGetPackedParamsNode) {
// make sure the foo in {foo}.{_packed_params} is a quantized conv
c10::optional<std::string> moduleName = getModuleName(n->inputs()[0]);
bool moduleNameIsQuantizedConv = moduleName.has_value() &&
(moduleName.value() ==
"__torch__.torch.ao.nn.quantized.modules.conv.Conv1d" ||
moduleName.value() ==
"__torch__.torch.ao.nn.quantized.modules.conv.Conv2d" ||
moduleName.value() ==
"__torch__.torch.ao.nn.quantized.modules.conv.Conv3d" ||
moduleName.value() ==
"__torch__.torch.nn.intrinsic.quantized.modules.conv_relu.ConvReLU1d" ||
moduleName.value() ==
"__torch__.torch.nn.intrinsic.quantized.modules.conv_relu.ConvReLU2d" ||
moduleName.value() ==
"__torch__.torch.nn.intrinsic.quantized.modules.conv_relu.ConvReLU3d" ||
// BC Stuff
moduleName.value() ==
"__torch__.torch.nn.quantized.modules.conv.Conv1d" ||
moduleName.value() ==
"__torch__.torch.nn.quantized.modules.conv.Conv2d" ||
moduleName.value() ==
"__torch__.torch.nn.quantized.modules.conv.Conv3d");
if (moduleNameIsQuantizedConv) {
GRAPH_UPDATE("Hoisting ", *n, " to root module.");
hoistConvPackedParams(m, n, attr_name_base, nameUniqueCounter);
}
}
for (Block* subblock : n->blocks()) {
blocks_to_visit.push(subblock);
}
} // for
} // while
}
} // namespace jit
} // namespace torch