Net类简介
Net类主要处理各个Layer之间的输入输出数据和参数数据共享等的关系。由于Net类的代码较多,本次主要介绍网络初始化部分的代码。Net类在初始化的时候将各个Layer的输出blob都统一保存在变量blobs_中,利用各个层的输入输出数据在blobs_的位置,可以方便地确定层之间数据传递的关系。Net中还利用类似的方法来存储各个层之间的学习参数,管理层之间的参数共享关系。
net.cpp部分源码
template
Net::Net(const NetParameter& param) { //构造函数,使用NetParameter类型的变量初始化网络
Init(param);
}
//param_file: Proto类型的文本文件名
//phase: 字段,网络的一种状态(TRAIN OR TEST) //这三种设置会在StateMeetsRule()函数中详细说明
//level: 级别,网络的一种状态
//stages: 同样为网络的一种状态,phase/level/stages三种均为网络的状态,定义在NetState消息中
template
Net::Net(const string& param_file, Phase phase,
const int level, const vector* stages) { //创建网络,并使用从param_file中读取的网络参数和三种网络状态初始化
NetParameter param;
ReadNetParamsFromTextFileOrDie(param_file, ¶m); //从ProtoTxt文件中读取网络参数(还做了些兼容旧版本的转换操作)
// Set phase, stages and level
param.mutable_state()->set_phase(phase); //设置网络状态,字段(TRAIN 或 TEST)
if (stages != NULL) {
for (int i = 0; i < stages->size(); i++) {
param.mutable_state()->add_stage((*stages)[i]); //设置网络的stage
}
}
param.mutable_state()->set_level(level); //设置网络的level
Init(param); //初始化网络
}
template
void Net::Init(const NetParameter& in_param) { //网络初始化
// Set phase from the state.
phase_ = in_param.state().phase(); //网络字段(TRAIN 或 TEST)
// Filter layers based on their include/exclude rules and
// the current NetState.
NetParameter filtered_param;
FilterNet(in_param, &filtered_param); //根据网络状态和layer的状态规则过滤部分layer,得到过滤后的网络参数
LOG_IF(INFO, Caffe::root_solver()) << "Initializing net from parameters: " << std::endl
<< filtered_param.DebugString();
// Create a copy of filtered_param with splits added where necessary.
NetParameter param;
InsertSplits(filtered_param, ¶m); //将filtered_param复制到param,并拆分其中一些复用的blob数据,参见insert_splits.cpp
// Basically, build all the layers and set up their connections.
name_ = param.name(); //设置网络的名称
map blob_name_to_idx; //<输出blob的名称, blob在blobs_中的索引>,只会保存输出blob的映射关系
//AppendBottom()会从中删除已被用作输入的输出blob,而AppendTop()会将layer的输出blob都加入其中.
//最终剩下的blob都认为是网络的输出,例如loss层的输出blob
set available_blobs; //当前还未被当作输入的所有输出blob
memory_used_ = 0; //内存使用大小,初始化为0
// For each layer, set up its input and output
bottom_vecs_.resize(param.layer_size()); //初始化net中各项参数的大小,layer的个数
top_vecs_.resize(param.layer_size());
bottom_id_vecs_.resize(param.layer_size());
param_id_vecs_.resize(param.layer_size());
top_id_vecs_.resize(param.layer_size());
bottom_need_backward_.resize(param.layer_size());
for (int layer_id = 0; layer_id < param.layer_size(); ++layer_id) { //网络中的各个层
// Inherit phase from net if unset.
if (!param.layer(layer_id).has_phase()) {
param.mutable_layer(layer_id)->set_phase(phase_); //当前layer未设置phase,则使用net本身的值设置
}
// Setup layer.
const LayerParameter& layer_param = param.layer(layer_id); //当前layer的参数
if (layer_param.propagate_down_size() > 0) { //设置了是否需要反向传播
CHECK_EQ(layer_param.propagate_down_size(),
layer_param.bottom_size())
<< "propagate_down param must be specified "
<< "either 0 or bottom_size times "; //检查设置的个数与layer的输入blob的个数是否相等
}
layers_.push_back(LayerRegistry::CreateLayer(layer_param)); //根据layer参数创建layer
layer_names_.push_back(layer_param.name()); //保存layer的名称
LOG_IF(INFO, Caffe::root_solver()) << "Creating Layer " << layer_param.name(); //主线程打印信息
bool need_backward = false; //该layer是否需要反向传播,下面会根据layer中输入blob和层的参数来决定该值
// Figure out this layer's input and output
for (int bottom_id = 0; bottom_id < layer_param.bottom_size(); ++bottom_id) { //处理层的输入blob
const int blob_id = AppendBottom(param, layer_id, bottom_id,
&available_blobs, &blob_name_to_idx); //将layer的第bottom_id个输入blob添加到param中
// If a blob needs backward, this layer should provide it.
need_backward |= blob_need_backward_[blob_id]; //如果输入blob需要反传,则所在的layer也设置成需要
}
int num_top = layer_param.top_size();
for (int top_id = 0; top_id < num_top; ++top_id) { //处理层的输出blob
AppendTop(param, layer_id, top_id, &available_blobs, &blob_name_to_idx); //将输出blob添加到net中
// Collect Input layer tops as Net inputs.
if (layer_param.type() == "Input") { //该layer为网络的数据输入层
const int blob_id = blobs_.size() - 1;
net_input_blob_indices_.push_back(blob_id); //保存该输出blob的位置
net_input_blobs_.push_back(blobs_[blob_id].get()); //保存该输出blob的指针
}
}
// If the layer specifies that AutoTopBlobs() -> true and the LayerParameter
// specified fewer than the required number (as specified by
// ExactNumTopBlobs() or MinTopBlobs()), allocate them here.
Layer* layer = layers_[layer_id].get();
//允许创建匿名blob.允许输出blob的个数比ExactNumTopBlobs()或者MinTopBlobs()的值小
if (layer->AutoTopBlobs()) {
const int needed_num_top =
std::max(layer->MinTopBlobs(), layer->ExactNumTopBlobs()); //需要的输出blob的个数
for (; num_top < needed_num_top; ++num_top) {
// Add "anonymous" top blobs -- do not modify available_blobs or
// blob_name_to_idx as we don't want these blobs to be usable as input
// to other layers.
//从此处来看,匿名blob不可用做其他层的输入,但会存储在blobs_中
AppendTop(param, layer_id, num_top, NULL, NULL); //添加到net中
}
}
// After this layer is connected, set it up.
layers_[layer_id]->SetUp(bottom_vecs_[layer_id], top_vecs_[layer_id]); //调用layer的SetUp函数,检查输入输出blob的个数,调整大小等
LOG_IF(INFO, Caffe::root_solver()) << "Setting up " << layer_names_[layer_id]; //打印
for (int top_id = 0; top_id < top_vecs_[layer_id].size(); ++top_id) {
if (blob_loss_weights_.size() <= top_id_vecs_[layer_id][top_id]) { //输出blob对应的blob_id比权重的个数大
blob_loss_weights_.resize(top_id_vecs_[layer_id][top_id] + 1, Dtype(0)); //权重的个数加1,初始为0
}
blob_loss_weights_[top_id_vecs_[layer_id][top_id]] = layer->loss(top_id); //根据layer中的权重设置
LOG_IF(INFO, Caffe::root_solver())
<< "Top shape: " << top_vecs_[layer_id][top_id]->shape_string(); //打印blob的形状信息
if (layer->loss(top_id)) {
LOG_IF(INFO, Caffe::root_solver()) << " with loss weight " << layer->loss(top_id); //loss权重不为0,也打印下
}
//TODO AppendTop()中同址计算的blob指针也会重复保存在top_vecs_中,那此处内存消耗的计算重复了???
memory_used_ += top_vecs_[layer_id][top_id]->count(); //累加输出blob中数据的个数
}
LOG_IF(INFO, Caffe::root_solver())
<< "Memory required for data: " << memory_used_ * sizeof(Dtype);
const int param_size = layer_param.param_size(); //layer参数中param参数的设置的个数
const int num_param_blobs = layers_[layer_id]->blobs().size(); //layer中可学习的参数blob的个数
CHECK_LE(param_size, num_param_blobs)
<< "Too many params specified for layer " << layer_param.name(); //设置的个数需小于等于参数blob的个数
ParamSpec default_param_spec; //layer中blob参数的默认配置,例如layer参数的lr_mult_=1和decay_mult_=1等
for (int param_id = 0; param_id < num_param_blobs; ++param_id) {
const ParamSpec* param_spec = (param_id < param_size) ?
&layer_param.param(param_id) : &default_param_spec; //使用设置的训练参数,或者使用默认配置
const bool param_need_backward = param_spec->lr_mult() != 0; //学习率系数不为0,则需要反向传播
need_backward |= param_need_backward; //如果layer的参数blob需要反传,则所在的layer也设置成需要
layers_[layer_id]->set_param_propagate_down(param_id,
param_need_backward); //将参数blob是否反传保存到net的layers_中
}
for (int param_id = 0; param_id < num_param_blobs; ++param_id) {
AppendParam(param, layer_id, param_id); //将layer的参数blob添加到net中
}
// Finally, set the backward flag
//设置layer是否需要反传(layer中输入blob和参数blob中任何一个需要,则该层都需要反传)
layer_need_backward_.push_back(need_backward);
if (need_backward) {
for (int top_id = 0; top_id < top_id_vecs_[layer_id].size(); ++top_id) {
//将layer中的输出blob也设置为需要反传
blob_need_backward_[top_id_vecs_[layer_id][top_id]] = true;
}
}
}
// Go through the net backwards to determine which blobs contribute to the
// loss. We can skip backward computation for blobs that don't contribute
// to the loss.
// Also checks if all bottom blobs don't need backward computation (possible
// because the skip_propagate_down param) and so we can skip backward
// computation for the entire layer
set blobs_under_loss; //参与loss计算的blob的名称的集合
set blobs_skip_backp; //可跳过反向传播的blob的名称的集合
for (int layer_id = layers_.size() - 1; layer_id >= 0; --layer_id) { //从后往前判断哪些layer对网络的loss计算有贡献
bool layer_contributes_loss = false; //初始设置为无用,可跳过
bool layer_skip_propagate_down = true;
for (int top_id = 0; top_id < top_vecs_[layer_id].size(); ++top_id) {
const string& blob_name = blob_names_[top_id_vecs_[layer_id][top_id]]; //第layer_id层的第top_id个输出blob的名称
if (layers_[layer_id]->loss(top_id) ||
(blobs_under_loss.find(blob_name) != blobs_under_loss.end())) { //该输出blob权重不为0,或者在参与计算loss的集合中能找到
layer_contributes_loss = true; //输出blob有用,则该layer设置为对loss计算有用
}
if (blobs_skip_backp.find(blob_name) == blobs_skip_backp.end()) { //该输出blob不在可跳过反传的集合中
layer_skip_propagate_down = false; //设置为不可跳过
}
//layer只要有任意一个输出blob参与loss计算,同时不可跳过反传步骤,则整个layer都是此状态,不必再判断其他输出blob了
if (layer_contributes_loss && !layer_skip_propagate_down)
break; //不用再判断其他输出blob
}
// If this layer can skip backward computation, also all his bottom blobs don't need backpropagation
//如果之前layer设置了需要反传,但是此处是可跳过反传的(此处可跳过,说明该层的所有输出blob均可跳过反传,均不参与loss计算)
if (layer_need_backward_[layer_id] && layer_skip_propagate_down) {
layer_need_backward_[layer_id] = false; //有限考虑此处的设置,设置为false
for (int bottom_id = 0; bottom_id < bottom_vecs_[layer_id].size();
++bottom_id) {
bottom_need_backward_[layer_id][bottom_id] = false; //所有输入blob也设置为不需要反传
}
}
//layer的输出blob对loss的计算无用,则layer设置为不需要反传(此处只设置了layer,后续才设置了输入参数不需要)
if (!layer_contributes_loss) { layer_need_backward_[layer_id] = false; }
if (Caffe::root_solver()) { //主线程打印该层是否需要反传的信息
if (layer_need_backward_[layer_id]) {
LOG(INFO) << layer_names_[layer_id] << " needs backward computation.";
} else {
LOG(INFO) << layer_names_[layer_id]
<< " does not need backward computation.";
}
}
for (int bottom_id = 0; bottom_id < bottom_vecs_[layer_id].size(); ++bottom_id) {
if (layer_contributes_loss) { //layer对loss计算有用
//把layer的所有输入blob的名称保存到blobs_under_loss中,表示与该输入blob对应的上层layer的
//输出blob也同样参与了loss的计算,保存名称方便设置这些上层layer的输出blob
const string& blob_name = blob_names_[bottom_id_vecs_[layer_id][bottom_id]];
blobs_under_loss.insert(blob_name);
} else {
bottom_need_backward_[layer_id][bottom_id] = false; //否则,设置输入blob为不需要反传
}
if (!bottom_need_backward_[layer_id][bottom_id]) { //当前输入blob不需要反传
//将名称保存在blobs_skip_backp中,表示与该输入blob对应的上层layer的输出blob可以跳过反传
const string& blob_name = blob_names_[bottom_id_vecs_[layer_id][bottom_id]];
blobs_skip_backp.insert(blob_name);
}
}
}
// Handle force_backward if needed.
//如果net参数中设置了需要强制反传,则根据layer的是否设置允许强制反传再设置一遍
if (param.force_backward()) {
for (int layer_id = 0; layer_id < layers_.size(); ++layer_id) {
layer_need_backward_[layer_id] = true; //默认需要
for (int bottom_id = 0; bottom_id < bottom_need_backward_[layer_id].size(); ++bottom_id) {
//输入blob设置了需要反传,或者该输入blob设置了允许强制反传
//参见layer.hpp,允许强制反传则优先遵从layer的强制设置,不允许的话则只考虑自身的设置
bottom_need_backward_[layer_id][bottom_id] =
bottom_need_backward_[layer_id][bottom_id] ||
layers_[layer_id]->AllowForceBackward(bottom_id);
//找到该输入blob在blobs_中的索引,并在blob_need_backward_中设置对应的值
blob_need_backward_[bottom_id_vecs_[layer_id][bottom_id]] =
blob_need_backward_[bottom_id_vecs_[layer_id][bottom_id]] ||
bottom_need_backward_[layer_id][bottom_id];
}
for (int param_id = 0; param_id < layers_[layer_id]->blobs().size(); ++param_id) {
layers_[layer_id]->set_param_propagate_down(param_id, true); //layer还是要遵守net的设置,全部置为需要反传
}
}
}
// In the end, all remaining blobs are considered output blobs.
//最后,将所有当前还未被当作输入的输出blob(保存在available_blobs中)全部认为是网络的输出
for (set::iterator it = available_blobs.begin();
it != available_blobs.end(); ++it) { //available_blobs中剩余的blob
LOG_IF(INFO, Caffe::root_solver())
<< "This network produces output " << *it; //打印信息
net_output_blobs_.push_back(blobs_[blob_name_to_idx[*it]].get()); //将这些输出blob在blobs_中的指针统一保存起来
net_output_blob_indices_.push_back(blob_name_to_idx[*it]); //保存在blobs_中的索引
}
for (size_t blob_id = 0; blob_id < blob_names_.size(); ++blob_id) {
blob_names_index_[blob_names_[blob_id]] = blob_id; //关联blob名称和在blobs_中的索引,方便由名称得到位置
}
for (size_t layer_id = 0; layer_id < layer_names_.size(); ++layer_id) {
layer_names_index_[layer_names_[layer_id]] = layer_id; //关联layer的名称和在layers_的索引,方便由名称得到位置
}
ShareWeights(); //设置共享参数的数据和梯度指针
debug_info_ = param.debug_info(); //是否计算和打印调试信息
LOG_IF(INFO, Caffe::root_solver()) << "Network initialization done."; //网路初始化完成
}
//根据网络设置的NetState类型状态和各层内部设置的NetStateRule,判断网络是否需要包含该层
//过滤掉这些层之后得到新的网络参数param_filtered
template
void Net::FilterNet(const NetParameter& param,
NetParameter* param_filtered) {
NetState net_state(param.state()); //读取网络的状态
param_filtered->CopyFrom(param); //先从param中拷贝所有数据
param_filtered->clear_layer(); //然后清空所有layer
for (int i = 0; i < param.layer_size(); ++i) { //所有layer一个一个进行判断
const LayerParameter& layer_param = param.layer(i); //当前layer的参数(LayerParameter类型的消息)
const string& layer_name = layer_param.name(); //layer的名称
//检查,include和exclude不能同时设置,参见caffe.proto中的LayerParameter
//include和exclude均为NetStateRule类型的消息,用于设置layer的状态
CHECK(layer_param.include_size() == 0 || layer_param.exclude_size() == 0)
<< "Specify either include rules or exclude rules; not both.";
// If no include rules are specified, the layer is included by default and
// only excluded if it meets one of the exclude rules.
bool layer_included = (layer_param.include_size() == 0); //不包含included消息 //include和exclude都不设置时默认包含该层
for (int j = 0; layer_included && j < layer_param.exclude_size(); ++j) {
if (StateMeetsRule(net_state, layer_param.exclude(j), layer_name)) { //检查网络参数是否满足layer的exclude参数的要求
layer_included = false; //满足layer设置的任意一个exclude,则net运行时不能包含该层layer
}
}
for (int j = 0; !layer_included && j < layer_param.include_size(); ++j) {
if (StateMeetsRule(net_state, layer_param.include(j), layer_name)) {
layer_included = true; //满足layer设置的任意一个included,则net运行时需要包含该层layer
}
}
if (layer_included) {
param_filtered->add_layer()->CopyFrom(layer_param); //将该层的参数加入到网络参数中
}
}
}
//检查state(NetParameter中的参数)中的设置是否满足rule(LayerParameter中的参数)中的所有规则
//网络的phase必须与layer的phase相同,网络的level必须在layer的min_level和max_level之间
//网络的stage必须包含layer的所有stage字符串,不能包含not_stage中的任意一个字符串,这些都满足才说明网络的设置满足layer的要求
template
bool Net::StateMeetsRule(const NetState& state,
const NetStateRule& rule, const string& layer_name) {
// Check whether the rule is broken due to phase.
if (rule.has_phase()) { //rule中设置了phase,那么state中的字段必须与之相等才符合
if (rule.phase() != state.phase()) {
LOG_IF(INFO, Caffe::root_solver())
<< "The NetState phase (" << state.phase()
<< ") differed from the phase (" << rule.phase()
<< ") specified by a rule in layer " << layer_name; //不等,记录信息
return false;
}
}
// Check whether the rule is broken due to min level.
if (rule.has_min_level()) { //state中的level必须大于rule中设置的最小level
if (state.level() < rule.min_level()) {
LOG_IF(INFO, Caffe::root_solver())
<< "The NetState level (" << state.level()
<< ") is above the min_level (" << rule.min_level()
<< ") specified by a rule in layer " << layer_name; //不满足,记录信息
return false;
}
}
// Check whether the rule is broken due to max level.
if (rule.has_max_level()) { //state中的level必须小于rule中设置的最大level
if (state.level() > rule.max_level()) {
LOG_IF(INFO, Caffe::root_solver())
<< "The NetState level (" << state.level()
<< ") is above the max_level (" << rule.max_level()
<< ") specified by a rule in layer " << layer_name;
return false;
}
}
// Check whether the rule is broken due to stage. The NetState must
// contain ALL of the rule's stages to meet it.
for (int i = 0; i < rule.stage_size(); ++i) { //rule中设置了一堆stage字符串
// Check that the NetState contains the rule's ith stage.
bool has_stage = false;
for (int j = 0; !has_stage && j < state.stage_size(); ++j) {
if (rule.stage(i) == state.stage(j)) { has_stage = true; } //检查rule.stage(i)是否在state.stage(...)中
}
if (!has_stage) { //state中必须包含rule中的所有stage,否则返回false
LOG_IF(INFO, Caffe::root_solver())
<< "The NetState did not contain stage '" << rule.stage(i)
<< "' specified by a rule in layer " << layer_name;
return false;
}
}
// Check whether the rule is broken due to not_stage. The NetState must
// contain NONE of the rule's not_stages to meet it.
for (int i = 0; i < rule.not_stage_size(); ++i) { //rule中设置了一堆not_stage字符串
// Check that the NetState contains the rule's ith not_stage.
bool has_stage = false;
for (int j = 0; !has_stage && j < state.stage_size(); ++j) {
if (rule.not_stage(i) == state.stage(j)) { has_stage = true; } //检查rule.not_stage(i)是否在state.stage(...)中
}
if (has_stage) { //state中不能包含rule中的任何一个not_stage,否则返回false
LOG_IF(INFO, Caffe::root_solver())
<< "The NetState contained a not_stage '" << rule.not_stage(i)
<< "' specified by a rule in layer " << layer_name;
return false;
}
}
return true; //全部满足
}
//根据第layer_id层的第top_id个输出blob,创建一个空的blob保存到网络参数param中的blobs_中,同址计算则不创建
//并且无论是否为同址计算,将该blob数据的指针保存到top_vecs_中,将其在blobs_中的索引保存在top_id_vecs_中
// Helper for Net::Init: add a new top blob to the net.
template
void Net::AppendTop(const NetParameter& param, const int layer_id,
const int top_id, set* available_blobs,
map* blob_name_to_idx) {
shared_ptr layer_param(new LayerParameter(param.layer(layer_id))); //拷贝当前layer的参数
const string& blob_name = (layer_param->top_size() > top_id) ?
layer_param->top(top_id) : "(automatic)"; //输出blob的名称,或者用"(automatic)"(匿名blob)替代
// Check if we are doing in-place computation
if (blob_name_to_idx && layer_param->bottom_size() > top_id &&
blob_name == layer_param->bottom(top_id)) { //当前输出blob的名称与对应位置的输入blob的名称相同,为同址计算
// In-place computation
LOG_IF(INFO, Caffe::root_solver())
<< layer_param->name() << " -> " << blob_name << " (in-place)"; //在主线程中打印信息
//(*blob_name_to_idx)[blob_name]为该名称对应的blob数据在blobs_中的索引
//top_vecs_中会保存第layer_id层第top_id个输出blob的指针位置,top_id_vecs_保存索引
top_vecs_[layer_id].push_back(blobs_[(*blob_name_to_idx)[blob_name]].get());
top_id_vecs_[layer_id].push_back((*blob_name_to_idx)[blob_name]);
} else if (blob_name_to_idx &&
blob_name_to_idx->find(blob_name) != blob_name_to_idx->end()) {
//blob_name_to_idx只会存储输出blob的名称和位置,此处找到同名的输出blob,说明layer的输出blob的名称设置有问题
// If we are not doing in-place computation but have duplicated blobs,
// raise an error.
LOG(FATAL) << "Top blob '" << blob_name << "' produced by multiple sources.";
} else { //基操勿六
// Normal output.
if (Caffe::root_solver()) {
LOG(INFO) << layer_param->name() << " -> " << blob_name; //主线程中打印信息
}
//出现普通的输出blob,将其信息存储到Net中
shared_ptr > blob_pointer(new Blob()); //创建一个空的blob,返回其指针
const int blob_id = blobs_.size();
blobs_.push_back(blob_pointer); //blob_id为该输出blob在blobs_中的索引,将指针存入
blob_names_.push_back(blob_name); //设置名称
blob_need_backward_.push_back(false); //初始设置为不需要反传
if (blob_name_to_idx) { (*blob_name_to_idx)[blob_name] = blob_id; } //保存名称与位置的映射关系
top_id_vecs_[layer_id].push_back(blob_id); //top_id_vecs_[layer_id][top_id]为其位置索引
top_vecs_[layer_id].push_back(blob_pointer.get()); //top_vecs_[layer_id][top_id]为其blob指针
}
//该输出blob可被后续的layer当成输入,用于AppendBottom()中
if (available_blobs) { available_blobs->insert(blob_name); }
}
//将第layer_id层的第bottom_id个输入blob加入到网络参数param中的对应layer的参数中
// Helper for Net::Init: add a new bottom blob to the net.
template
int Net::AppendBottom(const NetParameter& param, const int layer_id,
const int bottom_id, set* available_blobs,
map* blob_name_to_idx) {
const LayerParameter& layer_param = param.layer(layer_id); //当前层的参数
const string& blob_name = layer_param.bottom(bottom_id); //输入blob的名称
//available_blobs存放着第0至layer_id-1层的layer中所有还未被当作输入的输出blob的名称,
//在这里面找不到该层的输入blob的名称,说明该输入blob在之前层中找不到对应的输出blob,报错
if (available_blobs->find(blob_name) == available_blobs->end()) {
LOG(FATAL) << "Unknown bottom blob '" << blob_name << "' (layer '"
<< layer_param.name() << "', bottom index " << bottom_id << ")";
}
const int blob_id = (*blob_name_to_idx)[blob_name]; //名称得到其在blobs_中的索引id
LOG_IF(INFO, Caffe::root_solver()) << layer_names_[layer_id] << " <- " << blob_name; //主线程中打印信息
//每个输入blob都会调用一次AppendBottom(),所以bottom_vecs_[layer_id][bottom_id]存放着当前输入blob的指针
bottom_vecs_[layer_id].push_back(blobs_[blob_id].get()); //将其在blobs_中的数据指针保存在bottom_vecs_中
bottom_id_vecs_[layer_id].push_back(blob_id); //bottom_id_vecs_[layer_id][bottom_id]保存该输入blob在blobs_中的索引
available_blobs->erase(blob_name); //blob_name对应的输出blob已被当作输入blob,在available_blobs中删除
bool need_backward = blob_need_backward_[blob_id]; //blobs_[blob_id]是否需要反向传播
// Check if the backpropagation on bottom_id should be skipped
if (layer_param.propagate_down_size() > 0) { //该层设置了是否需要反向传播参数
need_backward = layer_param.propagate_down(bottom_id); //读取设置中第bottom_id个的值
}
//bottom_need_backward_[layer_id][bottom_id]保存该输入blob是否需要反向传播
bottom_need_backward_[layer_id].push_back(need_backward);
return blob_id; //返回该输入blob在blobs_中的位置
}
//将第layer_id层的第param_id个参数数据保存到网络中,并保存一些用于索引的值
//对于非共享的参数,还会将数据指针保存在learnable_params_中
//learnable_param_ids_保存第第layer_id层的第param_id个参数在learnable_params_的索引,共享参数保存其来源参数的索引
template
void Net::AppendParam(const NetParameter& param, const int layer_id,
const int param_id) {
const LayerParameter& layer_param = layers_[layer_id]->layer_param(); //layer的参数
const int param_size = layer_param.param_size(); //layer参数中param参数的设置的个数
//按照caffe.proto文件中ParamSpec消息的说明,当需要在layer之间共享参数时,可以param_name会是一个非空的名称
string param_name =
(param_size > param_id) ? layer_param.param(param_id).name() : ""; //参数blob的名称或者默认的""
if (param_name.size()) {
param_display_names_.push_back(param_name); //非空,保存其名称,方便其他layer查询
} else {
ostringstream param_display_name;
param_display_name << param_id;
param_display_names_.push_back(param_display_name.str()); //使用param_id作为其名称
}
const int net_param_id = params_.size(); //net_param_id为当前参数在net的params_中的索引
params_.push_back(layers_[layer_id]->blobs()[param_id]); //将第layer_id层的第param_id个参数数据保存在params_中
param_id_vecs_[layer_id].push_back(net_param_id); //param_id_vecs_[layer_id][param_id]为在params_中的索引net_param_id
//param_layer_indices_[net_param_id]为参数在网络中的位置(layer_id, param_id)
param_layer_indices_.push_back(make_pair(layer_id, param_id));
ParamSpec default_param_spec; //layer中blob参数的默认训练参数,lr_mult_=1和decay_mult_=1等
const ParamSpec* param_spec = (layer_param.param_size() > param_id) ?
&layer_param.param(param_id) : &default_param_spec; //使用设置的或者默认配置
if (!param_size || !param_name.size() || (param_name.size() &&
param_names_index_.find(param_name) == param_names_index_.end())) {
//layer中没有配置param参数,或者param参数中没有设置名称,或者设置了名称但是还不在param_names_index_中
//(设置了名称但还不在param_names_index_中,说明该参数首次出现,是源参数,但是需要共享给其他的layer)
// This layer "owns" this parameter blob -- it is either anonymous
// (i.e., not given a param_name) or explicitly given a name that we
// haven't already seen.
param_owners_.push_back(-1); //参数来自自身的layer,源参数,设置为-1
if (param_name.size()) { //param中设置了名称,说明需要共享给其他layer的某个参数
param_names_index_[param_name] = net_param_id; //将参数名称param_name与在params_的位置net_param_id关联起来
}
const int learnable_param_id = learnable_params_.size(); //当前参数在learnable_params_中的索引
learnable_params_.push_back(params_[net_param_id].get()); //保存参数指针
learnable_param_ids_.push_back(learnable_param_id); //保存索引
has_params_lr_.push_back(param_spec->has_lr_mult()); //保存参数是否设置了对应的学习率系数
has_params_decay_.push_back(param_spec->has_decay_mult()); //保存参数是否设置了对应的衰减率系数
params_lr_.push_back(param_spec->lr_mult()); //保存参数的学习率系数
params_weight_decay_.push_back(param_spec->decay_mult()); //保存参数的衰减率系数
} else {
//说明该参数来源于其他的layer参数,为共享参数
//共享参数配置的param中的name必须与源参数的param中的name相同
// Named param blob with name we've seen before: share params
const int owner_net_param_id = param_names_index_[param_name]; //先找到源参数在net的params_的索引
param_owners_.push_back(owner_net_param_id); //保存源参数的索引
const pair& owner_index =
param_layer_indices_[owner_net_param_id]; //得到源参数在网络中的位置(layer_id, param_id)
const int owner_layer_id = owner_index.first; //源参数所在的层
const int owner_param_id = owner_index.second; //源参数所在的位置
LOG_IF(INFO, Caffe::root_solver()) << "Sharing parameters '" << param_name
<< "' owned by "
<< "layer '" << layer_names_[owner_layer_id] << "', param "
<< "index " << owner_param_id; //打印信息
Blob* this_blob = layers_[layer_id]->blobs()[param_id].get(); //当前参数的blob指针,第layer_id层第param_id个
Blob* owner_blob =
layers_[owner_layer_id]->blobs()[owner_param_id].get(); //源参数在net中的blob指针,第owner_layer_id层第owner_param_id个
const int param_size = layer_param.param_size();
if (param_size > param_id && (layer_param.param(param_id).share_mode() ==
ParamSpec_DimCheckMode_PERMISSIVE)) {
//检查参数共享的模式,PERMISSIVE模式下只要求两个blob的总数据个数相同
// Permissive dimension checking -- only check counts are the same.
CHECK_EQ(this_blob->count(), owner_blob->count())
<< "Cannot share param '" << param_name << "' owned by layer '"
<< layer_names_[owner_layer_id] << "' with layer '"
<< layer_names_[layer_id] << "'; count mismatch. Owner layer param "
<< "shape is " << owner_blob->shape_string() << "; sharing layer "
<< "shape is " << this_blob->shape_string(); //检查数据的总个数是否相等
} else {
//STRICT模式下要求两个参数blob的数据的各个维度值都相等
// Strict dimension checking -- all dims must be the same.
CHECK(this_blob->shape() == owner_blob->shape())
<< "Cannot share param '" << param_name << "' owned by layer '"
<< layer_names_[owner_layer_id] << "' with layer '"
<< layer_names_[layer_id] << "'; shape mismatch. Owner layer param "
<< "shape is " << owner_blob->shape_string() << "; sharing layer "
<< "expects shape " << this_blob->shape_string(); //检查形状是否相同
}
//owner_net_param_id虽然是源参数在params_的索引,但是每次调用AppendParam()时,params_与learnable_param_ids_
//都会存入一个值,他们的大小一致,是逐个对应的(注意learnable_params_与params_不是),所以他们之间的索引可以通用
const int learnable_param_id = learnable_param_ids_[owner_net_param_id]; //得到参数在learnable_params_中的索引
learnable_param_ids_.push_back(learnable_param_id); //将源参数的索引当成当前的共享参数的索引,保存起来
if (param_spec->has_lr_mult()) { //当前的共享参数配置的param中有设置学习率系数
if (has_params_lr_[learnable_param_id]) { //源参数也配置了学习率系数
CHECK_EQ(param_spec->lr_mult(), params_lr_[learnable_param_id])
<< "Shared param '" << param_name << "' has mismatched lr_mult."; //要求两者相等,不然参数共享时更新的步长不一致
} else {
has_params_lr_[learnable_param_id] = true; //源参数未设置,则将共享参数的设置保存到源参数的设置中
params_lr_[learnable_param_id] = param_spec->lr_mult();
}
}
if (param_spec->has_decay_mult()) { //权重衰减系数同样处理
if (has_params_decay_[learnable_param_id]) {
CHECK_EQ(param_spec->decay_mult(),
params_weight_decay_[learnable_param_id])
<< "Shared param '" << param_name << "' has mismatched decay_mult."; //要求相等
} else {
has_params_decay_[learnable_param_id] = true;
params_weight_decay_[learnable_param_id] = param_spec->decay_mult(); //配置之前未设置的源参数
}
}
}
}
net.hpp源码
/// @brief The network name
string name_; //网络的名称
/// @brief The phase: TRAIN or TEST
Phase phase_; //网络的字段状态(TRAIN or TEST)
/// @brief Individual layers in the net
vector > > layers_; //网络的各个层,layers_[layer_id]
vector layer_names_; //网络的各个层的名称
map layer_names_index_; //
vector layer_need_backward_; //网络的各个层是否需要反向传播
/// @brief the blobs storing intermediate results between the layer.
vector > > blobs_; //存储网络的所有输出blob,blobs_[blob_id]
vector blob_names_; //blob_names_[blob_id]为blobs_[blob_id]存储的blob数据的名称
map blob_names_index_; //
vector blob_need_backward_; //blobs_中的各个数据是否需要反向传播
/// bottom_vecs stores the vectors containing the input for each layer.
/// They don't actually host the blobs (blobs_ does), so we simply store
/// pointers.
vector*> > bottom_vecs_; //bottom_vecs_[i][j]为第i层第j个输入blob数据的指针,数据在blobs_中
vector > bottom_id_vecs_; //bottom_id_vecs_[i][j]为第i层第j个输入blob在blobs_中的索引blob_id
vector > bottom_need_backward_; //bottom_need_backward_[i][j]为第i层第j个输入blob是否需要反传
/// top_vecs stores the vectors containing the output for each layer
vector*> > top_vecs_; //top_vecs_[i][j]为第i层第j个输出blob数据的指针,数据在blobs_中
vector > top_id_vecs_; //top_id_vecs_[i][j]为第i层第j个输出blob在blobs_中的索引blob_id
/// Vector of weight in the loss (or objective) function of each net blob, indexed by blob_id.
vector blob_loss_weights_; //blobs_[blob_id]对应的loss权重
vector > param_id_vecs_; //param_id_vecs_[i][j]为第i层第j个参数blob在params_中的位置
//param_owners_[i]为参数params_[i]的拥有者的id,源参数(参数来源于所在的layer)的值为-1,而共享参数的值为源参数在params_的索引
vector param_owners_;
//param_display_names_[i]为参数params_[i]显示的名称,需共享的参数会设置非一个非空的值,未设置则用参数的id替代
//由于AppendParam()中参数名称未设置时会用param_id替代,所以该变量中可能会存在重复的名称.因此该变量不可用于寻找或匹配与共享参数同名的源参数的位置
vector param_display_names_;
vector > param_layer_indices_; //param_layer_indices_[i]为参数params_[i]在网络中的位置(layer_id, param_id)
//<参数名称, 参数在params_的索引>,非空名称的参数只有在首次出现时才会保存在该变量中,后续出现都作为共享参数,所以该变量可用于寻找和匹配共享参数的源参数的位置
map param_names_index_;
/// blob indices for the input and the output of the net
vector net_input_blob_indices_; //net_input_blob_indices_[i]为网络的数据输入blob在blobs_中的索引
vector net_output_blob_indices_; //net_output_blob_indices_[i]为网络的数据输出blob在blobs_中的索引
vector*> net_input_blobs_; //网络的数据输入blob的指针
vector*> net_output_blobs_; //网络的数据输出blob的指针(所有未被当作其他层输入的输出blob)
/// The parameters in the network.
vector > > params_; //网络中的所有参数blob(包括源参数和共享参数)
vector*> learnable_params_; //网络中的所有学习参数,只会保存源参数
/**
* The mapping from params_ -> learnable_params_: we have
* learnable_param_ids_.size() == params_.size(),
* and learnable_params_[learnable_param_ids_[i]] == params_[i].get()
* if and only if params_[i] is an "owner"; otherwise, params_[i] is a sharer
* and learnable_params_[learnable_param_ids_[i]] gives its owner.
*/
vector learnable_param_ids_; //learnable_param_ids_[i]为参数params_[i]在learnable_params_中的索引
/// the learning rate multipliers for learnable_params_
vector params_lr_; //params_lr_[i]为参数learnable_params_[i]的学习率系数
vector has_params_lr_; //has_params_lr_[i]为参数learnable_params_[i]是否设置了学习率系数
/// the weight decay multipliers for learnable_params_
vector params_weight_decay_; //params_weight_decay_[i]为参数learnable_params_[i]的权重衰减系数
vector has_params_decay_; //has_params_decay_[i]为参数learnable_params_[i]是否设置了权重衰减系数
/// The bytes of memory used by this net
size_t memory_used_; //网络中所有输出blob的大小之和
/// Whether to compute and display debug info for the net.
bool debug_info_; //网络是否允许计算并打印调试信息,允许的话则会打印参与前向计算的layer的输出blob和参数blob的数据的绝对值之和的均值
// Callbacks
vector before_forward_; //网络执行前向计算之前的回调函数
vector after_forward_; //网络执行前向计算之后的回调函数
vector before_backward_; //网络执行反向传播之前的回调函数
vector after_backward_; //网络执行反向传播之后的回调函数
小结
- 初始化函数中的
available_blobs
,是指目前还未被当成输入blob使用的输出blob。在经过InsertSplits()操作后,除去部分被用于计算网络loss的输出blob外,剩余的输出blob与输入blob是一一对应的。所以可使用该变量快速确定某层的输入blob是否有效,以及网络的所有输出blob。 - 网络的所有输出blob都会保存在
blobs_
中,网络各层之间的数据传递依赖此变量进行。但初始化的时候只会创建一个空指针,实际训练时才会将数据填充进去。 - 网络的各Layer是否需要反传,首先是根据输入数据和参数数据的设置来判断该层是否需要反传,然后是根据该层的输出数据是否参与了网络loss的计算来判断,最后是根据是否设置了强制需要反传参数来确定,越往后的判断的优先级越高。
Caffe的源码笔者是第一次阅读,一边阅读一边记录,对代码的理解和分析可能会存在错误或遗漏,希望各位读者批评指正,谢谢支持!