tools/python: add script to convert TensorFlow model (.pb) to native model (.model)
For example, given TensorFlow model file espcn.pb,
to generate native model file espcn.model, just run:
python convert.py espcn.pb
In current implementation, the native model file is generated for
specific dnn network with hard-code python scripts maintained out of ffmpeg.
For example, srcnn network used by vf_sr is generated with
https://github.com/HighVoltageRocknRoll/sr/blob/master/generate_header_and_model.py#L85
In this patch, the script is designed as a general solution which
converts general TensorFlow model .pb file into .model file. The script
now has some tricky to be compatible with current implemention, will
be refined step by step.
The script is also added into ffmpeg source tree. It is expected there
will be many more patches and community needs the ownership of it.
Another technical direction is to do the conversion in c/c++ code within
ffmpeg source tree. While .pb file is organized with protocol buffers,
it is not easy to do such work with tiny c/c++ code, see more discussion
at http://ffmpeg.org/pipermail/ffmpeg-devel/2019-May/244496.html. So,
choose the python script.
Signed-off-by: Guo, Yejun <yejun.guo@intel.com>
6 years ago
|
|
|
# Copyright (c) 2019 Guo Yejun
|
|
|
|
#
|
|
|
|
# This file is part of FFmpeg.
|
|
|
|
#
|
|
|
|
# FFmpeg is free software; you can redistribute it and/or
|
|
|
|
# modify it under the terms of the GNU Lesser General Public
|
|
|
|
# License as published by the Free Software Foundation; either
|
|
|
|
# version 2.1 of the License, or (at your option) any later version.
|
|
|
|
#
|
|
|
|
# FFmpeg is distributed in the hope that it will be useful,
|
|
|
|
# but WITHOUT ANY WARRANTY; without even the implied warranty of
|
|
|
|
# MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the GNU
|
|
|
|
# Lesser General Public License for more details.
|
|
|
|
#
|
|
|
|
# You should have received a copy of the GNU Lesser General Public
|
|
|
|
# License along with FFmpeg; if not, write to the Free Software
|
|
|
|
# Foundation, Inc., 51 Franklin Street, Fifth Floor, Boston, MA 02110-1301 USA
|
|
|
|
# ==============================================================================
|
|
|
|
|
|
|
|
import tensorflow as tf
|
|
|
|
import numpy as np
|
|
|
|
import sys, struct
|
|
|
|
import convert_header as header
|
tools/python: add script to convert TensorFlow model (.pb) to native model (.model)
For example, given TensorFlow model file espcn.pb,
to generate native model file espcn.model, just run:
python convert.py espcn.pb
In current implementation, the native model file is generated for
specific dnn network with hard-code python scripts maintained out of ffmpeg.
For example, srcnn network used by vf_sr is generated with
https://github.com/HighVoltageRocknRoll/sr/blob/master/generate_header_and_model.py#L85
In this patch, the script is designed as a general solution which
converts general TensorFlow model .pb file into .model file. The script
now has some tricky to be compatible with current implemention, will
be refined step by step.
The script is also added into ffmpeg source tree. It is expected there
will be many more patches and community needs the ownership of it.
Another technical direction is to do the conversion in c/c++ code within
ffmpeg source tree. While .pb file is organized with protocol buffers,
it is not easy to do such work with tiny c/c++ code, see more discussion
at http://ffmpeg.org/pipermail/ffmpeg-devel/2019-May/244496.html. So,
choose the python script.
Signed-off-by: Guo, Yejun <yejun.guo@intel.com>
6 years ago
|
|
|
|
|
|
|
__all__ = ['convert_from_tensorflow']
|
|
|
|
|
|
|
|
class Operand(object):
|
|
|
|
IOTYPE_INPUT = 1
|
|
|
|
IOTYPE_OUTPUT = 2
|
|
|
|
IOTYPE_INTERMEDIATE = IOTYPE_INPUT | IOTYPE_OUTPUT
|
|
|
|
DTYPE_FLOAT = 1
|
|
|
|
DTYPE_UINT8 = 4
|
|
|
|
index = 0
|
|
|
|
def __init__(self, name, dtype, dims):
|
|
|
|
self.name = name
|
|
|
|
self.dtype = dtype
|
|
|
|
self.dims = dims
|
|
|
|
self.iotype = 0
|
|
|
|
self.used_count = 0
|
|
|
|
self.index = Operand.index
|
|
|
|
Operand.index = Operand.index + 1
|
|
|
|
self.iotype2str = {Operand.IOTYPE_INPUT: 'in', Operand.IOTYPE_OUTPUT: 'out', Operand.IOTYPE_INTERMEDIATE: 'inout'}
|
|
|
|
self.dtype2str = {Operand.DTYPE_FLOAT: 'DT_FLOAT', Operand.DTYPE_UINT8: 'DT_UINT8'}
|
|
|
|
|
|
|
|
def add_iotype(self, iotype):
|
|
|
|
self.iotype = self.iotype | iotype
|
|
|
|
if iotype == Operand.IOTYPE_INPUT:
|
|
|
|
self.used_count = self.used_count + 1
|
|
|
|
|
|
|
|
def __str__(self):
|
|
|
|
return "{}: (name: {}, iotype: {}, dtype: {}, dims: {}, used_count: {})".format(self.index,
|
|
|
|
self.name, self.iotype2str[self.iotype], self.dtype2str[self.dtype],
|
|
|
|
self.dims, self.used_count)
|
|
|
|
|
|
|
|
def __lt__(self, other):
|
|
|
|
return self.index < other.index
|
|
|
|
|
tools/python: add script to convert TensorFlow model (.pb) to native model (.model)
For example, given TensorFlow model file espcn.pb,
to generate native model file espcn.model, just run:
python convert.py espcn.pb
In current implementation, the native model file is generated for
specific dnn network with hard-code python scripts maintained out of ffmpeg.
For example, srcnn network used by vf_sr is generated with
https://github.com/HighVoltageRocknRoll/sr/blob/master/generate_header_and_model.py#L85
In this patch, the script is designed as a general solution which
converts general TensorFlow model .pb file into .model file. The script
now has some tricky to be compatible with current implemention, will
be refined step by step.
The script is also added into ffmpeg source tree. It is expected there
will be many more patches and community needs the ownership of it.
Another technical direction is to do the conversion in c/c++ code within
ffmpeg source tree. While .pb file is organized with protocol buffers,
it is not easy to do such work with tiny c/c++ code, see more discussion
at http://ffmpeg.org/pipermail/ffmpeg-devel/2019-May/244496.html. So,
choose the python script.
Signed-off-by: Guo, Yejun <yejun.guo@intel.com>
6 years ago
|
|
|
class TFConverter:
|
|
|
|
def __init__(self, graph_def, nodes, outfile, dump4tb):
|
tools/python: add script to convert TensorFlow model (.pb) to native model (.model)
For example, given TensorFlow model file espcn.pb,
to generate native model file espcn.model, just run:
python convert.py espcn.pb
In current implementation, the native model file is generated for
specific dnn network with hard-code python scripts maintained out of ffmpeg.
For example, srcnn network used by vf_sr is generated with
https://github.com/HighVoltageRocknRoll/sr/blob/master/generate_header_and_model.py#L85
In this patch, the script is designed as a general solution which
converts general TensorFlow model .pb file into .model file. The script
now has some tricky to be compatible with current implemention, will
be refined step by step.
The script is also added into ffmpeg source tree. It is expected there
will be many more patches and community needs the ownership of it.
Another technical direction is to do the conversion in c/c++ code within
ffmpeg source tree. While .pb file is organized with protocol buffers,
it is not easy to do such work with tiny c/c++ code, see more discussion
at http://ffmpeg.org/pipermail/ffmpeg-devel/2019-May/244496.html. So,
choose the python script.
Signed-off-by: Guo, Yejun <yejun.guo@intel.com>
6 years ago
|
|
|
self.graph_def = graph_def
|
|
|
|
self.nodes = nodes
|
|
|
|
self.outfile = outfile
|
|
|
|
self.dump4tb = dump4tb
|
tools/python: add script to convert TensorFlow model (.pb) to native model (.model)
For example, given TensorFlow model file espcn.pb,
to generate native model file espcn.model, just run:
python convert.py espcn.pb
In current implementation, the native model file is generated for
specific dnn network with hard-code python scripts maintained out of ffmpeg.
For example, srcnn network used by vf_sr is generated with
https://github.com/HighVoltageRocknRoll/sr/blob/master/generate_header_and_model.py#L85
In this patch, the script is designed as a general solution which
converts general TensorFlow model .pb file into .model file. The script
now has some tricky to be compatible with current implemention, will
be refined step by step.
The script is also added into ffmpeg source tree. It is expected there
will be many more patches and community needs the ownership of it.
Another technical direction is to do the conversion in c/c++ code within
ffmpeg source tree. While .pb file is organized with protocol buffers,
it is not easy to do such work with tiny c/c++ code, see more discussion
at http://ffmpeg.org/pipermail/ffmpeg-devel/2019-May/244496.html. So,
choose the python script.
Signed-off-by: Guo, Yejun <yejun.guo@intel.com>
6 years ago
|
|
|
self.layer_number = 0
|
|
|
|
self.output_names = []
|
|
|
|
self.name_node_dict = {}
|
|
|
|
self.edges = {}
|
|
|
|
self.conv_activations = {'Relu':0, 'Tanh':1, 'Sigmoid':2, 'None':3, 'LeakyRelu':4}
|
|
|
|
self.conv_paddings = {'VALID':0, 'SAME':1}
|
|
|
|
self.pool_paddings = {'VALID':0, 'SAME':1}
|
tools/python: add script to convert TensorFlow model (.pb) to native model (.model)
For example, given TensorFlow model file espcn.pb,
to generate native model file espcn.model, just run:
python convert.py espcn.pb
In current implementation, the native model file is generated for
specific dnn network with hard-code python scripts maintained out of ffmpeg.
For example, srcnn network used by vf_sr is generated with
https://github.com/HighVoltageRocknRoll/sr/blob/master/generate_header_and_model.py#L85
In this patch, the script is designed as a general solution which
converts general TensorFlow model .pb file into .model file. The script
now has some tricky to be compatible with current implemention, will
be refined step by step.
The script is also added into ffmpeg source tree. It is expected there
will be many more patches and community needs the ownership of it.
Another technical direction is to do the conversion in c/c++ code within
ffmpeg source tree. While .pb file is organized with protocol buffers,
it is not easy to do such work with tiny c/c++ code, see more discussion
at http://ffmpeg.org/pipermail/ffmpeg-devel/2019-May/244496.html. So,
choose the python script.
Signed-off-by: Guo, Yejun <yejun.guo@intel.com>
6 years ago
|
|
|
self.converted_nodes = set()
|
|
|
|
self.conv2d_scope_names = set()
|
|
|
|
self.conv2d_scopename_inputname_dict = {}
|
|
|
|
self.dense_scope_names = set()
|
|
|
|
self.dense_scopename_inputname_dict = {}
|
|
|
|
self.op2code = {'Conv2D':1, 'DepthToSpace':2, 'MirrorPad':3, 'Maximum':4,
|
|
|
|
'MathBinary':5, 'MathUnary':6, 'AvgPool':7, 'MatMul':8}
|
|
|
|
self.mathbin2code = {'Sub':0, 'Add':1, 'Mul':2, 'RealDiv':3, 'Minimum':4, 'FloorMod':5}
|
dnn_backend_native_layer_mathunary: add ceil support
It can be tested with the model generated with below python script:
import tensorflow as tf
import os
import numpy as np
import imageio
from tensorflow.python.framework import graph_util
name = 'ceil'
pb_file_path = os.getcwd()
if not os.path.exists(pb_file_path+'/{}_savemodel/'.format(name)):
os.mkdir(pb_file_path+'/{}_savemodel/'.format(name))
with tf.Session(graph=tf.Graph()) as sess:
in_img = imageio.imread('detection.jpg')
in_img = in_img.astype(np.float32)
in_data = in_img[np.newaxis, :]
input_x = tf.placeholder(tf.float32, shape=[1, None, None, 3], name='dnn_in')
y = tf.math.ceil( input_x, name='dnn_out')
sess.run(tf.global_variables_initializer())
constant_graph = graph_util.convert_variables_to_constants(sess, sess.graph_def, ['dnn_out'])
with tf.gfile.FastGFile(pb_file_path+'/{}_savemodel/model.pb'.format(name), mode='wb') as f:
f.write(constant_graph.SerializeToString())
print("model.pb generated, please in ffmpeg path use\n \n \
python tools/python/convert.py ceil_savemodel/model.pb --outdir=ceil_savemodel/ \n \n \
to generate model.model\n")
output = sess.run(y, feed_dict={ input_x: in_data})
imageio.imsave("out.jpg", np.squeeze(output))
print("To verify, please ffmpeg path use\n \n \
./ffmpeg -i detection.jpg -vf format=rgb24,dnn_processing=model=ceil_savemodel/model.pb:input=dnn_in:output=dnn_out:dnn_backend=tensorflow -f framemd5 ceil_savemodel/tensorflow_out.md5\n \n \
to generate output result of tensorflow model\n")
print("To verify, please ffmpeg path use\n \n \
./ffmpeg -i detection.jpg -vf format=rgb24,dnn_processing=model=ceil_savemodel/model.model:input=dnn_in:output=dnn_out:dnn_backend=native -f framemd5 ceil_savemodel/native_out.md5\n \n \
to generate output result of native model\n")
Signed-off-by: Mingyu Yin <mingyu.yin@intel.com>
Reviewed-by: Guo, Yejun <yejun.guo@intel.com>
4 years ago
|
|
|
self.mathun2code = {'Abs':0, 'Sin':1, 'Cos':2, 'Tan':3, 'Asin':4,
|
|
|
|
'Acos':5, 'Atan':6, 'Sinh':7, 'Cosh':8, 'Tanh':9, 'Asinh':10,
|
|
|
|
'Acosh':11, 'Atanh':12, 'Ceil':13, 'Floor':14, 'Round':15}
|
|
|
|
self.mirrorpad_mode = {'CONSTANT':0, 'REFLECT':1, 'SYMMETRIC':2}
|
|
|
|
self.name_operand_dict = {}
|
|
|
|
|
|
|
|
|
|
|
|
def add_operand(self, name, type):
|
|
|
|
node = self.name_node_dict[name]
|
|
|
|
if name not in self.name_operand_dict:
|
|
|
|
dtype = node.attr['dtype'].type
|
|
|
|
if dtype == 0:
|
|
|
|
dtype = node.attr['T'].type
|
|
|
|
dims = [-1,-1,-1,-1]
|
|
|
|
if 'shape' in node.attr:
|
|
|
|
dims[0] = node.attr['shape'].shape.dim[0].size
|
|
|
|
dims[1] = node.attr['shape'].shape.dim[1].size
|
|
|
|
dims[2] = node.attr['shape'].shape.dim[2].size
|
|
|
|
dims[3] = node.attr['shape'].shape.dim[3].size
|
|
|
|
operand = Operand(name, dtype, dims)
|
|
|
|
self.name_operand_dict[name] = operand;
|
|
|
|
self.name_operand_dict[name].add_iotype(type)
|
|
|
|
return self.name_operand_dict[name].index
|
tools/python: add script to convert TensorFlow model (.pb) to native model (.model)
For example, given TensorFlow model file espcn.pb,
to generate native model file espcn.model, just run:
python convert.py espcn.pb
In current implementation, the native model file is generated for
specific dnn network with hard-code python scripts maintained out of ffmpeg.
For example, srcnn network used by vf_sr is generated with
https://github.com/HighVoltageRocknRoll/sr/blob/master/generate_header_and_model.py#L85
In this patch, the script is designed as a general solution which
converts general TensorFlow model .pb file into .model file. The script
now has some tricky to be compatible with current implemention, will
be refined step by step.
The script is also added into ffmpeg source tree. It is expected there
will be many more patches and community needs the ownership of it.
Another technical direction is to do the conversion in c/c++ code within
ffmpeg source tree. While .pb file is organized with protocol buffers,
it is not easy to do such work with tiny c/c++ code, see more discussion
at http://ffmpeg.org/pipermail/ffmpeg-devel/2019-May/244496.html. So,
choose the python script.
Signed-off-by: Guo, Yejun <yejun.guo@intel.com>
6 years ago
|
|
|
|
|
|
|
|
|
|
|
def dump_for_tensorboard(self):
|
|
|
|
graph = tf.get_default_graph()
|
|
|
|
tf.import_graph_def(self.graph_def, name="")
|
|
|
|
tf.summary.FileWriter('/tmp/graph', graph)
|
|
|
|
print('graph saved, run "tensorboard --logdir=/tmp/graph" to see it')
|
tools/python: add script to convert TensorFlow model (.pb) to native model (.model)
For example, given TensorFlow model file espcn.pb,
to generate native model file espcn.model, just run:
python convert.py espcn.pb
In current implementation, the native model file is generated for
specific dnn network with hard-code python scripts maintained out of ffmpeg.
For example, srcnn network used by vf_sr is generated with
https://github.com/HighVoltageRocknRoll/sr/blob/master/generate_header_and_model.py#L85
In this patch, the script is designed as a general solution which
converts general TensorFlow model .pb file into .model file. The script
now has some tricky to be compatible with current implemention, will
be refined step by step.
The script is also added into ffmpeg source tree. It is expected there
will be many more patches and community needs the ownership of it.
Another technical direction is to do the conversion in c/c++ code within
ffmpeg source tree. While .pb file is organized with protocol buffers,
it is not easy to do such work with tiny c/c++ code, see more discussion
at http://ffmpeg.org/pipermail/ffmpeg-devel/2019-May/244496.html. So,
choose the python script.
Signed-off-by: Guo, Yejun <yejun.guo@intel.com>
6 years ago
|
|
|
|
|
|
|
|
|
|
|
def get_conv2d_params(self, conv2d_scope_name):
|
|
|
|
knode = self.name_node_dict[conv2d_scope_name + '/kernel']
|
|
|
|
bnode = self.name_node_dict[conv2d_scope_name + '/bias']
|
|
|
|
|
|
|
|
if conv2d_scope_name + '/dilation_rate' in self.name_node_dict:
|
|
|
|
dnode = self.name_node_dict[conv2d_scope_name + '/dilation_rate']
|
|
|
|
else:
|
|
|
|
dnode = None
|
|
|
|
|
|
|
|
# the BiasAdd name is possible be changed into the output name,
|
|
|
|
# if activation is None, and BiasAdd.next is the last op which is Identity
|
|
|
|
if conv2d_scope_name + '/BiasAdd' in self.edges:
|
|
|
|
anode = self.edges[conv2d_scope_name + '/BiasAdd'][0]
|
|
|
|
if anode.op not in self.conv_activations:
|
|
|
|
anode = None
|
|
|
|
else:
|
|
|
|
anode = None
|
|
|
|
return knode, bnode, dnode, anode
|
tools/python: add script to convert TensorFlow model (.pb) to native model (.model)
For example, given TensorFlow model file espcn.pb,
to generate native model file espcn.model, just run:
python convert.py espcn.pb
In current implementation, the native model file is generated for
specific dnn network with hard-code python scripts maintained out of ffmpeg.
For example, srcnn network used by vf_sr is generated with
https://github.com/HighVoltageRocknRoll/sr/blob/master/generate_header_and_model.py#L85
In this patch, the script is designed as a general solution which
converts general TensorFlow model .pb file into .model file. The script
now has some tricky to be compatible with current implemention, will
be refined step by step.
The script is also added into ffmpeg source tree. It is expected there
will be many more patches and community needs the ownership of it.
Another technical direction is to do the conversion in c/c++ code within
ffmpeg source tree. While .pb file is organized with protocol buffers,
it is not easy to do such work with tiny c/c++ code, see more discussion
at http://ffmpeg.org/pipermail/ffmpeg-devel/2019-May/244496.html. So,
choose the python script.
Signed-off-by: Guo, Yejun <yejun.guo@intel.com>
6 years ago
|
|
|
|
|
|
|
|
|
|
|
def get_dense_params(self, dense_scope_name):
|
|
|
|
knode = self.name_node_dict[dense_scope_name + '/kernel']
|
|
|
|
bnode = self.name_node_dict.get(dense_scope_name + '/bias')
|
|
|
|
# the BiasAdd name is possible be changed into the output name,
|
|
|
|
# if activation is None, and BiasAdd.next is the last op which is Identity
|
|
|
|
anode = None
|
|
|
|
if bnode:
|
|
|
|
if dense_scope_name + '/BiasAdd' in self.edges:
|
|
|
|
anode = self.edges[dense_scope_name + '/BiasAdd'][0]
|
|
|
|
if anode.op not in self.conv_activations:
|
|
|
|
anode = None
|
|
|
|
else:
|
|
|
|
anode = None
|
|
|
|
return knode, bnode, anode
|
|
|
|
|
|
|
|
|
|
|
|
def dump_complex_conv2d_to_file(self, node, f):
|
tools/python: add script to convert TensorFlow model (.pb) to native model (.model)
For example, given TensorFlow model file espcn.pb,
to generate native model file espcn.model, just run:
python convert.py espcn.pb
In current implementation, the native model file is generated for
specific dnn network with hard-code python scripts maintained out of ffmpeg.
For example, srcnn network used by vf_sr is generated with
https://github.com/HighVoltageRocknRoll/sr/blob/master/generate_header_and_model.py#L85
In this patch, the script is designed as a general solution which
converts general TensorFlow model .pb file into .model file. The script
now has some tricky to be compatible with current implemention, will
be refined step by step.
The script is also added into ffmpeg source tree. It is expected there
will be many more patches and community needs the ownership of it.
Another technical direction is to do the conversion in c/c++ code within
ffmpeg source tree. While .pb file is organized with protocol buffers,
it is not easy to do such work with tiny c/c++ code, see more discussion
at http://ffmpeg.org/pipermail/ffmpeg-devel/2019-May/244496.html. So,
choose the python script.
Signed-off-by: Guo, Yejun <yejun.guo@intel.com>
6 years ago
|
|
|
assert(node.op == 'Conv2D')
|
|
|
|
self.layer_number = self.layer_number + 1
|
|
|
|
self.converted_nodes.add(node.name)
|
|
|
|
|
|
|
|
scope_name = TFConverter.get_scope_name(node.name)
|
|
|
|
#knode for kernel, bnode for bias, dnode for dilation, anode for activation
|
|
|
|
knode, bnode, dnode, anode = self.get_conv2d_params(scope_name)
|
|
|
|
|
|
|
|
if dnode is not None:
|
|
|
|
dilation = struct.unpack('i', dnode.attr['value'].tensor.tensor_content[0:4])[0]
|
|
|
|
else:
|
|
|
|
dilation = 1
|
|
|
|
|
|
|
|
if anode is not None:
|
|
|
|
activation = anode.op
|
|
|
|
else:
|
|
|
|
activation = 'None'
|
|
|
|
|
|
|
|
padding = node.attr['padding'].s.decode("utf-8")
|
|
|
|
# conv2d with dilation > 1 generates tens of nodes, not easy to parse them, so use this tricky method.
|
|
|
|
if dilation > 1 and scope_name + '/stack' in self.name_node_dict:
|
|
|
|
if self.name_node_dict[scope_name + '/stack'].op == "Const":
|
|
|
|
padding = 'SAME'
|
|
|
|
padding = self.conv_paddings[padding]
|
tools/python: add script to convert TensorFlow model (.pb) to native model (.model)
For example, given TensorFlow model file espcn.pb,
to generate native model file espcn.model, just run:
python convert.py espcn.pb
In current implementation, the native model file is generated for
specific dnn network with hard-code python scripts maintained out of ffmpeg.
For example, srcnn network used by vf_sr is generated with
https://github.com/HighVoltageRocknRoll/sr/blob/master/generate_header_and_model.py#L85
In this patch, the script is designed as a general solution which
converts general TensorFlow model .pb file into .model file. The script
now has some tricky to be compatible with current implemention, will
be refined step by step.
The script is also added into ffmpeg source tree. It is expected there
will be many more patches and community needs the ownership of it.
Another technical direction is to do the conversion in c/c++ code within
ffmpeg source tree. While .pb file is organized with protocol buffers,
it is not easy to do such work with tiny c/c++ code, see more discussion
at http://ffmpeg.org/pipermail/ffmpeg-devel/2019-May/244496.html. So,
choose the python script.
Signed-off-by: Guo, Yejun <yejun.guo@intel.com>
6 years ago
|
|
|
|
|
|
|
ktensor = knode.attr['value'].tensor
|
|
|
|
filter_height = ktensor.tensor_shape.dim[0].size
|
|
|
|
filter_width = ktensor.tensor_shape.dim[1].size
|
|
|
|
in_channels = ktensor.tensor_shape.dim[2].size
|
|
|
|
out_channels = ktensor.tensor_shape.dim[3].size
|
|
|
|
kernel = np.frombuffer(ktensor.tensor_content, dtype=np.float32)
|
|
|
|
kernel = kernel.reshape(filter_height, filter_width, in_channels, out_channels)
|
|
|
|
kernel = np.transpose(kernel, [3, 0, 1, 2])
|
|
|
|
|
|
|
|
has_bias = 1
|
|
|
|
np.array([self.op2code[node.op], dilation, padding, self.conv_activations[activation], in_channels, out_channels, filter_height, has_bias], dtype=np.uint32).tofile(f)
|
tools/python: add script to convert TensorFlow model (.pb) to native model (.model)
For example, given TensorFlow model file espcn.pb,
to generate native model file espcn.model, just run:
python convert.py espcn.pb
In current implementation, the native model file is generated for
specific dnn network with hard-code python scripts maintained out of ffmpeg.
For example, srcnn network used by vf_sr is generated with
https://github.com/HighVoltageRocknRoll/sr/blob/master/generate_header_and_model.py#L85
In this patch, the script is designed as a general solution which
converts general TensorFlow model .pb file into .model file. The script
now has some tricky to be compatible with current implemention, will
be refined step by step.
The script is also added into ffmpeg source tree. It is expected there
will be many more patches and community needs the ownership of it.
Another technical direction is to do the conversion in c/c++ code within
ffmpeg source tree. While .pb file is organized with protocol buffers,
it is not easy to do such work with tiny c/c++ code, see more discussion
at http://ffmpeg.org/pipermail/ffmpeg-devel/2019-May/244496.html. So,
choose the python script.
Signed-off-by: Guo, Yejun <yejun.guo@intel.com>
6 years ago
|
|
|
kernel.tofile(f)
|
|
|
|
|
|
|
|
btensor = bnode.attr['value'].tensor
|
|
|
|
if btensor.tensor_shape.dim[0].size == 1:
|
|
|
|
bias = struct.pack("f", btensor.float_val[0])
|
|
|
|
else:
|
|
|
|
bias = btensor.tensor_content
|
|
|
|
f.write(bias)
|
|
|
|
|
|
|
|
input_name = self.conv2d_scopename_inputname_dict[scope_name]
|
|
|
|
input_operand_index = self.add_operand(input_name, Operand.IOTYPE_INPUT)
|
|
|
|
|
|
|
|
if anode is not None:
|
|
|
|
output_operand_index = self.add_operand(anode.name, Operand.IOTYPE_OUTPUT)
|
|
|
|
else:
|
|
|
|
output_operand_index = self.add_operand(self.edges[bnode.name][0].name, Operand.IOTYPE_OUTPUT)
|
|
|
|
np.array([input_operand_index, output_operand_index], dtype=np.uint32).tofile(f)
|
|
|
|
|
|
|
|
def dump_dense_to_file(self, node, f):
|
|
|
|
assert(node.op == 'MatMul')
|
|
|
|
self.layer_number = self.layer_number + 1
|
|
|
|
self.converted_nodes.add(node.name)
|
|
|
|
|
|
|
|
scope_name = TFConverter.get_scope_name(node.name)
|
|
|
|
#knode for kernel, bnode for bias, anode for activation
|
|
|
|
knode, bnode, anode = self.get_dense_params(scope_name.split('/')[0])
|
|
|
|
|
|
|
|
if bnode is not None:
|
|
|
|
has_bias = 1
|
|
|
|
btensor = bnode.attr['value'].tensor
|
|
|
|
if btensor.tensor_shape.dim[0].size == 1:
|
|
|
|
bias = struct.pack("f", btensor.float_val[0])
|
|
|
|
else:
|
|
|
|
bias = btensor.tensor_content
|
|
|
|
else:
|
|
|
|
has_bias = 0
|
|
|
|
|
|
|
|
if anode is not None:
|
|
|
|
activation = anode.op
|
|
|
|
else:
|
|
|
|
activation = 'None'
|
|
|
|
|
|
|
|
ktensor = knode.attr['value'].tensor
|
|
|
|
in_channels = ktensor.tensor_shape.dim[0].size
|
|
|
|
out_channels = ktensor.tensor_shape.dim[1].size
|
|
|
|
if in_channels * out_channels == 1:
|
|
|
|
kernel = np.float32(ktensor.float_val[0])
|
|
|
|
else:
|
|
|
|
kernel = np.frombuffer(ktensor.tensor_content, dtype=np.float32)
|
|
|
|
kernel = kernel.reshape(in_channels, out_channels)
|
|
|
|
kernel = np.transpose(kernel, [1, 0])
|
|
|
|
|
|
|
|
np.array([self.op2code[node.op], self.conv_activations[activation], in_channels, out_channels, has_bias], dtype=np.uint32).tofile(f)
|
|
|
|
kernel.tofile(f)
|
|
|
|
if has_bias:
|
|
|
|
f.write(bias)
|
|
|
|
|
|
|
|
input_name = self.dense_scopename_inputname_dict[scope_name.split('/')[0]]
|
|
|
|
input_operand_index = self.add_operand(input_name, Operand.IOTYPE_INPUT)
|
|
|
|
|
|
|
|
if anode is not None:
|
|
|
|
output_operand_index = self.add_operand(anode.name, Operand.IOTYPE_OUTPUT)
|
|
|
|
else:
|
|
|
|
if bnode is not None:
|
|
|
|
output_operand_index = self.add_operand(self.edges[bnode.name][0].name, Operand.IOTYPE_OUTPUT)
|
|
|
|
else:
|
|
|
|
output_operand_index = self.add_operand(self.edges[scope_name+'/concat_1'][0].name, Operand.IOTYPE_OUTPUT)
|
|
|
|
np.array([input_operand_index, output_operand_index], dtype=np.uint32).tofile(f)
|
|
|
|
|
tools/python: add script to convert TensorFlow model (.pb) to native model (.model)
For example, given TensorFlow model file espcn.pb,
to generate native model file espcn.model, just run:
python convert.py espcn.pb
In current implementation, the native model file is generated for
specific dnn network with hard-code python scripts maintained out of ffmpeg.
For example, srcnn network used by vf_sr is generated with
https://github.com/HighVoltageRocknRoll/sr/blob/master/generate_header_and_model.py#L85
In this patch, the script is designed as a general solution which
converts general TensorFlow model .pb file into .model file. The script
now has some tricky to be compatible with current implemention, will
be refined step by step.
The script is also added into ffmpeg source tree. It is expected there
will be many more patches and community needs the ownership of it.
Another technical direction is to do the conversion in c/c++ code within
ffmpeg source tree. While .pb file is organized with protocol buffers,
it is not easy to do such work with tiny c/c++ code, see more discussion
at http://ffmpeg.org/pipermail/ffmpeg-devel/2019-May/244496.html. So,
choose the python script.
Signed-off-by: Guo, Yejun <yejun.guo@intel.com>
6 years ago
|
|
|
|
|
|
|
def dump_simple_conv2d_to_file(self, node, f):
|
|
|
|
assert(node.op == 'Conv2D')
|
|
|
|
self.layer_number = self.layer_number + 1
|
|
|
|
self.converted_nodes.add(node.name)
|
|
|
|
|
|
|
|
node0 = self.name_node_dict[node.input[0]]
|
|
|
|
node1 = self.name_node_dict[node.input[1]]
|
|
|
|
if node0.op == 'Const':
|
|
|
|
knode = node0
|
|
|
|
input_name = node.input[1]
|
|
|
|
else:
|
|
|
|
knode = node1
|
|
|
|
input_name = node.input[0]
|
|
|
|
|
|
|
|
ktensor = knode.attr['value'].tensor
|
|
|
|
filter_height = ktensor.tensor_shape.dim[0].size
|
|
|
|
filter_width = ktensor.tensor_shape.dim[1].size
|
|
|
|
in_channels = ktensor.tensor_shape.dim[2].size
|
|
|
|
out_channels = ktensor.tensor_shape.dim[3].size
|
|
|
|
if filter_height * filter_width * in_channels * out_channels == 1:
|
|
|
|
kernel = np.float32(ktensor.float_val[0])
|
|
|
|
else:
|
|
|
|
kernel = np.frombuffer(ktensor.tensor_content, dtype=np.float32)
|
|
|
|
kernel = kernel.reshape(filter_height, filter_width, in_channels, out_channels)
|
|
|
|
kernel = np.transpose(kernel, [3, 0, 1, 2])
|
|
|
|
|
|
|
|
has_bias = 0
|
|
|
|
dilation = 1
|
|
|
|
padding = node.attr['padding'].s.decode("utf-8")
|
|
|
|
np.array([self.op2code[node.op], dilation, self.conv_paddings[padding], self.conv_activations['None'],
|
|
|
|
in_channels, out_channels, filter_height, has_bias], dtype=np.uint32).tofile(f)
|
|
|
|
kernel.tofile(f)
|
|
|
|
|
|
|
|
input_operand_index = self.add_operand(input_name, Operand.IOTYPE_INPUT)
|
|
|
|
output_operand_index = self.add_operand(node.name, Operand.IOTYPE_OUTPUT)
|
|
|
|
np.array([input_operand_index, output_operand_index], dtype=np.uint32).tofile(f)
|
|
|
|
|
|
|
|
|
tools/python: add script to convert TensorFlow model (.pb) to native model (.model)
For example, given TensorFlow model file espcn.pb,
to generate native model file espcn.model, just run:
python convert.py espcn.pb
In current implementation, the native model file is generated for
specific dnn network with hard-code python scripts maintained out of ffmpeg.
For example, srcnn network used by vf_sr is generated with
https://github.com/HighVoltageRocknRoll/sr/blob/master/generate_header_and_model.py#L85
In this patch, the script is designed as a general solution which
converts general TensorFlow model .pb file into .model file. The script
now has some tricky to be compatible with current implemention, will
be refined step by step.
The script is also added into ffmpeg source tree. It is expected there
will be many more patches and community needs the ownership of it.
Another technical direction is to do the conversion in c/c++ code within
ffmpeg source tree. While .pb file is organized with protocol buffers,
it is not easy to do such work with tiny c/c++ code, see more discussion
at http://ffmpeg.org/pipermail/ffmpeg-devel/2019-May/244496.html. So,
choose the python script.
Signed-off-by: Guo, Yejun <yejun.guo@intel.com>
6 years ago
|
|
|
def dump_depth2space_to_file(self, node, f):
|
|
|
|
assert(node.op == 'DepthToSpace')
|
|
|
|
self.layer_number = self.layer_number + 1
|
|
|
|
block_size = node.attr['block_size'].i
|
|
|
|
np.array([self.op2code[node.op], block_size], dtype=np.uint32).tofile(f)
|
|
|
|
self.converted_nodes.add(node.name)
|
|
|
|
input_operand_index = self.add_operand(node.input[0], Operand.IOTYPE_INPUT)
|
|
|
|
output_operand_index = self.add_operand(node.name, Operand.IOTYPE_OUTPUT)
|
|
|
|
np.array([input_operand_index, output_operand_index], dtype=np.uint32).tofile(f)
|
tools/python: add script to convert TensorFlow model (.pb) to native model (.model)
For example, given TensorFlow model file espcn.pb,
to generate native model file espcn.model, just run:
python convert.py espcn.pb
In current implementation, the native model file is generated for
specific dnn network with hard-code python scripts maintained out of ffmpeg.
For example, srcnn network used by vf_sr is generated with
https://github.com/HighVoltageRocknRoll/sr/blob/master/generate_header_and_model.py#L85
In this patch, the script is designed as a general solution which
converts general TensorFlow model .pb file into .model file. The script
now has some tricky to be compatible with current implemention, will
be refined step by step.
The script is also added into ffmpeg source tree. It is expected there
will be many more patches and community needs the ownership of it.
Another technical direction is to do the conversion in c/c++ code within
ffmpeg source tree. While .pb file is organized with protocol buffers,
it is not easy to do such work with tiny c/c++ code, see more discussion
at http://ffmpeg.org/pipermail/ffmpeg-devel/2019-May/244496.html. So,
choose the python script.
Signed-off-by: Guo, Yejun <yejun.guo@intel.com>
6 years ago
|
|
|
|
|
|
|
|
|
|
|
def dump_mirrorpad_to_file(self, node, f):
|
|
|
|
assert(node.op == 'MirrorPad')
|
|
|
|
self.layer_number = self.layer_number + 1
|
|
|
|
mode = node.attr['mode'].s
|
|
|
|
mode = self.mirrorpad_mode[mode.decode("utf-8")]
|
|
|
|
np.array([self.op2code[node.op], mode], dtype=np.uint32).tofile(f)
|
|
|
|
pnode = self.name_node_dict[node.input[1]]
|
|
|
|
self.converted_nodes.add(pnode.name)
|
|
|
|
paddings = pnode.attr['value'].tensor.tensor_content
|
|
|
|
f.write(paddings)
|
|
|
|
self.converted_nodes.add(node.name)
|
|
|
|
input_operand_index = self.add_operand(node.input[0], Operand.IOTYPE_INPUT)
|
|
|
|
output_operand_index = self.add_operand(node.name, Operand.IOTYPE_OUTPUT)
|
|
|
|
np.array([input_operand_index, output_operand_index], dtype=np.uint32).tofile(f)
|
|
|
|
|
|
|
|
|
|
|
|
def dump_maximum_to_file(self, node, f):
|
|
|
|
assert(node.op == 'Maximum')
|
|
|
|
self.layer_number = self.layer_number + 1
|
|
|
|
ynode = self.name_node_dict[node.input[1]]
|
|
|
|
y = ynode.attr['value'].tensor.float_val[0]
|
|
|
|
np.array([self.op2code[node.op]], dtype=np.uint32).tofile(f)
|
|
|
|
np.array([y], dtype=np.float32).tofile(f)
|
|
|
|
self.converted_nodes.add(node.name)
|
|
|
|
input_operand_index = self.add_operand(node.input[0], Operand.IOTYPE_INPUT)
|
|
|
|
output_operand_index = self.add_operand(node.name, Operand.IOTYPE_OUTPUT)
|
|
|
|
np.array([input_operand_index, output_operand_index], dtype=np.uint32).tofile(f)
|
|
|
|
|
|
|
|
|
dnn/native: add native support for 'add'
It can be tested with the model file generated with below python script:
import tensorflow as tf
import numpy as np
import imageio
in_img = imageio.imread('input.jpg')
in_img = in_img.astype(np.float32)/255.0
in_data = in_img[np.newaxis, :]
x = tf.placeholder(tf.float32, shape=[1, None, None, 3], name='dnn_in')
z1 = 0.039 + x
z2 = x + 0.042
z3 = z1 + z2
z4 = z3 - 0.381
z5 = z4 - x
y = tf.math.maximum(z5, 0.0, name='dnn_out')
sess=tf.Session()
sess.run(tf.global_variables_initializer())
graph_def = tf.graph_util.convert_variables_to_constants(sess, sess.graph_def, ['dnn_out'])
tf.train.write_graph(graph_def, '.', 'image_process.pb', as_text=False)
print("image_process.pb generated, please use \
path_to_ffmpeg/tools/python/convert.py to generate image_process.model\n")
output = sess.run(y, feed_dict={x: in_data})
imageio.imsave("out.jpg", np.squeeze(output))
Signed-off-by: Guo, Yejun <yejun.guo@intel.com>
5 years ago
|
|
|
def dump_mathbinary_to_file(self, node, f):
|
|
|
|
self.layer_number = self.layer_number + 1
|
|
|
|
self.converted_nodes.add(node.name)
|
|
|
|
i0_node = self.name_node_dict[node.input[0]]
|
|
|
|
i1_node = self.name_node_dict[node.input[1]]
|
|
|
|
np.array([self.op2code['MathBinary'], self.mathbin2code[node.op]], dtype=np.uint32).tofile(f)
|
|
|
|
if i0_node.op == 'Const':
|
|
|
|
scalar = i0_node.attr['value'].tensor.float_val[0]
|
dnn/native: add native support for 'add'
It can be tested with the model file generated with below python script:
import tensorflow as tf
import numpy as np
import imageio
in_img = imageio.imread('input.jpg')
in_img = in_img.astype(np.float32)/255.0
in_data = in_img[np.newaxis, :]
x = tf.placeholder(tf.float32, shape=[1, None, None, 3], name='dnn_in')
z1 = 0.039 + x
z2 = x + 0.042
z3 = z1 + z2
z4 = z3 - 0.381
z5 = z4 - x
y = tf.math.maximum(z5, 0.0, name='dnn_out')
sess=tf.Session()
sess.run(tf.global_variables_initializer())
graph_def = tf.graph_util.convert_variables_to_constants(sess, sess.graph_def, ['dnn_out'])
tf.train.write_graph(graph_def, '.', 'image_process.pb', as_text=False)
print("image_process.pb generated, please use \
path_to_ffmpeg/tools/python/convert.py to generate image_process.model\n")
output = sess.run(y, feed_dict={x: in_data})
imageio.imsave("out.jpg", np.squeeze(output))
Signed-off-by: Guo, Yejun <yejun.guo@intel.com>
5 years ago
|
|
|
np.array([1], dtype=np.uint32).tofile(f) # broadcast: 1
|
|
|
|
np.array([scalar], dtype=np.float32).tofile(f)
|
dnn/native: add native support for 'add'
It can be tested with the model file generated with below python script:
import tensorflow as tf
import numpy as np
import imageio
in_img = imageio.imread('input.jpg')
in_img = in_img.astype(np.float32)/255.0
in_data = in_img[np.newaxis, :]
x = tf.placeholder(tf.float32, shape=[1, None, None, 3], name='dnn_in')
z1 = 0.039 + x
z2 = x + 0.042
z3 = z1 + z2
z4 = z3 - 0.381
z5 = z4 - x
y = tf.math.maximum(z5, 0.0, name='dnn_out')
sess=tf.Session()
sess.run(tf.global_variables_initializer())
graph_def = tf.graph_util.convert_variables_to_constants(sess, sess.graph_def, ['dnn_out'])
tf.train.write_graph(graph_def, '.', 'image_process.pb', as_text=False)
print("image_process.pb generated, please use \
path_to_ffmpeg/tools/python/convert.py to generate image_process.model\n")
output = sess.run(y, feed_dict={x: in_data})
imageio.imsave("out.jpg", np.squeeze(output))
Signed-off-by: Guo, Yejun <yejun.guo@intel.com>
5 years ago
|
|
|
np.array([0], dtype=np.uint32).tofile(f) # broadcast: 0
|
|
|
|
input_operand_index = self.add_operand(i1_node.name, Operand.IOTYPE_INPUT)
|
|
|
|
np.array([input_operand_index], dtype=np.uint32).tofile(f)
|
|
|
|
elif i1_node.op == 'Const':
|
|
|
|
scalar = i1_node.attr['value'].tensor.float_val[0]
|
|
|
|
np.array([0], dtype=np.uint32).tofile(f)
|
|
|
|
input_operand_index = self.add_operand(i0_node.name, Operand.IOTYPE_INPUT)
|
|
|
|
np.array([input_operand_index], dtype=np.uint32).tofile(f)
|
|
|
|
np.array([1], dtype=np.uint32).tofile(f)
|
|
|
|
np.array([scalar], dtype=np.float32).tofile(f)
|
|
|
|
else:
|
|
|
|
np.array([0], dtype=np.uint32).tofile(f)
|
|
|
|
input_operand_index = self.add_operand(i0_node.name, Operand.IOTYPE_INPUT)
|
|
|
|
np.array([input_operand_index], dtype=np.uint32).tofile(f)
|
|
|
|
np.array([0], dtype=np.uint32).tofile(f)
|
|
|
|
input_operand_index = self.add_operand(i1_node.name, Operand.IOTYPE_INPUT)
|
|
|
|
np.array([input_operand_index], dtype=np.uint32).tofile(f)
|
|
|
|
output_operand_index = self.add_operand(node.name, Operand.IOTYPE_OUTPUT)
|
|
|
|
np.array([output_operand_index], dtype=np.uint32).tofile(f)
|
|
|
|
|
|
|
|
|
dnn_backend_native_layer_mathunary: add abs support
more math unary operations will be added here
It can be tested with the model file generated with below python scripy:
import tensorflow as tf
import numpy as np
import imageio
in_img = imageio.imread('input.jpeg')
in_img = in_img.astype(np.float32)/255.0
in_data = in_img[np.newaxis, :]
x = tf.placeholder(tf.float32, shape=[1, None, None, 3], name='dnn_in')
x1 = tf.subtract(x, 0.5)
x2 = tf.abs(x1)
y = tf.identity(x2, name='dnn_out')
sess=tf.Session()
sess.run(tf.global_variables_initializer())
graph_def = tf.graph_util.convert_variables_to_constants(sess, sess.graph_def, ['dnn_out'])
tf.train.write_graph(graph_def, '.', 'image_process.pb', as_text=False)
print("image_process.pb generated, please use \
path_to_ffmpeg/tools/python/convert.py to generate image_process.model\n")
output = sess.run(y, feed_dict={x: in_data})
imageio.imsave("out.jpg", np.squeeze(output))
Signed-off-by: Ting Fu <ting.fu@intel.com>
Signed-off-by: Guo, Yejun <yejun.guo@intel.com>
5 years ago
|
|
|
def dump_mathunary_to_file(self, node, f):
|
|
|
|
self.layer_number = self.layer_number + 1
|
|
|
|
self.converted_nodes.add(node.name)
|
|
|
|
i0_node = self.name_node_dict[node.input[0]]
|
|
|
|
np.array([self.op2code['MathUnary'], self.mathun2code[node.op]], dtype=np.uint32).tofile(f)
|
|
|
|
input_operand_index = self.add_operand(i0_node.name, Operand.IOTYPE_INPUT)
|
|
|
|
np.array([input_operand_index], dtype=np.uint32).tofile(f)
|
|
|
|
output_operand_index = self.add_operand(node.name, Operand.IOTYPE_OUTPUT)
|
|
|
|
np.array([output_operand_index],dtype=np.uint32).tofile(f)
|
|
|
|
|
|
|
|
|
|
|
|
def dump_avg_pool_to_file(self, node, f):
|
|
|
|
assert(node.op == 'AvgPool')
|
|
|
|
self.layer_number = self.layer_number + 1
|
|
|
|
self.converted_nodes.add(node.name)
|
|
|
|
node0 = self.name_node_dict[node.input[0]]
|
|
|
|
strides = node.attr['strides']
|
|
|
|
|
|
|
|
# Tensorflow do not support pooling strides in batch dimension and
|
|
|
|
# current native NN do not support pooling strides in channel dimension, added assert() here.
|
|
|
|
assert(strides.list.i[1]==strides.list.i[2])
|
|
|
|
assert(strides.list.i[0]==1)
|
|
|
|
assert(strides.list.i[3]==1)
|
|
|
|
strides = strides.list.i[1]
|
|
|
|
filter_node = node.attr['ksize']
|
|
|
|
input_name = node.input[0]
|
|
|
|
|
|
|
|
# Tensorflow do not support pooling ksize in batch dimension and channel dimension.
|
|
|
|
assert(filter_node.list.i[0]==1)
|
|
|
|
assert(filter_node.list.i[3]==1)
|
|
|
|
filter_height = filter_node.list.i[1]
|
|
|
|
filter_width = filter_node.list.i[2]
|
|
|
|
|
|
|
|
padding = node.attr['padding'].s.decode("utf-8")
|
|
|
|
np.array([self.op2code[node.op], strides, self.pool_paddings[padding], filter_height],
|
|
|
|
dtype=np.uint32).tofile(f)
|
|
|
|
|
|
|
|
input_operand_index = self.add_operand(input_name, Operand.IOTYPE_INPUT)
|
|
|
|
output_operand_index = self.add_operand(node.name, Operand.IOTYPE_OUTPUT)
|
|
|
|
np.array([input_operand_index, output_operand_index],dtype=np.uint32).tofile(f)
|
|
|
|
|
|
|
|
|
tools/python: add script to convert TensorFlow model (.pb) to native model (.model)
For example, given TensorFlow model file espcn.pb,
to generate native model file espcn.model, just run:
python convert.py espcn.pb
In current implementation, the native model file is generated for
specific dnn network with hard-code python scripts maintained out of ffmpeg.
For example, srcnn network used by vf_sr is generated with
https://github.com/HighVoltageRocknRoll/sr/blob/master/generate_header_and_model.py#L85
In this patch, the script is designed as a general solution which
converts general TensorFlow model .pb file into .model file. The script
now has some tricky to be compatible with current implemention, will
be refined step by step.
The script is also added into ffmpeg source tree. It is expected there
will be many more patches and community needs the ownership of it.
Another technical direction is to do the conversion in c/c++ code within
ffmpeg source tree. While .pb file is organized with protocol buffers,
it is not easy to do such work with tiny c/c++ code, see more discussion
at http://ffmpeg.org/pipermail/ffmpeg-devel/2019-May/244496.html. So,
choose the python script.
Signed-off-by: Guo, Yejun <yejun.guo@intel.com>
6 years ago
|
|
|
def dump_layers_to_file(self, f):
|
|
|
|
for node in self.nodes:
|
|
|
|
if node.name in self.converted_nodes:
|
|
|
|
continue
|
|
|
|
|
|
|
|
# conv2d with dilation generates very complex nodes, so handle it in special
|
|
|
|
if self.in_conv2d_scope(node.name):
|
|
|
|
if node.op == 'Conv2D':
|
|
|
|
self.dump_complex_conv2d_to_file(node, f)
|
|
|
|
continue
|
|
|
|
if self.in_dense_scope(node.name):
|
|
|
|
if node.op == 'MatMul':
|
|
|
|
self.dump_dense_to_file(node, f)
|
|
|
|
continue
|
|
|
|
|
|
|
|
|
|
|
|
if node.op == 'Conv2D':
|
|
|
|
self.dump_simple_conv2d_to_file(node, f)
|
|
|
|
continue
|
|
|
|
if node.name in self.output_names:
|
|
|
|
input_name = self.id_different_scope_dict[node.name]
|
|
|
|
if TFConverter.get_scope_name(input_name)!=TFConverter.get_scope_name(node.name):
|
|
|
|
continue
|
|
|
|
if node.op == 'AvgPool':
|
|
|
|
self.dump_avg_pool_to_file(node, f)
|
|
|
|
elif node.op == 'DepthToSpace':
|
tools/python: add script to convert TensorFlow model (.pb) to native model (.model)
For example, given TensorFlow model file espcn.pb,
to generate native model file espcn.model, just run:
python convert.py espcn.pb
In current implementation, the native model file is generated for
specific dnn network with hard-code python scripts maintained out of ffmpeg.
For example, srcnn network used by vf_sr is generated with
https://github.com/HighVoltageRocknRoll/sr/blob/master/generate_header_and_model.py#L85
In this patch, the script is designed as a general solution which
converts general TensorFlow model .pb file into .model file. The script
now has some tricky to be compatible with current implemention, will
be refined step by step.
The script is also added into ffmpeg source tree. It is expected there
will be many more patches and community needs the ownership of it.
Another technical direction is to do the conversion in c/c++ code within
ffmpeg source tree. While .pb file is organized with protocol buffers,
it is not easy to do such work with tiny c/c++ code, see more discussion
at http://ffmpeg.org/pipermail/ffmpeg-devel/2019-May/244496.html. So,
choose the python script.
Signed-off-by: Guo, Yejun <yejun.guo@intel.com>
6 years ago
|
|
|
self.dump_depth2space_to_file(node, f)
|
|
|
|
elif node.op == 'MirrorPad':
|
|
|
|
self.dump_mirrorpad_to_file(node, f)
|
|
|
|
elif node.op == 'Maximum':
|
|
|
|
self.dump_maximum_to_file(node, f)
|
dnn/native: add native support for minimum
it can be tested with model file generated with below python script:
import tensorflow as tf
import numpy as np
import imageio
in_img = imageio.imread('input.jpg')
in_img = in_img.astype(np.float32)/255.0
in_data = in_img[np.newaxis, :]
x = tf.placeholder(tf.float32, shape=[1, None, None, 3], name='dnn_in')
x1 = tf.minimum(0.7, x)
x2 = tf.maximum(x1, 0.4)
y = tf.identity(x2, name='dnn_out')
sess=tf.Session()
sess.run(tf.global_variables_initializer())
graph_def = tf.graph_util.convert_variables_to_constants(sess, sess.graph_def, ['dnn_out'])
tf.train.write_graph(graph_def, '.', 'image_process.pb', as_text=False)
print("image_process.pb generated, please use \
path_to_ffmpeg/tools/python/convert.py to generate image_process.model\n")
output = sess.run(y, feed_dict={x: in_data})
imageio.imsave("out.jpg", np.squeeze(output))
Signed-off-by: Guo, Yejun <yejun.guo@intel.com>
5 years ago
|
|
|
elif node.op in self.mathbin2code:
|
dnn/native: add native support for divide
it can be tested with model file generated with below python script:
import tensorflow as tf
import numpy as np
import imageio
in_img = imageio.imread('input.jpg')
in_img = in_img.astype(np.float32)/255.0
in_data = in_img[np.newaxis, :]
x = tf.placeholder(tf.float32, shape=[1, None, None, 3], name='dnn_in')
z1 = 2 / x
z2 = 1 / z1
z3 = z2 / 0.25 + 0.3
z4 = z3 - x * 1.5 - 0.3
y = tf.identity(z4, name='dnn_out')
sess=tf.Session()
sess.run(tf.global_variables_initializer())
graph_def = tf.graph_util.convert_variables_to_constants(sess, sess.graph_def, ['dnn_out'])
tf.train.write_graph(graph_def, '.', 'image_process.pb', as_text=False)
print("image_process.pb generated, please use \
path_to_ffmpeg/tools/python/convert.py to generate image_process.model\n")
output = sess.run(y, feed_dict={x: in_data})
imageio.imsave("out.jpg", np.squeeze(output))
Signed-off-by: Guo, Yejun <yejun.guo@intel.com>
5 years ago
|
|
|
self.dump_mathbinary_to_file(node, f)
|
dnn_backend_native_layer_mathunary: add abs support
more math unary operations will be added here
It can be tested with the model file generated with below python scripy:
import tensorflow as tf
import numpy as np
import imageio
in_img = imageio.imread('input.jpeg')
in_img = in_img.astype(np.float32)/255.0
in_data = in_img[np.newaxis, :]
x = tf.placeholder(tf.float32, shape=[1, None, None, 3], name='dnn_in')
x1 = tf.subtract(x, 0.5)
x2 = tf.abs(x1)
y = tf.identity(x2, name='dnn_out')
sess=tf.Session()
sess.run(tf.global_variables_initializer())
graph_def = tf.graph_util.convert_variables_to_constants(sess, sess.graph_def, ['dnn_out'])
tf.train.write_graph(graph_def, '.', 'image_process.pb', as_text=False)
print("image_process.pb generated, please use \
path_to_ffmpeg/tools/python/convert.py to generate image_process.model\n")
output = sess.run(y, feed_dict={x: in_data})
imageio.imsave("out.jpg", np.squeeze(output))
Signed-off-by: Ting Fu <ting.fu@intel.com>
Signed-off-by: Guo, Yejun <yejun.guo@intel.com>
5 years ago
|
|
|
elif node.op in self.mathun2code:
|
|
|
|
self.dump_mathunary_to_file(node, f)
|
tools/python: add script to convert TensorFlow model (.pb) to native model (.model)
For example, given TensorFlow model file espcn.pb,
to generate native model file espcn.model, just run:
python convert.py espcn.pb
In current implementation, the native model file is generated for
specific dnn network with hard-code python scripts maintained out of ffmpeg.
For example, srcnn network used by vf_sr is generated with
https://github.com/HighVoltageRocknRoll/sr/blob/master/generate_header_and_model.py#L85
In this patch, the script is designed as a general solution which
converts general TensorFlow model .pb file into .model file. The script
now has some tricky to be compatible with current implemention, will
be refined step by step.
The script is also added into ffmpeg source tree. It is expected there
will be many more patches and community needs the ownership of it.
Another technical direction is to do the conversion in c/c++ code within
ffmpeg source tree. While .pb file is organized with protocol buffers,
it is not easy to do such work with tiny c/c++ code, see more discussion
at http://ffmpeg.org/pipermail/ffmpeg-devel/2019-May/244496.html. So,
choose the python script.
Signed-off-by: Guo, Yejun <yejun.guo@intel.com>
6 years ago
|
|
|
|
dnn/native: add native support for minimum
it can be tested with model file generated with below python script:
import tensorflow as tf
import numpy as np
import imageio
in_img = imageio.imread('input.jpg')
in_img = in_img.astype(np.float32)/255.0
in_data = in_img[np.newaxis, :]
x = tf.placeholder(tf.float32, shape=[1, None, None, 3], name='dnn_in')
x1 = tf.minimum(0.7, x)
x2 = tf.maximum(x1, 0.4)
y = tf.identity(x2, name='dnn_out')
sess=tf.Session()
sess.run(tf.global_variables_initializer())
graph_def = tf.graph_util.convert_variables_to_constants(sess, sess.graph_def, ['dnn_out'])
tf.train.write_graph(graph_def, '.', 'image_process.pb', as_text=False)
print("image_process.pb generated, please use \
path_to_ffmpeg/tools/python/convert.py to generate image_process.model\n")
output = sess.run(y, feed_dict={x: in_data})
imageio.imsave("out.jpg", np.squeeze(output))
Signed-off-by: Guo, Yejun <yejun.guo@intel.com>
5 years ago
|
|
|
|
|
|
|
def dump_operands_to_file(self, f):
|
|
|
|
operands = sorted(self.name_operand_dict.values())
|
|
|
|
for operand in operands:
|
|
|
|
#print('{}'.format(operand))
|
|
|
|
np.array([operand.index, len(operand.name)], dtype=np.uint32).tofile(f)
|
|
|
|
f.write(operand.name.encode('utf-8'))
|
|
|
|
np.array([operand.iotype, operand.dtype], dtype=np.uint32).tofile(f)
|
|
|
|
np.array(operand.dims, dtype=np.uint32).tofile(f)
|
|
|
|
|
|
|
|
|
tools/python: add script to convert TensorFlow model (.pb) to native model (.model)
For example, given TensorFlow model file espcn.pb,
to generate native model file espcn.model, just run:
python convert.py espcn.pb
In current implementation, the native model file is generated for
specific dnn network with hard-code python scripts maintained out of ffmpeg.
For example, srcnn network used by vf_sr is generated with
https://github.com/HighVoltageRocknRoll/sr/blob/master/generate_header_and_model.py#L85
In this patch, the script is designed as a general solution which
converts general TensorFlow model .pb file into .model file. The script
now has some tricky to be compatible with current implemention, will
be refined step by step.
The script is also added into ffmpeg source tree. It is expected there
will be many more patches and community needs the ownership of it.
Another technical direction is to do the conversion in c/c++ code within
ffmpeg source tree. While .pb file is organized with protocol buffers,
it is not easy to do such work with tiny c/c++ code, see more discussion
at http://ffmpeg.org/pipermail/ffmpeg-devel/2019-May/244496.html. So,
choose the python script.
Signed-off-by: Guo, Yejun <yejun.guo@intel.com>
6 years ago
|
|
|
def dump_to_file(self):
|
|
|
|
with open(self.outfile, 'wb') as f:
|
|
|
|
f.write(header.str.encode('utf-8'))
|
|
|
|
np.array([header.major, header.minor], dtype=np.uint32).tofile(f)
|
tools/python: add script to convert TensorFlow model (.pb) to native model (.model)
For example, given TensorFlow model file espcn.pb,
to generate native model file espcn.model, just run:
python convert.py espcn.pb
In current implementation, the native model file is generated for
specific dnn network with hard-code python scripts maintained out of ffmpeg.
For example, srcnn network used by vf_sr is generated with
https://github.com/HighVoltageRocknRoll/sr/blob/master/generate_header_and_model.py#L85
In this patch, the script is designed as a general solution which
converts general TensorFlow model .pb file into .model file. The script
now has some tricky to be compatible with current implemention, will
be refined step by step.
The script is also added into ffmpeg source tree. It is expected there
will be many more patches and community needs the ownership of it.
Another technical direction is to do the conversion in c/c++ code within
ffmpeg source tree. While .pb file is organized with protocol buffers,
it is not easy to do such work with tiny c/c++ code, see more discussion
at http://ffmpeg.org/pipermail/ffmpeg-devel/2019-May/244496.html. So,
choose the python script.
Signed-off-by: Guo, Yejun <yejun.guo@intel.com>
6 years ago
|
|
|
self.dump_layers_to_file(f)
|
|
|
|
self.dump_operands_to_file(f)
|
|
|
|
np.array([self.layer_number, len(self.name_operand_dict)], dtype=np.uint32).tofile(f)
|
tools/python: add script to convert TensorFlow model (.pb) to native model (.model)
For example, given TensorFlow model file espcn.pb,
to generate native model file espcn.model, just run:
python convert.py espcn.pb
In current implementation, the native model file is generated for
specific dnn network with hard-code python scripts maintained out of ffmpeg.
For example, srcnn network used by vf_sr is generated with
https://github.com/HighVoltageRocknRoll/sr/blob/master/generate_header_and_model.py#L85
In this patch, the script is designed as a general solution which
converts general TensorFlow model .pb file into .model file. The script
now has some tricky to be compatible with current implemention, will
be refined step by step.
The script is also added into ffmpeg source tree. It is expected there
will be many more patches and community needs the ownership of it.
Another technical direction is to do the conversion in c/c++ code within
ffmpeg source tree. While .pb file is organized with protocol buffers,
it is not easy to do such work with tiny c/c++ code, see more discussion
at http://ffmpeg.org/pipermail/ffmpeg-devel/2019-May/244496.html. So,
choose the python script.
Signed-off-by: Guo, Yejun <yejun.guo@intel.com>
6 years ago
|
|
|
|
|
|
|
|
|
|
|
def generate_name_node_dict(self):
|
|
|
|
for node in self.nodes:
|
|
|
|
self.name_node_dict[node.name] = node
|
|
|
|
|
|
|
|
|
|
|
|
def generate_output_names(self):
|
|
|
|
used_names = []
|
|
|
|
for node in self.nodes:
|
|
|
|
for input in node.input:
|
|
|
|
used_names.append(input)
|
|
|
|
|
|
|
|
for node in self.nodes:
|
|
|
|
if node.name not in used_names:
|
|
|
|
self.output_names.append(node.name)
|
|
|
|
|
|
|
|
|
|
|
|
def remove_identity(self):
|
|
|
|
self.id_different_scope_dict = {}
|
tools/python: add script to convert TensorFlow model (.pb) to native model (.model)
For example, given TensorFlow model file espcn.pb,
to generate native model file espcn.model, just run:
python convert.py espcn.pb
In current implementation, the native model file is generated for
specific dnn network with hard-code python scripts maintained out of ffmpeg.
For example, srcnn network used by vf_sr is generated with
https://github.com/HighVoltageRocknRoll/sr/blob/master/generate_header_and_model.py#L85
In this patch, the script is designed as a general solution which
converts general TensorFlow model .pb file into .model file. The script
now has some tricky to be compatible with current implemention, will
be refined step by step.
The script is also added into ffmpeg source tree. It is expected there
will be many more patches and community needs the ownership of it.
Another technical direction is to do the conversion in c/c++ code within
ffmpeg source tree. While .pb file is organized with protocol buffers,
it is not easy to do such work with tiny c/c++ code, see more discussion
at http://ffmpeg.org/pipermail/ffmpeg-devel/2019-May/244496.html. So,
choose the python script.
Signed-off-by: Guo, Yejun <yejun.guo@intel.com>
6 years ago
|
|
|
id_nodes = []
|
|
|
|
id_dict = {}
|
|
|
|
for node in self.nodes:
|
|
|
|
if node.op == 'Identity':
|
|
|
|
name = node.name
|
|
|
|
input = node.input[0]
|
|
|
|
id_nodes.append(node)
|
|
|
|
# do not change the output name
|
|
|
|
if name in self.output_names:
|
|
|
|
self.name_node_dict[input].name = name
|
|
|
|
self.name_node_dict[name] = self.name_node_dict[input]
|
|
|
|
del self.name_node_dict[input]
|
|
|
|
self.id_different_scope_dict[name] = input
|
tools/python: add script to convert TensorFlow model (.pb) to native model (.model)
For example, given TensorFlow model file espcn.pb,
to generate native model file espcn.model, just run:
python convert.py espcn.pb
In current implementation, the native model file is generated for
specific dnn network with hard-code python scripts maintained out of ffmpeg.
For example, srcnn network used by vf_sr is generated with
https://github.com/HighVoltageRocknRoll/sr/blob/master/generate_header_and_model.py#L85
In this patch, the script is designed as a general solution which
converts general TensorFlow model .pb file into .model file. The script
now has some tricky to be compatible with current implemention, will
be refined step by step.
The script is also added into ffmpeg source tree. It is expected there
will be many more patches and community needs the ownership of it.
Another technical direction is to do the conversion in c/c++ code within
ffmpeg source tree. While .pb file is organized with protocol buffers,
it is not easy to do such work with tiny c/c++ code, see more discussion
at http://ffmpeg.org/pipermail/ffmpeg-devel/2019-May/244496.html. So,
choose the python script.
Signed-off-by: Guo, Yejun <yejun.guo@intel.com>
6 years ago
|
|
|
else:
|
|
|
|
id_dict[name] = input
|
|
|
|
|
|
|
|
for idnode in id_nodes:
|
|
|
|
self.nodes.remove(idnode)
|
|
|
|
|
|
|
|
for node in self.nodes:
|
|
|
|
for i in range(len(node.input)):
|
|
|
|
input = node.input[i]
|
|
|
|
if input in id_dict:
|
|
|
|
node.input[i] = id_dict[input]
|
|
|
|
|
|
|
|
|
|
|
|
def generate_edges(self):
|
|
|
|
for node in self.nodes:
|
|
|
|
for input in node.input:
|
|
|
|
if input in self.edges:
|
|
|
|
self.edges[input].append(node)
|
|
|
|
else:
|
|
|
|
self.edges[input] = [node]
|
|
|
|
|
|
|
|
|
|
|
|
@staticmethod
|
|
|
|
def get_scope_name(name):
|
|
|
|
index = name.rfind('/')
|
|
|
|
if index == -1:
|
|
|
|
return ""
|
|
|
|
return name[0:index]
|
|
|
|
|
|
|
|
|
|
|
|
def in_conv2d_scope(self, name):
|
|
|
|
inner_scope = TFConverter.get_scope_name(name)
|
|
|
|
if inner_scope == "":
|
|
|
|
return False;
|
|
|
|
for scope in self.conv2d_scope_names:
|
|
|
|
index = inner_scope.find(scope)
|
|
|
|
if index == 0:
|
|
|
|
return True
|
|
|
|
return False
|
|
|
|
|
|
|
|
|
|
|
|
def in_dense_scope(self, name):
|
|
|
|
inner_scope = TFConverter.get_scope_name(name)
|
|
|
|
if inner_scope == "":
|
|
|
|
return False;
|
|
|
|
for scope in self.dense_scope_names:
|
|
|
|
index = inner_scope.find(scope)
|
|
|
|
if index == 0:
|
|
|
|
return True
|
|
|
|
return False
|
|
|
|
|
|
|
|
def generate_sub_block_op_scope_info(self):
|
|
|
|
# mostly, conv2d/dense is a sub block in graph, get the scope name
|
|
|
|
for node in self.nodes:
|
|
|
|
if node.op == 'Conv2D':
|
|
|
|
scope = TFConverter.get_scope_name(node.name)
|
|
|
|
# for the case tf.nn.conv2d is called directly
|
|
|
|
if scope == '':
|
|
|
|
continue
|
|
|
|
# for the case tf.nn.conv2d is called within a scope
|
|
|
|
if scope + '/kernel' not in self.name_node_dict:
|
|
|
|
continue
|
|
|
|
self.conv2d_scope_names.add(scope)
|
|
|
|
elif node.op == 'MatMul':
|
|
|
|
scope = TFConverter.get_scope_name(node.name)
|
|
|
|
# for the case tf.nn.dense is called directly
|
|
|
|
if scope == '':
|
|
|
|
continue
|
|
|
|
# for the case tf.nn.dense is called within a scope
|
|
|
|
if scope + '/kernel' not in self.name_node_dict and scope.split('/Tensordot')[0] + '/kernel' not in self.name_node_dict:
|
|
|
|
continue
|
|
|
|
self.dense_scope_names.add(scope.split('/Tensordot')[0])
|
|
|
|
|
|
|
|
# get the input name to the conv2d/dense sub block
|
|
|
|
for node in self.nodes:
|
|
|
|
scope = TFConverter.get_scope_name(node.name)
|
|
|
|
if scope in self.conv2d_scope_names:
|
|
|
|
if node.op == 'Conv2D' or node.op == 'Shape':
|
|
|
|
for inp in node.input:
|
|
|
|
if TFConverter.get_scope_name(inp) != scope:
|
|
|
|
self.conv2d_scopename_inputname_dict[scope] = inp
|
|
|
|
elif scope in self.dense_scope_names:
|
|
|
|
if node.op == 'MatMul' or node.op == 'Shape':
|
|
|
|
for inp in node.input:
|
|
|
|
if TFConverter.get_scope_name(inp) != scope:
|
|
|
|
self.dense_scopename_inputname_dict[scope] = inp
|
|
|
|
elif scope.split('/Tensordot')[0] in self.dense_scope_names:
|
|
|
|
if node.op == 'Transpose':
|
|
|
|
for inp in node.input:
|
|
|
|
if TFConverter.get_scope_name(inp).find(scope)<0 and TFConverter.get_scope_name(inp).find(scope.split('/')[0])<0:
|
|
|
|
self.dense_scopename_inputname_dict[scope.split('/Tensordot')[0]] = inp
|
|
|
|
|
|
|
|
|
tools/python: add script to convert TensorFlow model (.pb) to native model (.model)
For example, given TensorFlow model file espcn.pb,
to generate native model file espcn.model, just run:
python convert.py espcn.pb
In current implementation, the native model file is generated for
specific dnn network with hard-code python scripts maintained out of ffmpeg.
For example, srcnn network used by vf_sr is generated with
https://github.com/HighVoltageRocknRoll/sr/blob/master/generate_header_and_model.py#L85
In this patch, the script is designed as a general solution which
converts general TensorFlow model .pb file into .model file. The script
now has some tricky to be compatible with current implemention, will
be refined step by step.
The script is also added into ffmpeg source tree. It is expected there
will be many more patches and community needs the ownership of it.
Another technical direction is to do the conversion in c/c++ code within
ffmpeg source tree. While .pb file is organized with protocol buffers,
it is not easy to do such work with tiny c/c++ code, see more discussion
at http://ffmpeg.org/pipermail/ffmpeg-devel/2019-May/244496.html. So,
choose the python script.
Signed-off-by: Guo, Yejun <yejun.guo@intel.com>
6 years ago
|
|
|
def run(self):
|
|
|
|
self.generate_name_node_dict()
|
|
|
|
self.generate_output_names()
|
|
|
|
self.remove_identity()
|
|
|
|
self.generate_edges()
|
|
|
|
self.generate_sub_block_op_scope_info()
|
tools/python: add script to convert TensorFlow model (.pb) to native model (.model)
For example, given TensorFlow model file espcn.pb,
to generate native model file espcn.model, just run:
python convert.py espcn.pb
In current implementation, the native model file is generated for
specific dnn network with hard-code python scripts maintained out of ffmpeg.
For example, srcnn network used by vf_sr is generated with
https://github.com/HighVoltageRocknRoll/sr/blob/master/generate_header_and_model.py#L85
In this patch, the script is designed as a general solution which
converts general TensorFlow model .pb file into .model file. The script
now has some tricky to be compatible with current implemention, will
be refined step by step.
The script is also added into ffmpeg source tree. It is expected there
will be many more patches and community needs the ownership of it.
Another technical direction is to do the conversion in c/c++ code within
ffmpeg source tree. While .pb file is organized with protocol buffers,
it is not easy to do such work with tiny c/c++ code, see more discussion
at http://ffmpeg.org/pipermail/ffmpeg-devel/2019-May/244496.html. So,
choose the python script.
Signed-off-by: Guo, Yejun <yejun.guo@intel.com>
6 years ago
|
|
|
|
|
|
|
if self.dump4tb:
|
|
|
|
self.dump_for_tensorboard()
|
tools/python: add script to convert TensorFlow model (.pb) to native model (.model)
For example, given TensorFlow model file espcn.pb,
to generate native model file espcn.model, just run:
python convert.py espcn.pb
In current implementation, the native model file is generated for
specific dnn network with hard-code python scripts maintained out of ffmpeg.
For example, srcnn network used by vf_sr is generated with
https://github.com/HighVoltageRocknRoll/sr/blob/master/generate_header_and_model.py#L85
In this patch, the script is designed as a general solution which
converts general TensorFlow model .pb file into .model file. The script
now has some tricky to be compatible with current implemention, will
be refined step by step.
The script is also added into ffmpeg source tree. It is expected there
will be many more patches and community needs the ownership of it.
Another technical direction is to do the conversion in c/c++ code within
ffmpeg source tree. While .pb file is organized with protocol buffers,
it is not easy to do such work with tiny c/c++ code, see more discussion
at http://ffmpeg.org/pipermail/ffmpeg-devel/2019-May/244496.html. So,
choose the python script.
Signed-off-by: Guo, Yejun <yejun.guo@intel.com>
6 years ago
|
|
|
|
|
|
|
self.dump_to_file()
|
|
|
|
|
|
|
|
|
|
|
|
def convert_from_tensorflow(infile, outfile, dump4tb):
|
tools/python: add script to convert TensorFlow model (.pb) to native model (.model)
For example, given TensorFlow model file espcn.pb,
to generate native model file espcn.model, just run:
python convert.py espcn.pb
In current implementation, the native model file is generated for
specific dnn network with hard-code python scripts maintained out of ffmpeg.
For example, srcnn network used by vf_sr is generated with
https://github.com/HighVoltageRocknRoll/sr/blob/master/generate_header_and_model.py#L85
In this patch, the script is designed as a general solution which
converts general TensorFlow model .pb file into .model file. The script
now has some tricky to be compatible with current implemention, will
be refined step by step.
The script is also added into ffmpeg source tree. It is expected there
will be many more patches and community needs the ownership of it.
Another technical direction is to do the conversion in c/c++ code within
ffmpeg source tree. While .pb file is organized with protocol buffers,
it is not easy to do such work with tiny c/c++ code, see more discussion
at http://ffmpeg.org/pipermail/ffmpeg-devel/2019-May/244496.html. So,
choose the python script.
Signed-off-by: Guo, Yejun <yejun.guo@intel.com>
6 years ago
|
|
|
with open(infile, 'rb') as f:
|
|
|
|
# read the file in .proto format
|
|
|
|
graph_def = tf.GraphDef()
|
|
|
|
graph_def.ParseFromString(f.read())
|
|
|
|
nodes = graph_def.node
|
|
|
|
|
|
|
|
converter = TFConverter(graph_def, nodes, outfile, dump4tb)
|
tools/python: add script to convert TensorFlow model (.pb) to native model (.model)
For example, given TensorFlow model file espcn.pb,
to generate native model file espcn.model, just run:
python convert.py espcn.pb
In current implementation, the native model file is generated for
specific dnn network with hard-code python scripts maintained out of ffmpeg.
For example, srcnn network used by vf_sr is generated with
https://github.com/HighVoltageRocknRoll/sr/blob/master/generate_header_and_model.py#L85
In this patch, the script is designed as a general solution which
converts general TensorFlow model .pb file into .model file. The script
now has some tricky to be compatible with current implemention, will
be refined step by step.
The script is also added into ffmpeg source tree. It is expected there
will be many more patches and community needs the ownership of it.
Another technical direction is to do the conversion in c/c++ code within
ffmpeg source tree. While .pb file is organized with protocol buffers,
it is not easy to do such work with tiny c/c++ code, see more discussion
at http://ffmpeg.org/pipermail/ffmpeg-devel/2019-May/244496.html. So,
choose the python script.
Signed-off-by: Guo, Yejun <yejun.guo@intel.com>
6 years ago
|
|
|
converter.run()
|