In the previous post as the below link, I mentioned that the default value of rewrite_config seems to change a little bit.
https://danny270degree.blogspot.com/2018/06/tensorflow-compare-memory-options-in.html
To clarify my doubt, I check the TensorFlow's memory_optimizer.cc and arrange the mapping table:
Friday, August 17, 2018
Thursday, August 16, 2018
[TensorFlow] How to print the timestamp of a node/operation of computation graph in run-time?
When some people first time tries to debug or print out information of some result from a node/operation in the computation graph in TensorFlow, they maybe confuse about how to do it. Fortunately, someone in Google gave a great explanation of the print function:
https://towardsdatascience.com/using-tf-print-in-tensorflow-aa26e1cff11e
After reading it, you should understand how tf.Print() function works and to use it.
https://towardsdatascience.com/using-tf-print-in-tensorflow-aa26e1cff11e
After reading it, you should understand how tf.Print() function works and to use it.
Wednesday, August 8, 2018
[ONNX] Use ONNX_TF and nGraph_ONNX to do inference/prediction with ONNX model
Here I try to use the pre-trained model from ONNX model zoo, which the models are already converted from some deep learning framework. So I download the Resnet50 model from the following URL and untar it:
wget https://s3.amazonaws.com/download.onnx/models/opset_8/resnet50.tar.gz
tar -xzvf resnet50.tar.gz
P.S: pre-trained ONNX models: https://github.com/onnx/modelsThen, I can do the inference/prediction using this ONNX model in two ways:
[ONNX] Train in Tensorflow and export to ONNX (Part I)
From my point of view, ONNX is a model description spec and ONNX model needs Deep Learning framework or backend tool/compiler which supports it to run.
The advantage of ONNX as I know is about portable and exchangeable between DL frameworks.
Here I will use this tutorial to convert TensorFlow's model to ONNX model by myself.
https://github.com/onnx/tutorials/blob/master/tutorials/OnnxTensorflowExport.ipynb
The advantage of ONNX as I know is about portable and exchangeable between DL frameworks.
Here I will use this tutorial to convert TensorFlow's model to ONNX model by myself.
https://github.com/onnx/tutorials/blob/master/tutorials/OnnxTensorflowExport.ipynb
Tuesday, July 31, 2018
[Fun] compress and composite dataset to one image file
Tuesday, July 17, 2018
[Confusion Matrix] How to calculate confusion matrix, precision and recall list from scratch
I directly give an example which is with 10 categories, such as CIFAR-10 and MNIST. It explains how to calculate the confusion matrix, precision and recall list from scratch in Python. My data is generated at random. You should replace by yours. Here it goes:
Saturday, July 14, 2018
[Qt5] How to develop Qt5 GUI with TensorFlow C++ library?
Here I give a simple and complete example of how to develop Qt5 GUI with TensorFlow C++ library on Linux platform. Please check out my GitHub's repository as follow:
https://github.com/teyenliu/tf_inference_gui
https://github.com/teyenliu/tf_inference_gui
Monday, July 9, 2018
[TensorFlow] How to implement LMDBDataset in tf.data API?
Subscribe to:
Posts (Atom)