The previous post about reinforcement learning:
[Reinforcement Learning] Get started to learn gradient method for reinforcement learning
For the Q-Learning tutorial, I refer to these as follows: ( sorry, they are written in Chinese )
https://morvanzhou.github.io/tutorials/machine-learning/reinforcement-learning/2-2-A-q-learning/
https://morvanzhou.github.io/tutorials/machine-learning/reinforcement-learning/2-2-tabular-q1/
https://morvanzhou.github.io/tutorials/machine-learning/reinforcement-learning/2-3-tabular-q2/
Thursday, November 22, 2018
Wednesday, November 21, 2018
[Reinforcement Learning] Get started to learn policy gradient method for reinforcement learning
This post is about my first time to learn policy gradient method for reinforcement learning. Basically, there are already a lot of materials on the internet, but in this time, I only want to focus on a tutorial as follows: ( sorry, they are written in Chinese )
https://morvanzhou.github.io/tutorials/machine-learning/reinforcement-learning/5-1-policy-gradient-softmax1/
https://morvanzhou.github.io/tutorials/machine-learning/reinforcement-learning/5-2-policy-gradient-softmax2/
https://morvanzhou.github.io/tutorials/machine-learning/reinforcement-learning/5-1-policy-gradient-softmax1/
https://morvanzhou.github.io/tutorials/machine-learning/reinforcement-learning/5-2-policy-gradient-softmax2/
Thursday, November 15, 2018
[RNN] What are the difference of input and output's tensor shape in dynamic_rnn and static_rnn using TensorFlow
When studying RNN, my first issue encountered in my program is about the shape of input and output tensors. Shape is a very important information to connect between layers. Here I just directly point out what are differences in input/output shape of static RNN and dynamic RNN.
P.S: If you use Keras to write your RNN model, you won't need to deal with these details.
P.S: If you use Keras to write your RNN model, you won't need to deal with these details.
Tuesday, November 13, 2018
[TensorFlow] The explanation of average gradients by example in data parallelism
When studying some examples of training model using Multi-GPUs ( in data parallelism ), the average gradients function always exists in some kind of ways, and here is a simple version as follows:
Thursday, November 8, 2018
[Dynamic Control Flow] Whitepaper: Implementation of Control Flow in TensorFlow
In the following whitepaper, we can understand more dynamic control flow in details.
Whitepaper: Implementation of Control Flow in TensorFlow
http://download.tensorflow.org/paper/white_paper_tf_control_flow_implementation_2017_11_1.pdf
Whitepaper: Implementation of Control Flow in TensorFlow
http://download.tensorflow.org/paper/white_paper_tf_control_flow_implementation_2017_11_1.pdf
Tuesday, October 30, 2018
[TensorFlow] Train in Tensorflow and do inference with the trained model
If you want to train your model in Tensorflow and do inference with the trained model, you can refer to this post.
[ONNX] Train in Tensorflow and export to ONNX (Part II)
https://danny270degree.blogspot.com/2018/08/onnx-train-in-tensorflow-and-export-to_20.html
So, after training, you will get these files:
1. Train your model
I will use the simple CNN model in my previous post:[ONNX] Train in Tensorflow and export to ONNX (Part II)
https://danny270degree.blogspot.com/2018/08/onnx-train-in-tensorflow-and-export-to_20.html
So, after training, you will get these files:
my_mnist/
├── checkpoint
├── graph.pbtxt
├── my_mnist_model.data-00000-of-00001
├── my_mnist_model.index
└── my_mnist_model.meta
Wednesday, October 24, 2018
[LLVM] LLVM studying list for newbie
If you are an LLVM newbie and are interested in LLVM like me, you may take a look at my LLVM studying list. It takes time for me to search the related resources and documents. So, I think it will help somehow. By the way, most of my list items are written in Chinese so that those who are native Engish speakers may not suit for this.
Tuesday, October 23, 2018
[TensorFlow] Does it help the processing time and transmission time if increasing CUDA Steam number in TensorFlow?
Before starting to increase the CUDA Steam number in TensorFlow, I want to recap some ideas about the Executor module. When TensorFlow session runs, it will build Executor. Meanwhile, if you enable CUDA in TensorFlow build configuration, the Executor will add visible GPU devices and create TF device object (GPUDevice object) mapping to physical GPU device. There are 4 kinds of streams inside GPUDevice:
- CUDA stream
- Host_to_Device stream
- Device_to_Host stream
- Device_to_Device stream
Subscribe to:
Posts (Atom)