Created by: reyoung
It is easy to make gradient exploration by dynamic RNN. We should give a standard gradient check for Dynamic RNN(i.e. the combination of while op, shrink mem, etc.).