Notes taken while working through the 'Deep MNIST for Experts' tutorial.

- 'Placeholders' in TensorFlow feel like a nicer abstraction than 'symbols' in Theano.
- Automatic tensor shape checking is a great idea. In fact, any kind of optional static analysis of your computation graph is a great idea.
`cross_entropy = -tf.reduce_sum(y_*tf.log(y))`

: sweet`train_step = tf.train.GradientDescentOptimizer(0.01).minimize(cross_entropy)`

: sweet- What does it mean for a tensor to have -1 for a shape dimension?
- Dropout seems easier to apply than in Lasagne - no 'deterministic' flag equivalent needed when predicting.
- Running the conv net example on my Macbook (where only the CPU version of TensorFlow is available) is wicked slow (dude). Not unexpected I guess.

All in all, Tensorflow feels lot like Theano + Lasagne, but with a much cleaner API. I'll need to try it out on the lab's GPU machine before I can say anything about speed, but if it's quick enough, I'd consider switching over.