Martyn Garcia
https://github.com/255BITS/HyperGAN
Supervised data for a dog dataset:
* each image labelled by breed
* hard, error prone
Unsupervised data for a dog dataset:
* scrape reddit/r/dogpics
What type of layer does this function create?
tf.nn.xw_plus_b
What type of layer does this function create?
tf.nn.xw_plus_b
Linear layer
Character
vector
RNN
Softmax
Character vectors example
a | (1 0 0) |
b | (0 1 0) |
c | (0 0 1) |
x = np.eye(vocab)[char]
tf_x = tf.constant(x)
sess.run(tf_x)
RNN
tf.nn.rnn_cell.LSTMCell
tf.nn.rnn_cell.GRUCell
Softmax loss
net = rnn(input)
net = linear(net, vocab_size)
loss = tf.nn.softmax_cross_entropy_with_logits(net, y)
Example output
Drawbacks:
Sampling
Invalid character combinations
No higher level meaning
What is this called?
net = tf.maximum(0, net)
tf.nn.relu
https://github.com/hardmaru/sketch-rnn
https://arxiv.org/pdf/1308.0850v5.pdf
Stroke
vector
RNN
Mixture
Density
Network
Locally connected layers with shared weights
https://github.com/vdumoulin/conv_arithmetic
What is this called?
net = tf.maximum(net, 0.2*net)
Leaky Relu!
def lrelu(x, leak=0.2, name="lrelu"): | |
return tf.maximum(x, leak*x) | |
https://github.com/ibab/tensorflow-wavenet
Generates audio using a pixel-cnn inspired architecture.
Incredible results from Google
https://deepmind.com/blog/wavenet-generative-model-raw-audio/
Encoder
z
Generator
Loss
MSE(G(z), x)
x
tf.square(G(z)-x)
MSE in Tensorflow
64x64x3
128
64x64x3
Encoder
z
Generator
x
Problem: This doesn't work well
z = 2
http://github.com/255bits/hyperchamber
z
Generator
gloss
Discriminator
x
dloss
Xent(D(G(z)),1)
Xent(D(x),1) + Xent(D(G(z)),0)
z
Generator
Discriminator
x
This, and the variations that are now being proposed is the most interesting idea in the last 10 years in ML, in my opinion.
LeCun https://www.quora.com/What-are-some-recent-and-potentially-upcoming-breakthroughs-in-deep-learning