Files
exercises-in-programming-style/35-dnn-no-learning/tf-35.py
2019-12-26 10:35:28 -08:00

94 lines
2.7 KiB
Python

from keras.models import Model
from keras import layers
from keras.layers import Input, Dense
from keras.utils import plot_model
import numpy as np
import sys, os, string
characters = sorted(string.printable)
char_indices = dict((c, i) for i, c in enumerate(characters))
indices_char = dict((i, c) for i, c in enumerate(characters))
INPUT_VOCAB_SIZE = len(characters)
LINE_SIZE = 100
def encode_one_hot(s):
all = []
for c in s:
if c not in characters:
continue
x = np.zeros((INPUT_VOCAB_SIZE))
index = char_indices[c]
x[index] = 1
all.append(x)
return all
def decode_one_hot(x):
s = []
for onehot in x:
one_index = np.where(onehot == 1) # tuple of two things
if len(one_index[1]) > 0:
n = one_index[1][0]
c = indices_char[n]
s.append(c)
return ''.join(s)
def normalization_layer_set_weights(n_layer):
wb = []
b = np.zeros((INPUT_VOCAB_SIZE), dtype=np.float32)
w = np.zeros((INPUT_VOCAB_SIZE, INPUT_VOCAB_SIZE), dtype=np.float32)
# Let lower case letters go through
for c in string.ascii_lowercase:
i = char_indices[c]
w[i, i] = 1
# Map capitals to lower case
for c in string.ascii_uppercase:
i = char_indices[c]
il = char_indices[c.lower()]
w[i, il] = 1
# Map all non-letters to space
sp_idx = char_indices[' ']
for c in [c for c in list(string.printable) if c not in list(string.ascii_letters)]:
i = char_indices[c]
w[i, sp_idx] = 1
wb.append(w)
wb.append(b)
n_layer.set_weights(wb)
return n_layer
def build_model():
# Normalize characters using a shared dense model
n_layer = Dense(INPUT_VOCAB_SIZE)
raw_inputs = []
normalized_outputs = []
for _ in range(0, LINE_SIZE):
input_char = Input(shape=(INPUT_VOCAB_SIZE, ))
filtered_char = n_layer(input_char)
raw_inputs.append(input_char)
normalized_outputs.append(filtered_char)
normalization_layer_set_weights(n_layer)
model = Model(inputs=raw_inputs, outputs=normalized_outputs)
return model
model = build_model()
plot_model(model, to_file='normalization.png', show_shapes=True)
with open(sys.argv[1]) as f:
for line in f:
if line.isspace(): continue
onehots = encode_one_hot(line)
data = [[] for _ in range(LINE_SIZE)]
for i, c in enumerate(onehots):
data[i].append(c)
for j in range(len(onehots), LINE_SIZE):
data[j].append(np.zeros((INPUT_VOCAB_SIZE)))
inputs = [np.array(e) for e in data]
preds = model.predict(inputs)
normal = decode_one_hot(preds)
print(normal)