Skip to content

Commit

Permalink
remove tokenizer filter
Browse files Browse the repository at this point in the history
  • Loading branch information
demdecuong committed Sep 29, 2020
1 parent 9b2ebaf commit e44ff41
Showing 1 changed file with 1 addition and 2 deletions.
3 changes: 1 addition & 2 deletions src/loader.py
Original file line number Diff line number Diff line change
Expand Up @@ -47,8 +47,7 @@ def load_data(data_path):
train_images, valid_images, train_texts, valid_texts, all_text = load_csv(data_path)

tokenizer = tf.keras.preprocessing.text.Tokenizer(num_words=config.top_k,
oov_token="<unk>",
filters='!"#$%&()*+.,-/:;=?@[\]^_`{|}~ ')
oov_token="<unk>")

tokenizer.fit_on_texts(all_text)
tokenizer.word_index['<pad>'] = 0
Expand Down

0 comments on commit e44ff41

Please sign in to comment.