Commit
·
c5dc1d4
1
Parent(s):
6543d58
Seq2Seq model implemented
Browse files- .idea/.gitignore +3 -0
- .idea/bn_multi_tribe_mt.iml +8 -0
- .idea/inspectionProfiles/profiles_settings.xml +6 -0
- .idea/misc.xml +7 -0
- .idea/modules.xml +8 -0
- .idea/vcs.xml +6 -0
- src/__init__.py +0 -0
- src/pipes/data.py +14 -13
- src/{seq2seqtrainer.py → seq2seq_trainer.py} +25 -9
.idea/.gitignore
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
# Default ignored files
|
2 |
+
/shelf/
|
3 |
+
/workspace.xml
|
.idea/bn_multi_tribe_mt.iml
ADDED
@@ -0,0 +1,8 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
1 |
+
<?xml version="1.0" encoding="UTF-8"?>
|
2 |
+
<module type="PYTHON_MODULE" version="4">
|
3 |
+
<component name="NewModuleRootManager">
|
4 |
+
<content url="file://$MODULE_DIR$" />
|
5 |
+
<orderEntry type="inheritedJdk" />
|
6 |
+
<orderEntry type="sourceFolder" forTests="false" />
|
7 |
+
</component>
|
8 |
+
</module>
|
.idea/inspectionProfiles/profiles_settings.xml
ADDED
@@ -0,0 +1,6 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
1 |
+
<component name="InspectionProjectProfileManager">
|
2 |
+
<settings>
|
3 |
+
<option name="USE_PROJECT_PROFILE" value="false" />
|
4 |
+
<version value="1.0" />
|
5 |
+
</settings>
|
6 |
+
</component>
|
.idea/misc.xml
ADDED
@@ -0,0 +1,7 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
1 |
+
<?xml version="1.0" encoding="UTF-8"?>
|
2 |
+
<project version="4">
|
3 |
+
<component name="Black">
|
4 |
+
<option name="sdkName" value="C:\ProgramData\miniconda3" />
|
5 |
+
</component>
|
6 |
+
<component name="ProjectRootManager" version="2" project-jdk-name="C:\ProgramData\miniconda3" project-jdk-type="Python SDK" />
|
7 |
+
</project>
|
.idea/modules.xml
ADDED
@@ -0,0 +1,8 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
1 |
+
<?xml version="1.0" encoding="UTF-8"?>
|
2 |
+
<project version="4">
|
3 |
+
<component name="ProjectModuleManager">
|
4 |
+
<modules>
|
5 |
+
<module fileurl="file://$PROJECT_DIR$/.idea/bn_multi_tribe_mt.iml" filepath="$PROJECT_DIR$/.idea/bn_multi_tribe_mt.iml" />
|
6 |
+
</modules>
|
7 |
+
</component>
|
8 |
+
</project>
|
.idea/vcs.xml
ADDED
@@ -0,0 +1,6 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
1 |
+
<?xml version="1.0" encoding="UTF-8"?>
|
2 |
+
<project version="4">
|
3 |
+
<component name="VcsDirectoryMappings">
|
4 |
+
<mapping directory="" vcs="Git" />
|
5 |
+
</component>
|
6 |
+
</project>
|
src/__init__.py
ADDED
File without changes
|
src/pipes/data.py
CHANGED
@@ -1,8 +1,9 @@
|
|
1 |
import random
|
2 |
-
import const
|
3 |
-
import utils
|
4 |
import string
|
5 |
-
|
|
|
6 |
|
7 |
class SequenceLoader:
|
8 |
def __init__(self):
|
@@ -40,8 +41,9 @@ class SequenceLoader:
|
|
40 |
|
41 |
|
42 |
def serialize(src_seq, tar_seq):
|
43 |
-
tar_seq_in = tar_seq[:, :-1]
|
44 |
-
tar_seq_out = tar_seq[:, 1:]
|
|
|
45 |
return (src_seq, tar_seq_in), tar_seq_out
|
46 |
|
47 |
|
@@ -165,28 +167,27 @@ class Dataset:
|
|
165 |
self.dataset_dict = seq_processor.get_dict()
|
166 |
|
167 |
def pull(self):
|
168 |
-
src_lang_train_seqs = self.dataset_dict[self.langs[0]]["train"]
|
169 |
-
tar_lang_train_seqs = self.dataset_dict[self.langs[1]]["train"]
|
170 |
|
171 |
-
src_lang_val_seqs = self.dataset_dict[self.langs[0]]["val"]
|
172 |
-
tar_lang_val_seqs = self.dataset_dict[self.langs[1]]["val"]
|
173 |
|
174 |
train_ds = ((tf.data.Dataset
|
175 |
.from_tensor_slices((src_lang_train_seqs, tar_lang_train_seqs)))
|
176 |
.shuffle(const.BUFFER_SIZE)
|
177 |
.batch(const.BATCH_SIZE))
|
178 |
|
179 |
-
val_ds = (tf.data.Dataset
|
180 |
-
|
181 |
.shuffle(const.BUFFER_SIZE)
|
182 |
.batch(const.BATCH_SIZE))
|
183 |
|
184 |
train_ds = train_ds.map(serialize, tf.data.AUTOTUNE)
|
185 |
val_ds = val_ds.map(serialize, tf.data.AUTOTUNE)
|
186 |
|
187 |
-
return
|
188 |
|
189 |
-
@staticmethod
|
190 |
def get_dict(self):
|
191 |
return self.dataset_dict
|
192 |
|
|
|
1 |
import random
|
2 |
+
from pipes import const
|
3 |
+
from pipes import utils
|
4 |
import string
|
5 |
+
import tensorflow as tf
|
6 |
+
import numpy as np
|
7 |
|
8 |
class SequenceLoader:
|
9 |
def __init__(self):
|
|
|
41 |
|
42 |
|
43 |
def serialize(src_seq, tar_seq):
|
44 |
+
tar_seq_in = tf.convert_to_tensor(tar_seq[:, :-1])
|
45 |
+
tar_seq_out = tf.convert_to_tensor(tar_seq[:, 1:])
|
46 |
+
src_seq = tf.convert_to_tensor(src_seq)
|
47 |
return (src_seq, tar_seq_in), tar_seq_out
|
48 |
|
49 |
|
|
|
167 |
self.dataset_dict = seq_processor.get_dict()
|
168 |
|
169 |
def pull(self):
|
170 |
+
src_lang_train_seqs = np.array(self.dataset_dict[self.langs[0]]["train"])
|
171 |
+
tar_lang_train_seqs = np.array(self.dataset_dict[self.langs[1]]["train"])
|
172 |
|
173 |
+
src_lang_val_seqs = np.array(self.dataset_dict[self.langs[0]]["val"])
|
174 |
+
tar_lang_val_seqs = np.array(self.dataset_dict[self.langs[1]]["val"])
|
175 |
|
176 |
train_ds = ((tf.data.Dataset
|
177 |
.from_tensor_slices((src_lang_train_seqs, tar_lang_train_seqs)))
|
178 |
.shuffle(const.BUFFER_SIZE)
|
179 |
.batch(const.BATCH_SIZE))
|
180 |
|
181 |
+
val_ds = ((tf.data.Dataset
|
182 |
+
.from_tensor_slices((src_lang_val_seqs, tar_lang_val_seqs)))
|
183 |
.shuffle(const.BUFFER_SIZE)
|
184 |
.batch(const.BATCH_SIZE))
|
185 |
|
186 |
train_ds = train_ds.map(serialize, tf.data.AUTOTUNE)
|
187 |
val_ds = val_ds.map(serialize, tf.data.AUTOTUNE)
|
188 |
|
189 |
+
return train_ds, val_ds
|
190 |
|
|
|
191 |
def get_dict(self):
|
192 |
return self.dataset_dict
|
193 |
|
src/{seq2seqtrainer.py → seq2seq_trainer.py}
RENAMED
@@ -1,5 +1,8 @@
|
|
1 |
-
from pipes import
|
|
|
|
|
2 |
from pipes.data import Dataset
|
|
|
3 |
|
4 |
if __name__ == "__main__":
|
5 |
input_lang = 'gr'
|
@@ -8,17 +11,30 @@ if __name__ == "__main__":
|
|
8 |
dataset_object = Dataset([input_lang, output_lang])
|
9 |
dataset_object.pack()
|
10 |
dataset_object.process()
|
|
|
11 |
dataset_dict = dataset_object.get_dict()
|
12 |
|
13 |
-
|
14 |
input_vocab_size=dataset_dict[input_lang]["vocab_size"],
|
15 |
output_vocab_size=dataset_dict[output_lang]["vocab_size"],
|
16 |
embedding_dim=256,
|
17 |
-
hidden_units=
|
18 |
-
|
19 |
-
|
20 |
-
|
21 |
-
|
22 |
-
|
23 |
-
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
24 |
)
|
|
|
1 |
+
from pipes import utils
|
2 |
+
from pipes import const
|
3 |
+
from pipes import models
|
4 |
from pipes.data import Dataset
|
5 |
+
import tensorflow as tf
|
6 |
|
7 |
if __name__ == "__main__":
|
8 |
input_lang = 'gr'
|
|
|
11 |
dataset_object = Dataset([input_lang, output_lang])
|
12 |
dataset_object.pack()
|
13 |
dataset_object.process()
|
14 |
+
train_ds, val_ds = dataset_object.pull()
|
15 |
dataset_dict = dataset_object.get_dict()
|
16 |
|
17 |
+
model_object = models.Seq2Seq(
|
18 |
input_vocab_size=dataset_dict[input_lang]["vocab_size"],
|
19 |
output_vocab_size=dataset_dict[output_lang]["vocab_size"],
|
20 |
embedding_dim=256,
|
21 |
+
hidden_units=512
|
22 |
+
)
|
23 |
+
|
24 |
+
model_object.build()
|
25 |
+
model = model_object.get()
|
26 |
+
|
27 |
+
model.compile(
|
28 |
+
optimizer=tf.keras.optimizers.Adam(),
|
29 |
+
loss=tf.keras.losses.SparseCategoricalCrossentropy(),
|
30 |
+
metrics=['accuracy', 'val_accuracy'],
|
31 |
+
)
|
32 |
+
|
33 |
+
history = model.fit(
|
34 |
+
train_ds.repeat(),
|
35 |
+
epochs=10,
|
36 |
+
steps_per_epoch=100,
|
37 |
+
validation_steps=20,
|
38 |
+
validation_data=val_ds,
|
39 |
+
callbacks=[tf.keras.callbacks.EarlyStopping(patience=3)]
|
40 |
)
|