blob: 7e8dc5eca85a92f9542b85b9f5be2e5554055eff (
plain)
1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
20
21
22
23
24
25
26
27
28
29
30
31
32
33
34
35
36
37
38
39
40
41
42
43
44
45
46
47
48
49
50
51
52
53
54
55
56
57
58
59
60
61
62
63
64
65
66
67
|
#!/bin/bash -x
#####################################################################
# SUMMARY:
# TAGS: mini-batch-fit
#####################################################################
# Exit on error
set -e
# Test code goes here
rm -rf corpus_fit_transformer_sync corpus_fit_transformer_sync*.log
mkdir -p corpus_fit_transformer_sync
test -e vocab.de.yml
test -e vocab.en.yml
extra_opts="--seed 5555 --maxi-batch 8 --maxi-batch-sort src --mini-batch 32 --mini-batch-fit -w 100 --optimizer sgd --dim-emb 128 --transformer-dim-ffn 256 --disp-freq 4 --type transformer --sync-sgd"
# Added because default options has changes
extra_opts="$extra_opts --cost-type ce-mean --disp-label-counts false"
# Step 1: Train a model in one go, up to the update no. 70, and save training logs
$MRT_MARIAN/marian \
-m corpus_fit_transformer_sync/model_full.npz -t $MRT_DATA/train.max50.{en,de} -v vocab.en.yml vocab.de.yml \
--after-batches 80 $extra_opts \
--log corpus_fit_transformer_sync.log
test -e corpus_fit_transformer_sync/model_full.npz
test -e corpus_fit_transformer_sync.log
cat corpus_fit_transformer_sync.log | $MRT_TOOLS/strip-timestamps.sh | grep "Ep\. " | sed 's/ : Time.*//' > corpus_fit_transformer_sync.expected
# Step 2: Train a new model from scratch, but only to the update no. 40, and save the model
$MRT_MARIAN/marian \
-m corpus_fit_transformer_sync/model.npz -t $MRT_DATA/train.max50.{en,de} -v vocab.en.yml vocab.de.yml \
--after-batches 60 $extra_opts \
--log corpus_fit_transformer_sync_1.log
test -e corpus_fit_transformer_sync/model.npz
test -e corpus_fit_transformer_sync_1.log
cat corpus_fit_transformer_sync_1.log | $MRT_TOOLS/strip-timestamps.sh | grep "Ep\. " | sed 's/ : Time.*//' > corpus_fit_transformer_sync_1.out
cp corpus_fit_transformer_sync/model.npz.yml corpus_fit_transformer_sync/model.npz.1.yml
cp corpus_fit_transformer_sync/model.npz.progress.yml corpus_fit_transformer_sync/model.npz.progress.1.yml
# Step 3: Restart the training from step 2 and continue up to the update no. 70, and save training logs
$MRT_MARIAN/marian \
-m corpus_fit_transformer_sync/model.npz -t $MRT_DATA/train.max50.{en,de} -v vocab.en.yml vocab.de.yml \
--after-batches 80 $extra_opts \
--log corpus_fit_transformer_sync_2.log
test -e corpus_fit_transformer_sync/model.npz
test -e corpus_fit_transformer_sync_2.log
# Step 4: Combine training logs from steps 2 and 3 and compare them with logs from step 1
cat corpus_fit_transformer_sync_2.log | $MRT_TOOLS/strip-timestamps.sh | grep "Ep\. " | sed 's/ : Time.*//' > corpus_fit_transformer_sync_2.out
cat corpus_fit_transformer_sync_1.out corpus_fit_transformer_sync_2.out > corpus_fit_transformer_sync.out
$MRT_TOOLS/diff-nums.py corpus_fit_transformer_sync.out corpus_fit_transformer_sync.expected -p 0.1 -o corpus_fit_transformer_sync.diff
# Exit with success code
exit 0
|