blob: 476c2e239fbd18e5902a0c66711daff6e73722fa (
plain)
1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
20
21
22
23
24
25
26
27
28
29
30
31
32
33
34
35
36
37
|
#!/bin/bash -x
#####################################################################
# SUMMARY: Train a Transformer language model
# AUTHOR: snukky
# TAGS: lm transformer
#####################################################################
# Exit on error
set -e
# Test code goes here
rm -rf lm-transformer lm-transformer.log
mkdir -p lm-transformer
$MRT_MARIAN/marian \
--seed 1111 --no-shuffle \
--type lm-transformer --dim-emb 128 --dim-rnn 256 --cost-type ce-mean \
-m lm-transformer/model.npz -t $MRT_DATA/europarl.de-en/corpus.bpe.en -v vocab.en.yml \
--disp-freq 20 --after-batches 100 \
--log lm-transformer.log
test -e lm-transformer/model.npz
test -e lm-transformer/model.npz.yml
test -e lm-transformer.log
cat lm-transformer.log | grep 'Ep\. 1 :' | $MRT_TOOLS/extract-costs.sh > lm-transformer.out
$MRT_TOOLS/diff-nums.py lm-transformer.out lm-transformer.expected -p 0.02 -o lm-transformer.diff
# Scoring with LM
test -s temp.bpe.en || tail $MRT_DATA/europarl.de-en/corpus.bpe.en > test.bpe.en
$MRT_MARIAN/marian-scorer -m lm-transformer/model.npz -t test.bpe.en -v vocab.en.yml > lm-transformer.scores.out
$MRT_TOOLS/diff-nums.py lm-transformer.scores.out lm-transformer.scores.expected -p 0.002 -o lm-transformer.scores.diff
# Exit with success code
exit 0
|