----- Forwarded Message -----
From: Hieu Hoang <Hieu.Hoang@ed.ac.uk>
To: samira tofighi <samiratofighi@gmail.com>
Cc: moses-support <moses-support@mit.edu>
Sent: Wednesday, 14 August 2013, 19:57
Subject: Re: [Moses-support] using multiple language models
From: Hieu Hoang <Hieu.Hoang@ed.ac.uk>
To: samira tofighi <samiratofighi@gmail.com>
Cc: moses-support <moses-support@mit.edu>
Sent: Wednesday, 14 August 2013, 19:57
Subject: Re: [Moses-support] using multiple language models
hi samira
how recent is the moses source code you're using? Can you please 'git pull' and see it it's still a problem.
On 14 August 2013 13:15, samira tofighi <samiratofighi@gmail.com> wrote:
Problem is exist still and the decode file output is attached.thanks for your answers.
On Wed, Aug 14, 2013 at 3:43 PM, Hieu Hoang <Hieu.Hoang@ed.ac.uk> wrote:
hi samiracan you try using this ini file instead of your version. The ini file format has recently changed. It should be able to read your ini file, but there might have been a bug in the conversion from old to new format.
Please tell me if it works--
On 14 August 2013 11:51, samira tofighi <samiratofighi@gmail.com> wrote:
I got these STDERR output file:Hi,with this configuration file:
# MERT optimized configuration
# decoder /home/s.tofighi/mosesdecoder/bin/moses
# BLEU 0.261756 on dev /Share/local/s.tofighi/program_moses/bridge_defa/fa_en/corpora/dev/corpus0
# We were before running iteration 6
# finished Tue Apr 9 17:28:22 IRDT 2013
### MOSES CONFIG FILE ###
#########################
# input factors
[input-factors]
0
# mapping steps
[mapping]
0 T 0
1 T 1
# translation tables: table type (hierarchical(0), textual (0), binary (1)), source-factors, target-factors, number of scores, file
# OLD FORMAT is still handled for back-compatibility
# OLD FORMAT translation tables: source-factors, target-factors, number of scores, file
# OLD FORMAT a binary table type (1) is assumed
[ttable-file]
0 0 0 5 /home/samira/myprogs/moses_progs/cnn_28000/working2/train/model/phrase-table.gz
0 0 0 5 /home/samira/myprogs/snover_sst/sys4/working/train/model/phrase-table.gz
# no generation models, no generation-file section
# language models: type(srilm/irstlm), factors, order, file
[lmodel-file]
8 0 3 /home/samira/myprogs/moses_progs/cnn_28000/lm2/cnn5_28000.blm.en
8 0 3 /home/samira/myprogs/moses_progs/cnn_28000/lm2/cnn5_28000.blm2.en
# limit on how many phrase translations e for each phrase f are loaded
# 0 = all elements loaded
[ttable-limit]
40
# distortion (reordering) files
[distortion-file]
0-0 wbe-msd-bidirectional-fe-allff 6 /home/samira/myprogs/moses_progs/cnn_28000/working2/train/model/reordering-table.wbe-msd-bidirectional-fe.gz
# distortion (reordering) weight
[weight-d]
0.0153922
0.072167
0.0335072
0.0432684
0.0614601
0.0181157
0.0878144
# language model weights
[weight-l]
0.1022989
0.1222989
# translation model weights
[weight-t]
0.000153608
0.122284
0.057549
0.0621673
0.0209181
0.00153608
0.122284
0.057549
0.0621673
0.0209181
# no generation models, no weight-generation section
# word penalty
[weight-w]
-0.332904
[distortion-limit]
6
[decoding-graph-backoff]
0
1
nohup: ignoring input
Defined parameters (per moses.ini or switch):
config: working2/train/model/moses.ini
decoding-graph-backoff: 0 1
distortion-file: 0-0 wbe-msd-bidirectional-fe-allff 6 /home/samira/myprogs/moses_progs/cnn_28000/working2/train/model/reordering-table.wbe-msd-bidirectional-fe.gzdistortion-limit: 5
drop-unknown:
input-factors: 0
input-file: corpora/test/test.fa
lmodel-file: 8 0 3 /home/samira/myprogs/moses_progs/cnn_28000/lm2/cnn5_28000.blm.en 8 0 3 /home/samira/myprogs/moses_progs/cnn_28000/lm2/cnn5_28000.blm2.en
mapping: 0 T 0 1 T 1translation-details: working2/o
ttable-file: 0 0 0 5 /home/samira/myprogs/moses_progs/cnn_28000/working2/train/model/phrase-table.gz 0 0 0 5 /home/samira/myprogs/snover_sst/sys4/working/train/model/phrase-table.gzttable-limit: 6s
weight-d: 0.0153922 0.072167 0.0335072 0.0432684 0.0614601 0.0181157 0.0878144weight-l: 0.1022989 0.1222989
weight-t: 0.000153608 0.122284 0.057549 0.0621673 0.0209181 0.00153608 0.122284 0.057549 0.0621673 0.0209181weight-w: -0.332904
/home/samira/mosesdecoder/bin
line=KENLM factor=0 order=3 lazyken=0 path=/home/samira/myprogs/moses_progs/cnn_28000/lm2/cnn5_28000.blm.en
FeatureFunction: KENLM0 start: 0 end: 1
WEIGHT KENLM0=0.122,
line=KENLM factor=0 order=3 lazyken=0 path=/home/samira/myprogs/moses_progs/cnn_28000/lm2/cnn5_28000.blm2.en
FeatureFunction: KENLM1 start: 1 end: 2
WEIGHT KENLM1=
Check scores.size() == indexes.second - indexes.first failed in moses/ScoreComponentCollection.h:235
On Wed, Aug 14, 2013 at 1:27 AM, Philipp Koehn <pkoehn@inf.ed.ac.uk> wrote:
Hi,
this should work, and I do not see anything obviously wrong with
your configuration file.
Is it possible that you run out of memory
Can you post your STDERR output file
testOut/translate_snoverSst1.decode.out ?
-phi
> _______________________________________________
On Tue, Aug 13, 2013 at 9:43 PM, samira tofighi <samiratofighi@gmail.com> wrote:
> i want to use two language models in my system but i get these errors:
>
> ./test: line 10: 4951 Aborted (core dumped) nohup nice
> ~/mosesdecoder/bin/moses -dl 5 -ttl 6s -config
> working2/train/model/moses.ini -drop-unknown -translation-details working2/o
> -input-file corpora/test/test.fa > testOut/translates_snoverSSt1.output 2>
> testOut/translate_snoverSst1.decode.out
>
> any idea to solve it?
> my moses.ini file is like below:
>
>
> # MERT optimized configuration
> # decoder /home/s.tofighi/mosesdecoder/bin/moses
> # BLEU 0.261756 on dev
> /Share/local/s.tofighi/program_moses/bridge_defa/fa_en/corpora/dev/corpus0
> # We were before running iteration 6
> # finished Tue Apr 9 17:28:22 IRDT 2013
> ### MOSES CONFIG FILE ###
> #########################
>
> # input factors
> [input-factors]
> 0
>
> # mapping steps
> [mapping]
> 0 T 0
> 1 T 1
>
> # translation tables: table type (hierarchical(0), textual (0), binary (1)),
> source-factors, target-factors, number of scores, file
> # OLD FORMAT is still handled for back-compatibility
> # OLD FORMAT translation tables: source-factors, target-factors, number of
> scores, file
> # OLD FORMAT a binary table type (1) is assumed
> [ttable-file]
> 0 0 0 5
> /home/samira/myprogs/moses_progs/cnn_28000/working2/train/model/phrase-table.gz
> 0 0 0 5
> /home/samira/myprogs/snover_sst/sys4/working/train/model/phrase-table.gz
>
> # no generation models, no generation-file section
>
> # language models: type(srilm/irstlm), factors, order, file
> [lmodel-file]
> 0 0 3 /home/samira/myprogs/moses_progs/cnn_28000/lm2/cnn5_28000.blm.en
> 0 0 3 /home/samira/myprogs/moses_progs/cnn_28000/lm2/cnn5_28000.blm2.en
>
>
> # limit on how many phrase translations e for each phrase f are loaded
> # 0 = all elements loaded
> [ttable-limit]
> 40
>
> # distortion (reordering) files
> [distortion-file]
> 0-0 wbe-msd-bidirectional-fe-allff 6
> /home/samira/myprogs/moses_progs/cnn_28000/working2/train/model/reordering-table.wbe-msd-bidirectional-fe.gz
>
> # distortion (reordering) weight
> [weight-d]
> 0.0153922
> 0.072167
> 0.0335072
> 0.0432684
> 0.0614601
> 0.0181157
> 0.0878144
>
> # language model weights
> [weight-l]
> 0.1022989
> 0.1222989
>
> # translation model weights
> [weight-t]
> 0.000153608
> 0.122284
> 0.057549
> 0.0621673
> 0.0209181
> 0.00153608
> 0.122284
> 0.057549
> 0.0621673
> 0.0209181
>
> # no generation models, no weight-generation section
>
> # word penalty
> [weight-w]
> -0.332904
>
> [distortion-limit]
> 6
>
> [decoding-graph-backoff]
> 0
> 1
>
> Moses-support mailing list
> Moses-support@mit.edu
> http://mailman.mit.edu/mailman/listinfo/moses-support
>
_______________________________________________
Moses-support mailing list
Moses-support@mit.edu
http://mailman.mit.edu/mailman/listinfo/moses-support
Hieu Hoang
Research Associate
University of Edinburgh
http://www.hoang.co.uk/hieu
--
Hieu Hoang
Research Associate
University of Edinburgh
http://www.hoang.co.uk/hieu
_______________________________________________
Moses-support mailing list
Moses-support@mit.edu
http://mailman.mit.edu/mailman/listinfo/moses-support
0 Response to "using multiple language models"
Post a Comment