Send Moses-support mailing list submissions to
moses-support@mit.edu
To subscribe or unsubscribe via the World Wide Web, visit
http://mailman.mit.edu/mailman/listinfo/moses-support
or, via email, send a message with subject or body 'help' to
moses-support-request@mit.edu
You can reach the person managing the list at
moses-support-owner@mit.edu
When replying, please edit your Subject line so it is more specific
than "Re: Contents of Moses-support digest..."
Today's Topics:
1. Call for Papers: INTERACTIVE AND ADAPTIVE MACHINE TRANSLATION
(Philipp Koehn)
2. Re: Regarding language model (Philipp Koehn)
3. Seeking for a co-author of a paper for INTERACTIVE AND
ADAPTIVE MACHINE TRANSLATION (Wei JIANG [PT-COM])
4. Re: Filtering Issue (Hieu Hoang)
----------------------------------------------------------------------
Message: 1
Date: Tue, 8 Jul 2014 12:51:25 -0400
From: Philipp Koehn <pkoehn@inf.ed.ac.uk>
Subject: [Moses-support] Call for Papers: INTERACTIVE AND ADAPTIVE
MACHINE TRANSLATION
To: Multiple recipients of list <mt_list@nist.gov>,
"moses-support@mit.edu" <moses-support@mit.edu>, "corpora@uib.no"
<CORPORA@uib.no>, mt-list@eamt.org
Message-ID:
<CAAFADDCGVuTYgqpVwdwx3F2wpOTwA6qzkcgKJEO_73crj5QZHA@mail.gmail.com>
Content-Type: text/plain; charset=UTF-8
First Call for Papers:
AMTA 2014 WORKSHOP ON INTERACTIVE AND ADAPTIVE MACHINE TRANSLATION
22 October 2014, Vancouver, Canada
http://www.statmt.org/iamt/
The increasing use of machine translation in the workflow of
professional translators creates demand for machine translation
technology that provides more interactive collaboration, learns from
its errors and adapts to the translators' style and adapts the
underlying machine translation system online to the specific needs of
the translator for the given task.
The next generation of computer aided translation (CAT) tools has to
move beyond the use of static machine translation for human
post-editing into a much richer division of labor between man and
machine that takes full advantage of man's understanding of content
and machine's greater ability to quickly process large amounts of
data.
On the other hand, these tools will allow a more friendly interaction
between the human and the machine through the use of different
modalities of interactions as speech, gaze tracking, e-pen, etc.
Finally, all of these issues will lead to an increase of the
productivity of the professional translators.
Such tools are in development in a number of research labs across the
world, one example is the open source workbench developed by the
EU-funded projects Matecat and Casmacat, led by the organizers of this
workshop.
This workshop brings to together researchers in this nascent subfield
of machine translation. The workshop will divide its schedule about
equally between invited talks by leading researchers and paper
presentations on more recent advances.
We encourage submissions including, but not limited to:
- rapid user and project adaptation
- online learning methods
- active learning
- domain adaptation
- interactive translation prediction
- use of confidence measures to inform translators
- quality estimation to filter/rank translation suggestions
- automatic terminology support
- integration of translation memory and machine translation
- computer assisted writing in the context of translation
- novel types of assistance
- multimodal interaction: speech, writing, gestures, gaze.
- user studies on advanced computer aided translation
INVITED TALKS
- Spence Green, Stanford
- Michael Denkowski, CMU
- Lane Schwartz, Air Force Research Laboratory
- Michel Simard, National Research Council Canada
Additional speakers will be announced shortly.
IMPORTANT DATES
Paper submission deadline: August 29, 2014
Notification of acceptance: September 12, 2014
Camera-ready deadline: September 26, 2014
ORGANIZERS
Francisco Casacuberta, Universitat Polit?cnica de Val?ncia
Marcello Federico, Fondazione Bruno Kessler
Philipp Koehn, University of Edinburgh / Johns Hopkins University
------------------------------
Message: 2
Date: Tue, 8 Jul 2014 13:48:35 -0400
From: Philipp Koehn <pkoehn@inf.ed.ac.uk>
Subject: Re: [Moses-support] Regarding language model
To: Pranjal Das <pranjal4456@gmail.com>
Cc: moses-support <moses-support@mit.edu>
Message-ID:
<CAAFADDBB-0qdm5qdtb=FkAFLvYU62uv+GP3O3WDw43FCXO9JQQ@mail.gmail.com>
Content-Type: text/plain; charset=UTF-8
Hi,
an excellent summary of various smoothing methods for n-gram models is:
http://www.cs.berkeley.edu/~klein/cs294-5/chen_goodman.pdf
-phi
On Thu, Jul 3, 2014 at 3:20 AM, Pranjal Das <pranjal4456@gmail.com> wrote:
> While calculating bigram probability, suppose for a sentence "He likes
> to play football", does P(He|<start>) and P(<end>|football) calculated
> ?
> --
>
> Regards,
>
> *Pranjal Das*
> Department of Information Technology,
> Gauhati University Institute of Science and Technology,
> Phone- +91-8399879454
> _______________________________________________
> Moses-support mailing list
> Moses-support@mit.edu
> http://mailman.mit.edu/mailman/listinfo/moses-support
------------------------------
Message: 3
Date: Wed, 09 Jul 2014 15:23:36 +0800
From: "Wei JIANG [PT-COM]" <jiangw@polytrans.com>
Subject: [Moses-support] Seeking for a co-author of a paper for
INTERACTIVE AND ADAPTIVE MACHINE TRANSLATION
To: moses-support@mit.edu
Message-ID: <53BCEDF8.70506@polytrans.com>
Content-Type: text/plain; charset=UTF-8; format=flowed
Hello all,
I am a professional Chinese translator and multilingual project
manager. I am much interested in submitting a paper on the subject of
QA/QC of human translations based on machine translations.
The paper would be a co-authored one. Basically, I would be focusing on
human translation (HT), and the other co-author would need to focus on
machine translation (MT). The concerned languages would be C2JK, ie,
Traditional Chinese & Simplified Chinese, Japanese, and Korean. And we
could be using Microsoft bilingual parallel corpus for MT, and real
and/or semi-real projects for HT.
I live in Beijing, China, and I will be traveling to Toronto Canada
late August and stay there until the end of this year, so I will be very
happy to attend the Workshop and read the co-authored paper and/or
demonstrate the sample QA/QC tool.
Anyone interested, please email me off-list.
Thank you!
Wei
On 7/9/2014 12:51 AM, Philipp Koehn wrote:
> First Call for Papers:
>
> AMTA 2014 WORKSHOP ON INTERACTIVE AND ADAPTIVE MACHINE TRANSLATION
> 22 October 2014, Vancouver, Canada
>
> http://www.statmt.org/iamt/
>
> The increasing use of machine translation in the workflow of
> professional translators creates demand for machine translation
> technology that provides more interactive collaboration, learns from
> its errors and adapts to the translators' style and adapts the
> underlying machine translation system online to the specific needs of
> the translator for the given task.
>
> The next generation of computer aided translation (CAT) tools has to
> move beyond the use of static machine translation for human
> post-editing into a much richer division of labor between man and
> machine that takes full advantage of man's understanding of content
> and machine's greater ability to quickly process large amounts of
> data.
>
> On the other hand, these tools will allow a more friendly interaction
> between the human and the machine through the use of different
> modalities of interactions as speech, gaze tracking, e-pen, etc.
> Finally, all of these issues will lead to an increase of the
> productivity of the professional translators.
>
> Such tools are in development in a number of research labs across the
> world, one example is the open source workbench developed by the
> EU-funded projects Matecat and Casmacat, led by the organizers of this
> workshop.
>
> This workshop brings to together researchers in this nascent subfield
> of machine translation. The workshop will divide its schedule about
> equally between invited talks by leading researchers and paper
> presentations on more recent advances.
>
> We encourage submissions including, but not limited to:
>
> - rapid user and project adaptation
> - online learning methods
> - active learning
> - domain adaptation
> - interactive translation prediction
> - use of confidence measures to inform translators
> - quality estimation to filter/rank translation suggestions
> - automatic terminology support
> - integration of translation memory and machine translation
> - computer assisted writing in the context of translation
> - novel types of assistance
> - multimodal interaction: speech, writing, gestures, gaze.
> - user studies on advanced computer aided translation
>
> INVITED TALKS
>
> - Spence Green, Stanford
> - Michael Denkowski, CMU
> - Lane Schwartz, Air Force Research Laboratory
> - Michel Simard, National Research Council Canada
>
> Additional speakers will be announced shortly.
>
> IMPORTANT DATES
>
> Paper submission deadline: August 29, 2014
> Notification of acceptance: September 12, 2014
> Camera-ready deadline: September 26, 2014
>
> ORGANIZERS
>
> Francisco Casacuberta, Universitat Polit?cnica de Val?ncia
> Marcello Federico, Fondazione Bruno Kessler
> Philipp Koehn, University of Edinburgh / Johns Hopkins University
>
> _______________________________________________
> Moses-support mailing list
> Moses-support@mit.edu
> http://mailman.mit.edu/mailman/listinfo/moses-support
------------------------------
Message: 4
Date: Wed, 9 Jul 2014 17:02:09 +0100
From: Hieu Hoang <Hieu.Hoang@ed.ac.uk>
Subject: Re: [Moses-support] Filtering Issue
To: Judah Schvimer <judah.schvimer@mongodb.com>
Cc: moses-support <moses-support@mit.edu>, "user-irstlm@list.fbk.eu"
<user-irstlm@list.fbk.eu>
Message-ID:
<CAEKMkbg2pR2VE9Ntv_6F4-yujPWQxeSOerJBRx+EtB_CnPdDwg@mail.gmail.com>
Content-Type: text/plain; charset="utf-8"
Sorry for late reply.
Did you clean the data before you created the language model? Or you can
try binarizing with KenLM instead
On 3 July 2014 14:50, Judah Schvimer <judah.schvimer@mongodb.com> wrote:
> Hi,
>
> I have a script that basically does all of the training steps for me and
> it's worked many times before, but suddenly I got this error with a
> different corpus. It doesn't occur when I use the moses.ini file in
> train/model/ or in mert-work/, just when I use the one in filtered/. Any
> idea what's wrong? Something appears to be weird around the 39-grams step.
>
> Defined parameters (per moses.ini or switch):
> config: /home/judah/archive8/3/working/filtered/moses.ini
> distortion-limit: 6
> feature: UnknownWordPenalty WordPenalty PhrasePenalty
> PhraseDictionaryBinary name=TranslationModel0 table-limit=20 num-features=4
> path=/home/judah/archive8/3/working/filtered/phrase-table.0-0.1.1
> input-factor=0 output-factor=0 LexicalReordering name=LexicalReordering0
> num-features=6 type=hier-msd-bidirectional-fe-allff input-factor=0
> output-factor=0
> path=/home/judah/archive8/3/working/filtered/reordering-table.hier-msd-bidirectional-fe
> Distortion IRSTLM name=LM0 factor=0 path=/home/judah/archive8/3/lm/
> train.en-es.blm.es order=3
> input-factors: 0
> mapping: 0 T 0
> weight: LexicalReordering0= 0.389589 0.0418995 0.0286706 0.0187875
> 0.0287628 0.00958344 Distortion0= 0.0583275 LM0= -0.00740405 WordPenalty0=
> 0.0123212 PhrasePenalty0= 0.0413057 TranslationModel0= 0.0360237 0.040722
> 0.266723 0.0198795 UnknownWordPenalty0= 1
> /home/judah/mosesdecoder/bin
> line=UnknownWordPenalty
> FeatureFunction: UnknownWordPenalty0 start: 0 end: 0
> line=WordPenalty
> FeatureFunction: WordPenalty0 start: 1 end: 1
> line=PhrasePenalty
> FeatureFunction: PhrasePenalty0 start: 2 end: 2
> line=PhraseDictionaryBinary name=TranslationModel0 table-limit=20
> num-features=4
> path=/home/judah/archive8/3/working/filtered/phrase-table.0-0.1.1
> input-factor=0 output-factor=0
> FeatureFunction: TranslationModel0 start: 3 end: 6
> line=LexicalReordering name=LexicalReordering0 num-features=6
> type=hier-msd-bidirectional-fe-allff input-factor=0 output-factor=0
> path=/home/judah/archive8/3/working/filtered/reordering-table.hier-msd-bidirectional-fe
> FeatureFunction: LexicalReordering0 start: 7 end: 12
> Initializing LexicalReordering..
> line=Distortion
> FeatureFunction: Distortion0 start: 13 end: 13
> line=IRSTLM name=LM0 factor=0 path=/home/judah/archive8/3/lm/
> train.en-es.blm.es order=3
> FeatureFunction: LM0 start: 14 end: 14
> Loading UnknownWordPenalty0
> Loading WordPenalty0
> Loading PhrasePenalty0
> Loading LexicalReordering0
> binary file loaded, default OFF_T: -1
> Loading Distortion0
> Loading LM0
> In LanguageModelIRST::Load: nGramOrder = 3
> Language Model Type of /home/judah/archive8/3/lm/train.en-es.blm.es is 1
> Language Model Type is 1
> mmap
> loadtxt_ram()
> 3-grams: reading 0 entries
> done level 3
> 1-grams: reading 0 entries
> done level 1
> 8-grams: reading 0 entries
> done level 8
> 2-grams: reading 0 entries
> done level 2
> 39-grams: reading 3991252117 entries
> moses: util.cpp:289: int parseline(std::istream&, int, ngram&, float&,
> float&): Assertion `howmany == (Order+ 1) || howmany == (Order + 2)' failed.
> [1] 5062 abort (core dumped) /home/judah/mosesdecoder/bin/moses -f <
> >
>
>
> Thanks,
> Judah
>
> _______________________________________________
> Moses-support mailing list
> Moses-support@mit.edu
> http://mailman.mit.edu/mailman/listinfo/moses-support
>
>
--
Hieu Hoang
Research Associate
University of Edinburgh
http://www.hoang.co.uk/hieu
-------------- next part --------------
An HTML attachment was scrubbed...
URL: http://mailman.mit.edu/mailman/private/moses-support/attachments/20140709/a1b3770e/attachment.htm
------------------------------
_______________________________________________
Moses-support mailing list
Moses-support@mit.edu
http://mailman.mit.edu/mailman/listinfo/moses-support
End of Moses-support Digest, Vol 93, Issue 10
*********************************************
Subscribe to:
Post Comments (Atom)
0 Response to "Moses-support Digest, Vol 93, Issue 10"
Post a Comment