UnsupervisedMT 架构的理解

Hi, Mr Kim. Recently, I re-read the two papers
<Unsupervised Machine Translation Using Monolingual Corpora Only>
and
<Phrase based & Neural Unsupervised Machine Translation>
I have some ideas may not right and some questions.

The improvement from the first paper to the second paper comes from:
i) adding language model training before and during the MT training process, since the lm consists of encoder and decoder, so the better encoder and decoder parameters can help the translation results get more smooth.


ii) adding on-the-fly back translation


image.png

MT will be improved iteratively


Question:
i) in lm.py, the language model using shared layers (which is LSTM layer)of the encoder or decoder, so this means there is no implementation for Transformer-LM. Why? any reference says the RNN-LM works betters than Transformer-LM?


ii) in transformer.py


image.png

one_hot=True has not been implemented for transformer, why? I think Transformer should also use one-hot for loss and training


iii) in trainer.py

image.png

there are three ways to train the encoder/decoder LM, I did not see the reason why we need train lm_enc_rev, and also here we do not add_noise for LM training, which is different from
image.png


iiii) add_noise is only called in autoencoder training, not in lm training as written in main.py, though they may work the same but I think the authors should better notice this in this paper.

What's your thought on my ideas and questions, if you have comments, tell me please~

©著作权归作者所有,转载或内容合作请联系作者
平台声明:文章内容(如有图片或视频亦包括在内)由作者上传并发布,文章内容仅代表作者本人观点,简书系信息发布平台,仅提供信息存储服务。

推荐阅读更多精彩内容

  • 一笑 人间陌路 二叹 往事在目 三悲 明妃青阳成冢 紫台朱门一场豪赌 那段仓皇北顾 那封烽火家书 那魂已成白骨 谁...
    安吉丽娜乔阅读 284评论 0 0
  • 成都的旅行完毕,开始写攻略,给各位亲朋好友一些小小的建议,在自由行的时候避免一些不必要的麻烦,节省一些不必要的费用...
    eileen_li阅读 824评论 2 20
  • 丰腴的柳枝揽风入怀 轻轻撞击不可安置的妖娆 过来吧 快过来 我亲爱的孩儿 嘟囔囊的双手 在鼓胀的乳头上游走 母亲啊...
    抱一阅读 365评论 0 1