We used 10,000 test sentences in this experiment. Among these 10,000 sentences, 1,143 sentences matched the Japanese-English patterns. The results of ``English''-English translation revealed that 725 out of the 1,143 sentences were different compared to the standard SMT system (Moses). The other 8,857 sentences (10,000 - 1,143) did not match the Japanese-English pattern.
We used the BLEU [Papineni
2002] and NIST [NIST2003] and METEOR
[Banerjee and Lavie2005] for evaluation tools. Table 11
summarizes the automatic evaluation results of our machine translation
evaluation for the Japanese-English simple sentences. This table shows
the results of 1,143 sentences that were matched with the
Japanese-English patterns. ``Proposed'' indicates our proposed system
(PBMT+SMT), and ``Moses'' indicates a standard SMT system.
We obtained a BLEU score of 0.1821 in the Japanese-English simple sentences using our proposed system. In contrast, we obtained a BLEU score of 0.2218 in the Japanese-English simple sentences using the standard SMT system (Moses). This means that our proposed system was not effective for automatic evaluation in the Japanese-English simple sentences.
BLEU | NIST | METEOR | |
Proposed | 0.1821 | 4.817 | 0.4426 |
Moses | 0.2218 | 5.239 | 0.4363 |
Table 12 shows the all test sentences(10,000 sentences). The 1,143 sentences were translated with the proposed method. The rest of 8,857 sentences were translated with the standard SMT system (Moses).
BLEU | NIST | METEOR | |
Proposed | 0.1101 | 4.4511 | 0.3175 |
Moses | 0.1130 | 4.5131 | 0.3160 |