Inference with NMT

Learn to do inference with the trained NMT model.

How inferencing is different from the training model

Inferencing is slightly different from the training process for NMT. Because we don’t have a target sentence at the inference time, we need a way to trigger the decoder at the end of the encoding phase. It’s not difficult because we’ve already done the groundwork for this in the data we have. We simply kick off the decoder by using <s> as the first input to the decoder. Then, we recursively call the decoder using the predicted word as the input for the next time step. We continue this way until the model outputs </s> as the predicted token or reaches a predefined sentence length .

How to work with inference models

To do this, we have to define a new model using the existing weights of the training model. This is because our trained model is designed to consume a sequence of decoder inputs at once. We need a mechanism to recursively call the decoder. Here’s how we can define the inference model:

  • Define an encoder model that outputs the encoder’s hidden state sequence and the last encoder state.

  • Define a new decoder that takes a decoder input having a time dimension of 1 and a new input, to which we’ll input the previously hidden state value of the decoder (initialized with the encoder’s last state).

With that, we can start feeding data to generate predictions as follows:

  • Preprocess xsx_s as in data processing.

  • Feed xsx_s into GRUenc\text{GRU}_{enc} ...