You signed in with another tab or window. Reload to refresh your session.You signed out in another tab or window. Reload to refresh your session.You switched accounts on another tab or window. Reload to refresh your session.Dismiss alert
{{ message }}
This repository has been archived by the owner on Dec 29, 2022. It is now read-only.
If I am not using the length_penalty_weight (second parameter above) while just specifying the beam_width (first parameter), the model is working fine with no truncations.
When I am using length_penalty_weight (with beam width 10), my predictions text file (predictions.txt as specified in Decoding with beam search ) is truncated while the dump numpy array (.npz file) contains the full length results.
Taking cue from model parameters specified here, I tried using the following parameters at inference:
but getting the error "ValueError: inference.max_decode_length is not a valid model parameter"
Do, I have to specify any other model parameter (during training, I have used both max_decode_length:300 as well as target.max_seq_len:300)? It was strange, as the truncations are only happening when using the 'length_penalty_weight' (along with 'beam_width')
Should the results in 'predictions.txt' be the first (best/most probable) prediction if we produce the top-k predictions using the script in How to infer top-k output sequences? #158?
Thank you.
The text was updated successfully, but these errors were encountered:
Sign up for freeto subscribe to this conversation on GitHub.
Already have an account?
Sign in.
Hi
I am using Beam search for inference as described here.
The relevant 'model parameters' are:
If I am not using the length_penalty_weight (second parameter above) while just specifying the beam_width (first parameter), the model is working fine with no truncations.
When I am using length_penalty_weight (with beam width 10), my predictions text file (predictions.txt as specified in Decoding with beam search ) is truncated while the dump numpy array (.npz file) contains the full length results.
Taking cue from model parameters specified here, I tried using the following parameters at inference:
but getting the error "ValueError: inference.max_decode_length is not a valid model parameter"
Do, I have to specify any other model parameter (during training, I have used both max_decode_length:300 as well as target.max_seq_len:300)? It was strange, as the truncations are only happening when using the 'length_penalty_weight' (along with 'beam_width')
Should the results in 'predictions.txt' be the first (best/most probable) prediction if we produce the top-k predictions using the script in How to infer top-k output sequences? #158?
Thank you.
The text was updated successfully, but these errors were encountered: