Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

Summarize operators used in ConvS2S #7312

Closed
lcy-seso opened this issue Jan 8, 2018 · 0 comments
Closed

Summarize operators used in ConvS2S #7312

lcy-seso opened this issue Jan 8, 2018 · 0 comments
Labels

Comments

@lcy-seso
Copy link
Contributor

lcy-seso commented Jan 8, 2018

Here I summarize operators will be used in ConvS2S:

  1. positional embedding

  2. convolution block structure: one-dimensional convolution followed by a GLU.

    Is it necessary to implement GLU in one operator to optimize the time efficiency. This can be determined later.

    • sequence convolution
      • 2D convolution: sequence_conv_op
    • GLU
      • offset operator ?? (To be determined later)
      • sigmoid
      • element-wise multiplication
      • addition
    • attention
    • weight normalization

The missing operator:

  1. weight normalization, related to Implement weight normalization in Fluid. #6914

Need to be enhanced: (This enhancement is also needed by both Transformer and ConvS2S)

  1. look_up_table : related to Support padding_idx in the lookup_table_op. #7309
@lcy-seso lcy-seso added the NMT label Jan 8, 2018
@lcy-seso lcy-seso changed the title Operators used in ConvS2S Summarize operators used in ConvS2S Jan 8, 2018
Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
Projects
None yet
Development

No branches or pull requests

1 participant