Dimension Validation Checks for Head/Embedding Divisions

Add dimension validation checks in a vanilla multi-head attention module. The goal is to ensure that the multi-head attention mechanism can be applied correctly.

Constraints

  • The embedding dimension and the number of heads are positive integers.

Examples

Example 1

{
  "input": "same as the basic transformer module",
  "output": "same as the basic transformer module"
}

</>Code

Test

Input:

use python data or natural language description

Output: