Concept

Embedding Layer (A Self-Attentive model for Knowledge Tracing)

Input sequence is transformed so that they would have the same fixed size n. If the size of sample is less than n than padding is applied otherwise inputs are divided and separate samples are generated. Then the interaction embedding matrix is trained MR2E×dM \in \R^{2E \times d} which is used to get sis_is. Exercise embedding matrix is trained in a similar manner ERE×dE \in \R^{E \times d}. Position encoding is used in the model in order to encode the order of sequence. Position embedding is learned in the process of training PRn×dP \in R^{n \times d}. The PiP_is are added to the interaction embedding matrix. Therefore as a result of embedding layer embedded interaction and exercise matrices are generated.

0

1

Updated 2020-11-27

Tags

Data Science