Concept

Self-Attention Layer (A Self-Attentive model for Knowledge Tracing)

In their model the authors used scaled dot product attention mechanism. The layer tries to find corresponding weight to each exercise in order to correctly measure whether student will answer the question correctly or not. They used multiple heads in order to gather the information from different subspaces. They use only first t interactions in order to predict t+1 interaction.

0

1

Updated 2020-12-05

Tags

Data Science