Concept

Training (Knowledge Tracing Machines: Factorization Machines for Knowledge Tracing)

In training following function was minimized:

NLL(p(X),y)=i=1Syilogp(xi)+(1yi)log(1p(xi))NLL(p(X), y) = \sum_{i=1}^S y_i \log p(x_i) + (1-y_i) \log (1-p(x_i)) here ys are outcomes and xs are inputs.

Biases, embeddings follow hyperpriors - wkN(μ,1/ λ)w_k \sim N ( \mu, 1 /\ \lambda) vkf simN(μ,1/ λ)v_{kf} \ sim N ( \mu, 1 /\ \lambda) μ\mu, λ\lambda follow hyperpriors - μN(0,1) \mu \sim N(0, 1) λΓ(0,1) \lambda \sim \Gamma (0, 1)

Due to these hyperpriors the authors don't need finetuning.

0

1

Updated 2020-11-26

Tags

Data Science