Relation

Why is it better to use ReLU by default?

ReLU is very simple compared with many other activation functions. If you are not sure what activation function to use, it's better to start with ReLU and then adjust the model to be better. This will be computationally less expensive because too complex model may take too long (e.g. a few months) to train and it's not wise to spend such a long time when you start.

0

3

Updated 2026-05-06

Tags

Data Science

D2L

Dive into Deep Learning @ D2L