Anonymous
ML Knowledge
Between ReLU and sigmoid functions, which one mitigates the vanishing gradient issue more efficiently?
Answers
Unlock Community Insights
Contribute your knowledge to access all answers
#Give&Take - Share to unlock
Try Free AI Interview
Interview question asked to Data Scientists and Machine Learning Engineers interviewing at Course Hero, Asana, Rivian and others: Between ReLU and sigmoid functions, which one mitigates the vanishing gradient issue more efficiently?.