All public logs
Combined display of all available logs of AI Wiki. You can narrow down the view by selecting a log type, the username (case-sensitive), or the affected page (also case-sensitive).
- 13:13, 18 March 2023 Walle talk contribs created page Rectified Linear Unit (ReLU) (Created page with "{{see also|Machine learning terms}} ==Rectified Linear Unit (ReLU)== The Rectified Linear Unit (ReLU) is a widely-used activation function in the field of machine learning and deep learning. It is a non-linear function that helps to model complex patterns and relationships in data. ReLU has gained significant popularity because of its simplicity and efficiency in training deep neural networks. ===History of ReLU=== The concept of ReLU can be traced back to t...")