About 217,000 results
Open links in new tab
  1. ReLU Activation Function in Deep Learning - GeeksforGeeks

    Jul 23, 2025 · The ReLU function is a piecewise linear function that outputs the input directly if it is positive; otherwise, it outputs zero. In simpler terms, ReLU allows positive values to pass through …

  2. ReLU - Desmos

    Graph functions, plot points, visualize algebraic equations, add sliders, animate graphs, and more.

  3. Rectified linear unit - Wikipedia

    In the context of artificial neural networks, the rectifier or ReLU (rectified linear unit) activation function[1][2] is an activation function defined as the non-negative part of its argument, i.e., the ramp …

  4. A Beginner’s Guide to the Rectified Linear Unit (ReLU)

    Jan 28, 2025 · f (x) = max (0, x) The ReLU function outputs the maximum between its input and zero, as shown by the graph. For positive inputs, the output of the function is equal to the input. For strictly …

  5. Rectified Linear Unit (ReLU) Function in Deep Learning

    Learn how the rectified linear unit (ReLU) function works, how to implement it in Python, and its variations, advantages, and disadvantages.

  6. ReLU Function, Online Calculator, Graph and Examples

    Online calculator and explanation of the ReLU (Rectified Linear Unit) activation function - the modern alternative to Sigmoid

  7. ReLU Activation Function Explained | Built In

    Feb 26, 2024 · What Is the ReLU Activation Function? The diagram below with the blue line is the representation of the Rectified Linear Unit (ReLU), whereas the green line is a variant of ReLU called …

  8. ReLU Activation Function and Its Variants - pythonkitchen.com

    Dec 13, 2022 · In the above image, we can see that all of the negative values passing through the ReLU activation function will be converted to zero and there will be a straight line in the graph of the …

  9. Tanh vs. Sigmoid vs. ReLU - GeeksforGeeks

    Nov 27, 2025 · In this graph, the x-axis represents the input values ranging from ∞ −∞ to + ∞ +∞, while the y-axis shows the ReLU output, which remains at 0 for negative inputs and increases linearly for …

  10. Understanding ReLU in PyTorch: A Comprehensive Guide

    ReLU creates a “hinge” at x=0, allowing the function to: To better understand the behavior of the ReLU activation function, we can plot it over a range of input values using PyTorch and Matplotlib. This …