From 570f69f750a385d3776331ac026035b7754059e0 Mon Sep 17 00:00:00 2001 From: guanjihuan Date: Fri, 10 May 2024 05:28:33 +0800 Subject: [PATCH] Update activation_function.py --- .../activation_function.py | 14 ++++++++++++++ 1 file changed, 14 insertions(+) diff --git a/2024.03.26_activation_function/activation_function.py b/2024.03.26_activation_function/activation_function.py index fda5f9d..a0d651c 100644 --- a/2024.03.26_activation_function/activation_function.py +++ b/2024.03.26_activation_function/activation_function.py @@ -12,16 +12,30 @@ x_array_torch_tensor = torch.from_numpy(x_array) y_array_torch_tensor = torch.nn.functional.sigmoid(x_array_torch_tensor) plt.plot(x_array_torch_tensor, y_array_torch_tensor) +plt.title('sigmoid') plt.show() y_array_torch_tensor = torch.nn.functional.tanh(x_array_torch_tensor) plt.plot(x_array_torch_tensor, y_array_torch_tensor) +plt.title('tanh') plt.show() y_array_torch_tensor = torch.nn.functional.relu(x_array_torch_tensor) plt.plot(x_array_torch_tensor, y_array_torch_tensor) +plt.title('relu') plt.show() y_array_torch_tensor = torch.nn.functional.leaky_relu(x_array_torch_tensor) plt.plot(x_array_torch_tensor, y_array_torch_tensor) +plt.title('leaky_relu') +plt.show() + +y_array_torch_tensor = torch.nn.functional.gelu(x_array_torch_tensor) +plt.plot(x_array_torch_tensor, y_array_torch_tensor) +plt.title('gelu') +plt.show() + +y_array_torch_tensor = torch.nn.functional.silu(x_array_torch_tensor) +plt.plot(x_array_torch_tensor, y_array_torch_tensor) +plt.title('silu') plt.show() \ No newline at end of file