From 53510ce427160674268199349b22bbb62cd2b9ba Mon Sep 17 00:00:00 2001 From: Laurent Mazare Date: Fri, 29 Sep 2023 08:06:54 +0200 Subject: Use a silu activation in mistral. (#991) --- candle-nn/src/activation.rs | 4 ++++ 1 file changed, 4 insertions(+) (limited to 'candle-nn/src') diff --git a/candle-nn/src/activation.rs b/candle-nn/src/activation.rs index 1e67ed53..ddc211a7 100644 --- a/candle-nn/src/activation.rs +++ b/candle-nn/src/activation.rs @@ -9,6 +9,8 @@ pub enum Activation { #[serde(rename = "gated-gelu")] NewGelu, Relu, + Silu, + Sigmoid, Elu(f64), LeakyRelu(f64), } @@ -20,6 +22,8 @@ impl super::Module for Activation { // https://github.com/huggingface/transformers/blob/12f043eaeaabfef6f6efea411d98e6f6d3c094b7/src/transformers/activations.py#L49-L78 Self::NewGelu => xs.gelu(), Self::Relu => xs.relu(), + Self::Silu => crate::ops::silu(xs), + Self::Sigmoid => crate::ops::sigmoid(xs), &Self::Elu(alpha) => xs.elu(alpha), &Self::LeakyRelu(negative_slope) => crate::ops::leaky_relu(xs, negative_slope), } -- cgit v1.2.3