summaryrefslogtreecommitdiff
path: root/candle-transformers/src/models/quantized_mistral.rs
diff options
context:
space:
mode:
authorLaurent Mazare <laurent.mazare@gmail.com>2023-10-18 21:55:46 +0100
committerGitHub <noreply@github.com>2023-10-18 21:55:46 +0100
commit902d0b91664df1e9074cc365de5eba6b578d6692 (patch)
tree9b67d33a6f56cf6391f4756de0cad1f97182f3a2 /candle-transformers/src/models/quantized_mistral.rs
parent185b54a33bae51410a667dbb212ba6f29bb6104f (diff)
downloadcandle-902d0b91664df1e9074cc365de5eba6b578d6692.tar.gz
candle-902d0b91664df1e9074cc365de5eba6b578d6692.tar.bz2
candle-902d0b91664df1e9074cc365de5eba6b578d6692.zip
More model cloning. (#1126)
* More model cloning. * More cloning on quantized models.
Diffstat (limited to 'candle-transformers/src/models/quantized_mistral.rs')
-rw-r--r--candle-transformers/src/models/quantized_mistral.rs10
1 files changed, 5 insertions, 5 deletions
diff --git a/candle-transformers/src/models/quantized_mistral.rs b/candle-transformers/src/models/quantized_mistral.rs
index 00c80209..9e306c67 100644
--- a/candle-transformers/src/models/quantized_mistral.rs
+++ b/candle-transformers/src/models/quantized_mistral.rs
@@ -6,7 +6,7 @@ use std::sync::Arc;
pub use crate::models::mistral::Config;
-#[derive(Debug)]
+#[derive(Debug, Clone)]
struct RotaryEmbedding {
sin: Tensor,
cos: Tensor,
@@ -57,7 +57,7 @@ impl RotaryEmbedding {
}
}
-#[derive(Debug)]
+#[derive(Debug, Clone)]
#[allow(clippy::upper_case_acronyms)]
struct MLP {
gate_proj: Linear,
@@ -90,7 +90,7 @@ impl Module for MLP {
}
}
-#[derive(Debug)]
+#[derive(Debug, Clone)]
struct Attention {
q_proj: Linear,
k_proj: Linear,
@@ -200,7 +200,7 @@ impl Attention {
}
}
-#[derive(Debug)]
+#[derive(Debug, Clone)]
struct DecoderLayer {
self_attn: Attention,
mlp: MLP,
@@ -243,7 +243,7 @@ impl DecoderLayer {
}
}
-#[derive(Debug)]
+#[derive(Debug, Clone)]
pub struct Model {
embed_tokens: Embedding,
layers: Vec<DecoderLayer>,