summaryrefslogtreecommitdiff
diff options
context:
space:
mode:
authorLaurent Mazare <laurent.mazare@gmail.com>2024-11-15 16:46:55 +0100
committerGitHub <noreply@github.com>2024-11-15 16:46:55 +0100
commit00d8a0c178f588b6454c02e66b709917628c2bae (patch)
tree218ad1e2c5125d9dd2cb09a7fd3ff3b7ecaa78d0
parentf689ce5d39c6f1475dfc71503288ea2905c8f685 (diff)
downloadcandle-00d8a0c178f588b6454c02e66b709917628c2bae.tar.gz
candle-00d8a0c178f588b6454c02e66b709917628c2bae.tar.bz2
candle-00d8a0c178f588b6454c02e66b709917628c2bae.zip
Remove some unused macros. (#2618)
* Remove some unused macros. * More unused fixes.
-rw-r--r--candle-examples/Cargo.toml2
-rw-r--r--candle-examples/examples/reinforcement-learning/ddpg.rs8
-rw-r--r--candle-examples/examples/reinforcement-learning/gym_env.rs1
-rw-r--r--candle-examples/examples/reinforcement-learning/main.rs2
-rw-r--r--candle-examples/examples/reinforcement-learning/policy_gradient.rs2
-rw-r--r--candle-examples/examples/reinforcement-learning/vec_gym_env.rs5
-rw-r--r--candle-pyo3/Cargo.toml2
-rw-r--r--candle-transformers/src/models/encodec.rs4
-rw-r--r--candle-transformers/src/models/starcoder2.rs1
9 files changed, 13 insertions, 14 deletions
diff --git a/candle-examples/Cargo.toml b/candle-examples/Cargo.toml
index 0c1219d7..df85302d 100644
--- a/candle-examples/Cargo.toml
+++ b/candle-examples/Cargo.toml
@@ -27,7 +27,7 @@ intel-mkl-src = { workspace = true, optional = true }
num-traits = { workspace = true }
palette = { version = "0.7.6", optional = true }
enterpolation = { version = "0.2.1", optional = true}
-pyo3 = { version = "0.22.0", features = ["auto-initialize"], optional = true }
+pyo3 = { version = "0.22.0", features = ["auto-initialize", "abi3-py311"], optional = true }
rayon = { workspace = true }
rubato = { version = "0.15.0", optional = true }
safetensors = { workspace = true }
diff --git a/candle-examples/examples/reinforcement-learning/ddpg.rs b/candle-examples/examples/reinforcement-learning/ddpg.rs
index 5309eaf6..389caac1 100644
--- a/candle-examples/examples/reinforcement-learning/ddpg.rs
+++ b/candle-examples/examples/reinforcement-learning/ddpg.rs
@@ -1,5 +1,4 @@
use std::collections::VecDeque;
-use std::fmt::Display;
use candle::{DType, Device, Error, Module, Result, Tensor, Var};
use candle_nn::{
@@ -167,6 +166,7 @@ fn track(
Ok(())
}
+#[allow(unused)]
struct Actor<'a> {
varmap: VarMap,
vb: VarBuilder<'a>,
@@ -211,7 +211,7 @@ impl Actor<'_> {
let target_network = make_network("target-actor")?;
// this sets the two networks to be equal to each other using tau = 1.0
- track(&mut varmap, &vb, "target-actor", "actor", &dims, 1.0);
+ track(&mut varmap, &vb, "target-actor", "actor", &dims, 1.0)?;
Ok(Self {
varmap,
@@ -244,6 +244,7 @@ impl Actor<'_> {
}
}
+#[allow(unused)]
struct Critic<'a> {
varmap: VarMap,
vb: VarBuilder<'a>,
@@ -287,7 +288,7 @@ impl Critic<'_> {
let target_network = make_network("target-critic")?;
// this sets the two networks to be equal to each other using tau = 1.0
- track(&mut varmap, &vb, "target-critic", "critic", &dims, 1.0);
+ track(&mut varmap, &vb, "target-critic", "critic", &dims, 1.0)?;
Ok(Self {
varmap,
@@ -322,6 +323,7 @@ impl Critic<'_> {
}
}
+#[allow(unused)]
#[allow(clippy::upper_case_acronyms)]
pub struct DDPG<'a> {
actor: Actor<'a>,
diff --git a/candle-examples/examples/reinforcement-learning/gym_env.rs b/candle-examples/examples/reinforcement-learning/gym_env.rs
index a2b6652f..05518b1b 100644
--- a/candle-examples/examples/reinforcement-learning/gym_env.rs
+++ b/candle-examples/examples/reinforcement-learning/gym_env.rs
@@ -1,4 +1,3 @@
-#![allow(unused)]
//! Wrappers around the Python API of Gymnasium (the new version of OpenAI gym)
use candle::{Device, Result, Tensor};
use pyo3::prelude::*;
diff --git a/candle-examples/examples/reinforcement-learning/main.rs b/candle-examples/examples/reinforcement-learning/main.rs
index 1a25cd93..34115b22 100644
--- a/candle-examples/examples/reinforcement-learning/main.rs
+++ b/candle-examples/examples/reinforcement-learning/main.rs
@@ -1,5 +1,3 @@
-#![allow(unused)]
-
#[cfg(feature = "mkl")]
extern crate intel_mkl_src;
diff --git a/candle-examples/examples/reinforcement-learning/policy_gradient.rs b/candle-examples/examples/reinforcement-learning/policy_gradient.rs
index 6c355fe6..3ae2617d 100644
--- a/candle-examples/examples/reinforcement-learning/policy_gradient.rs
+++ b/candle-examples/examples/reinforcement-learning/policy_gradient.rs
@@ -14,7 +14,7 @@ fn new_model(
) -> Result<(impl Module, VarMap)> {
let input_size = input_shape.iter().product();
- let mut varmap = VarMap::new();
+ let varmap = VarMap::new();
let var_builder = VarBuilder::from_varmap(&varmap, dtype, device);
let model = seq()
diff --git a/candle-examples/examples/reinforcement-learning/vec_gym_env.rs b/candle-examples/examples/reinforcement-learning/vec_gym_env.rs
index e382ad76..a985d9e9 100644
--- a/candle-examples/examples/reinforcement-learning/vec_gym_env.rs
+++ b/candle-examples/examples/reinforcement-learning/vec_gym_env.rs
@@ -1,9 +1,8 @@
-#![allow(unused)]
//! Vectorized version of the gym environment.
use candle::{DType, Device, Result, Tensor};
use pyo3::prelude::*;
-use pyo3::types::PyDict;
+#[allow(unused)]
#[derive(Debug)]
pub struct Step {
pub obs: Tensor,
@@ -11,6 +10,7 @@ pub struct Step {
pub is_done: Tensor,
}
+#[allow(unused)]
pub struct VecGymEnv {
env: PyObject,
action_space: usize,
@@ -21,6 +21,7 @@ fn w(res: PyErr) -> candle::Error {
candle::Error::wrap(res)
}
+#[allow(unused)]
impl VecGymEnv {
pub fn new(name: &str, img_dir: Option<&str>, nprocesses: usize) -> Result<VecGymEnv> {
Python::with_gil(|py| {
diff --git a/candle-pyo3/Cargo.toml b/candle-pyo3/Cargo.toml
index 2776a3f7..d91619fb 100644
--- a/candle-pyo3/Cargo.toml
+++ b/candle-pyo3/Cargo.toml
@@ -20,7 +20,7 @@ candle-nn = { workspace = true }
candle-onnx = { workspace = true, optional = true }
half = { workspace = true }
intel-mkl-src = { workspace = true, optional = true }
-pyo3 = { version = "0.22.0", features = ["extension-module", "abi3-py38"] }
+pyo3 = { version = "0.22.0", features = ["extension-module", "abi3-py311"] }
[build-dependencies]
pyo3-build-config = "0.22"
diff --git a/candle-transformers/src/models/encodec.rs b/candle-transformers/src/models/encodec.rs
index a8d509ce..517b9b1d 100644
--- a/candle-transformers/src/models/encodec.rs
+++ b/candle-transformers/src/models/encodec.rs
@@ -4,9 +4,8 @@
//!
//! Based on implementation from [huggingface/transformers](https://github.com/huggingface/transformers/blob/main/src/transformers/models/encodec/modeling_encodec.py)
-#![allow(unused)]
use candle::{DType, IndexOp, Layout, Module, Result, Shape, Tensor, D};
-use candle_nn::{conv1d, Conv1d, Conv1dConfig, ConvTranspose1d, VarBuilder};
+use candle_nn::{conv1d, Conv1d, ConvTranspose1d, VarBuilder};
// Encodec Model
// https://github.com/huggingface/transformers/blob/main/src/transformers/models/encodec/modeling_encodec.py
@@ -226,6 +225,7 @@ impl candle::CustomOp2 for CodebookEncode {
}
// https://github.com/huggingface/transformers/blob/abaca9f9432a84cfaa95531de4c72334f38a42f2/src/transformers/models/encodec/modeling_encodec.py#L340
+#[allow(unused)]
#[derive(Clone, Debug)]
pub struct EuclideanCodebook {
inited: Tensor,
diff --git a/candle-transformers/src/models/starcoder2.rs b/candle-transformers/src/models/starcoder2.rs
index 833cb067..0df5990b 100644
--- a/candle-transformers/src/models/starcoder2.rs
+++ b/candle-transformers/src/models/starcoder2.rs
@@ -15,7 +15,6 @@
//! - [Model Card](https://huggingface.co/bigcode/starcoder)
//!
-#![allow(unused)]
use candle::{DType, Device, Module, Result, Tensor, D};
use candle_nn::{layer_norm, linear_b, LayerNorm, Linear, VarBuilder};
use std::sync::Arc;