summaryrefslogtreecommitdiff
path: root/candle-transformers/src/models/chinese_clip/mod.rs
diff options
context:
space:
mode:
Diffstat (limited to 'candle-transformers/src/models/chinese_clip/mod.rs')
-rw-r--r--candle-transformers/src/models/chinese_clip/mod.rs5
1 files changed, 2 insertions, 3 deletions
diff --git a/candle-transformers/src/models/chinese_clip/mod.rs b/candle-transformers/src/models/chinese_clip/mod.rs
index 86616baa..1edc9031 100644
--- a/candle-transformers/src/models/chinese_clip/mod.rs
+++ b/candle-transformers/src/models/chinese_clip/mod.rs
@@ -3,10 +3,9 @@
//! Chinese contrastive Language-Image Pre-Training (CLIP) is an architecture trained on
//! pairs of images with related texts.
//!
-//! - [GH Link](https://github.com/OFA-Sys/Chinese-CLIP)
-//! - Transformers Python [reference implementation](https://github.com/huggingface/transformers/blob/5af7d41e49bbfc8319f462eb45253dcb3863dfb7/src/transformers/models/chinese_clip/modeling_chinese_clip.py)
+//! - 💻 [GH Link](https://github.com/OFA-Sys/Chinese-CLIP)
+//! - 💻 Transformers Python [reference implementation](https://github.com/huggingface/transformers/blob/5af7d41e49bbfc8319f462eb45253dcb3863dfb7/src/transformers/models/chinese_clip/modeling_chinese_clip.py)
//!
-
use candle::{Module, Result, Tensor, D};
use candle_nn as nn;