TheCluster commited on
Commit
9d84ec7
·
verified ·
1 Parent(s): 0bb3870

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +17 -6
README.md CHANGED
@@ -7,18 +7,24 @@ pipeline_tag: image-text-to-text
7
  tags:
8
  - chat
9
  - mlx
 
10
  - uncensored
11
  - apple
12
  - 8bit
 
 
 
 
 
 
 
 
13
  library_name: mlx
14
  ---
15
  # Gemma-3-27B Instruct Uncensored 8-bit MLX
16
  Uncensored version of **Gemma 3 27B**.
17
 
18
- This model was converted to MLX format from [`nidum/Nidum-gemma-3-27B-it-Uncensored`]() using mlx-vlm version **0.1.19**.
19
-
20
- Refer to the [original model card](https://huggingface.co/google/gemma-3-27b-it) and [uncensored model](https://huggingface.co/nidum/Nidum-gemma-3-27B-it-Uncensored) for more details on the model.
21
-
22
 
23
  ## Technical Details
24
 
@@ -26,6 +32,8 @@ Supports a context length of 128k tokens, with a max output of 8192.
26
 
27
  Multimodal supporting images normalized to 896 x 896 resolution.
28
 
 
 
29
 
30
  ## Use with mlx
31
 
@@ -34,5 +42,8 @@ pip install -U mlx-vlm
34
  ```
35
 
36
  ```bash
37
- python -m mlx_vlm.generate --model TheCluster/gemma-3-27b-it-uncensored-mlx-8bit --max-tokens 128 --temperature 0.0 --prompt "Describe this image." --image <path_to_image>
38
- ```
 
 
 
 
7
  tags:
8
  - chat
9
  - mlx
10
+ - gemma3
11
  - uncensored
12
  - apple
13
  - 8bit
14
+ language:
15
+ - en
16
+ - fr
17
+ - es
18
+ - de
19
+ - it
20
+ - hi
21
+ - ru
22
  library_name: mlx
23
  ---
24
  # Gemma-3-27B Instruct Uncensored 8-bit MLX
25
  Uncensored version of **Gemma 3 27B**.
26
 
27
+ Also you can try new uncensored version: [Amoral Gemma-3 27B 8-bit MLX](https://huggingface.co/TheCluster/amoral-gemma-3-27B-v2-mlx-8bit)
 
 
 
28
 
29
  ## Technical Details
30
 
 
32
 
33
  Multimodal supporting images normalized to 896 x 896 resolution.
34
 
35
+ Refer to the [original model card](https://huggingface.co/google/gemma-3-27b-it) and [uncensored model](https://huggingface.co/nidum/Nidum-gemma-3-27B-it-Uncensored) for more details on the model.
36
+
37
 
38
  ## Use with mlx
39
 
 
42
  ```
43
 
44
  ```bash
45
+ python -m mlx_vlm.generate --model TheCluster/gemma-3-27b-it-uncensored-mlx-8bit --max-tokens 256 --temperature 0.4 --prompt "Describe this image." --image <path_to_image>
46
+ ```
47
+
48
+ ### Source
49
+ This model was converted to MLX format from [`nidum/Nidum-gemma-3-27B-it-Uncensored`]() using mlx-vlm version **0.1.19**.