Spaces:
Sleeping
Sleeping
Commit
·
32074ae
1
Parent(s):
1ea681e
Fix model initialization by removing unsupported parameters
Browse files
app.py
CHANGED
@@ -85,34 +85,34 @@ def initialize_model():
|
|
85 |
device = "cuda" if torch.cuda.is_available() else "cpu"
|
86 |
logger.info(f"Using device: {device}")
|
87 |
|
88 |
-
# Initialize new model with
|
89 |
model = LLaVA(
|
90 |
vision_model_path="openai/clip-vit-base-patch32",
|
91 |
language_model_path="TinyLlama/TinyLlama-1.1B-Chat-v1.0",
|
92 |
-
device=device,
|
93 |
projection_hidden_dim=2048,
|
94 |
-
|
95 |
-
low_cpu_mem_usage=True,
|
96 |
-
use_safetensors=True,
|
97 |
-
load_in_8bit=True if device == "cuda" else False,
|
98 |
-
trust_remote_code=True
|
99 |
)
|
100 |
|
101 |
# Configure model for inference
|
102 |
if hasattr(model, 'language_model'):
|
103 |
-
model
|
104 |
model.language_model.eval()
|
105 |
|
106 |
-
#
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
107 |
if hasattr(model.language_model, 'generation_config'):
|
108 |
model.language_model.generation_config.do_sample = True
|
109 |
model.language_model.generation_config.max_new_tokens = 256
|
110 |
model.language_model.generation_config.temperature = 0.7
|
111 |
model.language_model.generation_config.top_p = 0.9
|
112 |
-
|
113 |
-
|
114 |
-
# Move model to device
|
115 |
-
model = model.to(device)
|
116 |
|
117 |
model_status.update({
|
118 |
"initialized": True,
|
@@ -121,11 +121,10 @@ def initialize_model():
|
|
121 |
"model_info": {
|
122 |
"vision_model": "openai/clip-vit-base-patch32",
|
123 |
"language_model": "TinyLlama/TinyLlama-1.1B-Chat-v1.0",
|
124 |
-
"dtype": str(model.dtype),
|
125 |
"device": str(model.device)
|
126 |
}
|
127 |
})
|
128 |
-
logger.info(f"Model successfully initialized on {model.device}
|
129 |
return True
|
130 |
|
131 |
except Exception as e:
|
|
|
85 |
device = "cuda" if torch.cuda.is_available() else "cpu"
|
86 |
logger.info(f"Using device: {device}")
|
87 |
|
88 |
+
# Initialize new model with basic parameters
|
89 |
model = LLaVA(
|
90 |
vision_model_path="openai/clip-vit-base-patch32",
|
91 |
language_model_path="TinyLlama/TinyLlama-1.1B-Chat-v1.0",
|
|
|
92 |
projection_hidden_dim=2048,
|
93 |
+
device=device
|
|
|
|
|
|
|
|
|
94 |
)
|
95 |
|
96 |
# Configure model for inference
|
97 |
if hasattr(model, 'language_model'):
|
98 |
+
# Set model to evaluation mode
|
99 |
model.language_model.eval()
|
100 |
|
101 |
+
# Configure model for inference
|
102 |
+
if hasattr(model.language_model, 'config'):
|
103 |
+
model.language_model.config.use_cache = False
|
104 |
+
|
105 |
+
# Move model to device
|
106 |
+
model = model.to(device)
|
107 |
+
|
108 |
+
# Set generation config if available
|
109 |
if hasattr(model.language_model, 'generation_config'):
|
110 |
model.language_model.generation_config.do_sample = True
|
111 |
model.language_model.generation_config.max_new_tokens = 256
|
112 |
model.language_model.generation_config.temperature = 0.7
|
113 |
model.language_model.generation_config.top_p = 0.9
|
114 |
+
if hasattr(model.language_model.config, 'eos_token_id'):
|
115 |
+
model.language_model.generation_config.pad_token_id = model.language_model.config.eos_token_id
|
|
|
|
|
116 |
|
117 |
model_status.update({
|
118 |
"initialized": True,
|
|
|
121 |
"model_info": {
|
122 |
"vision_model": "openai/clip-vit-base-patch32",
|
123 |
"language_model": "TinyLlama/TinyLlama-1.1B-Chat-v1.0",
|
|
|
124 |
"device": str(model.device)
|
125 |
}
|
126 |
})
|
127 |
+
logger.info(f"Model successfully initialized on {model.device}")
|
128 |
return True
|
129 |
|
130 |
except Exception as e:
|