Prashant26am commited on
Commit
32074ae
·
1 Parent(s): 1ea681e

Fix model initialization by removing unsupported parameters

Browse files
Files changed (1) hide show
  1. app.py +14 -15
app.py CHANGED
@@ -85,34 +85,34 @@ def initialize_model():
85
  device = "cuda" if torch.cuda.is_available() else "cpu"
86
  logger.info(f"Using device: {device}")
87
 
88
- # Initialize new model with Hugging Face specific parameters
89
  model = LLaVA(
90
  vision_model_path="openai/clip-vit-base-patch32",
91
  language_model_path="TinyLlama/TinyLlama-1.1B-Chat-v1.0",
92
- device=device,
93
  projection_hidden_dim=2048,
94
- torch_dtype=torch.float16 if device == "cuda" else torch.float32,
95
- low_cpu_mem_usage=True,
96
- use_safetensors=True,
97
- load_in_8bit=True if device == "cuda" else False,
98
- trust_remote_code=True
99
  )
100
 
101
  # Configure model for inference
102
  if hasattr(model, 'language_model'):
103
- model.language_model.config.use_cache = False
104
  model.language_model.eval()
105
 
106
- # Set generation config
 
 
 
 
 
 
 
107
  if hasattr(model.language_model, 'generation_config'):
108
  model.language_model.generation_config.do_sample = True
109
  model.language_model.generation_config.max_new_tokens = 256
110
  model.language_model.generation_config.temperature = 0.7
111
  model.language_model.generation_config.top_p = 0.9
112
- model.language_model.generation_config.pad_token_id = model.language_model.config.eos_token_id
113
-
114
- # Move model to device
115
- model = model.to(device)
116
 
117
  model_status.update({
118
  "initialized": True,
@@ -121,11 +121,10 @@ def initialize_model():
121
  "model_info": {
122
  "vision_model": "openai/clip-vit-base-patch32",
123
  "language_model": "TinyLlama/TinyLlama-1.1B-Chat-v1.0",
124
- "dtype": str(model.dtype),
125
  "device": str(model.device)
126
  }
127
  })
128
- logger.info(f"Model successfully initialized on {model.device} with dtype {model.dtype}")
129
  return True
130
 
131
  except Exception as e:
 
85
  device = "cuda" if torch.cuda.is_available() else "cpu"
86
  logger.info(f"Using device: {device}")
87
 
88
+ # Initialize new model with basic parameters
89
  model = LLaVA(
90
  vision_model_path="openai/clip-vit-base-patch32",
91
  language_model_path="TinyLlama/TinyLlama-1.1B-Chat-v1.0",
 
92
  projection_hidden_dim=2048,
93
+ device=device
 
 
 
 
94
  )
95
 
96
  # Configure model for inference
97
  if hasattr(model, 'language_model'):
98
+ # Set model to evaluation mode
99
  model.language_model.eval()
100
 
101
+ # Configure model for inference
102
+ if hasattr(model.language_model, 'config'):
103
+ model.language_model.config.use_cache = False
104
+
105
+ # Move model to device
106
+ model = model.to(device)
107
+
108
+ # Set generation config if available
109
  if hasattr(model.language_model, 'generation_config'):
110
  model.language_model.generation_config.do_sample = True
111
  model.language_model.generation_config.max_new_tokens = 256
112
  model.language_model.generation_config.temperature = 0.7
113
  model.language_model.generation_config.top_p = 0.9
114
+ if hasattr(model.language_model.config, 'eos_token_id'):
115
+ model.language_model.generation_config.pad_token_id = model.language_model.config.eos_token_id
 
 
116
 
117
  model_status.update({
118
  "initialized": True,
 
121
  "model_info": {
122
  "vision_model": "openai/clip-vit-base-patch32",
123
  "language_model": "TinyLlama/TinyLlama-1.1B-Chat-v1.0",
 
124
  "device": str(model.device)
125
  }
126
  })
127
+ logger.info(f"Model successfully initialized on {model.device}")
128
  return True
129
 
130
  except Exception as e: