Spaces:
Running
Running
Upload folder using huggingface_hub
Browse files- auto_diffusers.log +907 -0
- gradio_app.py +331 -29
auto_diffusers.log
CHANGED
@@ -18538,3 +18538,910 @@ IMPORTANT GUIDELINES:
|
|
18538 |
2025-06-05 01:24:20,475 - simple_memory_calculator - DEBUG - Model memory: 24.0GB, Inference memory: 36.0GB
|
18539 |
2025-06-05 01:24:20,475 - simple_memory_calculator - INFO - Getting memory requirements for model: black-forest-labs/FLUX.1-schnell
|
18540 |
2025-06-05 01:24:20,475 - simple_memory_calculator - DEBUG - Using cached memory data for black-forest-labs/FLUX.1-schnell
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
18538 |
2025-06-05 01:24:20,475 - simple_memory_calculator - DEBUG - Model memory: 24.0GB, Inference memory: 36.0GB
|
18539 |
2025-06-05 01:24:20,475 - simple_memory_calculator - INFO - Getting memory requirements for model: black-forest-labs/FLUX.1-schnell
|
18540 |
2025-06-05 01:24:20,475 - simple_memory_calculator - DEBUG - Using cached memory data for black-forest-labs/FLUX.1-schnell
|
18541 |
+
2025-06-11 10:50:13,595 - __main__ - INFO - Initializing GradioAutodiffusers
|
18542 |
+
2025-06-11 10:50:13,595 - __main__ - DEBUG - API key found, length: 39
|
18543 |
+
2025-06-11 10:50:13,595 - auto_diffusers - INFO - Initializing AutoDiffusersGenerator
|
18544 |
+
2025-06-11 10:50:13,595 - auto_diffusers - DEBUG - API key length: 39
|
18545 |
+
2025-06-11 10:50:13,595 - auto_diffusers - WARNING - Tool calling dependencies not available, running without tools
|
18546 |
+
2025-06-11 10:50:13,596 - hardware_detector - INFO - Initializing HardwareDetector
|
18547 |
+
2025-06-11 10:50:13,596 - hardware_detector - DEBUG - Starting system hardware detection
|
18548 |
+
2025-06-11 10:50:13,596 - hardware_detector - DEBUG - Platform: Darwin, Architecture: arm64
|
18549 |
+
2025-06-11 10:50:13,596 - hardware_detector - DEBUG - CPU cores: 16, Python: 3.11.11
|
18550 |
+
2025-06-11 10:50:13,596 - hardware_detector - DEBUG - Attempting GPU detection via nvidia-smi
|
18551 |
+
2025-06-11 10:50:13,599 - hardware_detector - DEBUG - nvidia-smi not found, no NVIDIA GPU detected
|
18552 |
+
2025-06-11 10:50:13,599 - hardware_detector - DEBUG - Checking PyTorch availability
|
18553 |
+
2025-06-11 10:50:15,597 - hardware_detector - INFO - PyTorch 2.7.0 detected
|
18554 |
+
2025-06-11 10:50:15,597 - hardware_detector - DEBUG - CUDA available: False, MPS available: True
|
18555 |
+
2025-06-11 10:50:15,597 - hardware_detector - INFO - Hardware detection completed successfully
|
18556 |
+
2025-06-11 10:50:15,597 - hardware_detector - DEBUG - Detected specs: {'platform': 'Darwin', 'architecture': 'arm64', 'cpu_count': 16, 'python_version': '3.11.11', 'gpu_info': None, 'cuda_available': False, 'mps_available': True, 'torch_version': '2.7.0'}
|
18557 |
+
2025-06-11 10:50:15,597 - auto_diffusers - INFO - Hardware detector initialized successfully
|
18558 |
+
2025-06-11 10:50:15,597 - __main__ - INFO - AutoDiffusersGenerator initialized successfully
|
18559 |
+
2025-06-11 10:50:15,597 - simple_memory_calculator - INFO - Initializing SimpleMemoryCalculator
|
18560 |
+
2025-06-11 10:50:15,597 - simple_memory_calculator - DEBUG - HuggingFace API initialized
|
18561 |
+
2025-06-11 10:50:15,597 - simple_memory_calculator - DEBUG - Known models in database: 4
|
18562 |
+
2025-06-11 10:50:15,597 - __main__ - INFO - SimpleMemoryCalculator initialized successfully
|
18563 |
+
2025-06-11 10:50:15,597 - __main__ - DEBUG - Default model settings: gemini-2.5-flash-preview-05-20, temp=0.7
|
18564 |
+
2025-06-11 10:50:15,599 - asyncio - DEBUG - Using selector: KqueueSelector
|
18565 |
+
2025-06-11 10:50:15,611 - httpcore.connection - DEBUG - connect_tcp.started host='api.gradio.app' port=443 local_address=None timeout=3 socket_options=None
|
18566 |
+
2025-06-11 10:50:15,620 - urllib3.connectionpool - DEBUG - Starting new HTTPS connection (1): huggingface.co:443
|
18567 |
+
2025-06-11 10:50:15,847 - httpcore.connection - DEBUG - connect_tcp.complete return_value=<httpcore._backends.sync.SyncStream object at 0x140ba7210>
|
18568 |
+
2025-06-11 10:50:15,847 - httpcore.connection - DEBUG - start_tls.started ssl_context=<ssl.SSLContext object at 0x140aa1eb0> server_hostname='api.gradio.app' timeout=3
|
18569 |
+
2025-06-11 10:50:15,853 - urllib3.connectionpool - DEBUG - https://huggingface.co:443 "HEAD /api/telemetry/gradio/initiated HTTP/1.1" 200 0
|
18570 |
+
2025-06-11 10:50:16,123 - httpcore.connection - DEBUG - start_tls.complete return_value=<httpcore._backends.sync.SyncStream object at 0x140d913d0>
|
18571 |
+
2025-06-11 10:50:16,123 - httpcore.http11 - DEBUG - send_request_headers.started request=<Request [b'GET']>
|
18572 |
+
2025-06-11 10:50:16,124 - httpcore.http11 - DEBUG - send_request_headers.complete
|
18573 |
+
2025-06-11 10:50:16,124 - httpcore.http11 - DEBUG - send_request_body.started request=<Request [b'GET']>
|
18574 |
+
2025-06-11 10:50:16,124 - httpcore.http11 - DEBUG - send_request_body.complete
|
18575 |
+
2025-06-11 10:50:16,125 - httpcore.http11 - DEBUG - receive_response_headers.started request=<Request [b'GET']>
|
18576 |
+
2025-06-11 10:50:16,263 - httpcore.http11 - DEBUG - receive_response_headers.complete return_value=(b'HTTP/1.1', 200, b'OK', [(b'Date', b'Wed, 11 Jun 2025 01:50:16 GMT'), (b'Content-Type', b'application/json'), (b'Content-Length', b'21'), (b'Connection', b'keep-alive'), (b'Server', b'nginx/1.18.0'), (b'Access-Control-Allow-Origin', b'*')])
|
18577 |
+
2025-06-11 10:50:16,263 - httpx - INFO - HTTP Request: GET https://api.gradio.app/pkg-version "HTTP/1.1 200 OK"
|
18578 |
+
2025-06-11 10:50:16,264 - httpcore.http11 - DEBUG - receive_response_body.started request=<Request [b'GET']>
|
18579 |
+
2025-06-11 10:50:16,264 - httpcore.http11 - DEBUG - receive_response_body.complete
|
18580 |
+
2025-06-11 10:50:16,264 - httpcore.http11 - DEBUG - response_closed.started
|
18581 |
+
2025-06-11 10:50:16,264 - httpcore.http11 - DEBUG - response_closed.complete
|
18582 |
+
2025-06-11 10:50:16,264 - httpcore.connection - DEBUG - close.started
|
18583 |
+
2025-06-11 10:50:16,265 - httpcore.connection - DEBUG - close.complete
|
18584 |
+
2025-06-11 10:51:30,295 - __main__ - INFO - Initializing GradioAutodiffusers
|
18585 |
+
2025-06-11 10:51:30,295 - __main__ - DEBUG - API key found, length: 39
|
18586 |
+
2025-06-11 10:51:30,295 - auto_diffusers - INFO - Initializing AutoDiffusersGenerator
|
18587 |
+
2025-06-11 10:51:30,296 - auto_diffusers - DEBUG - API key length: 39
|
18588 |
+
2025-06-11 10:51:30,296 - auto_diffusers - WARNING - Tool calling dependencies not available, running without tools
|
18589 |
+
2025-06-11 10:51:30,296 - hardware_detector - INFO - Initializing HardwareDetector
|
18590 |
+
2025-06-11 10:51:30,296 - hardware_detector - DEBUG - Starting system hardware detection
|
18591 |
+
2025-06-11 10:51:30,296 - hardware_detector - DEBUG - Platform: Darwin, Architecture: arm64
|
18592 |
+
2025-06-11 10:51:30,296 - hardware_detector - DEBUG - CPU cores: 16, Python: 3.11.11
|
18593 |
+
2025-06-11 10:51:30,296 - hardware_detector - DEBUG - Attempting GPU detection via nvidia-smi
|
18594 |
+
2025-06-11 10:51:30,299 - hardware_detector - DEBUG - nvidia-smi not found, no NVIDIA GPU detected
|
18595 |
+
2025-06-11 10:51:30,299 - hardware_detector - DEBUG - Checking PyTorch availability
|
18596 |
+
2025-06-11 10:51:30,748 - hardware_detector - INFO - PyTorch 2.7.0 detected
|
18597 |
+
2025-06-11 10:51:30,748 - hardware_detector - DEBUG - CUDA available: False, MPS available: True
|
18598 |
+
2025-06-11 10:51:30,749 - hardware_detector - INFO - Hardware detection completed successfully
|
18599 |
+
2025-06-11 10:51:30,749 - hardware_detector - DEBUG - Detected specs: {'platform': 'Darwin', 'architecture': 'arm64', 'cpu_count': 16, 'python_version': '3.11.11', 'gpu_info': None, 'cuda_available': False, 'mps_available': True, 'torch_version': '2.7.0'}
|
18600 |
+
2025-06-11 10:51:30,749 - auto_diffusers - INFO - Hardware detector initialized successfully
|
18601 |
+
2025-06-11 10:51:30,749 - __main__ - INFO - AutoDiffusersGenerator initialized successfully
|
18602 |
+
2025-06-11 10:51:30,749 - simple_memory_calculator - INFO - Initializing SimpleMemoryCalculator
|
18603 |
+
2025-06-11 10:51:30,749 - simple_memory_calculator - DEBUG - HuggingFace API initialized
|
18604 |
+
2025-06-11 10:51:30,749 - simple_memory_calculator - DEBUG - Known models in database: 4
|
18605 |
+
2025-06-11 10:51:30,749 - __main__ - INFO - SimpleMemoryCalculator initialized successfully
|
18606 |
+
2025-06-11 10:51:30,749 - __main__ - DEBUG - Default model settings: gemini-2.5-flash-preview-05-20, temp=0.7
|
18607 |
+
2025-06-11 10:51:30,751 - asyncio - DEBUG - Using selector: KqueueSelector
|
18608 |
+
2025-06-11 10:51:30,764 - httpcore.connection - DEBUG - connect_tcp.started host='api.gradio.app' port=443 local_address=None timeout=3 socket_options=None
|
18609 |
+
2025-06-11 10:51:30,771 - urllib3.connectionpool - DEBUG - Starting new HTTPS connection (1): huggingface.co:443
|
18610 |
+
2025-06-11 10:51:30,850 - asyncio - DEBUG - Using selector: KqueueSelector
|
18611 |
+
2025-06-11 10:51:30,883 - httpcore.connection - DEBUG - connect_tcp.started host='localhost' port=7860 local_address=None timeout=None socket_options=None
|
18612 |
+
2025-06-11 10:51:30,884 - httpcore.connection - DEBUG - connect_tcp.complete return_value=<httpcore._backends.sync.SyncStream object at 0x119eae950>
|
18613 |
+
2025-06-11 10:51:30,884 - httpcore.http11 - DEBUG - send_request_headers.started request=<Request [b'GET']>
|
18614 |
+
2025-06-11 10:51:30,884 - httpcore.http11 - DEBUG - send_request_headers.complete
|
18615 |
+
2025-06-11 10:51:30,884 - httpcore.http11 - DEBUG - send_request_body.started request=<Request [b'GET']>
|
18616 |
+
2025-06-11 10:51:30,884 - httpcore.http11 - DEBUG - send_request_body.complete
|
18617 |
+
2025-06-11 10:51:30,885 - httpcore.http11 - DEBUG - receive_response_headers.started request=<Request [b'GET']>
|
18618 |
+
2025-06-11 10:51:30,885 - httpcore.http11 - DEBUG - receive_response_headers.complete return_value=(b'HTTP/1.1', 200, b'OK', [(b'date', b'Wed, 11 Jun 2025 01:51:30 GMT'), (b'server', b'uvicorn'), (b'content-length', b'4'), (b'content-type', b'application/json')])
|
18619 |
+
2025-06-11 10:51:30,885 - httpx - INFO - HTTP Request: GET http://localhost:7860/gradio_api/startup-events "HTTP/1.1 200 OK"
|
18620 |
+
2025-06-11 10:51:30,885 - httpcore.http11 - DEBUG - receive_response_body.started request=<Request [b'GET']>
|
18621 |
+
2025-06-11 10:51:30,885 - httpcore.http11 - DEBUG - receive_response_body.complete
|
18622 |
+
2025-06-11 10:51:30,885 - httpcore.http11 - DEBUG - response_closed.started
|
18623 |
+
2025-06-11 10:51:30,885 - httpcore.http11 - DEBUG - response_closed.complete
|
18624 |
+
2025-06-11 10:51:30,885 - httpcore.connection - DEBUG - close.started
|
18625 |
+
2025-06-11 10:51:30,885 - httpcore.connection - DEBUG - close.complete
|
18626 |
+
2025-06-11 10:51:30,886 - httpcore.connection - DEBUG - connect_tcp.started host='localhost' port=7860 local_address=None timeout=3 socket_options=None
|
18627 |
+
2025-06-11 10:51:30,886 - httpcore.connection - DEBUG - connect_tcp.complete return_value=<httpcore._backends.sync.SyncStream object at 0x11fb5e210>
|
18628 |
+
2025-06-11 10:51:30,886 - httpcore.http11 - DEBUG - send_request_headers.started request=<Request [b'HEAD']>
|
18629 |
+
2025-06-11 10:51:30,886 - httpcore.http11 - DEBUG - send_request_headers.complete
|
18630 |
+
2025-06-11 10:51:30,886 - httpcore.http11 - DEBUG - send_request_body.started request=<Request [b'HEAD']>
|
18631 |
+
2025-06-11 10:51:30,886 - httpcore.http11 - DEBUG - send_request_body.complete
|
18632 |
+
2025-06-11 10:51:30,886 - httpcore.http11 - DEBUG - receive_response_headers.started request=<Request [b'HEAD']>
|
18633 |
+
2025-06-11 10:51:30,894 - httpcore.http11 - DEBUG - receive_response_headers.complete return_value=(b'HTTP/1.1', 200, b'OK', [(b'date', b'Wed, 11 Jun 2025 01:51:30 GMT'), (b'server', b'uvicorn'), (b'content-length', b'138668'), (b'content-type', b'text/html; charset=utf-8')])
|
18634 |
+
2025-06-11 10:51:30,894 - httpx - INFO - HTTP Request: HEAD http://localhost:7860/ "HTTP/1.1 200 OK"
|
18635 |
+
2025-06-11 10:51:30,894 - httpcore.http11 - DEBUG - receive_response_body.started request=<Request [b'HEAD']>
|
18636 |
+
2025-06-11 10:51:30,894 - httpcore.http11 - DEBUG - receive_response_body.complete
|
18637 |
+
2025-06-11 10:51:30,894 - httpcore.http11 - DEBUG - response_closed.started
|
18638 |
+
2025-06-11 10:51:30,894 - httpcore.http11 - DEBUG - response_closed.complete
|
18639 |
+
2025-06-11 10:51:30,894 - httpcore.connection - DEBUG - close.started
|
18640 |
+
2025-06-11 10:51:30,894 - httpcore.connection - DEBUG - close.complete
|
18641 |
+
2025-06-11 10:51:30,906 - httpcore.connection - DEBUG - connect_tcp.started host='api.gradio.app' port=443 local_address=None timeout=30 socket_options=None
|
18642 |
+
2025-06-11 10:51:30,915 - httpcore.connection - DEBUG - connect_tcp.complete return_value=<httpcore._backends.sync.SyncStream object at 0x119317990>
|
18643 |
+
2025-06-11 10:51:30,915 - httpcore.connection - DEBUG - start_tls.started ssl_context=<ssl.SSLContext object at 0x111d311c0> server_hostname='api.gradio.app' timeout=3
|
18644 |
+
2025-06-11 10:51:31,035 - urllib3.connectionpool - DEBUG - https://huggingface.co:443 "HEAD /api/telemetry/gradio/initiated HTTP/1.1" 200 0
|
18645 |
+
2025-06-11 10:51:31,064 - httpcore.connection - DEBUG - connect_tcp.complete return_value=<httpcore._backends.sync.SyncStream object at 0x119c67f90>
|
18646 |
+
2025-06-11 10:51:31,064 - httpcore.connection - DEBUG - start_tls.started ssl_context=<ssl.SSLContext object at 0x11fae7f50> server_hostname='api.gradio.app' timeout=30
|
18647 |
+
2025-06-11 10:51:31,200 - httpcore.connection - DEBUG - start_tls.complete return_value=<httpcore._backends.sync.SyncStream object at 0x11a1284d0>
|
18648 |
+
2025-06-11 10:51:31,201 - httpcore.http11 - DEBUG - send_request_headers.started request=<Request [b'GET']>
|
18649 |
+
2025-06-11 10:51:31,201 - httpcore.http11 - DEBUG - send_request_headers.complete
|
18650 |
+
2025-06-11 10:51:31,201 - httpcore.http11 - DEBUG - send_request_body.started request=<Request [b'GET']>
|
18651 |
+
2025-06-11 10:51:31,202 - httpcore.http11 - DEBUG - send_request_body.complete
|
18652 |
+
2025-06-11 10:51:31,202 - httpcore.http11 - DEBUG - receive_response_headers.started request=<Request [b'GET']>
|
18653 |
+
2025-06-11 10:51:31,345 - httpcore.http11 - DEBUG - receive_response_headers.complete return_value=(b'HTTP/1.1', 200, b'OK', [(b'Date', b'Wed, 11 Jun 2025 01:51:31 GMT'), (b'Content-Type', b'application/json'), (b'Content-Length', b'21'), (b'Connection', b'keep-alive'), (b'Server', b'nginx/1.18.0'), (b'Access-Control-Allow-Origin', b'*')])
|
18654 |
+
2025-06-11 10:51:31,346 - httpx - INFO - HTTP Request: GET https://api.gradio.app/pkg-version "HTTP/1.1 200 OK"
|
18655 |
+
2025-06-11 10:51:31,346 - httpcore.http11 - DEBUG - receive_response_body.started request=<Request [b'GET']>
|
18656 |
+
2025-06-11 10:51:31,346 - httpcore.http11 - DEBUG - receive_response_body.complete
|
18657 |
+
2025-06-11 10:51:31,346 - httpcore.http11 - DEBUG - response_closed.started
|
18658 |
+
2025-06-11 10:51:31,346 - httpcore.http11 - DEBUG - response_closed.complete
|
18659 |
+
2025-06-11 10:51:31,347 - httpcore.connection - DEBUG - close.started
|
18660 |
+
2025-06-11 10:51:31,347 - httpcore.connection - DEBUG - close.complete
|
18661 |
+
2025-06-11 10:51:31,382 - httpcore.connection - DEBUG - start_tls.complete return_value=<httpcore._backends.sync.SyncStream object at 0x1198c9010>
|
18662 |
+
2025-06-11 10:51:31,382 - httpcore.http11 - DEBUG - send_request_headers.started request=<Request [b'GET']>
|
18663 |
+
2025-06-11 10:51:31,383 - httpcore.http11 - DEBUG - send_request_headers.complete
|
18664 |
+
2025-06-11 10:51:31,383 - httpcore.http11 - DEBUG - send_request_body.started request=<Request [b'GET']>
|
18665 |
+
2025-06-11 10:51:31,383 - httpcore.http11 - DEBUG - send_request_body.complete
|
18666 |
+
2025-06-11 10:51:31,383 - httpcore.http11 - DEBUG - receive_response_headers.started request=<Request [b'GET']>
|
18667 |
+
2025-06-11 10:51:31,544 - httpcore.http11 - DEBUG - receive_response_headers.complete return_value=(b'HTTP/1.1', 200, b'OK', [(b'Date', b'Wed, 11 Jun 2025 01:51:31 GMT'), (b'Content-Type', b'text/html; charset=utf-8'), (b'Transfer-Encoding', b'chunked'), (b'Connection', b'keep-alive'), (b'Server', b'nginx/1.18.0'), (b'ContentType', b'application/json'), (b'Access-Control-Allow-Origin', b'*'), (b'Content-Encoding', b'gzip')])
|
18668 |
+
2025-06-11 10:51:31,544 - httpx - INFO - HTTP Request: GET https://api.gradio.app/v3/tunnel-request "HTTP/1.1 200 OK"
|
18669 |
+
2025-06-11 10:51:31,545 - httpcore.http11 - DEBUG - receive_response_body.started request=<Request [b'GET']>
|
18670 |
+
2025-06-11 10:51:31,545 - httpcore.http11 - DEBUG - receive_response_body.complete
|
18671 |
+
2025-06-11 10:51:31,545 - httpcore.http11 - DEBUG - response_closed.started
|
18672 |
+
2025-06-11 10:51:31,546 - httpcore.http11 - DEBUG - response_closed.complete
|
18673 |
+
2025-06-11 10:51:31,546 - httpcore.connection - DEBUG - close.started
|
18674 |
+
2025-06-11 10:51:31,546 - httpcore.connection - DEBUG - close.complete
|
18675 |
+
2025-06-11 10:51:32,314 - urllib3.connectionpool - DEBUG - Starting new HTTPS connection (1): huggingface.co:443
|
18676 |
+
2025-06-11 10:51:32,526 - urllib3.connectionpool - DEBUG - https://huggingface.co:443 "HEAD /api/telemetry/gradio/launched HTTP/1.1" 200 0
|
18677 |
+
2025-06-11 10:51:41,746 - simple_memory_calculator - INFO - Getting memory requirements for model: black-forest-labs/FLUX.1-schnell
|
18678 |
+
2025-06-11 10:51:41,746 - hardware_detector - INFO - Initializing HardwareDetector
|
18679 |
+
2025-06-11 10:51:41,746 - simple_memory_calculator - INFO - Using known memory data for black-forest-labs/FLUX.1-schnell
|
18680 |
+
2025-06-11 10:51:41,746 - hardware_detector - DEBUG - Starting system hardware detection
|
18681 |
+
2025-06-11 10:51:41,747 - simple_memory_calculator - DEBUG - Known data: {'params_billions': 12.0, 'fp16_gb': 24.0, 'inference_fp16_gb': 36.0}
|
18682 |
+
2025-06-11 10:51:41,747 - hardware_detector - DEBUG - Platform: Darwin, Architecture: arm64
|
18683 |
+
2025-06-11 10:51:41,747 - simple_memory_calculator - INFO - Generating memory recommendations for black-forest-labs/FLUX.1-schnell with 8.0GB VRAM
|
18684 |
+
2025-06-11 10:51:41,747 - hardware_detector - DEBUG - CPU cores: 16, Python: 3.11.11
|
18685 |
+
2025-06-11 10:51:41,747 - simple_memory_calculator - INFO - Getting memory requirements for model: black-forest-labs/FLUX.1-schnell
|
18686 |
+
2025-06-11 10:51:41,747 - hardware_detector - DEBUG - Attempting GPU detection via nvidia-smi
|
18687 |
+
2025-06-11 10:51:41,747 - simple_memory_calculator - DEBUG - Using cached memory data for black-forest-labs/FLUX.1-schnell
|
18688 |
+
2025-06-11 10:51:41,747 - simple_memory_calculator - DEBUG - Model memory: 24.0GB, Inference memory: 36.0GB
|
18689 |
+
2025-06-11 10:51:41,748 - simple_memory_calculator - INFO - Getting memory requirements for model: black-forest-labs/FLUX.1-schnell
|
18690 |
+
2025-06-11 10:51:41,748 - simple_memory_calculator - DEBUG - Using cached memory data for black-forest-labs/FLUX.1-schnell
|
18691 |
+
2025-06-11 10:51:41,754 - hardware_detector - DEBUG - nvidia-smi not found, no NVIDIA GPU detected
|
18692 |
+
2025-06-11 10:51:41,755 - hardware_detector - DEBUG - Checking PyTorch availability
|
18693 |
+
2025-06-11 10:51:41,755 - hardware_detector - INFO - PyTorch 2.7.0 detected
|
18694 |
+
2025-06-11 10:51:41,755 - hardware_detector - DEBUG - CUDA available: False, MPS available: True
|
18695 |
+
2025-06-11 10:51:41,755 - hardware_detector - INFO - Hardware detection completed successfully
|
18696 |
+
2025-06-11 10:51:41,755 - hardware_detector - DEBUG - Detected specs: {'platform': 'Darwin', 'architecture': 'arm64', 'cpu_count': 16, 'python_version': '3.11.11', 'gpu_info': None, 'cuda_available': False, 'mps_available': True, 'torch_version': '2.7.0'}
|
18697 |
+
2025-06-11 10:58:51,900 - __main__ - INFO - Initializing GradioAutodiffusers
|
18698 |
+
2025-06-11 10:58:51,900 - __main__ - DEBUG - API key found, length: 39
|
18699 |
+
2025-06-11 10:58:51,900 - auto_diffusers - INFO - Initializing AutoDiffusersGenerator
|
18700 |
+
2025-06-11 10:58:51,900 - auto_diffusers - DEBUG - API key length: 39
|
18701 |
+
2025-06-11 10:58:51,900 - auto_diffusers - WARNING - Tool calling dependencies not available, running without tools
|
18702 |
+
2025-06-11 10:58:51,900 - hardware_detector - INFO - Initializing HardwareDetector
|
18703 |
+
2025-06-11 10:58:51,900 - hardware_detector - DEBUG - Starting system hardware detection
|
18704 |
+
2025-06-11 10:58:51,900 - hardware_detector - DEBUG - Platform: Darwin, Architecture: arm64
|
18705 |
+
2025-06-11 10:58:51,900 - hardware_detector - DEBUG - CPU cores: 16, Python: 3.11.11
|
18706 |
+
2025-06-11 10:58:51,900 - hardware_detector - DEBUG - Attempting GPU detection via nvidia-smi
|
18707 |
+
2025-06-11 10:58:51,904 - hardware_detector - DEBUG - nvidia-smi not found, no NVIDIA GPU detected
|
18708 |
+
2025-06-11 10:58:51,904 - hardware_detector - DEBUG - Checking PyTorch availability
|
18709 |
+
2025-06-11 10:58:52,365 - hardware_detector - INFO - PyTorch 2.7.0 detected
|
18710 |
+
2025-06-11 10:58:52,365 - hardware_detector - DEBUG - CUDA available: False, MPS available: True
|
18711 |
+
2025-06-11 10:58:52,365 - hardware_detector - INFO - Hardware detection completed successfully
|
18712 |
+
2025-06-11 10:58:52,365 - hardware_detector - DEBUG - Detected specs: {'platform': 'Darwin', 'architecture': 'arm64', 'cpu_count': 16, 'python_version': '3.11.11', 'gpu_info': None, 'cuda_available': False, 'mps_available': True, 'torch_version': '2.7.0'}
|
18713 |
+
2025-06-11 10:58:52,365 - auto_diffusers - INFO - Hardware detector initialized successfully
|
18714 |
+
2025-06-11 10:58:52,365 - __main__ - INFO - AutoDiffusersGenerator initialized successfully
|
18715 |
+
2025-06-11 10:58:52,365 - simple_memory_calculator - INFO - Initializing SimpleMemoryCalculator
|
18716 |
+
2025-06-11 10:58:52,365 - simple_memory_calculator - DEBUG - HuggingFace API initialized
|
18717 |
+
2025-06-11 10:58:52,365 - simple_memory_calculator - DEBUG - Known models in database: 4
|
18718 |
+
2025-06-11 10:58:52,365 - __main__ - INFO - SimpleMemoryCalculator initialized successfully
|
18719 |
+
2025-06-11 10:58:52,365 - __main__ - DEBUG - Default model settings: gemini-2.5-flash-preview-05-20, temp=0.7
|
18720 |
+
2025-06-11 10:58:52,368 - asyncio - DEBUG - Using selector: KqueueSelector
|
18721 |
+
2025-06-11 10:58:52,381 - httpcore.connection - DEBUG - connect_tcp.started host='api.gradio.app' port=443 local_address=None timeout=3 socket_options=None
|
18722 |
+
2025-06-11 10:58:52,388 - urllib3.connectionpool - DEBUG - Starting new HTTPS connection (1): huggingface.co:443
|
18723 |
+
2025-06-11 10:58:52,468 - asyncio - DEBUG - Using selector: KqueueSelector
|
18724 |
+
2025-06-11 10:58:52,500 - httpcore.connection - DEBUG - connect_tcp.started host='localhost' port=7860 local_address=None timeout=None socket_options=None
|
18725 |
+
2025-06-11 10:58:52,501 - httpcore.connection - DEBUG - connect_tcp.complete return_value=<httpcore._backends.sync.SyncStream object at 0x11979ae90>
|
18726 |
+
2025-06-11 10:58:52,501 - httpcore.http11 - DEBUG - send_request_headers.started request=<Request [b'GET']>
|
18727 |
+
2025-06-11 10:58:52,501 - httpcore.http11 - DEBUG - send_request_headers.complete
|
18728 |
+
2025-06-11 10:58:52,501 - httpcore.http11 - DEBUG - send_request_body.started request=<Request [b'GET']>
|
18729 |
+
2025-06-11 10:58:52,502 - httpcore.http11 - DEBUG - send_request_body.complete
|
18730 |
+
2025-06-11 10:58:52,502 - httpcore.http11 - DEBUG - receive_response_headers.started request=<Request [b'GET']>
|
18731 |
+
2025-06-11 10:58:52,502 - httpcore.http11 - DEBUG - receive_response_headers.complete return_value=(b'HTTP/1.1', 200, b'OK', [(b'date', b'Wed, 11 Jun 2025 01:58:52 GMT'), (b'server', b'uvicorn'), (b'content-length', b'4'), (b'content-type', b'application/json')])
|
18732 |
+
2025-06-11 10:58:52,502 - httpx - INFO - HTTP Request: GET http://localhost:7860/gradio_api/startup-events "HTTP/1.1 200 OK"
|
18733 |
+
2025-06-11 10:58:52,502 - httpcore.http11 - DEBUG - receive_response_body.started request=<Request [b'GET']>
|
18734 |
+
2025-06-11 10:58:52,502 - httpcore.http11 - DEBUG - receive_response_body.complete
|
18735 |
+
2025-06-11 10:58:52,502 - httpcore.http11 - DEBUG - response_closed.started
|
18736 |
+
2025-06-11 10:58:52,502 - httpcore.http11 - DEBUG - response_closed.complete
|
18737 |
+
2025-06-11 10:58:52,502 - httpcore.connection - DEBUG - close.started
|
18738 |
+
2025-06-11 10:58:52,502 - httpcore.connection - DEBUG - close.complete
|
18739 |
+
2025-06-11 10:58:52,503 - httpcore.connection - DEBUG - connect_tcp.started host='localhost' port=7860 local_address=None timeout=3 socket_options=None
|
18740 |
+
2025-06-11 10:58:52,503 - httpcore.connection - DEBUG - connect_tcp.complete return_value=<httpcore._backends.sync.SyncStream object at 0x11e0c3990>
|
18741 |
+
2025-06-11 10:58:52,503 - httpcore.http11 - DEBUG - send_request_headers.started request=<Request [b'HEAD']>
|
18742 |
+
2025-06-11 10:58:52,503 - httpcore.http11 - DEBUG - send_request_headers.complete
|
18743 |
+
2025-06-11 10:58:52,503 - httpcore.http11 - DEBUG - send_request_body.started request=<Request [b'HEAD']>
|
18744 |
+
2025-06-11 10:58:52,503 - httpcore.http11 - DEBUG - send_request_body.complete
|
18745 |
+
2025-06-11 10:58:52,503 - httpcore.http11 - DEBUG - receive_response_headers.started request=<Request [b'HEAD']>
|
18746 |
+
2025-06-11 10:58:52,510 - httpcore.http11 - DEBUG - receive_response_headers.complete return_value=(b'HTTP/1.1', 200, b'OK', [(b'date', b'Wed, 11 Jun 2025 01:58:52 GMT'), (b'server', b'uvicorn'), (b'content-length', b'145792'), (b'content-type', b'text/html; charset=utf-8')])
|
18747 |
+
2025-06-11 10:58:52,510 - httpx - INFO - HTTP Request: HEAD http://localhost:7860/ "HTTP/1.1 200 OK"
|
18748 |
+
2025-06-11 10:58:52,510 - httpcore.http11 - DEBUG - receive_response_body.started request=<Request [b'HEAD']>
|
18749 |
+
2025-06-11 10:58:52,510 - httpcore.http11 - DEBUG - receive_response_body.complete
|
18750 |
+
2025-06-11 10:58:52,510 - httpcore.http11 - DEBUG - response_closed.started
|
18751 |
+
2025-06-11 10:58:52,511 - httpcore.http11 - DEBUG - response_closed.complete
|
18752 |
+
2025-06-11 10:58:52,511 - httpcore.connection - DEBUG - close.started
|
18753 |
+
2025-06-11 10:58:52,511 - httpcore.connection - DEBUG - close.complete
|
18754 |
+
2025-06-11 10:58:52,522 - httpcore.connection - DEBUG - connect_tcp.started host='api.gradio.app' port=443 local_address=None timeout=30 socket_options=None
|
18755 |
+
2025-06-11 10:58:52,627 - httpcore.connection - DEBUG - connect_tcp.complete return_value=<httpcore._backends.sync.SyncStream object at 0x119093c50>
|
18756 |
+
2025-06-11 10:58:52,627 - httpcore.connection - DEBUG - start_tls.started ssl_context=<ssl.SSLContext object at 0x107d31370> server_hostname='api.gradio.app' timeout=3
|
18757 |
+
2025-06-11 10:58:52,662 - httpcore.connection - DEBUG - connect_tcp.complete return_value=<httpcore._backends.sync.SyncStream object at 0x11e139450>
|
18758 |
+
2025-06-11 10:58:52,662 - httpcore.connection - DEBUG - start_tls.started ssl_context=<ssl.SSLContext object at 0x11e0c8a70> server_hostname='api.gradio.app' timeout=30
|
18759 |
+
2025-06-11 10:58:52,671 - urllib3.connectionpool - DEBUG - https://huggingface.co:443 "HEAD /api/telemetry/gradio/initiated HTTP/1.1" 200 0
|
18760 |
+
2025-06-11 10:58:52,933 - httpcore.connection - DEBUG - start_tls.complete return_value=<httpcore._backends.sync.SyncStream object at 0x11956d810>
|
18761 |
+
2025-06-11 10:58:52,933 - httpcore.http11 - DEBUG - send_request_headers.started request=<Request [b'GET']>
|
18762 |
+
2025-06-11 10:58:52,933 - httpcore.http11 - DEBUG - send_request_headers.complete
|
18763 |
+
2025-06-11 10:58:52,933 - httpcore.http11 - DEBUG - send_request_body.started request=<Request [b'GET']>
|
18764 |
+
2025-06-11 10:58:52,933 - httpcore.http11 - DEBUG - send_request_body.complete
|
18765 |
+
2025-06-11 10:58:52,933 - httpcore.http11 - DEBUG - receive_response_headers.started request=<Request [b'GET']>
|
18766 |
+
2025-06-11 10:58:52,944 - httpcore.connection - DEBUG - start_tls.complete return_value=<httpcore._backends.sync.SyncStream object at 0x118fc5fd0>
|
18767 |
+
2025-06-11 10:58:52,944 - httpcore.http11 - DEBUG - send_request_headers.started request=<Request [b'GET']>
|
18768 |
+
2025-06-11 10:58:52,944 - httpcore.http11 - DEBUG - send_request_headers.complete
|
18769 |
+
2025-06-11 10:58:52,944 - httpcore.http11 - DEBUG - send_request_body.started request=<Request [b'GET']>
|
18770 |
+
2025-06-11 10:58:52,944 - httpcore.http11 - DEBUG - send_request_body.complete
|
18771 |
+
2025-06-11 10:58:52,945 - httpcore.http11 - DEBUG - receive_response_headers.started request=<Request [b'GET']>
|
18772 |
+
2025-06-11 10:58:53,082 - httpcore.http11 - DEBUG - receive_response_headers.complete return_value=(b'HTTP/1.1', 200, b'OK', [(b'Date', b'Wed, 11 Jun 2025 01:58:53 GMT'), (b'Content-Type', b'application/json'), (b'Content-Length', b'21'), (b'Connection', b'keep-alive'), (b'Server', b'nginx/1.18.0'), (b'Access-Control-Allow-Origin', b'*')])
|
18773 |
+
2025-06-11 10:58:53,082 - httpx - INFO - HTTP Request: GET https://api.gradio.app/pkg-version "HTTP/1.1 200 OK"
|
18774 |
+
2025-06-11 10:58:53,083 - httpcore.http11 - DEBUG - receive_response_body.started request=<Request [b'GET']>
|
18775 |
+
2025-06-11 10:58:53,083 - httpcore.http11 - DEBUG - receive_response_body.complete
|
18776 |
+
2025-06-11 10:58:53,083 - httpcore.http11 - DEBUG - response_closed.started
|
18777 |
+
2025-06-11 10:58:53,083 - httpcore.http11 - DEBUG - response_closed.complete
|
18778 |
+
2025-06-11 10:58:53,083 - httpcore.connection - DEBUG - close.started
|
18779 |
+
2025-06-11 10:58:53,084 - httpcore.connection - DEBUG - close.complete
|
18780 |
+
2025-06-11 10:58:53,087 - httpcore.http11 - DEBUG - receive_response_headers.complete return_value=(b'HTTP/1.1', 200, b'OK', [(b'Date', b'Wed, 11 Jun 2025 01:58:53 GMT'), (b'Content-Type', b'text/html; charset=utf-8'), (b'Transfer-Encoding', b'chunked'), (b'Connection', b'keep-alive'), (b'Server', b'nginx/1.18.0'), (b'ContentType', b'application/json'), (b'Access-Control-Allow-Origin', b'*'), (b'Content-Encoding', b'gzip')])
|
18781 |
+
2025-06-11 10:58:53,087 - httpx - INFO - HTTP Request: GET https://api.gradio.app/v3/tunnel-request "HTTP/1.1 200 OK"
|
18782 |
+
2025-06-11 10:58:53,088 - httpcore.http11 - DEBUG - receive_response_body.started request=<Request [b'GET']>
|
18783 |
+
2025-06-11 10:58:53,089 - httpcore.http11 - DEBUG - receive_response_body.complete
|
18784 |
+
2025-06-11 10:58:53,089 - httpcore.http11 - DEBUG - response_closed.started
|
18785 |
+
2025-06-11 10:58:53,089 - httpcore.http11 - DEBUG - response_closed.complete
|
18786 |
+
2025-06-11 10:58:53,089 - httpcore.connection - DEBUG - close.started
|
18787 |
+
2025-06-11 10:58:53,090 - httpcore.connection - DEBUG - close.complete
|
18788 |
+
2025-06-11 10:58:53,778 - urllib3.connectionpool - DEBUG - Starting new HTTPS connection (1): huggingface.co:443
|
18789 |
+
2025-06-11 10:58:54,003 - urllib3.connectionpool - DEBUG - https://huggingface.co:443 "HEAD /api/telemetry/gradio/launched HTTP/1.1" 200 0
|
18790 |
+
2025-06-11 10:58:59,726 - hardware_detector - INFO - Initializing HardwareDetector
|
18791 |
+
2025-06-11 10:58:59,726 - simple_memory_calculator - INFO - Getting memory requirements for model: black-forest-labs/FLUX.1-schnell
|
18792 |
+
2025-06-11 10:58:59,727 - hardware_detector - DEBUG - Starting system hardware detection
|
18793 |
+
2025-06-11 10:58:59,727 - simple_memory_calculator - INFO - Using known memory data for black-forest-labs/FLUX.1-schnell
|
18794 |
+
2025-06-11 10:58:59,727 - hardware_detector - DEBUG - Platform: Darwin, Architecture: arm64
|
18795 |
+
2025-06-11 10:58:59,727 - simple_memory_calculator - DEBUG - Known data: {'params_billions': 12.0, 'fp16_gb': 24.0, 'inference_fp16_gb': 36.0}
|
18796 |
+
2025-06-11 10:58:59,727 - hardware_detector - DEBUG - CPU cores: 16, Python: 3.11.11
|
18797 |
+
2025-06-11 10:58:59,727 - simple_memory_calculator - INFO - Generating memory recommendations for black-forest-labs/FLUX.1-schnell with 8.0GB VRAM
|
18798 |
+
2025-06-11 10:58:59,727 - hardware_detector - DEBUG - Attempting GPU detection via nvidia-smi
|
18799 |
+
2025-06-11 10:58:59,727 - simple_memory_calculator - INFO - Getting memory requirements for model: black-forest-labs/FLUX.1-schnell
|
18800 |
+
2025-06-11 10:58:59,727 - simple_memory_calculator - DEBUG - Using cached memory data for black-forest-labs/FLUX.1-schnell
|
18801 |
+
2025-06-11 10:58:59,727 - simple_memory_calculator - DEBUG - Model memory: 24.0GB, Inference memory: 36.0GB
|
18802 |
+
2025-06-11 10:58:59,731 - simple_memory_calculator - INFO - Getting memory requirements for model: black-forest-labs/FLUX.1-schnell
|
18803 |
+
2025-06-11 10:58:59,732 - simple_memory_calculator - DEBUG - Using cached memory data for black-forest-labs/FLUX.1-schnell
|
18804 |
+
2025-06-11 10:58:59,734 - hardware_detector - DEBUG - nvidia-smi not found, no NVIDIA GPU detected
|
18805 |
+
2025-06-11 10:58:59,734 - hardware_detector - DEBUG - Checking PyTorch availability
|
18806 |
+
2025-06-11 10:58:59,734 - hardware_detector - INFO - PyTorch 2.7.0 detected
|
18807 |
+
2025-06-11 10:58:59,734 - hardware_detector - DEBUG - CUDA available: False, MPS available: True
|
18808 |
+
2025-06-11 10:58:59,734 - hardware_detector - INFO - Hardware detection completed successfully
|
18809 |
+
2025-06-11 10:58:59,734 - hardware_detector - DEBUG - Detected specs: {'platform': 'Darwin', 'architecture': 'arm64', 'cpu_count': 16, 'python_version': '3.11.11', 'gpu_info': None, 'cuda_available': False, 'mps_available': True, 'torch_version': '2.7.0'}
|
18810 |
+
2025-06-11 10:59:29,945 - hardware_detector - INFO - Initializing HardwareDetector
|
18811 |
+
2025-06-11 10:59:29,946 - hardware_detector - DEBUG - Starting system hardware detection
|
18812 |
+
2025-06-11 10:59:29,946 - hardware_detector - DEBUG - Platform: Darwin, Architecture: arm64
|
18813 |
+
2025-06-11 10:59:29,946 - hardware_detector - DEBUG - CPU cores: 16, Python: 3.11.11
|
18814 |
+
2025-06-11 10:59:29,946 - hardware_detector - DEBUG - Attempting GPU detection via nvidia-smi
|
18815 |
+
2025-06-11 10:59:29,955 - hardware_detector - DEBUG - nvidia-smi not found, no NVIDIA GPU detected
|
18816 |
+
2025-06-11 10:59:29,956 - hardware_detector - DEBUG - Checking PyTorch availability
|
18817 |
+
2025-06-11 10:59:29,956 - hardware_detector - INFO - PyTorch 2.7.0 detected
|
18818 |
+
2025-06-11 10:59:29,956 - hardware_detector - DEBUG - CUDA available: False, MPS available: True
|
18819 |
+
2025-06-11 10:59:29,956 - hardware_detector - INFO - Hardware detection completed successfully
|
18820 |
+
2025-06-11 10:59:29,956 - hardware_detector - DEBUG - Detected specs: {'platform': 'Darwin', 'architecture': 'arm64', 'cpu_count': 16, 'python_version': '3.11.11', 'gpu_info': None, 'cuda_available': False, 'mps_available': True, 'torch_version': '2.7.0'}
|
18821 |
+
2025-06-11 11:02:06,705 - __main__ - INFO - Initializing GradioAutodiffusers
|
18822 |
+
2025-06-11 11:02:06,705 - __main__ - DEBUG - API key found, length: 39
|
18823 |
+
2025-06-11 11:02:06,705 - auto_diffusers - INFO - Initializing AutoDiffusersGenerator
|
18824 |
+
2025-06-11 11:02:06,705 - auto_diffusers - DEBUG - API key length: 39
|
18825 |
+
2025-06-11 11:02:06,705 - auto_diffusers - WARNING - Tool calling dependencies not available, running without tools
|
18826 |
+
2025-06-11 11:02:06,705 - hardware_detector - INFO - Initializing HardwareDetector
|
18827 |
+
2025-06-11 11:02:06,705 - hardware_detector - DEBUG - Starting system hardware detection
|
18828 |
+
2025-06-11 11:02:06,705 - hardware_detector - DEBUG - Platform: Darwin, Architecture: arm64
|
18829 |
+
2025-06-11 11:02:06,705 - hardware_detector - DEBUG - CPU cores: 16, Python: 3.11.11
|
18830 |
+
2025-06-11 11:02:06,705 - hardware_detector - DEBUG - Attempting GPU detection via nvidia-smi
|
18831 |
+
2025-06-11 11:02:06,708 - hardware_detector - DEBUG - nvidia-smi not found, no NVIDIA GPU detected
|
18832 |
+
2025-06-11 11:02:06,708 - hardware_detector - DEBUG - Checking PyTorch availability
|
18833 |
+
2025-06-11 11:02:07,162 - hardware_detector - INFO - PyTorch 2.7.0 detected
|
18834 |
+
2025-06-11 11:02:07,162 - hardware_detector - DEBUG - CUDA available: False, MPS available: True
|
18835 |
+
2025-06-11 11:02:07,162 - hardware_detector - INFO - Hardware detection completed successfully
|
18836 |
+
2025-06-11 11:02:07,162 - hardware_detector - DEBUG - Detected specs: {'platform': 'Darwin', 'architecture': 'arm64', 'cpu_count': 16, 'python_version': '3.11.11', 'gpu_info': None, 'cuda_available': False, 'mps_available': True, 'torch_version': '2.7.0'}
|
18837 |
+
2025-06-11 11:02:07,162 - auto_diffusers - INFO - Hardware detector initialized successfully
|
18838 |
+
2025-06-11 11:02:07,162 - __main__ - INFO - AutoDiffusersGenerator initialized successfully
|
18839 |
+
2025-06-11 11:02:07,162 - simple_memory_calculator - INFO - Initializing SimpleMemoryCalculator
|
18840 |
+
2025-06-11 11:02:07,162 - simple_memory_calculator - DEBUG - HuggingFace API initialized
|
18841 |
+
2025-06-11 11:02:07,162 - simple_memory_calculator - DEBUG - Known models in database: 4
|
18842 |
+
2025-06-11 11:02:07,162 - __main__ - INFO - SimpleMemoryCalculator initialized successfully
|
18843 |
+
2025-06-11 11:02:07,162 - __main__ - DEBUG - Default model settings: gemini-2.5-flash-preview-05-20, temp=0.7
|
18844 |
+
2025-06-11 11:02:07,164 - asyncio - DEBUG - Using selector: KqueueSelector
|
18845 |
+
2025-06-11 11:02:07,177 - httpcore.connection - DEBUG - connect_tcp.started host='api.gradio.app' port=443 local_address=None timeout=3 socket_options=None
|
18846 |
+
2025-06-11 11:02:07,184 - urllib3.connectionpool - DEBUG - Starting new HTTPS connection (1): huggingface.co:443
|
18847 |
+
2025-06-11 11:02:07,259 - asyncio - DEBUG - Using selector: KqueueSelector
|
18848 |
+
2025-06-11 11:02:07,291 - httpcore.connection - DEBUG - connect_tcp.started host='localhost' port=7860 local_address=None timeout=None socket_options=None
|
18849 |
+
2025-06-11 11:02:07,292 - httpcore.connection - DEBUG - connect_tcp.complete return_value=<httpcore._backends.sync.SyncStream object at 0x12cca70d0>
|
18850 |
+
2025-06-11 11:02:07,292 - httpcore.http11 - DEBUG - send_request_headers.started request=<Request [b'GET']>
|
18851 |
+
2025-06-11 11:02:07,292 - httpcore.http11 - DEBUG - send_request_headers.complete
|
18852 |
+
2025-06-11 11:02:07,292 - httpcore.http11 - DEBUG - send_request_body.started request=<Request [b'GET']>
|
18853 |
+
2025-06-11 11:02:07,292 - httpcore.http11 - DEBUG - send_request_body.complete
|
18854 |
+
2025-06-11 11:02:07,292 - httpcore.http11 - DEBUG - receive_response_headers.started request=<Request [b'GET']>
|
18855 |
+
2025-06-11 11:02:07,293 - httpcore.http11 - DEBUG - receive_response_headers.complete return_value=(b'HTTP/1.1', 200, b'OK', [(b'date', b'Wed, 11 Jun 2025 02:02:07 GMT'), (b'server', b'uvicorn'), (b'content-length', b'4'), (b'content-type', b'application/json')])
|
18856 |
+
2025-06-11 11:02:07,293 - httpx - INFO - HTTP Request: GET http://localhost:7860/gradio_api/startup-events "HTTP/1.1 200 OK"
|
18857 |
+
2025-06-11 11:02:07,293 - httpcore.http11 - DEBUG - receive_response_body.started request=<Request [b'GET']>
|
18858 |
+
2025-06-11 11:02:07,293 - httpcore.http11 - DEBUG - receive_response_body.complete
|
18859 |
+
2025-06-11 11:02:07,293 - httpcore.http11 - DEBUG - response_closed.started
|
18860 |
+
2025-06-11 11:02:07,293 - httpcore.http11 - DEBUG - response_closed.complete
|
18861 |
+
2025-06-11 11:02:07,293 - httpcore.connection - DEBUG - close.started
|
18862 |
+
2025-06-11 11:02:07,293 - httpcore.connection - DEBUG - close.complete
|
18863 |
+
2025-06-11 11:02:07,293 - httpcore.connection - DEBUG - connect_tcp.started host='localhost' port=7860 local_address=None timeout=3 socket_options=None
|
18864 |
+
2025-06-11 11:02:07,294 - httpcore.connection - DEBUG - connect_tcp.complete return_value=<httpcore._backends.sync.SyncStream object at 0x12f9793d0>
|
18865 |
+
2025-06-11 11:02:07,294 - httpcore.http11 - DEBUG - send_request_headers.started request=<Request [b'HEAD']>
|
18866 |
+
2025-06-11 11:02:07,294 - httpcore.http11 - DEBUG - send_request_headers.complete
|
18867 |
+
2025-06-11 11:02:07,294 - httpcore.http11 - DEBUG - send_request_body.started request=<Request [b'HEAD']>
|
18868 |
+
2025-06-11 11:02:07,294 - httpcore.http11 - DEBUG - send_request_body.complete
|
18869 |
+
2025-06-11 11:02:07,294 - httpcore.http11 - DEBUG - receive_response_headers.started request=<Request [b'HEAD']>
|
18870 |
+
2025-06-11 11:02:07,301 - httpcore.http11 - DEBUG - receive_response_headers.complete return_value=(b'HTTP/1.1', 200, b'OK', [(b'date', b'Wed, 11 Jun 2025 02:02:07 GMT'), (b'server', b'uvicorn'), (b'content-length', b'145791'), (b'content-type', b'text/html; charset=utf-8')])
|
18871 |
+
2025-06-11 11:02:07,301 - httpx - INFO - HTTP Request: HEAD http://localhost:7860/ "HTTP/1.1 200 OK"
|
18872 |
+
2025-06-11 11:02:07,302 - httpcore.http11 - DEBUG - receive_response_body.started request=<Request [b'HEAD']>
|
18873 |
+
2025-06-11 11:02:07,302 - httpcore.http11 - DEBUG - receive_response_body.complete
|
18874 |
+
2025-06-11 11:02:07,302 - httpcore.http11 - DEBUG - response_closed.started
|
18875 |
+
2025-06-11 11:02:07,302 - httpcore.http11 - DEBUG - response_closed.complete
|
18876 |
+
2025-06-11 11:02:07,302 - httpcore.connection - DEBUG - close.started
|
18877 |
+
2025-06-11 11:02:07,302 - httpcore.connection - DEBUG - close.complete
|
18878 |
+
2025-06-11 11:02:07,313 - httpcore.connection - DEBUG - connect_tcp.started host='api.gradio.app' port=443 local_address=None timeout=30 socket_options=None
|
18879 |
+
2025-06-11 11:02:07,341 - httpcore.connection - DEBUG - connect_tcp.complete return_value=<httpcore._backends.sync.SyncStream object at 0x12d6ce890>
|
18880 |
+
2025-06-11 11:02:07,341 - httpcore.connection - DEBUG - start_tls.started ssl_context=<ssl.SSLContext object at 0x12d5aa8d0> server_hostname='api.gradio.app' timeout=3
|
18881 |
+
2025-06-11 11:02:07,459 - urllib3.connectionpool - DEBUG - https://huggingface.co:443 "HEAD /api/telemetry/gradio/initiated HTTP/1.1" 200 0
|
18882 |
+
2025-06-11 11:02:07,459 - httpcore.connection - DEBUG - connect_tcp.complete return_value=<httpcore._backends.sync.SyncStream object at 0x12cf2e010>
|
18883 |
+
2025-06-11 11:02:07,459 - httpcore.connection - DEBUG - start_tls.started ssl_context=<ssl.SSLContext object at 0x12f96cb00> server_hostname='api.gradio.app' timeout=30
|
18884 |
+
2025-06-11 11:02:07,614 - httpcore.connection - DEBUG - start_tls.complete return_value=<httpcore._backends.sync.SyncStream object at 0x12f877590>
|
18885 |
+
2025-06-11 11:02:07,615 - httpcore.http11 - DEBUG - send_request_headers.started request=<Request [b'GET']>
|
18886 |
+
2025-06-11 11:02:07,615 - httpcore.http11 - DEBUG - send_request_headers.complete
|
18887 |
+
2025-06-11 11:02:07,616 - httpcore.http11 - DEBUG - send_request_body.started request=<Request [b'GET']>
|
18888 |
+
2025-06-11 11:02:07,616 - httpcore.http11 - DEBUG - send_request_body.complete
|
18889 |
+
2025-06-11 11:02:07,616 - httpcore.http11 - DEBUG - receive_response_headers.started request=<Request [b'GET']>
|
18890 |
+
2025-06-11 11:02:07,752 - httpcore.http11 - DEBUG - receive_response_headers.complete return_value=(b'HTTP/1.1', 200, b'OK', [(b'Date', b'Wed, 11 Jun 2025 02:02:07 GMT'), (b'Content-Type', b'application/json'), (b'Content-Length', b'21'), (b'Connection', b'keep-alive'), (b'Server', b'nginx/1.18.0'), (b'Access-Control-Allow-Origin', b'*')])
|
18891 |
+
2025-06-11 11:02:07,753 - httpx - INFO - HTTP Request: GET https://api.gradio.app/pkg-version "HTTP/1.1 200 OK"
|
18892 |
+
2025-06-11 11:02:07,753 - httpcore.http11 - DEBUG - receive_response_body.started request=<Request [b'GET']>
|
18893 |
+
2025-06-11 11:02:07,753 - httpcore.http11 - DEBUG - receive_response_body.complete
|
18894 |
+
2025-06-11 11:02:07,753 - httpcore.http11 - DEBUG - response_closed.started
|
18895 |
+
2025-06-11 11:02:07,753 - httpcore.http11 - DEBUG - response_closed.complete
|
18896 |
+
2025-06-11 11:02:07,753 - httpcore.connection - DEBUG - close.started
|
18897 |
+
2025-06-11 11:02:07,753 - httpcore.connection - DEBUG - close.complete
|
18898 |
+
2025-06-11 11:02:07,761 - httpcore.connection - DEBUG - start_tls.complete return_value=<httpcore._backends.sync.SyncStream object at 0x11b064750>
|
18899 |
+
2025-06-11 11:02:07,761 - httpcore.http11 - DEBUG - send_request_headers.started request=<Request [b'GET']>
|
18900 |
+
2025-06-11 11:02:07,761 - httpcore.http11 - DEBUG - send_request_headers.complete
|
18901 |
+
2025-06-11 11:02:07,761 - httpcore.http11 - DEBUG - send_request_body.started request=<Request [b'GET']>
|
18902 |
+
2025-06-11 11:02:07,761 - httpcore.http11 - DEBUG - send_request_body.complete
|
18903 |
+
2025-06-11 11:02:07,761 - httpcore.http11 - DEBUG - receive_response_headers.started request=<Request [b'GET']>
|
18904 |
+
2025-06-11 11:02:07,910 - httpcore.http11 - DEBUG - receive_response_headers.complete return_value=(b'HTTP/1.1', 200, b'OK', [(b'Date', b'Wed, 11 Jun 2025 02:02:07 GMT'), (b'Content-Type', b'text/html; charset=utf-8'), (b'Transfer-Encoding', b'chunked'), (b'Connection', b'keep-alive'), (b'Server', b'nginx/1.18.0'), (b'ContentType', b'application/json'), (b'Access-Control-Allow-Origin', b'*'), (b'Content-Encoding', b'gzip')])
|
18905 |
+
2025-06-11 11:02:07,910 - httpx - INFO - HTTP Request: GET https://api.gradio.app/v3/tunnel-request "HTTP/1.1 200 OK"
|
18906 |
+
2025-06-11 11:02:07,910 - httpcore.http11 - DEBUG - receive_response_body.started request=<Request [b'GET']>
|
18907 |
+
2025-06-11 11:02:07,910 - httpcore.http11 - DEBUG - receive_response_body.complete
|
18908 |
+
2025-06-11 11:02:07,910 - httpcore.http11 - DEBUG - response_closed.started
|
18909 |
+
2025-06-11 11:02:07,910 - httpcore.http11 - DEBUG - response_closed.complete
|
18910 |
+
2025-06-11 11:02:07,910 - httpcore.connection - DEBUG - close.started
|
18911 |
+
2025-06-11 11:02:07,910 - httpcore.connection - DEBUG - close.complete
|
18912 |
+
2025-06-11 11:02:08,533 - urllib3.connectionpool - DEBUG - Starting new HTTPS connection (1): huggingface.co:443
|
18913 |
+
2025-06-11 11:02:08,757 - urllib3.connectionpool - DEBUG - https://huggingface.co:443 "HEAD /api/telemetry/gradio/launched HTTP/1.1" 200 0
|
18914 |
+
2025-06-11 11:02:13,522 - simple_memory_calculator - INFO - Getting memory requirements for model: black-forest-labs/FLUX.1-schnell
|
18915 |
+
2025-06-11 11:02:13,522 - simple_memory_calculator - INFO - Using known memory data for black-forest-labs/FLUX.1-schnell
|
18916 |
+
2025-06-11 11:02:13,522 - hardware_detector - INFO - Initializing HardwareDetector
|
18917 |
+
2025-06-11 11:02:13,522 - simple_memory_calculator - DEBUG - Known data: {'params_billions': 12.0, 'fp16_gb': 24.0, 'inference_fp16_gb': 36.0}
|
18918 |
+
2025-06-11 11:02:13,522 - hardware_detector - DEBUG - Starting system hardware detection
|
18919 |
+
2025-06-11 11:02:13,522 - simple_memory_calculator - INFO - Generating memory recommendations for black-forest-labs/FLUX.1-schnell with 8.0GB VRAM
|
18920 |
+
2025-06-11 11:02:13,522 - hardware_detector - DEBUG - Platform: Darwin, Architecture: arm64
|
18921 |
+
2025-06-11 11:02:13,522 - simple_memory_calculator - INFO - Getting memory requirements for model: black-forest-labs/FLUX.1-schnell
|
18922 |
+
2025-06-11 11:02:13,522 - hardware_detector - DEBUG - CPU cores: 16, Python: 3.11.11
|
18923 |
+
2025-06-11 11:02:13,522 - simple_memory_calculator - DEBUG - Using cached memory data for black-forest-labs/FLUX.1-schnell
|
18924 |
+
2025-06-11 11:02:13,522 - hardware_detector - DEBUG - Attempting GPU detection via nvidia-smi
|
18925 |
+
2025-06-11 11:02:13,522 - simple_memory_calculator - DEBUG - Model memory: 24.0GB, Inference memory: 36.0GB
|
18926 |
+
2025-06-11 11:02:13,523 - simple_memory_calculator - INFO - Getting memory requirements for model: black-forest-labs/FLUX.1-schnell
|
18927 |
+
2025-06-11 11:02:13,523 - simple_memory_calculator - DEBUG - Using cached memory data for black-forest-labs/FLUX.1-schnell
|
18928 |
+
2025-06-11 11:02:13,528 - hardware_detector - DEBUG - nvidia-smi not found, no NVIDIA GPU detected
|
18929 |
+
2025-06-11 11:02:13,529 - hardware_detector - DEBUG - Checking PyTorch availability
|
18930 |
+
2025-06-11 11:02:13,529 - hardware_detector - INFO - PyTorch 2.7.0 detected
|
18931 |
+
2025-06-11 11:02:13,529 - hardware_detector - DEBUG - CUDA available: False, MPS available: True
|
18932 |
+
2025-06-11 11:02:13,529 - hardware_detector - INFO - Hardware detection completed successfully
|
18933 |
+
2025-06-11 11:02:13,529 - hardware_detector - DEBUG - Detected specs: {'platform': 'Darwin', 'architecture': 'arm64', 'cpu_count': 16, 'python_version': '3.11.11', 'gpu_info': None, 'cuda_available': False, 'mps_available': True, 'torch_version': '2.7.0'}
|
18934 |
+
2025-06-11 11:02:13,529 - __main__ - INFO - 🔍 DEBUG: Full detected specs: {'platform': 'Darwin', 'architecture': 'arm64', 'cpu_count': 16, 'python_version': '3.11.11', 'gpu_info': None, 'cuda_available': False, 'mps_available': True, 'torch_version': '2.7.0'}
|
18935 |
+
2025-06-11 11:02:13,529 - __main__ - INFO - 🔍 DEBUG: Platform: Darwin
|
18936 |
+
2025-06-11 11:02:13,529 - __main__ - INFO - 🔍 DEBUG: Architecture: arm64
|
18937 |
+
2025-06-11 11:02:13,529 - __main__ - INFO - 🔍 DEBUG: MPS Available: True
|
18938 |
+
2025-06-11 11:02:13,529 - __main__ - INFO - 🔍 DEBUG: Is Apple Silicon: True
|
18939 |
+
2025-06-11 11:02:13,529 - __main__ - INFO - 🔍 DEBUG: GPU Info: None
|
18940 |
+
2025-06-11 11:02:13,529 - __main__ - INFO - 🔍 DEBUG: Detected RAM (GB): 16
|
18941 |
+
2025-06-11 11:02:13,783 - __main__ - INFO - 🔍 DEBUG: PSUtil detected 128GB unified memory
|
18942 |
+
2025-06-11 11:02:13,878 - __main__ - INFO - 🔍 DEBUG: System profiler output: hardware:
|
18943 |
+
|
18944 |
+
hardware overview:
|
18945 |
+
|
18946 |
+
model name: macbook pro
|
18947 |
+
model identifier: mac16,5
|
18948 |
+
model number: z1fs0002mkh/a
|
18949 |
+
chip: apple m4 max
|
18950 |
+
total number of cores: 16 (12 performance and 4 efficiency)
|
18951 |
+
memory: 128 gb
|
18952 |
+
system firmware version: 11881.61.3
|
18953 |
+
os loader version: 11881.61.3
|
18954 |
+
serial number (system): kkvqcmx4dt
|
18955 |
+
hardware uuid: 92bfa67e-d816-5b1d-99bb-f9a2824999ba
|
18956 |
+
provisioning udid: 00006041-001269200280801c
|
18957 |
+
activation lock status: enabled
|
18958 |
+
|
18959 |
+
|
18960 |
+
2025-06-11 11:02:13,878 - __main__ - INFO - 🔍 DEBUG: Detected Apple chip: M4 Max
|
18961 |
+
2025-06-11 11:02:13,878 - __main__ - INFO - 🔍 DEBUG: Final values - Platform: Apple Silicon, GPU Vendor: Apple Silicon, GPU Name: M4 Max, Unified RAM: 128GB
|
18962 |
+
2025-06-11 11:04:13,169 - __main__ - INFO - Initializing GradioAutodiffusers
|
18963 |
+
2025-06-11 11:04:13,169 - __main__ - DEBUG - API key found, length: 39
|
18964 |
+
2025-06-11 11:04:13,169 - auto_diffusers - INFO - Initializing AutoDiffusersGenerator
|
18965 |
+
2025-06-11 11:04:13,169 - auto_diffusers - DEBUG - API key length: 39
|
18966 |
+
2025-06-11 11:04:13,169 - auto_diffusers - WARNING - Tool calling dependencies not available, running without tools
|
18967 |
+
2025-06-11 11:04:13,169 - hardware_detector - INFO - Initializing HardwareDetector
|
18968 |
+
2025-06-11 11:04:13,169 - hardware_detector - DEBUG - Starting system hardware detection
|
18969 |
+
2025-06-11 11:04:13,169 - hardware_detector - DEBUG - Platform: Darwin, Architecture: arm64
|
18970 |
+
2025-06-11 11:04:13,169 - hardware_detector - DEBUG - CPU cores: 16, Python: 3.11.11
|
18971 |
+
2025-06-11 11:04:13,169 - hardware_detector - DEBUG - Attempting GPU detection via nvidia-smi
|
18972 |
+
2025-06-11 11:04:13,173 - hardware_detector - DEBUG - nvidia-smi not found, no NVIDIA GPU detected
|
18973 |
+
2025-06-11 11:04:13,173 - hardware_detector - DEBUG - Checking PyTorch availability
|
18974 |
+
2025-06-11 11:04:13,635 - hardware_detector - INFO - PyTorch 2.7.0 detected
|
18975 |
+
2025-06-11 11:04:13,635 - hardware_detector - DEBUG - CUDA available: False, MPS available: True
|
18976 |
+
2025-06-11 11:04:13,636 - hardware_detector - INFO - Hardware detection completed successfully
|
18977 |
+
2025-06-11 11:04:13,636 - hardware_detector - DEBUG - Detected specs: {'platform': 'Darwin', 'architecture': 'arm64', 'cpu_count': 16, 'python_version': '3.11.11', 'gpu_info': None, 'cuda_available': False, 'mps_available': True, 'torch_version': '2.7.0'}
|
18978 |
+
2025-06-11 11:04:13,636 - auto_diffusers - INFO - Hardware detector initialized successfully
|
18979 |
+
2025-06-11 11:04:13,636 - __main__ - INFO - AutoDiffusersGenerator initialized successfully
|
18980 |
+
2025-06-11 11:04:13,636 - simple_memory_calculator - INFO - Initializing SimpleMemoryCalculator
|
18981 |
+
2025-06-11 11:04:13,636 - simple_memory_calculator - DEBUG - HuggingFace API initialized
|
18982 |
+
2025-06-11 11:04:13,636 - simple_memory_calculator - DEBUG - Known models in database: 4
|
18983 |
+
2025-06-11 11:04:13,636 - __main__ - INFO - SimpleMemoryCalculator initialized successfully
|
18984 |
+
2025-06-11 11:04:13,636 - __main__ - DEBUG - Default model settings: gemini-2.5-flash-preview-05-20, temp=0.7
|
18985 |
+
2025-06-11 11:04:13,638 - asyncio - DEBUG - Using selector: KqueueSelector
|
18986 |
+
2025-06-11 11:04:13,652 - httpcore.connection - DEBUG - connect_tcp.started host='api.gradio.app' port=443 local_address=None timeout=3 socket_options=None
|
18987 |
+
2025-06-11 11:04:13,657 - urllib3.connectionpool - DEBUG - Starting new HTTPS connection (1): huggingface.co:443
|
18988 |
+
2025-06-11 11:04:13,732 - asyncio - DEBUG - Using selector: KqueueSelector
|
18989 |
+
2025-06-11 11:04:13,765 - httpcore.connection - DEBUG - connect_tcp.started host='localhost' port=7860 local_address=None timeout=None socket_options=None
|
18990 |
+
2025-06-11 11:04:13,766 - httpcore.connection - DEBUG - connect_tcp.complete return_value=<httpcore._backends.sync.SyncStream object at 0x11f95aed0>
|
18991 |
+
2025-06-11 11:04:13,766 - httpcore.http11 - DEBUG - send_request_headers.started request=<Request [b'GET']>
|
18992 |
+
2025-06-11 11:04:13,766 - httpcore.http11 - DEBUG - send_request_headers.complete
|
18993 |
+
2025-06-11 11:04:13,766 - httpcore.http11 - DEBUG - send_request_body.started request=<Request [b'GET']>
|
18994 |
+
2025-06-11 11:04:13,766 - httpcore.http11 - DEBUG - send_request_body.complete
|
18995 |
+
2025-06-11 11:04:13,766 - httpcore.http11 - DEBUG - receive_response_headers.started request=<Request [b'GET']>
|
18996 |
+
2025-06-11 11:04:13,767 - httpcore.http11 - DEBUG - receive_response_headers.complete return_value=(b'HTTP/1.1', 200, b'OK', [(b'date', b'Wed, 11 Jun 2025 02:04:13 GMT'), (b'server', b'uvicorn'), (b'content-length', b'4'), (b'content-type', b'application/json')])
|
18997 |
+
2025-06-11 11:04:13,767 - httpx - INFO - HTTP Request: GET http://localhost:7860/gradio_api/startup-events "HTTP/1.1 200 OK"
|
18998 |
+
2025-06-11 11:04:13,767 - httpcore.http11 - DEBUG - receive_response_body.started request=<Request [b'GET']>
|
18999 |
+
2025-06-11 11:04:13,767 - httpcore.http11 - DEBUG - receive_response_body.complete
|
19000 |
+
2025-06-11 11:04:13,767 - httpcore.http11 - DEBUG - response_closed.started
|
19001 |
+
2025-06-11 11:04:13,767 - httpcore.http11 - DEBUG - response_closed.complete
|
19002 |
+
2025-06-11 11:04:13,767 - httpcore.connection - DEBUG - close.started
|
19003 |
+
2025-06-11 11:04:13,767 - httpcore.connection - DEBUG - close.complete
|
19004 |
+
2025-06-11 11:04:13,767 - httpcore.connection - DEBUG - connect_tcp.started host='localhost' port=7860 local_address=None timeout=3 socket_options=None
|
19005 |
+
2025-06-11 11:04:13,768 - httpcore.connection - DEBUG - connect_tcp.complete return_value=<httpcore._backends.sync.SyncStream object at 0x11f9707d0>
|
19006 |
+
2025-06-11 11:04:13,768 - httpcore.http11 - DEBUG - send_request_headers.started request=<Request [b'HEAD']>
|
19007 |
+
2025-06-11 11:04:13,768 - httpcore.http11 - DEBUG - send_request_headers.complete
|
19008 |
+
2025-06-11 11:04:13,768 - httpcore.http11 - DEBUG - send_request_body.started request=<Request [b'HEAD']>
|
19009 |
+
2025-06-11 11:04:13,768 - httpcore.http11 - DEBUG - send_request_body.complete
|
19010 |
+
2025-06-11 11:04:13,768 - httpcore.http11 - DEBUG - receive_response_headers.started request=<Request [b'HEAD']>
|
19011 |
+
2025-06-11 11:04:13,775 - httpcore.http11 - DEBUG - receive_response_headers.complete return_value=(b'HTTP/1.1', 200, b'OK', [(b'date', b'Wed, 11 Jun 2025 02:04:13 GMT'), (b'server', b'uvicorn'), (b'content-length', b'144016'), (b'content-type', b'text/html; charset=utf-8')])
|
19012 |
+
2025-06-11 11:04:13,775 - httpx - INFO - HTTP Request: HEAD http://localhost:7860/ "HTTP/1.1 200 OK"
|
19013 |
+
2025-06-11 11:04:13,775 - httpcore.http11 - DEBUG - receive_response_body.started request=<Request [b'HEAD']>
|
19014 |
+
2025-06-11 11:04:13,775 - httpcore.http11 - DEBUG - receive_response_body.complete
|
19015 |
+
2025-06-11 11:04:13,775 - httpcore.http11 - DEBUG - response_closed.started
|
19016 |
+
2025-06-11 11:04:13,775 - httpcore.http11 - DEBUG - response_closed.complete
|
19017 |
+
2025-06-11 11:04:13,775 - httpcore.connection - DEBUG - close.started
|
19018 |
+
2025-06-11 11:04:13,775 - httpcore.connection - DEBUG - close.complete
|
19019 |
+
2025-06-11 11:04:13,787 - httpcore.connection - DEBUG - connect_tcp.started host='api.gradio.app' port=443 local_address=None timeout=30 socket_options=None
|
19020 |
+
2025-06-11 11:04:13,828 - httpcore.connection - DEBUG - connect_tcp.complete return_value=<httpcore._backends.sync.SyncStream object at 0x11cd75e50>
|
19021 |
+
2025-06-11 11:04:13,828 - httpcore.connection - DEBUG - start_tls.started ssl_context=<ssl.SSLContext object at 0x109d35370> server_hostname='api.gradio.app' timeout=3
|
19022 |
+
2025-06-11 11:04:13,937 - httpcore.connection - DEBUG - connect_tcp.complete return_value=<httpcore._backends.sync.SyncStream object at 0x11f9d0f90>
|
19023 |
+
2025-06-11 11:04:13,937 - httpcore.connection - DEBUG - start_tls.started ssl_context=<ssl.SSLContext object at 0x11f960b00> server_hostname='api.gradio.app' timeout=30
|
19024 |
+
2025-06-11 11:04:13,939 - urllib3.connectionpool - DEBUG - https://huggingface.co:443 "HEAD /api/telemetry/gradio/initiated HTTP/1.1" 200 0
|
19025 |
+
2025-06-11 11:04:14,124 - httpcore.connection - DEBUG - start_tls.complete return_value=<httpcore._backends.sync.SyncStream object at 0x11b820d10>
|
19026 |
+
2025-06-11 11:04:14,124 - httpcore.http11 - DEBUG - send_request_headers.started request=<Request [b'GET']>
|
19027 |
+
2025-06-11 11:04:14,125 - httpcore.http11 - DEBUG - send_request_headers.complete
|
19028 |
+
2025-06-11 11:04:14,125 - httpcore.http11 - DEBUG - send_request_body.started request=<Request [b'GET']>
|
19029 |
+
2025-06-11 11:04:14,125 - httpcore.http11 - DEBUG - send_request_body.complete
|
19030 |
+
2025-06-11 11:04:14,126 - httpcore.http11 - DEBUG - receive_response_headers.started request=<Request [b'GET']>
|
19031 |
+
2025-06-11 11:04:14,240 - httpcore.connection - DEBUG - start_tls.complete return_value=<httpcore._backends.sync.SyncStream object at 0x119caf6d0>
|
19032 |
+
2025-06-11 11:04:14,241 - httpcore.http11 - DEBUG - send_request_headers.started request=<Request [b'GET']>
|
19033 |
+
2025-06-11 11:04:14,241 - httpcore.http11 - DEBUG - send_request_headers.complete
|
19034 |
+
2025-06-11 11:04:14,241 - httpcore.http11 - DEBUG - send_request_body.started request=<Request [b'GET']>
|
19035 |
+
2025-06-11 11:04:14,241 - httpcore.http11 - DEBUG - send_request_body.complete
|
19036 |
+
2025-06-11 11:04:14,241 - httpcore.http11 - DEBUG - receive_response_headers.started request=<Request [b'GET']>
|
19037 |
+
2025-06-11 11:04:14,272 - httpcore.http11 - DEBUG - receive_response_headers.complete return_value=(b'HTTP/1.1', 200, b'OK', [(b'Date', b'Wed, 11 Jun 2025 02:04:14 GMT'), (b'Content-Type', b'application/json'), (b'Content-Length', b'21'), (b'Connection', b'keep-alive'), (b'Server', b'nginx/1.18.0'), (b'Access-Control-Allow-Origin', b'*')])
|
19038 |
+
2025-06-11 11:04:14,274 - httpx - INFO - HTTP Request: GET https://api.gradio.app/pkg-version "HTTP/1.1 200 OK"
|
19039 |
+
2025-06-11 11:04:14,275 - httpcore.http11 - DEBUG - receive_response_body.started request=<Request [b'GET']>
|
19040 |
+
2025-06-11 11:04:14,275 - httpcore.http11 - DEBUG - receive_response_body.complete
|
19041 |
+
2025-06-11 11:04:14,275 - httpcore.http11 - DEBUG - response_closed.started
|
19042 |
+
2025-06-11 11:04:14,275 - httpcore.http11 - DEBUG - response_closed.complete
|
19043 |
+
2025-06-11 11:04:14,275 - httpcore.connection - DEBUG - close.started
|
19044 |
+
2025-06-11 11:04:14,276 - httpcore.connection - DEBUG - close.complete
|
19045 |
+
2025-06-11 11:04:14,391 - httpcore.http11 - DEBUG - receive_response_headers.complete return_value=(b'HTTP/1.1', 200, b'OK', [(b'Date', b'Wed, 11 Jun 2025 02:04:14 GMT'), (b'Content-Type', b'text/html; charset=utf-8'), (b'Transfer-Encoding', b'chunked'), (b'Connection', b'keep-alive'), (b'Server', b'nginx/1.18.0'), (b'ContentType', b'application/json'), (b'Access-Control-Allow-Origin', b'*'), (b'Content-Encoding', b'gzip')])
|
19046 |
+
2025-06-11 11:04:14,391 - httpx - INFO - HTTP Request: GET https://api.gradio.app/v3/tunnel-request "HTTP/1.1 200 OK"
|
19047 |
+
2025-06-11 11:04:14,392 - httpcore.http11 - DEBUG - receive_response_body.started request=<Request [b'GET']>
|
19048 |
+
2025-06-11 11:04:14,392 - httpcore.http11 - DEBUG - receive_response_body.complete
|
19049 |
+
2025-06-11 11:04:14,392 - httpcore.http11 - DEBUG - response_closed.started
|
19050 |
+
2025-06-11 11:04:14,392 - httpcore.http11 - DEBUG - response_closed.complete
|
19051 |
+
2025-06-11 11:04:14,392 - httpcore.connection - DEBUG - close.started
|
19052 |
+
2025-06-11 11:04:14,392 - httpcore.connection - DEBUG - close.complete
|
19053 |
+
2025-06-11 11:04:14,994 - urllib3.connectionpool - DEBUG - Starting new HTTPS connection (1): huggingface.co:443
|
19054 |
+
2025-06-11 11:04:15,564 - urllib3.connectionpool - DEBUG - https://huggingface.co:443 "HEAD /api/telemetry/gradio/launched HTTP/1.1" 200 0
|
19055 |
+
2025-06-11 11:05:42,002 - simple_memory_calculator - INFO - Getting memory requirements for model: black-forest-labs/FLUX.1-schnell
|
19056 |
+
2025-06-11 11:05:42,003 - simple_memory_calculator - INFO - Using known memory data for black-forest-labs/FLUX.1-schnell
|
19057 |
+
2025-06-11 11:05:42,003 - simple_memory_calculator - DEBUG - Known data: {'params_billions': 12.0, 'fp16_gb': 24.0, 'inference_fp16_gb': 36.0}
|
19058 |
+
2025-06-11 11:05:42,003 - simple_memory_calculator - INFO - Generating memory recommendations for black-forest-labs/FLUX.1-schnell with 8.0GB VRAM
|
19059 |
+
2025-06-11 11:05:42,003 - simple_memory_calculator - INFO - Getting memory requirements for model: black-forest-labs/FLUX.1-schnell
|
19060 |
+
2025-06-11 11:05:42,003 - simple_memory_calculator - DEBUG - Using cached memory data for black-forest-labs/FLUX.1-schnell
|
19061 |
+
2025-06-11 11:05:42,003 - simple_memory_calculator - DEBUG - Model memory: 24.0GB, Inference memory: 36.0GB
|
19062 |
+
2025-06-11 11:05:42,003 - simple_memory_calculator - INFO - Getting memory requirements for model: black-forest-labs/FLUX.1-schnell
|
19063 |
+
2025-06-11 11:05:42,004 - simple_memory_calculator - DEBUG - Using cached memory data for black-forest-labs/FLUX.1-schnell
|
19064 |
+
2025-06-11 11:07:51,377 - __main__ - INFO - Initializing GradioAutodiffusers
|
19065 |
+
2025-06-11 11:07:51,377 - __main__ - DEBUG - API key found, length: 39
|
19066 |
+
2025-06-11 11:07:51,377 - auto_diffusers - INFO - Initializing AutoDiffusersGenerator
|
19067 |
+
2025-06-11 11:07:51,377 - auto_diffusers - DEBUG - API key length: 39
|
19068 |
+
2025-06-11 11:07:51,377 - auto_diffusers - WARNING - Tool calling dependencies not available, running without tools
|
19069 |
+
2025-06-11 11:07:51,377 - hardware_detector - INFO - Initializing HardwareDetector
|
19070 |
+
2025-06-11 11:07:51,377 - hardware_detector - DEBUG - Starting system hardware detection
|
19071 |
+
2025-06-11 11:07:51,377 - hardware_detector - DEBUG - Platform: Darwin, Architecture: arm64
|
19072 |
+
2025-06-11 11:07:51,377 - hardware_detector - DEBUG - CPU cores: 16, Python: 3.11.11
|
19073 |
+
2025-06-11 11:07:51,377 - hardware_detector - DEBUG - Attempting GPU detection via nvidia-smi
|
19074 |
+
2025-06-11 11:07:51,380 - hardware_detector - DEBUG - nvidia-smi not found, no NVIDIA GPU detected
|
19075 |
+
2025-06-11 11:07:51,381 - hardware_detector - DEBUG - Checking PyTorch availability
|
19076 |
+
2025-06-11 11:07:51,843 - hardware_detector - INFO - PyTorch 2.7.0 detected
|
19077 |
+
2025-06-11 11:07:51,843 - hardware_detector - DEBUG - CUDA available: False, MPS available: True
|
19078 |
+
2025-06-11 11:07:51,843 - hardware_detector - INFO - Hardware detection completed successfully
|
19079 |
+
2025-06-11 11:07:51,843 - hardware_detector - DEBUG - Detected specs: {'platform': 'Darwin', 'architecture': 'arm64', 'cpu_count': 16, 'python_version': '3.11.11', 'gpu_info': None, 'cuda_available': False, 'mps_available': True, 'torch_version': '2.7.0'}
|
19080 |
+
2025-06-11 11:07:51,843 - auto_diffusers - INFO - Hardware detector initialized successfully
|
19081 |
+
2025-06-11 11:07:51,843 - __main__ - INFO - AutoDiffusersGenerator initialized successfully
|
19082 |
+
2025-06-11 11:07:51,843 - simple_memory_calculator - INFO - Initializing SimpleMemoryCalculator
|
19083 |
+
2025-06-11 11:07:51,843 - simple_memory_calculator - DEBUG - HuggingFace API initialized
|
19084 |
+
2025-06-11 11:07:51,843 - simple_memory_calculator - DEBUG - Known models in database: 4
|
19085 |
+
2025-06-11 11:07:51,843 - __main__ - INFO - SimpleMemoryCalculator initialized successfully
|
19086 |
+
2025-06-11 11:07:51,843 - __main__ - DEBUG - Default model settings: gemini-2.5-flash-preview-05-20, temp=0.7
|
19087 |
+
2025-06-11 11:07:51,845 - asyncio - DEBUG - Using selector: KqueueSelector
|
19088 |
+
2025-06-11 11:07:51,859 - httpcore.connection - DEBUG - connect_tcp.started host='api.gradio.app' port=443 local_address=None timeout=3 socket_options=None
|
19089 |
+
2025-06-11 11:07:51,859 - urllib3.connectionpool - DEBUG - Starting new HTTPS connection (1): huggingface.co:443
|
19090 |
+
2025-06-11 11:07:51,938 - asyncio - DEBUG - Using selector: KqueueSelector
|
19091 |
+
2025-06-11 11:07:51,968 - httpcore.connection - DEBUG - connect_tcp.started host='localhost' port=7860 local_address=None timeout=None socket_options=None
|
19092 |
+
2025-06-11 11:07:51,969 - httpcore.connection - DEBUG - connect_tcp.complete return_value=<httpcore._backends.sync.SyncStream object at 0x1266d1610>
|
19093 |
+
2025-06-11 11:07:51,969 - httpcore.http11 - DEBUG - send_request_headers.started request=<Request [b'GET']>
|
19094 |
+
2025-06-11 11:07:51,969 - httpcore.http11 - DEBUG - send_request_headers.complete
|
19095 |
+
2025-06-11 11:07:51,969 - httpcore.http11 - DEBUG - send_request_body.started request=<Request [b'GET']>
|
19096 |
+
2025-06-11 11:07:51,969 - httpcore.http11 - DEBUG - send_request_body.complete
|
19097 |
+
2025-06-11 11:07:51,969 - httpcore.http11 - DEBUG - receive_response_headers.started request=<Request [b'GET']>
|
19098 |
+
2025-06-11 11:07:51,970 - httpcore.http11 - DEBUG - receive_response_headers.complete return_value=(b'HTTP/1.1', 200, b'OK', [(b'date', b'Wed, 11 Jun 2025 02:07:51 GMT'), (b'server', b'uvicorn'), (b'content-length', b'4'), (b'content-type', b'application/json')])
|
19099 |
+
2025-06-11 11:07:51,970 - httpx - INFO - HTTP Request: GET http://localhost:7860/gradio_api/startup-events "HTTP/1.1 200 OK"
|
19100 |
+
2025-06-11 11:07:51,970 - httpcore.http11 - DEBUG - receive_response_body.started request=<Request [b'GET']>
|
19101 |
+
2025-06-11 11:07:51,970 - httpcore.http11 - DEBUG - receive_response_body.complete
|
19102 |
+
2025-06-11 11:07:51,970 - httpcore.http11 - DEBUG - response_closed.started
|
19103 |
+
2025-06-11 11:07:51,970 - httpcore.http11 - DEBUG - response_closed.complete
|
19104 |
+
2025-06-11 11:07:51,970 - httpcore.connection - DEBUG - close.started
|
19105 |
+
2025-06-11 11:07:51,970 - httpcore.connection - DEBUG - close.complete
|
19106 |
+
2025-06-11 11:07:51,971 - httpcore.connection - DEBUG - connect_tcp.started host='localhost' port=7860 local_address=None timeout=3 socket_options=None
|
19107 |
+
2025-06-11 11:07:51,971 - httpcore.connection - DEBUG - connect_tcp.complete return_value=<httpcore._backends.sync.SyncStream object at 0x127b5b750>
|
19108 |
+
2025-06-11 11:07:51,971 - httpcore.http11 - DEBUG - send_request_headers.started request=<Request [b'HEAD']>
|
19109 |
+
2025-06-11 11:07:51,971 - httpcore.http11 - DEBUG - send_request_headers.complete
|
19110 |
+
2025-06-11 11:07:51,971 - httpcore.http11 - DEBUG - send_request_body.started request=<Request [b'HEAD']>
|
19111 |
+
2025-06-11 11:07:51,971 - httpcore.http11 - DEBUG - send_request_body.complete
|
19112 |
+
2025-06-11 11:07:51,971 - httpcore.http11 - DEBUG - receive_response_headers.started request=<Request [b'HEAD']>
|
19113 |
+
2025-06-11 11:07:51,978 - httpcore.http11 - DEBUG - receive_response_headers.complete return_value=(b'HTTP/1.1', 200, b'OK', [(b'date', b'Wed, 11 Jun 2025 02:07:51 GMT'), (b'server', b'uvicorn'), (b'content-length', b'140375'), (b'content-type', b'text/html; charset=utf-8')])
|
19114 |
+
2025-06-11 11:07:51,978 - httpx - INFO - HTTP Request: HEAD http://localhost:7860/ "HTTP/1.1 200 OK"
|
19115 |
+
2025-06-11 11:07:51,978 - httpcore.http11 - DEBUG - receive_response_body.started request=<Request [b'HEAD']>
|
19116 |
+
2025-06-11 11:07:51,978 - httpcore.http11 - DEBUG - receive_response_body.complete
|
19117 |
+
2025-06-11 11:07:51,978 - httpcore.http11 - DEBUG - response_closed.started
|
19118 |
+
2025-06-11 11:07:51,978 - httpcore.http11 - DEBUG - response_closed.complete
|
19119 |
+
2025-06-11 11:07:51,978 - httpcore.connection - DEBUG - close.started
|
19120 |
+
2025-06-11 11:07:51,978 - httpcore.connection - DEBUG - close.complete
|
19121 |
+
2025-06-11 11:07:51,990 - httpcore.connection - DEBUG - connect_tcp.started host='api.gradio.app' port=443 local_address=None timeout=30 socket_options=None
|
19122 |
+
2025-06-11 11:07:52,029 - httpcore.connection - DEBUG - connect_tcp.complete return_value=<httpcore._backends.sync.SyncStream object at 0x1264bdd50>
|
19123 |
+
2025-06-11 11:07:52,029 - httpcore.connection - DEBUG - start_tls.started ssl_context=<ssl.SSLContext object at 0x1263aa840> server_hostname='api.gradio.app' timeout=3
|
19124 |
+
2025-06-11 11:07:52,122 - httpcore.connection - DEBUG - connect_tcp.complete return_value=<httpcore._backends.sync.SyncStream object at 0x127bd9110>
|
19125 |
+
2025-06-11 11:07:52,122 - httpcore.connection - DEBUG - start_tls.started ssl_context=<ssl.SSLContext object at 0x127b64b00> server_hostname='api.gradio.app' timeout=30
|
19126 |
+
2025-06-11 11:07:52,188 - urllib3.connectionpool - DEBUG - https://huggingface.co:443 "HEAD /api/telemetry/gradio/initiated HTTP/1.1" 200 0
|
19127 |
+
2025-06-11 11:07:52,325 - httpcore.connection - DEBUG - start_tls.complete return_value=<httpcore._backends.sync.SyncStream object at 0x125acf010>
|
19128 |
+
2025-06-11 11:07:52,325 - httpcore.http11 - DEBUG - send_request_headers.started request=<Request [b'GET']>
|
19129 |
+
2025-06-11 11:07:52,325 - httpcore.http11 - DEBUG - send_request_headers.complete
|
19130 |
+
2025-06-11 11:07:52,326 - httpcore.http11 - DEBUG - send_request_body.started request=<Request [b'GET']>
|
19131 |
+
2025-06-11 11:07:52,326 - httpcore.http11 - DEBUG - send_request_body.complete
|
19132 |
+
2025-06-11 11:07:52,326 - httpcore.http11 - DEBUG - receive_response_headers.started request=<Request [b'GET']>
|
19133 |
+
2025-06-11 11:07:52,388 - httpcore.connection - DEBUG - start_tls.complete return_value=<httpcore._backends.sync.SyncStream object at 0x125dc7210>
|
19134 |
+
2025-06-11 11:07:52,388 - httpcore.http11 - DEBUG - send_request_headers.started request=<Request [b'GET']>
|
19135 |
+
2025-06-11 11:07:52,389 - httpcore.http11 - DEBUG - send_request_headers.complete
|
19136 |
+
2025-06-11 11:07:52,389 - httpcore.http11 - DEBUG - send_request_body.started request=<Request [b'GET']>
|
19137 |
+
2025-06-11 11:07:52,389 - httpcore.http11 - DEBUG - send_request_body.complete
|
19138 |
+
2025-06-11 11:07:52,389 - httpcore.http11 - DEBUG - receive_response_headers.started request=<Request [b'GET']>
|
19139 |
+
2025-06-11 11:07:52,474 - httpcore.http11 - DEBUG - receive_response_headers.complete return_value=(b'HTTP/1.1', 200, b'OK', [(b'Date', b'Wed, 11 Jun 2025 02:07:52 GMT'), (b'Content-Type', b'application/json'), (b'Content-Length', b'21'), (b'Connection', b'keep-alive'), (b'Server', b'nginx/1.18.0'), (b'Access-Control-Allow-Origin', b'*')])
|
19140 |
+
2025-06-11 11:07:52,475 - httpx - INFO - HTTP Request: GET https://api.gradio.app/pkg-version "HTTP/1.1 200 OK"
|
19141 |
+
2025-06-11 11:07:52,475 - httpcore.http11 - DEBUG - receive_response_body.started request=<Request [b'GET']>
|
19142 |
+
2025-06-11 11:07:52,475 - httpcore.http11 - DEBUG - receive_response_body.complete
|
19143 |
+
2025-06-11 11:07:52,475 - httpcore.http11 - DEBUG - response_closed.started
|
19144 |
+
2025-06-11 11:07:52,475 - httpcore.http11 - DEBUG - response_closed.complete
|
19145 |
+
2025-06-11 11:07:52,475 - httpcore.connection - DEBUG - close.started
|
19146 |
+
2025-06-11 11:07:52,476 - httpcore.connection - DEBUG - close.complete
|
19147 |
+
2025-06-11 11:07:52,524 - httpcore.http11 - DEBUG - receive_response_headers.complete return_value=(b'HTTP/1.1', 200, b'OK', [(b'Date', b'Wed, 11 Jun 2025 02:07:52 GMT'), (b'Content-Type', b'text/html; charset=utf-8'), (b'Transfer-Encoding', b'chunked'), (b'Connection', b'keep-alive'), (b'Server', b'nginx/1.18.0'), (b'ContentType', b'application/json'), (b'Access-Control-Allow-Origin', b'*'), (b'Content-Encoding', b'gzip')])
|
19148 |
+
2025-06-11 11:07:52,524 - httpx - INFO - HTTP Request: GET https://api.gradio.app/v3/tunnel-request "HTTP/1.1 200 OK"
|
19149 |
+
2025-06-11 11:07:52,524 - httpcore.http11 - DEBUG - receive_response_body.started request=<Request [b'GET']>
|
19150 |
+
2025-06-11 11:07:52,525 - httpcore.http11 - DEBUG - receive_response_body.complete
|
19151 |
+
2025-06-11 11:07:52,525 - httpcore.http11 - DEBUG - response_closed.started
|
19152 |
+
2025-06-11 11:07:52,525 - httpcore.http11 - DEBUG - response_closed.complete
|
19153 |
+
2025-06-11 11:07:52,526 - httpcore.connection - DEBUG - close.started
|
19154 |
+
2025-06-11 11:07:52,526 - httpcore.connection - DEBUG - close.complete
|
19155 |
+
2025-06-11 11:07:53,161 - urllib3.connectionpool - DEBUG - Starting new HTTPS connection (1): huggingface.co:443
|
19156 |
+
2025-06-11 11:07:53,383 - urllib3.connectionpool - DEBUG - https://huggingface.co:443 "HEAD /api/telemetry/gradio/launched HTTP/1.1" 200 0
|
19157 |
+
2025-06-11 11:07:56,235 - simple_memory_calculator - INFO - Getting memory requirements for model: black-forest-labs/FLUX.1-schnell
|
19158 |
+
2025-06-11 11:07:56,235 - simple_memory_calculator - INFO - Using known memory data for black-forest-labs/FLUX.1-schnell
|
19159 |
+
2025-06-11 11:07:56,235 - simple_memory_calculator - DEBUG - Known data: {'params_billions': 12.0, 'fp16_gb': 24.0, 'inference_fp16_gb': 36.0}
|
19160 |
+
2025-06-11 11:07:56,235 - simple_memory_calculator - INFO - Generating memory recommendations for black-forest-labs/FLUX.1-schnell with 8.0GB VRAM
|
19161 |
+
2025-06-11 11:07:56,235 - simple_memory_calculator - INFO - Getting memory requirements for model: black-forest-labs/FLUX.1-schnell
|
19162 |
+
2025-06-11 11:07:56,235 - simple_memory_calculator - DEBUG - Using cached memory data for black-forest-labs/FLUX.1-schnell
|
19163 |
+
2025-06-11 11:07:56,236 - simple_memory_calculator - DEBUG - Model memory: 24.0GB, Inference memory: 36.0GB
|
19164 |
+
2025-06-11 11:07:56,236 - simple_memory_calculator - INFO - Getting memory requirements for model: black-forest-labs/FLUX.1-schnell
|
19165 |
+
2025-06-11 11:07:56,236 - simple_memory_calculator - DEBUG - Using cached memory data for black-forest-labs/FLUX.1-schnell
|
19166 |
+
2025-06-11 11:08:06,275 - __main__ - INFO - Initializing GradioAutodiffusers
|
19167 |
+
2025-06-11 11:08:06,275 - __main__ - DEBUG - API key found, length: 39
|
19168 |
+
2025-06-11 11:08:06,275 - auto_diffusers - INFO - Initializing AutoDiffusersGenerator
|
19169 |
+
2025-06-11 11:08:06,275 - auto_diffusers - DEBUG - API key length: 39
|
19170 |
+
2025-06-11 11:08:06,275 - auto_diffusers - WARNING - Tool calling dependencies not available, running without tools
|
19171 |
+
2025-06-11 11:08:06,275 - hardware_detector - INFO - Initializing HardwareDetector
|
19172 |
+
2025-06-11 11:08:06,275 - hardware_detector - DEBUG - Starting system hardware detection
|
19173 |
+
2025-06-11 11:08:06,275 - hardware_detector - DEBUG - Platform: Darwin, Architecture: arm64
|
19174 |
+
2025-06-11 11:08:06,275 - hardware_detector - DEBUG - CPU cores: 16, Python: 3.11.11
|
19175 |
+
2025-06-11 11:08:06,275 - hardware_detector - DEBUG - Attempting GPU detection via nvidia-smi
|
19176 |
+
2025-06-11 11:08:06,278 - hardware_detector - DEBUG - nvidia-smi not found, no NVIDIA GPU detected
|
19177 |
+
2025-06-11 11:08:06,278 - hardware_detector - DEBUG - Checking PyTorch availability
|
19178 |
+
2025-06-11 11:08:06,730 - hardware_detector - INFO - PyTorch 2.7.0 detected
|
19179 |
+
2025-06-11 11:08:06,730 - hardware_detector - DEBUG - CUDA available: False, MPS available: True
|
19180 |
+
2025-06-11 11:08:06,730 - hardware_detector - INFO - Hardware detection completed successfully
|
19181 |
+
2025-06-11 11:08:06,730 - hardware_detector - DEBUG - Detected specs: {'platform': 'Darwin', 'architecture': 'arm64', 'cpu_count': 16, 'python_version': '3.11.11', 'gpu_info': None, 'cuda_available': False, 'mps_available': True, 'torch_version': '2.7.0'}
|
19182 |
+
2025-06-11 11:08:06,730 - auto_diffusers - INFO - Hardware detector initialized successfully
|
19183 |
+
2025-06-11 11:08:06,730 - __main__ - INFO - AutoDiffusersGenerator initialized successfully
|
19184 |
+
2025-06-11 11:08:06,730 - simple_memory_calculator - INFO - Initializing SimpleMemoryCalculator
|
19185 |
+
2025-06-11 11:08:06,730 - simple_memory_calculator - DEBUG - HuggingFace API initialized
|
19186 |
+
2025-06-11 11:08:06,730 - simple_memory_calculator - DEBUG - Known models in database: 4
|
19187 |
+
2025-06-11 11:08:06,730 - __main__ - INFO - SimpleMemoryCalculator initialized successfully
|
19188 |
+
2025-06-11 11:08:06,730 - __main__ - DEBUG - Default model settings: gemini-2.5-flash-preview-05-20, temp=0.7
|
19189 |
+
2025-06-11 11:08:06,732 - asyncio - DEBUG - Using selector: KqueueSelector
|
19190 |
+
2025-06-11 11:08:06,745 - httpcore.connection - DEBUG - connect_tcp.started host='api.gradio.app' port=443 local_address=None timeout=3 socket_options=None
|
19191 |
+
2025-06-11 11:08:06,753 - urllib3.connectionpool - DEBUG - Starting new HTTPS connection (1): huggingface.co:443
|
19192 |
+
2025-06-11 11:08:06,825 - asyncio - DEBUG - Using selector: KqueueSelector
|
19193 |
+
2025-06-11 11:08:06,858 - httpcore.connection - DEBUG - connect_tcp.started host='localhost' port=7860 local_address=None timeout=None socket_options=None
|
19194 |
+
2025-06-11 11:08:06,858 - httpcore.connection - DEBUG - connect_tcp.complete return_value=<httpcore._backends.sync.SyncStream object at 0x11a17f410>
|
19195 |
+
2025-06-11 11:08:06,858 - httpcore.http11 - DEBUG - send_request_headers.started request=<Request [b'GET']>
|
19196 |
+
2025-06-11 11:08:06,859 - httpcore.http11 - DEBUG - send_request_headers.complete
|
19197 |
+
2025-06-11 11:08:06,859 - httpcore.http11 - DEBUG - send_request_body.started request=<Request [b'GET']>
|
19198 |
+
2025-06-11 11:08:06,859 - httpcore.http11 - DEBUG - send_request_body.complete
|
19199 |
+
2025-06-11 11:08:06,859 - httpcore.http11 - DEBUG - receive_response_headers.started request=<Request [b'GET']>
|
19200 |
+
2025-06-11 11:08:06,859 - httpcore.http11 - DEBUG - receive_response_headers.complete return_value=(b'HTTP/1.1', 200, b'OK', [(b'date', b'Wed, 11 Jun 2025 02:08:06 GMT'), (b'server', b'uvicorn'), (b'content-length', b'4'), (b'content-type', b'application/json')])
|
19201 |
+
2025-06-11 11:08:06,859 - httpx - INFO - HTTP Request: GET http://localhost:7860/gradio_api/startup-events "HTTP/1.1 200 OK"
|
19202 |
+
2025-06-11 11:08:06,859 - httpcore.http11 - DEBUG - receive_response_body.started request=<Request [b'GET']>
|
19203 |
+
2025-06-11 11:08:06,860 - httpcore.http11 - DEBUG - receive_response_body.complete
|
19204 |
+
2025-06-11 11:08:06,860 - httpcore.http11 - DEBUG - response_closed.started
|
19205 |
+
2025-06-11 11:08:06,860 - httpcore.http11 - DEBUG - response_closed.complete
|
19206 |
+
2025-06-11 11:08:06,860 - httpcore.connection - DEBUG - close.started
|
19207 |
+
2025-06-11 11:08:06,860 - httpcore.connection - DEBUG - close.complete
|
19208 |
+
2025-06-11 11:08:06,860 - httpcore.connection - DEBUG - connect_tcp.started host='localhost' port=7860 local_address=None timeout=3 socket_options=None
|
19209 |
+
2025-06-11 11:08:06,860 - httpcore.connection - DEBUG - connect_tcp.complete return_value=<httpcore._backends.sync.SyncStream object at 0x11fb5f990>
|
19210 |
+
2025-06-11 11:08:06,860 - httpcore.http11 - DEBUG - send_request_headers.started request=<Request [b'HEAD']>
|
19211 |
+
2025-06-11 11:08:06,861 - httpcore.http11 - DEBUG - send_request_headers.complete
|
19212 |
+
2025-06-11 11:08:06,861 - httpcore.http11 - DEBUG - send_request_body.started request=<Request [b'HEAD']>
|
19213 |
+
2025-06-11 11:08:06,861 - httpcore.http11 - DEBUG - send_request_body.complete
|
19214 |
+
2025-06-11 11:08:06,861 - httpcore.http11 - DEBUG - receive_response_headers.started request=<Request [b'HEAD']>
|
19215 |
+
2025-06-11 11:08:06,867 - httpcore.http11 - DEBUG - receive_response_headers.complete return_value=(b'HTTP/1.1', 200, b'OK', [(b'date', b'Wed, 11 Jun 2025 02:08:06 GMT'), (b'server', b'uvicorn'), (b'content-length', b'142574'), (b'content-type', b'text/html; charset=utf-8')])
|
19216 |
+
2025-06-11 11:08:06,868 - httpx - INFO - HTTP Request: HEAD http://localhost:7860/ "HTTP/1.1 200 OK"
|
19217 |
+
2025-06-11 11:08:06,868 - httpcore.http11 - DEBUG - receive_response_body.started request=<Request [b'HEAD']>
|
19218 |
+
2025-06-11 11:08:06,868 - httpcore.http11 - DEBUG - receive_response_body.complete
|
19219 |
+
2025-06-11 11:08:06,868 - httpcore.http11 - DEBUG - response_closed.started
|
19220 |
+
2025-06-11 11:08:06,868 - httpcore.http11 - DEBUG - response_closed.complete
|
19221 |
+
2025-06-11 11:08:06,868 - httpcore.connection - DEBUG - close.started
|
19222 |
+
2025-06-11 11:08:06,868 - httpcore.connection - DEBUG - close.complete
|
19223 |
+
2025-06-11 11:08:06,879 - httpcore.connection - DEBUG - connect_tcp.started host='api.gradio.app' port=443 local_address=None timeout=30 socket_options=None
|
19224 |
+
2025-06-11 11:08:06,892 - httpcore.connection - DEBUG - connect_tcp.complete return_value=<httpcore._backends.sync.SyncStream object at 0x11a1c8390>
|
19225 |
+
2025-06-11 11:08:06,892 - httpcore.connection - DEBUG - start_tls.started ssl_context=<ssl.SSLContext object at 0x107835370> server_hostname='api.gradio.app' timeout=3
|
19226 |
+
2025-06-11 11:08:07,028 - httpcore.connection - DEBUG - connect_tcp.complete return_value=<httpcore._backends.sync.SyncStream object at 0x11fbd9090>
|
19227 |
+
2025-06-11 11:08:07,028 - httpcore.connection - DEBUG - start_tls.started ssl_context=<ssl.SSLContext object at 0x11fb68a70> server_hostname='api.gradio.app' timeout=30
|
19228 |
+
2025-06-11 11:08:07,032 - urllib3.connectionpool - DEBUG - https://huggingface.co:443 "HEAD /api/telemetry/gradio/initiated HTTP/1.1" 200 0
|
19229 |
+
2025-06-11 11:08:07,169 - httpcore.connection - DEBUG - start_tls.complete return_value=<httpcore._backends.sync.SyncStream object at 0x118e975d0>
|
19230 |
+
2025-06-11 11:08:07,170 - httpcore.http11 - DEBUG - send_request_headers.started request=<Request [b'GET']>
|
19231 |
+
2025-06-11 11:08:07,170 - httpcore.http11 - DEBUG - send_request_headers.complete
|
19232 |
+
2025-06-11 11:08:07,170 - httpcore.http11 - DEBUG - send_request_body.started request=<Request [b'GET']>
|
19233 |
+
2025-06-11 11:08:07,171 - httpcore.http11 - DEBUG - send_request_body.complete
|
19234 |
+
2025-06-11 11:08:07,171 - httpcore.http11 - DEBUG - receive_response_headers.started request=<Request [b'GET']>
|
19235 |
+
2025-06-11 11:08:07,310 - httpcore.http11 - DEBUG - receive_response_headers.complete return_value=(b'HTTP/1.1', 200, b'OK', [(b'Date', b'Wed, 11 Jun 2025 02:08:07 GMT'), (b'Content-Type', b'application/json'), (b'Content-Length', b'21'), (b'Connection', b'keep-alive'), (b'Server', b'nginx/1.18.0'), (b'Access-Control-Allow-Origin', b'*')])
|
19236 |
+
2025-06-11 11:08:07,311 - httpx - INFO - HTTP Request: GET https://api.gradio.app/pkg-version "HTTP/1.1 200 OK"
|
19237 |
+
2025-06-11 11:08:07,311 - httpcore.http11 - DEBUG - receive_response_body.started request=<Request [b'GET']>
|
19238 |
+
2025-06-11 11:08:07,312 - httpcore.http11 - DEBUG - receive_response_body.complete
|
19239 |
+
2025-06-11 11:08:07,312 - httpcore.http11 - DEBUG - response_closed.started
|
19240 |
+
2025-06-11 11:08:07,312 - httpcore.http11 - DEBUG - response_closed.complete
|
19241 |
+
2025-06-11 11:08:07,312 - httpcore.connection - DEBUG - close.started
|
19242 |
+
2025-06-11 11:08:07,312 - httpcore.connection - DEBUG - close.complete
|
19243 |
+
2025-06-11 11:08:07,326 - httpcore.connection - DEBUG - start_tls.complete return_value=<httpcore._backends.sync.SyncStream object at 0x119af4110>
|
19244 |
+
2025-06-11 11:08:07,326 - httpcore.http11 - DEBUG - send_request_headers.started request=<Request [b'GET']>
|
19245 |
+
2025-06-11 11:08:07,326 - httpcore.http11 - DEBUG - send_request_headers.complete
|
19246 |
+
2025-06-11 11:08:07,326 - httpcore.http11 - DEBUG - send_request_body.started request=<Request [b'GET']>
|
19247 |
+
2025-06-11 11:08:07,326 - httpcore.http11 - DEBUG - send_request_body.complete
|
19248 |
+
2025-06-11 11:08:07,326 - httpcore.http11 - DEBUG - receive_response_headers.started request=<Request [b'GET']>
|
19249 |
+
2025-06-11 11:08:07,477 - httpcore.http11 - DEBUG - receive_response_headers.complete return_value=(b'HTTP/1.1', 200, b'OK', [(b'Date', b'Wed, 11 Jun 2025 02:08:07 GMT'), (b'Content-Type', b'text/html; charset=utf-8'), (b'Transfer-Encoding', b'chunked'), (b'Connection', b'keep-alive'), (b'Server', b'nginx/1.18.0'), (b'ContentType', b'application/json'), (b'Access-Control-Allow-Origin', b'*'), (b'Content-Encoding', b'gzip')])
|
19250 |
+
2025-06-11 11:08:07,478 - httpx - INFO - HTTP Request: GET https://api.gradio.app/v3/tunnel-request "HTTP/1.1 200 OK"
|
19251 |
+
2025-06-11 11:08:07,478 - httpcore.http11 - DEBUG - receive_response_body.started request=<Request [b'GET']>
|
19252 |
+
2025-06-11 11:08:07,478 - httpcore.http11 - DEBUG - receive_response_body.complete
|
19253 |
+
2025-06-11 11:08:07,479 - httpcore.http11 - DEBUG - response_closed.started
|
19254 |
+
2025-06-11 11:08:07,479 - httpcore.http11 - DEBUG - response_closed.complete
|
19255 |
+
2025-06-11 11:08:07,479 - httpcore.connection - DEBUG - close.started
|
19256 |
+
2025-06-11 11:08:07,479 - httpcore.connection - DEBUG - close.complete
|
19257 |
+
2025-06-11 11:08:08,331 - urllib3.connectionpool - DEBUG - Starting new HTTPS connection (1): huggingface.co:443
|
19258 |
+
2025-06-11 11:08:08,553 - urllib3.connectionpool - DEBUG - https://huggingface.co:443 "HEAD /api/telemetry/gradio/launched HTTP/1.1" 200 0
|
19259 |
+
2025-06-11 11:08:11,643 - hardware_detector - INFO - Initializing HardwareDetector
|
19260 |
+
2025-06-11 11:08:11,643 - hardware_detector - DEBUG - Starting system hardware detection
|
19261 |
+
2025-06-11 11:08:11,643 - hardware_detector - DEBUG - Platform: Darwin, Architecture: arm64
|
19262 |
+
2025-06-11 11:08:11,643 - hardware_detector - DEBUG - CPU cores: 16, Python: 3.11.11
|
19263 |
+
2025-06-11 11:08:11,643 - hardware_detector - DEBUG - Attempting GPU detection via nvidia-smi
|
19264 |
+
2025-06-11 11:08:11,650 - hardware_detector - DEBUG - nvidia-smi not found, no NVIDIA GPU detected
|
19265 |
+
2025-06-11 11:08:11,650 - hardware_detector - DEBUG - Checking PyTorch availability
|
19266 |
+
2025-06-11 11:08:11,650 - hardware_detector - INFO - PyTorch 2.7.0 detected
|
19267 |
+
2025-06-11 11:08:11,651 - hardware_detector - DEBUG - CUDA available: False, MPS available: True
|
19268 |
+
2025-06-11 11:08:11,651 - hardware_detector - INFO - Hardware detection completed successfully
|
19269 |
+
2025-06-11 11:08:11,651 - hardware_detector - DEBUG - Detected specs: {'platform': 'Darwin', 'architecture': 'arm64', 'cpu_count': 16, 'python_version': '3.11.11', 'gpu_info': None, 'cuda_available': False, 'mps_available': True, 'torch_version': '2.7.0'}
|
19270 |
+
2025-06-11 11:08:11,651 - __main__ - INFO - 🔍 DEBUG: Full detected specs: {'platform': 'Darwin', 'architecture': 'arm64', 'cpu_count': 16, 'python_version': '3.11.11', 'gpu_info': None, 'cuda_available': False, 'mps_available': True, 'torch_version': '2.7.0'}
|
19271 |
+
2025-06-11 11:08:11,651 - __main__ - INFO - 🔍 DEBUG: Platform: Darwin
|
19272 |
+
2025-06-11 11:08:11,651 - __main__ - INFO - 🔍 DEBUG: Architecture: arm64
|
19273 |
+
2025-06-11 11:08:11,651 - __main__ - INFO - 🔍 DEBUG: MPS Available: True
|
19274 |
+
2025-06-11 11:08:11,651 - __main__ - INFO - 🔍 DEBUG: Is Apple Silicon: True
|
19275 |
+
2025-06-11 11:08:11,651 - __main__ - INFO - 🔍 DEBUG: GPU Info: None
|
19276 |
+
2025-06-11 11:08:11,651 - __main__ - INFO - 🔍 DEBUG: Detected RAM (GB): 16
|
19277 |
+
2025-06-11 11:08:11,658 - __main__ - INFO - 🔍 DEBUG: PSUtil detected 128GB unified memory
|
19278 |
+
2025-06-11 11:08:11,694 - simple_memory_calculator - INFO - Getting memory requirements for model: black-forest-labs/FLUX.1-schnell
|
19279 |
+
2025-06-11 11:08:11,694 - simple_memory_calculator - INFO - Using known memory data for black-forest-labs/FLUX.1-schnell
|
19280 |
+
2025-06-11 11:08:11,694 - simple_memory_calculator - DEBUG - Known data: {'params_billions': 12.0, 'fp16_gb': 24.0, 'inference_fp16_gb': 36.0}
|
19281 |
+
2025-06-11 11:08:11,694 - simple_memory_calculator - INFO - Generating memory recommendations for black-forest-labs/FLUX.1-schnell with 8.0GB VRAM
|
19282 |
+
2025-06-11 11:08:11,694 - simple_memory_calculator - INFO - Getting memory requirements for model: black-forest-labs/FLUX.1-schnell
|
19283 |
+
2025-06-11 11:08:11,695 - simple_memory_calculator - DEBUG - Using cached memory data for black-forest-labs/FLUX.1-schnell
|
19284 |
+
2025-06-11 11:08:11,695 - simple_memory_calculator - DEBUG - Model memory: 24.0GB, Inference memory: 36.0GB
|
19285 |
+
2025-06-11 11:08:11,695 - simple_memory_calculator - INFO - Getting memory requirements for model: black-forest-labs/FLUX.1-schnell
|
19286 |
+
2025-06-11 11:08:11,695 - simple_memory_calculator - DEBUG - Using cached memory data for black-forest-labs/FLUX.1-schnell
|
19287 |
+
2025-06-11 11:08:11,809 - __main__ - INFO - 🔍 DEBUG: System profiler output: hardware:
|
19288 |
+
|
19289 |
+
hardware overview:
|
19290 |
+
|
19291 |
+
model name: macbook pro
|
19292 |
+
model identifier: mac16,5
|
19293 |
+
model number: z1fs0002mkh/a
|
19294 |
+
chip: apple m4 max
|
19295 |
+
total number of cores: 16 (12 performance and 4 efficiency)
|
19296 |
+
memory: 128 gb
|
19297 |
+
system firmware version: 11881.61.3
|
19298 |
+
os loader version: 11881.61.3
|
19299 |
+
serial number (system): kkvqcmx4dt
|
19300 |
+
hardware uuid: 92bfa67e-d816-5b1d-99bb-f9a2824999ba
|
19301 |
+
provisioning udid: 00006041-001269200280801c
|
19302 |
+
activation lock status: enabled
|
19303 |
+
|
19304 |
+
|
19305 |
+
2025-06-11 11:08:11,809 - __main__ - INFO - 🔍 DEBUG: Detected Apple chip: M4 Max
|
19306 |
+
2025-06-11 11:08:11,809 - __main__ - INFO - 🔍 DEBUG: Final values - Platform: Apple Silicon, GPU Vendor: Apple Silicon, GPU Name: M4 Max, Unified RAM: 128GB
|
19307 |
+
2025-06-11 11:09:25,254 - __main__ - INFO - Initializing GradioAutodiffusers
|
19308 |
+
2025-06-11 11:09:25,254 - __main__ - DEBUG - API key found, length: 39
|
19309 |
+
2025-06-11 11:09:25,254 - auto_diffusers - INFO - Initializing AutoDiffusersGenerator
|
19310 |
+
2025-06-11 11:09:25,254 - auto_diffusers - DEBUG - API key length: 39
|
19311 |
+
2025-06-11 11:09:25,254 - auto_diffusers - WARNING - Tool calling dependencies not available, running without tools
|
19312 |
+
2025-06-11 11:09:25,254 - hardware_detector - INFO - Initializing HardwareDetector
|
19313 |
+
2025-06-11 11:09:25,254 - hardware_detector - DEBUG - Starting system hardware detection
|
19314 |
+
2025-06-11 11:09:25,254 - hardware_detector - DEBUG - Platform: Darwin, Architecture: arm64
|
19315 |
+
2025-06-11 11:09:25,254 - hardware_detector - DEBUG - CPU cores: 16, Python: 3.11.11
|
19316 |
+
2025-06-11 11:09:25,254 - hardware_detector - DEBUG - Attempting GPU detection via nvidia-smi
|
19317 |
+
2025-06-11 11:09:25,258 - hardware_detector - DEBUG - nvidia-smi not found, no NVIDIA GPU detected
|
19318 |
+
2025-06-11 11:09:25,258 - hardware_detector - DEBUG - Checking PyTorch availability
|
19319 |
+
2025-06-11 11:09:25,724 - hardware_detector - INFO - PyTorch 2.7.0 detected
|
19320 |
+
2025-06-11 11:09:25,724 - hardware_detector - DEBUG - CUDA available: False, MPS available: True
|
19321 |
+
2025-06-11 11:09:25,724 - hardware_detector - INFO - Hardware detection completed successfully
|
19322 |
+
2025-06-11 11:09:25,724 - hardware_detector - DEBUG - Detected specs: {'platform': 'Darwin', 'architecture': 'arm64', 'cpu_count': 16, 'python_version': '3.11.11', 'gpu_info': None, 'cuda_available': False, 'mps_available': True, 'torch_version': '2.7.0'}
|
19323 |
+
2025-06-11 11:09:25,724 - auto_diffusers - INFO - Hardware detector initialized successfully
|
19324 |
+
2025-06-11 11:09:25,724 - __main__ - INFO - AutoDiffusersGenerator initialized successfully
|
19325 |
+
2025-06-11 11:09:25,724 - simple_memory_calculator - INFO - Initializing SimpleMemoryCalculator
|
19326 |
+
2025-06-11 11:09:25,724 - simple_memory_calculator - DEBUG - HuggingFace API initialized
|
19327 |
+
2025-06-11 11:09:25,724 - simple_memory_calculator - DEBUG - Known models in database: 4
|
19328 |
+
2025-06-11 11:09:25,724 - __main__ - INFO - SimpleMemoryCalculator initialized successfully
|
19329 |
+
2025-06-11 11:09:25,724 - __main__ - DEBUG - Default model settings: gemini-2.5-flash-preview-05-20, temp=0.7
|
19330 |
+
2025-06-11 11:09:25,726 - asyncio - DEBUG - Using selector: KqueueSelector
|
19331 |
+
2025-06-11 11:09:25,738 - httpcore.connection - DEBUG - connect_tcp.started host='api.gradio.app' port=443 local_address=None timeout=3 socket_options=None
|
19332 |
+
2025-06-11 11:09:25,747 - urllib3.connectionpool - DEBUG - Starting new HTTPS connection (1): huggingface.co:443
|
19333 |
+
2025-06-11 11:09:25,821 - asyncio - DEBUG - Using selector: KqueueSelector
|
19334 |
+
2025-06-11 11:09:25,852 - httpcore.connection - DEBUG - connect_tcp.started host='localhost' port=7860 local_address=None timeout=None socket_options=None
|
19335 |
+
2025-06-11 11:09:25,853 - httpcore.connection - DEBUG - connect_tcp.complete return_value=<httpcore._backends.sync.SyncStream object at 0x12867ed90>
|
19336 |
+
2025-06-11 11:09:25,853 - httpcore.http11 - DEBUG - send_request_headers.started request=<Request [b'GET']>
|
19337 |
+
2025-06-11 11:09:25,853 - httpcore.http11 - DEBUG - send_request_headers.complete
|
19338 |
+
2025-06-11 11:09:25,853 - httpcore.http11 - DEBUG - send_request_body.started request=<Request [b'GET']>
|
19339 |
+
2025-06-11 11:09:25,853 - httpcore.http11 - DEBUG - send_request_body.complete
|
19340 |
+
2025-06-11 11:09:25,854 - httpcore.http11 - DEBUG - receive_response_headers.started request=<Request [b'GET']>
|
19341 |
+
2025-06-11 11:09:25,854 - httpcore.http11 - DEBUG - receive_response_headers.complete return_value=(b'HTTP/1.1', 200, b'OK', [(b'date', b'Wed, 11 Jun 2025 02:09:25 GMT'), (b'server', b'uvicorn'), (b'content-length', b'4'), (b'content-type', b'application/json')])
|
19342 |
+
2025-06-11 11:09:25,854 - httpx - INFO - HTTP Request: GET http://localhost:7860/gradio_api/startup-events "HTTP/1.1 200 OK"
|
19343 |
+
2025-06-11 11:09:25,854 - httpcore.http11 - DEBUG - receive_response_body.started request=<Request [b'GET']>
|
19344 |
+
2025-06-11 11:09:25,854 - httpcore.http11 - DEBUG - receive_response_body.complete
|
19345 |
+
2025-06-11 11:09:25,854 - httpcore.http11 - DEBUG - response_closed.started
|
19346 |
+
2025-06-11 11:09:25,854 - httpcore.http11 - DEBUG - response_closed.complete
|
19347 |
+
2025-06-11 11:09:25,854 - httpcore.connection - DEBUG - close.started
|
19348 |
+
2025-06-11 11:09:25,854 - httpcore.connection - DEBUG - close.complete
|
19349 |
+
2025-06-11 11:09:25,855 - httpcore.connection - DEBUG - connect_tcp.started host='localhost' port=7860 local_address=None timeout=3 socket_options=None
|
19350 |
+
2025-06-11 11:09:25,855 - httpcore.connection - DEBUG - connect_tcp.complete return_value=<httpcore._backends.sync.SyncStream object at 0x12c65f3d0>
|
19351 |
+
2025-06-11 11:09:25,855 - httpcore.http11 - DEBUG - send_request_headers.started request=<Request [b'HEAD']>
|
19352 |
+
2025-06-11 11:09:25,855 - httpcore.http11 - DEBUG - send_request_headers.complete
|
19353 |
+
2025-06-11 11:09:25,855 - httpcore.http11 - DEBUG - send_request_body.started request=<Request [b'HEAD']>
|
19354 |
+
2025-06-11 11:09:25,855 - httpcore.http11 - DEBUG - send_request_body.complete
|
19355 |
+
2025-06-11 11:09:25,855 - httpcore.http11 - DEBUG - receive_response_headers.started request=<Request [b'HEAD']>
|
19356 |
+
2025-06-11 11:09:25,862 - httpcore.http11 - DEBUG - receive_response_headers.complete return_value=(b'HTTP/1.1', 200, b'OK', [(b'date', b'Wed, 11 Jun 2025 02:09:25 GMT'), (b'server', b'uvicorn'), (b'content-length', b'142574'), (b'content-type', b'text/html; charset=utf-8')])
|
19357 |
+
2025-06-11 11:09:25,862 - httpx - INFO - HTTP Request: HEAD http://localhost:7860/ "HTTP/1.1 200 OK"
|
19358 |
+
2025-06-11 11:09:25,862 - httpcore.http11 - DEBUG - receive_response_body.started request=<Request [b'HEAD']>
|
19359 |
+
2025-06-11 11:09:25,862 - httpcore.http11 - DEBUG - receive_response_body.complete
|
19360 |
+
2025-06-11 11:09:25,862 - httpcore.http11 - DEBUG - response_closed.started
|
19361 |
+
2025-06-11 11:09:25,862 - httpcore.http11 - DEBUG - response_closed.complete
|
19362 |
+
2025-06-11 11:09:25,862 - httpcore.connection - DEBUG - close.started
|
19363 |
+
2025-06-11 11:09:25,862 - httpcore.connection - DEBUG - close.complete
|
19364 |
+
2025-06-11 11:09:25,874 - httpcore.connection - DEBUG - connect_tcp.started host='api.gradio.app' port=443 local_address=None timeout=30 socket_options=None
|
19365 |
+
2025-06-11 11:09:25,904 - httpcore.connection - DEBUG - connect_tcp.complete return_value=<httpcore._backends.sync.SyncStream object at 0x12868d1d0>
|
19366 |
+
2025-06-11 11:09:25,904 - httpcore.connection - DEBUG - start_tls.started ssl_context=<ssl.SSLContext object at 0x10cd35370> server_hostname='api.gradio.app' timeout=3
|
19367 |
+
2025-06-11 11:09:26,016 - httpcore.connection - DEBUG - connect_tcp.complete return_value=<httpcore._backends.sync.SyncStream object at 0x12c6d3790>
|
19368 |
+
2025-06-11 11:09:26,016 - httpcore.connection - DEBUG - start_tls.started ssl_context=<ssl.SSLContext object at 0x12c668a70> server_hostname='api.gradio.app' timeout=30
|
19369 |
+
2025-06-11 11:09:26,028 - urllib3.connectionpool - DEBUG - https://huggingface.co:443 "HEAD /api/telemetry/gradio/initiated HTTP/1.1" 200 0
|
19370 |
+
2025-06-11 11:09:26,177 - httpcore.connection - DEBUG - start_tls.complete return_value=<httpcore._backends.sync.SyncStream object at 0x12c3e0950>
|
19371 |
+
2025-06-11 11:09:26,177 - httpcore.http11 - DEBUG - send_request_headers.started request=<Request [b'GET']>
|
19372 |
+
2025-06-11 11:09:26,177 - httpcore.http11 - DEBUG - send_request_headers.complete
|
19373 |
+
2025-06-11 11:09:26,177 - httpcore.http11 - DEBUG - send_request_body.started request=<Request [b'GET']>
|
19374 |
+
2025-06-11 11:09:26,177 - httpcore.http11 - DEBUG - send_request_body.complete
|
19375 |
+
2025-06-11 11:09:26,177 - httpcore.http11 - DEBUG - receive_response_headers.started request=<Request [b'GET']>
|
19376 |
+
2025-06-11 11:09:26,306 - httpcore.connection - DEBUG - start_tls.complete return_value=<httpcore._backends.sync.SyncStream object at 0x11c64e6d0>
|
19377 |
+
2025-06-11 11:09:26,307 - httpcore.http11 - DEBUG - send_request_headers.started request=<Request [b'GET']>
|
19378 |
+
2025-06-11 11:09:26,307 - httpcore.http11 - DEBUG - send_request_headers.complete
|
19379 |
+
2025-06-11 11:09:26,308 - httpcore.http11 - DEBUG - send_request_body.started request=<Request [b'GET']>
|
19380 |
+
2025-06-11 11:09:26,308 - httpcore.http11 - DEBUG - send_request_body.complete
|
19381 |
+
2025-06-11 11:09:26,308 - httpcore.http11 - DEBUG - receive_response_headers.started request=<Request [b'GET']>
|
19382 |
+
2025-06-11 11:09:26,315 - httpcore.http11 - DEBUG - receive_response_headers.complete return_value=(b'HTTP/1.1', 200, b'OK', [(b'Date', b'Wed, 11 Jun 2025 02:09:26 GMT'), (b'Content-Type', b'application/json'), (b'Content-Length', b'21'), (b'Connection', b'keep-alive'), (b'Server', b'nginx/1.18.0'), (b'Access-Control-Allow-Origin', b'*')])
|
19383 |
+
2025-06-11 11:09:26,315 - httpx - INFO - HTTP Request: GET https://api.gradio.app/pkg-version "HTTP/1.1 200 OK"
|
19384 |
+
2025-06-11 11:09:26,315 - httpcore.http11 - DEBUG - receive_response_body.started request=<Request [b'GET']>
|
19385 |
+
2025-06-11 11:09:26,316 - httpcore.http11 - DEBUG - receive_response_body.complete
|
19386 |
+
2025-06-11 11:09:26,316 - httpcore.http11 - DEBUG - response_closed.started
|
19387 |
+
2025-06-11 11:09:26,316 - httpcore.http11 - DEBUG - response_closed.complete
|
19388 |
+
2025-06-11 11:09:26,316 - httpcore.connection - DEBUG - close.started
|
19389 |
+
2025-06-11 11:09:26,316 - httpcore.connection - DEBUG - close.complete
|
19390 |
+
2025-06-11 11:09:26,454 - httpcore.http11 - DEBUG - receive_response_headers.complete return_value=(b'HTTP/1.1', 200, b'OK', [(b'Date', b'Wed, 11 Jun 2025 02:09:26 GMT'), (b'Content-Type', b'text/html; charset=utf-8'), (b'Transfer-Encoding', b'chunked'), (b'Connection', b'keep-alive'), (b'Server', b'nginx/1.18.0'), (b'ContentType', b'application/json'), (b'Access-Control-Allow-Origin', b'*'), (b'Content-Encoding', b'gzip')])
|
19391 |
+
2025-06-11 11:09:26,455 - httpx - INFO - HTTP Request: GET https://api.gradio.app/v3/tunnel-request "HTTP/1.1 200 OK"
|
19392 |
+
2025-06-11 11:09:26,455 - httpcore.http11 - DEBUG - receive_response_body.started request=<Request [b'GET']>
|
19393 |
+
2025-06-11 11:09:26,455 - httpcore.http11 - DEBUG - receive_response_body.complete
|
19394 |
+
2025-06-11 11:09:26,456 - httpcore.http11 - DEBUG - response_closed.started
|
19395 |
+
2025-06-11 11:09:26,456 - httpcore.http11 - DEBUG - response_closed.complete
|
19396 |
+
2025-06-11 11:09:26,456 - httpcore.connection - DEBUG - close.started
|
19397 |
+
2025-06-11 11:09:26,457 - httpcore.connection - DEBUG - close.complete
|
19398 |
+
2025-06-11 11:09:27,037 - urllib3.connectionpool - DEBUG - Starting new HTTPS connection (1): huggingface.co:443
|
19399 |
+
2025-06-11 11:09:27,127 - simple_memory_calculator - INFO - Getting memory requirements for model: black-forest-labs/FLUX.1-schnell
|
19400 |
+
2025-06-11 11:09:27,127 - hardware_detector - INFO - Initializing HardwareDetector
|
19401 |
+
2025-06-11 11:09:27,127 - simple_memory_calculator - INFO - Using known memory data for black-forest-labs/FLUX.1-schnell
|
19402 |
+
2025-06-11 11:09:27,127 - hardware_detector - DEBUG - Starting system hardware detection
|
19403 |
+
2025-06-11 11:09:27,127 - simple_memory_calculator - DEBUG - Known data: {'params_billions': 12.0, 'fp16_gb': 24.0, 'inference_fp16_gb': 36.0}
|
19404 |
+
2025-06-11 11:09:27,127 - hardware_detector - DEBUG - Platform: Darwin, Architecture: arm64
|
19405 |
+
2025-06-11 11:09:27,127 - simple_memory_calculator - INFO - Generating memory recommendations for black-forest-labs/FLUX.1-schnell with 8.0GB VRAM
|
19406 |
+
2025-06-11 11:09:27,127 - hardware_detector - DEBUG - CPU cores: 16, Python: 3.11.11
|
19407 |
+
2025-06-11 11:09:27,127 - simple_memory_calculator - INFO - Getting memory requirements for model: black-forest-labs/FLUX.1-schnell
|
19408 |
+
2025-06-11 11:09:27,127 - hardware_detector - DEBUG - Attempting GPU detection via nvidia-smi
|
19409 |
+
2025-06-11 11:09:27,128 - simple_memory_calculator - DEBUG - Using cached memory data for black-forest-labs/FLUX.1-schnell
|
19410 |
+
2025-06-11 11:09:27,128 - simple_memory_calculator - DEBUG - Model memory: 24.0GB, Inference memory: 36.0GB
|
19411 |
+
2025-06-11 11:09:27,132 - simple_memory_calculator - INFO - Getting memory requirements for model: black-forest-labs/FLUX.1-schnell
|
19412 |
+
2025-06-11 11:09:27,133 - simple_memory_calculator - DEBUG - Using cached memory data for black-forest-labs/FLUX.1-schnell
|
19413 |
+
2025-06-11 11:09:27,134 - hardware_detector - DEBUG - nvidia-smi not found, no NVIDIA GPU detected
|
19414 |
+
2025-06-11 11:09:27,135 - hardware_detector - DEBUG - Checking PyTorch availability
|
19415 |
+
2025-06-11 11:09:27,135 - hardware_detector - INFO - PyTorch 2.7.0 detected
|
19416 |
+
2025-06-11 11:09:27,135 - hardware_detector - DEBUG - CUDA available: False, MPS available: True
|
19417 |
+
2025-06-11 11:09:27,135 - hardware_detector - INFO - Hardware detection completed successfully
|
19418 |
+
2025-06-11 11:09:27,135 - hardware_detector - DEBUG - Detected specs: {'platform': 'Darwin', 'architecture': 'arm64', 'cpu_count': 16, 'python_version': '3.11.11', 'gpu_info': None, 'cuda_available': False, 'mps_available': True, 'torch_version': '2.7.0'}
|
19419 |
+
2025-06-11 11:09:27,135 - __main__ - INFO - 🔍 DEBUG: Full detected specs: {'platform': 'Darwin', 'architecture': 'arm64', 'cpu_count': 16, 'python_version': '3.11.11', 'gpu_info': None, 'cuda_available': False, 'mps_available': True, 'torch_version': '2.7.0'}
|
19420 |
+
2025-06-11 11:09:27,135 - __main__ - INFO - 🔍 DEBUG: Platform: Darwin
|
19421 |
+
2025-06-11 11:09:27,135 - __main__ - INFO - 🔍 DEBUG: Architecture: arm64
|
19422 |
+
2025-06-11 11:09:27,135 - __main__ - INFO - 🔍 DEBUG: MPS Available: True
|
19423 |
+
2025-06-11 11:09:27,135 - __main__ - INFO - 🔍 DEBUG: Is Apple Silicon: True
|
19424 |
+
2025-06-11 11:09:27,135 - __main__ - INFO - 🔍 DEBUG: GPU Info: None
|
19425 |
+
2025-06-11 11:09:27,135 - __main__ - INFO - 🔍 DEBUG: Detected RAM (GB): 16
|
19426 |
+
2025-06-11 11:09:27,140 - __main__ - INFO - 🔍 DEBUG: PSUtil detected 128GB unified memory
|
19427 |
+
2025-06-11 11:09:27,218 - __main__ - INFO - 🔍 DEBUG: System profiler output: hardware:
|
19428 |
+
|
19429 |
+
hardware overview:
|
19430 |
+
|
19431 |
+
model name: macbook pro
|
19432 |
+
model identifier: mac16,5
|
19433 |
+
model number: z1fs0002mkh/a
|
19434 |
+
chip: apple m4 max
|
19435 |
+
total number of cores: 16 (12 performance and 4 efficiency)
|
19436 |
+
memory: 128 gb
|
19437 |
+
system firmware version: 11881.61.3
|
19438 |
+
os loader version: 11881.61.3
|
19439 |
+
serial number (system): kkvqcmx4dt
|
19440 |
+
hardware uuid: 92bfa67e-d816-5b1d-99bb-f9a2824999ba
|
19441 |
+
provisioning udid: 00006041-001269200280801c
|
19442 |
+
activation lock status: enabled
|
19443 |
+
|
19444 |
+
|
19445 |
+
2025-06-11 11:09:27,218 - __main__ - INFO - 🔍 DEBUG: Detected Apple chip: M4 Max
|
19446 |
+
2025-06-11 11:09:27,218 - __main__ - INFO - 🔍 DEBUG: Final values - Platform: Apple Silicon, GPU Vendor: Apple Silicon, GPU Name: M4 Max, Unified RAM: 128GB
|
19447 |
+
2025-06-11 11:09:27,258 - urllib3.connectionpool - DEBUG - https://huggingface.co:443 "HEAD /api/telemetry/gradio/launched HTTP/1.1" 200 0
|
gradio_app.py
CHANGED
@@ -1,10 +1,12 @@
|
|
1 |
import os
|
2 |
import logging
|
|
|
3 |
import gradio as gr
|
4 |
from dotenv import load_dotenv
|
5 |
import google.generativeai as genai
|
6 |
from auto_diffusers import AutoDiffusersGenerator
|
7 |
from simple_memory_calculator import SimpleMemoryCalculator
|
|
|
8 |
|
9 |
load_dotenv()
|
10 |
|
@@ -170,8 +172,12 @@ class GradioAutodiffusers:
|
|
170 |
return f"Error generating code: {str(e)}"
|
171 |
|
172 |
|
173 |
-
def create_gradio_interface():
|
174 |
-
"""Create and configure the Gradio interface.
|
|
|
|
|
|
|
|
|
175 |
|
176 |
app = GradioAutodiffusers()
|
177 |
|
@@ -1641,7 +1647,7 @@ def create_gradio_interface():
|
|
1641 |
|
1642 |
# Hardware Selection Section
|
1643 |
with gr.Group(elem_classes="glass-card"):
|
1644 |
-
with gr.Accordion("⚙️ Hardware Specifications", open=
|
1645 |
gr.HTML("""
|
1646 |
<div class="section-header" style="text-align: center;">
|
1647 |
<p style="margin: 0; color: #6a443a; font-size: 1.1rem; font-weight: 600; font-family: 'Georgia', serif;">
|
@@ -1653,7 +1659,7 @@ def create_gradio_interface():
|
|
1653 |
# Platform, VRAM, and RAM in a single row
|
1654 |
with gr.Row():
|
1655 |
platform = gr.Dropdown(
|
1656 |
-
choices=["Linux", "Darwin", "Windows"],
|
1657 |
label="🖥️ Platform",
|
1658 |
value="Linux",
|
1659 |
info="Your operating system"
|
@@ -1672,6 +1678,14 @@ def create_gradio_interface():
|
|
1672 |
maximum=256,
|
1673 |
info="Total system memory"
|
1674 |
)
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
1675 |
|
1676 |
# GPU configuration on separate lines
|
1677 |
gpu_vendor = gr.Dropdown(
|
@@ -1711,6 +1725,7 @@ def create_gradio_interface():
|
|
1711 |
visible=False
|
1712 |
)
|
1713 |
|
|
|
1714 |
# Model Configuration Section
|
1715 |
with gr.Group(elem_classes="glass-card"):
|
1716 |
with gr.Accordion("🤖 Model Configuration", open=True) as model_accordion:
|
@@ -1819,6 +1834,25 @@ def create_gradio_interface():
|
|
1819 |
value=""
|
1820 |
)
|
1821 |
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
1822 |
def on_gpu_vendor_change(vendor):
|
1823 |
"""Handle GPU vendor selection and update series dropdown."""
|
1824 |
if vendor == "Custom (Manual Input)":
|
@@ -1944,12 +1978,16 @@ def create_gradio_interface():
|
|
1944 |
else:
|
1945 |
return vendor if vendor != "Custom (Manual Input)" else "Custom GPU"
|
1946 |
|
1947 |
-
def update_hardware_accordion_title(platform, gpu_vendor, gpu_model, vram_gb, ram_gb):
|
1948 |
"""Update hardware accordion title with current configuration."""
|
1949 |
final_gpu = get_final_gpu_name(gpu_vendor, "", gpu_model)
|
1950 |
if not final_gpu:
|
1951 |
final_gpu = gpu_vendor if gpu_vendor != "Custom (Manual Input)" else "Custom GPU"
|
1952 |
|
|
|
|
|
|
|
|
|
1953 |
# Extract GPU name and VRAM for cleaner display
|
1954 |
gpu_display = final_gpu
|
1955 |
if gpu_model and "(" in gpu_model and "GB" in gpu_model:
|
@@ -2014,6 +2052,13 @@ def create_gradio_interface():
|
|
2014 |
error_msg = str(e)
|
2015 |
return f"**Note:** {error_msg} (API error)", model_name
|
2016 |
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
2017 |
# Connect GPU dropdown change handlers with memory analysis updates
|
2018 |
gpu_vendor.change(
|
2019 |
on_gpu_vendor_change,
|
@@ -2069,18 +2114,195 @@ def create_gradio_interface():
|
|
2069 |
)
|
2070 |
|
2071 |
# Create wrapper functions that return gr.update for accordion labels
|
2072 |
-
def update_hardware_accordion(platform, gpu_vendor, gpu_model, vram_gb, ram_gb):
|
2073 |
-
title = update_hardware_accordion_title(platform, gpu_vendor, gpu_model, vram_gb, ram_gb)
|
2074 |
return gr.update(label=title)
|
2075 |
|
2076 |
def update_model_accordion(model_name, dtype_selection, width, height, inference_steps, memory_analysis_text=""):
|
2077 |
title = update_model_accordion_title(model_name, dtype_selection, width, height, inference_steps, memory_analysis_text)
|
2078 |
return gr.update(label=title)
|
2079 |
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
2080 |
# Load initial accordion titles on startup
|
2081 |
interface.load(
|
2082 |
update_hardware_accordion,
|
2083 |
-
inputs=[platform, gpu_vendor, gpu_model, vram_gb, ram_gb],
|
2084 |
outputs=hardware_accordion
|
2085 |
)
|
2086 |
|
@@ -2093,10 +2315,10 @@ def create_gradio_interface():
|
|
2093 |
# Accordion title update event handlers
|
2094 |
|
2095 |
# Hardware accordion title updates
|
2096 |
-
for component in [platform, gpu_vendor, gpu_model, vram_gb, ram_gb]:
|
2097 |
component.change(
|
2098 |
update_hardware_accordion,
|
2099 |
-
inputs=[platform, gpu_vendor, gpu_model, vram_gb, ram_gb],
|
2100 |
outputs=hardware_accordion
|
2101 |
)
|
2102 |
|
@@ -2191,20 +2413,28 @@ def create_gradio_interface():
|
|
2191 |
|
2192 |
return '\n'.join(filtered_lines)
|
2193 |
|
2194 |
-
def generate_with_combined_gpu_name(gpu_vendor, gpu_series, gpu_model, vram_gb, ram_gb, platform, model_name, dtype_selection, width, height, inference_steps):
|
2195 |
"""Generate code with the correct GPU name from multi-level selection, including memory analysis."""
|
2196 |
final_gpu_name = get_final_gpu_name(gpu_vendor, gpu_series, gpu_model)
|
2197 |
|
2198 |
# Constant prompt text
|
2199 |
prompt_text = "A cat holding a sign that says hello world"
|
2200 |
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
2201 |
# STEP 1: Perform memory analysis BEFORE code generation
|
2202 |
memory_analysis_data = None
|
2203 |
memory_header = ""
|
2204 |
|
2205 |
try:
|
2206 |
-
if model_name and
|
2207 |
-
memory_info, recommendations, _ = app.analyze_model_memory(model_name,
|
2208 |
|
2209 |
# Package memory analysis for Gemini API
|
2210 |
memory_analysis_data = {
|
@@ -2241,8 +2471,17 @@ def create_gradio_interface():
|
|
2241 |
|
2242 |
optimization_strategy = get_optimization_strategy(recommendations)
|
2243 |
|
2244 |
-
|
2245 |
-
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
2246 |
# Model Memory Requirements: {memory_info.get('estimated_inference_memory_fp16_gb', 'Unknown')} GB
|
2247 |
# Recommendation: {', '.join(recommendations.get('recommendations', ['N/A']))}
|
2248 |
# Optimization Strategy: {optimization_strategy}
|
@@ -2256,9 +2495,25 @@ def create_gradio_interface():
|
|
2256 |
"""
|
2257 |
|
2258 |
# STEP 2: Get the actual prompt that will be sent to Gemini
|
2259 |
-
|
2260 |
-
|
2261 |
-
{
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
2262 |
'platform': platform,
|
2263 |
'architecture': 'manual_input',
|
2264 |
'cpu_count': 8,
|
@@ -2267,16 +2522,20 @@ def create_gradio_interface():
|
|
2267 |
'mps_available': platform == 'Darwin' and 'apple' in final_gpu_name.lower() if final_gpu_name else False,
|
2268 |
'torch_version': '2.0+',
|
2269 |
'manual_input': True,
|
2270 |
-
'ram_gb': int(
|
2271 |
'user_dtype': f"torch.{dtype_selection}" if dtype_selection != "Auto" else None,
|
2272 |
-
'gpu_info': [{'name': final_gpu_name, 'memory_mb': int(
|
2273 |
-
}
|
|
|
|
|
|
|
|
|
2274 |
optimization_profile="balanced"
|
2275 |
)
|
2276 |
|
2277 |
# STEP 3: Generate the optimized code WITH memory analysis information
|
2278 |
generated_code = app.generate_code_with_manual_specs(
|
2279 |
-
final_gpu_name,
|
2280 |
model_name, prompt_text, dtype_selection, width, height, inference_steps,
|
2281 |
memory_analysis_data
|
2282 |
)
|
@@ -2294,11 +2553,11 @@ def create_gradio_interface():
|
|
2294 |
full_code_storage = gr.State(value="")
|
2295 |
prompt_storage = gr.State(value="")
|
2296 |
|
2297 |
-
def generate_and_store_code(gpu_vendor, gpu_series, gpu_model, vram_gb, ram_gb, platform, model_name, dtype_selection, width, height, inference_steps):
|
2298 |
"""Generate code and return code for display and full code for storage."""
|
2299 |
summary, full_code, actual_prompt = generate_with_combined_gpu_name(
|
2300 |
gpu_vendor, gpu_series, gpu_model, vram_gb, ram_gb, platform,
|
2301 |
-
model_name, dtype_selection, width, height, inference_steps
|
2302 |
)
|
2303 |
# Show the view prompt button after generation and close accordions
|
2304 |
return (full_code, full_code, False, gr.update(visible=True), actual_prompt,
|
@@ -2317,7 +2576,7 @@ def create_gradio_interface():
|
|
2317 |
generate_and_store_code,
|
2318 |
inputs=[
|
2319 |
gpu_vendor, gpu_series, gpu_model, vram_gb, ram_gb, platform,
|
2320 |
-
model_name, dtype_selection, width, height, inference_steps
|
2321 |
],
|
2322 |
outputs=[code_output, full_code_storage, code_collapsed, view_prompt_btn, prompt_storage, hardware_accordion, model_accordion]
|
2323 |
)
|
@@ -2334,6 +2593,7 @@ def create_gradio_interface():
|
|
2334 |
hide_modal,
|
2335 |
outputs=[prompt_modal, prompt_display]
|
2336 |
)
|
|
|
2337 |
|
2338 |
|
2339 |
|
@@ -2342,12 +2602,54 @@ def create_gradio_interface():
|
|
2342 |
|
2343 |
def main():
|
2344 |
"""Launch the Gradio application."""
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
2345 |
try:
|
2346 |
-
interface = create_gradio_interface()
|
2347 |
interface.launch(
|
2348 |
-
server_name=
|
2349 |
-
server_port=
|
2350 |
-
share=
|
2351 |
show_error=True
|
2352 |
)
|
2353 |
except Exception as e:
|
|
|
1 |
import os
|
2 |
import logging
|
3 |
+
import argparse
|
4 |
import gradio as gr
|
5 |
from dotenv import load_dotenv
|
6 |
import google.generativeai as genai
|
7 |
from auto_diffusers import AutoDiffusersGenerator
|
8 |
from simple_memory_calculator import SimpleMemoryCalculator
|
9 |
+
from hardware_detector import HardwareDetector
|
10 |
|
11 |
load_dotenv()
|
12 |
|
|
|
172 |
return f"Error generating code: {str(e)}"
|
173 |
|
174 |
|
175 |
+
def create_gradio_interface(auto_hardware_detection=False):
|
176 |
+
"""Create and configure the Gradio interface.
|
177 |
+
|
178 |
+
Args:
|
179 |
+
auto_hardware_detection (bool): Whether to enable automatic hardware detection on startup
|
180 |
+
"""
|
181 |
|
182 |
app = GradioAutodiffusers()
|
183 |
|
|
|
1647 |
|
1648 |
# Hardware Selection Section
|
1649 |
with gr.Group(elem_classes="glass-card"):
|
1650 |
+
with gr.Accordion("⚙️ Hardware Specifications", open=True) as hardware_accordion:
|
1651 |
gr.HTML("""
|
1652 |
<div class="section-header" style="text-align: center;">
|
1653 |
<p style="margin: 0; color: #6a443a; font-size: 1.1rem; font-weight: 600; font-family: 'Georgia', serif;">
|
|
|
1659 |
# Platform, VRAM, and RAM in a single row
|
1660 |
with gr.Row():
|
1661 |
platform = gr.Dropdown(
|
1662 |
+
choices=["Linux", "Darwin", "Windows", "Apple Silicon"],
|
1663 |
label="🖥️ Platform",
|
1664 |
value="Linux",
|
1665 |
info="Your operating system"
|
|
|
1678 |
maximum=256,
|
1679 |
info="Total system memory"
|
1680 |
)
|
1681 |
+
unified_ram_gb = gr.Number(
|
1682 |
+
label="🔗 Unified RAM (GB)",
|
1683 |
+
value=16,
|
1684 |
+
minimum=8,
|
1685 |
+
maximum=256,
|
1686 |
+
visible=False,
|
1687 |
+
info="Total unified memory (shared by CPU and GPU)"
|
1688 |
+
)
|
1689 |
|
1690 |
# GPU configuration on separate lines
|
1691 |
gpu_vendor = gr.Dropdown(
|
|
|
1725 |
visible=False
|
1726 |
)
|
1727 |
|
1728 |
+
|
1729 |
# Model Configuration Section
|
1730 |
with gr.Group(elem_classes="glass-card"):
|
1731 |
with gr.Accordion("🤖 Model Configuration", open=True) as model_accordion:
|
|
|
1834 |
value=""
|
1835 |
)
|
1836 |
|
1837 |
+
def on_platform_change(platform_value):
|
1838 |
+
"""Handle platform selection and update field visibility."""
|
1839 |
+
if platform_value == "Apple Silicon":
|
1840 |
+
# Hide VRAM, System RAM, and GPU Series; show Unified RAM
|
1841 |
+
return (
|
1842 |
+
gr.update(visible=False), # vram_gb
|
1843 |
+
gr.update(visible=False), # ram_gb
|
1844 |
+
gr.update(visible=True), # unified_ram_gb
|
1845 |
+
gr.update(visible=False) # gpu_series
|
1846 |
+
)
|
1847 |
+
else:
|
1848 |
+
# Show VRAM and System RAM, hide Unified RAM, show GPU Series
|
1849 |
+
return (
|
1850 |
+
gr.update(visible=True), # vram_gb
|
1851 |
+
gr.update(visible=True), # ram_gb
|
1852 |
+
gr.update(visible=False), # unified_ram_gb
|
1853 |
+
gr.update(visible=True) # gpu_series
|
1854 |
+
)
|
1855 |
+
|
1856 |
def on_gpu_vendor_change(vendor):
|
1857 |
"""Handle GPU vendor selection and update series dropdown."""
|
1858 |
if vendor == "Custom (Manual Input)":
|
|
|
1978 |
else:
|
1979 |
return vendor if vendor != "Custom (Manual Input)" else "Custom GPU"
|
1980 |
|
1981 |
+
def update_hardware_accordion_title(platform, gpu_vendor, gpu_model, vram_gb, ram_gb, unified_ram_gb=None):
|
1982 |
"""Update hardware accordion title with current configuration."""
|
1983 |
final_gpu = get_final_gpu_name(gpu_vendor, "", gpu_model)
|
1984 |
if not final_gpu:
|
1985 |
final_gpu = gpu_vendor if gpu_vendor != "Custom (Manual Input)" else "Custom GPU"
|
1986 |
|
1987 |
+
# Handle Apple Silicon unified memory display
|
1988 |
+
if platform == "Apple Silicon":
|
1989 |
+
return f"⚙️ Hardware: Apple Silicon | {unified_ram_gb}GB Unified Memory"
|
1990 |
+
|
1991 |
# Extract GPU name and VRAM for cleaner display
|
1992 |
gpu_display = final_gpu
|
1993 |
if gpu_model and "(" in gpu_model and "GB" in gpu_model:
|
|
|
2052 |
error_msg = str(e)
|
2053 |
return f"**Note:** {error_msg} (API error)", model_name
|
2054 |
|
2055 |
+
# Connect platform change handler
|
2056 |
+
platform.change(
|
2057 |
+
on_platform_change,
|
2058 |
+
inputs=[platform],
|
2059 |
+
outputs=[vram_gb, ram_gb, unified_ram_gb, gpu_series]
|
2060 |
+
)
|
2061 |
+
|
2062 |
# Connect GPU dropdown change handlers with memory analysis updates
|
2063 |
gpu_vendor.change(
|
2064 |
on_gpu_vendor_change,
|
|
|
2114 |
)
|
2115 |
|
2116 |
# Create wrapper functions that return gr.update for accordion labels
|
2117 |
+
def update_hardware_accordion(platform, gpu_vendor, gpu_model, vram_gb, ram_gb, unified_ram_gb):
|
2118 |
+
title = update_hardware_accordion_title(platform, gpu_vendor, gpu_model, vram_gb, ram_gb, unified_ram_gb)
|
2119 |
return gr.update(label=title)
|
2120 |
|
2121 |
def update_model_accordion(model_name, dtype_selection, width, height, inference_steps, memory_analysis_text=""):
|
2122 |
title = update_model_accordion_title(model_name, dtype_selection, width, height, inference_steps, memory_analysis_text)
|
2123 |
return gr.update(label=title)
|
2124 |
|
2125 |
+
def auto_detect_hardware():
|
2126 |
+
"""Auto-detect hardware specifications and return values for form fields."""
|
2127 |
+
try:
|
2128 |
+
detector = HardwareDetector()
|
2129 |
+
specs = detector.specs
|
2130 |
+
|
2131 |
+
# DEBUG: Log all detected specs
|
2132 |
+
logger.info(f"🔍 DEBUG: Full detected specs: {specs}")
|
2133 |
+
|
2134 |
+
# Extract platform
|
2135 |
+
detected_platform = specs.get('platform', 'Linux')
|
2136 |
+
logger.info(f"🔍 DEBUG: Platform: {detected_platform}")
|
2137 |
+
logger.info(f"🔍 DEBUG: Architecture: {specs.get('architecture')}")
|
2138 |
+
logger.info(f"🔍 DEBUG: MPS Available: {specs.get('mps_available')}")
|
2139 |
+
|
2140 |
+
# Check for Apple Silicon specifically
|
2141 |
+
is_apple_silicon = (
|
2142 |
+
detected_platform == 'Darwin' and
|
2143 |
+
specs.get('architecture') == 'arm64' and
|
2144 |
+
specs.get('mps_available', False)
|
2145 |
+
)
|
2146 |
+
|
2147 |
+
logger.info(f"🔍 DEBUG: Is Apple Silicon: {is_apple_silicon}")
|
2148 |
+
|
2149 |
+
if is_apple_silicon:
|
2150 |
+
detected_platform = "Apple Silicon"
|
2151 |
+
|
2152 |
+
# Extract GPU info
|
2153 |
+
gpu_info = specs.get('gpu_info', [])
|
2154 |
+
logger.info(f"🔍 DEBUG: GPU Info: {gpu_info}")
|
2155 |
+
|
2156 |
+
detected_gpu_name = ""
|
2157 |
+
detected_vram = 8 # default
|
2158 |
+
detected_unified_ram = 16 # default for Apple Silicon
|
2159 |
+
|
2160 |
+
if gpu_info and len(gpu_info) > 0:
|
2161 |
+
gpu = gpu_info[0] # Take first GPU
|
2162 |
+
detected_gpu_name = gpu.get('name', '')
|
2163 |
+
detected_vram = gpu.get('memory_mb', 8192) // 1024 # Convert MB to GB
|
2164 |
+
logger.info(f"🔍 DEBUG: GPU Name: {detected_gpu_name}")
|
2165 |
+
logger.info(f"🔍 DEBUG: GPU Memory MB: {gpu.get('memory_mb')}")
|
2166 |
+
|
2167 |
+
# Extract RAM - try multiple methods for Apple Silicon
|
2168 |
+
detected_ram = specs.get('ram_gb', 16)
|
2169 |
+
logger.info(f"🔍 DEBUG: Detected RAM (GB): {detected_ram}")
|
2170 |
+
|
2171 |
+
# For Apple Silicon, try to get more accurate memory info
|
2172 |
+
if is_apple_silicon:
|
2173 |
+
# Try to get system memory in different ways
|
2174 |
+
try:
|
2175 |
+
import psutil
|
2176 |
+
total_memory_bytes = psutil.virtual_memory().total
|
2177 |
+
total_memory_gb = total_memory_bytes / (1024**3)
|
2178 |
+
detected_unified_ram = round(total_memory_gb)
|
2179 |
+
logger.info(f"🔍 DEBUG: PSUtil detected {detected_unified_ram}GB unified memory")
|
2180 |
+
except:
|
2181 |
+
# Fallback to specs value
|
2182 |
+
detected_unified_ram = detected_ram
|
2183 |
+
logger.info(f"🔍 DEBUG: Using specs RAM value: {detected_unified_ram}GB")
|
2184 |
+
|
2185 |
+
# Map detected GPU to vendor categories
|
2186 |
+
gpu_vendor = "Custom (Manual Input)"
|
2187 |
+
|
2188 |
+
# For Apple Silicon, try to detect the chip series
|
2189 |
+
if is_apple_silicon:
|
2190 |
+
# Try to detect Apple Silicon chip from system info
|
2191 |
+
try:
|
2192 |
+
import subprocess
|
2193 |
+
# Get chip info from system profiler
|
2194 |
+
result = subprocess.run(['system_profiler', 'SPHardwareDataType'],
|
2195 |
+
capture_output=True, text=True, timeout=5)
|
2196 |
+
chip_info = result.stdout.lower()
|
2197 |
+
logger.info(f"🔍 DEBUG: System profiler output: {chip_info}")
|
2198 |
+
|
2199 |
+
if 'm1' in chip_info:
|
2200 |
+
if 'ultra' in chip_info:
|
2201 |
+
detected_gpu_name = "M1 Ultra"
|
2202 |
+
elif 'max' in chip_info:
|
2203 |
+
detected_gpu_name = "M1 Max"
|
2204 |
+
elif 'pro' in chip_info:
|
2205 |
+
detected_gpu_name = "M1 Pro"
|
2206 |
+
else:
|
2207 |
+
detected_gpu_name = "M1"
|
2208 |
+
elif 'm2' in chip_info:
|
2209 |
+
if 'ultra' in chip_info:
|
2210 |
+
detected_gpu_name = "M2 Ultra"
|
2211 |
+
elif 'max' in chip_info:
|
2212 |
+
detected_gpu_name = "M2 Max"
|
2213 |
+
elif 'pro' in chip_info:
|
2214 |
+
detected_gpu_name = "M2 Pro"
|
2215 |
+
else:
|
2216 |
+
detected_gpu_name = "M2"
|
2217 |
+
elif 'm3' in chip_info:
|
2218 |
+
if 'ultra' in chip_info:
|
2219 |
+
detected_gpu_name = "M3 Ultra"
|
2220 |
+
elif 'max' in chip_info:
|
2221 |
+
detected_gpu_name = "M3 Max"
|
2222 |
+
elif 'pro' in chip_info:
|
2223 |
+
detected_gpu_name = "M3 Pro"
|
2224 |
+
else:
|
2225 |
+
detected_gpu_name = "M3"
|
2226 |
+
elif 'm4' in chip_info:
|
2227 |
+
if 'ultra' in chip_info:
|
2228 |
+
detected_gpu_name = "M4 Ultra"
|
2229 |
+
elif 'max' in chip_info:
|
2230 |
+
detected_gpu_name = "M4 Max"
|
2231 |
+
elif 'pro' in chip_info:
|
2232 |
+
detected_gpu_name = "M4 Pro"
|
2233 |
+
else:
|
2234 |
+
detected_gpu_name = "M4"
|
2235 |
+
|
2236 |
+
logger.info(f"🔍 DEBUG: Detected Apple chip: {detected_gpu_name}")
|
2237 |
+
except Exception as e:
|
2238 |
+
logger.warning(f"🔍 DEBUG: Could not detect Apple chip: {e}")
|
2239 |
+
detected_gpu_name = "Apple Silicon GPU"
|
2240 |
+
|
2241 |
+
gpu_vendor = "Apple Silicon"
|
2242 |
+
elif detected_gpu_name:
|
2243 |
+
gpu_lower = detected_gpu_name.lower()
|
2244 |
+
if 'rtx' in gpu_lower and ('geforce' in gpu_lower or '30' in gpu_lower or '40' in gpu_lower or '50' in gpu_lower):
|
2245 |
+
gpu_vendor = "NVIDIA Consumer (GeForce RTX)"
|
2246 |
+
elif 'rtx' in gpu_lower and ('a' in gpu_lower or 'quadro' in gpu_lower):
|
2247 |
+
gpu_vendor = "NVIDIA Professional (RTX A-Series)"
|
2248 |
+
elif any(x in gpu_lower for x in ['h100', 'a100', 'v100', 'l40', 'l4']):
|
2249 |
+
gpu_vendor = "NVIDIA Data Center"
|
2250 |
+
elif 'm1' in gpu_lower or 'm2' in gpu_lower or 'm3' in gpu_lower or 'm4' in gpu_lower or 'apple' in gpu_lower:
|
2251 |
+
gpu_vendor = "Apple Silicon"
|
2252 |
+
elif 'radeon' in gpu_lower or 'rx ' in gpu_lower or 'instinct' in gpu_lower:
|
2253 |
+
gpu_vendor = "AMD"
|
2254 |
+
elif 'arc' in gpu_lower:
|
2255 |
+
gpu_vendor = "Intel"
|
2256 |
+
|
2257 |
+
logger.info(f"🔍 DEBUG: Final values - Platform: {detected_platform}, GPU Vendor: {gpu_vendor}, GPU Name: {detected_gpu_name}, Unified RAM: {detected_unified_ram}GB")
|
2258 |
+
|
2259 |
+
return detected_platform, gpu_vendor, detected_gpu_name, detected_vram, detected_ram, detected_unified_ram
|
2260 |
+
|
2261 |
+
except Exception as e:
|
2262 |
+
logger.warning(f"Auto hardware detection failed: {e}")
|
2263 |
+
# Return defaults if detection fails
|
2264 |
+
return "Linux", "Custom (Manual Input)", "", 8, 16, 16
|
2265 |
+
|
2266 |
+
# Auto hardware detection on startup (if enabled)
|
2267 |
+
if auto_hardware_detection:
|
2268 |
+
def populate_hardware_on_load():
|
2269 |
+
"""Populate hardware fields with auto-detected values on load."""
|
2270 |
+
detected_platform, detected_gpu_vendor, detected_gpu_name, detected_vram, detected_ram, detected_unified_ram = auto_detect_hardware()
|
2271 |
+
|
2272 |
+
# Handle Apple Silicon field visibility
|
2273 |
+
if detected_platform == "Apple Silicon":
|
2274 |
+
vram_visible = False
|
2275 |
+
ram_visible = False
|
2276 |
+
unified_visible = True
|
2277 |
+
gpu_series_visible = False
|
2278 |
+
else:
|
2279 |
+
vram_visible = True
|
2280 |
+
ram_visible = True
|
2281 |
+
unified_visible = False
|
2282 |
+
gpu_series_visible = True
|
2283 |
+
|
2284 |
+
# Update all hardware fields
|
2285 |
+
updates = [
|
2286 |
+
gr.update(value=detected_platform), # platform
|
2287 |
+
gr.update(value=detected_gpu_vendor), # gpu_vendor
|
2288 |
+
gr.update(value=detected_gpu_name), # gpu_name
|
2289 |
+
gr.update(value=detected_vram, visible=vram_visible), # vram_gb
|
2290 |
+
gr.update(value=detected_ram, visible=ram_visible), # ram_gb
|
2291 |
+
gr.update(value=detected_unified_ram, visible=unified_visible), # unified_ram_gb
|
2292 |
+
gr.update(visible=gpu_series_visible), # gpu_series
|
2293 |
+
]
|
2294 |
+
|
2295 |
+
return updates
|
2296 |
+
|
2297 |
+
interface.load(
|
2298 |
+
populate_hardware_on_load,
|
2299 |
+
outputs=[platform, gpu_vendor, gpu_name, vram_gb, ram_gb, unified_ram_gb, gpu_series]
|
2300 |
+
)
|
2301 |
+
|
2302 |
# Load initial accordion titles on startup
|
2303 |
interface.load(
|
2304 |
update_hardware_accordion,
|
2305 |
+
inputs=[platform, gpu_vendor, gpu_model, vram_gb, ram_gb, unified_ram_gb],
|
2306 |
outputs=hardware_accordion
|
2307 |
)
|
2308 |
|
|
|
2315 |
# Accordion title update event handlers
|
2316 |
|
2317 |
# Hardware accordion title updates
|
2318 |
+
for component in [platform, gpu_vendor, gpu_model, vram_gb, ram_gb, unified_ram_gb]:
|
2319 |
component.change(
|
2320 |
update_hardware_accordion,
|
2321 |
+
inputs=[platform, gpu_vendor, gpu_model, vram_gb, ram_gb, unified_ram_gb],
|
2322 |
outputs=hardware_accordion
|
2323 |
)
|
2324 |
|
|
|
2413 |
|
2414 |
return '\n'.join(filtered_lines)
|
2415 |
|
2416 |
+
def generate_with_combined_gpu_name(gpu_vendor, gpu_series, gpu_model, vram_gb, ram_gb, platform, model_name, dtype_selection, width, height, inference_steps, unified_ram_gb=None):
|
2417 |
"""Generate code with the correct GPU name from multi-level selection, including memory analysis."""
|
2418 |
final_gpu_name = get_final_gpu_name(gpu_vendor, gpu_series, gpu_model)
|
2419 |
|
2420 |
# Constant prompt text
|
2421 |
prompt_text = "A cat holding a sign that says hello world"
|
2422 |
|
2423 |
+
# Handle Apple Silicon unified memory
|
2424 |
+
effective_vram = vram_gb
|
2425 |
+
effective_ram = ram_gb
|
2426 |
+
if platform == "Apple Silicon" and unified_ram_gb:
|
2427 |
+
# For Apple Silicon, unified memory serves as both VRAM and RAM
|
2428 |
+
effective_vram = unified_ram_gb
|
2429 |
+
effective_ram = unified_ram_gb
|
2430 |
+
|
2431 |
# STEP 1: Perform memory analysis BEFORE code generation
|
2432 |
memory_analysis_data = None
|
2433 |
memory_header = ""
|
2434 |
|
2435 |
try:
|
2436 |
+
if model_name and effective_vram and effective_vram > 0:
|
2437 |
+
memory_info, recommendations, _ = app.analyze_model_memory(model_name, effective_vram)
|
2438 |
|
2439 |
# Package memory analysis for Gemini API
|
2440 |
memory_analysis_data = {
|
|
|
2471 |
|
2472 |
optimization_strategy = get_optimization_strategy(recommendations)
|
2473 |
|
2474 |
+
if platform == "Apple Silicon":
|
2475 |
+
memory_header = f"""# Memory Analysis for {model_name}:
|
2476 |
+
# Platform: Apple Silicon with {unified_ram_gb}GB Unified Memory (shared CPU/GPU)
|
2477 |
+
# Model Memory Requirements: {memory_info.get('estimated_inference_memory_fp16_gb', 'Unknown')} GB
|
2478 |
+
# Recommendation: {', '.join(recommendations.get('recommendations', ['N/A']))}
|
2479 |
+
# Optimization Strategy: {optimization_strategy}
|
2480 |
+
|
2481 |
+
"""
|
2482 |
+
else:
|
2483 |
+
memory_header = f"""# Memory Analysis for {model_name}:
|
2484 |
+
# GPU: {final_gpu_name if final_gpu_name else 'Not specified'} ({effective_vram}GB VRAM)
|
2485 |
# Model Memory Requirements: {memory_info.get('estimated_inference_memory_fp16_gb', 'Unknown')} GB
|
2486 |
# Recommendation: {', '.join(recommendations.get('recommendations', ['N/A']))}
|
2487 |
# Optimization Strategy: {optimization_strategy}
|
|
|
2495 |
"""
|
2496 |
|
2497 |
# STEP 2: Get the actual prompt that will be sent to Gemini
|
2498 |
+
# Create hardware specs based on platform
|
2499 |
+
if platform == "Apple Silicon":
|
2500 |
+
hardware_specs = {
|
2501 |
+
'platform': 'Darwin', # Use Darwin for backend compatibility
|
2502 |
+
'architecture': 'arm64',
|
2503 |
+
'cpu_count': 8,
|
2504 |
+
'python_version': '3.11',
|
2505 |
+
'cuda_available': False,
|
2506 |
+
'mps_available': True, # Apple Silicon supports MPS
|
2507 |
+
'torch_version': '2.0+',
|
2508 |
+
'manual_input': True,
|
2509 |
+
'ram_gb': int(effective_ram) if effective_ram else 16,
|
2510 |
+
'unified_memory': True, # Special flag for Apple Silicon
|
2511 |
+
'unified_memory_gb': int(unified_ram_gb) if unified_ram_gb else 16,
|
2512 |
+
'user_dtype': f"torch.{dtype_selection}" if dtype_selection != "Auto" else None,
|
2513 |
+
'gpu_info': None # Apple Silicon doesn't report discrete GPU info
|
2514 |
+
}
|
2515 |
+
else:
|
2516 |
+
hardware_specs = {
|
2517 |
'platform': platform,
|
2518 |
'architecture': 'manual_input',
|
2519 |
'cpu_count': 8,
|
|
|
2522 |
'mps_available': platform == 'Darwin' and 'apple' in final_gpu_name.lower() if final_gpu_name else False,
|
2523 |
'torch_version': '2.0+',
|
2524 |
'manual_input': True,
|
2525 |
+
'ram_gb': int(effective_ram) if effective_ram else 16,
|
2526 |
'user_dtype': f"torch.{dtype_selection}" if dtype_selection != "Auto" else None,
|
2527 |
+
'gpu_info': [{'name': final_gpu_name, 'memory_mb': int(effective_vram) * 1024}] if final_gpu_name and effective_vram else None
|
2528 |
+
}
|
2529 |
+
|
2530 |
+
actual_prompt = app.get_generation_prompt(
|
2531 |
+
model_name, prompt_text, (int(height), int(width)), int(inference_steps),
|
2532 |
+
hardware_specs,
|
2533 |
optimization_profile="balanced"
|
2534 |
)
|
2535 |
|
2536 |
# STEP 3: Generate the optimized code WITH memory analysis information
|
2537 |
generated_code = app.generate_code_with_manual_specs(
|
2538 |
+
final_gpu_name, effective_vram, effective_ram, platform,
|
2539 |
model_name, prompt_text, dtype_selection, width, height, inference_steps,
|
2540 |
memory_analysis_data
|
2541 |
)
|
|
|
2553 |
full_code_storage = gr.State(value="")
|
2554 |
prompt_storage = gr.State(value="")
|
2555 |
|
2556 |
+
def generate_and_store_code(gpu_vendor, gpu_series, gpu_model, vram_gb, ram_gb, platform, model_name, dtype_selection, width, height, inference_steps, unified_ram_gb):
|
2557 |
"""Generate code and return code for display and full code for storage."""
|
2558 |
summary, full_code, actual_prompt = generate_with_combined_gpu_name(
|
2559 |
gpu_vendor, gpu_series, gpu_model, vram_gb, ram_gb, platform,
|
2560 |
+
model_name, dtype_selection, width, height, inference_steps, unified_ram_gb
|
2561 |
)
|
2562 |
# Show the view prompt button after generation and close accordions
|
2563 |
return (full_code, full_code, False, gr.update(visible=True), actual_prompt,
|
|
|
2576 |
generate_and_store_code,
|
2577 |
inputs=[
|
2578 |
gpu_vendor, gpu_series, gpu_model, vram_gb, ram_gb, platform,
|
2579 |
+
model_name, dtype_selection, width, height, inference_steps, unified_ram_gb
|
2580 |
],
|
2581 |
outputs=[code_output, full_code_storage, code_collapsed, view_prompt_btn, prompt_storage, hardware_accordion, model_accordion]
|
2582 |
)
|
|
|
2593 |
hide_modal,
|
2594 |
outputs=[prompt_modal, prompt_display]
|
2595 |
)
|
2596 |
+
|
2597 |
|
2598 |
|
2599 |
|
|
|
2602 |
|
2603 |
def main():
|
2604 |
"""Launch the Gradio application."""
|
2605 |
+
parser = argparse.ArgumentParser(description="Auto-Diffusers Code Generator")
|
2606 |
+
parser.add_argument(
|
2607 |
+
"--auto-hardware-detection",
|
2608 |
+
action="store_true",
|
2609 |
+
default=False,
|
2610 |
+
help="Enable automatic hardware detection on startup"
|
2611 |
+
)
|
2612 |
+
parser.add_argument(
|
2613 |
+
"--no-auto-hardware-detection",
|
2614 |
+
action="store_true",
|
2615 |
+
help="Disable automatic hardware detection on startup (default: True)"
|
2616 |
+
)
|
2617 |
+
parser.add_argument(
|
2618 |
+
"--port",
|
2619 |
+
type=int,
|
2620 |
+
default=7860,
|
2621 |
+
help="Port to run the server on (default: 7860)"
|
2622 |
+
)
|
2623 |
+
parser.add_argument(
|
2624 |
+
"--host",
|
2625 |
+
type=str,
|
2626 |
+
default="0.0.0.0",
|
2627 |
+
help="Host to run the server on (default: 0.0.0.0)"
|
2628 |
+
)
|
2629 |
+
parser.add_argument(
|
2630 |
+
"--no-share",
|
2631 |
+
action="store_true",
|
2632 |
+
help="Disable public sharing of the interface"
|
2633 |
+
)
|
2634 |
+
|
2635 |
+
args = parser.parse_args()
|
2636 |
+
|
2637 |
+
# Handle auto hardware detection flag
|
2638 |
+
auto_hardware_detection = args.auto_hardware_detection and not args.no_auto_hardware_detection
|
2639 |
+
|
2640 |
+
# Log startup configuration
|
2641 |
+
print(f"🚀 Starting Auto-Diffusers Code Generator...")
|
2642 |
+
print(f" Host: {args.host}")
|
2643 |
+
print(f" Port: {args.port}")
|
2644 |
+
print(f" Share: {not args.no_share}")
|
2645 |
+
print(f" Auto Hardware Detection: {'Enabled' if auto_hardware_detection else 'Disabled'}")
|
2646 |
+
|
2647 |
try:
|
2648 |
+
interface = create_gradio_interface(auto_hardware_detection=auto_hardware_detection)
|
2649 |
interface.launch(
|
2650 |
+
server_name=args.host,
|
2651 |
+
server_port=args.port,
|
2652 |
+
share=not args.no_share,
|
2653 |
show_error=True
|
2654 |
)
|
2655 |
except Exception as e:
|