Exported with
from optimum.neuron import NeuronFluxPipeline
if __name__ == "__main__":
compiler_args = {"auto_cast": "none"}
input_shapes = {"batch_size": 1, "height": 1024, "width": 1024}
pipe = NeuronFluxPipeline.from_pretrained(
"black-forest-labs/FLUX.1-dev",
torch_dtype=torch.bfloat16,
export=True,
tensor_parallel_size=8,
# disable_neuron_cache=True,
**compiler_args,
**input_shapes
)
# Save locally
pipe.save_pretrained("flux_dev_neuron_1024_tp8/")
# Upload to the HuggingFace Hub
pipe.push_to_hub(
"flux_dev_neuron_1024_tp8/", repository_id="Jingya/FLUX.1-dev-neuronx-1024x1024-tp8" # Replace with your HF Hub repo id
)
Inference Providers
NEW
This model isn't deployed by any Inference Provider.
๐
Ask for provider support