中文版模型说明

image/png

Introduction

Neta Lumina is a high‑quality anime‑style image‑generation model developed by Neta.art Lab.
Building on the open‑source Lumina‑Image‑2.0 released by the Alpha‑VLLM team at Shanghai AI Laboratory, we fine‑tuned the model with a vast corpus of high‑quality anime images and multilingual tag data. The preliminary result is a compelling model with powerful comprehension and interpretation abilities (thanks to Gemma text encoder), ideal for illustration, posters, storyboards, character design, and more.

Key Features

  • Optimized for diverse creative scenarios such as Furry, Guofeng (traditional‑Chinese aesthetics), pets, etc.
  • Wide coverage of characters and styles, from popular to niche concepts. (Still support danbooru tags!)
  • Accurate natural‑language understanding with excellent adherence to complex prompts.
  • Native multilingual support, with Chinese, English, and Japanese recommended first.

Model Versions

For models in alpha tests, requst access at https://huggingface.co/neta-art/NetaLumina_Alpha if you are interested. We will keep updating.

neta-lumina-beta-0624-raw

  • Primary Goal: General knowledge and anime‑style optimization
  • Data Set: >13 million anime‑style images
  • >46,000 A100 Hours
  • Higher upper limit, suitable for pro users. Check Neta Lumina Prompt Book for better results.

neta-lumina-beta-0624-aes

  • First beta release candidate
  • Primary Goal: Enhanced aesthetics, pose accuracy, and scene detail
  • Data Set: Hundreds of thousands of handpicked high‑quality anime images (fine‑tuned on an older version of raw model)
  • User-friendly, suitable for most people.

How  to  Use

Try it at Hugging Face playground

ComfyUI

Neta Lumina is built on the Lumina2 Diffusion Transformer (DiT) framework, please follow these steps precisely.

Environment Requirements

Currently Neta Lumina runs only on ComfyUI:

  • Latest ComfyUI installation
  • ≥ 8 GB VRAM

Downloads & Installation

Original (component) release

  1. Neta Lumina-Beta
  2. Text Encoder (Gemma-2B)
  3. VAE Model (16-Channel FLUX VAE)

Workflow: load lumina_workflow.json in ComfyUI.

image/png

  • UNETLoader – loads the .pth
  • VAELoader – loads ae.safetensors
  • CLIPLoader – loads gemma_2_2b_fp16.safetensors
  • Text Encoder – connects positive /negative prompts to K Sampler

Simple merged release
Download neta-lumina-beta-0624-all-in-one.safetensors,
md5sum = dca54fef3c64e942c1a62a741c4f9d8a,
you may use ComfyUI’s simple checkpoint loader workflow.

Recommended Settings

  • Sampler: res_multistep
  • Scheduler: linear_quadratic
  • Steps: 30
  • CFG (guidance): 4 – 5.5
  • EmptySD3LatentImage resolution: 1024 × 1024, 768 × 1532, or 968 × 1322

Prompt Book

Detailed prompt guidelines: Neta Lumina Prompt Book

Community

Roadmap

Model

  • Continous base‑model training to raise reasoning capability.
  • Aesthetic‑dataset iteration to improve anatomy, background richness, and overall appealness.
  • Smarter, more versatile tagging tools to lower the creative barrier.

Ecosystem

  • LoRA training tutorials and components
    • Experienced users may already fine‑tune via Lumina‑Image‑2.0’s open code.
  • Development of advanced control / style‑consistency features (e.g., Omini Control). Call for Collaboration!

License & Disclaimer

Participants & Contributors

Community Contributors

Evaluators & developers: 二小姐, spawner, Rnglg2
Other contributors: 沉迷摸鱼, poi氵, ashan, 十分无奈, GHOSTLXH, wenaka, iiiiii, 年糕特工队, 恩匹希, 奶冻美宣集, mumu, yizyin, smile

Appendix & Resources

Downloads last month

-

Downloads are not tracked for this model. How to track
Inference Providers NEW
This model isn't deployed by any Inference Provider. 🙋 Ask for provider support

Model tree for neta-art/Neta-Lumina

Finetuned
(6)
this model
Merges
1 model

Space using neta-art/Neta-Lumina 1

Collection including neta-art/Neta-Lumina