Ovarra-v1 β€” Business Planning Assistant (Fine-tuned Mistral 7B)

Ovarra-v1 is a fine-tuned version of Mistral-7B-v0.1, designed to help users plan, clarify, and launch AI-powered business ideas.

This model acts like a startup co-pilot, guiding users from rough ideas to concrete action steps β€” with AI tools, strategy, and marketing guidance embedded in every response.


πŸ“Œ Model Details

Attribute Value
Base model mistralai/Mistral-7B-v0.1
Finetuned on ~1500 GPT-4-generated prompt-response pairs, validated by human reviews of the best AI tools for each sector.
Format Instruction-tuned with [INST] ... [/INST]
Finetuning method QLoRA (4-bit), Flash Attention 2
Trainer Axolotl + RunPod (A40 GPU)

Training Objective

The model was trained to:

  • Ask clarifying questions when input is vague
  • Plan structured roadmaps (4–6 steps)
  • Recommend relevant AI tools for each stage
  • Support follow-up questions with contextual awareness

Training categories included:

  • AI tool integration
  • Product planning
  • MVP prototyping
  • Marketing strategy
  • Startup operations

Example Usage

from transformers import AutoTokenizer, AutoModelForCausalLM

model = AutoModelForCausalLM.from_pretrained("vitalune/ovarra-v1", device_map="auto")
tokenizer = AutoTokenizer.from_pretrained("vitalune/ovarra-v1")

prompt = "[INST] I want to launch an AI copywriting SaaS. Help me plan it. [/INST]"
inputs = tokenizer(prompt, return_tensors="pt").to(model.device)

output = model.generate(**inputs, max_new_tokens=512)
print(tokenizer.decode(output[0], skip_special_tokens=True))

Credits

  • Fine-tuned and built by @vitalune
  • Based on prompt logic and data generation co-created using GPT-4 API
  • Hosted and accelerated via RunPod
Downloads last month
10
Safetensors
Model size
7.24B params
Tensor type
BF16
Β·
Inference Providers NEW
This model isn't deployed by any Inference Provider. πŸ™‹ Ask for provider support

Model tree for vitalune/ovarra-v1

Finetuned
(965)
this model

Dataset used to train vitalune/ovarra-v1