Model Details

Model Description

Repository

Usage

Method 1: llama.cpp Backend Server + Chatbox

Step 1: Start .llama.cpp Server

./llama-server \
  -m /path/to/model.gguf \
  -c 2048 \          # Context length
  --host 0.0.0.0 \   # Allow remote connections
  --port 8080 \      # Server port
  --n-gpu-layers 35  # GPU acceleration (if available)

Step 2: Connect via Chatbox

  1. Download Chatbox
  2. Configure API endpoint:
    API URL: http://localhost:8080
    Model: (leave empty)
    API Type: llama.cpp
    
  3. Set generation parameters:
    {
      "temperature": 0.7,
      "max_tokens": 512,
      "top_p": 0.9
    }
    

Method 2: LM Studio

  1. Download LM Studio
  2. Load GGUF file:
    • Launch LM Studio
    • Search Slipstream-Max/Emollm-InternLM2.5-7B-chat-GGUF-fp16
  3. Configure settings:
    Context Length: 2048
    GPU Offload: Recommended (enable if available)
    Batch Size: 512
    
  4. Start chatting through the built-in UI

Precision Details

Filename Precision Size Characteristics
CPsyCounX.gguf FP16 [15.5GB] Full original model precision

Hardware Requirements

Minimum:

  • 24GB RAM (for 7B model)
  • CPU with AVX/AVX2 instruction set support

Recommended:

  • 32GB RAM
  • CUDA-capable GPU (for acceleration)
  • Fast SSD storage (due to large model size)

Key Notes

  1. Requires latest llama.cpp (v3+ recommended)
  2. Use --n-gpu-layers 35 for GPU acceleration (requires CUDA-enabled build)
  3. Initial loading takes longer (2-5 minutes)
  4. Requires more memory/storage than quantized versions
  5. Use --mlock to prevent swapping

Advantages

  • Preserves original model precision
  • Ideal for precision-sensitive applications
  • No quantization loss
  • Suitable for continued fine-tuning
Downloads last month
25
GGUF
Model size
7.74B params
Architecture
internlm2
Hardware compatibility
Log In to view the estimation
Inference Providers NEW
This model isn't deployed by any Inference Provider. 🙋 Ask for provider support

Model tree for Slipstream-Max/CPsyCounX-InternLM2-Chat-7B-GGUF-fp16

Quantized
(27)
this model

Dataset used to train Slipstream-Max/CPsyCounX-InternLM2-Chat-7B-GGUF-fp16