openwebui/setup.sh
MyeonghoeLee 1597a9140f setup.sh VRAM 기반 판단 로직 추가 및 완전 자동화
- NVIDIA GPU VRAM 20GB 미만이면 vLLM 대신 Ollama 선택
- 사용자 확인 단계 제거, 감지 후 바로 실행

Co-Authored-By: Claude Opus 4.6 (1M context) <noreply@anthropic.com>
2026-03-26 15:13:53 +09:00

110 lines
3.5 KiB
Bash
Executable File

#!/bin/bash
set -e
#====================================================================
# Qwen3.5 + Open WebUI 통합 셋업
# 시스템 환경을 자동 감지하여 최적의 방식을 선택합니다.
#
# Apple Silicon + RAM 32GB↑ → MLX (vllm-mlx)
# Apple Silicon + RAM 부족 → Ollama
# NVIDIA GPU → vLLM (Docker)
# 그 외 → Ollama
#====================================================================
PROJECT_DIR="$(cd "$(dirname "$0")" && pwd)"
echo "============================================"
echo " Qwen3.5 + Open WebUI 통합 셋업"
echo " 시스템 환경을 자동 감지합니다..."
echo "============================================"
echo ""
#--------------------------------------------------------------------
# 시스템 감지
#--------------------------------------------------------------------
OS=$(uname -s)
ARCH=$(uname -m)
HAS_APPLE_SILICON=false
HAS_NVIDIA_GPU=false
RAM_GB=0
# Apple Silicon 감지
if [ "$OS" = "Darwin" ] && [ "$ARCH" = "arm64" ]; then
if sysctl -n machdep.cpu.brand_string 2>/dev/null | grep -q "Apple"; then
HAS_APPLE_SILICON=true
RAM_GB=$(sysctl -n hw.memsize | awk '{printf "%.0f", $1/1024/1024/1024}')
fi
fi
# Linux RAM 감지
if [ "$OS" = "Linux" ]; then
RAM_GB=$(free -g | awk '/^Mem:/{print $2}')
fi
# NVIDIA GPU 감지 + VRAM 확인
VRAM_GB=0
if command -v nvidia-smi &>/dev/null; then
if nvidia-smi &>/dev/null; then
HAS_NVIDIA_GPU=true
VRAM_MB=$(nvidia-smi --query-gpu=memory.total --format=csv,noheader,nounits 2>/dev/null | head -1)
VRAM_GB=$((VRAM_MB / 1024))
fi
fi
echo " OS: $OS ($ARCH)"
echo " Apple Silicon: $HAS_APPLE_SILICON"
echo " NVIDIA GPU: $HAS_NVIDIA_GPU"
if [ "$HAS_NVIDIA_GPU" = true ]; then
echo " VRAM: ${VRAM_GB}GB"
fi
echo " RAM: ${RAM_GB}GB"
echo ""
#--------------------------------------------------------------------
# 방식 결정
# - NVIDIA GPU + VRAM 20GB 이상 → vLLM
# - NVIDIA GPU + VRAM 부족 → Ollama (GPU로는 모델 못 올림)
# - Apple Silicon + RAM 32GB↑ → MLX
# - Apple Silicon + RAM 부족 → Ollama
# - 그 외 → Ollama
#--------------------------------------------------------------------
SELECTED=""
if [ "$HAS_NVIDIA_GPU" = true ] && [ "$VRAM_GB" -ge 20 ]; then
SELECTED="vllm"
echo " → NVIDIA GPU (VRAM ${VRAM_GB}GB) → vLLM 선택"
elif [ "$HAS_NVIDIA_GPU" = true ] && [ "$VRAM_GB" -lt 20 ]; then
SELECTED="ollama"
echo " → NVIDIA GPU 감지되었으나 VRAM ${VRAM_GB}GB (20GB 미만) → Ollama 선택"
elif [ "$HAS_APPLE_SILICON" = true ] && [ "$RAM_GB" -ge 32 ]; then
SELECTED="mlx"
echo " → Apple Silicon + RAM ${RAM_GB}GB → MLX (vllm-mlx) 선택"
elif [ "$HAS_APPLE_SILICON" = true ]; then
SELECTED="ollama"
echo " → Apple Silicon + RAM ${RAM_GB}GB (32GB 미만) → Ollama 선택"
else
SELECTED="ollama"
echo " → 범용 환경 → Ollama 선택"
fi
echo ""
echo "============================================"
echo " $SELECTED 방식으로 셋업을 시작합니다"
echo "============================================"
echo ""
#--------------------------------------------------------------------
# 선택된 스크립트 실행
#--------------------------------------------------------------------
case "$SELECTED" in
ollama)
exec "$PROJECT_DIR/setup-ollama.sh"
;;
mlx)
exec "$PROJECT_DIR/setup-mlx.sh"
;;
vllm)
exec "$PROJECT_DIR/setup-vllm.sh"
;;
esac