openai>=1.0.0
httpx>=0.27.0
pyyaml>=6.0
keyring>=25.0
cryptography>=42.0
pystray>=0.19
pyperclip>=1.9
packaging>=24.0
Pillow>=10.0
tiktoken>=0.7.0
pyqtgraph>=0.13.0
darkdetect>=0.8.0
mss>=9.0
# Optional: llama-cpp-python for in-process GGUF inference (no network, CSCA-safe)
# Install with CUDA: pip install llama-cpp-python --extra-index-url https://abetlen.github.io/llama-cpp-python/whl/cu124
# llama-cpp-python>=0.3.0
