LLM Quantization (Ollama, LM Studio): Any Performance Drop? TEST | code_your_own_AI | Podwise