App Version
3.16.6
API Provider
LM Studio
Model Used
Qwen3-32B
🔁 Steps to Reproduce
I am trying to use a local Qwen3-32B model via llama.cpp. To do so, I use the LMStudio integration that I point to the local server. Everything works fine, but after 10 minutes (600 seconds), the connection is dropped and I get an API Request Failed message. The inference is on cpu and quite slow, but I would be happy to let it crunch while I'm doing something else. If I use the tiny Qwen3 0.6B model, the inference is fast enough and everything works as expected (although with very mediocre results).
When it fails, llama.cpp finishes processing the prompt anyway. It succeeds on retry, the prompt being already cached.
💥 Outcome Summary (Optional)
No response
📄 Relevant Logs or Errors
App Version
3.16.6
API Provider
LM Studio
Model Used
Qwen3-32B
🔁 Steps to Reproduce
I am trying to use a local Qwen3-32B model via llama.cpp. To do so, I use the LMStudio integration that I point to the local server. Everything works fine, but after 10 minutes (600 seconds), the connection is dropped and I get an API Request Failed message. The inference is on cpu and quite slow, but I would be happy to let it crunch while I'm doing something else. If I use the tiny Qwen3 0.6B model, the inference is fast enough and everything works as expected (although with very mediocre results).
When it fails, llama.cpp finishes processing the prompt anyway. It succeeds on retry, the prompt being already cached.
💥 Outcome Summary (Optional)
No response
📄 Relevant Logs or Errors