Revert "Fix chat completions response handling"
This reverts commit 02593e56d0.
This commit is contained in:
parent
22f293b4d6
commit
07db57d106
4 changed files with 7 additions and 66 deletions
|
|
@ -15,7 +15,7 @@ Classification requests ask the model for structured JSON output with a required
|
|||
- **Configurable endpoint** – set the classification service base URL on the options page.
|
||||
- **Model selection** – load available models from the endpoint and choose one (or omit the model field).
|
||||
- **Optional OpenAI auth headers** – provide an API key plus optional organization/project headers when needed.
|
||||
- **Request formats** – built-in formats use native chat messages; a custom format can still send one templated user message when needed.
|
||||
- **Request formats** – use native OpenAI chat messages or choose Qwen, Mistral, Harmony (gpt-oss), or a custom templated message format.
|
||||
- **Custom system prompts** – tailor the instructions sent to the model for more precise results.
|
||||
- **Persistent result caching** – classification results and reasoning are saved to disk so messages aren't re-evaluated across restarts.
|
||||
- **Advanced parameters** – tune generation settings like temperature and top‑p from the options page. Unsupported OpenAI sampling fields are filtered out automatically.
|
||||
|
|
|
|||
Loading…
Add table
Add a link
Reference in a new issue