ZIm/crates/language_models/src
Max Brunsfeld 04772bf17d
Add support for queuing status updates in cloud language model provider (#29818)
This sets us up to display queue position information to the user, once
our language model backend is updated to support request queuing.

The JSON returned by the LLM backend will need to look like this:

```json
{"queue": {"status": "queued", "position": 1}}
{"queue": {"status": "started"}}
{"event": {"THE_UPSTREAM_MODEL_PROVIDER_EVENT": "..."}} 
```

Release Notes:

- N/A

---------

Co-authored-by: Marshall Bowers <git@maxdeviant.com>
2025-05-02 20:36:39 +00:00
..
provider Add support for queuing status updates in cloud language model provider (#29818) 2025-05-02 20:36:39 +00:00
ui chore: Bump Rust edition to 2024 (#27800) 2025-03-31 20:55:27 +02:00
language_models.rs language_models: Remove language-models feature flag (#29416) 2025-04-25 14:18:48 +00:00
provider.rs assistant: Add Bedrock support (#21092) 2025-02-24 18:10:12 -05:00
settings.rs VSCode Settings import (#29018) 2025-04-23 20:54:09 +00:00
ui.rs assistant: Refine settings view's instruction visuals (#25812) 2025-02-28 12:06:47 -03:00