cthomas
|
e72f0f68ef
|
fix: gracefully handle too long responses from llm provider (#2677)
|
2025-06-06 13:13:32 -07:00 |
|
Andy Li
|
80f6e97ca9
|
feat: otel metrics and expanded collecting (#2647)
(passed tests in last run)
|
2025-06-05 17:20:14 -07:00 |
|
Sarah Wooders
|
729414e7a9
|
feat: concurrently make embedding request and use async client for OpenAI (#2482)
Co-authored-by: Matthew Zhou <mattzh1314@gmail.com>
|
2025-05-28 11:35:22 -07:00 |
|
Andy Li
|
0a54b998a8
|
feat: track llm provider traces and tracking steps in async agent loop (#2219)
|
2025-05-19 15:50:56 -07:00 |
|
cthomas
|
d8151d76e2
|
feat: add provider_category field to distinguish byok (#2038)
|
2025-05-06 17:31:36 -07:00 |
|
cthomas
|
9280d85ba4
|
feat: always add user id to openai requests (#1969)
|
2025-04-30 23:23:01 -07:00 |
|
cthomas
|
25c188689a
|
feat: byok 2.0 (#1963)
|
2025-04-30 21:26:50 -07:00 |
|
cthomas
|
7605675f23
|
feat: add llm config per request (#1866)
|
2025-04-23 16:37:05 -07:00 |
|
Matthew Zhou
|
68fbcf33d8
|
feat: Finish step_until_request in new batch agent loop (#1656)
|
2025-04-10 10:19:06 -07:00 |
|
Matthew Zhou
|
6a2312a02d
|
chore: Inject LLM config directly to batch api request func (#1652)
|
2025-04-09 15:56:54 -07:00 |
|
Matthew Zhou
|
7565ad4a8d
|
feat: Write batch request on base LLM client (#1646)
|
2025-04-09 14:58:26 -07:00 |
|
Matthew Zhou
|
59ca61da58
|
feat: Simplify arguments for LLM clients (#1536)
|
2025-04-02 14:26:27 -07:00 |
|
cthomas
|
df899ee05b
|
fix: anthropic system message parse (#1467)
|
2025-03-30 18:44:55 -07:00 |
|
Matthew Zhou
|
60ffc9e8ec
|
fix: Fix message_id ordering in agent serialization (#1458)
|
2025-03-28 15:13:33 -07:00 |
|
cthomas
|
78e0641055
|
feat: anthropic class improvements (#1425)
|
2025-03-27 08:47:54 -07:00 |
|
cthomas
|
ea5113489c
|
chore: migrate anthropic to llm client class (#1409)
|
2025-03-26 09:37:27 -07:00 |
|
cthomas
|
6a20f85189
|
feat: add new llm client framework and migrate google apis (#1209)
|
2025-03-07 16:34:06 -08:00 |
|