cthomas
|
a9771424fe
|
fix: duplicate message stream error (#3834)
|
2025-08-11 14:27:35 -07:00 |
|
cthomas
|
cab621ac14
|
feat: continue stream processing on client cancel (#3796)
|
2025-08-07 13:17:36 -07:00 |
|
Andy Li
|
76679e3ecc
|
feat: track metrics for runs in db
|
2025-08-06 15:46:50 -07:00 |
|
cthomas
|
f3316a0d8a
|
feat: log stream cancellation to sentry (#3759)
|
2025-08-05 16:07:30 -07:00 |
|
jnjpng
|
7f12230c21
|
fix: manually count tokens for streaming lmstudio models
Co-authored-by: Jin Peng <jinjpeng@Jins-MacBook-Pro.local>
Co-authored-by: Charles Packer <packercharles@gmail.com>
|
2025-07-29 18:12:42 -07:00 |
|
Andy Li
|
f9bb757a98
|
feat: support for agent loop job cancelation (#2837)
|
2025-07-02 14:31:16 -07:00 |
|
Kevin Lin
|
efae2231ea
|
feat: add omitted reasoning to streaming openai reasoning (#2846)
Co-authored-by: Charles Packer <packercharles@gmail.com>
|
2025-06-24 18:47:38 -07:00 |
|
Sarah Wooders
|
78e81391ba
|
fix: avoid calling model_dump on stop reason messages twice (#2811)
|
2025-06-13 18:25:35 -07:00 |
|
cthomas
|
97986b0f03
|
feat: send stop reason in letta APIs (#2789)
|
2025-06-13 16:04:48 -07:00 |
|
Andy Li
|
bf7bd3b905
|
fix: metric tracking (#2785)
|
2025-06-13 13:53:10 -07:00 |
|
cthomas
|
ccb203b264
|
feat: consolidate logic for finish tokens (#2779)
|
2025-06-12 15:24:06 -07:00 |
|
Kevin Lin
|
78dad0db7a
|
fix: patch reasoning models (#2703)
Co-authored-by: Charles Packer <packercharles@gmail.com>
|
2025-06-11 17:20:04 -07:00 |
|
Andy Li
|
80f6e97ca9
|
feat: otel metrics and expanded collecting (#2647)
(passed tests in last run)
|
2025-06-05 17:20:14 -07:00 |
|
cthomas
|
aa564fc847
|
fix: add temp hack to gracefully handle parallel tool calling (#2654)
|
2025-06-05 14:43:46 -07:00 |
|
Kevin Lin
|
081746b74a
|
fix: set openai streaming interface letta_message_id (#2648)
Co-authored-by: Caren Thomas <carenthomas@gmail.com>
|
2025-06-05 12:26:01 -07:00 |
|
cthomas
|
f8c0002fbc
|
feat: populate tool call name and id in when token streaming (#2639)
|
2025-06-04 17:06:44 -07:00 |
|
Matthew Zhou
|
c9b047249e
|
fix: Change enum to fix composio tests (#2488)
|
2025-05-28 10:24:22 -07:00 |
|
Matthew Zhou
|
f18f236399
|
feat: Add TTFT latency from provider in traces (#2481)
|
2025-05-28 10:06:16 -07:00 |
|
cthomas
|
982cf0c6fe
|
feat: add tracing to streaming interface (#2477)
|
2025-05-27 16:20:05 -07:00 |
|
Matthew Zhou
|
8e9307c289
|
feat: Asyncify insert archival memories (#2430)
Co-authored-by: Caren Thomas <carenthomas@gmail.com>
|
2025-05-25 22:28:35 -07:00 |
|
Shangyin Tan
|
19efa1a89a
|
fix: do not pass temperature to request if model is oai reasoning model (#2189)
Co-authored-by: Charles Packer <packercharles@gmail.com>
|
2025-05-24 21:34:18 -07:00 |
|
cthomas
|
012e1cab4d
|
ci: use experimental for send message tests (#2290)
Co-authored-by: Sarah Wooders <sarahwooders@gmail.com>
|
2025-05-20 18:39:27 -07:00 |
|
Andy Li
|
0a54b998a8
|
feat: track llm provider traces and tracking steps in async agent loop (#2219)
|
2025-05-19 15:50:56 -07:00 |
|
cthomas
|
8463903d13
|
feat(asyncify): migrate actors(users) endpoints (#2211)
|
2025-05-16 00:37:08 -07:00 |
|
Sarah Wooders
|
69e21e850b
|
feat: add OpenAI streaming interface for new agent loop (#2191)
|
2025-05-15 22:20:08 -07:00 |
|