Commit Graph

35 Commits

Author SHA1 Message Date
Charles Packer
af62e22b76 fix: patch send message partial json extractor (#2868) 2025-09-13 12:52:13 -07:00
cthomas
56801bcf17 fix: 4o mini assistant streaming error (#2865) 2025-09-12 20:06:00 -07:00
Charles Packer
21df8e9394 fix: patch stray chunks on streaming w/ models with tight tokenization (#2830) 2025-09-11 11:51:00 -07:00
cthomas
3ede465b53 fix: null check on span before log event (#2810) 2025-09-09 16:23:59 -07:00
cthomas
9c098e8912 fix: logger error in streaming interface (#2807) 2025-09-09 16:02:27 -07:00
cthomas
6029c80446 fix: approval request for streaming (#4445)
* fix: approval request for streaming

* fix: claude code attempt, unit test passing (add on to #4445) (#4448)

* fix: claude code attempt, unit test passing

* chore: update locks to 0.1.314 from 0.1.312

* chore: just stage-api && just publish-api

* chore: drop dead poetry lock

---------

Co-authored-by: Charles Packer <packercharles@gmail.com>
2025-09-05 17:43:21 -07:00
Charles Packer
e40c0512c2 fix: patch streaming hidden reasoning event [LET-4167] (#4367)
* fix: patch streaming hidden reasoning event

* fix: patch reasoning_effort not getting passed to openai
2025-09-02 16:21:18 -07:00
Charles Packer
e3eb598353 fix: patch the streaming issue in the openai client for when inner_thoughts_in_kwargs is off [LET-4146] (#4350)
fix: patch the streaming issue in the openai client for when inner_thoughts_in_kwargs is off
2025-09-02 12:44:25 -07:00
cthomas
7c88470705 feat: support filtering out messages when converting to openai dict (#4337)
* feat: support filtering out messages when converting to openai dict

* fix imports
2025-09-01 12:48:45 -07:00
Kian Jones
ca6cfa5914 chore: migrate to ruff (#4305)
* base requirements

* autofix

* Configure ruff for Python linting and formatting

- Set up minimal ruff configuration with basic checks (E, W, F, I)
- Add temporary ignores for common issues during migration
- Configure pre-commit hooks to use ruff with pass_filenames
- This enables gradual migration from black to ruff

* Delete sdj

* autofixed only

* migrate lint action

* more autofixed

* more fixes

* change precommit

* try changing the hook

* try this stuff
2025-08-29 11:11:19 -07:00
cthomas
a9771424fe fix: duplicate message stream error (#3834) 2025-08-11 14:27:35 -07:00
cthomas
cab621ac14 feat: continue stream processing on client cancel (#3796) 2025-08-07 13:17:36 -07:00
Andy Li
76679e3ecc feat: track metrics for runs in db 2025-08-06 15:46:50 -07:00
cthomas
f3316a0d8a feat: log stream cancellation to sentry (#3759) 2025-08-05 16:07:30 -07:00
jnjpng
7f12230c21 fix: manually count tokens for streaming lmstudio models
Co-authored-by: Jin Peng <jinjpeng@Jins-MacBook-Pro.local>
Co-authored-by: Charles Packer <packercharles@gmail.com>
2025-07-29 18:12:42 -07:00
Andy Li
f9bb757a98 feat: support for agent loop job cancelation (#2837) 2025-07-02 14:31:16 -07:00
Kevin Lin
efae2231ea feat: add omitted reasoning to streaming openai reasoning (#2846)
Co-authored-by: Charles Packer <packercharles@gmail.com>
2025-06-24 18:47:38 -07:00
Sarah Wooders
78e81391ba fix: avoid calling model_dump on stop reason messages twice (#2811) 2025-06-13 18:25:35 -07:00
cthomas
97986b0f03 feat: send stop reason in letta APIs (#2789) 2025-06-13 16:04:48 -07:00
Andy Li
bf7bd3b905 fix: metric tracking (#2785) 2025-06-13 13:53:10 -07:00
cthomas
ccb203b264 feat: consolidate logic for finish tokens (#2779) 2025-06-12 15:24:06 -07:00
Kevin Lin
78dad0db7a fix: patch reasoning models (#2703)
Co-authored-by: Charles Packer <packercharles@gmail.com>
2025-06-11 17:20:04 -07:00
Andy Li
80f6e97ca9 feat: otel metrics and expanded collecting (#2647)
(passed tests in last run)
2025-06-05 17:20:14 -07:00
cthomas
aa564fc847 fix: add temp hack to gracefully handle parallel tool calling (#2654) 2025-06-05 14:43:46 -07:00
Kevin Lin
081746b74a fix: set openai streaming interface letta_message_id (#2648)
Co-authored-by: Caren Thomas <carenthomas@gmail.com>
2025-06-05 12:26:01 -07:00
cthomas
f8c0002fbc feat: populate tool call name and id in when token streaming (#2639) 2025-06-04 17:06:44 -07:00
Matthew Zhou
c9b047249e fix: Change enum to fix composio tests (#2488) 2025-05-28 10:24:22 -07:00
Matthew Zhou
f18f236399 feat: Add TTFT latency from provider in traces (#2481) 2025-05-28 10:06:16 -07:00
cthomas
982cf0c6fe feat: add tracing to streaming interface (#2477) 2025-05-27 16:20:05 -07:00
Matthew Zhou
8e9307c289 feat: Asyncify insert archival memories (#2430)
Co-authored-by: Caren Thomas <carenthomas@gmail.com>
2025-05-25 22:28:35 -07:00
Shangyin Tan
19efa1a89a fix: do not pass temperature to request if model is oai reasoning model (#2189)
Co-authored-by: Charles Packer <packercharles@gmail.com>
2025-05-24 21:34:18 -07:00
cthomas
012e1cab4d ci: use experimental for send message tests (#2290)
Co-authored-by: Sarah Wooders <sarahwooders@gmail.com>
2025-05-20 18:39:27 -07:00
Andy Li
0a54b998a8 feat: track llm provider traces and tracking steps in async agent loop (#2219) 2025-05-19 15:50:56 -07:00
cthomas
8463903d13 feat(asyncify): migrate actors(users) endpoints (#2211) 2025-05-16 00:37:08 -07:00
Sarah Wooders
69e21e850b feat: add OpenAI streaming interface for new agent loop (#2191) 2025-05-15 22:20:08 -07:00