Commit Graph

53 Commits

Author SHA1 Message Date
Matthew Zhou
73e8ee73fc feat: Robustify openai embedding [LET-4256] (#4478)
* Robustify embedding

* Remove unecessary imports

* Add test embeddings
2025-09-08 17:18:54 -07:00
Charles Packer
6ae49ab3c5 fix: properly throw context window exceeded error on the new style of… [LET-4179] (#4380)
fix: properly throw context window exceeded error on the new style of context window overflow from gpt-5
2025-09-02 22:20:46 -07:00
Charles Packer
e40c0512c2 fix: patch streaming hidden reasoning event [LET-4167] (#4367)
* fix: patch streaming hidden reasoning event

* fix: patch reasoning_effort not getting passed to openai
2025-09-02 16:21:18 -07:00
Charles Packer
3296471d09 fix: various fixes to make gpt-5 work better [LET-4138] (#4344)
* fix: patch gpt-5 compat

* feat: add verbosity level dropdown selector to llm config panel in ade

* fix: patch duplicated logic

* fix: make default verbosity None, just stage just publish

* fix: missing files
2025-09-02 10:30:25 -07:00
cthomas
7c88470705 feat: support filtering out messages when converting to openai dict (#4337)
* feat: support filtering out messages when converting to openai dict

* fix imports
2025-09-01 12:48:45 -07:00
Kian Jones
ca6cfa5914 chore: migrate to ruff (#4305)
* base requirements

* autofix

* Configure ruff for Python linting and formatting

- Set up minimal ruff configuration with basic checks (E, W, F, I)
- Add temporary ignores for common issues during migration
- Configure pre-commit hooks to use ruff with pass_filenames
- This enables gradual migration from black to ruff

* Delete sdj

* autofixed only

* migrate lint action

* more autofixed

* more fixes

* change precommit

* try changing the hook

* try this stuff
2025-08-29 11:11:19 -07:00
cthomas
d0ddc5545c feat: add new xai llm client (#3936) 2025-08-20 15:49:09 -07:00
Kevin Lin
0583eec2f1 feat: GPT-5 support (#3924)
Co-authored-by: Sarah Wooders <sarahwooders@gmail.com>
2025-08-15 19:01:50 -07:00
cthomas
6289ca55a2 feat: refactor byok logic in llm clients (#3880) 2025-08-12 14:19:02 -07:00
cthomas
227e472282 feat: add new together llm client (#3875) 2025-08-12 13:37:20 -07:00
cthomas
639641137d feat: consolidate reasoning model checks (#3862) 2025-08-11 16:55:45 -07:00
Sarah Wooders
ce44386011 feat: support opus 4.1 and gpt-5 (#3806) 2025-08-07 15:11:57 -07:00
jnjpng
de3c99d7f5 fix: lmstudio support for qwen and llama
Co-authored-by: Jin Peng <jinjpeng@Jins-MacBook-Pro.local>
Co-authored-by: Charles Packer <packercharles@gmail.com>
2025-07-29 15:57:20 -07:00
Andy Li
904d9ba5a2 chore: strings lint cleanup (#3374) 2025-07-18 09:20:45 -07:00
Charles Packer
72b472896d fix: add frequency penalty for gpt-4o-mini (#3166) 2025-07-06 11:05:31 -07:00
Matthew Zhou
804ec12ee2 feat: Only add suffix on duplication (#3120) 2025-07-01 13:48:38 -07:00
Matthew Zhou
241f653eb6 fix: Fix constraints and also implement bulk attach (#3107) 2025-06-30 14:27:57 -07:00
Andy Li
343dbb5359 feat: timeout configuration for LLM clients + vertex (#2972) 2025-06-23 16:55:23 -07:00
Kevin Lin
6ea0a7e74e fix: add exceptions to accept_developer_role (#2848) 2025-06-16 15:14:40 -07:00
Kevin Lin
a8895ad017 feat: add reasoning models to integration_test_send_message (#2710) 2025-06-13 14:54:37 -07:00
cthomas
654f6b5053 feat: make tool calls required for model proxy (#2756) 2025-06-11 11:36:45 -07:00
cthomas
484a6f1d37 fix: parallel tool calling OpenAI (#2738) 2025-06-10 14:27:01 -07:00
cthomas
6a23bc743d feat: support multi content part input (#2717) 2025-06-10 13:36:17 -07:00
Matthew Zhou
530f9488b6 fix: Turn parallel tool calling off for OpenAI (#2737) 2025-06-10 13:27:00 -07:00
cthomas
2823e4447a feat: add multi-modal input support (#2590) 2025-06-08 18:28:01 -07:00
cthomas
e72f0f68ef fix: gracefully handle too long responses from llm provider (#2677) 2025-06-06 13:13:32 -07:00
Andy Li
80f6e97ca9 feat: otel metrics and expanded collecting (#2647)
(passed tests in last run)
2025-06-05 17:20:14 -07:00
Matthew Zhou
3ccd79efa5 fix: Make OpenAI context window exceeded error more specific (#2624) 2025-06-04 12:57:51 -07:00
Matthew Zhou
285acd0324 feat: Add summarization for more scenarios (#2499) 2025-05-29 11:10:13 -07:00
Sarah Wooders
729414e7a9 feat: concurrently make embedding request and use async client for OpenAI (#2482)
Co-authored-by: Matthew Zhou <mattzh1314@gmail.com>
2025-05-28 11:35:22 -07:00
Sarah Wooders
cab3733280 feat: add more trace methods (#2471) 2025-05-27 16:56:30 -07:00
cthomas
20470844a7 feat(asyncify): byok in async loop (#2421) 2025-05-25 19:47:20 -07:00
Shangyin Tan
19efa1a89a fix: do not pass temperature to request if model is oai reasoning model (#2189)
Co-authored-by: Charles Packer <packercharles@gmail.com>
2025-05-24 21:34:18 -07:00
cthomas
ed9f3e6abf feat: add tracing to llm clients (#2340) 2025-05-22 13:55:32 -07:00
cthomas
012e1cab4d ci: use experimental for send message tests (#2290)
Co-authored-by: Sarah Wooders <sarahwooders@gmail.com>
2025-05-20 18:39:27 -07:00
Matthew Zhou
068f27d83d feat: Add tavily search builtin tool (#2257) 2025-05-19 16:38:11 -07:00
Andy Li
0a54b998a8 feat: track llm provider traces and tracking steps in async agent loop (#2219) 2025-05-19 15:50:56 -07:00
Sarah Wooders
f076964bd1 feat: support together in new agent loop and add tests (#2231) 2025-05-17 19:17:08 -07:00
Kevin Lin
6a33859224 fix: use auto function calling for together models (#2097) 2025-05-09 17:46:35 -07:00
Charles Packer
8bb194541e fix: make togetherai nebius xai etc usable via the openaiprovider (#1981)
Co-authored-by: Kevin Lin <klin5061@gmail.com>
Co-authored-by: Kevin Lin <kl2806@columbia.edu>
2025-05-09 10:50:55 -07:00
cthomas
d8151d76e2 feat: add provider_category field to distinguish byok (#2038) 2025-05-06 17:31:36 -07:00
Charles Packer
77f9e2d127 fix: patch o1 support (#1978) 2025-05-02 14:54:25 -07:00
cthomas
9280d85ba4 feat: always add user id to openai requests (#1969) 2025-04-30 23:23:01 -07:00
cthomas
25c188689a feat: byok 2.0 (#1963) 2025-04-30 21:26:50 -07:00
cthomas
5187e79a43 feat: use new model-proxy in production (#1908) 2025-04-30 15:20:54 -07:00
cthomas
83157c1949 feat: add letta-free endpoint constant (#1907) 2025-04-27 12:57:06 -07:00
cthomas
7605675f23 feat: add llm config per request (#1866) 2025-04-23 16:37:05 -07:00
Charles Packer
e1fa7f0b40 fix: patch o-series (#1699) 2025-04-23 13:41:34 -07:00
Sarah Wooders
1dbf6cd4e5 feat: translate system to developer or o-series models (#1692)
Co-authored-by: cpacker <packercharles@gmail.com>
2025-04-11 21:59:48 -07:00
Matthew Zhou
4536800b3d fix: Fix build request data for OpenAI (#1654) 2025-04-09 16:31:20 -07:00