Commit Graph

69 Commits

Author SHA1 Message Date
Ari Webb
48cc73175b feat: parallel tool calling for openai non streaming [LET-4593] (#5773)
* first hack

* clean up

* first implementation working

* revert package-lock

* remove openai test

* error throw

* typo

* Update integration_test_send_message_v2.py

* Update integration_test_send_message_v2.py

* refine test

* Only make changes for openai non streaming

* Add tests

---------

Co-authored-by: Ari Webb <ari@letta.com>
Co-authored-by: Matt Zhou <mattzh1314@gmail.com>
2025-11-13 15:36:14 -08:00
Ari Webb
f3a40a41f5 feat: updated backend to not allow minimal for codex [LET-5883] (#5760)
* updated backend

* add function in openai_client

* remove values before error

* remove test

---------

Co-authored-by: Ari Webb <ari@letta.com>
2025-11-13 15:35:34 -08:00
Matthew Zhou
09ba075cfa feat: Modify embedding strategy to first halve the batch size v.s. the batc… [LET-5510] (#5434)
Modify embedding strategy to first halve the batch size v.s. the batch size
2025-10-24 15:12:11 -07:00
Matthew Zhou
0543a60538 chore: Restore chunk size for openai embeddings (#5431)
Restore chunk size
2025-10-24 15:12:11 -07:00
Ari Webb
624c591820 Ari/let 5486 badrequesterror error code 400 error message requested [LET-5486] (#5422)
* letta agent v2 throw exception not error

* warning instead of error or exception

* decrease min_chunk_size

---------

Co-authored-by: Ari Webb <ari@letta.com>
2025-10-24 15:12:11 -07:00
Kevin Lin
08da1a64bb feat: parse reasoning_content from OAI proxies (eg. vLLM / OpenRouter) (#5372)
* reasonig_content support

* fix

* comment

* fix

* rm comment

---------

Co-authored-by: Charles Packer <packercharles@gmail.com>
2025-10-24 15:11:31 -07:00
cthomas
a1e771877d feat: add model name patch for openrouter (#5303)
* feat: add model name patch for openrouter

* add comment
2025-10-09 15:25:21 -07:00
cthomas
f8dce88ce4 feat: support for models that do not allow None content (#5218) 2025-10-07 17:50:50 -07:00
cthomas
a3545110cf feat: add full responses api support in new agent loop (#5051)
* feat: add full responses api support in new agent loop

* update matrix in workflow

* relax check for reasoning messages for high effort gpt 5

* fix indent

* one more relax
2025-10-07 17:50:48 -07:00
Matthew Zhou
df5c997da0 feat: Enable dynamic toggling of tool choice in v3 agent loop for OpenAI [LET-4564] (#5042)
* Add subsequent flag

* Finish integrating constrained/unconstrained toggling on v3 agent loop

* Update tests to run on v3

* Run lint
2025-10-07 17:50:47 -07:00
Sarah Wooders
b5de42fefd fix: patch summarizers for integration_test_send_message.py (#4919)
* fix: integration_test_send_message.py

* patch summarizer

* remove print
2025-10-07 17:50:46 -07:00
Charles Packer
8da15aaf08 fix(core): patch issue where LLM may generate a 'noop' call [PRO-1340] (#4944)
fix(core): patch issue where LLM may generate a 'noop' call
2025-10-07 17:50:46 -07:00
Charles Packer
a4041879a4 feat: add new agent loop (squash rebase of OSS PR) (#4815)
* feat: squash rebase of OSS PR

* fix: revert changes that weren't on manual rebase

* fix: caught another one

* fix: disable force

* chore: drop print

* fix: just stage-api && just publish-api

* fix: make agent_type consistently an arg in the client

* fix: patch multi-modal support

* chore: put in todo stub

* fix: disable hardcoding for tests

* fix: patch validate agent sync (#4882)

patch validate agent sync

* fix: strip bad merge diff

* fix: revert unrelated diff

* fix: react_v2 naming -> letta_v1 naming

* fix: strip bad merge

---------

Co-authored-by: Kevin Lin <klin5061@gmail.com>
2025-10-07 17:50:45 -07:00
Charles Packer
9edc7f4d64 feat: add OpenRouterProvider (#4848)
* feat: init add of openrouter provider, doesn't work bc no key pass and no header pass

* fix: working
2025-10-07 17:50:45 -07:00
Kian Jones
b8e9a80d93 merge this (#4759)
* wait I forgot to comit locally

* cp the entire core directory and then rm the .git subdir
2025-09-17 15:47:40 -07:00
Kian Jones
22f70ca07c chore: officially migrate to submodule (#4502)
* remove apps/core and apps/fern

* fix precommit

* add submodule updates in workflows

* submodule

* remove core tests

* update core revision

* Add submodules: true to all GitHub workflows

- Ensure all workflows can access git submodules
- Add submodules support to deployment, test, and CI workflows
- Fix YAML syntax issues in workflow files

🤖 Generated with [Claude Code](https://claude.ai/code)

Co-Authored-By: Claude <noreply@anthropic.com>

* remove core-lint

* upgrade core with latest main of oss

---------

Co-authored-by: Claude <noreply@anthropic.com>
2025-09-09 12:45:53 -07:00
Matthew Zhou
0eed3722cc feat: Robustify openai embedding [LET-4256] (#4478)
* Robustify embedding

* Remove unecessary imports

* Add test embeddings
2025-09-08 17:18:54 -07:00
Charles Packer
1ae0ab1790 fix: properly throw context window exceeded error on the new style of… [LET-4179] (#4380)
fix: properly throw context window exceeded error on the new style of context window overflow from gpt-5
2025-09-02 22:20:46 -07:00
Charles Packer
264171f327 fix: patch streaming hidden reasoning event [LET-4167] (#4367)
* fix: patch streaming hidden reasoning event

* fix: patch reasoning_effort not getting passed to openai
2025-09-02 16:21:18 -07:00
Charles Packer
a696d9e3d5 fix: various fixes to make gpt-5 work better [LET-4138] (#4344)
* fix: patch gpt-5 compat

* feat: add verbosity level dropdown selector to llm config panel in ade

* fix: patch duplicated logic

* fix: make default verbosity None, just stage just publish

* fix: missing files
2025-09-02 10:30:25 -07:00
cthomas
1edcc13778 feat: support filtering out messages when converting to openai dict (#4337)
* feat: support filtering out messages when converting to openai dict

* fix imports
2025-09-01 12:48:45 -07:00
Kian Jones
fecf6decfb chore: migrate to ruff (#4305)
* base requirements

* autofix

* Configure ruff for Python linting and formatting

- Set up minimal ruff configuration with basic checks (E, W, F, I)
- Add temporary ignores for common issues during migration
- Configure pre-commit hooks to use ruff with pass_filenames
- This enables gradual migration from black to ruff

* Delete sdj

* autofixed only

* migrate lint action

* more autofixed

* more fixes

* change precommit

* try changing the hook

* try this stuff
2025-08-29 11:11:19 -07:00
cthomas
c7b71ad8a9 feat: add new xai llm client (#3936) 2025-08-20 15:49:09 -07:00
Kevin Lin
a22d6b1315 feat: GPT-5 support (#3924)
Co-authored-by: Sarah Wooders <sarahwooders@gmail.com>
2025-08-15 19:01:50 -07:00
cthomas
9f84fb8500 feat: refactor byok logic in llm clients (#3880) 2025-08-12 14:19:02 -07:00
cthomas
2e19f2a801 feat: add new together llm client (#3875) 2025-08-12 13:37:20 -07:00
cthomas
5cf807574f feat: consolidate reasoning model checks (#3862) 2025-08-11 16:55:45 -07:00
Sarah Wooders
b85f8aa43c feat: support opus 4.1 and gpt-5 (#3806) 2025-08-07 15:11:57 -07:00
jnjpng
9d8a122da0 fix: lmstudio support for qwen and llama
Co-authored-by: Jin Peng <jinjpeng@Jins-MacBook-Pro.local>
Co-authored-by: Charles Packer <packercharles@gmail.com>
2025-07-29 15:57:20 -07:00
Andy Li
04e9f43220 chore: strings lint cleanup (#3374) 2025-07-18 09:20:45 -07:00
Charles Packer
12c2b49461 fix: add frequency penalty for gpt-4o-mini (#3166) 2025-07-06 11:05:31 -07:00
Matthew Zhou
efca9d8ea0 feat: Only add suffix on duplication (#3120) 2025-07-01 13:48:38 -07:00
Matthew Zhou
5dccccec21 fix: Fix constraints and also implement bulk attach (#3107) 2025-06-30 14:27:57 -07:00
Andy Li
734680db81 feat: timeout configuration for LLM clients + vertex (#2972) 2025-06-23 16:55:23 -07:00
Kevin Lin
65530e8380 fix: add exceptions to accept_developer_role (#2848) 2025-06-16 15:14:40 -07:00
Kevin Lin
93c15244ab feat: add reasoning models to integration_test_send_message (#2710) 2025-06-13 14:54:37 -07:00
cthomas
c1255dc9d1 feat: make tool calls required for model proxy (#2756) 2025-06-11 11:36:45 -07:00
cthomas
5ecd8a706c fix: parallel tool calling OpenAI (#2738) 2025-06-10 14:27:01 -07:00
cthomas
b332ebfa85 feat: support multi content part input (#2717) 2025-06-10 13:36:17 -07:00
Matthew Zhou
b53be62e7a fix: Turn parallel tool calling off for OpenAI (#2737) 2025-06-10 13:27:00 -07:00
cthomas
20e6732f36 feat: add multi-modal input support (#2590) 2025-06-08 18:28:01 -07:00
cthomas
e4da78fce7 fix: gracefully handle too long responses from llm provider (#2677) 2025-06-06 13:13:32 -07:00
Andy Li
d2252f2953 feat: otel metrics and expanded collecting (#2647)
(passed tests in last run)
2025-06-05 17:20:14 -07:00
Matthew Zhou
82b3222a52 fix: Make OpenAI context window exceeded error more specific (#2624) 2025-06-04 12:57:51 -07:00
Matthew Zhou
87f4bcad9a feat: Add summarization for more scenarios (#2499) 2025-05-29 11:10:13 -07:00
Sarah Wooders
3354f5fe50 feat: concurrently make embedding request and use async client for OpenAI (#2482)
Co-authored-by: Matthew Zhou <mattzh1314@gmail.com>
2025-05-28 11:35:22 -07:00
Sarah Wooders
4cc075f1fc feat: add more trace methods (#2471) 2025-05-27 16:56:30 -07:00
cthomas
e813a65351 feat(asyncify): byok in async loop (#2421) 2025-05-25 19:47:20 -07:00
Shangyin Tan
2199d8fdda fix: do not pass temperature to request if model is oai reasoning model (#2189)
Co-authored-by: Charles Packer <packercharles@gmail.com>
2025-05-24 21:34:18 -07:00
cthomas
b554171d41 feat: add tracing to llm clients (#2340) 2025-05-22 13:55:32 -07:00