Commit Graph

6949 Commits

Author SHA1 Message Date
Matthew Zhou
129dd97902 feat: Add fetch webpage tool [LET-4188] (#4395)
* Add fetch webpage tool

* Use trafilatura for web extraction
2025-09-03 13:34:35 -07:00
Matthew Zhou
e6c9b4d708 feat: Add fetch webpage tool [LET-4188] (#4395)
* Add fetch webpage tool

* Use trafilatura for web extraction
2025-09-03 13:34:35 -07:00
Matthew Zhou
051a5cde6a feat: Add archival search endpoint [LET-4184] (#4390)
* Add archival search endpoint

* Run fern autogen

* Add de-dupe logic
2025-09-03 10:55:20 -07:00
Matthew Zhou
6dccc1961b feat: Add archival search endpoint [LET-4184] (#4390)
* Add archival search endpoint

* Run fern autogen

* Add de-dupe logic
2025-09-03 10:55:20 -07:00
Matthew Zhou
6e633bd8f9 feat: Change namespace to be org scoped and filter on agent_id [LET-4163] (#4368)
* Change to org scoped and agent_id filtering

* Finish modifying conversation search tool

* Fix failing tests

* Get rid of bad imports
2025-09-03 10:00:19 -07:00
Matthew Zhou
88831c7256 feat: Change namespace to be org scoped and filter on agent_id [LET-4163] (#4368)
* Change to org scoped and agent_id filtering

* Finish modifying conversation search tool

* Fix failing tests

* Get rid of bad imports
2025-09-03 10:00:19 -07:00
Charles Packer
507784f292 fix: try and patch the PATCH/update issue with MCP server URL [LET-3933]
* fix: patch update path

* autogenerate

* update

* update

---------

Co-authored-by: Jin Peng <jinjpeng@Jins-MacBook-Pro.local>
Co-authored-by: Letta Bot <noreply@letta.com>
Co-authored-by: jnjpng <jin@letta.com>
2025-09-03 09:42:57 -07:00
Charles Packer
77cab00cb8 fix: try and patch the PATCH/update issue with MCP server URL [LET-3933]
* fix: patch update path

* autogenerate

* update

* update

---------

Co-authored-by: Jin Peng <jinjpeng@Jins-MacBook-Pro.local>
Co-authored-by: Letta Bot <noreply@letta.com>
Co-authored-by: jnjpng <jin@letta.com>
2025-09-03 09:42:57 -07:00
cthomas
93ef177ab6 feat: always default to openai client in llm client factory [LET-4176] (#4377)
feat: always default to openai client in llm client factory
2025-09-03 09:40:59 -07:00
cthomas
40f03b9780 feat: always default to openai client in llm client factory [LET-4176] (#4377)
feat: always default to openai client in llm client factory
2025-09-03 09:40:59 -07:00
jnjpng
cfe47489d8 fix: clean property schema reference
fix

Co-authored-by: Letta Bot <noreply@letta.com>
2025-09-03 08:42:46 -07:00
jnjpng
1d0d44d68a fix: clean property schema reference
fix

Co-authored-by: Letta Bot <noreply@letta.com>
2025-09-03 08:42:46 -07:00
Charles Packer
1ae0ab1790 fix: properly throw context window exceeded error on the new style of… [LET-4179] (#4380)
fix: properly throw context window exceeded error on the new style of context window overflow from gpt-5
2025-09-02 22:20:46 -07:00
Charles Packer
6ae49ab3c5 fix: properly throw context window exceeded error on the new style of… [LET-4179] (#4380)
fix: properly throw context window exceeded error on the new style of context window overflow from gpt-5
2025-09-02 22:20:46 -07:00
Charles Packer
361c50bce4 fix: patch context window of gpt-5 to 272k not 400k [LET-4177] (#4379)
fix: patch context window of gpt-5 to 272k not 400k
2025-09-02 21:10:40 -07:00
Charles Packer
8475f5b316 fix: patch context window of gpt-5 to 272k not 400k [LET-4177] (#4379)
fix: patch context window of gpt-5 to 272k not 400k
2025-09-02 21:10:40 -07:00
jnjpng
a9e20097bd fix: mcp tool schema formatting for anthropic streaming [LET-4165]
* fix

* update commnet

---------

Co-authored-by: Letta Bot <noreply@letta.com>
2025-09-02 20:44:43 -07:00
jnjpng
ea311d2e22 fix: mcp tool schema formatting for anthropic streaming [LET-4165]
* fix

* update commnet

---------

Co-authored-by: Letta Bot <noreply@letta.com>
2025-09-02 20:44:43 -07:00
cthomas
11d2507877 feat: refactor streaming route logic (#4369) 2025-09-02 19:27:13 -07:00
cthomas
1e0c9a710b feat: refactor streaming route logic (#4369) 2025-09-02 19:27:13 -07:00
Charles Packer
bcfd88ae98 fix: patch cache bust in the system prompt via timestamp [LET-4155] (#4375)
fix: patch cache bust in the system prompt via timestamp
2025-09-02 18:07:13 -07:00
Charles Packer
46672de2d6 fix: patch cache bust in the system prompt via timestamp [LET-4155] (#4375)
fix: patch cache bust in the system prompt via timestamp
2025-09-02 18:07:13 -07:00
Charles Packer
264171f327 fix: patch streaming hidden reasoning event [LET-4167] (#4367)
* fix: patch streaming hidden reasoning event

* fix: patch reasoning_effort not getting passed to openai
2025-09-02 16:21:18 -07:00
Charles Packer
e40c0512c2 fix: patch streaming hidden reasoning event [LET-4167] (#4367)
* fix: patch streaming hidden reasoning event

* fix: patch reasoning_effort not getting passed to openai
2025-09-02 16:21:18 -07:00
Shubham Naik
98e8fcdee6 feat: allow list tools by tool type [PRO-870] (#4036)
* feat: allow list tools by tool type

* chore: update list

* chore: respond to comments

* chore: refactor tools hella

* Add tests to managers

* chore: branch

---------

Co-authored-by: Shubham Naik <shub@memgpt.ai>
Co-authored-by: Matt Zhou <mattzh1314@gmail.com>
2025-09-02 15:53:36 -07:00
Shubham Naik
6716f5b5e8 feat: allow list tools by tool type [PRO-870] (#4036)
* feat: allow list tools by tool type

* chore: update list

* chore: respond to comments

* chore: refactor tools hella

* Add tests to managers

* chore: branch

---------

Co-authored-by: Shubham Naik <shub@memgpt.ai>
Co-authored-by: Matt Zhou <mattzh1314@gmail.com>
2025-09-02 15:53:36 -07:00
Matthew Zhou
c69fc95870 feat: Extend crud lifecycle of messages [LET-4158] (#4364)
Extend crud lifecycle of messages
2025-09-02 14:38:30 -07:00
Matthew Zhou
75d444c335 feat: Extend crud lifecycle of messages [LET-4158] (#4364)
Extend crud lifecycle of messages
2025-09-02 14:38:30 -07:00
Matthew Zhou
8f425aa024 feat: Modify conversation search tool to be hybrid (#4362)
* Modify conversation search functionality

* Gate the roles
2025-09-02 13:45:53 -07:00
Matthew Zhou
8e2a53b956 feat: Modify conversation search tool to be hybrid (#4362)
* Modify conversation search functionality

* Gate the roles
2025-09-02 13:45:53 -07:00
Matthew Zhou
6b625cb039 feat: Extract more text from messages [LET-4154] (#4359)
* Extract more message types

* Add extra tests testing for multiple parts
2025-09-02 13:25:32 -07:00
Matthew Zhou
7d9d8b0d50 feat: Extract more text from messages [LET-4154] (#4359)
* Extract more message types

* Add extra tests testing for multiple parts
2025-09-02 13:25:32 -07:00
Charles Packer
12d39c083e fix: change the default reasoning effort for gpt-5 to be minimal, due to extremely high latency on medium (#4356)
fix: change the default reasoning effort for gpt-5 to be minimal, due to extremely high latency on it
2025-09-02 13:23:13 -07:00
Charles Packer
7a74778394 fix: change the default reasoning effort for gpt-5 to be minimal, due to extremely high latency on medium (#4356)
fix: change the default reasoning effort for gpt-5 to be minimal, due to extremely high latency on it
2025-09-02 13:23:13 -07:00
Matthew Zhou
29502ac3aa feat: Add association for the vector namespace [LET-4148] (#4354)
Add association for the namespace
2025-09-02 13:21:42 -07:00
Matthew Zhou
00fd8617a6 feat: Add association for the vector namespace [LET-4148] (#4354)
Add association for the namespace
2025-09-02 13:21:42 -07:00
Charles Packer
9d49eff204 fix: patch the streaming issue in the openai client for when inner_thoughts_in_kwargs is off [LET-4146] (#4350)
fix: patch the streaming issue in the openai client for when inner_thoughts_in_kwargs is off
2025-09-02 12:44:25 -07:00
Charles Packer
e3eb598353 fix: patch the streaming issue in the openai client for when inner_thoughts_in_kwargs is off [LET-4146] (#4350)
fix: patch the streaming issue in the openai client for when inner_thoughts_in_kwargs is off
2025-09-02 12:44:25 -07:00
Matthew Zhou
b8a198688f feat: Embed all messages in turbopuffer [LET-4143] (#4352)
* wip

* Finish embedding

* Fix ruff and tests
2025-09-02 12:43:48 -07:00
Matthew Zhou
20ac2923ff feat: Embed all messages in turbopuffer [LET-4143] (#4352)
* wip

* Finish embedding

* Fix ruff and tests
2025-09-02 12:43:48 -07:00
Charles Packer
a696d9e3d5 fix: various fixes to make gpt-5 work better [LET-4138] (#4344)
* fix: patch gpt-5 compat

* feat: add verbosity level dropdown selector to llm config panel in ade

* fix: patch duplicated logic

* fix: make default verbosity None, just stage just publish

* fix: missing files
2025-09-02 10:30:25 -07:00
Charles Packer
3296471d09 fix: various fixes to make gpt-5 work better [LET-4138] (#4344)
* fix: patch gpt-5 compat

* feat: add verbosity level dropdown selector to llm config panel in ade

* fix: patch duplicated logic

* fix: make default verbosity None, just stage just publish

* fix: missing files
2025-09-02 10:30:25 -07:00
cthomas
9b315e9f54 feat: support filtering out approval messages for anthropic api (#4342) 2025-09-01 21:39:50 -07:00
cthomas
5ca9b5e774 feat: support filtering out approval messages for anthropic api (#4342) 2025-09-01 21:39:50 -07:00
Charles Packer
e741f84add fix: patch bug w/ extended thinking mode involving text leaking into reasoning (#4341)
* fix: patch for bad native reasoning behavior w/ sonnet

* fix: cleanup

* fix: cleanup

* fix: another prompt tune for less flaking
2025-09-01 20:26:24 -07:00
Charles Packer
dc5da0ba36 fix: patch bug w/ extended thinking mode involving text leaking into reasoning (#4341)
* fix: patch for bad native reasoning behavior w/ sonnet

* fix: cleanup

* fix: cleanup

* fix: another prompt tune for less flaking
2025-09-01 20:26:24 -07:00
cthomas
f3df0433ae feat: add more tests for hitl (#4339) 2025-09-01 16:23:06 -07:00
cthomas
fecff4c12c feat: add more tests for hitl (#4339) 2025-09-01 16:23:06 -07:00
Charles Packer
71a5eaa262 fix(core): change the backend mid-stream error packing to match what the FE expects [PRO-1107] (#4340)
fix(core): change the backend mid-stream error packing to match what the FE expects
2025-09-01 14:59:42 -07:00
Charles Packer
7d388bff32 fix(core): change the backend mid-stream error packing to match what the FE expects [PRO-1107] (#4340)
fix(core): change the backend mid-stream error packing to match what the FE expects
2025-09-01 14:59:42 -07:00