Log in
Log into community
Find answers from the community
View all posts
Related posts
Did this answer your question?
π
π
π
Powered by
Hall
Inactive
Updated last year
0
Follow
Logan M When I m using Multi hop Agent
Logan M When I m using Multi hop Agent
Inactive
0
Follow
V
Vish
last year
Β·
@Logan M When I'm using Multi-hop Agent calls, there's increased response times for queries especially when calling multiple tools across agents. Is there a caching solution that I can use to help reduce response times?
L
V
8 comments
Share
Open in Discord
L
Logan M
last year
We've been meaning to integrate something like gpt cache, but we haven't gotten around to it yet lol
L
Logan M
last year
I agree it would probably be helpful
V
Vish
last year
I see
V
Vish
last year
What about implementations like Langchain's InMemoryCache or something
L
Logan M
last year
I think you'd have to use a langchain LLM class (which then should have caching)
L
Logan M
last year
And that should be compatible with llama index π«°
V
Vish
last year
Well Langchain LLMs aren't compatible with OpenAI Agents iirc
V
Vish
last year
What I meant was more so porting what they are doing for caching, natively to llama index somehow
Add a reply
Sign up and join the conversation on Discord
Join on Discord