I’m trying out this new agentic search thing with the GPT-5 model through the OpenAI API. The issue is it’s very slow – usually takes at least 25 seconds for a simple search on one text field (not even multiple ones). And it often hits the 30-second timeout. For comparison, my own RESTful API handles the same search (no LLM involved) in just over 100ms. I get that hitting the LLM, getting the DSL response, and running it on opensearch cluster all adds up, but is my 20+ seconds wait time typical? If so, I guess it is literally “experimental”.
@summerist.l I’ve tried that with AWS Bedrock us.anthropic.claude-sonnet-4-20250514-v1:0 and openai.gpt-oss-120b-1:0, and they were running fine. (below 10 sec)
Thanks for looking into agentic search @summerist.l .We are aware of the latency, this is expected, as it comes from the multiple LLM calls made by the conversation agent.
If your primary goal is intelligent retrieval with contextual understanding, we recommend using the conversation agent.
@summerist.l@pablo@jakabasej5 we are looking for gathering use cases for agentic search. If you can let us know your use cases for using the same would be really helpful. Thank you.