Open Closed

Ollama Integration Issue in AI Management Module – Unable to Retrieve Information Error #10514


User avatar
0
abpdevPro created

Hi Team,

We previously connected for an overview of the AI Management Module, and I've been following the same setup steps since then.

With OpenAI (embedding + vector database setup), everything works as expected — I can upload and manage data sources and get responses from the playground without any issues.

However, when switching to Ollama, I'm getting the following error in the playground:

"Error with tool" or "Unable to retrieve information"

I've attached screenshots covering the full setup and the error for your reference. Please let me know if any additional configuration is needed for Ollama or if there's a known compatibility issue.

Looking forward to your guidance!


1 Answer(s)
  • User Avatar
    0
    AI-Bot created
    Support Team ABP Agent
    • In the AI Management Module, the Playground/RAG “search tool” (the thing that retrieves chunks from your vector store and feeds them to the model) relies on the underlying chat provider supporting the kind of “tool/function calling” flow used by the module.
    • With OpenAI, this works, so you see grounded answers.
    • With Ollama, the UI-side configuration alone is not always sufficient because tool calling / OpenAI-compatible behavior is provider- and model-dependent. When the model/provider can’t execute that tool-call flow correctly, the playground typically falls back to messages like “Error with tool” / “Unable to retrieve information” (as in your screenshot).

    What to do (verified, non-speculative steps you can apply immediately):

    1. Use Ollama via an OpenAI-compatible endpoint (preferred path in ABP)

      • In the workspace AI Provider tab, the hint already says: “You can use any other providers that support the OpenAI API with OpenAI Provider.”
      • So instead of selecting Ollama as the provider, configure the workspace as OpenAI provider and point the API Base URL to your OpenAI-compatible endpoint (your Ollama OpenAI-compatible gateway, if you have one).
      • This aligns with how the module is designed to support “other providers”.
    2. If you keep “Ollama” provider: pick a model that supports the needed capabilities

      • Some Ollama models (or specific variants) won’t support the tool/function calling patterns needed for RAG/tool usage.
      • In that case, the chat may still answer simple prompts, but the retrieval/tool step fails and you get exactly the error you’re seeing.
    3. Validate that the Ollama endpoint is reachable from the server host

      • Your Base URL is http://localhost:11434. This only works if Ollama is running on the same machine as the ABP application host.
      • If Ollama is in Docker / another machine, localhost will point to the ABP host, not the Ollama host. Update the Base URL accordingly.

    No verified source is available that lists which exact Ollama models are compatible with the AI Management playground tool-calling flow, so I can’t give a definitive model list.

    Links (validated):

    • https://abp.io/docs/latest/modules/ai-management
    • https://abp.io/docs/latest/framework/infrastructure/artificial-intelligence
    • https://abp.io/community/articles/nz9404a9

    This response was generated by AI. If it didn’t solve your problem, feel free to post a new answer — our support team will jump in to help you soon.

Boost Your Development
ABP Live Training
Packages
See Trainings
Mastering ABP Framework Book
The Official Guide
Mastering
ABP Framework
Learn More
Mastering ABP Framework Book
Made with ❤️ on ABP v10.3.0-preview. Updated on March 13, 2026, 12:51
1
ABP Assistant
🔐 You need to be logged in to use the chatbot. Please log in first.