graph LR
Completion_Orchestrator["Completion Orchestrator"]
Response_Cache_Manager["Response Cache Manager"]
LLM_Request_Sender["LLM Request Sender"]
Completion_Orchestrator -- "consults" --> Response_Cache_Manager
Completion_Orchestrator -- "delegates request to" --> LLM_Request_Sender
LLM_Request_Sender -- "passes response to" --> Response_Cache_Manager
Manages interactions with the external llama.cpp server and optimizes performance through caching, acting as the interface between the Vim/Neovim environment and the external LLM service.
This is the central component responsible for initiating and coordinating the LLM completion process. It acts as the primary entry point for completion requests within the Vim plugin, deciding whether to use cached data or send a new request to the LL LLM server.
Related Classes/Methods:
Manages the in-memory cache for previously generated LLM completions. Its core responsibility is to improve performance by reducing redundant calls to the external llama.cpp server. It provides mechanisms for storing new completions and retrieving existing ones based on request parameters.
Related Classes/Methods:
This component is solely responsible for handling the direct communication with the external llama.cpp server. It encapsulates the logic for formatting the request payload, sending the HTTP request, and processing the raw response received from the LLM server. It serves as the direct interface to the external AI service.
Related Classes/Methods: