Skip to content

Latest commit

 

History

History
48 lines (27 loc) · 2.83 KB

File metadata and controls

48 lines (27 loc) · 2.83 KB
graph LR
    Completion_Orchestrator["Completion Orchestrator"]
    Response_Cache_Manager["Response Cache Manager"]
    LLM_Request_Sender["LLM Request Sender"]
    Completion_Orchestrator -- "consults" --> Response_Cache_Manager
    Completion_Orchestrator -- "delegates request to" --> LLM_Request_Sender
    LLM_Request_Sender -- "passes response to" --> Response_Cache_Manager
Loading

CodeBoardingDemoContact

Details

Manages interactions with the external llama.cpp server and optimizes performance through caching, acting as the interface between the Vim/Neovim environment and the external LLM service.

Completion Orchestrator

This is the central component responsible for initiating and coordinating the LLM completion process. It acts as the primary entry point for completion requests within the Vim plugin, deciding whether to use cached data or send a new request to the LL LLM server.

Related Classes/Methods:

Response Cache Manager

Manages the in-memory cache for previously generated LLM completions. Its core responsibility is to improve performance by reducing redundant calls to the external llama.cpp server. It provides mechanisms for storing new completions and retrieving existing ones based on request parameters.

Related Classes/Methods:

LLM Request Sender

This component is solely responsible for handling the direct communication with the external llama.cpp server. It encapsulates the logic for formatting the request payload, sending the HTTP request, and processing the raw response received from the LLM server. It serves as the direct interface to the external AI service.

Related Classes/Methods: