To access the latest features keep your code editor plug-in up to date.
By default, Navie uses an AppMap proxy of the latest OpenAI supported AI models. If you would like to customize your own model, you can leverage a variety of other AI model providers such as Azure OpenAI, Fireworks.ai, LM Studio, and more.
If you have an active GitHub Copilot subscription, you can use Navie with the Copilot Language Model as a supported backend. Refer to the Navie Copilot documentation for instructions on how to enable.
After you complete the installation of AppMap for your code editor. Open the Navie Chat Window to ask Navie about your application.
To open the Navie Chat, open the AppMap plugin in the sidebar menu for your code editor, and select the New Navie Chat
option.
You can ask questions about your application with Navie immediately after installing the plugin. Navie will answer questions based on analysis of your project code. For increased accuracy of more complex projects, you can record AppMap data and Navie will utilize this information as well.
By default, Navie will utilize an OpenAI service hosted by AppMap. If, for data privacy or other reasons, you are do not wish to use the AppMap OpenAI proxy, you can bring your own OpenAI API key, or use an entirely different AI Model, hosted in your environment or hosted locally.
When you ask a question to Navie, it will search through all the available AppMap data for your project to pull in relevant traces, sequence diagrams, and code snippets for analysis. it will send the selected context to your preferred LLM provider.
To achieve the highest quality results, we suggest using the available command modes when prompting Navie. Simply type @
into the chat input to access the list of available command modes.
By default, Navie chat is in a default mode called @explain
. Other specialized modes are available for generating diagrams, planning work, generating code and tests, and more. Consult Using Navie docs for more details on Navie commands.
The Navie UI includes a standard chat window, and a context panel which will include all the context that is included in the query to the AI provider. This context can include things such as:
Always available:
Note: Only code files and other files that are tracked by Git will be included as context and sent to the LLM when you ask a question. AppMap Navie will respect your .gitignore
and will not access files which are not tracked or ignored by Git (such as passwords or sensitive data).
If AppMap Data exists:
Navie will look for the files listed above in the following locations:
Generate AppMap Data and you will greatly improve the quality of your Navie AI responses. With AppMap Data for your project, you can now ask much deeper architectural questions about your application. This is possible because of the additional context from AppMap Data and the higher accuracy and relevance of the code snippets which are relevant to your question.
View the Navie AI examples page to see some examples of Navie fixing complex architectural issues, performance issues, and adding new features to your application.
After your AppMap Data is generated, the Navie window will indicate the AppMap Data that exists for your project.
With this AppMap Data in your project, asking questions to Navie will now include data flows, sequence diagrams, traces, in addition to the relevant code snippets for the project.
Continue to ask questions to Navie, creating new code for your application, and continue to generate additional AppMap Data as your code changes. On each subsequent question asked, Navie will re-query your AppMap Data, traces, data flows, and source code for your project to power the context for the answer.
Learn more about making AppMap Data to improve Navie response accuracy