Skip to content

sfabio01/custom-pilot

Folders and files

NameName
Last commit message
Last commit date

Latest commit

 

History

4 Commits
 
 
 
 

Repository files navigation

Custom Pilot

Custom Pilot is a Visual Studio Code extension framework that lets you easily integrate your custom code completion model into VS Code.

Getting started

You can use any API server following OpenAI's API format. In particular, the API server should serve the endpoints:

  • GET /v1/models
  • POST /v1/completions

You can set your configuration in the Sidebar panel of the extension:

  1. Set the API server URL
  2. Choose an inference model from the dropdown menu
  3. If your API requires authentication, enter your API Key. It will be sent in the Authorization header of every request
  4. Optionally, adjust the advanced settings as you like

Sidebar Overview

The model you selected will be used to provide completions in any programming language and on any file.

Use Custom Pilot with LM Studio

With LM Studio, you can run LLMs offline on your machine. You can use Custom Pilot to integrate your LLM model running in LM Studio into VS Code.

Firstly, follow this guide to run a local webserver in LM Studio.

Then, set the API server URL to http://<ip_address>:1234 in the Sidebar panel of Custom Pilot, where <ip_address> is the IP address of the machine running LM Studio. For example, if you are running LM Studio on your local machine, you can set the URL to http://localhost:1234.

Choose an inference model from the dropdown menu, and you are ready to use your LLM model in VS Code!

About

Integrate your custom code completion model into VS Code

Resources

Stars

Watchers

Forks

Releases

No releases published

Packages

No packages published