ctrl+shift+p filters: :st2 :st3 :win :osx :linux
Browse

Open​AI completion

First class Sublime Text AI assistant with o3, o4-mini, gpt-4.1 and ollama support!

Details

Installs

  • Total 7K
  • Win 3K
  • Mac 3K
  • Linux 1K
Jun 3 Jun 2 Jun 1 May 31 May 30 May 29 May 28 May 27 May 26 May 25 May 24 May 23 May 22 May 21 May 20 May 19 May 18 May 17 May 16 May 15 May 14 May 13 May 12 May 11 May 10 May 9 May 8 May 7 May 6 May 5 May 4 May 3 May 2 May 1 Apr 30 Apr 29 Apr 28 Apr 27 Apr 26 Apr 25 Apr 24 Apr 23 Apr 22 Apr 21 Apr 20 Apr 19
Windows 6 7 2 5 8 15 10 8 4 3 1 7 9 5 12 5 4 2 10 3 4 6 2 11 3 14 4 10 11 7 2 2 3 0 8 8 7 3 3 7 6 8 5 8 5 5
Mac 4 6 4 4 5 3 13 7 4 6 9 9 14 6 5 3 5 2 5 13 4 6 6 4 1 5 7 5 3 7 5 4 6 4 10 2 8 4 0 7 7 6 7 7 5 5
Linux 4 3 2 6 5 3 6 7 2 7 6 4 7 3 7 5 8 6 3 3 3 1 5 2 2 2 4 4 4 6 3 4 3 2 1 1 3 2 2 1 5 4 2 2 2 2

Readme

Source
raw.​githubusercontent.​com

Star on GitHub Package Control

OpenAI Sublime Text Plugin

tldr;

Cursor level of AI assistance for Sublime Text. I mean it.

Works with all OpenAI'ish API: llama.cpp server, ollama or whatever third party LLM hosting. Claude API support coming soon.

Features

  • Chat mode powered by whatever model you'd like.
  • o4-mini and o3 support.
  • gpt-4.1 support
  • llama.cpp's server, ollama and all the rest OpenAI'ish API compatible.
  • Dedicated chats histories and assistant settings for a projects.
  • Ability to send whole files or their parts as a context expanding.
  • Phantoms Get non-disruptive inline right in view answers from the model.
  • Markdown syntax with code languages syntax highlight (Chat mode only).
  • Server Side Streaming (SSE) streaming support.
  • Status bar various info: model name, mode, sent/received tokens.
  • Proxy support.

Requirements

  • Sublime Text 4
  • llama.cpp, ollama installed OR
  • Remote llm service provider API key, e.g. OpenAI
  • Anthropic API key [coming soon].

Installation

Via Package Control

  1. Install the Sublime Text Package Control plugin if you haven't done this before.
  2. Open the command palette and type Package Control: Install Package.
  3. Type OpenAI and press Enter.

Via Git Clone

  1. Go to your packages folder: Preferences: Browse Packages.
  2. Run git clone https://github.com/yaroslavyaroslav/OpenAI-sublime-text.git OpenAI\ completion in that folder that Sublime opened.
  3. Open Sublime Text and let it installed the dependencies.
  4. It may ask you to restart Sublime, do that if it does.
  5. Open Sublime again and type OpenAI and press Enter.

[!NOTE] Highly recommended complimentary packages: - https://github.com/SublimeText-Markdown/MarkdownCodeExporter - https://sublimetext-markdown.github.io/MarkdownEditing

Usage

Interacting with the AI

You can interact with the AI in several ways, primarily through commands available in the Sublime Text Command Palette:

  1. Select Text (Optional): You can select a region of text in your current file to be included as part of the context for your prompt.
  2. Choose Your Command:
    • OpenAI: Chat Model Select: This is the most flexible command. It opens a panel allowing you to:
      • Choose a specific “assistant” (which defines the model, API key, temperature, etc.).
      • Select an “output_mode” (inline Phantom or a chat View in a panel/new tab). This command automatically includes any files you've marked for context (see “Additional Request Context Management” below).
    • OpenAI: New Message: This command sends your input directly using the assistant and output mode that were last selected or are currently active. It's quicker if you're consistently using the same settings. This command also includes any files marked for context.
  3. Input Your Prompt: An input panel will appear, allowing you to type your question or instruction for the AI.
  4. View Response:
    • The AI's response will typically appear in the OpenAI output panel.
    • If you chose “Phantom” mode (with OpenAI: Chat Model Select), the response will appear as an inline overlay.
    • You can switch to a dedicated tab for the chat using the OpenAI: Open in Tab command.

Including Build/LSP Output: For more specific contexts, especially when coding, you can use commands that automatically include output from Sublime Text's diagnostic panels: * OpenAI: New Message With Build Output * OpenAI: Chat Model Select With Build Output * OpenAI: New Message With LSP Output * OpenAI: Chat Model Select With LSP Output

These commands will append recent lines from the respective output panels (Build results or LSP diagnostics) to your request. The number of lines included can be configured with the build_output_limit setting in openAI.sublime-settings. This is useful for asking the AI to explain errors, debug code, or summarize diagnostics.

Managing Chat Sessions: * OpenAI: Refresh Chat*: Reloads the chat history into the output panel or tab. * *OpenAI: Reset Chat History: Clears the chat history for the current context (project-specific or global).

Chat history management

You can separate a chat history and assistant settings for a given project by appending the following snippet to its settings:

{
    "settings": {
        "ai_assistant": {
            "cache_prefix": "/absolute/path/to/project/"
        }
    }
}

Additional Request Context Management

You can include the content of specific files as context for the AI. Files marked for context will have their content sent along with your prompt. There are several ways to manage this:

  • Using the Command Palette:
    • Run the OpenAI: Add Sheets to Context command. If you run this while one or more tabs are selected (e.g., using Ctrl+Click or Cmd+Click on tabs, or by selecting files in the sidebar that get focused as tabs), it will toggle their inclusion in the AI context.
  • Using the Tab Context Menu:
    • Right-click on an open tab and select OpenAI: Add File to Context from the context menu to toggle its inclusion.
  • Using the Sidebar Context Menu:
    • Right-click on a file or a selection of files in the sidebar and choose OpenAI: Add File to Context from the context menu to toggle their inclusion.

Once files are added to the context: * You can see the number of currently included sheets in the status bar (if this option is enabled in the status_hint setting). * The OpenAI: Chat Model Select command preview panel will also list the files currently included. * To view all files currently marked for context in the current window, run the OpenAI: Show All Selected Sheets command from the Command Palette. This will select these files in their respective views/groups.

Files can be deselected using the same methods (the commands effectively toggle the inclusion status).

Image handling

Image handle can be called by OpenAI: Handle Image command.

It expects an absolute path to image to be selected in a buffer or stored in clipboard on the command call (smth like /Users/username/Documents/Project/image.png). In addition command can be passed by input panel to proceed the image with special treatment. png and jpg images are only supported.

[!NOTE] Currently plugin expects the link or the list of links separated by a new line to be selected in buffer or stored in clipboard only.

In-buffer llm use case

Phantom use case

Phantom is the overlay UI placed inline in the editor view (see the picture below). It doesn't affects content of the view.

  1. [optional] Select some text to pass in context in to manipulate with.
  2. Pick Phantom as an output mode in quick panel OpenAI: Chat Model Select.
  3. After the AI responds, the phantom will display actions like:
    • x: Dismisses the phantom.
    • Copy: Copies the AI's response (or just the code, if phantom_integrate_code_only is true) to the clipboard.
    • Append: Appends the AI's response to the end of your current selection in the editor (or at the cursor position if no selection).
    • Replace: Replaces your current selection with the AI's response.
    • In New Tab: Opens the AI's full response in a new tab.
    • Add to History: Saves the current interaction (your prompt and the AI's response) to the chat history panel/view.
  4. You can hit ctrl+c to stop prompting same as with in panel mode.

Other features

Open Source models support (llama.cpp, ollama)

  1. Replace "url" setting of a given model to point to whatever host you're server running on (e.g.http://localhost:8080/v1/chat/completions).
  2. Provide a "token" if your provider required one.
  3. Tweak "chat_model" to a model of your choice and you're set.

Google Gemini models

  1. Replace "url" setting of a given model to point to the Google Gemini OpenAI compatible API: https://generativelanguage.googleapis.com/v1beta/openai/chat/completions.
  2. Provide a "token" if your provider required one.
  3. Tweak "chat_model" to a model from the list of supported models.

You can read more about OpenAI compatibility in the Gemini documentation.

[!NOTE] You can set both url and token either global or on per assistant instance basis, thus being capable to freely switching between closed source and open sourced models within a single session.

Settings

The OpenAI Completion plugin has a settings file where youcan set your OpenAI API key. This is required for the most of providers to work. To set your API key, open the settings with the Preferences: OpenAI Settings command and paste your API key in the token property, as follows: You can also access these settings and the default keybindings via the main menu: Preferences -> Package Settings -> OpenAI completion.

{
    "token": "sk-your-token",
}

Advertisement disabling

To disable advertisement you have to add "advertisement": false line into an assistant setting where you wish it to be disabled.

Key Bindings

You can create custom keybindings for OpenAI commands by adding entries to your user keymap file. Access this file via the Command Palette (Preferences: Key Bindings) or the main menu (Preferences -> Key Bindings). Sublime Text keybindings often use sequences, for example, pressing super+k (or ctrl+k on Windows/Linux) followed by another key.

Here are some examples to get you started:

1. New Message with current assistant (includes files marked for context): “json { "keys”: [“super+k”, “m”], // macOS: Cmd+k, then m “command”: “openai” }

**2. Open Chat Model Select panel (includes files marked for context):**
```json
{
    "keys": ["super+k", "super+m"], // macOS: Cmd+k, then Cmd+m
    "command": "openai_panel"
}

3. New Message with Build Output (using current assistant): “json { "keys”: [“super+k”, “b”], // macOS: Cmd+k, then b “command”: “openai”, “args”: { “build_output”: true } }

**4. Toggle current file's inclusion in AI Context (matches "OpenAI: Add Sheets to Context" command):**
```json
{
    "keys": ["super+k", "c"], // macOS: Cmd+k, then c
    "command": "toggle_view_ai_context_included"
}

5. Show the AI Chat output panel: “json { "keys”: [“super+k”, “p”], // macOS: Cmd+k, then p “command”: “show_panel”, “args”: { “panel”: “output.AI Chat” } }

### Proxy support

You can setup it up by overriding the proxy property in the `OpenAI completion` settings like follow:

```js
"proxy": {
    "address": "127.0.0.1", // required
    "port": 9898, // required
    "username": "account",
    "password": "sOmEpAsSwOrD"
}

Disclaimers

[!WARNING] All selected code will be sent to the OpenAI servers (if not using custom API provider) for processing, so make sure you have all necessary permissions to do so.

[!NOTE] Dedicated to GPT3.5 that one the one who initially written at 80% of this back then. This was felt like a pure magic!