Coding Assistants and Artificial Intelligence for the Rest of Us


Introduction

Everyone is familiar with AI (acronym for Artificial Intelligence), at least the term. It’s a field that has been around for decades and it’s still growing. One of the most exciting things about AI is how it can be used to unlock the Developer Productivity with the invent of coding assistants and AI-powered tools.

The first coding assistant came out of nowhere from GitHub named Copilot, though in the beginning it generated a lot of controversy about how it utilizes the open source repositories to train itself. And the second one is called ChatGPT which has been around since 2018 and it’s still growing. It can answer questions, write and debug code, summarize text, translate languages, play games, etc. This list of AI engines continues to grow and evolve.

But the true power of coding assistants is that they can provide the inline completion as you are typing the code. For example; let’s say we have some Rust code that creates a variable finland of type Country that in turns holds a vector of City, the Code assistant can give you the following suggestion:

finland-autocomplete

As you can witness, the AI automatically figured out that you most likely are creating Finland country, and as the AI has the code context, it automatically suggested a vector of City for you.

Here are two more examples where AI helped to complete the code you most likely want to write:

  1. Assistant automatically suggesting that you might like to create a new function to instantiate a new instance of City.

    city-impl

  2. As in the first screenshot we can see that AI is suggesting to use from function to create a Country out of a vector of City. That is happening because I implemented From trait for the Country struct, again with the help of Code Assistant. Here is a step by step Assistant suggestions which I accepted just by pressing Tab key.

a. impl-from-for-country b. country-from-suggestion c. from-impl-suggestion

What to choose? Local or Remote?

We saw how much helpful a code assistant can be, but the real question is should you use service like Supermaven, Codium, Copilot or Cursor and others ? Or just run the open source LLMs on your local computer?

💡 BTW LLM stands for Large Language Model while LLaMA stands for Large Language Model Meta AI.

Running a local LLM can help if you are super conscious that you don’t want to share your information in real time to a third party company, though some claim that they delete it after few days. But, it incurs resources usage cost as LLMs are RAM and GPU/CPU incentive and can quickly dwindle down your computer’s performance in milliseconds if you are running a potato 🥔 computer.

On the other hand, using a Remote service frees you from such restrictions but may put a hole in your wallet/credit card.

Here I’ll show you how to use Open Source LLMs running on your own computer.

Ollama and Continue.dev

  1. Ollama is an open source LLM initiative that allows users to run LLMss locally, without relying on cloud providers or third-party services. It’s a great tool if you want to use powerful AI capabilities but don’t want to share your data with any third party.

    a. Let’s install it on a MacBook Pro using Homebrew:

     brew install ollama
    

    b. Now run it using the following command. It’ll spin up a GO web server, that would pass the response to a running LLM and will give the response back to you.

     ollama serve
    

    c. Install an LLM. But wait, which LLM to choose? There are various LLMs available for various tasks. We would focus on two LLMs, one is Meta’s llama3.1:8b and the other one is codeqwen:7b. These are reasonably sized and performant models. Moreover, these will not consume much of your computer’s resources.

    There is a reason I chose two models as both are specialized in different tasks. llama3.1 is good for general use but codeqwen is good as a coding assistant. Think of them as ChatGPT and Copilot respectively. You can easily switch between them while chatting inside Continue extension/plugin.

    Run LLMs using the following commands. Of course, these commands will download the models first if not available locally.

     ollama run llama3.1:8b
    
     ollama run codeqwen:7b
    

    d. Assuming your codeqwen model is installed and running we can interact with it in the terminal just like you do with a normal ChatGPT client.

    codeqwen-cmd-iteraction

    Or, you can even choose to use a GUI application in a browser with a ChatGPT like interface.

    codeqwen-open-web-ui-interaction

    I’ll tell you later how to run this GUI application below. For now, let’s talk about Continue.dev.

  2. Running an LLM is different than to interact with it in an IDE or TextEditor. Even if you have a running LLM which may respond to your queries just like ChatGPT does, it is not necessary that these will provide editor integrations, this is where Continue.dev comes into picture. Continue provides VSCode extension and JetBrains plugin to interact with your local LLM. All we need to do is install and configure it.

    a. From the Extension’s dropdown select the Add Chat Model option and select the values shown in the screenshot:

    continue-extension-model-setup continue-extension-select-provider

    b. If required you can even change the settings manually by editing the Continue’s ~/.continue/config.json file. Here is an important excerpt of the configuration. Notice that the tabAutocompleteModel configuration selects the LLM that will be used for autocompletion in the editor.

     {
         "models": [
             {
               "title": "Llama 3.1",
               "provider": "ollama",
               "model": "llama3.1-70b"
             },
             {
               "title": "Codeqwen",
               "provider": "ollama",
               "model": "codeqwen:latest"
             },
         ],
         "tabAutocompleteModel": {
             "title": "Codeqwen",
             "provider": "ollama",
             "model": "codeqwen:latest"
         }
     }
    

That concludes the Ollama and Continue setup. Now let’s look at how you can get the ChatGPT like browser application running on your computer so that you don’t need to deal with the command line.

Open WebUI

Open WebUI provides the ChatGPT like interface and can be easily installed. It’s recommended that you use Docker to run this web app in a container.

  1. Install Docker from the official website or Using Homebrew.
  2. Use following command to pull and run the Open WebUI image. Notice that we are passing WEBUI_AUTH=false environment variable which will disable the authentication. If you need the authentication then just remove this environment variable from the command.
    docker run -e WEBUI_AUTH=false -d -p 5000:8080 --add-host=host.docker.internal:host-gateway -v open-webui:/app/backend/data --name open-webui --restart always ghcr.io/open-webui/open-webui:main
    
  3. The application will now readily be available to chat on http://localhost:5000 as we are mapping the internal 8080 port to host’s 5000 port.

Ollama CLI

Ollama CLI has some helpful commands that can be accessed by running ollama --help command. Here are two of the most used ones:

  1. Ollama ls will list all the installed models.

    ollama-ls

  2. ollama ps will list all the explicitly running models.

    ollama-ps

This concludes the Coding Assistants and Artificial Intelligence for the Rest of Us.




Thanks for stopping by. See you next time.


© 2023. Creative Commons. Thakur Manvendra Singh.

Powered by Hydejack v6.3.0