cmdh (short for Command Helper) is a tool that invokes LLM models provided by ollama or OpenAI to convert a command request into a desired command.
Use it to look up commands and flags that that you don't know offhand or generate complex commands with chaining.
cmdh_demonstration_video.webm
- Generate Linux commands from natural language
- Interactively run the commands using a hotkey menu system
- Differentiates between shell command types: interactive and non-interactive
- Supports ChatGPT, ollama, and text-generation-webui
- NodeJS - installation guide
- npm - installation guide
- tsc - installation guide
- Set up and configure cmdh using the following command:
git clone https://github.com/pgibler/cmdh.git && cd cmdh && ./install.sh
- Run it like so:
cmdh 'Output the number of lines of code committed to git last month'
- Interact with the result interface to run the setup commands, desired command, all of the commands, or quit.
NOTE: You will have to reload your .bashrc / .zshrc / etc. or open a new terminal to make the cmdh command available in the shell. In Debian / Ubuntu, this is done by running source ~/.bashrc
.
Before running cmdh, you will need to configure an LLM host and set configuration options required of that host.
- Run
cmdh configure
to start the configuration wizard. You will be asked to select an LLM host and input settings required by that host. - Run
cmdh configure show
to display your current configuration.
- Generate an OpenAI key here.
- Run
cmdh configure
and select the OpenAI option. - Select a model & input your OpenAI key.
- Install & run the ollama service & pull the codellama model using the following commands:
curl https://ollama.ai/install.sh | sh
ollama pull codellama
- Run
cmdh configure
, select the ollama option, and set 'codellama' as the model.
- Clone the repo:
git clone https://github.com/oobabooga/text-generation-webui
- Navigate to the cloned text-generation-webui folder and start the server by running
./start_linux --api --listen
. - Open the web UI for text-generation-webui (http://localhost:7860), open the "Model" tab, and in the "Download model or LoRA" form, input
Trelis/Llama-2-7b-chat-hf-function-calling-v2
and press "Download". - Click the reload button next to the Model dropdown menu under the model tab & select
llama-2-7b-function-calling.Q3_K_M.gguf
. Then click "Load" to load the model. - Run
cmdh configure
and choose the 'text-generation-webui' option.
cmdh will automatically send the prompts to whichever model is loaded by text-generation-webui.
HuggingFace model URL: https://huggingface.co/Trelis/Llama-2-7b-chat-hf-function-calling-v2
The issue tracker is mostly feature requests I have put in so I don't forget them. If you have any bug reports or good ideas, please include them in the tracker.
If you run into any issues installing or running cmdh, please open a ticket in the project tracker. Include a detailed bug report with stacktraces and inputs and the mode of operation (OpenAI or ollama).