Content
<div align="center">
<h1>
console-chat-gpt v6
</h1>
**Your Ultimate CLI Companion for Chatting with AI Models**
_Enjoy seamless interactions with **OpenAI**, **MistralAI**, **Anthropic**, **xAI**, **Google AI**, **DeepSeek**, **Alibaba**, **Inception** or **Ollama**-hosted models directly from your command line. </br>Elevate your chat experience with efficiency and ease._
<h3>
[Homepage](https://github.com/amidabuddha/consoleChatGPT) | [Examples](/examples)
</h3>
<h4 align="center">
<a href="https://github.com/amidabuddha/consoleChatGPT/blob/main/LICENSE.md">
<img src="https://img.shields.io/github/license/amidabuddha/consoleChatGPT" alt="Released under the Apache license." />
</a>
<img src="https://img.shields.io/badge/Python-3.10+-blue" alt="Working on Python 3.10+" />
<img src="https://img.shields.io/github/stars/amidabuddha/consoleChatGPT"/>
<img src="https://img.shields.io/github/issues/amidabuddha/consoleChatGPT"/>
<img src="https://img.shields.io/github/forks/amidabuddha/consoleChatGPT"/>
<img src="https://img.shields.io/badge/platform-Linux%20%7C%20macOS-blue"/>
</h4>
</div>
---
# Table of Contents
- [Features](#features)
- [Installation and Usage](#installation-and-usage)
- [Examples](#examples)
- [Contributing](#contributing)
---
**DISCLAIMER:**
The intention and implementation of this code are entirely unconnected and unrelated to OpenAI, MistralAI, Anthropic, xAI, Google AI, DeepSeek, Alibaba, Inception, or any other related parties. There is no affiliation
or relationship with OpenAI, MistralAI, Anthropic, xAI, Google, DeepSeek, Alibaba, Inception or their subsidiaries in any form.
---
## Features
- :new: [**OpenAI Responses API**](https://platform.openai.com/docs/api-reference/responses) supported. :new:
- :new: Run any OpenAI SDK compatible model - just add the model structure with the relevant `model_name` and `base_url` to the `config.toml` file. :new:
- :star: Run [**Ollama**](https://ollama.com) hosted models locally. _Ollama should be installed and the selected models already downloaded_ :star:
- :star: [**Anthropic Prompt caching**](https://docs.anthropic.com/en/docs/build-with-claude/prompt-caching) Fully supported :star:
- :star: [Model Context Protocol (MCP)](https://modelcontextprotocol.io) supported! If you are already using MCP servers just copy your `claude_desktop_config.json` to the root directory and rename to `mcp_config.json` to start using with any model! :star:
- Unified chat completion function separated as independent library to be used in any application for seamless cross-provider API experience. The source code is available in [Python](https://github.com/amidabuddha/unichat) and [TypeScript](https://github.com/amidabuddha/unichat-ts)
- Streaming with all supported models, disabled by default, may be enabled in `settings` menu
- [**OpenAI Assistants Beta**](https://platform.openai.com/docs/assistants/overview) fully supported
- **AI Managed mode** Based on the complexity of the task, automatically determines which model to use.
- **Configuration File**: Easily customize the app's settings through the `config.toml` file for complete control over
how the app works. Also supported in-app via the `settings` command.
- **Role selection**: Users can define the role of the AI in the conversation, allowing for a more personalized and
interactive experience.
- **Temperature Control**: Adjust the temperature of generated responses to control creativity and randomness in the
conversation.
- **Command Handling**: The app responds to various commands entered by the user for easy and intuitive interaction.
- **Image input**: with selected models.
- **Error Handling**: Clear and helpful error messages to easily understand and resolve any issues.
- **Conversation History**: Review previous interactions and save conversations for future reference, providing context
and continuity.
- **Graceful Exit**: Smoothly handle interruptions, ensuring conversations are saved before exiting to avoid loss of
progress.
- **A nice team**: Actively adding features, open for ideas and fixing bugs.
Overall, this app focuses on providing a user-friendly and customizable experience with features that enhance
personalization, control, and convenience.
---
## Installation and Usage
> The script works fine on Linux and MacOS terminals. For Windows it's recommended to use [WSL](https://learn.microsoft.com/en-us/windows/wsl/).
1. Clone the repository:
```shell
git clone https://github.com/amidabuddha/console-chat-gpt.git
```
2. Go inside the folder:
```shell
cd console-chat-gpt
```
3. Install the necessary dependencies:
```shell
python3 -m pip install -r requirements.txt
```
4. Get your API key from [OpenAI](https://platform.openai.com/account/api-keys), [MistralAI](https://console.mistral.ai/user/api-keys/), [Anthropic](https://console.anthropic.com/settings/keys), [xAI](https://console.x.ai/), [Google AI Studio](https://aistudio.google.com/apikey), [DeepSeek](https://platform.deepseek.com/api_keys), [Alibaba](https://bailian.console.alibabacloud.com/?apiKey=1#/api-key), [Inception](https://platform.inceptionlabs.ai/dashboard/api-keys) depending on your selected LLM.
5. The `config.toml.sample` will be automatically copied into `config.toml` upon first run, with a prompt to enter your API key/s. Feel free to change any of the other defaults that are not available in the `settings` in-app menu as per your needs.
6. Run the executable:
```shell
python3 main.py
```
> Pro-tip:
> Create an alias for the executable to run from anywhere.
7. Use the `help` command within the chat to check the available options.
8. Enjoy
---
### Configurable Options in `config.toml`
| [chat.defaults] | Main properties to generate a chat completion/response. |
|-|-|
| temperature | What sampling temperature to use, between 0 and 2. Higher values like 0.8 will make the output more random, while lower values like 0.2 more focused. May be set for each new chat session if `adjust_temperature` in [chat.features] is **true**. |
| system_role | A system (or developer) message inserted into the model's context. Should be one of the listed in [chat.roles] section. May be set for each new chat session if `role_selector` in [chat.features] is **true**. |
| model | Model ID used to generate the chat completion/response, like gpt-4o or o3. Should be listed in the [chat.models] section, with relevant parameters. May be set for each new chat session if `model_selector` in [chat.features] is **true**. |
| [chat.features] | Configurable options of the chat application. Some are accessible from within a chat session via the `settings` command.|
|-|-|
| model_selector | A selection list of models available in section [chat.models] of `config.toml`. When **true** this list may be modified at the beginning of each new chat session. |
| adjust_temperature | Prompt to change the temperature for each chat session. When **true** the temperature value may be modified at the beginning of each new chat session. |
| role_selector | A selection list of roles available in section [chat.roles] of `config.toml`. When **true** this list may be modified at the beginning of each new chat session.|
| save_chat_on_exit | When **true** automatically save the chat session upon using the `exit` command in chat.|
| continue_chat | When **true** offers a list of previously saved chat sessions to be continued in a new session. The list may be modified from within a chat session via the `chats` command.|
| ~~debug~~ | Application logging - not yet implemented. |
| disable_intro_help_message | All chat commands available in `help` are printed upon chat initialization. This is targeted at new users and may be disabled by setting to **false**. |
| assistant_mode | Enable **Open AI Assistants API** as an available selection upon chat initialization. |
| ai_managed | Enable **AI Managed mode** to allow a model to automatically select the best model according to your prompt. Detailed settings below. |
| streaming | If set to **true**, the model response data will be streamed to the client. |
| mcp_client | Setting to **false** will prevent the default initialization of MCP servers for each chat if not needed. |
| [chat.managed] | Settings dedicated to the AI Managed mode. *Not available to be edited from within a chat session*. |
|-|-|
| assistant | The preferred model that will evaluate your prompt and select the best available model out of the four configured below to handle it. Should be listed in the [chat.models] section, with relevant parameters. |
| assistant_role | Custom instruction to the evaluation model. *Change this only if you know exactly what you are doing!* |
| assistant_generalist | Your preferred general purpose model, typically the one you use the most for any type of queries. Should be listed in the [chat.models] section, with relevant parameters. |
| assistant_fast | When speed is preferred to accuracy. Should be listed in the [chat.models] section, with relevant parameters. |
| assistant_thinker | A reasoning model for complex tasks. Should be listed in the [chat.models] section, with relevant parameters. |
| assistant_coder | Your preferred model to handle Coding and Math questions. Should be listed in the [chat.models] section, with relevant parameters. |
| prompt | When **AI Managed mode** is used frequently the Y/N prompt may be disabled by changing this to **false**. |
---
## Examples
- Prompt example:

- Markdown visualization example:

- Settings and help:

> You can find more examples on our [Examples page](EXAMPLES.md).
---
## Contributing
Contributions are welcome! If you find any bugs, have feature requests, or want to contribute improvements, please open an issue or submit a pull request.