NoLlama is a terminal-based interface for interacting with Google's Gemini API directly from your terminal. Inspired by Ollama, NoLlama provides a streamlined experience for chatting with Gemini models like Gemini 2.0 Flash, Gemini 2.5 Flash Preview, and Gemini 2.5 Pro Preview, etc. Ollama, Groq and OpenRouter support will be added soon.
NoLlama offers a neat terminal interface for powerful language models that aren't easily available for local execution, complete with colorful markdown rendering, multiple model choices, and efficient memory usage.
- Google Gemini Models: Access to powerful models like Gemini 2.0 Flash, Gemini 2.5 Flash Preview, and Gemini 2.5 Pro Preview.
- Multi-turn Conversations: Maintain context between prompts for more coherent conversations.
- Neat Terminal UI: Enjoy a clean and intuitive interface for your interactions.
- Live Streaming Responses: Watch responses appear in real-time as they're generated.
- Colorful Markdown Rendering: Rich text formatting and syntax highlighting in your terminal.
- Low Memory Usage: Efficient memory management makes it lightweight compared to using a browser.
- Easy Model Switching: Simply type
model
in the chat to switch between models. - Clear Chat History: Type
clear
to clear the chat history. - Exit Commands: Type
q
,quit
, orexit
to leave the chat, or use Ctrl+C or Ctrl+D.
-
API Key Configuration:
Create a
.nollama
file in your home directory with your Gemini API key:echo "GEMINI=your_api_key_here" > ~/.nollama
You can get a free API key from Google AI Studio.
-
Installation:
a. Install from PyPI (recommended):
pip install nollama
b. Or clone and install from source:
git clone https://github.com/spignelon/nollama.git cd nollama pip install -e .
-
Run NoLlama:
nollama
- Select a Model: At startup, choose from available Gemini models.
- Chat Normally: Type your questions and see the responses with rich formatting.
- Switch Models: Type
model
in the chat to choose a different model. - Clear Chat: Type
clear
to clear the chat history. - Exit: Type
q
,quit
, orexit
to leave the chat, or press Ctrl+C or Ctrl+D.
- Add context window
- Web interface
- Add support for Groq
- Add support for OpenRouter
- Add support for Ollama API
- Support for custom APIs
Contributions are welcome! If you have suggestions for new features or improvements, feel free to open an issue or submit a pull request.
NoLlama is not affiliated with Ollama. It is an independent project inspired by the concept of providing a neat terminal interface for interacting with language models.
This project is licensed under the GPL-3.0 License.