Skip to content

bamarler/Larry

Folders and files

NameName
Last commit message
Last commit date

Latest commit

 

History

41 Commits
 
 
 
 
 
 
 
 
 
 
 
 
 
 

Repository files navigation

Larry

Description

Larry is an AI assistant that runs opensource Large Language Models locally on consumer grade hardware through Ollama. It is designed to be an assistant for any task you might have such as coding, research, writing, etc. The window stays on top of all of your applications so you can easily reference and chat with your model.

image

How to Use

  • Before attempting to use, make sure you have a working GPU and NVIDIA CUDA installed. Otherwise the models will run strictly using your CPU and performance will falter.
  • Clone the repository
  • Create a new python environment
  • Run pip install -r requirements.txt
  • Run the following script: src/backend/main.py
  • Additionally, you can run pyinstaller --onefile build-larry.spec to generate an executable.

Features

  • You can install any available model from Ollama depending on your available memory and computer specs. You can also remove any old models that you don't use anymore to make space.

image

  • All chats are stored in sqlite3 database and the model is fed all previous context from the current chat each time.

  • If the model you are using supports markdown features, they will be rendered appropriately in an HTML window

  • You can choose to pin the window on the left or right side depending on your preference

  • You can create multiple personas with custom names, roles, personalities, and behaviors to be masked over any model

image

  • For those who want to modify specific model details, you have the option to change the models the temperature, penalties, and other attributes.

image

Tech Stack

The majority of the Program is coded in Python, and frontend is designed using the TKinter library. Models and model requests are handled via Ollama's python api library. Chats are stored in a sqlite3 database.

About

Larry is an AI assistant that runs opensource Large Language Models locally on consumer grade hardware through Ollama. It is designed to be an assistant for any task you might have such as coding, research, writing, etc.

Resources

License

Stars

Watchers

Forks

Releases

No releases published

Packages

 
 
 

Contributors

Languages