From cde66da9838ba31d81909029c13bbf24a8c9be81 Mon Sep 17 00:00:00 2001 From: retoor Date: Mon, 20 Jan 2025 13:36:59 +0100 Subject: [PATCH] Update docs. --- README.md | 5 ++++- 1 file changed, 4 insertions(+), 1 deletion(-) diff --git a/README.md b/README.md index 96d3fc9..2c83f49 100644 --- a/README.md +++ b/README.md @@ -40,7 +40,10 @@ It's also possible to give `api_key` as parameter to the initiation of the `rage --- -## Costs +## For free! +But there is a small catch. It's very easy to replace Open AI with a local hosted LLM like Olama. Olama is installed within minutes using a one liner. You will figure the replacement of the URL out in 20 minutes or so. So, with a good hour + time to download your favorite Olama LLM you have a chatbot for free in a few hours wherefrom most is waiting. I recommend models above 3b of even above 7b. My personal experience with Olama LLM's is that llama models / qwen (3b+) / gemma2 works the best. Gemma2 is made by Google and is only 2b and 4gb or so. Gemma2 is probably the most for less. You can try it out with the `python -m ragent.demo_olama` command. Just kiddin'. You really have to do this small thing yourself. I don't have the right hardware to run a decent LLM so i just didn't implement it. Don't be cheap AND lazy. It's worth it. + +## Costs if you use Open AI You can chat the whole day with the bots for just ten cents or so. See here how much it costed to test it extensively. ![Intense usage costs.](costs.png)