|
# Yura LLM Client for Katya server
|
|
|
|
Part of project with as target replacing the native ollama protocol. This protocol supports streaming and is usable trough https and it is possible to directly attach a web client to the backend.
|
|
|
|
## Install
|
|
```bash
|
|
pip install -e .
|
|
```
|
|
|
|
## Build
|
|
```bash
|
|
make build
|
|
```
|
|
|
|
## Command line usage
|
|
```bash
|
|
yura ws://[host]:[port]/[path]/
|
|
```
|
|
|
|
## Python
|
|
```python
|
|
import asyncio
|
|
|
|
from yura.client import AsyncClient
|
|
|
|
|
|
async def communicate():
|
|
client = AsyncClient("ws://[host]:[port]/[path]/")
|
|
async for response in client.chat("Your prompt"):
|
|
print(response)
|
|
|
|
|
|
asyncio.run(communicate())
|
|
```
|