From 92a1cd92e36118748ce9a2293a774396f9dc094e Mon Sep 17 00:00:00 2001 From: ParthSareen Date: Wed, 20 Nov 2024 14:34:09 -0800 Subject: [PATCH] Delete extra async examples --- examples/README.md | 1 - examples/async-list.py | 21 --------------------- examples/async-ps.py | 40 ---------------------------------------- 3 files changed, 62 deletions(-) delete mode 100644 examples/async-list.py delete mode 100644 examples/async-ps.py diff --git a/examples/README.md b/examples/README.md index 96cbe97..38e114f 100644 --- a/examples/README.md +++ b/examples/README.md @@ -33,7 +33,6 @@ python3 .py ### Ollama List - List all downloaded models and their properties - [list.py](list.py) -- [async-list.py](async-list.py) ### Ollama ps - Show model status with CPU/GPU usage diff --git a/examples/async-list.py b/examples/async-list.py deleted file mode 100644 index 1b69a94..0000000 --- a/examples/async-list.py +++ /dev/null @@ -1,21 +0,0 @@ -import asyncio -import ollama - - -async def main(): - client = ollama.AsyncClient() - - response = await client.list() - for model in response.models: - if model.details: - print(f'Name: {model.model}') - print(f'Size (MB): {(model.size.real / 1024 / 1024):.2f}') - print(f'Format: {model.details.format}') - print(f'Family: {model.details.family}') - print(f'Parameter Size: {model.details.parameter_size}') - print(f'Quantization Level: {model.details.quantization_level}') - print('-' * 50) - - -if __name__ == '__main__': - asyncio.run(main()) diff --git a/examples/async-ps.py b/examples/async-ps.py deleted file mode 100644 index 843ae34..0000000 --- a/examples/async-ps.py +++ /dev/null @@ -1,40 +0,0 @@ -import asyncio -from ollama import AsyncClient - - -async def main(): - client = AsyncClient() - - response = await client.pull('llama3.1', stream=True) - progress_states = set() - async for progress in response: - if progress.get('status') in progress_states: - continue - progress_states.add(progress.get('status')) - print(progress.get('status')) - - print('\n') - - response = await client.chat('llama3.1', messages=[{'role': 'user', 'content': 'Hello!'}]) - print(response['message']['content']) - - print('\n') - - response = await client.ps() - - name = response['models'][0]['name'] - size = response['models'][0]['size'] - size_vram = response['models'][0]['size_vram'] - - if size == size_vram: - print(f'{name}: 100% GPU') - elif not size_vram: - print(f'{name}: 100% CPU') - else: - size_cpu = size - size_vram - cpu_percent = round(size_cpu / size * 100) - print(f'{name}: {cpu_percent}% CPU/{100 - cpu_percent}% GPU') - - -if __name__ == '__main__': - asyncio.run(main())