import asyncio
from openai import AsyncOpenAI
client = AsyncOpenAI(
base_url="http://api.telaos.com/v1",
api_key="YOUR_API_KEY",
)
MODEL = "llama3.2:3b"
async def get_response_stream(model, prompt):
messages = [
{"role": "system", "content": "You are a helpful assistant."},
{"role": "user", "content": "What is deep learning?"},
]
try:
response_generator = await client.chat.completions.create(
model=model,
messages=messages,
stream=True,
max_tokens=1000,
temperature=0.1,
)
complete_response = (
""
)
async for response_chunk in response_generator:
complete_response += response_chunk.chunk
print("Complete response:", complete_response)
except Exception as e:
print(f"An error occurred: {e}")
async def main():
prompts = [f"What is 1+{i}" for i in range(1)]
tasks = [get_response_stream(MODEL, prompt) for prompt in prompts]
await asyncio.gather(*tasks)
if __name__ == "__main__":
asyncio.run(main())