Local chatbot testing.
This commit is contained in:
42
test/advanced_example.py
Normal file
42
test/advanced_example.py
Normal file
@@ -0,0 +1,42 @@
|
||||
# advanced_example.py
|
||||
from ollama_deepseek_client import OllamaDeepSeekClient
|
||||
|
||||
def advanced_example():
|
||||
client = OllamaDeepSeekClient()
|
||||
|
||||
# Example 1: Code generation
|
||||
code_prompt = "Write a Python function to calculate fibonacci numbers"
|
||||
print("Code Generation Example:")
|
||||
code_response = client.generate_response(
|
||||
code_prompt,
|
||||
temperature=0.3, # Lower temperature for more deterministic code
|
||||
max_tokens=1000
|
||||
)
|
||||
print(code_response)
|
||||
print("\n" + "="*50 + "\n")
|
||||
|
||||
# Example 2: Streaming response
|
||||
print("Streaming Response Example:")
|
||||
stream_response = client.generate_response(
|
||||
"Explain quantum computing in simple terms",
|
||||
stream=True,
|
||||
temperature=0.8
|
||||
)
|
||||
print("\n" + "="*50 + "\n")
|
||||
|
||||
# Example 3: Chat with context
|
||||
conversation = []
|
||||
|
||||
messages = [
|
||||
"What's the capital of France?",
|
||||
"Tell me more about its history",
|
||||
"What are the main tourist attractions there?"
|
||||
]
|
||||
|
||||
for msg in messages:
|
||||
print(f"You: {msg}")
|
||||
response, conversation = client.chat(msg, conversation)
|
||||
print(f"Assistant: {response}\n")
|
||||
|
||||
if __name__ == "__main__":
|
||||
advanced_example()
|
||||
Reference in New Issue
Block a user