Spaces:
Paused
Paused
| import sys | |
| import os | |
| import io | |
| sys.path.insert(0, os.path.abspath("../..")) | |
| from litellm import completion | |
| import litellm | |
| litellm.success_callback = ["langsmith"] | |
| litellm.set_verbose = True | |
| import time | |
| def test_langsmith_logging(): | |
| try: | |
| response = completion( | |
| model="claude-instant-1.2", | |
| messages=[{"role": "user", "content": "what llm are u"}], | |
| max_tokens=10, | |
| temperature=0.2, | |
| ) | |
| print(response) | |
| except Exception as e: | |
| print(e) | |
| # test_langsmith_logging() | |
| def test_langsmith_logging_with_metadata(): | |
| try: | |
| response = completion( | |
| model="gpt-3.5-turbo", | |
| messages=[{"role": "user", "content": "what llm are u"}], | |
| max_tokens=10, | |
| temperature=0.2, | |
| metadata={ | |
| "run_name": "litellmRUN", | |
| "project_name": "litellm-completion", | |
| }, | |
| ) | |
| print(response) | |
| except Exception as e: | |
| print(e) | |
| # test_langsmith_logging_with_metadata() | |
| def test_langsmith_logging_with_streaming_and_metadata(): | |
| try: | |
| response = completion( | |
| model="gpt-3.5-turbo", | |
| messages=[{"role": "user", "content": "what llm are u"}], | |
| max_tokens=10, | |
| temperature=0.2, | |
| metadata={ | |
| "run_name": "litellmRUN", | |
| "project_name": "litellm-completion", | |
| }, | |
| stream=True, | |
| ) | |
| for chunk in response: | |
| continue | |
| except Exception as e: | |
| print(e) | |
| test_langsmith_logging_with_streaming_and_metadata() | |