Spaces:
Running
Running
| # TODO: implement ollama client | |
| from typing import Any, List, Optional | |
| from graphgen.bases import BaseLLMClient, Token | |
| class OllamaClient(BaseLLMClient): | |
| async def generate_answer( | |
| self, text: str, history: Optional[List[str]] = None, **extra: Any | |
| ) -> str: | |
| pass | |
| async def generate_topk_per_token( | |
| self, text: str, history: Optional[List[str]] = None, **extra: Any | |
| ) -> List[Token]: | |
| pass | |
| async def generate_inputs_prob( | |
| self, text: str, history: Optional[List[str]] = None, **extra: Any | |
| ) -> List[Token]: | |
| pass | |