Spaces:
Sleeping
Sleeping
Commit
·
ee609d9
1
Parent(s):
81a3ac1
Switch to GPT-2 model for better API reliability
Browse files
app.py
CHANGED
|
@@ -333,7 +333,7 @@ class TextilindoAI:
|
|
| 333 |
def __init__(self):
|
| 334 |
self.api_key = os.getenv('HUGGINGFAC_API_KEY_2')
|
| 335 |
# Use available model with your API key
|
| 336 |
-
self.model = os.getenv('DEFAULT_MODEL', '
|
| 337 |
self.system_prompt = self.load_system_prompt()
|
| 338 |
self.data_loader = TrainingDataLoader()
|
| 339 |
|
|
@@ -491,11 +491,13 @@ Minimum purchase is 1 roll (67-70 yards)."""
|
|
| 491 |
return self.get_fallback_response(user_message)
|
| 492 |
|
| 493 |
try:
|
| 494 |
-
# Use
|
| 495 |
if "dialogpt" in self.model.lower():
|
| 496 |
prompt = f"User: {user_message}\nAssistant:"
|
|
|
|
|
|
|
| 497 |
else:
|
| 498 |
-
# Fallback
|
| 499 |
prompt = f"User: {user_message}\nAssistant:"
|
| 500 |
|
| 501 |
logger.info(f"Using model: {self.model}")
|
|
@@ -531,8 +533,8 @@ Minimum purchase is 1 roll (67-70 yards)."""
|
|
| 531 |
logger.info(f"Raw AI response: {response[:200]}...")
|
| 532 |
|
| 533 |
# Clean up the response based on model type
|
| 534 |
-
if "dialogpt" in self.model.lower():
|
| 535 |
-
# Clean up DialoGPT response
|
| 536 |
if "Assistant:" in response:
|
| 537 |
assistant_response = response.split("Assistant:")[-1].strip()
|
| 538 |
else:
|
|
@@ -541,7 +543,7 @@ Minimum purchase is 1 roll (67-70 yards)."""
|
|
| 541 |
# Remove any remaining conversation markers
|
| 542 |
assistant_response = assistant_response.replace("User:", "").replace("Assistant:", "").strip()
|
| 543 |
else:
|
| 544 |
-
# Clean up
|
| 545 |
if "Assistant:" in response:
|
| 546 |
assistant_response = response.split("Assistant:")[-1].strip()
|
| 547 |
else:
|
|
|
|
| 333 |
def __init__(self):
|
| 334 |
self.api_key = os.getenv('HUGGINGFAC_API_KEY_2')
|
| 335 |
# Use available model with your API key
|
| 336 |
+
self.model = os.getenv('DEFAULT_MODEL', 'gpt2')
|
| 337 |
self.system_prompt = self.load_system_prompt()
|
| 338 |
self.data_loader = TrainingDataLoader()
|
| 339 |
|
|
|
|
| 491 |
return self.get_fallback_response(user_message)
|
| 492 |
|
| 493 |
try:
|
| 494 |
+
# Use appropriate conversation format
|
| 495 |
if "dialogpt" in self.model.lower():
|
| 496 |
prompt = f"User: {user_message}\nAssistant:"
|
| 497 |
+
elif "gpt2" in self.model.lower():
|
| 498 |
+
prompt = f"User: {user_message}\nAssistant:"
|
| 499 |
else:
|
| 500 |
+
# Fallback format for other models
|
| 501 |
prompt = f"User: {user_message}\nAssistant:"
|
| 502 |
|
| 503 |
logger.info(f"Using model: {self.model}")
|
|
|
|
| 533 |
logger.info(f"Raw AI response: {response[:200]}...")
|
| 534 |
|
| 535 |
# Clean up the response based on model type
|
| 536 |
+
if "dialogpt" in self.model.lower() or "gpt2" in self.model.lower():
|
| 537 |
+
# Clean up DialoGPT/GPT-2 response
|
| 538 |
if "Assistant:" in response:
|
| 539 |
assistant_response = response.split("Assistant:")[-1].strip()
|
| 540 |
else:
|
|
|
|
| 543 |
# Remove any remaining conversation markers
|
| 544 |
assistant_response = assistant_response.replace("User:", "").replace("Assistant:", "").strip()
|
| 545 |
else:
|
| 546 |
+
# Clean up other model responses
|
| 547 |
if "Assistant:" in response:
|
| 548 |
assistant_response = response.split("Assistant:")[-1].strip()
|
| 549 |
else:
|