feat: update Venice AI model to glm-4.7-flash-heretic, increase tokens for reasoning model, inline SITREP source links
This commit is contained in:
parent
3d6a6cfdc9
commit
00e3d07743
2 changed files with 12 additions and 6 deletions
|
|
@ -723,14 +723,17 @@ def venice_chat():
|
||||||
json={
|
json={
|
||||||
'model': venice_model,
|
'model': venice_model,
|
||||||
'messages': messages,
|
'messages': messages,
|
||||||
'max_tokens': 512,
|
'max_tokens': 4096,
|
||||||
'temperature': 0.7
|
'temperature': 0.7
|
||||||
},
|
},
|
||||||
timeout=30
|
timeout=60
|
||||||
)
|
)
|
||||||
resp.raise_for_status()
|
resp.raise_for_status()
|
||||||
result = resp.json()
|
result = resp.json()
|
||||||
reply = result['choices'][0]['message']['content']
|
reply = result['choices'][0]['message'].get('content', '') or ''
|
||||||
|
# Reasoning models may put output in reasoning_content
|
||||||
|
if not reply.strip():
|
||||||
|
reply = result['choices'][0]['message'].get('reasoning_content', '') or ''
|
||||||
return jsonify({'reply': reply})
|
return jsonify({'reply': reply})
|
||||||
except req.exceptions.RequestException as e:
|
except req.exceptions.RequestException as e:
|
||||||
return jsonify({'error': f'Venice API error: {str(e)}'}), 502
|
return jsonify({'error': f'Venice API error: {str(e)}'}), 502
|
||||||
|
|
|
||||||
|
|
@ -320,14 +320,17 @@ def generate_with_venice(user_prompt):
|
||||||
{'role': 'system', 'content': SITREP_SYSTEM_PROMPT},
|
{'role': 'system', 'content': SITREP_SYSTEM_PROMPT},
|
||||||
{'role': 'user', 'content': user_prompt}
|
{'role': 'user', 'content': user_prompt}
|
||||||
],
|
],
|
||||||
'max_tokens': 2048,
|
'max_tokens': 8192,
|
||||||
'temperature': 0.7
|
'temperature': 0.7
|
||||||
},
|
},
|
||||||
timeout=60
|
timeout=120
|
||||||
)
|
)
|
||||||
resp.raise_for_status()
|
resp.raise_for_status()
|
||||||
result = resp.json()
|
result = resp.json()
|
||||||
content = result['choices'][0]['message']['content']
|
content = result['choices'][0]['message'].get('content', '') or ''
|
||||||
|
# Reasoning models may put output in reasoning_content
|
||||||
|
if not content.strip():
|
||||||
|
content = result['choices'][0]['message'].get('reasoning_content', '') or ''
|
||||||
log.info(f'Venice AI response received ({len(content)} chars)')
|
log.info(f'Venice AI response received ({len(content)} chars)')
|
||||||
return content, venice_model
|
return content, venice_model
|
||||||
except Exception as e:
|
except Exception as e:
|
||||||
|
|
|
||||||
Loading…
Add table
Reference in a new issue