-
Notifications
You must be signed in to change notification settings - Fork 49
/
assist_local.py
54 lines (44 loc) · 2.6 KB
/
assist_local.py
1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
20
21
22
23
24
25
26
27
28
29
30
31
32
33
34
35
36
37
38
39
40
41
42
43
44
45
46
47
48
49
50
51
52
53
54
from openai import OpenAI
import ollama
import time
from pygame import mixer
import os
# Initialize the OpenAI client and mixer
client = OpenAI()
mixer.init()
# Global variable to store conversation history
conversation_history = []
def ask_question_memory(question):
try:
system_message = """You are Jarvis, the AI assistant from Iron Man. Remember, I am not Tony Stark, just your commander. You are formal and helpful, and you don't make up facts, you only comply to the user requests. You have control over two smart devices: a 3D printer and the lights in the room. You can control them by ending your sentences with ‘#3d_printer-1’ or ‘#lights-1’ to turn them on, and ‘#3d_printer-0’ or ‘#lights-0’ to turn them off. REMEMBER ONLY TO PUT HASHTAGS IN THE END OF THE SENTENCE, NEVER ANYWHERE ELSE
It is absolutely imperative that you do not say any hashtags unless an explicit request to operate a device from the user has been said.
NEVER MENTION THE TIME! Only mention the time upon being asked about it. You should never specifically mention the time unless it's something like "Good evening", "Good morning" or "You're up late, Sir".
Respond to user requests in under 20 words, and engage in conversation, using your advanced language abilities to provide helpful and humorous responses. Call the user by 'Sir'"""
# Add the new question to the conversation history
conversation_history.append({'role': 'user', 'content': question})
# Include the system message and conversation history in the request
response = ollama.chat(model='llama3.1', messages=[
{'role': 'system', 'content': system_message},
*conversation_history
])
# Add the AI response to the conversation history
conversation_history.append({'role': 'assistant', 'content': response['message']['content']})
return response['message']['content']
except ollama.ResponseError as e:
print(f"An error occurred: {e}")
return f"The request failed: {e}"
def generate_tts(sentence, speech_file_path):
response = client.audio.speech.create(model="tts-1", voice="echo", input=sentence)
response.stream_to_file(speech_file_path)
return str(speech_file_path)
def play_sound(file_path):
mixer.music.load(file_path)
mixer.music.play()
def TTS(text):
speech_file_path = generate_tts(text, "speech.mp3")
play_sound(speech_file_path)
while mixer.music.get_busy():
time.sleep(1)
mixer.music.unload()
os.remove(speech_file_path)
return "done"