Maga222006
commited on
Commit
·
8cb3d2a
1
Parent(s):
2feeb1e
MultiagentPersonalAssistant
Browse files- agent/__pycache__/models.cpython-312.pyc +0 -0
- agent/__pycache__/prompts.cpython-312.pyc +0 -0
- agent/models.py +6 -3
- agent/prompts.py +15 -13
- database_files/main.db +0 -0
- mediafiles/167bb9e0-f228-4c14-b13e-6f042b3b610c.ogg +0 -0
- mediafiles/8d7ccab5-a3ae-4c32-b1d5-bc1f877eee40.ogg +0 -0
- requirements.txt +1 -0
agent/__pycache__/models.cpython-312.pyc
CHANGED
|
Binary files a/agent/__pycache__/models.cpython-312.pyc and b/agent/__pycache__/models.cpython-312.pyc differ
|
|
|
agent/__pycache__/prompts.cpython-312.pyc
CHANGED
|
Binary files a/agent/__pycache__/prompts.cpython-312.pyc and b/agent/__pycache__/prompts.cpython-312.pyc differ
|
|
|
agent/models.py
CHANGED
|
@@ -15,13 +15,16 @@ llm_supervisor = init_chat_model(
|
|
| 15 |
)
|
| 16 |
|
| 17 |
llm_peripheral = init_chat_model(
|
| 18 |
-
|
|
|
|
| 19 |
)
|
| 20 |
|
| 21 |
llm_agents = init_chat_model(
|
| 22 |
-
model="groq:qwen/qwen3-32b"
|
|
|
|
| 23 |
)
|
| 24 |
|
| 25 |
llm_image = init_chat_model(
|
| 26 |
-
model="groq:meta-llama/llama-4-scout-17b-16e-instruct"
|
|
|
|
| 27 |
)
|
|
|
|
| 15 |
)
|
| 16 |
|
| 17 |
llm_peripheral = init_chat_model(
|
| 18 |
+
model="groq:gemma2-9b-it",
|
| 19 |
+
max_tokens=4000
|
| 20 |
)
|
| 21 |
|
| 22 |
llm_agents = init_chat_model(
|
| 23 |
+
model="groq:qwen/qwen3-32b",
|
| 24 |
+
max_tokens=4000
|
| 25 |
)
|
| 26 |
|
| 27 |
llm_image = init_chat_model(
|
| 28 |
+
model="groq:meta-llama/llama-4-scout-17b-16e-instruct",
|
| 29 |
+
max_tokens=3000
|
| 30 |
)
|
agent/prompts.py
CHANGED
|
@@ -23,22 +23,24 @@ deep_research_instructions = f"""You are a deep research expert. Follow these st
|
|
| 23 |
"""
|
| 24 |
|
| 25 |
def supervisor_instructions(tools: list, agents: list):
|
| 26 |
-
return f"""You are the
|
| 27 |
-
Your role is to interpret the user’s request and decide the best way to fulfill it by delegating tasks to available tools and agents.
|
| 28 |
|
| 29 |
-
You have access to the following:
|
|
|
|
|
|
|
| 30 |
|
| 31 |
-
|
| 32 |
-
-
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 33 |
|
| 34 |
-
|
| 35 |
-
1. Analyze the user’s request carefully.
|
| 36 |
-
2. Choose the most suitable tools and/or agents for the task.
|
| 37 |
-
3. Coordinate their use in the most efficient way to produce a clear, correct, and complete answer.
|
| 38 |
-
4. If multiple steps are required, break them down logically.
|
| 39 |
-
5. Always ensure the final output is directly useful to the user.
|
| 40 |
-
|
| 41 |
-
Your primary objective: **satisfy the user’s query as effectively as possible using the available resources.**
|
| 42 |
"""
|
| 43 |
|
| 44 |
def coder_system_message(state: dict):
|
|
|
|
| 23 |
"""
|
| 24 |
|
| 25 |
def supervisor_instructions(tools: list, agents: list):
|
| 26 |
+
return f"""You are the Supervisor Agent. Your job is to interpret the user’s request and delegate tasks to the available tools and agents to complete the task.
|
|
|
|
| 27 |
|
| 28 |
+
You have access to the following agents and tools:
|
| 29 |
+
- Tools: {[tool.name for tool in tools]}
|
| 30 |
+
- Agents: {[agent.name for agent in agents]}
|
| 31 |
|
| 32 |
+
1. Understand and break down user’s request into smaller sub-tasks.
|
| 33 |
+
2. Always delegate each sub-task to the best fitting agent/tool (e.g. always delegate any kind of coding or GitHub related task to ‘coder_agent’).
|
| 34 |
+
3. Coordinate tools/agents efficiently, ensuring minimal redundancy.
|
| 35 |
+
4. Validate results before presenting them.
|
| 36 |
+
5. Respond in Telegram Markdown.
|
| 37 |
+
- Use `*bold*`, `_italic_`, `[text](http://example.com)`, `` `inline code` ``, and ``` fenced code blocks ```.
|
| 38 |
+
- Never leave an opening `*`, `_`, `` ` ``, `[`, or ``` without its closing pair.
|
| 39 |
+
- Do not nest Markdown entities.
|
| 40 |
+
- Escape reserved characters (`* _ [ ] ( ) \ ``) if they are not part of a valid entity.
|
| 41 |
+
- Escaping inside entities is not allowed, so entity must be closed first and reopened again: use _snake_\__case_ for italic snake_case and *2*\**2=4* for bold 2*2=4.
|
| 42 |
|
| 43 |
+
Your top priority: satisfy the user’s query as effectively and efficiently as possible using the resources at your disposal, while formatting the final output in valid Markdown.
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 44 |
"""
|
| 45 |
|
| 46 |
def coder_system_message(state: dict):
|
database_files/main.db
CHANGED
|
Binary files a/database_files/main.db and b/database_files/main.db differ
|
|
|
mediafiles/167bb9e0-f228-4c14-b13e-6f042b3b610c.ogg
ADDED
|
Binary file (4.64 kB). View file
|
|
|
mediafiles/8d7ccab5-a3ae-4c32-b1d5-bc1f877eee40.ogg
ADDED
|
Binary file (9.02 kB). View file
|
|
|
requirements.txt
CHANGED
|
@@ -34,6 +34,7 @@ PyGithub>=2.7.0
|
|
| 34 |
langchain-openai>=0.3.28
|
| 35 |
langchain-google-genai
|
| 36 |
gtts>=2.5.4
|
|
|
|
| 37 |
timezonefinder>=7.0.1
|
| 38 |
geocoder>=1.38.1
|
| 39 |
aiosqlite>=0.21.0
|
|
|
|
| 34 |
langchain-openai>=0.3.28
|
| 35 |
langchain-google-genai
|
| 36 |
gtts>=2.5.4
|
| 37 |
+
jiter
|
| 38 |
timezonefinder>=7.0.1
|
| 39 |
geocoder>=1.38.1
|
| 40 |
aiosqlite>=0.21.0
|