Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

Add SelfConsciousReplies component for context aware message response ignore #647

Open
wants to merge 16 commits into
base: dev
Choose a base branch
from

Conversation

AbdurrehmanSubhani
Copy link
Contributor

This PR adds a SelfConsciousReplies to the DefaultComponents of an Agent.

The SelfConsciousReplies component is a PerceptionModifier which is aimed to provide the agent with the capability of choosing not reply to an incoming "say" perception based on the following parameters:

  • Conversation chat history context
  • Conversation Members
  • Prev 6 messages to check if the agent has been back and forth (helps to reduce agent's interest in the conversation)
  • Agent's personality (bio)
  • Set of guidelines for the Agent to consider provided the context above

Currently getting in response to the thinking process above the following schema:

  • shouldRespond: boolean
  • reason: string (this is for helping debugging cases in the future to see what the thought process was and how the prompt can be improved later on)
  • confidence: number (the overall confidence score the Agent thinks it has to send a reply to this message)

If the agent has been back and forth in the conversation a lot, a backAndForthPenalty is applied to the agent's reply reasoning to lower down its overall confidence for replying to this message

Copy link

vercel bot commented Nov 23, 2024

The latest updates on your projects. Learn more about Vercel for Git ↗︎

Name Status Preview Comments Updated (UTC)
chat ✅ Ready (Inspect) Visit Preview 💬 Add feedback Nov 25, 2024 1:22pm
docs ✅ Ready (Inspect) Visit Preview 💬 Add feedback Nov 25, 2024 1:22pm

…versation-heat-say-perception-reply-decision
…ased architecture, need to create another complete hueristic based appraoch with the intened behaviour
@avaer
Copy link
Contributor

avaer commented Nov 25, 2024

Additional thoughts/questions:

  • How much latency does gpt-4o-mini add to this pipeline?
  • Can we run an OODA loop in the main action pipeline to one-shot this action decision to eliminate the latency?
  • Can we make the OODA loop asynchronous from the action process to reduce latency?

@AbdurrehmanSubhani
Copy link
Contributor Author

Additional thoughts/questions:

  • How much latency does gpt-4o-mini add to this pipeline?
  • Can we run an OODA loop in the main action pipeline to one-shot this action decision to eliminate the latency?
  • Can we make the OODA loop asynchronous from the action process to reduce latency?
  • The average latency with gpt-4o-mini is ~3 seconds
  • That has been implemented in Null chat action support for Agent choosing not to reply in a conversation #699 removing the latency however maintaining the complete cost of the main action pipeline inference. This cost is cut-off as the perception is aborted before the action pipeline inference is triggered hence saving around ~26 upstreet credits for users
  • would be possible, the only downside being 2 inference costs for a chat message (~0.26 + ~26-30 upstreet credits)

Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
None yet
Projects
None yet
Development

Successfully merging this pull request may close these issues.

2 participants