Skip to content

VertexAiRagMemoryService - implementation may not be optimal / questions. #827

@ksachdeva

Description

@ksachdeva

Describe the bug
In vertex_ai_rag_memory_service, the entire session conversation (the text parts) are being combined
and uploaded as a file.

async def add_session_to_memory(self, session: Session):

Q1:
I would have thought that only the "user" prompt/input would be recoded as the memory and not the output/response of the AI.

Q2:
Assuming one vector embedding is being generated for the entire file, the quality of the feature vector would be low. I don't know the underlying behavior/implementation of VertexAiRag so it is possible that some chunking is happening. Even if it is the case you are storing event as json and chunking etc may not be optimal.

Q3:
Wouldn't it make sense to summarize the conversation or extract key elements and then store them as memory?

Regards & thanks
Kapil

Metadata

Metadata

Assignees

No one assigned

    Labels

    No labels
    No labels

    Type

    No type

    Projects

    No projects

    Milestone

    No milestone

    Relationships

    None yet

    Development

    No branches or pull requests

    Issue actions