Cool stuff. Feature: enable history to be a trailing contenxt of n messages. #4
Replies: 10 comments 1 reply
-
One other idea is for long messages to get the ai to summarize and then save that as the history |
Beta Was this translation helpful? Give feedback.
-
History is already saved in a log file: https://github.com/drorm/gish/blob/main/README.md#log-file in JSON format, both the request and the response.
|
Beta Was this translation helpful? Give feedback.
-
Sorry for the confusion. That may be useful but I meant a way to support a trailing history context being passed with the requests to openai. the problem is when engaging in an interactive season there is no history or context being fead to the model. It would be interesting to have a way of defining this. For example a way of seating a system message and then also providing a trailing history. For example options
this would result in the request being made with
Here is an example illustrating the problem with the lack of context.
Hope this makes sense. I think the psudo code is close, needs to update the interactive request execution or somthing. The system feature would involve extra though. |
Beta Was this translation helpful? Give feedback.
-
Ah yes, It's on my roadmap. I was thinking of it as "chat mode". But pretty much the same idea. I'll probably generalize it so that it'll take arguments like to a printer where you can specify either a range or specific pages to print. Need to think a bit about what the right syntax to support that. |
Beta Was this translation helpful? Give feedback.
-
On further thought, it feels like the UI for this could get complicated quite quickly.
You edit it to your heart's content, and when you exit, gish submits it. Also, as far as summaries, my inclination is to summarize when needed rather than summarize everything so that we don't waste money on time unless it's needed. So for long content, there could be a summarizing step prior to the editing.. |
Beta Was this translation helpful? Give feedback.
-
json would be the most useful for a simple lib. then the end developer would be able to import in their lang of choice. In the editor mode that would be cool if it just happened a new blank message. Ie on file save it submits and adds one message and the empt space for a new one.. This is where message trimming could be useful. but it could keep extending the file and only submit the last n_responses or max_request_tokens. |
Beta Was this translation helpful? Give feedback.
-
Yeah, it is all an optimization problem at the end of the day. I think there is some threshold LEN such that it is cost-effective to summarize. I think the natural solution will be to get gpt4 access and just submit large texts to that to summarize into an ideal input for 3.5-turbo. |
Beta Was this translation helpful? Give feedback.
-
I migrated this conversation to a discussion, and created #5, the initial implementation that I'm looking at. |
Beta Was this translation helpful? Give feedback.
-
@syonfox this is now implemented in the dev branch. You're welcome to try it there or wait a bit, till I wrap up 0.2 and deploy it. |
Beta Was this translation helpful? Give feedback.
-
I just release V0.2
result in the last request showing up in the log as:
Because it carries the previous one which in turn carried the one before. |
Beta Was this translation helpful? Give feedback.
-
First of all nice project, I appreciate a clean cli template for js :)
I think a useful addition would be to allow history.
this adds to the requested cost but will improve performance. I think you can just pass in the array of messages. should be minimal tweak to API call
Beta Was this translation helpful? Give feedback.
All reactions