So I have the text-generation-ui by oogabooga running at one place, then I also have stable diffusion in the other tab. But I’m looking for ways to expose these project’s APIs, and then combine them to then produce output like what GPT-4 does, where it can call APIs when it needs to, to other models.
I’m also looking for a solution where the text generation output is also able to execute the said code, and then infer from its results to do next things. (iknow the risks but yeah).
Two off the top of my head: https://heyamica.com/ and silly tavern for fun stuff.
For agents there are https://github.com/spyglass-search/talos or https://github.com/Josh-XT/AGiXT
I think the problems is work + play aren’t really the same goals.
Really just a UX problem. Work is a subset of play. Play is how we simulate and practice anything and everything
Thanks for your answer! I get it. These projects do give me some ideas. I didn’t know such things are called ‘agents’ in this space