Hacker Newsnew | past | comments | ask | show | jobs | submitlogin

I've been manually copying responses between chats when I hit token limits, and I'm wondering - have you considered a multi-AI consensus approach instead of persistent memory?

The idea: multiple AIs (Claude, GPT, Gemini, Grok) brainstorm simultaneously and produce one agreed response. This might solve the context problem more elegantly because:

- No token limit anxiety - you get comprehensive answers upfront - Better quality through AI cross-validation - The consensus answer naturally becomes your context - Simpler to implement - just parallel API calls vs memory tree management

Just curious if you've explored this direction or if there's a reason the memory persistence approach works better for your use case?



Guidelines | FAQ | Lists | API | Security | Legal | Apply to YC | Contact

Search: