I’m working in Triple Minds and I’m working on a Candy AI Clone project like this. The clone is coming together nicely, but I’m facing a weird issue with real-time responses. Sometimes the AI takes too long to reply, making the chat feel laggy and unnatural.
Here’s what I’ve tried so far:
Has anyone encountered this while building Candy AI or other chatbot clones? I’d love to hear how you managed fast, smooth AI responses without losing context or conversation quality.
Here’s what I’ve tried so far:
- Optimizing API calls to OpenAI GPT
- Reducing context size for faster processing
- Implementing async handling on the frontend
- Still, occasional delays of 5–10 seconds persist
Has anyone encountered this while building Candy AI or other chatbot clones? I’d love to hear how you managed fast, smooth AI responses without losing context or conversation quality.