For AI agents, what's the bigger problem: context sharing or prompting?

1 exclusivewombat 0 5/20/2025, 3:02:51 PM
Been building with LLM-based agents lately and keep running into a few recurring challenges:

1/ Prompting - making sure agents behave how you want without overly long, fragile instructions

2/ Context sharing – passing memory, results, and state across time or between agents w/o flooding the system

3/ Cost – tokens get expensive fast, especially as things scale

Curious what others think is the real bottleneck here, and any tips/tricks for solving this. Are you optimizing around token limits, memory persistence, better prompt engineering?

Would love to hear how you’re thinking about this or if there’s a smarter approach we’re all missing. ty in advance!

Comments (0)

No comments yet