I’m in the process of loading a JSON dataset of roughly 3 million documents into our PostgreSQL/vector store via JSONKnowledgeBase and PgVector. At this scale, the ingestion is taking much longer than is practical.
If you have any methods, best practices, or configuration tweaks that can significantly accelerate this bulk import—whether through parallelization, specialized bulk‐load paths, client settings, or other optimizations—I’d greatly appreciate any pointers or examples you can share.
Hey @Kenniferm thanks for reaching out and supporting Agno. I’ve shared this with the team, we’re working through all requests one by one and will get back to you soon.
If it’s urgent, please let us know. We appreciate your patience!
Frankly, it feels like you’re just running ChatGPT on autopilot and tossing out examples without ever verifying they work. It’s painfully obvious there is no async_read method in your JSONReader, so your “async ingestion” snippet is fundamentally broken. Could you point out a clear solution, please?