The (incorrect) interpretation I got reading JerryLiu's response was only the nodes that contribute to the response will be included in source_nodes, and since default chunk size is large, it could mean many (all in some cases) documents. However, from what I see, it's a combination of item: chunk size and top_k setting. Top_K will always return the number of nodes specified in the setting. Specifically, in my case, I had multiple (different) documents and I knew only one of the document contained a specific subject that isn't found in the other documents; yet the result in source_nodes was more affected by the top_k value so that didn't help.
So, my thought, is that this is about probability scoring and I'm looking at the scores in the NodeWithScore.
Currently, node[x].score all return "None" but perhaps because of a setting I'm using.
I'm looking here:
https://github.com/run-llama/llama_index/issues/14157Looking at the Fuzzy-citation that @WhiteFang_Jr shared, seems this is the post processing similarity scoring. Also read that certain types of query_engines such as TreeIndex would need post processing. I already did a documentSummaryIndex and will be using a retriever engine so hopefully, the score will already be in the NodesWithScoring