This series measures how much local LLM response quality changes when Retrieval-Augmented Generation (RAG) is enabled.
Built a Gemma3-based local pipeline and compared output quality between RAG-enabled and non-RAG generations.
Automated comparison report of Gemma3 outputs with and without RAG for the same topic prompt.
Comparison focused on switching the query-rewrite phase to Gemma3 in the benchmark pipeline.
Benchmark focused on vectorization flow derived from Wikipedia dump data.
Benchmark article with a practical MCP server implementation scenario.
Comparison centered on implementation patterns for dynamic LLM model selection.
Benchmark article using Bun and Biome as the explanatory target domain.
Comparison focused on outputs for prompts about Wikipedia API usage patterns.