Close Menu
    Facebook LinkedIn YouTube WhatsApp X (Twitter) Pinterest
    Trending
    • The 11 Best Fans to Buy Before It Gets Hot Again (2026)
    • A look at Dylan Patel’s SemiAnalysis, an AI newsletter and research firm that expects $100M+ in 2026 revenue from subscriptions and AI supply chain research (Abram Brown/The Information)
    • ‘Euphoria’ Season 3 Release Schedule: When Does Episode 2 Come Out?
    • Francis Bacon and the Scientific Method
    • Proxy-Pointer RAG: Structure Meets Scale at 100% Accuracy with Smarter Retrieval
    • Sulfur lava exoplanet L 98-59 d defies classification
    • Hisense U7SG TV Review (2026): Better Design, Great Value
    • Google is in talks with Marvell Technology to develop a memory processing unit that works alongside TPUs, and a new TPU for running AI models (Qianer Liu/The Information)
    Facebook LinkedIn WhatsApp
    Times FeaturedTimes Featured
    Sunday, April 19
    • Home
    • Founders
    • Startups
    • Technology
    • Profiles
    • Entrepreneurs
    • Leaders
    • Students
    • VC Funds
    • More
      • AI
      • Robotics
      • Industries
      • Global
    Times FeaturedTimes Featured
    Home»Artificial Intelligence»Proxy-Pointer RAG: Structure Meets Scale at 100% Accuracy with Smarter Retrieval
    Artificial Intelligence

    Proxy-Pointer RAG: Structure Meets Scale at 100% Accuracy with Smarter Retrieval

    Editor Times FeaturedBy Editor Times FeaturedApril 19, 2026No Comments16 Mins Read
    Facebook Twitter Pinterest Telegram LinkedIn Tumblr WhatsApp Email
    Share
    Facebook Twitter LinkedIn Pinterest Telegram Email WhatsApp Copy Link


    In my previous article, I launched Proxy-Pointer RAG — a retrieval doc construction straight right into a vector index, reaching the surgical precision of “Vectorless RAG” methods like PageIndex, with out their scalability and price penalties. That article laid the muse: the why, the how, and a promising 10-query comparability on a single World Financial institution report.

    Though a helpful proof-of-concept, that didn’t show manufacturing readiness. This text goals to handle that.

    In an enterprise, the overwhelming majority of paperwork on which RAG is utilized — technical manuals, analysis papers, authorized contracts, coverage studies, annual filings, compliance paperwork — have part headings. That makes them structured. And that construction encapsulates that means, exactly how a human comprehends and searches a posh doc by organizing the part circulation within the thoughts. A normal vector RAG throws away the construction, when it shreds a doc right into a flat bag of chunks resulting in sub-optimal responses. Proxy-Pointer as an alternative exploits that ubiquitous construction to dramatically enhance retrieval accuracy, at minimal further value.

    To emphasize-test the structure, we would have liked probably the most demanding structured paperwork we might discover — the sort the place a single misplaced decimal level or a missed footnote can invalidate a whole evaluation. That’s monetary filings. 10-Ok annual studies are deeply nested, cross-referenced throughout a number of monetary statements, and demand exact numerical reasoning. If Proxy-Pointer can deal with these, it could deal with something with headings.

    This text gives proof to help Proxy-Pointer’s functionality. I took 4 publicly available FY2022 10-Ok filings — AMD (121 pages), American Specific (260 pages), Boeing (190 pages), and PepsiCo (500 pages) — and examined Proxy-Pointer on 66 questions throughout two distinct benchmarks, together with adversarial queries particularly designed to interrupt naive retrieval methods. The outcomes have been decisive and summarized under.

    Additionally on this article, I’m open-sourcing the entire pipeline — so you may run it by yourself paperwork, reproduce the outcomes, and push it additional.

    Fast Recap: What’s Proxy-Pointer?

    Customary vector RAG splits paperwork into blind chunks, embeds them, and retrieves the top-Ok by cosine similarity. The synthesizer LLM sees fragmented, context-less textual content — and continuously hallucinates or misses the reply solely.

    Within the earlier article, Proxy-Pointer mounted this with 5 zero-cost engineering strategies:

    1. Skeleton Tree — Parse Markdown headings right into a hierarchical tree (pure Python, no LLM wanted)
    2. Breadcrumb Injection — Prepend the total structural path (AMD > Monetary Statements > Money Flows) to each chunk earlier than embedding
    3. Construction-Guided Chunking — Break up textual content inside part boundaries, by no means throughout them
    4. Noise Filtering — Take away distracting sections (TOC, glossary, govt summaries) from the index
    5. Pointer-Primarily based Context — Use retrieved chunks as pointers to load the full, unbroken doc part for the synthesizer

    The outcome: each chunk is aware of the place it lives within the doc, and the synthesizer sees full sections — not fragments.

    Refinements For the reason that First Article

    A number of important enhancements have been made to the pipeline earlier than benchmarking. These are outlined as follows:

    Indexing Pipeline Adjustments

    Standalone Structure. The unique implementation relied on PageIndex as a dependency for skeleton tree era. This has been fully eliminated. Proxy-Pointer now ships a self-contained, ~150-line pure-Python tree builder that parses Markdown headings right into a hierarchical JSON construction — zero exterior dependencies, zero LLM calls, runs in milliseconds.

    LLM-Powered Noise Filter. The primary model used a hardcoded listing of noise titles (NOISE_TITLES = {"contents", "foreword", ...}). This broke on variations like “Be aware of Thanks” vs. “Acknowledgments” or “TOC” vs. “TABLE OF CONTENTS.” On this model, the brand new pipeline sends the light-weight skeleton tree to gemini-flash-lite and asks it to establish noise nodes throughout six classes. This catches semantic equivalents that regex couldn’t.

    The up to date pipeline is the next:

    Retrieval Pipeline Adjustments

    Two-Stage Retrieval: Semantic + LLM Re-Ranker. The primary article used a easy top-Ok retrieval from FAISS. The refined pipeline now operates in two phases:

    • Stage 1 (Broad Recall): FAISS returns the highest 200 chunks by embedding similarity, that are deduplicated by (doc_id,node_id) and shortlisted to 50 distinctive candidate nodes.
    • Stage 2 (Structural Re-Rating): The hierarchical breadcrumb paths of all 50 candidates are despatched to a Gemini LLM, which re-ranks them by structural relevance — not embedding similarity — and returns the highest 5. That is the important thing differentiator: a question about “AMD’s money circulation” now appropriately prioritizes AMD > Monetary Statements > Money Flows over a paragraph that merely mentions money circulation.

    Right here is the up to date pipeline:

    These refinements remodeled Proxy-Pointer from a promising prototype right into a production-grade retrieval engine.

    Benchmarking: Two Exams, 66 Questions, 4 Corporations

    To carefully consider the pipeline, I downloaded the FY2022 10-Ok annual filings of AMD, American Specific (AMEX), Boeing, and PepsiCo. These have been extracted to Markdown utilizing LlamaParse and listed by the Proxy-Pointer pipeline.

    I then examined in opposition to two distinct benchmarks:

    Benchmark 1: FinanceBench (26 Questions)

    FinanceBench is a longtime benchmark of qualitative and quantitative questions throughout monetary filings. I chosen all 26 questions spanning the 4 firms within the dataset — overlaying numerical reasoning, data extraction and logical reasoning. Resulting from dataset licensing restrictions, the FinanceBench questions and floor reality solutions are usually not included within the github repository. Nevertheless, the scorecards with Proxy-Pointer responses are supplied for reference.

    To make sure reproducibility, the repository features a benchmark.py script that permits you to run the total analysis your self on FinanceBench (or any customized dataset), producing each detailed logs and scorecards utilizing the identical pipeline.

    Benchmark 2: Complete Stress Check (40 Questions)

    FinanceBench, whereas helpful, primarily exams factual recall. I wanted one thing tougher — queries that might break a system counting on surface-level chunk matching. So I created 40 customized questions, 10 per firm, particularly designed to stress-test numerical reasoning, multi-hop retrieval, adversarial robustness, and cross-statement reconciliation.

    The total Q&A logs utilizing the bot and scorecards evaluating with the bottom reality are included within the github repository.

    Listed below are 5 examples that illustrate the complexity:

    Multi-hop Numerical (AMEX): “Calculate the proportion of web curiosity earnings to whole revenues web of curiosity expense for 2022 and examine it to 2021. Did dependence improve?”
    This requires finding two completely different line objects throughout two fiscal years, computing ratios for every, after which evaluating them — a three-step reasoning chain that calls for exact retrieval of the earnings assertion.

    Adversarial Numerical (AMD): “Estimate whether or not stock buildup contributed considerably to money circulation decline in FY2022.”
    That is intentionally adversarial: it presupposes that money circulation declined (which it did, marginally) and requires the mannequin to quantify a stability sheet merchandise’s affect on a money circulation assertion metric. A naive retriever would fetch stability sheet knowledge however miss the money circulation context.

    Reinvestment Price (PepsiCo): “Calculate the reinvestment charge outlined as Capex divided by (Working Money Stream minus Dividends).”
    This requires pulling three distinct figures from the money circulation assertion, performing a non-standard calculation that isn’t reported wherever within the 10-Ok, and arriving at a exact ratio (1.123).

    Money Stream High quality (Boeing): “What proportion of working money circulation in FY2022 was consumed by modifications in working capital?”
    The reply right here is counterintuitive: 0% — as a result of working capital was truly a supply of money (+$4,139M vs. OCF of $3,512M), contributing 118% of working money circulation. Any system that retrieves the improper part or misinterprets the signal will fail.

    Attribution (AMEX): “Estimate how a lot of whole income progress is attributable to low cost income improve.”
    This requires computing two deltas (low cost income change and whole income change), then expressing one as a proportion of the opposite — a calculation nowhere current within the submitting itself.

    Each query has a pre-computed floor reality reply with particular numerical values, making analysis unambiguous.

    Outcomes

    ok=5 Configuration (Main)

    On this configuration, the retriever selects a set of 5 nodes (k_final = 5), and the corresponding sections despatched to the synthesizer LLM for response.

    Benchmark Rating Accuracy
    FinanceBench (26 questions) 26 / 26 100%
    Complete (40 questions) 40 / 40 100%
    Whole 66 / 66 100%

    An ideal rating throughout all 66 questions. Each numerical worth matched the bottom reality. Each qualitative evaluation aligned with the submitting knowledge.

    As an example, listed here are two precise bot responses from the benchmark run — exhibiting the retrieval path and the synthesized reply for queries cited within the earlier part:

    PepsiCo: Reinvestment Price — “Calculate the reinvestment charge outlined as Capex divided by (Working Money Stream minus Dividends).”

    The bot retrieved Consolidated Assertion of Money Flows (ID: 0080) and Consolidated Assertion of Money Flows (continued) (ID: 0081), then computed:

    Capital spending: $5,207M / (Working Money Stream: $10,811M − Dividends: $6,172M) = 5,207 / 4,639 = 112.24% (i.e., 1.1224)

    Floor reality: 1.123. ✅ Match. However the bot went additional — unprompted, it computed the identical ratio for FY2021 (79.73%) and FY2020 (83.07%), surfacing a transparent development of accelerating reinvestment that wasn’t even requested for.

    Boeing: Money Stream High quality — “What proportion of working money circulation in FY2022 was consumed by modifications in working capital?”

    The bot retrieved Boeing > Liquidity and Capital Sources > Money Stream Abstract (ID: 0082) and Consolidated Statements of Money Flows (ID: 0107), then responded:

    Adjustments in working capital supplied $4,139M in money, whereas web working money circulation was $3,512M. As a result of the change in working capital was a supply of money relatively than a consumption, it didn’t eat any proportion of OCF.

    Floor reality: 0% consumed; working capital was a supply of money. ✅ Match. This can be a intentionally counterintuitive query — most methods would pressure a proportion relatively than recognizing the logical inversion. The structural retrieval ensured the bot had the total money circulation assertion, and the synthesizer appropriately interpreted the signal.

    ok=3 Configuration (Stress Check)

    To know the system’s failure boundaries, I re-ran each benchmarks with k_final=3 — retrieving solely 3 doc sections as an alternative of 5. This intentionally constrains the context window to check whether or not the retrieval precision is powerful sufficient to work with fewer nodes.

    Benchmark Rating Accuracy
    FinanceBench (26 questions) 25 / 26 96.2%
    Complete (40 questions) 37 / 40 92.5%
    Whole 62 / 66 93.9%

    The failures within the ok=3 run have been insightful and supply extra perspective than the proper ok=5 scores:

    • FinanceBench: One query for AMD, suffered a calculation hallucination — the mannequin retrieved the proper inputs ($9,981M / $6,369M) however computed the division incorrectly (outputting 1.78 as an alternative of 1.57). The retrieval was appropriate; the LLM’s arithmetic was not. This isn’t a retrieval failure — it’s a synthesis failure, and a bigger LLM (than flash-lite used right here) doubtless wouldn’t have made an error.
    • Complete (AMEX Q7): The question “Did provisions for credit score losses improve quicker than whole income?” requires each the provisions line and the whole income line from the earnings assertion. With ok=3, the ranker prioritized the credit score loss notes over the income abstract, leaving the synthesizer with out the denominator for its comparability.
    • Complete (AMEX Q10): Working leverage evaluation requires evaluating income progress and expense progress aspect by aspect. At ok=3, the expense breakdowns have been excluded.
    • Complete (PepsiCo Q9): “What proportion of Russia-Ukraine prices have been as a result of intangible asset impairments?” requires a particular footnote that, at ok=3, was displaced by higher-ranking money circulation nodes.

    The sample is constant: each ok=3 failure was attributable to inadequate context protection, not incorrect retrieval. The ranker selected the proper main sections; it merely didn’t have room for the secondary ones that complicated reconciliation queries demand.

    This confirms an vital architectural perception: when questions require cross-referencing a number of components of monetary statements, ok=5 gives the required protection, whereas ok=3 introduces retrieval gaps for probably the most complicated reconciliations. For many sensible purposes — the place the vast majority of queries goal a single part or assertion — ok=3 could be completely sufficient and quicker.

    What the Scorecards Don’t Present

    Past the uncooked numbers, the benchmark revealed qualitative strengths value highlighting:

    Supply Grounding. Each response cited particular sources utilizing their structural breadcrumbs (e.g., AMD > Monetary Situation > Liquidity and Capital Sources). An analyst receiving these solutions can hint them on to the submitting part, creating an audit path.

    Adversarial Robustness. When requested about crypto income at AMEX (which doesn’t exist), the system appropriately returned “No proof” relatively than hallucinating a determine. When requested about Boeing’s Debt/Fairness ratio (which is mathematically undefined as a result of detrimental fairness), it defined why the metric is just not significant relatively than forcing a quantity. These are the queries that journey up methods with poor retrieval — they floor plausible-looking however irrelevant context, and the LLM invents a solution.

    Outperforming Floor Reality. In a number of instances, the bot’s reply was arguably higher than our pre-computed floor reality. Boeing’s backlog change was estimated as “mid-single digit %” within the floor reality, however the bot computed the exact determine: +7.12%. AMD’s stock affect was ground-truthed as “$1B+ drag,” however the bot recognized the precise $1.4B buildup. These aren’t errors — they’re enhancements made doable as a result of the synthesizer noticed the total, unedited part textual content, not a truncated chunk.

    For detailed question-by-question outcomes, comparability logs, and traffic-light scorecards, consult with the open supply repo talked about under.

    Open-Supply Repository

    Proxy-Pointer is now absolutely open-source (MIT License) and might be accessed at Proxy-Pointer Github repository.

    It’s designed for a 5-minute quickstart:

    Proxy-Pointer/
    ├── src/
    │   ├── config.py                      # Centralized configuration
    │   ├── extraction/                    # PDF → Markdown (LlamaParse)
    │   ├── indexing/
    │   │   ├── build_skeleton_trees.py    # Pure-Python tree builder
    │   │   └── build_pp_index.py          # Noise filter + chunking + FAISS
    │   └── agent/
    │       ├── pp_rag_bot.py              # Interactive RAG bot
    │       └── benchmark.py               # Automated benchmarking with LLM-as-a-judge
    ├── knowledge/
    │   ├── pdf/                           # 4 FinanceBench 10-Ks included
    │   ├── paperwork/                     # Pre-extracted AMD Markdown (able to index)
    │   └── Benchmark/                     # Full scorecards and comparability logs

    What’s included out of the field

    • A pre-extracted Markdown file for AMD’s FY2022 10-Ok — simply configure your Gemini API key, construct the index, and begin querying in beneath 5 minutes.
    • Three further 10-Ok PDFs (AMEX, Boeing, PepsiCo) for customers who wish to increase their corpus.
    • Scorecards for each benchmarks and full question-answer logs for 40 questions of Complete benchmark.
    • An automatic benchmarking script with LLM-as-a-judge analysis — convey your individual Excel file with questions and floor truths, and the system generates timestamped scorecards robotically.

    The complete pipeline runs on a single Gemini API key utilizing gemini-flash-lite, probably the most value efficient mannequin in Google’s lineup. No GPU required. No complicated infrastructure or token-hungry and costly indexing tree to be constructed.

    Simply clone, configure, vector index, question.

    Conclusion

    Once I revealed the primary article on Proxy-Pointer, it was an attention-grabbing speculation: you don’t want costly LLM-navigated timber to make retrieval structurally conscious — you simply have to be intelligent about what you embed. The proof was a 10-query comparability on a single doc.

    This text strikes past speculation to proof.

    66 questions. 4 Fortune 500 firms. 100% accuracy at ok=5. Multi-hop numerical reasoning, cross-statement reconciliation, adversarial edge instances, and counterintuitive monetary metrics — Proxy-Pointer dealt with all of them. And after we intentionally starved it of context at ok=3, it nonetheless delivered 93.9% accuracy, failing solely when complicated queries genuinely required greater than three doc sections.

    Here’s what we achieved — and why it issues for manufacturing RAG methods:

    • One structure, all doc varieties. Enterprises now not want to keep up two separate retrieval pathways — Vectorless for complicated, structured, high-value paperwork (monetary filings, authorized contracts, analysis studies) and the confirmed Vector RAG for routine data bases. Proxy-Pointer handles each inside a single, unified vector RAG pipeline. If a doc has structural headings, the system exploits them. If it doesn’t, it degrades gracefully to plain chunking. No routing logic, no particular instances.
    • Full vector RAG scalability on the identical worth level. Tree-based “Vectorless” approaches require costly LLM calls throughout indexing (one abstract per node for all nodes in a doc) and throughout retrieval (LLM tree navigation per question). Proxy-Pointer eliminates each. The skeleton tree is a particularly light-weight construction constructed with pure Python in milliseconds. Indexing makes use of solely an embedding mannequin — an identical to plain vector RAG. The one LLM calls are the noise filter (as soon as per doc at indexing) and the re-ranker, which takes solely 50 (and may even work effectively with half that quantity) breadcrumbs (tree paths) + synthesizer at question time.
    • Funds-friendly fashions, premium outcomes. The complete benchmark was run utilizing gemini-embedding-001 at half its default dimensionality (1536 as an alternative of 3072) to optimise vector database storage and gemini-flash-lite — probably the most cost-efficient mannequin in Google’s lineup — for noise filtering, re-ranking, and synthesis. No GPT-5, no Claude Opus, no fine-tuned fashions. The structure compensates for mannequin simplicity by delivering higher context to the synthesizer.
    • Clear, auditable, explainable. Each response comes with a structural hint: which nodes have been retrieved, their hierarchical breadcrumbs, and the precise line ranges within the supply doc. An analyst can confirm any reply by opening the Markdown file and studying the cited part. This isn’t a black-box system — it’s a glass-box retrieval engine.
    • Open-source and instantly usable. The entire pipeline — from PDF extraction to benchmarking — is obtainable as a single repository with a 5-minute quickstart. Clone, configure a LLM API key, construct the index and begin querying. No GPU, no Docker, no complicated infrastructure.

    In case your retrieval system is scuffling with complicated, structured paperwork, the issue might be not your embedding mannequin. It’s that your index has no concept the place something lives within the doc. Give it construction, and the accuracy follows.

    Clone the repo. Strive your individual paperwork. Let me know your ideas.

    Join with me and share your feedback at www.linkedin.com/in/partha-sarkar-lets-talk-AI

    All paperwork used on this benchmark are publicly accessible FY2022 10-Ok filings at SEC.gov. Code and benchmark outcomes are open-source beneath the MIT License. Photos used on this article are generated utilizing Google Gemini.



    Source link

    Share. Facebook Twitter Pinterest LinkedIn Tumblr Email
    Editor Times Featured
    • Website

    Related Posts

    Dreaming in Cubes | Towards Data Science

    April 19, 2026

    AI Agents Need Their Own Desk, and Git Worktrees Give Them One

    April 18, 2026

    Your RAG System Retrieves the Right Data — But Still Produces Wrong Answers. Here’s Why (and How to Fix It).

    April 18, 2026

    Europe Warns of a Next-Gen Cyber Threat

    April 18, 2026

    How to Learn Python for Data Science Fast in 2026 (Without Wasting Time)

    April 18, 2026

    A Practical Guide to Memory for Autonomous LLM Agents

    April 17, 2026
    Leave A Reply Cancel Reply

    Editors Picks

    The 11 Best Fans to Buy Before It Gets Hot Again (2026)

    April 19, 2026

    A look at Dylan Patel’s SemiAnalysis, an AI newsletter and research firm that expects $100M+ in 2026 revenue from subscriptions and AI supply chain research (Abram Brown/The Information)

    April 19, 2026

    ‘Euphoria’ Season 3 Release Schedule: When Does Episode 2 Come Out?

    April 19, 2026

    Francis Bacon and the Scientific Method

    April 19, 2026
    Categories
    • Founders
    • Startups
    • Technology
    • Profiles
    • Entrepreneurs
    • Leaders
    • Students
    • VC Funds
    About Us
    About Us

    Welcome to Times Featured, an AI-driven entrepreneurship growth engine that is transforming the future of work, bridging the digital divide and encouraging younger community inclusion in the 4th Industrial Revolution, and nurturing new market leaders.

    Empowering the growth of profiles, leaders, entrepreneurs businesses, and startups on international landscape.

    Asia-Middle East-Europe-North America-Australia-Africa

    Facebook LinkedIn WhatsApp
    Featured Picks

    Taylor Swift Launches a New Instagram Feature Inviting Fans to Remix Others’ Stories

    October 7, 2025

    Key cryptocurrency terms and what they mean

    July 14, 2025

    Epirus Leonidas microwave weapon downs 49 drones at once

    September 28, 2025
    Categories
    • Founders
    • Startups
    • Technology
    • Profiles
    • Entrepreneurs
    • Leaders
    • Students
    • VC Funds
    Copyright © 2024 Timesfeatured.com IP Limited. All Rights.
    • Privacy Policy
    • Disclaimer
    • Terms and Conditions
    • About us
    • Contact us

    Type above and press Enter to search. Press Esc to cancel.