Neural Document Preprocessing // x402 Protocol Enabled
API.ONLINE
SOLANA-MAINNET
x402.v2
3
Intelligence Tiers
99%
Cost Reduction
<1ms
Response Time
$0.0001
Starting Price

System.Demo

📁
DROP FILE OR CLICK TO UPLOAD
[SUPPORTED: PDF, DOCX, TXT, RTF, HTML, MD]
PAYMENT REQUIRED
$0.001
1000 USDC micro-units
WALLET NOT CONNECTED
Network: SOLANA-MAINNET

System.Benefits

⚡ ZERO_SETUP
No libraries to install. No Docker containers. No infrastructure. Just hit the API and get chunks. Agents can discover and use us autonomously.
💰 MICROPAYMENTS
Pay only for what you use via x402 protocol. No subscriptions. No monthly fees. $0.0001 to start. Cash-flow positive from transaction #1.
🧠 CONTEXT_INJECTION
Our Tier 3 solves the "who is 'He'?" problem. We inject document context and entity metadata so your AI never loses the thread. No open-source tool does this.
⚡ INSTANT_RESPONSE
Millisecond latency. No 30-second LLM context loading. Get your chunks immediately and move on. Built for real-time agentic workflows.

System.Info

>> We built CHUNKER.CC for the Agentic Economy. In 2026, AI agents have wallets and make autonomous decisions. They don't "install Haystack" or "sign up for a SaaS plan." They find a service, pay $0.001 via x402, and get results instantly.

>> The problem we solve: Context Window Limits. When an agent hits a 500-page PDF but only has 8K tokens of memory, it needs smart chunking. Our Tier 3 contextual enrichment ensures the agent never loses track of "who," "what," or "when."

>> Built on Solana for near-zero transaction fees. Powered by x402 V2 for internet-native payments. Designed for autonomous agents, not humans clicking buttons.

Competitive.Edge

vs. GEMINI
99.75%
CHEAPER
$0.001 vs $2-4 per doc
vs. HAYSTACK
0 MIN
SETUP TIME
No 2-4hr infrastructure setup
vs. FIRECRAWL
CONTEXT
PRESERVATION
They scrape. We understand.
[TECHNICAL_ADVANTAGE] Our contextual injection solves the pronoun resolution problem. When your chunk says "He died in 1867," we prepend: [Referring to: Sir John A. Macdonald]. Open-source tokenizers can't do this. Gemini long-context is too expensive. We're the sweet spot.

Pricing.Matrix

TIER_1: DEMO
FREE
  • Recursive character splitting
  • Up to 100 pages/day
  • Basic paragraph detection
  • Quality: 6/10
TIER_2: STANDARD
$0.001/page
  • Smart sentence boundaries
  • Topic shift detection
  • Never breaks mid-sentence
  • Abbreviation-aware (Dr., U.S.)
  • Quality: 8.5/10 ⭐ RECOMMENDED
TIER_3: PROFESSIONAL
$0.008/page
  • All Standard features +
  • Document context injection
  • Entity extraction & classification
  • Cross-reference detection
  • Quality: 9.5/10

Integration.Protocol

# Demo tier (FREE - no payment required)
curl -X POST https://api.adepti.cc/chunk/demo \
  -F "file=@document.txt"
# Paid tiers (include Solana TX signature)
curl -X POST https://api.adepti.cc/chunk/standard \
  -H "X-PAYMENT: <solana-tx-signature>" \
  -F "file=@document.txt"
// Browser: Connect Phantom, pay USDC, upload
const { signature } = await phantom.signAndSendTransaction(tx);
fetch('/chunk/standard', { headers: { 'X-PAYMENT': signature }, body: formData });

FAQ.Database

What is document chunking and why do I need it? [+]
Document chunking splits large documents into smaller, semantically meaningful pieces for AI processing. LLMs have context window limits (8K-128K tokens), so a 500-page PDF needs to be chunked before it can be searched, embedded, or analyzed. Poor chunking = lost context = bad AI responses. Our smart chunking ensures each chunk is a complete, coherent unit of meaning.
What makes your chunking algorithm better? [+]
Our algorithm uses multiple techniques for superior results:

Smart Sentence Boundaries: We detect sentence endings accurately, handling abbreviations (Dr., U.S., Inc., etc.), initials (J.K. Rowling), and decimal numbers (3.14).

Topic Shift Detection: We analyze keyword density to detect when content shifts to a new topic, creating natural break points.

Never Breaks Mid-Sentence: Unlike character-based splitters, we always end chunks at complete sentences.

Sentence Overlap: Each chunk includes the last sentence of the previous chunk for context continuity.
How does payment work? [+]
We accept USDC on Solana with two payment methods:

1. Manual (Web UI): Connect your Phantom wallet on this page, select a paid tier, and approve the USDC transfer. The transaction signature is automatically sent with your file.

2. Programmatic (x402 API): For AI agents and developers - call /estimate to get pricing, execute a USDC transfer on Solana, then include the TX signature in the X-PAYMENT header when calling the chunking endpoint.

Payment is per-page based on document size (~500 chars = 1 page). Demo tier is free (100 pages/day limit).
What's the difference between the tiers? [+]
Demo (Free): Basic recursive character splitting with paragraph detection. Good for testing and simple documents. Quality: 6/10.

Standard ($0.001/page): Smart sentence boundary detection, topic shift analysis, abbreviation handling (Dr., U.S., Inc., etc.), and sentence-level overlap. Never breaks mid-sentence. Best value for most RAG use cases. Quality: 8.5/10.

Professional ($0.008/page): All Standard features plus document context injection, entity extraction and classification, cross-reference detection. When a chunk mentions "He," we prepend who "He" refers to. Best for legal, medical, and complex documents. Quality: 9.5/10.
What file formats are supported? [+]
We support: PDF, DOCX, TXT, RTF, HTML, and Markdown.

PDF: Text-based PDFs (not scanned images). We use PyMuPDF for accurate extraction.
DOCX: Full extraction including headers, footers, text boxes, and structured content.
TXT/RTF/HTML/MD: Direct text processing with format-specific parsing.

Pages are calculated at ~500 characters per page.
Can AI agents use this API autonomously? [+]
Yes! That's exactly what we're built for. The workflow is:

1. Call POST /estimate with the file to get page count and pricing
2. Execute a USDC transfer on Solana to our payment address
3. Call the chunking endpoint with the TX signature in X-PAYMENT header

Fully autonomous, no human intervention needed. Demo tier requires no payment for testing.
What are the chunk size parameters? [+]
Our chunking is optimized for RAG (Retrieval Augmented Generation):

Target Size: ~1200 characters (ideal for embedding models)
Minimum Size: 400 characters (avoids tiny fragments)
Maximum Size: 2000 characters (prevents oversized chunks)
Overlap: 1 sentence (maintains context between chunks)

These parameters are tuned for optimal performance with OpenAI, Cohere, and other embedding APIs.
Is my data stored or logged? [+]
No. Documents are processed in memory and immediately discarded. We don't store your files, chunks, or content. Only basic request logs (IP, timestamp, file size) are kept for rate limiting and abuse prevention. Your documents never touch disk storage.