{"id":25898,"library":"llm-wiki-compiler","title":"llmwiki","description":"LLM Wiki Compiler is a CLI tool in active development (v0.5.1) that compiles raw sources (URLs, files) into an interlinked markdown wiki with claim-level provenance, semantic search, and MCP server integration for AI agents. It supports Anthropic, OpenAI, and Ollama providers, with configurable timeouts. Key differentiators: incremental compilation, Obsidian-compatible wikilinks, and provider-agnostic design.","status":"active","version":"0.5.1","language":"javascript","source_language":"en","source_url":"https://github.com/atomicmemory/llm-wiki-compiler","tags":["javascript","knowledge","wiki","llm","compiler","markdown","obsidian"],"install":[{"cmd":"npm install llm-wiki-compiler","lang":"bash","label":"npm"},{"cmd":"yarn add llm-wiki-compiler","lang":"bash","label":"yarn"},{"cmd":"pnpm add llm-wiki-compiler","lang":"bash","label":"pnpm"}],"dependencies":[{"reason":"Used for fetching YouTube transcripts as sources","package":"youtube-transcript","optional":true}],"imports":[{"note":"This is a CLI tool primarily used via npx or global install, not as a library. Import only works if package exports a programmatic API (currently not documented).","wrong":"const llmwiki = require('llm-wiki-compiler')","symbol":"default","correct":"import llmwiki from 'llm-wiki-compiler'"},{"note":"No programmatic API is officially exported for v0.5.x; use CLI commands instead.","wrong":"","symbol":"compile","correct":"import { compile } from 'llm-wiki-compiler'"},{"note":"No programmatic API is officially exported for v0.5.x; use CLI commands instead.","wrong":"","symbol":"ingest","correct":"import { ingest } from 'llm-wiki-compiler'"}],"quickstart":{"code":"npm install -g llm-wiki-compiler@0.5.1\nexport ANTHROPIC_API_KEY=sk-...\nllmwiki ingest https://example.com/article\nllmwiki compile\nllmwiki query \"What is the main point?\"","lang":"javascript","description":"Global install, set Anthropic API key, ingest a URL, compile sources into wiki, then ask a question."},"warnings":[{"fix":"Upgrade to v0.5.1: npm install -g llm-wiki-compiler@0.5.1","message":"v0.5.0 crashes on startup with 'ERR_PACKAGE_PATH_NOT_EXPORTED' due to deep import into youtube-transcript.","severity":"breaking","affected_versions":"0.5.0"},{"fix":"Set ANTHROPIC_API_KEY in shell or .env file.","message":"Anthropic provider requires either ANTHROPIC_API_KEY or ANTHROPIC_AUTH_TOKEN; order of precedence: shell env, .env, then Claude Code settings.","severity":"gotcha","affected_versions":">=0.1.0"},{"fix":"Ensure OPENAI_BASE_URL ends with /v1.","message":"OpenAI provider requires OPENAI_BASE_URL to include '/v1' (e.g., http://host:port/v1).","severity":"gotcha","affected_versions":">=0.1.0"}],"env_vars":null,"last_verified":"2026-05-01T00:00:00.000Z","next_check":"2026-07-30T00:00:00.000Z","problems":[{"fix":"npm install -g llm-wiki-compiler@0.5.1","cause":"v0.5.0 deep imports into youtube-transcript that does not expose that subpath.","error":"ERR_PACKAGE_PATH_NOT_EXPORTED"},{"fix":"Set LLMWIKI_PROVIDER and corresponding API key environment variable.","cause":"Missing LLMWIKI_PROVIDER or provider-specific API key.","error":"Error: No LLM provider configured"}],"ecosystem":"npm","meta_description":null,"install_score":null,"install_tag":null,"quickstart_score":null,"quickstart_tag":null}