aboutsummaryrefslogtreecommitdiffstats
path: root/docker/docker-compose.build.yml
diff options
context:
space:
mode:
authorMohamed Bassem <me@mbassem.com>2026-02-01 22:57:11 +0000
committerGitHub <noreply@github.com>2026-02-01 22:57:11 +0000
commit3fcccb858ee3ef22fe9ce479af4ce458ac9a0fe1 (patch)
tree0d6ae299126a581f0ccc58afa89b2dd16a9a0925 /docker/docker-compose.build.yml
parent54243b8cc5ccd76fe23821f6e159b954a2166578 (diff)
downloadkarakeep-3fcccb858ee3ef22fe9ce479af4ce458ac9a0fe1.tar.zst
feat: Add LLM-based OCR as alternative to Tesseract (#2442)
* feat(ocr): add LLM-based OCR support alongside Tesseract Add support for using configured LLM inference providers (OpenAI or Ollama) for OCR text extraction from images as an alternative to Tesseract. Changes: - Add OCR_USE_LLM environment variable flag (default: false) - Add buildOCRPrompt function for LLM-based text extraction - Add readImageTextWithLLM function in asset preprocessing worker - Update extractAndSaveImageText to route between Tesseract and LLM OCR - Update documentation with the new configuration option When OCR_USE_LLM is enabled, the system uses the configured inference model to extract text from images. If no inference provider is configured, it falls back to Tesseract. https://claude.ai/code/session_01Y7h7kDAmqXKXEWDmWbVkDs * format --------- Co-authored-by: Claude <noreply@anthropic.com>
Diffstat (limited to 'docker/docker-compose.build.yml')
0 files changed, 0 insertions, 0 deletions