KamiYomu is a powerful, extensible manga download built for manga enthusiasts who want full control over their collection. It scans and downloads manga from supported websites, stores them locally, and lets you host your own private manga reader—no ads, no subscriptions, no limits.
Join the conversation and be part of the KamiYomu community:
- 🔍 Automated Crawling — Fetch chapters from supported manga sites with ease
- 💾 Local Storage — Keep your manga files on your own server or device
- 🧩 Plugin Architecture — Add support for new sources or customize crawling logic
- 🛠️ Built with .NET 8 — Lightweight, maintainable, and easy to extend
Whether you're cataloging rare series, powering a personal manga dashboard, or seeking a cleaner alternative to bloated online readers, KamiYomu puts you in control. It's a lightweight, developer-friendly crawler built for clarity, extensibility, and respectful use of publicly accessible sources.
- Save the following
docker-compose.ymlfile:
services:
kamiyomu:
image: marcoscostadev/kamiyomu:latest
ports:
- "8080:8080"
restart: unless-stopped
healthcheck:
test: ["CMD", "curl", "-f", "http://localhost:8080/healthz"]
interval: 30s
timeout: 10s
retries: 3
volumes:
- kamiyomu_manga:/manga
- kamiyomu_database:/db
- kamiyomu_agents:/agents
- kamiyomu_logs:/logs
volumes:
kamiyomu_manga:
kamiyomu_database:
kamiyomu_agents:
kamiyomu_logs:- Run the following command in the directory containing
docker-compose.yml:
docker-compose up -d- Access the web interface at
http://localhost:8080
{: .note }
Map volumes to local paths as needed. Check releases for available versions. {: .note }
- .NET 8 Razor Pages
- Hangfire for job scheduling
- LiteDB for lightweight persistence
- HTMX + Bootstrap for dynamic UI
- Plugin-based architecture for extensibility
- Orchestrate and schedule crawler agent tasks (search, list, get, etc.)
- Download images from websites that the crawler agent specify
- Create local archives from downloaded images in specified folder location.
- Allow you to export archives as PDF, CBZ, or ZIP files to your local storage
- Provide a user-friendly web interface for managing your manga library.
- Invade sites or exploit vulnerabilities to access sensitive information
- Extract database, protected data information from websites
- Extract non-public information (such as website login credentials, API keys, etc)
- Store any files, data, or metadata on KamiYomu servers
- Track or monitor crawler activity across KamiYomu instances
- Enable peer-to-peer sharing or inter-instance communication
- Cache or retain downloaded content in KamiYomu servers
- Distribute, share, or redistribute downloaded content from KamiYomu instances to other users or platforms.
By using KamiYomu, you agree that you are solely responsible for all downloaded content, file management, and usage. KamiYomu is provided as-is for local use only. The developers and maintainers of KamiYomu assume no responsibility for your use of this tool, including but not limited to compliance with copyright laws, licensing agreements, or the terms of service of any source websites. You assume full liability for any legal consequences arising from your use of KamiYomu.
The KamiYomu project is licensed under the AGPL-3.0 (Affero General Public License version 3.0). This license allows users to freely use, modify, and distribute the software, provided that any modified versions are also distributed under the same license.
- Freedom to Use: Users can run the software for any purpose.
- Freedom to Study and Modify: Users can access the source code and modify it to suit their needs.
- Freedom to Distribute Copies: Users can share the original software with others.
- Freedom to Distribute Modified Versions: Users can distribute modified versions of the software, but they must also be licensed under AGPL-3.0, ensuring that the same freedoms are preserved for all users.
This license is particularly important for software that is intended to be run on servers, as it requires that the source code be made available to users who interact with the software over a network.
Pull requests are welcome! See the development guide to get started cloning the project and running it in Visual Studio or VS Code.
Create your own crawler agents by following the Crawler Agent development guide.
Questions, feedback, or bug reports? Open an issue or start a discussion.