Rapid AI-Powered Text and Object Recognition
Open-Source Content Insight Engine for Enterprise AI Applications
About β’ Features β’ Installation β’ Quick Start β’ Documentation β’ Contributing β’ Website
RAPTOR is an advanced AI framework developed by the DHT Taiwan Team at DHT Solutions.
Aigle 0.1 - First Community Beta Release
This is the initial open-source release of the RAPTOR framework, codenamed "Aigle". We're excited to share this with the community and look forward to your feedback and contributions.
To help developers get started with the RAPTOR framework quickly and easily, we've deployed a test run API on DHT's development infrastructure. This evaluation API allows developers to:
- Test and evaluate RAPTOR capabilities without setting up infrastructure
- Develop AI applications using the RAPTOR framework with zero deployment overhead
- Utilize DHT resources for testing and development purposes
- Prototype faster by accessing pre-configured AI services
This is an excellent way to explore RAPTOR's features, build proof-of-concepts, and validate your use cases before deploying your own infrastructure.
π Access the Evaluation API:
http://raptor_open_0_1_api.dhtsolution.com:8012/
For detailed API documentation, usage examples, and access instructions, please visit the link above.
Note: This is a development environment intended for evaluation and testing purposes. For production deployments, please refer to the Installation and Development sections below.
- About RAPTOR
- Features
- Installation
- Quick Start
- Documentation
- Community & Support
- Contributing
- License
- Acknowledgments
RAPTOR (Rapid AI-Powered Text and Object Recognition) is a Content Insight Engine that represents a paradigm shift in digital asset management, transforming passive media storage into an intelligent knowledge platform. By leveraging cutting-edge AI technologies including large language models, vector search, and semantic understanding, RAPTOR enables organizations to unlock the full value of their media assets through automated analysis, intelligent search, and actionable insights.
- 85% reduction in manual content tagging and metadata generation
- 10x faster content discovery through semantic search
- 60% improvement in content reuse and operational efficiency
- Real-time insights from video, audio, and document content
- Enterprise-grade security, scalability, and integration capabilities
- AI-Native Architecture: Built from the ground up around LLM orchestration and vector search
- Multi-Modal Understanding: Unified analysis across video, audio, image, and text
- Semantic Intelligence: Context-aware search that understands intent, not just keywords
- Open + Enterprise Model: Open-source core with premium enterprise features
- Production-Ready: Kubernetes-native with auto-scaling, fault tolerance, and 99.9% uptime
This first community release includes:
- Multi-Modal Content Analysis: Process and understand video, audio, images, and text
- Semantic Search Engine: Context-aware search using vector embeddings
- AI-Powered Metadata Generation: Automated tagging and content classification
- LLM Orchestration: Flexible integration with multiple language models
- Vector Database Integration: High-performance similarity search and retrieval
- Content Understanding: Extract insights from unstructured media
- Entity Recognition: Identify people, places, objects, and concepts
- Sentiment Analysis: Understand emotional context in content
- Topic Modeling: Automatic categorization and clustering
- Temporal Analysis: Track content evolution over time
- Scalable Architecture: Kubernetes-native deployment
- API-First Design: RESTful APIs for seamless integration
- Security: Enterprise-grade authentication and authorization
- Monitoring: Built-in observability and logging
- Extensible: Plugin architecture for custom processors
For detailed release notes, see CHANGELOG.md.
# Clone the repository
git clone https://github.com/DHT-AI-Studio/RAPTOR.git
cd RAPTOR/Aigle/0.1
# Create virtual environment (recommended)
conda create -n CIE python=3.10
conda activate CIE
# Install required dependencies:
pip install -r requirements.txtcd raptor
chmod +x check-services.sh deploy.sh logs.sh rollback.sh stop-all.sh tag-backup.sh
./deploy.sh-
Check container status
./check-services.sh
-
Test API connectivity
# Modellifecycle ζε curl -s http://192.168.157.165:8086/docs # Assetmanagement curl -s http://192.168.157.165:8010/docs
-
View service logs
./logs.sh <service_name>
-
Create a new user and assign a new branch to the user
curl -X 'POST' \ 'http://192.168.157.165:8086/users' \ -H 'accept: application/json' \ -H 'Content-Type: application/json' \ -d '{ "username": "user1", "password": "dht888888", "password_hash": "", "branch": "", "permissions": [ "upload", "download", "list" ] }'
-
Create a new access token for the user
curl -X 'POST' \ 'http://192.168.157.165:8086/token' \ -H 'accept: application/json' \ -H 'Content-Type: application/x-www-form-urlencoded' \ -d 'grant_type=password&username=user1&password=dht888888&scope=&client_id=string&client_secret=********'
-
Access RedisInsight π http://192.168.157.165:5540
Add a new connection:
- Connection Type: Redis Cluster
- Host:
redis1 - Port:
7000 - Name:
Redis Cluster - Authentication:
dht888888
-
Check if the local Ollama model includes qwen2.5:7b
ollama list
If not present
ollama pull qwen2.5:7b
-
Register MLflow with local Ollama
curl -X 'POST' \ 'http://192.168.157.165:8010/models/register_ollama' \ -H 'accept: application/json' \ -H 'Content-Type: application/json' \ -d '{ "local_model_name": "qwen2.5:7b", "model_params": 7, "registered_name": "qwenforsummary", "set_priority_to_one": false, "stage": "production", "task": "text-generation-ollama", "version_description": "Register qwen2.5:7b local model" }'
-
Check if the registration was successful
curl -X 'GET' \ 'http://192.168.157.165:8010/models/registered_in_mlflow?show_all=false' \ -H 'accept: application/json'
-
Start audio/video/image/document service
9.1 Create Kafka topics:cd path/to/kafka chmod +x create_topic.sh sudo ./create_topic.sh9.2 Starting Services
cd services chmod +x start_services.sh ./start_services.sh9.3 Check if all services are still running
./check_services.sh
-
Produce a test requests topic
cd /path/to/kafka/test_service python test.py -
View Service Logs
cd path/to/kafka tail -f service_name.logAvailable service names include:
document_orchestrator_service,document_analysis_service,document_summary_service,document_save2qdrant_service, etc. You can replacedocumentwithaudio,video, orimagebased on the required service, for example:audio_orchestrator_service,video_analysis_service,image_summary_service, etc -
Check Redis Data
sudo docker exec -it redis-kafka_dev redis-cli --raw GET "document_orchestrator:correlation_id" GET "video_orchestrator:correlation_id" GET "audio_orchestrator:correlation_id" GET "image_orchestrator:correlation_id"
Note: Replace service_name and correlation_id with actual values.
-
Use the Qdrant Search API to query data
(8821/video_search, 8822/audio_search, 8823/document_search, 8824/image_search)curl -X POST "http://192.168.157.165:8822/audio_search" \ -H "Content-Type: application/json" \ -d '{ "query_text": "OpenAI", "embedding_type": "text", "limit": 5 }'
- π System Design & Architecture - High-level system design
- π§ Technical Implementation Guide - Detailed implementation
- π API Documentation - API references and examples
- π Quick Start Guide - Get started in minutes
- π CHANGELOG - Version history and updates
- GitHub Wiki (Coming soon) - Tutorials, guides, and best practices
- API Reference (Coming soon) - Complete API documentation
- Video Tutorials (Coming soon) - Step-by-step video guides
- Examples Repository (Coming soon) - Sample projects and use cases
RAPTOR leverages cutting-edge technologies:
AI & Machine Learning:
- π€ Large Language Models (LLM) - Multi-provider support
- π§ LangChain - LLM orchestration framework
- π Qdrant - High-performance vector database
- π MLflow - ML lifecycle management
- π― Sentence Transformers - Text embeddings
Backend & Infrastructure:
- β‘ FastAPI - Modern Python web framework
- π Python 3.8+ - Core programming language
- π³ Docker & Docker Compose - Containerization
- βΈοΈ Kubernetes - Container orchestration (v1.0+)
- π¨ Apache Kafka - Event streaming platform
- πΎ Redis Cluster - High-performance caching
Processing & Analysis:
- π₯ FFmpeg - Video/audio processing
- π Whisper - Speech recognition
- πΌοΈ OpenCV - Computer vision
- π PyPDF2, python-docx - Document processing
- π΅ Librosa - Audio analysis
Observability:
- π Prometheus - Metrics collection
- π Grafana - Metrics visualization
- π ELK Stack - Logging (roadmap)
We value your feedback and encourage community participation!
Please use GitHub Issues to:
- π Report bugs
- β¨ Request new features
- β Ask questions
- π‘ Share suggestions
Before opening an issue:
- Check existing issues to avoid duplicates
- Use issue templates when available
- Provide detailed information and steps to reproduce
Join our community on multiple platforms:
Follow us for updates, announcements, and community discussions!
Coming soon: Discord server, LinkedIn group, and monthly community calls!
- π Check our Documentation first
- π Browse closed issues for solutions
- π Start a Discussion for questions
- π§ Email us through dhtsolution.com
We'll post updates, respond to questions, and collaborate with users across these platforms!
We welcome contributions from the community! Please read our CONTRIBUTING.md guide to get started.
- Fork the repository
- Create a feature branch (
git checkout -b feature/AmazingFeature) - Commit your changes (
git commit -m 'Add some AmazingFeature') - Push to the branch (
git push origin feature/AmazingFeature) - Open a Pull Request
Please read our Code of Conduct before contributing.
This project is licensed under the Apache License 2.0 - see the LICENSE file for details.
Copyright 2025 DHT Taiwan Team
Licensed under the Apache License, Version 2.0 (the "License");
you may not use this file except in compliance with the License.
You may obtain a copy of the License at
http://www.apache.org/licenses/LICENSE-2.0
Unless required by applicable law or agreed to in writing, software
distributed under the License is distributed on an "AS IS" BASIS,
WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
See the License for the specific language governing permissions and
limitations under the License.
RAPTOR is developed and maintained by the DHT Taiwan Team.
About DHT Solutions
DHT Solutions is a technology company specializing in AI and software development solutions. Learn more at https://dhtsolution.com/.
Meet the talented developers behind RAPTOR:
![]() titanh |
![]() Cing-dht |
![]() fungdht |
![]() GeorgeDHT |
![]() NelsonYou1026 |
![]() tianyu0223 |
![]() Robertdht |
![]() QuinnChueh |
![]() Matthew20040407 |
![]() minnie-dhtsolution |
![]() lunar8386 |
![]() Joe-DHT |
![]() benjamin-dhtsolution |
The following features are planned for upcoming releases to transform RAPTOR into a production-ready, enterprise-grade platform:
- API Gateway: Production-ready gateway with rate limiting and API documentation portal
- Webhook System: Event-driven integration capability with webhook event system
- Implement temporal reasoning models for event sequences
- Add action recognition and activity detection
- Build scene relationship graphs
- Create timeline-based navigation interface
- Note: Docker Compose deployment won't scale to production needs
Implement Model Context Protocol (MCP) interfaces for core services:
- Document Processing
- Video Analysis
- Audio Processing
- Image Analysis
- Semantic Search
- Vector Database queries
- Model Management
- MCP Prompts
- Build fine-tuning workflows for domain-specific models
- Optimized Training: Automated pipelines with DeepSpeed optimization, model evaluation framework, customer data isolation
- Federated Learning: Privacy-preserving distributed training where data stays local, only model updates transferred
- Integrate Prometheus/Grafana metrics for real-time monitoring console
- Create user management UI with role-based access control
- Permission management and audit logging
- Migrate from basic JWT to OAuth2/OIDC standards
- Enterprise SSO support (Azure AD, Google Workspace, Okta)
- Multi-factor authentication
- Train content moderation models (NSFW, violence, hate speech)
- Implement automated flagging system
- Build GDPR/CCPA compliance workflows
- Create comprehensive audit reporting
- Production-ready Kubernetes deployment with Helm charts
- Automated horizontal and vertical scaling
- Service mesh integration for resilience
- Multi-environment support (dev/staging/prod)
- Critical: Docker Compose is not suitable for production scale
- Deploy ELK Stack (Elasticsearch, Logstash, Kibana)
- Configure log shipping from all 25+ services
- Centralized log aggregation and retention
- Advanced log search and analytics
- Distributed tracing integration
| Version | Target | Focus | Key Features |
|---|---|---|---|
| v0.2 | Q4 2025 | API & Integration | API Gateway, Webhooks |
| v0.3 | Q4 2025 | AI Enhancement | Video Understanding, MCP Interface |
| v0.4 | Q1 2026 | Enterprise Features | Training Pipeline, Admin Dashboard, OAuth2 |
| v0.5 | Q1 2026 | Compliance | Content Moderation, GDPR/CCPA |
| v1.0 | March 2026 | Production Ready | Kubernetes, ELK Stack, 99.9% SLA |
Current Status: Aigle 0.1 (Community Beta) - October 2025 β
Next Milestone: v0.2 with API Gateway (Q4 2025)
Production Target: v1.0 by March 2026
Made with β€οΈ by the DHT Taiwan Team
For business inquiries: https://dhtsolution.com/













