DataCROP (Data Collection Routing & Processing) is a configurable framework for real-time data collection, transformation, filtering, and management across IoT and cybersecurity domains. It emphasizes interoperability through a specialized data model for sources, processors, and results, enabling flexible workflow-driven analytics.
- Barley (v1.0): MongoDB, Apache Kafka, RabbitMQ, Kafka Streams, Node.js, React, optional Hyperledger Fabric.
- Farro (v2.0): Builds on Barley; MongoDB, Apache Kafka, RabbitMQ, Node.js, React, and algorithm support (Java, Python, R).
- Maize (v3.0, in progress): MongoDB, Apache Kafka, ELK stack; expanding observability and data services.
Deployable Farro demo: https://github.com/datacrop/farro-demo-deployment-scripts.
- Ensure Ruby and Bundler are installed.
- Install dependencies:
bundle install - Serve the site:
bundle exec jekyll serve - Visit
http://localhost:4000(adjustbaseurlif configured).
_content/home/: High-level framework overview and roadmap._content/overview/: Getting started and authentication with Keycloak._content/airflow/: Airflow processing engine deployment steps._content/creating-workflows/,_content/creating-data-models/,_content/worker/: Building workflows, data models, and workers._content/dev-guide/,_content/editor/,_content/user-guide/: Guidance for developers and end users.