A multi-cloud message processor.
Mercury is a reference implementation of a multi-cloud queue processing service which utilizes the following technologies:
- Cloud PaaS Queues
- Cloud PaaS Storage
- Kubernetes
- KEDA
- OAUTH 2.0 and Open Identity
Mercury abstracts the cloud-specific implementation with a user-accessible API. This API provides a consistent interface for users regardless of the cloud the solution is running on, and currently supports Azure and Google Cloud Platform.
Mercury is a baseline implementation which should be forked and customized to fit your specific needs.
This solution uses managed identities to authenticate and access PaaS resources. This eliminates the security risk of storing PaaS keys. To set up managed identities see here.
IMPORTANT
For the Ingestion API to work properly, it must be configured with values for resources (e.g., Azure Queue). To configure these, follow the documentation here.
This is a lightweight cloud agnostic REST service written in .NET core. It accepts file and metadata payloads and stores them in cloud storage for later processing in a job queue. The API further provides access to retrieval of files in cloud storage.
The full README can be found here.
The Message Processor is a lightweight cloud agnostic service written in .NET
Core. The Message Processor watches a queue (e.g., Azure Storage Queue, GCP
Pub/Sub subscription), when a new item arrives, the Message Processor, reads
the message contents, downloads the correlated file, and performs a pre-defined,
task on the file. Currently, when a file is downloaded, the Message Processor
only uploads an acknowledgment file (NotifyReceived.txt
) to the cloud
agnostic storage, however, this application is designed to be higly
customizable.
The Message Processor is currently used in coordination with KEDA. KEDA is a Kubernetes Event Driven Autoscaler, which is used to automatically auto-scale when an event happens. In this scenario, when a new item is added to the queue.
The full README can be found here.
KEDA allows for fine grained autoscaling (including to/from zero) for event driven Kubernetes workloads. KEDA serves as a Kubernetes Metrics Server and allows users to define autoscaling rules using a dedicated Kubernetes custom resource definition.
KEDA can run on both the cloud and the edge, integrates natively with Kubernetes components such as the Horizontal Pod Autoscaler, and has no external dependencies.
The full README can be found here.
docker-compose build
The steps to run the solution locally for development can be found here
The Jenkins pipeline that builds and deploys the infrastructure and application for this solution is described here
Infrastructure is deployed using Terraform. Instructions to deploy can be found here.
Pull requests and issues are appreciated.
This project has adopted the Microsoft Open Source Code of Conduct. For more information see the Code of Conduct FAQ or contact opencode@microsoft.com with any additional questions or comments.