GCS Connector

High-throughput Go microservice for migrating and consolidating third-party data from Azure Data Lake into Google Cloud Storage for analytics.

GoAzureKubernetesGoogle Cloud Storage

Problem

Third-party operational data was being delivered into Azure Data Lake in inconsistent formats and needed to be reliably migrated into Google Cloud to support downstream analytics and reporting. Existing processes were manual, fragile, and did not scale as data volume and source diversity increased.

🛠️Solution

Designed and built a business-critical Go microservice responsible for high-throughput data migration from Azure Blob Storage into Google Cloud Storage. The service normalised disparate third-party data feeds, enforced schema and delivery guarantees, and provided a reliable ingestion layer for the analytics platform. The architecture was deliberately extensible, allowing additional third-party data sources to be onboarded with minimal change while maintaining performance and operational simplicity.

📈Impact

Enabled a scalable and reliable data pipeline feeding Google BigQuery and downstream analytics workloads. Reduced operational risk by replacing ad-hoc data transfers with a monitored, testable service. The connector became a core part of the organisation’s data ingestion strategy and was extended to support multiple third-party providers beyond the initial use case.

🎯Key Takeaways

Hands-on experience architecting and delivering production data pipelines in Go, from dependency selection and cloud service integration to performance tuning and operational concerns. Reinforced a delivery-focused approach to backend engineering, balancing maintainability, scalability, and real-world business requirements.