Solutions Architect-L3
1 day ago
Irving
Job DescriptionAs a Solution Architect (L3), you will be the technical authority bridging business requirements and engineering execution. You will design end-to-end data and application architectures on Microsoft Azure, leveraging Databricks for large-scale data engineering and Power BI for enterprise analytics and reporting. You will lead cross-functional technical discussions, define data platform strategy, and ensure that solutions are scalable, resilient, secure, and aligned with organisational technology goals. This is a high-impact senior individual contributor role requiring deep cloud and data engineering expertise alongside strong stakeholder communication skills. Key Responsibilities Solution Architecture & Design • Define and own end-to-end solution architecture for complex, data-intensive, multi-service systems on Microsoft Azure., • Create and maintain architecture blueprints, HLDs, LLDs, data flow diagrams, and Architecture Decision Records (ADRs)., • Evaluate build-vs-buy decisions and drive technology selection with clear trade-off analysis across the Azure ecosystem., • Ensure solutions satisfy non-functional requirements: performance, scalability, reliability, data freshness SLAs, and security.Data Engineering & Platform Architecture, • Architect end-to-end data pipelines using Azure Databricks spanning ingestion, transformation (Bronze / Silver / Gold layers), and serving., • Design and govern data lakehouse solutions using Azure Data Lake Storage Gen2 (ADLS Gen2) and Delta Lake., • Define data modelling standards (star schema, data vault, OBT) aligned to business reporting requirements in Power BI., • Architect real-time and batch ingestion patterns using Azure Event Hubs, Azure Data Factory (ADF), and Databricks Structured Streaming., • Drive adoption of Databricks Unity Catalog for data governance, lineage tracking, and fine-grained access control., • Design cost-efficient Databricks cluster strategies (job clusters, instance pools, spot instances) and monitor DBU consumption., • Establish data quality frameworks using Great Expectations, Databricks Lakehouse Monitoring, or equivalent tooling.Analytics & Reporting Architecture, • Design scalable semantic layer and data models powering Power BI enterprise reports and dashboards., • Define DirectQuery vs Import vs Composite model strategies for optimal Power BI performance and data freshness., • Architect Power BI Premium / Fabric workspace strategies, including capacity planning and deployment pipelines (Dev / Test / Prod)., • Guide teams on DAX optimisation, row-level security (RLS), and Power BI paginated reports for operational reporting., • Integrate Power BI with Azure Databricks via Partner Connect or Lakehouse connector for live analytics on Delta tables.Azure Cloud & Infrastructure, • Architect cloud-native solutions leveraging Azure services: Azure Synapse Analytics, Azure Purview, Azure Key Vault, Azure Monitor, and APIM., • Define Infrastructure-as-Code (IaC) standards using Terraform or Bicep for repeatable, auditable deployments., • Design Azure networking topology: VNets, Private Endpoints, NSGs, and hub-spoke patterns to secure data flows., • Guide teams on containerisation (Docker, Azure Kubernetes Service) and Databricks-native MLflow for ML model management., • Define disaster recovery, backup, and high-availability strategies across the Azure data platform stack.Technical Leadership, • Serve as the primary technical point of contact across data engineering, analytics, product, and business teams., • Provide technical mentorship to senior data engineers, analytics engineers, and tech leads., • Lead architecture reviews, proof-of-concepts, and data platform migration spikes., • Drive engineering best practices: modular pipeline design, idempotency, schema evolution, and CI/CD for data workflows.Stakeholder Engagement & Governance, • Partner with data owners, business analysts, and C-suite stakeholders to translate analytical requirements into robust data products., • Present platform architecture, roadmaps, and data strategy to executive leadership with clarity and confidence., • Establish data governance frameworks covering cataloguing (Azure Purview), data classification, and lineage., • Ensure compliance with data privacy regulations (PDPB, GDPR) and security standards (ISO 27001, SOC 2) across the data platform. Required Qualifications Experience • 8 12 years of overall IT experience, with at least 4 years in a Solution Architect, Principal Data Engineer, or Senior Data Architect role., • Proven track record of architecting and delivering large-scale data platforms on Microsoft Azure in production., • Hands-on experience with Azure Databricks including workspace administration, cluster management, Delta Live Tables, and Unity Catalog., • Strong experience delivering Power BI enterprise analytics solutions semantic modelling, RLS, Premium capacities, and deployment pipelines., • Experience in Agile / SAFe delivery environments with strong DataOps and CI/CD practices for data pipelines.Azure Data Platform Proficiency, • Deep expertise in Azure Data Lake Storage Gen2, Delta Lake, and the Medallion (Bronze / Silver / Gold) architecture pattern., • Proficient in Azure Data Factory for pipeline orchestration, and Azure Event Hubs / Azure Service Bus for streaming ingestion., • Strong knowledge of Azure Synapse Analytics, Azure SQL Database / Managed Instance, and Cosmos DB., • Hands-on with Azure Purview (Microsoft Purview) for data cataloguing, classification, and lineage., • Experience with Azure Key Vault, Azure Active Directory / Entra ID, and Azure Private Link for platform security., • Familiarity with Azure Monitor, Log Analytics, and Databricks monitoring integrations for observability.Databricks & Data Engineering, • Expert-level Spark programming in PySpark and/or Scala for large-scale batch and streaming transformations., • Deep knowledge of Delta Lake features: ACID transactions, time travel, Z-ordering, liquid clustering, and schema enforcement., • Experience with Databricks Workflows (formerly Jobs), Delta Live Tables (DLT), and Databricks Asset Bundles for CI/CD., • Familiarity with Databricks MLflow for experiment tracking and model registry (ML workloads are a plus, not mandatory)., • Ability to design cost-optimised cluster policies, autoscaling strategies, and spot-instance configurations.Power BI & Analytics, • Strong Power BI development skills: data modelling (star schema), DAX measures, calculated columns, and report/dashboard design., • Experience with Power BI Premium / Fabric capacity management, paginated reports, and dataflows Gen2., • Proficiency in defining composite model strategies (DirectQuery + Import) and optimising query performance with aggregations., • Knowledge of Power BI REST APIs, XMLA endpoints, and Tabular Editor for advanced semantic model management.Soft Skills, • Exceptional written and verbal communication ability to author crisp architecture documents and present confidently to diverse audiences., • Strong analytical and problem-solving mindset with the ability to navigate ambiguity and competing priorities., • Collaborative leader who can influence without authority across data engineering, analytics, product, and business stakeholders.Education & Certifications, • E. / B.Tech / M.Tech in Computer Science, Information Technology, or a related field (or equivalent industry experience)., • Microsoft Certified: Azure Data Engineer Associate (DP-203) required., • Databricks Certified Associate / Professional Data Engineer strongly preferred., • Microsoft Certified: Azure Solutions Architect Expert (AZ-305) or Power BI Data Analyst (PL-300) a strong plus. Preferred Qualifications • Experience with Microsoft Fabric (OneLake, Fabric Data Engineering, Real-Time Analytics) the strategic evolution of the Azure data platform., • Exposure to dbt (data build tool) for SQL-based transformation layers and analytics engineering workflows., • Familiarity with Apache Iceberg or Apache Hudi as alternative open table formats alongside Delta Lake., • Prior experience with Azure Machine Learning or Databricks Model Serving for operationalising ML models., • Background in FinTech, HealthTech, Retail, or Manufacturing verticals with complex regulatory data requirements., • Contributions to open-source data projects or published technical content (blogs, conference talks, white papers)., • Exposure to data mesh principles and federated data ownership models.