Job DescriptionAzure Data Architect
Brampton - Hybrid (2-3 days from Office)
Experience Range: 8+ years
1) Architecture & Solution Design
• Define end to end data platform reference architectures using Azure services such as Azure Synapse/SQL, Microsoft Fabric (Lakehouse, Warehouse, Semantic Models), Databricks, Azure Data Lake Storage Gen2, Data Factory/Synapse Pipelines, Event Hubs/Kafka, and Stream Analytics.
• Select appropriate storage and compute patterns (e.g., lakehouse, data mesh, medallion zones, Delta format) aligned to use cases, SLAs, and scalability needs.
• Produce and maintain HLD/LLD, Architecture Decision Records (ADRs), ERDs, and interface/API specifications.
________________________________________
2) Data Modeling & Engineering Patterns
• Establish dimensional (Kimball), 3NF, and Data Vault models where appropriate; promote semantic models for BI (Power BI / Fabric).
• Define ingestion and processing patterns (batch, streaming, CDC, micro batch) and file format standards (Parquet, Delta Lake, Avro).
• Set coding standards and performance guidelines for SQL, T SQL, PySpark, and Spark SQL.
________________________________________
3) Data Integration & Pipelines
• Architect robust ETL/ELT pipelines with ADF/Synapse Pipelines, Fabric Data Factory, Databricks Workflows, and orchestration using ADF triggers and Logic Apps.
• Integrate data from ERPs/CRMs, SaaS applications, APIs, and on prem systems via Self hosted Integration Runtime, Event Hubs, Kafka, Service Bus, and Change Data Capture tools.
________________________________________
4) Governance, Security & Compliance
• Implement data governance (catalogs, lineage, glossary, ownership) using Microsoft Purview; establish data mesh guardrails and federated governance practices.
• Enforce security using Azure AD (Entra ID), RBAC/ABAC, Managed Identities (MSI), Key Vault, Private Endpoints, VNet integration, row/column level security, dynamic data masking, and encryption at rest/in transit.
• Define data quality, master/reference data practices, and support audit, retention, and regulatory requirements.
________________________________________
5) Performance, Reliability & Cost Optimization
• Design for scalability and resilience (partitioning, caching, Z ordering, concurrency control, workload isolation).
• Tune Spark and SQL workloads; optimize storage layouts; configure auto optimize/auto compaction strategies (e.g., Delta).
• Own FinOps for data platforms: capacity planning, cost governance, tagging, budgets/alerts, and rightsizing.
________________________________________
6) DevOps, CI/CD & Observability
• Establish CI/CD for data and infrastructure using Azure DevOps or GitHub Actions, with Terraform/Bicep and deployment strategies (blue green, canary).
• Implement monitoring and alerting with Azure Monitor, Log Analytics, Fabric/Databricks observability, and custom dashboards; define SLOs/SLIs for data pipelines.
• Define backup & disaster recovery strategies and RPO/RTO targets for critical datasets.
________________________________________
7) Leadership & Collaboration
• Lead data architecture governance and design reviews; mentor data engineers and analysts.
• Partner with Product, Security, Enterprise Architecture, and Business stakeholders to align roadmaps and priorities.
• Drive vendor/tool evaluations, proofs of concept, and build vs buy decisions.
________________________________________
Skills & Experience
Required
• 7+ years in data architecture/engineering, including 3–5 years with Azure data services.
• Hands on expertise with several of the following:
o Azure Synapse Analytics / Azure SQL
o Microsoft Fabric (Lakehouse, Warehouse, Semantic Models / DAX)
o Databricks (Spark, Delta Lake, Unity Catalog), PySpark/Spark SQL
o ADLS Gen2, Data Factory / Synapse Pipelines / Fabric Data Factory
o Event Hubs, Service Bus, Kafka, Stream Analytics, CDC tools
• Strong data modeling (dimensional, 3NF, Data Vault) and ELT/ETL design.
• Proficient in SQL, T SQL, PySpark; solid understanding of metadata management and lineage.
• Security & networking basics: AAD, RBAC, VNets, Private Endpoints, Key Vault, Managed Identities.
• IaC & DevOps: Terraform/Bicep, Git, Azure DevOps/GitHub, automated testing for data pipelines.
Preferred
• Experience with data mesh and lakehouse patterns; MLOps handoffs (Feature Store, Model Registry).
• Performance engineering at large scale (TB–PB).
• Knowledge of data quality frameworks (e.g., Great Expectations, Deequ) and DQ scorecards.
• Exposure to Microsoft Purview (catalog, classifications, policies).
• Familiarity with privacy/compliance (SOX, HIPAA, GDPR, PCI).
RequirementsExperience (Years): 4-6 Essential Skills: • Work with project teams throughout the organization to design, implement, and manage CDN infrastructure using Akamai, to ensure high availability, performance, and scalability for customer facing applications and business processes. • Handle multiple priorities and assignments with excellence and precision. • Be a part of a 24/7/365 organization (some after hours support is expected as part of normal on-call rotation). • Directly support line of business development teams, provide guidance to them on implementation and changes for customer facing applications • Develop and maintain security protocols and measures to protect CDN infrastructure from cyber threats. • Monitor and analyze network performance, identifying and resolving issues to optimize content delivery of critical applications. • Collaborate with cross-functional teams to integrate Akamai CDN solutions with existing systems and applications. • Collaborate with information security teams to implement DDoS protection strategies and other security measures in the CDN. • Provide technical support and guidance to clients and internal teams regarding CDN and security best practices. • Work closely with vendor and professional service teams on delivery related activities and strategy. Qualifications: • Bachelor’s degree in Computer Science, Information Technology, or a related field. OR similar work experience. • Strong understanding of network protocols (HTTP/HTTPS, DNS, TCP/IP). • Proven experience as a CDN Engineer or similar role, with a strong focus on cybersecurity. • In-depth knowledge of Content Delivery Network technologies, including caching, load balancing, and content optimization. • Excellent problem-solving skills and attention to detail. • Strong communication and teamwork abilities. • Experience supporting 24/7/365 customer facing applications at enterprise scale. • Awareness and experience with cybersecurity tools and practices, such as firewalls, intrusion detection/prevention systems, and encryption. • Proficiency in scripting and automation (e.g., Python, Bash) a plus. • Relevant certifications (e.g., CISSP, CEH) are a plus but not required.