Learn Design data storage solutions (AZ-305) with Interactive Flashcards

Master key concepts in Design data storage solutions through our interactive flashcard system. Click on each card to reveal detailed explanations and enhance your understanding.

Recommend a solution for storing relational data

When recommending a solution for storing relational data in Azure, several key services should be considered based on workload requirements. Azure SQL Database is the primary Platform-as-a-Service (PaaS) offering for relational data, providing a fully managed database engine with built-in high availability, automated backups, and intelligent performance tuning. It supports multiple deployment options including single databases, elastic pools for managing multiple databases with shared resources, and hyperscale tier for databases up to 100 TB. For organizations requiring SQL Server compatibility with minimal code changes during migration, Azure SQL Managed Instance offers near-complete SQL Server feature parity while maintaining PaaS benefits. This is ideal for lift-and-shift scenarios where legacy applications depend on SQL Server-specific features like SQL Agent, cross-database queries, or CLR integration. Azure SQL Database serverless tier is cost-effective for intermittent, unpredictable workloads as it automatically scales compute and bills per-second usage. For multi-region deployments requiring active geo-replication, Azure SQL Database supports readable secondary replicas across regions. When evaluating solutions, consider these factors: performance requirements (DTU vs vCore purchasing models), scalability needs, compliance requirements, disaster recovery objectives (RPO/RTO), and budget constraints. Elastic pools are recommended when managing multiple databases with varying usage patterns to optimize costs. For hybrid scenarios, Azure Arc-enabled SQL Server extends Azure management capabilities to on-premises SQL instances. Security features include transparent data encryption, Always Encrypted for sensitive columns, Advanced Threat Protection, and Azure Active Directory authentication. For analytical workloads combined with transactional data, consider Azure Synapse Link for near real-time analytics. The architect should assess data volume, concurrent user requirements, geographic distribution needs, and integration requirements with other Azure services before making final recommendations. Cost optimization strategies include reserved capacity purchases for predictable workloads and right-sizing based on actual performance metrics.

Recommend a database service tier and compute tier

When recommending database service and compute tiers in Azure, architects must evaluate workload requirements, performance needs, and budget constraints to select the optimal configuration.

For Azure SQL Database, three primary service tiers exist:

**DTU-based tiers:**
- **Basic**: Suitable for small databases with light workloads, offering limited performance at low cost
- **Standard**: Ideal for most business applications requiring moderate performance and storage
- **Premium**: Designed for mission-critical applications demanding high I/O throughput and low latency

**vCore-based tiers:**
- **General Purpose**: Balances compute and storage for typical business workloads with standard availability
- **Business Critical**: Provides highest resilience with built-in high availability replicas and fastest storage
- **Hyperscale**: Supports databases up to 100TB with rapid scale-out capabilities and near-instantaneous backups

**Compute tier selection** involves choosing between:
- **Provisioned compute**: Best for predictable workloads where you specify exact vCores needed, paying per hour
- **Serverless compute**: Optimal for intermittent usage patterns with auto-scaling and per-second billing during active periods

**Key considerations for recommendations:**
1. **Performance requirements**: Analyze DTU/vCore needs based on CPU, memory, and I/O demands
2. **Storage size**: Evaluate current data volume and growth projections
3. **Availability requirements**: Higher tiers offer better SLAs and redundancy options
4. **Latency sensitivity**: Business Critical tier provides in-memory OLTP and faster storage
5. **Cost optimization**: Match tier to actual usage patterns; avoid over-provisioning
6. **Scaling needs**: Consider whether workloads require elastic pools for multiple databases

Architects should analyze existing workload metrics, conduct performance testing, and consider future growth when making tier recommendations. Starting with lower tiers and scaling up based on monitoring data often proves more cost-effective than initial over-provisioning.

Recommend a solution for database scalability

Database scalability is crucial for Azure solutions to handle growing workloads efficiently. Azure offers several approaches to achieve scalability based on your specific requirements.

**Vertical Scaling (Scale Up)**: Increase compute resources like CPU, memory, and storage on a single database instance. Azure SQL Database and Azure Database for PostgreSQL/MySQL allow you to change service tiers seamlessly. This approach suits applications with unpredictable growth patterns.

**Horizontal Scaling (Scale Out)**: Distribute data across multiple database instances. Azure provides several options:

1. **Azure SQL Database Elastic Pools**: Share resources among multiple databases, ideal for SaaS applications with varying usage patterns. Databases can burst when needed while sharing pooled resources cost-effectively.

2. **Read Replicas**: Azure SQL Database, PostgreSQL, and MySQL support read replicas to offload read-heavy workloads. This distributes query load across multiple instances while maintaining a single write endpoint.

3. **Sharding**: Implement horizontal partitioning using Azure SQL Database Elastic Database tools. Data is distributed across multiple databases based on a sharding key, enabling massive scale for multi-tenant applications.

4. **Azure Cosmos DB**: For global-scale applications, Cosmos DB offers automatic horizontal partitioning, multi-region writes, and elastic throughput scaling. It handles millions of requests per second with guaranteed low latency.

**Hyperscale Service Tier**: Azure SQL Database Hyperscale supports databases up to 100TB with rapid scale-out read replicas and instant backups, perfect for large transactional workloads.

**Recommendations**:
- Use Elastic Pools for multi-tenant scenarios with variable workloads
- Implement read replicas for read-intensive applications
- Choose Cosmos DB for globally distributed applications requiring unlimited scale
- Consider Hyperscale for very large databases with demanding performance requirements
- Implement caching layers like Azure Cache for Redis to reduce database load

Monitor performance using Azure Monitor and configure auto-scaling policies to adjust resources based on actual demand patterns.

Recommend a solution for data protection

Data protection in Azure requires a comprehensive approach combining multiple strategies and services. For Azure Solutions Architect Expert certification, understanding these recommendations is essential.

**Backup Solutions:**
Azure Backup provides centralized backup management for various workloads including VMs, SQL databases, Azure Files, and on-premises resources. Configure backup policies with appropriate retention periods based on business requirements. Use Recovery Services Vaults to store backup data with geo-redundancy options.

**Replication Strategies:**
Implement Azure Site Recovery (ASR) for disaster recovery scenarios, enabling VM replication across regions. Choose appropriate storage redundancy: LRS (Locally Redundant Storage) for cost-effective protection, ZRS (Zone Redundant Storage) for availability zone failures, GRS (Geo-Redundant Storage) for regional disasters, and RA-GRS for read access to replicated data.

**Encryption:**
Enable encryption at rest using Azure Storage Service Encryption with Microsoft-managed or customer-managed keys stored in Azure Key Vault. Implement encryption in transit using TLS 1.2 or higher. For sensitive workloads, consider Azure Confidential Computing for data protection during processing.

**Access Control:**
Implement Azure RBAC with least privilege principles. Use Azure AD authentication for storage accounts. Configure private endpoints to restrict network access and enable Azure Private Link for secure connectivity.

**Soft Delete and Versioning:**
Enable soft delete for blob storage, file shares, and containers to protect against accidental deletion. Configure blob versioning to maintain previous versions of data for recovery purposes.

**Monitoring and Compliance:**
Use Azure Monitor and Azure Security Center to track data protection status. Implement Azure Policy for compliance enforcement and enable diagnostic logging for audit trails.

**Immutable Storage:**
For regulatory compliance, configure immutable blob storage with time-based retention or legal hold policies to prevent data modification or deletion.

The optimal solution combines these elements based on RPO/RTO requirements, compliance needs, and budget constraints.

Recommend a solution for storing semi-structured data

When recommending a solution for storing semi-structured data in Azure, Azure Cosmos DB stands out as the premier choice for most scenarios. Semi-structured data includes formats like JSON, XML, and key-value pairs that don't conform to rigid relational schemas but still maintain some organizational structure. Azure Cosmos DB offers multiple APIs including SQL, MongoDB, Cassandra, Gremlin, and Table, providing flexibility in how you interact with your data. It delivers single-digit millisecond latency, automatic indexing, and global distribution capabilities. For applications requiring massive scale and low latency across geographic regions, Cosmos DB excels with its turnkey global replication and guaranteed 99.999% availability. Azure Blob Storage with JSON files represents a cost-effective alternative for scenarios where you need to store large volumes of semi-structured data with less frequent access patterns. Combined with Azure Data Lake Storage Gen2, this approach works well for analytics workloads. Azure Table Storage provides a NoSQL key-attribute store suitable for simpler semi-structured datasets. It offers cost efficiency for applications needing flexible schemas and fast key-based lookups, though it lacks the advanced querying capabilities of Cosmos DB. When selecting between these options, consider throughput requirements, consistency needs, query complexity, global distribution requirements, and budget constraints. For mission-critical applications requiring low latency and global presence, Cosmos DB is ideal. For analytical workloads with batch processing, Data Lake Storage proves more appropriate. For simpler applications with basic query needs, Table Storage offers excellent value. Additionally, Azure SQL Database supports JSON data types, making it suitable when you need to combine relational and semi-structured data within a single database solution. This hybrid approach works well for applications transitioning from traditional relational models while incorporating flexible schema elements.

Recommend a solution for storing unstructured data

When designing storage solutions for unstructured data in Azure, Azure Blob Storage stands as the primary recommendation due to its scalability, cost-effectiveness, and versatility. Unstructured data includes files such as images, videos, documents, logs, and backups that lack a predefined data model.

Azure Blob Storage offers three access tiers to optimize costs based on data access patterns. The Hot tier suits frequently accessed data, providing lowest access costs but higher storage costs. The Cool tier works well for infrequently accessed data stored for at least 30 days, offering lower storage costs with slightly higher access costs. The Archive tier provides the most economical storage for rarely accessed data retained for at least 180 days.

For specific use cases, consider these alternatives within the Azure ecosystem. Azure Data Lake Storage Gen2 combines Blob Storage capabilities with hierarchical namespace, making it ideal for big data analytics workloads. It provides enhanced performance for analytical operations and integrates seamlessly with Azure Synapse Analytics and Azure Databricks.

Azure Files offers fully managed file shares accessible via SMB and NFS protocols, suitable for lift-and-shift scenarios where applications require traditional file system semantics. This service supports hybrid deployments through Azure File Sync.

When selecting a solution, evaluate factors including data volume, access frequency, latency requirements, and integration needs with other Azure services. Implement lifecycle management policies to automatically transition data between tiers based on age or access patterns, reducing overall storage costs.

Security considerations should include enabling encryption at rest using Microsoft-managed or customer-managed keys, implementing Azure Private Link for network isolation, and configuring appropriate access controls using Azure RBAC and shared access signatures. Additionally, enable soft delete and versioning to protect against accidental deletion and support data recovery scenarios. Geo-redundant storage options ensure business continuity by replicating data across Azure regions.

Recommend a data storage solution to balance features, performance, and costs

When recommending a data storage solution in Azure, architects must carefully evaluate features, performance, and costs to achieve optimal balance. The process involves analyzing workload requirements, data access patterns, and budget constraints. For structured transactional data with ACID compliance needs, Azure SQL Database offers multiple service tiers. The Basic and Standard tiers suit development and light workloads at lower costs, while Premium and Business Critical tiers provide higher IOPS and memory-optimized performance for mission-critical applications. Consider serverless compute for unpredictable workloads to optimize spending. For unstructured data, Azure Blob Storage presents tiered options. Hot tier serves frequently accessed data with higher storage costs but lower access fees. Cool tier reduces storage costs for data accessed monthly, while Archive tier offers the lowest storage pricing for rarely accessed data with higher retrieval latency and costs. Implementing lifecycle management policies automates tier transitions based on access patterns. NoSQL requirements benefit from Azure Cosmos DB, which provides multiple consistency levels affecting both performance and cost. Choosing eventual consistency over strong consistency reduces request unit consumption. Provisioned throughput works well for predictable workloads, while autoscale adapts to variable demand. For analytical workloads, Azure Synapse Analytics offers dedicated SQL pools for consistent high-performance queries and serverless options for intermittent analysis, allowing cost optimization based on usage patterns. Key recommendations include right-sizing resources based on actual performance metrics, implementing data tiering strategies, using reserved capacity for predictable workloads to achieve significant discounts, and regularly reviewing Azure Advisor recommendations. Architects should also consider data redundancy requirements, selecting between locally redundant, zone-redundant, or geo-redundant storage based on availability needs and budget. Monitoring tools like Azure Monitor and Cost Management help track performance metrics against spending, enabling continuous optimization of the storage solution over time.

Recommend a data solution for protection and durability

When recommending a data solution for protection and durability in Azure, architects must consider multiple layers of redundancy, backup strategies, and disaster recovery mechanisms. Azure provides several options to ensure data remains protected and available.

For storage redundancy, Azure offers four primary options: Locally Redundant Storage (LRS) maintains three copies within a single datacenter, Zone-Redundant Storage (ZRS) replicates across three availability zones, Geo-Redundant Storage (GRS) provides cross-region replication with six total copies, and Geo-Zone-Redundant Storage (GZRS) combines zone and geographic redundancy for maximum durability.

For databases, Azure SQL Database offers built-in automated backups with point-in-time restore capabilities, supporting retention periods up to 35 days. Long-term retention policies can extend this to 10 years. Active geo-replication enables readable secondary databases in different regions for disaster recovery scenarios.

Azure Cosmos DB provides automatic replication across multiple regions with configurable consistency levels. Its multi-master capability allows writes in any region, ensuring high availability during regional outages.

Azure Backup service offers centralized backup management for virtual machines, SQL databases, Azure Files, and on-premises workloads. It supports application-consistent snapshots and stores backups in Recovery Services vaults with built-in encryption.

Azure Site Recovery enables business continuity by orchestrating replication and failover of virtual machines between regions or from on-premises to Azure. This ensures minimal downtime during disasters.

For blob storage, soft delete protects against accidental deletions, while blob versioning maintains previous versions automatically. Immutable storage with WORM (Write Once Read Many) policies prevents modification or deletion for compliance requirements.

When designing solutions, architects should assess Recovery Time Objectives (RTO) and Recovery Point Objectives (RPO) to select appropriate redundancy levels. Combining multiple protection mechanisms—such as GRS storage with Azure Backup and Site Recovery—creates comprehensive data protection strategies that ensure business continuity and regulatory compliance.

Recommend a solution for data integration

Data integration in Azure is crucial for creating unified, accessible data solutions across an enterprise. As an Azure Solutions Architect, recommending the right data integration solution requires understanding various Azure services and their optimal use cases.

Azure Data Factory (ADF) serves as the primary orchestration service for data integration. It enables you to create data-driven workflows for moving and transforming data at scale. ADF supports over 90 built-in connectors, allowing seamless connections between on-premises and cloud data sources. It excels at ETL (Extract, Transform, Load) and ELT operations, making it ideal for data warehouse population and batch processing scenarios.

For real-time data integration, Azure Event Hubs and Azure Stream Analytics provide powerful streaming capabilities. Event Hubs can ingest millions of events per second, while Stream Analytics processes and analyzes streaming data using SQL-like queries. This combination suits IoT scenarios, live dashboards, and real-time analytics requirements.

Azure Synapse Analytics offers an integrated approach by combining data integration, enterprise data warehousing, and big data analytics. Its Synapse Pipelines feature, built on ADF technology, enables data movement and transformation within the same analytical workspace, reducing complexity and improving developer productivity.

For hybrid integration scenarios involving applications and APIs, Azure Logic Apps and Azure API Management provide low-code solutions. Logic Apps connects SaaS applications and enterprise systems through pre-built connectors, while API Management secures and manages API traffic.

When recommending a solution, consider factors such as data volume, velocity, variety, latency requirements, existing infrastructure, and team expertise. For complex enterprise scenarios, a combination of these services often provides the most comprehensive solution. Cost optimization, security compliance, monitoring capabilities, and disaster recovery requirements should also influence your architectural decisions. Proper data governance through Azure Purview ensures data quality and lineage tracking across all integration points.

Recommend a solution for data analysis

When recommending a solution for data analysis in Azure, architects must consider several key components to build a comprehensive analytics platform. Azure Synapse Analytics serves as the cornerstone for enterprise data warehousing and big data analytics, combining data integration, enterprise data warehousing, and big data analytics into a single unified platform. For real-time streaming data analysis, Azure Stream Analytics provides powerful capabilities to process millions of events per second from various sources like IoT devices, applications, and social media feeds. This serverless offering enables complex event processing with SQL-based queries. Azure Databricks offers an Apache Spark-based analytics platform optimized for Azure, ideal for machine learning workloads and collaborative data science projects. It integrates seamlessly with Azure Data Lake Storage Gen2, which provides hierarchical namespace capabilities and optimized performance for analytics workloads. For data orchestration and ETL processes, Azure Data Factory enables the creation of data pipelines that move and transform data across various sources and destinations. It supports both code-free and code-based approaches for building data workflows. Power BI completes the analytics stack by providing business intelligence capabilities with interactive visualizations and self-service reporting. It connects to multiple data sources and enables sharing insights across organizations. The recommended architecture typically follows a medallion pattern with bronze, silver, and gold layers in the data lake, progressively refining data quality. Azure Purview adds data governance capabilities, providing data cataloging and lineage tracking across the entire data estate. Cost optimization strategies include using dedicated SQL pools for predictable workloads, serverless options for ad-hoc queries, and implementing proper data lifecycle management policies. Security considerations encompass Azure Active Directory integration, managed identities, encryption at rest and in transit, and network isolation through private endpoints and virtual network service endpoints.

More Design data storage solutions questions
610 questions (total)