Microsoft Certification

DP-900 — Microsoft Azure Data Fundamentals Study Guide

59 practice questions with correct answers and detailed explanations. Use this guide to review concepts before taking the practice exam.

▶ Take Practice Exam 59 questions  ·  Free  ·  No registration

About the DP-900 Exam

The Microsoft Microsoft Azure Data Fundamentals (DP-900) certification validates professional expertise in Microsoft technologies. This study guide covers all 59 practice questions from our DP-900 practice test, complete with correct answers and explanations to help you understand each concept thoroughly.

Review each question and explanation below, then test yourself with the full interactive practice exam to measure your readiness.

59 Practice Questions & Answers

Q1 Medium

What is the primary purpose of Azure Synapse Analytics?

  • A To manage only relational databases
  • B To replace Azure Storage accounts
  • C To provide unified analytics platform for big data and data warehousing ✓ Correct
  • D To handle real-time chat applications
Explanation

Azure Synapse Analytics integrates big data and data warehouse analytics into a single platform, enabling organizations to analyze data at scale.

Q2 Easy

Which Azure service is best suited for storing large volumes of unstructured data such as videos and images?

  • A Azure SQL Database
  • B Azure Blob Storage ✓ Correct
  • C Azure Table Storage
  • D Azure Queue Storage
Explanation

Azure Blob Storage is designed specifically for storing massive amounts of unstructured data including documents, videos, images, and backups.

Q3 Medium

What type of database model does Azure Cosmos DB primarily support?

  • A Hierarchical model exclusively
  • B Network model only
  • C Relational model only
  • D Multi-model including document, key-value, graph, and column-family ✓ Correct
Explanation

Azure Cosmos DB supports multiple data models including documents, key-value pairs, graphs, and column-family data, making it highly flexible for various application scenarios.

Q4 Easy

In the context of data classification, what does the term 'hot' tier refer to in Azure Storage?

  • A Data that has been corrupted and needs recovery
  • B Data that is accessed frequently and requires fast retrieval ✓ Correct
  • C Data that is accessed infrequently and stored for compliance
  • D Data that is encrypted with the highest security level
Explanation

The 'hot' tier in Azure Storage is optimized for frequently accessed data, providing the fastest retrieval speeds and highest access costs.

Q5 Medium

Which of the following statements about normalization in relational databases is correct?

  • A Normalization is only applicable to NoSQL databases
  • B Normalization eliminates the need for primary keys and foreign keys
  • C Normalization increases data redundancy to improve query performance
  • D Normalization reduces data redundancy and dependency by organizing data into separate tables ✓ Correct
Explanation

Database normalization is a process that reduces data redundancy and improves data integrity by organizing data into logical, related tables with proper relationships.

Q6 Medium

What is the primary advantage of using polyglot persistence in cloud architectures?

  • A It eliminates the need for data backups
  • B It guarantees lower costs than using traditional relational databases
  • C It ensures all data must be stored in a single database type
  • D It allows organizations to select the most appropriate database technology for each specific workload ✓ Correct
Explanation

Polyglot persistence enables teams to choose different database technologies optimized for specific use cases, such as document stores for flexible schemas and relational databases for structured data.

Q7 Hard

In Azure Data Lake Storage Gen2, what feature enables fine-grained access control at the file and directory level?

  • A Encryption at rest
  • B Hierarchical namespace
  • C Access Control Lists (ACLs) ✓ Correct
  • D Blob versioning
Explanation

Azure Data Lake Storage Gen2 uses Access Control Lists (ACLs) to provide fine-grained access control at the file and directory level, similar to POSIX permissions.

Q8 Medium

What is OLTP primarily used for in database systems?

  • A Processing large analytical queries across historical data for reporting
  • B Archiving data for long-term compliance storage
  • C Creating machine learning models from unstructured data
  • D Handling day-to-day transactional operations with frequent small updates ✓ Correct
Explanation

OLTP (Online Transactional Processing) is optimized for handling frequent, small transactions such as sales orders, inventory updates, and user registrations in real-time.

Q9 Hard

Which Azure service provides serverless SQL query capabilities for querying data in Azure Storage?

  • A Azure Data Factory
  • B Azure SQL Managed Instance
  • C Azure Synapse Analytics SQL on-demand ✓ Correct
  • D Azure Stream Analytics
Explanation

Azure Synapse Analytics SQL on-demand (now called serverless SQL pool) allows you to query files in Azure Storage using T-SQL without provisioning dedicated compute resources.

Q10 Medium

In data modeling, what is a dimension table primarily used for?

  • A Storing large numerical transaction amounts
  • B Managing user authentication and security credentials
  • C Containing descriptive attributes and characteristics for analytical queries ✓ Correct
  • D Recording high-frequency real-time events
Explanation

Dimension tables in star schemas or data warehouses store descriptive information like dates, locations, products, and customers that provide context for analyzing facts.

Q11 Hard

What is the primary difference between Azure SQL Database and Azure SQL Managed Instance?

  • A SQL Database supports larger data sizes than Managed Instance
  • B Managed Instance provides more control over database engine features and instance-level functionality compared to SQL Database ✓ Correct
  • C Managed Instance can only run on-premises while SQL Database is cloud-only
  • D SQL Database is serverless while Managed Instance requires server provisioning
Explanation

Azure SQL Managed Instance provides greater compatibility with SQL Server and instance-level features like cross-database queries and SQL Agent, whereas SQL Database is a fully managed PaaS offering.

Q12 Medium

Which characteristic defines a data warehouse?

  • A It stores historical data optimized for complex analytical queries and reporting ✓ Correct
  • B It exclusively stores unstructured data in its original format
  • C It primarily handles operational transactions and real-time updates
  • D It replaces the need for operational databases in organizations
Explanation

A data warehouse is specifically designed to store integrated, historical data optimized for analytical queries, reporting, and business intelligence rather than transactional operations.

Q13 Hard

What does eventual consistency mean in the context of distributed databases like Azure Cosmos DB?

  • A The database automatically deletes inconsistent records
  • B Data will eventually become consistent across all replicas after some time, though temporary inconsistencies may exist ✓ Correct
  • C Consistency is guaranteed only for backup copies
  • D All replicas are immediately synchronized after every write operation
Explanation

Eventual consistency is a model where updates to distributed replicas propagate over time, meaning brief inconsistencies may exist temporarily, but all replicas will converge to the same state.

Q14 Medium

In the Azure data platform, what is the primary purpose of Azure Data Factory?

  • A To orchestrate and automate data movement and transformation workflows ✓ Correct
  • B To provide real-time event streaming and processing
  • C To process large-scale analytical queries on data warehouses
  • D To create statistical models for predictive analytics
Explanation

Azure Data Factory is a cloud-based data integration service that enables creating, scheduling, and orchestrating data pipelines for ETL/ELT operations across various sources and destinations.

Q15 Easy

Which of the following best describes the 'archive' tier in Azure Blob Storage?

  • A For frequently accessed data requiring immediate retrieval with minimal latency
  • B For data that is encrypted and cannot be retrieved until decryption keys are provided
  • C For data that must be deleted automatically after a specified period
  • D For data rarely accessed and designed for long-term retention with lowest storage cost but higher retrieval latency ✓ Correct
Explanation

The archive tier is the most cost-effective storage option for data accessed infrequently, designed for long-term retention with retrieval latencies of hours (compared to seconds for hot tier).

Q16 Medium

What is a key characteristic of a fact table in a dimensional data model?

  • A It maintains historical snapshots of organizational structure
  • B It contains slowly changing descriptive information about business entities
  • C It stores measurable events or transactions with foreign keys linking to dimension tables ✓ Correct
  • D It exclusively handles user authentication and authorization data
Explanation

Fact tables in dimensional models contain quantitative measures and metrics of business events, with foreign keys that link to dimension tables to provide context for analysis.

Q17 Medium

Which Azure service is specifically designed for batch processing of large volumes of data using Apache Spark?

  • A Azure Stream Analytics
  • B Azure Logic Apps
  • C Azure Event Hubs
  • D Azure Synapse Spark pools ✓ Correct
Explanation

Azure Synapse Spark pools provide managed Apache Spark clusters within Azure Synapse Analytics, enabling batch processing and data transformation of large datasets.

Q18 Medium

In relational database design, what does referential integrity ensure?

  • A That all data values are formatted consistently across tables
  • B That all tables contain the same number of rows
  • C That foreign key values in a table correspond to valid primary key values in the referenced table ✓ Correct
  • D That backup copies are identical to the production database
Explanation

Referential integrity is a database constraint that ensures every foreign key value references an existing primary key value, maintaining valid relationships between tables.

Q19 Hard

What is the primary benefit of using partitioning in large tables?

  • A It replaces the need for backup and recovery procedures
  • B It improves query performance and enables efficient data management by dividing tables into smaller, more manageable segments ✓ Correct
  • C It automatically encrypts sensitive data columns
  • D It eliminates the need for database indexes
Explanation

Table partitioning divides large tables into smaller segments based on criteria like date ranges, enabling faster queries, more efficient maintenance, and better resource utilization.

Q20 Easy

Which statement accurately describes the difference between structured and unstructured data?

  • A Structured data requires more storage space than unstructured data of equivalent information content
  • B Structured data is always stored in relational databases while unstructured data is stored in cloud services only
  • C Structured data follows a predefined schema with organized fields, while unstructured data lacks a fixed format ✓ Correct
  • D Unstructured data is always larger in volume than structured data
Explanation

Structured data conforms to a defined schema with organized columns and types (like databases), while unstructured data lacks predefined organization (like documents, images, videos).

Q21 Medium

In Azure Synapse Analytics, what is the purpose of a dedicated SQL pool?

  • A To enable data warehousing with provisioned compute resources for complex analytical queries ✓ Correct
  • B To manage only small reference tables with minimal data volume
  • C To provide temporary storage for streaming data
  • D To handle real-time event processing with microsecond latency
Explanation

Dedicated SQL pools in Azure Synapse provide provisioned compute resources optimized for data warehousing and running complex analytical queries on large datasets.

Q22 Medium

What is a primary advantage of using Azure Data Lake Storage for big data analytics?

  • A It eliminates the need for data cleansing and validation procedures
  • B It automatically generates reports without requiring any data transformation
  • C It stores massive volumes of data in any format with hierarchical namespace and fine-grained access control suitable for analytics ✓ Correct
  • D It provides the fastest possible transaction processing for operational systems
Explanation

Azure Data Lake Storage is optimized for storing massive amounts of diverse data formats with hierarchical organization and security features ideal for big data analytics and machine learning workloads.

Q23 Hard

Which of the following scenarios would benefit most from using a document database like Azure Cosmos DB?

  • A Archiving historical logs that require sequential access patterns
  • B Managing structured employee payroll records with fixed fields and strict validation rules
  • C Storing user profiles that vary significantly in structure with flexible schema requirements and global distribution needs ✓ Correct
  • D Handling high-frequency financial transactions requiring strict ACID compliance
Explanation

Document databases like Azure Cosmos DB excel with flexible, semi-structured data like user profiles with varying attributes and provide global distribution with low latency.

Q24 Medium

What is the primary purpose of data classification in a data governance framework?

  • A To compress data files to reduce storage consumption automatically
  • B To randomly assign data to different storage locations for load balancing
  • C To categorize data based on sensitivity and compliance requirements to determine appropriate protection and handling ✓ Correct
  • D To identify which databases need to be upgraded to newer versions
Explanation

Data classification organizes data by sensitivity levels (public, internal, confidential) and compliance requirements, enabling organizations to apply appropriate security controls and handling procedures.

Q25 Medium

What is the primary purpose of a data warehouse?

  • A To replace all operational databases in an organization
  • B To consolidate data from multiple sources for analytical reporting and business intelligence ✓ Correct
  • C To store real-time transactional data for immediate processing
  • D To provide temporary storage during data migration processes
Explanation

A data warehouse is specifically designed to consolidate and integrate data from multiple operational sources to support analytical queries, reporting, and business intelligence rather than handle real-time transactions.

Q26 Medium

Which Azure service is best suited for serverless data analytics with SQL queries on files in Azure Storage?

  • A Azure SQL Database
  • B Azure Data Factory
  • C Azure Synapse Analytics ✓ Correct
  • D Azure Database for PostgreSQL
Explanation

Azure Synapse Analytics provides serverless SQL pools that allow querying data directly from files in Azure Storage without provisioning dedicated infrastructure.

Q27 Easy

What does OLTP primarily optimize for?

  • A Large-scale analytical queries across historical data
  • B Long-running batch processing jobs with high latency tolerance
  • C Fast, concurrent read and write operations on current operational data ✓ Correct
  • D Complex data transformations and aggregations
Explanation

OLTP (Online Transaction Processing) systems are optimized for rapid, concurrent transactions on current operational data with requirements for ACID compliance and low latency.

Q28 Medium

In Azure, which type of database is optimized for handling large volumes of unstructured data such as documents, images, and videos?

  • A Time-series databases like Azure Data Explorer
  • B NoSQL databases like Azure Cosmos DB or Azure Blob Storage ✓ Correct
  • C Graph databases exclusively available through third-party integrations
  • D Relational databases like Azure SQL Database
Explanation

Azure Cosmos DB and Azure Blob Storage are designed to handle unstructured data at scale, while Azure SQL Database is optimized for structured relational data.

Q29 Medium

What is the main difference between structured and semi-structured data?

  • A Structured data has a defined schema, while semi-structured data contains metadata tags but lacks a rigid schema ✓ Correct
  • B Structured data is always stored in relational databases, while semi-structured data is always stored in NoSQL systems
  • C Semi-structured data is older than structured data and less commonly used today
  • D There is no meaningful difference; the terms are interchangeable in modern database systems
Explanation

Structured data follows a rigid, predefined schema (like tables in SQL), while semi-structured data has some organizational structure through metadata or tags (like JSON or XML) but not a strict schema.

Q30 Medium

Which Azure service is designed for ingesting, processing, and analyzing streaming data in real-time?

  • A Azure Archive Storage
  • B Azure Event Hubs or Azure Stream Analytics ✓ Correct
  • C Azure Blob Storage
  • D Azure Table Storage
Explanation

Azure Event Hubs ingests high-volume streaming data while Azure Stream Analytics processes and analyzes it in real-time, making them ideal for streaming scenarios.

Q31 Medium

What is a key characteristic of NoSQL databases like Azure Cosmos DB?

  • A They provide flexible schemas and can scale horizontally across multiple servers ✓ Correct
  • B They enforce strict ACID transactions across all operations
  • C They are only suitable for small datasets under 1 GB
  • D They require complex JOIN operations to retrieve related data
Explanation

NoSQL databases like Cosmos DB offer flexible schemas, horizontal scalability, and are designed to handle large, distributed datasets without requiring complex JOIN operations.

Q32 Easy

Azure Data Factory is primarily used for which data operation?

  • A Interactive ad-hoc analytics on large datasets
  • B Extracting, transforming, and loading (ETL) data between sources ✓ Correct
  • C Storing unstructured binary files with high redundancy
  • D Real-time transactional processing of operational data
Explanation

Azure Data Factory is designed for orchestrating ETL pipelines that move and transform data from multiple sources to destinations at scheduled intervals.

Q33 Hard

Which normalization form ensures that all non-key attributes are functionally dependent only on the primary key?

  • A Boyce-Codd Normal Form (BCNF)
  • B Third Normal Form (3NF) ✓ Correct
  • C First Normal Form (1NF)
  • D Second Normal Form (2NF)
Explanation

Third Normal Form (3NF) requires that all non-key attributes depend only on the primary key, eliminating transitive dependencies that exist in lower normal forms.

Q34 Medium

What is the primary benefit of using Azure SQL Database managed backups?

  • A They provide unlimited storage capacity at no additional cost
  • B They automatically create point-in-time restore points and reduce manual backup administration ✓ Correct
  • C They guarantee zero data loss in all failure scenarios
  • D They eliminate the need for any disaster recovery planning
Explanation

Azure SQL Database automatically manages backups, creating point-in-time restore capabilities and reducing the operational overhead of manual backup management.

Q35 Hard

In the context of data analytics, what does 'denormalization' typically mean?

  • A Removing all relationships between database tables
  • B Intentionally introducing redundancy to improve query performance in analytical systems ✓ Correct
  • C Converting unstructured data into structured tables
  • D The process of fixing corrupted data in a warehouse
Explanation

Denormalization deliberately adds redundant data to reduce the need for complex JOINs, improving query performance in analytical scenarios where read speed is prioritized over update efficiency.

Q36 Medium

Which Azure service provides a distributed data processing framework for big data analytics similar to Hadoop or Spark?

  • A Azure Cognitive Services
  • B Azure Data Explorer
  • C Azure Machine Learning
  • D Azure HDInsight ✓ Correct
Explanation

Azure HDInsight is a managed service that provides Apache Hadoop, Spark, Hive, and other distributed processing frameworks for big data analytics.

Q37 Medium

What is the main advantage of using Azure Blob Storage's hot, cool, and archive tiers?

  • A They provide different performance levels that are invisible to applications
  • B They provide different encryption levels for security compliance
  • C They optimize costs by storing frequently accessed data in hot tier and infrequently accessed data in cooler tiers with lower costs ✓ Correct
  • D They automatically backup data to multiple geographic regions
Explanation

Azure Blob Storage tiers allow cost optimization by assigning different access patterns to appropriate tiers—hot for frequent access, cool and archive for less frequent access at lower costs.

Q38 Hard

Which statement best describes the CAP Theorem in distributed databases?

  • A Cloud systems should prioritize cost, availability, and performance in that order
  • B Consistency is always more important than availability in modern database design
  • C It guarantees that all systems can achieve perfect consistency, availability, and partition tolerance simultaneously
  • D A distributed system can guarantee at most two of three properties: consistency, availability, and partition tolerance ✓ Correct
Explanation

The CAP Theorem states that distributed databases must make trade-offs between consistency, availability, and partition tolerance—no system can perfectly guarantee all three simultaneously.

Q39 Easy

What is the primary purpose of indexing in a relational database?

  • A To improve query performance by reducing the amount of data that must be scanned ✓ Correct
  • B To increase storage capacity and reduce disk space requirements
  • C To eliminate the need for database normalization
  • D To enforce data type validation across all columns
Explanation

Indexes allow the database engine to quickly locate data without scanning every row, significantly improving query performance on frequently searched columns.

Q40 Medium

In Azure Cosmos DB, what does the term 'Request Unit (RU)' represent?

  • A The physical storage space allocated to a container in bytes
  • B The maximum number of concurrent users that can access a database
  • C A measure of throughput that represents the cost of performing read, write, and query operations ✓ Correct
  • D The replication factor used for global distribution across regions
Explanation

Request Units (RUs) in Cosmos DB are a normalized measure of throughput cost, where every operation consumes a certain number of RUs based on its computational complexity.

Q41 Hard

Which type of database schema design would be most appropriate for a data warehouse serving analytical queries?

  • A Document-based schema with no predefined structure to allow maximum flexibility
  • B Star or snowflake schema with dimension tables and fact tables optimized for query performance ✓ Correct
  • C A single denormalized table containing all business data to maximize simplicity
  • D Highly normalized third normal form to minimize storage and update operations
Explanation

Data warehouses typically use star or snowflake schemas that balance query performance through denormalization while maintaining some organizational structure around dimensions and facts.

Q42 Medium

What is the primary difference between Azure SQL Database and SQL Server running on Azure Virtual Machines?

  • A SQL Database cannot support large databases over 100 GB in size
  • B SQL Database supports more advanced features than SQL Server
  • C SQL Database is fully managed with automatic updates, while VM-based SQL Server requires manual administration ✓ Correct
  • D VM-based SQL Server is always more cost-effective for all workload types
Explanation

Azure SQL Database is a fully managed Platform-as-a-Service (PaaS) offering with automatic patching and maintenance, while SQL Server on Azure VMs (IaaS) requires manual administration by the user.

Q43 Hard

In data modeling, what is a 'grain' in the context of a fact table?

  • A The encryption algorithm used to secure sensitive data
  • B The physical location where data is stored in a data warehouse
  • C The process of cleaning and validating data quality
  • D The level of detail or granularity represented by each row in the fact table ✓ Correct
Explanation

The grain of a fact table defines the level of detail for each row—for example, 'daily sales by product by store' has a finer grain than 'monthly sales by region'.

Q44 Medium

Which Azure service allows you to query data across multiple Azure data sources using a single SQL-like interface?

  • A Azure Data Explorer
  • B Azure Portal
  • C Azure Synapse Analytics ✓ Correct
  • D Azure Data Catalog
Explanation

Azure Synapse Analytics provides a unified platform to query and analyze data across multiple sources including Data Lake, SQL databases, and other data stores with a single SQL interface.

Q45 Hard

What is the primary challenge of using polyglot persistence (multiple database technologies) in an organization?

  • A It increases operational complexity, requiring expertise in multiple technologies and managing data consistency across systems ✓ Correct
  • B It automatically optimizes all queries across all database systems
  • C It eliminates the need for data backup and disaster recovery
  • D It guarantees unlimited scalability without any performance trade-offs
Explanation

While polyglot persistence allows using the right database for each use case, it increases operational overhead by requiring teams to maintain expertise in multiple technologies and manage data consistency.

Q46 Medium

Which of the following best describes the relationship between data lakes and data warehouses?

  • A Data lakes completely replace the need for data warehouses in modern organizations
  • B Data lakes are exclusively for non-business use cases like machine learning experiments
  • C Data lakes store raw, unstructured data, while data warehouses store processed, structured data optimized for analytics ✓ Correct
  • D Data warehouses and data lakes are identical and the terms are used interchangeably
Explanation

Data lakes store raw, diverse data in its native format for flexibility and future use, while data warehouses store cleaned, structured data optimized for specific analytical queries and business intelligence.

Q47 Medium

In Azure, what is the purpose of Azure Private Link for databases?

  • A To encrypt all data in transit between client and server
  • B To provide private, secure network connectivity to Azure services without exposing them to the public internet ✓ Correct
  • C To automatically backup databases to multiple geographic regions
  • D To replicate data in real-time across multiple availability zones
Explanation

Azure Private Link establishes a private network connection between clients and Azure services, eliminating public internet exposure and improving security posture.

Q48 Medium

What does the principle of 'data sovereignty' require in cloud deployments?

  • A Every database must have a backup located at least 1000 miles away
  • B Data must be stored and processed within the geographic boundaries of specific countries or regions as required by regulations ✓ Correct
  • C All data must be encrypted using government-approved algorithms
  • D Organizations must use only government-owned and operated cloud infrastructure
Explanation

Data sovereignty requires that data be stored and processed within specific geographic boundaries as mandated by laws like GDPR and various country regulations.

Q49 Hard

Which normal form eliminates the possibility of anomalies caused by non-key attributes being dependent on other non-key attributes?

  • A Second Normal Form (2NF)
  • B Third Normal Form (3NF) ✓ Correct
  • C Fourth Normal Form (4NF)
  • D First Normal Form (1NF)
Explanation

Third Normal Form (3NF) removes transitive dependencies where non-key attributes depend on other non-key attributes, preventing update, insert, and delete anomalies.

Q50 Easy

Your organization needs to store large volumes of unstructured data such as videos, images, and documents. Which Azure service is best suited for this requirement?

  • A Azure Data Lake Storage Gen2
  • B Azure Blob Storage ✓ Correct
  • C Azure SQL Database
  • D Azure Cosmos DB
Explanation

Azure Blob Storage is optimized for storing massive amounts of unstructured data such as images, videos, and documents. While Data Lake Storage Gen2 is also suitable, Blob Storage is the primary service for general unstructured data.

Q51 Medium

You are designing a data warehouse solution for a retail company. The data warehouse needs to support complex analytical queries on historical sales data spanning five years. Which type of database would be most appropriate?

  • A Relational data warehouse optimized for analytical queries ✓ Correct
  • B Graph database
  • C Time-series database
  • D NoSQL document database
Explanation

A relational data warehouse (like Azure Synapse Analytics) is specifically designed for complex analytical queries on large volumes of historical data and supports OLAP operations effectively.

Q52 Medium

What is the primary advantage of using Azure Cosmos DB for globally distributed applications?

  • A It guarantees single-region data consistency
  • B It provides lower cost than traditional SQL databases
  • C It offers multi-region replication with guaranteed low-latency access globally ✓ Correct
  • D It automatically encrypts all data at rest without additional configuration
Explanation

Azure Cosmos DB is designed specifically for globally distributed applications, providing automatic multi-region replication and guaranteed low-latency access through its global distribution capabilities.

Q53 Medium

An organization wants to implement a data retention policy where data older than 7 years must be automatically deleted. Which Azure Storage feature enables this?

  • A Access tiers
  • B Encryption at rest
  • C Soft delete protection
  • D Lifecycle management policies ✓ Correct
Explanation

Lifecycle management policies in Azure Storage allow you to automatically transition blobs between access tiers or delete blobs based on age and other criteria, enabling automated data retention enforcement.

Q54 Easy

Which of the following best describes the difference between OLTP and OLAP workloads?

  • A OLTP handles numerous small, quick transactions while OLAP focuses on complex analytical queries across large datasets ✓ Correct
  • B OLTP is used exclusively for data warehousing while OLAP supports transactional applications
  • C OLTP requires more storage capacity than OLAP for the same volume of data
  • D OLTP and OLAP are the same type of workload with different naming conventions in Azure
Explanation

OLTP (Online Transaction Processing) optimizes for many small, fast transactions typical of operational systems, while OLAP (Online Analytical Processing) handles complex analytical queries on large historical datasets.

Q55 Medium

You need to query data in an Azure Data Lake using standard SQL syntax without moving the data. Which service should you use?

  • A Azure Stream Analytics
  • B Azure Data Factory
  • C Azure Databricks
  • D Azure Synapse Analytics ✓ Correct
Explanation

Azure Synapse Analytics allows you to query data in a data lake using standard SQL (via Serverless SQL pool) without requiring data movement or complex transformations.

Q56 Hard

An organization processes millions of events per second from IoT devices. The data must be ingested, processed in real-time, and stored for later analysis. Which combination of Azure services is most suitable?

  • A Azure Event Hubs for ingestion, Azure Stream Analytics for real-time processing, and Azure Blob Storage for storage ✓ Correct
  • B Azure Service Bus for ingestion, Azure Databricks for processing, and Azure Data Lake Storage for storage alone
  • C Azure Queue Storage for ingestion, Azure Functions for processing, and Azure Cosmos DB for storage only
  • D Azure Data Factory for ingestion, Azure Logic Apps for processing, and Azure SQL Database for storage
Explanation

Event Hubs is designed for high-throughput event ingestion, Stream Analytics provides real-time processing of streaming data, and Blob Storage (or Data Lake Storage) efficiently stores the processed results for analysis.

Q57 Medium

What is the primary purpose of normalization in a relational database?

  • A To reduce data redundancy and ensure data integrity through organized table structure ✓ Correct
  • B To automatically backup all database tables at regular intervals
  • C To encrypt sensitive data within database columns
  • D To increase query performance by denormalizing tables
Explanation

Normalization is a database design process that reduces data redundancy, minimizes anomalies, and ensures data integrity by organizing data into properly structured tables with appropriate relationships.

Q58 Hard

A company uses Azure SQL Database and experiences unpredictable performance during peak hours due to resource constraints. Which feature allows automatic scaling based on workload demand?

  • A Transparent data encryption
  • B Automatic failover groups
  • C Autoscale (serverless compute tier) ✓ Correct
  • D Query Store
Explanation

The serverless compute tier in Azure SQL Database includes autoscaling capabilities that automatically adjust compute resources based on workload demand, providing flexibility during peak hours.

Q59 Medium

Which Azure data service is specifically designed for time-series data and IoT analytics?

  • A Azure Event Grid
  • B Azure Time Series Insights ✓ Correct
  • C Azure Table Storage
  • D Azure Queue Storage
Explanation

Azure Time Series Insights is purpose-built for collecting, processing, storing, and analyzing time-series data from IoT devices and sensors at scale with specialized capabilities for temporal analysis.

Ready to test your knowledge?

You've reviewed all 59 questions. Take the interactive practice exam to simulate the real test environment.

▶ Start Practice Exam — Free