DEA-C02 무료 덤프문제 온라인 액세스
| 시험코드: | DEA-C02 |
| 시험이름: | SnowPro Advanced: Data Engineer (DEA-C02) |
| 인증사: | Snowflake |
| 무료 덤프 문항수: | 354 |
| 업로드 날짜: | 2026-01-13 |
You are developing a data pipeline that uses Snowpipe Streaming to ingest JSON data into a Snowflake table. Some JSON documents contain nested arrays and complex structures. You need to flatten the JSON structure during ingestion to simplify querying. Consider the following JSON document: { "order id": 123, "customer": { "id": "cust123", "name": "John Doe", "address": { "street": "123 Main St", "city": "Anytown" } }, "items": [ {"product_id": "prodl", "quantity": 2}, {"product_id": "prod2", "quantity": 1} ] } Which approach would you use within the 'COPY INTO' statement of your Snowpipe to flatten this JSON structure during ingestion?
You are using Snowpark Python to perform data transformation on a large dataset stored in a Snowflake table named customer transactions'. This table contains columns such as 'customer id', 'transaction date', 'transaction amount', and product_category'. Your task is to identify customers who have made transactions in more than one product category within the last 30 days. Which of the following Snowpark Python snippets is the most efficient way to achieve this, minimizing data shuffling and maximizing query performance?
You're configuring a Kafka connector to load data into Snowflake. You need to ensure that only records with a specific 'event_type' field equal to 'payment_processed' are loaded into the 'PAYMENT TABLE in Snowflake. Given the following connector configuration snippet, which additional property must be added to the 'transforms' section to achieve this filtering?
A data engineering team is running a series of complex analytical queries against a large Snowflake table. They notice that query performance is inconsistent, with some queries running much slower than others. After investigation, they determine that the queries are not properly leveraging the data clustering. Which of the following actions could improve the query performance related to the data clustering? Select all that apply.
You are designing a data pipeline that uses the Snowflake SQLAPI to execute a series of complex SQL queries. These queries involve multiple joins, aggregations, and user-defined functions (UDFs). You need to ensure that the pipeline is resilient to transient network errors and can handle a large volume of concurrent requests. Which of the following strategies would you implement to enhance the reliability and performance of your pipeline?