Questions and Answers
Question 2o6NH0MHvNPsfsoZHcpn
Question
A data engineer is using the following code block as part of a batch ingestion pipeline to read from a composable table:
//IMG//
Which of the following changes needs to be made so this code block will work when the transactions table is a stream source?
Choices
- A: Replace predict with a stream-friendly prediction function
- B: Replace schema(schema) with option (“maxFilesPerTrigger”, 1)
- C: Replace “transactions” with the path to the location of the Delta table
- D: Replace format(“delta”) with format(“stream”)
- E: Replace spark.read with spark.readStream
answer?
Answer: E Answer_ET: E Community answer E (100%) Discussion
Comment 1203854 by benni_ale
- Upvotes: 2
Selected Answer: E E is ok
Comment 1101644 by AndreFR
- Upvotes: 2
Selected Answer: E https://docs.databricks.com/en/structured-streaming/tutorial.html#use-auto-loader-to-read-streaming-data-from-object-storage
Comment 1084511 by 55f31c8
- Upvotes: 4
Selected Answer: E Example from https://docs.databricks.com/en/structured-streaming/delta-lake.html
spark.readStream.table(“table_name”)
spark.readStream.load(“/path/to/table”)
Comment 1050185 by meow_akk
- Upvotes: 3
Ans E; for streaming source you use readstream.
https://docs.databricks.com/en/structured-streaming/delta-lake.html
Question yQC9PZHxOUVrea9Jxt7X
Question
Which of the following queries is performing a streaming hop from raw data to a Bronze table?
Choices
- A:
- B:
- C:
- D:
- E:
answer?
Answer: E Answer_ET: E Community answer E (100%) Discussion
Comment 1064854 by mokrani
- Upvotes: 6
Selected Answer: E answer E: Raw to Bronze is simply an integration of source data in the lakehouse without any schema needed nor extra operationss (e;g filtering, aggregation, joins etc..) Please refer to this Medaillon Architecture article https://www.databricks.com/glossary/medallion-architecture
Comment 1203857 by benni_ale
- Upvotes: 1
Selected Answer: E E is ok , all others are incorrect
Comment 1101648 by AndreFR
- Upvotes: 2
sourcename is “rawSalesLocation” (bronze tables contain raw data) and code includes “readStream” to indicate that it is a streaming hop
Comment 1084513 by 55f31c8
- Upvotes: 2
Selected Answer: E https://docs.databricks.com/en/lakehouse/medallion.html#ingest-raw-data-to-the-bronze-layer
Comment 1061912 by sodere
- Upvotes: 1
Answer is B
Question FrR59kSDWjqPw8LYP5lc
Question
Which of the following Git operations must be performed outside of Databricks Repos?
Choices
- A: Commit
- B: Pull
- C: Push
- D: Clone
- E: Merge
answer?
Answer: E Answer_ET: E Community answer E (68%) D (32%) Discussion
Comment 922506 by ZSun
- Upvotes: 24
According to the most recent one, all command is feasible in Repos
Comment 994992 by NickWerbung
- Upvotes: 18
Not valid anymore… https://docs.databricks.com/en/repos/ci-cd-techniques-with-repos.html
Comment 1411231 by devbila
- Upvotes: 1
Selected Answer: E It is E
Comment 1365866 by chinhuy
- Upvotes: 1
Selected Answer: E https://docs.databricks.com/aws/en/repos/
Comment 1339435 by sakis213
- Upvotes: 2
Selected Answer: D Clone us correct
Comment 1334379 by danishanis
- Upvotes: 2
Selected Answer: D Option D - Clone. The Clone operation must be performed outside of Databricks Repos. Cloning a repository involves creating a copy of an existing remote repository to a local machine. In the context of Databricks, you would typically clone the repository to your local development environment first and then connect it to the Databricks repos
Comment 1328147 by MultiCloudIronMan
- Upvotes: 3
Selected Answer: D The reason D. Clone is the correct answer is because cloning a repository involves creating a copy of the entire repository, including all of its history, branches, and files, on your local machine. This operation is typically performed outside of Databricks Repos, using Git commands in a terminal or a Git client.
On the other hand, operations like Commit, Pull, Push, and Merge can be performed within Databricks Repos, as they involve interacting with the repository’s content and history that is already cloned and available in the Databricks environment.
Comment 1303998 by peadar_pa
- Upvotes: 1
For clarity, is the consensus that you can do all of the options, therefore making this an invalid question?
Comment 1272800 by 9d4d68a
- Upvotes: 4
D. Clone
Cloning a repository creates a local copy of the repository on your machine and must be done using a local Git client or command line. Once the repository is cloned, you can work with it in Databricks Repos, but the initial clone operation itself is outside the Databricks interface. Other operations like commit, pull, push, and merge can be managed within Databricks Repos or through other Git tools.
Comment 1262390 by 80370eb
- Upvotes: 2
Selected Answer: D D: Cloning a repository is typically done outside of Databricks Repos, often using a Git client or command line interface before the repository is linked to Databricks Repos.
Comment 1202790 by Isio05
- Upvotes: 4
Confirmed on live environment - merging is now possible directly in Databricks Repos
Comment 1177171 by Itmma
- Upvotes: 1
Selected Answer: E E is correct
Comment 1156957 by Bob123456
- Upvotes: 1
Why not option B Pull
The following tasks are not supported by Databricks Repos, and must be performed in your Git provider: Create a pull request Delete branches Merge and rebase branches *
Comment 1115828 by vvg130
- Upvotes: 7
The new answer is F - Delete .
Comment 1104707 by SerGrey
- Upvotes: 1
Selected Answer: E E is correct
Comment 1085129 by nedlo
- Upvotes: 1
i think it supports merge now https://docs.databricks.com/en/repos/git-operations-with-repos.html “If an operation such as pull, rebase, or merge causes a merge conflict, the Repos UI shows a list of files with conflicts and options for resolving the conflicts.
You have two primary options:
Use the Repos UI to resolve the conflict.”
Comment 1073385 by Huroye
- Upvotes: 1
E is the correct answer given the selections. You can clone.
Comment 1057877 by mokrani
- Upvotes: 2
According to the recent version, all commands are supported under Repos !
Comment 1017344 by KalavathiP
- Upvotes: 2
Selected Answer: E E is correct
Comment 916151 by fred_camargo
- Upvotes: 2
Selected Answer: E Merge is the correct answer
Comment 889039 by Majjjj
- Upvotes: 2
Selected Answer: E For following tasks, work in your Git provider:
Create a pull request. Resolve merge conflicts. Merge or delete branches. Rebase a branch.
https://docs.databricks.com/repos/index.html
Comment 876195 by Varma_Saraswathula
- Upvotes: 1
Merge - A
Comment 875851 by naxacod574
- Upvotes: 1
merge is not supported
Comment 867432 by SireeJ
- Upvotes: 1
Option: C
Comment 863859 by Data_4ever
- Upvotes: 3
Selected Answer: E MERGE is the only git operation that is listed in the options that cannot be performed with Databricks repos. CLONE is absolutely possible
Comment 861149 by upliftinghut
- Upvotes: 3
Selected Answer: E wrong answer, Clone can be done in Databricks Repo. Merge not in Repos, need to be in Git. Link here: https://learn.microsoft.com/en-us/azure/databricks/repos/
Comment 860618 by sdas1
- Upvotes: 1
option E
Comment 859452 by knivesz
- Upvotes: 3
Selected Answer: E La correcta es la E, Databricks Repose no tiene soporte para MERGE
Comment 857963 by XiltroX
- Upvotes: 1
Correct answer
Question 7F5ka1vah4isfsQVpQLz
Question
A dataset has been defined using Delta Live Tables and includes an expectations clause:
CONSTRAINT valid_timestamp EXPECT (timestamp > ‘2020-01-01’) ON VIOLATION FAIL UPDATE
What is the expected behavior when a batch of data containing data that violates these constraints is processed?
Choices
- A: Records that violate the expectation are dropped from the target dataset and recorded as invalid in the event log.
- B: Records that violate the expectation cause the job to fail.
- C: Records that violate the expectation are dropped from the target dataset and loaded into a quarantine table.
- D: Records that violate the expectation are added to the target dataset and recorded as invalid in the event log.
- E: Records that violate the expectation are added to the target dataset and flagged as invalid in a field added to the target dataset.
answer?
Answer: B Answer_ET: B Community answer B (100%) Discussion
Comment 1282234 by CommanderBigMac
- Upvotes: 1
Selected Answer: B B is the way
Comment 1203859 by benni_ale
- Upvotes: 1
Selected Answer: B b is ok
Comment 1084529 by 55f31c8
- Upvotes: 2
Selected Answer: B https://docs.databricks.com/en/delta-live-tables/sql-ref.html#sql-properties ON VIOLATION Optional action to take for failed rows: FAIL UPDATE: Immediately stop pipeline execution. DROP ROW: Drop the record and continue processing.
Comment 1065250 by Bakhtiyor
- Upvotes: 2
ON VIOLATION FAIL UPDATE: Immediately stop pipeline execution. DROP ROW: Drop the record and continue processing.
Comment 1050187 by meow_akk
- Upvotes: 4
Ans B : delta live tables data quality expectations . - https://docs.databricks.com/en/delta-live-tables/expectations.html Action
Result
warn (default)
Invalid records are written to the target; failure is reported as a metric for the dataset.
drop
Invalid records are dropped before data is written to the target; failure is reported as a metrics for the dataset.
fail
Invalid records prevent the update from succeeding. Manual intervention is required before re-processing.
Question C8IDYeCMeQrnzfxljjrT
Question
Which of the following statements regarding the relationship between Silver tables and Bronze tables is always true?
Choices
- A: Silver tables contain a less refined, less clean view of data than Bronze data.
- B: Silver tables contain aggregates while Bronze data is unaggregated.
- C: Silver tables contain more data than Bronze tables.
- D: Silver tables contain a more refined and cleaner view of data than Bronze tables.
- E: Silver tables contain less data than Bronze tables.
answer?
Answer: D Answer_ET: D Community answer D (100%) Discussion
Comment 1282237 by CommanderBigMac
- Upvotes: 1
Selected Answer: D Silver tables are used to clean the raw imported data from a bronze table
Comment 1203860 by benni_ale
- Upvotes: 1
Selected Answer: D d is ok
Comment 1127424 by azure_bimonster
- Upvotes: 2
Selected Answer: D D is the right answer
Comment 1050188 by meow_akk
- Upvotes: 2
Ans D : medallion arch databricks https://www.databricks.com/glossary/medallion-architecture