This page was exported from Free valid test braindumps [ http://free.validbraindumps.com ]
Export date: Sat Apr 5 14:25:38 2025 / +0000 GMT

Pass Your Exam With 100% Verified Associate-Developer-Apache-Spark Exam Questions [Q20-Q44]




Pass Your Exam With 100% Verified Associate-Developer-Apache-Spark Exam Questions

Associate-Developer-Apache-Spark Dumps PDF - Associate-Developer-Apache-Spark Real Exam Questions Answers

Q20. Which of the following statements about the differences between actions and transformations is correct?

 
 
 
 
 

Q21. Which of the following statements about Spark’s execution hierarchy is correct?

 
 
 
 
 

Q22. Which of the following code blocks creates a new DataFrame with two columns season and wind_speed_ms where column season is of data type string and column wind_speed_ms is of data type double?

 
 
 
 
 
 

Q23. Which of the following describes a way for resizing a DataFrame from 16 to 8 partitions in the most efficient way?

 
 
 
 

Q24. Which of the following describes slots?

 
 
 
 

Q25. Which of the following code blocks displays various aggregated statistics of all columns in DataFrame transactionsDf, including the standard deviation and minimum of values in each column?

 
 
 
 
 

Q26. Which of the following code blocks returns a new DataFrame with the same columns as DataFrame transactionsDf, except for columns predError and value which should be removed?

 
 
 
 
 

Q27. Which of the following describes Spark’s way of managing memory?

 
 
 
 
 

Q28. Which of the following code blocks returns a single row from DataFrame transactionsDf?
Full DataFrame transactionsDf:
1.+————-+———+—–+——-+———+—-+
2.|transactionId|predError|value|storeId|productId| f|
3.+————-+———+—–+——-+———+—-+
4.| 1| 3| 4| 25| 1|null|
5.| 2| 6| 7| 2| 2|null|
6.| 3| 3| null| 25| 3|null|
7.| 4| null| null| 3| 2|null|
8.| 5| null| null| null| 2|null|
9.| 6| 3| 2| 25| 2|null|
10.+————-+———+—–+——-+———+—-+

 
 
 
 
 

Q29. The code block displayed below contains an error. The code block is intended to join DataFrame itemsDf with the larger DataFrame transactionsDf on column itemId. Find the error.
Code block:
transactionsDf.join(itemsDf, “itemId”, how=”broadcast”)

 
 
 
 
 

Q30. The code block displayed below contains an error. When the code block below has executed, it should have divided DataFrame transactionsDf into 14 parts, based on columns storeId and transactionDate (in this order). Find the error.
Code block:
transactionsDf.coalesce(14, (“storeId”, “transactionDate”))

 
 
 
 
 

Q31. The code block shown below should read all files with the file ending .png in directory path into Spark.
Choose the answer that correctly fills the blanks in the code block to accomplish this.
spark.__1__.__2__(__3__).option(__4__, “*.png”).__5__(path)

 
 
 
 
 

Q32. The code block displayed below contains an error. The code block is intended to return all columns of DataFrame transactionsDf except for columns predError, productId, and value. Find the error.
Excerpt of DataFrame transactionsDf:
transactionsDf.select(~col(“predError”), ~col(“productId”), ~col(“value”))

 
 
 
 
 

Q33. Which of the following describes characteristics of the Spark driver?

 
 
 
 
 

Q34. Which of the following options describes the responsibility of the executors in Spark?

 
 
 
 
 

Q35. The code block shown below should return a copy of DataFrame transactionsDf with an added column cos.
This column should have the values in column value converted to degrees and having the cosine of those converted values taken, rounded to two decimals. Choose the answer that correctly fills the blanks in the code block to accomplish this.
Code block:
transactionsDf.__1__(__2__, round(__3__(__4__(__5__)),2))

 
 
 
 

Q36. The code block shown below should store DataFrame transactionsDf on two different executors, utilizing the executors’ memory as much as possible, but not writing anything to disk. Choose the answer that correctly fills the blanks in the code block to accomplish this.
1.from pyspark import StorageLevel
2.transactionsDf.__1__(StorageLevel.__2__).__3__

 
 
 
 
 

Q37. The code block displayed below contains an error. The code block should count the number of rows that have a predError of either 3 or 6. Find the error.
Code block:
transactionsDf.filter(col(‘predError’).in([3, 6])).count()

 
 
 
 
 

Q38. In which order should the code blocks shown below be run in order to assign articlesDf a DataFrame that lists all items in column attributes ordered by the number of times these items occur, from most to least often?
Sample of DataFrame articlesDf:
1.+——+—————————–+——————-+
2.|itemId|attributes |supplier |
3.+——+—————————–+——————-+
4.|1 |[blue, winter, cozy] |Sports Company Inc.|
5.|2 |[red, summer, fresh, cooling]|YetiX |
6.|3 |[green, summer, travel] |Sports Company Inc.|
7.+——+—————————–+——————-+

 
 
 
 
 
 

Q39. The code block shown below should write DataFrame transactionsDf to disk at path csvPath as a single CSV file, using tabs (t characters) as separators between columns, expressing missing values as string n/a, and omitting a header row with column names. Choose the answer that correctly fills the blanks in the code block to accomplish this.
transactionsDf.__1__.write.__2__(__3__, ” “).__4__.__5__(csvPath)

 
 
 
 

Q40. Which of the following is a viable way to improve Spark’s performance when dealing with large amounts of data, given that there is only a single application running on the cluster?

 
 
 
 
 

Q41. Which of the following describes the conversion of a computational query into an execution plan in Spark?

 
 
 
 
 

Q42. Which of the following is a characteristic of the cluster manager?

 
 
 
 
 

Q43. Which of the following code blocks returns all unique values of column storeId in DataFrame transactionsDf?

 
 
 
 
 

Q44. Which of the following statements about broadcast variables is correct?

 
 
 
 
 

Associate-Developer-Apache-Spark Dumps 100 Pass Guarantee With Latest Demo: https://www.validbraindumps.com/Associate-Developer-Apache-Spark-exam-prep.html

Post date: 2023-03-10 11:08:55
Post date GMT: 2023-03-10 11:08:55
Post modified date: 2023-03-10 11:08:55
Post modified date GMT: 2023-03-10 11:08:55