Alibaba aca-bigdata1 practice test

ACA Big Data Certification Exam Exam


Question 1

In order to ensure smooth processing of tasks in the Dataworks data development kit, you must
create an AccessKey. An AccessKey is primarily used for access permission verification between
various
Alibaba Cloud products. The AccessKey has two parts, they are ____. (Number of correct answers: 2)
Score 2

  • A. Access Username
  • B. Access Key ID
  • C. Access Key Secret
  • D. Access Password
Answer:

B, C

Discussions

Question 2

Resource is a particular concept of MaxCompute. If you want to use user-defined
function UDF or MapReduce, resource is needed. For example: After you have prepared UDF, you
must
upload the compiled jar package to MaxCompute as resource. Which of the following objects are
MaxCompute resources? (Number of correct answers: 4)
Score 2

  • A. Files
  • B. Tables: Tables in MaxCompute
  • C. Jar: Compiled Java jar package
  • D. Archive: Recognize the compression type according to the postfix in the resource name
  • E. ACL Policy
Answer:

A, B, C, E

Discussions

Question 3

MaxCompute is a fast and fully-managed TB/PB-level data warehousing solution provided by Alibaba
Cloud. Which of the following product features are correct? ______ (Number of correct answers: 3)
Score 2

  • A. Distributed architecture
  • B. High security and reliability
  • C. Multi-level management and authorization
  • D. Efficient transaction processing
  • E. Fast real-time response
Answer:

A, B, E

Discussions

Question 4

DataWorks provides powerful scheduling capabilities including time-based or dependency-based
task trigger mechanisms to perform tens of millions of tasks accurately and punctually each day
based
on DAG relationships. It supports multiple scheduling frequency configurations like: (Number of
correct
answers: 4)
Score 2

  • A. By Minute
  • B. By Hour
  • C. By Day
  • D. By Week
  • E. By Second
Answer:

A, B, C, D

Discussions

Question 5

The data development mode in DataWorks has been upgraded to the three-level structure
comprising of _____, _____, and ______. (Number of correct answers: 3)
Score 2

  • A. Project
  • B. Solution
  • C. Business flow
  • D. Directory
Answer:

A, B

Discussions

Question 6

The data development mode in DataWorks has been upgraded to the three-level structure
comprising of _____, _____, and ______. (Number of correct answers: 3)
Score 2

  • A. Project
  • B. Solution
  • C. Business flow
  • D. Directory
Answer:

A, B, D

Discussions

Question 7

DataWorks can be used to develop and configure data sync tasks. Which of the following statements
are correct? (Number of correct answers: 3)
Score 2

  • A. The data source configuration in the project management is required to add data source
  • B. Some of the columns in source tables can be extracted to create a mapping relationship between fields, and constants or variables can't be added
  • C. For the extraction of source data, "where" filtering clause can be referenced as the criteria of incremental synchronization
  • D. Clean-up rules can be set to clear or preserve existing data before data write
Answer:

A, B, D

Discussions

Question 8

MaxCompute can coordinate multiple users to operate one project through ACL authorization. The
objects that can be authorized by ACL include ______. (Number of correct answers: 3)
Score 2

  • A. Project
  • B. Table
  • C. Resource
  • D. Procedure
Answer:

A, B, C

Discussions

Question 9

A distributed file system like GFS and Hadoop are design to have much larger block(or chunk) size
like 64MB or 128MB, which of the following descriptions are correct? (Number of correct answers: 4)
Score 2

  • A. It reduces clients' need to interact with the master because reads and writes on the same block( or chunck) require only one initial request to the master for block location information
  • B. Since on a large block(or chunk), a client is more likely to perform many operations on a given block, it can reduce network overhead by keeping a persistent TCP connection to the metadata server over an extended period of time
  • C. It reduces the size of the metadata stored on the master
  • D. The servers storing those blocks may become hot spots if many clients are accessing the same small files
  • E. If necessary to support even larger file systems, the cost of adding extra memory to the meta data server is a big price
Answer:

A, B, C, D

Discussions

Question 10

Which of the following Hadoop ecosystem componets can you choose to setup a streaming log
analysis system?(Number of correct answers: 3)
Score 2

  • A. Apache Flume
  • B. Apache Kafka
  • C. Apache Spark
  • D. Apache Lucene
Answer:

A, C, D

Discussions
To page 2