Weekend Sale Limited Time 70% Discount Offer - Ends in 0d 00h 00m 00s - Coupon code: xmas50

Huawei H13-321_V2.5 - HCIP - AI EI Developer V2.5 Exam

Page: 1 / 2
Total 60 questions

In an image preprocessing experiment, the cv2.imread("lena.png", 1) function provided by OpenCV is used to read images. The parameter "1" in this function represents a --------- -channel image. (Fill in the blank with a number.)

In the field of deep learning, which of the following activation functions has a derivative not greater than 0.5?

A.

SeLU

B.

Sigmoid

C.

ReLU

D.

Tanh

Which of the following applications are supported by ModelArts ExeML?

A.

Predictive maintenance of manufacturing equipment

B.

Dress code conformance monitoring in campuses

C.

Anomalous sound detection in production or security scenarios

D.

Automatic offering classification

The attention mechanism in foundation model architectures allows the model to focus on specific parts of the input data. Which of the following steps are key components of a standard attention mechanism?

A.

Calculate the dot product similarity between the query and key vectors to obtain attention scores.

B.

Compute the weighted sum of the value vectors using the attention weights.

C.

Apply a non-linear mapping to the result obtained after the weighted summation.

D.

Normalize the attention scores to obtain attention weights.

Which of the following statements about the standard normal distribution are true?

A.

The variance is 0.

B.

The mean is 1.

C.

The variance is 1.

D.

The mean is 0.

The natural language processing field usually uses distributed semantic representation to represent words. Each word is no longer a completely orthogonal 0-1 vector, but a point in a multi-dimensional real number space, which is specifically represented as a real number vector.

A.

TRUE

B.

FALSE

Maximum likelihood estimation (MLE) requires knowledge of the sample data's distribution type.

A.

TRUE

B.

FALSE

In 2017, the Google machine translation team proposed the Transformer in their paperAttention is All You Need. The Transformer consists of an encoder and a(n) --------. (Fill in the blank.)

Which of the following is not an algorithm for training word vectors?

A.

TextCNN

B.

BERT

C.

FastText

D.

Word2Vec

In NLP tasks, transformer models perform well in multiple tasks due to their self-attention mechanism and parallel computing capability. Which of the following statements about transformer models are true?

A.

Transformer models outperform RNN and CNN in processing long texts because they can effectively capture global dependencies.

B.

Multi-head attention is the core component of a transformer model. It computes multiple attention heads in parallel to capture semantic information in different subspaces.

C.

A transformer model directly captures the dependency between different positions in the input sequence through the self-attention mechanism, without using the recurrent neural network (RNN) or convolutional neural network (CNN).

D.

Positional encoding is optional in a transformer model because the self-attention mechanism can naturally process the order information of sequences.