AI & Machine Learning

Neural Networks — Practice MCQs for CCAT

20 Questions Section B: Programming AI & Machine Learning

Practice 20 Neural Networks multiple-choice questions designed for CDAC CCAT exam preparation. Click "Show Answer" to reveal the correct option with detailed explanation.

Q1.
Neural networks are inspired by:
AComputer circuits
BBiological neurons in the brain
CDatabase systems
DNetwork protocols
Show Answer & Explanation

Correct Answer: B — Biological neurons in the brain

Artificial neural networks are inspired by the structure and function of biological neural networks in the brain.

Q2.
A perceptron is:
AMulti-layer network
BSingle layer neural network with one neuron
CDeep network
DRecurrent network
Show Answer & Explanation

Correct Answer: B — Single layer neural network with one neuron

A perceptron is the simplest neural network - a single neuron that makes binary classifications.

Q3.
Activation function introduces:
ALinearity
BNon-linearity to the network
CMore layers
DMore neurons
Show Answer & Explanation

Correct Answer: B — Non-linearity to the network

Activation functions introduce non-linearity, allowing networks to learn complex patterns.

Q4.
ReLU activation function is:
Af(x) = sigmoid(x)
Bf(x) = max(0, x)
Cf(x) = tanh(x)
Df(x) = x^2
Show Answer & Explanation

Correct Answer: B — f(x) = max(0, x)

ReLU (Rectified Linear Unit) returns x if positive, otherwise 0. Simple yet effective.

Q5.
Sigmoid function outputs values between:
A-1 and 1
B0 and 1
C-infinity and infinity
D0 and infinity
Show Answer & Explanation

Correct Answer: B — 0 and 1

Sigmoid squashes values to range (0, 1), useful for probability outputs.

Q6.
Backpropagation is used to:
AInitialize weights
BCalculate gradients and update weights
CAdd more layers
DRemove neurons
Show Answer & Explanation

Correct Answer: B — Calculate gradients and update weights

Backpropagation calculates gradients of the loss with respect to weights, enabling weight updates.

Q7.
Gradient descent minimizes:
ANumber of neurons
BLoss/error function
CNumber of layers
DTraining time
Show Answer & Explanation

Correct Answer: B — Loss/error function

Gradient descent iteratively adjusts weights to minimize the loss function.

Q8.
Learning rate in neural networks:
ADetermines number of epochs
BControls step size in weight updates
CSets batch size
DDefines layer count
Show Answer & Explanation

Correct Answer: B — Controls step size in weight updates

Learning rate controls how much weights are adjusted in each update step during training.

Q9.
Epoch in training means:
ASingle weight update
BOne complete pass through entire training dataset
CModel deployment
DTest evaluation
Show Answer & Explanation

Correct Answer: B — One complete pass through entire training dataset

An epoch is one complete pass through the entire training dataset during training.

Q10.
Deep learning refers to:
AShallow networks
BNeural networks with many layers
CSingle perceptron
DLinear regression
Show Answer & Explanation

Correct Answer: B — Neural networks with many layers

Deep learning uses neural networks with multiple hidden layers (deep networks) to learn hierarchical features.

Q11.
Vanishing gradient problem occurs when:
AGradients become very large
BGradients become very small in deep networks
CNetwork is too shallow
DLearning rate is too high
Show Answer & Explanation

Correct Answer: B — Gradients become very small in deep networks

In deep networks, gradients can become extremely small during backpropagation, preventing weight updates in early layers.

Q12.
Dropout is a technique to:
AAdd more neurons
BPrevent overfitting by randomly dropping neurons
CSpeed up training
DIncrease model size
Show Answer & Explanation

Correct Answer: B — Prevent overfitting by randomly dropping neurons

Dropout randomly deactivates neurons during training, preventing overfitting by reducing co-adaptation.

Q13.
Batch normalization:
AIncreases batch size
BNormalizes layer inputs for faster training
CRemoves batches
DOnly used in testing
Show Answer & Explanation

Correct Answer: B — Normalizes layer inputs for faster training

Batch normalization normalizes layer inputs, accelerating training and reducing internal covariate shift.

Q14.
Softmax activation is used in:
AHidden layers only
BOutput layer for multi-class classification
CInput layer
DRegularization
Show Answer & Explanation

Correct Answer: B — Output layer for multi-class classification

Softmax converts raw outputs to probability distribution over multiple classes (summing to 1).

Q15.
Loss function measures:
ANumber of layers
BDifference between predicted and actual values
CTraining speed
DMemory usage
Show Answer & Explanation

Correct Answer: B — Difference between predicted and actual values

Loss function quantifies how far model predictions are from actual target values.

Q16.
Adam optimizer combines:
AOnly SGD
BMomentum and adaptive learning rates
COnly batch normalization
DOnly dropout
Show Answer & Explanation

Correct Answer: B — Momentum and adaptive learning rates

Adam combines momentum (from RMSprop) and adaptive learning rates for efficient optimization.

Q17.
Weights in neural networks are typically initialized:
ATo zero
BRandomly with small values
CTo very large values
DAll to one
Show Answer & Explanation

Correct Answer: B — Randomly with small values

Weights are randomly initialized with small values to break symmetry and enable learning.

Q18.
Hidden layers in neural networks:
AAre always visible
BAre between input and output layers
COnly output results
DOnly receive input
Show Answer & Explanation

Correct Answer: B — Are between input and output layers

Hidden layers are intermediate layers between input and output that learn internal representations.

Q19.
Cross-entropy loss is commonly used for:
ARegression tasks
BClassification tasks
CClustering
DDimensionality reduction
Show Answer & Explanation

Correct Answer: B — Classification tasks

Cross-entropy loss measures difference between predicted probability distribution and actual class labels.

Q20.
Transfer learning involves:
ATraining from scratch
BUsing pre-trained model for new task
CDeleting all weights
DOnly testing
Show Answer & Explanation

Correct Answer: B — Using pre-trained model for new task

Transfer learning uses knowledge (weights) from a model trained on one task for a related task.