Loss Functions and optimizers and its type?

With modelling, there’s a particular goal that the model needs to achieve. It’s just as important to achieve the best possible values of the model parameters as it is to find out what each parameter means in terms of that goal. The loss function (cost function) is minimized, therefore getting unknown values for weight and…

Read More

Top 40 Data Science Interview Questions and Answers

1 – What is F1 score? F1 score is a measure of the accuracy of a model. It is defined as the harmonic mean of precision and recall. F1 score is one of the most popular metrics for assessing how well a machine learning algorithm performs on predicting a target variable. F1 score ranges from…

Read More

Deep Learning interview questions Part -2

1 – What are autoencoders? Explain the different layers of autoencoders. Autoencoders are neural networks that are trained to reconstruct an input data into a desired output data. They can be thought of as the opposite of a traditional classifier, which is trained to classify inputs into pre-defined classes. Autoencoders can be seen as a…

Read More

Deep Learning interview questions Part -1

1 – What is data normalization? What’s the need for it? Data normalization is a process of transforming data from one format to another in order to improve the quality of the data and make it more usable for analysis. In this process data is organized and formatted in such a way that it’s easier…

Read More

What is Tanh activation function?

The Tanh Activation function is a scaled and shifted version of the hyperbolic tangent function, a mathematical function frequently encountered in trigonometry and calculus. The Tanh function squashes input values within the range of -1 to 1, making it a useful choice for activation functions in neural networks. Defining the Tanh Function Mathematically The mathematical…

Read More

What is PReLU and ELU activation function?

PReLU(Parametric ReLU) – PReLU is vital to the success of deep learning. It solves the problem with activation functions like sigmoid, where gradients would often vanish. This approach is finding more and more success in deep learning environments. But, we can still improve upon ReLU. Leaky ReLU was introduced, which does not zero out the…

Read More

Difference between Leaky ReLU and ReLU activation function?

What is an Activation Function? An activation function is a critical component in neural networks. It determines a neuron’s output after the neuron processes its inputs by computing a weighted sum. The activation function decides whether the neuron should be activated or not, introducing nonlinearity to the model. This nonlinearity enables the model to learn…

Read More

What is ReLU and Sigmoid activation function?

The activation function is a nonlinear function that takes in the weighted sum and produces the output. They are used to provide a more simplified model of neuron behavior which can be used as an input to deep neural networks. There are many different activation functions that can be used, including sigmoid, hyperbolic tangent, logistic,…

Read More

What is Vanishing and exploding gradient descent?

Vanishing and exploding gradient descent is a type of optimization algorithm used in deep learning. Vanishing Gradient Vanishing Gradient occurs when the gradient is smaller than expected. It causes the earlier layers to start degrading before the later ones do, causing a decrease in the overall learning rate of that subset of layers. The weights…

Read More

What is forward and backward propagation in Deep Learning?

Forward propagation is a process in which the network’s weights are updated according to the input, output and gradient of the neural network. In order to update the weights, we need to find the input and output values. The input value is found by taking the difference between the current hidden-state value and that of…

Read More