在线时间:8:00-16:00
迪恩网络APP
随时随地掌握行业动态
扫描二维码
关注迪恩网络微信公众号
开源软件名称(OpenSource Name):andrewekhalel/MLQuestions开源软件地址(OpenSource Url):https://github.com/andrewekhalel/MLQuestions开源编程语言(OpenSource Language):开源软件介绍(OpenSource Introduction):Machine Learning Interview QuestionsA collection of technical interview questions for machine learning and computer vision engineering positions. src]1) What's the trade-off between bias and variance? [If our model is too simple and has very few parameters then it may have high bias and low variance. On the other hand if our model has large number of parameters then it’s going to have high variance and low bias. So we need to find the right/good balance without overfitting and underfitting the data. [src] src]2) What is gradient descent? [Gradient descent is an optimization algorithm used to find the values of parameters (coefficients) of a function (f) that minimizes a cost function (cost). Gradient descent is best used when the parameters cannot be calculated analytically (e.g. using linear algebra) and must be searched for by an optimization algorithm. src]3) Explain over- and under-fitting and how to combat them? [src]4) How do you combat the curse of dimensionality? [
src]5) What is regularization, why do we use it, and give some examples of common methods? [A technique that discourages learning a more complex or flexible model, so as to avoid the risk of overfitting. Examples
src]6) Explain Principal Component Analysis (PCA)? [src]7) Why is ReLU better and more often used than Sigmoid in Neural Networks? [Imagine a network with random initialized weights ( or normalised ) and almost 50% of the network yields 0 activation because of the characteristic of ReLu ( output 0 for negative values of x ). This means a fewer neurons are firing ( sparse activation ) and the network is lighter. [src] receptive field of a particular node in the network. This is just finding how many input nodes actually connect through to a neuron in a CNN. [src]8) Given stride S and kernel sizes for each layer of a (1-dimensional) CNN, create a function to compute theThe receptive field are defined portion of space within an inputs that will be used during an operation to generate an output. Considering a CNN filter of size k, the receptive field of a peculiar layer is only the number of input used by the filter, in this case k, multiplied by the dimension of the input that is not being reduced by the convolutionnal filter a. This results in a receptive field of k*a. More visually, in the case of an image of size 32x32x3, with a CNN with a filter size of 5x5, the corresponding recpetive field will be the the filter size, 5 multiplied by the depth of the input volume (the RGB colors) which is the color dimensio. This thus gives us a recpetive field of dimension 5x5x3. connected components on an image/matrix. [src]9) Implementsrc]10) Implement a sparse matrix class in C++. [integral image, and create another function to get area sums from the integral image.[src]11) Create a function to compute ansrc]12) How would you remove outliers when trying to estimate a flat plane from noisy samples? [Random sample consensus (RANSAC) is an iterative method to estimate parameters of a mathematical model from a set of observed data that contains outliers, when outliers are to be accorded no influence on the values of the estimates. [src] CBIR work? [src]13) How does[Answer] Content-based image retrieval is the concept of using images to gather metadata on their content. Compared to the current image retrieval approach based on the keywords associated to the images, this technique generates its metadata from computer vision techniques to extract the relevant informations that will be used during the querying step. Many approach are possible from feature detection to retrieve keywords to the usage of CNN to extract dense features that will be associated to a known distribution of keywords. With this last approach, we care less about what is shown on the image but more about the similarity between the metadata generated by a known image and a list of known label and or tags projected into this metadata space. optical flow and so on. [src]14) How does image registration work? Sparse vs. densesrc]15) Describe how convolution works. What about if your inputs are grayscale vs RGB imagery? What determines the shape of the next layer? [src]16) Talk me through how you would create a 3D model of an object from imagery and depth sensor measurements taken at all angles around the object. [src]17) Implement SQRT(const double & x) without using any special functions, just fundamental arithmetic. [The taylor series can be used for this step by providing an approximation of sqrt(x): src]18) Reverse a bitstring. [If you are using python3 :
src]19) Implement non maximal suppression as efficiently as you can. [src]20) Reverse a linked list in place. [src]21) What is data normalization and why do we need it? [Data normalization is very important preprocessing step, used to rescale values to fit in a specific range to assure better convergence during backpropagation. In general, it boils down to subtracting the mean of each data point and dividing by its standard deviation. If we don't do this then some of the features (those with high magnitude) will be weighted more in the cost function (if a higher-magnitude feature changes by 1%, then that change is pretty big, but for smaller features it's quite insignificant). The data normalization makes all features weighted equally. src]22) Why do we use convolutions for images rather than just FC layers? [Firstly, convolutions preserve, encode, and actually use the spatial information from the image. If we used only FC layers we would have no relative spatial information. Secondly, Convolutional Neural Networks (CNNs) have a partially built-in translation in-variance, since each convolution kernel acts as it's own filter/feature detector. src]23) What makes CNNs translation invariant? [As explained above, each convolution kernel acts as it's own filter/feature detector. So let's say you're doing object detection, it doesn't matter where in the image the object is since we're going to apply the convolution in a sliding window fashion across the entire image anyways. src]24) Why do we have max-pooling in classification CNNs? [for a role in Computer Vision. Max-pooling in a CNN allows you to reduce computation since your feature maps are smaller after the pooling. You don't lose too much semantic information since you're taking the maximum activation. There's also a theory that max-pooling contributes a bit to giving CNNs more translation in-variance. Check out this great video from Andrew Ng on the benefits of max-pooling. src]25) Why do segmentation CNNs typically have an encoder-decoder style / structure? [The encoder CNN can basically be thought of as a feature extraction network, while the decoder uses that information to predict the image segments by "decoding" the features and upscaling to the original image size. src]26) What is the significance of Residual Networks? [The main thing that residual connections did was allow for direct feature access from previous layers. This makes information propagation throughout the network much easier. One very interesting paper about this shows how using local skip connections gives the network a type of ensemble multi-path structure, giving features multiple paths to propagate throughout the network. src]27) What is batch normalization and why does it work? [Training Deep Neural Networks is complicated by the fact that the distribution of each layer's inputs changes during training, as the parameters of the previous layers change. The idea is then to normalize the inputs of each layer in such a way that they have a mean output activation of zero and standard deviation of one. This is done for each individual mini-batch at each layer i.e compute the mean and variance of that mini-batch alone, then normalize. This is analogous to how the inputs to networks are standardized. How does this help? We know that normalizing the inputs to a network helps it learn. But a network is just a series of layers, where the output of one layer becomes the input to the next. That means we can think of any layer in a neural network as the first layer of a smaller subsequent network. Thought of as a series of neural networks feeding into each other, we normalize the output of one layer before applying the activation function, and then feed it into the following layer (sub-network). src]28) Why would you use many small convolutional kernels such as 3x3 rather than a few large ones? [This is very well explained in the VGGNet paper. There are 2 reasons: First, you can use several smaller kernels rather than few large ones to get the same receptive field and capture more spatial context, but with the smaller kernels you are using less parameters and computations. Secondly, because with smaller kernels you will be using more filters, you'll be able to use more activation functions and thus have a more discriminative mapping function being learned by your CNN. src]29) Why do we need a validation set and test set? What is the difference between them? [When training a model, we divide the available data into three separate sets:
So if we omit the test set and only use a validation set, the validation score won’t be a good estimate of the generalization of the model. src]30) What is stratified cross-validation and when should we use it? [Cross-validation is a technique for dividing data between training and validation sets. On typical cross-validation this split is done randomly. But in stratified cross-validation, the split preserves the ratio of the categories on both the training and validation datasets. For example, if we have a dataset with 10% of category A and 90% of category B, and we use stratified cross-validation, we will have the same proportions in training and validation. In contrast, if we use simple cross-validation, in the worst case we may find that there are no samples of category A in the validation set. Stratified cross-validation may be applied in the following scenarios:
src]31) Why do ensembles typically have higher scores than individual models? [An ensemble is the combination of multiple models to create a single prediction. The key idea for making better predictions is that the models should make different errors. That way the errors of one model will be compensated by the right guesses of the other models and thus the score of the ensemble will be higher. We need diverse models for creating an ensemble. Diversity can be achieved by:
src]32) What is an imbalanced dataset? Can you list some ways to deal with it? [An imbalanced dataset is one that has different proportions of target categories. For example, a dataset with medical images where we have to detect some illness will typically have many more negative samples than positive samples—say, 98% of images are without the illness and 2% of images are with the illness. There are different options to deal with imbalanced datasets:
src]33) Can you explain the differences between supervised, unsupervised, and reinforcement learning? [In supervised learning, we train a model to learn the relationship between input data and output data. We need to have labeled data to be able to do supervised learning. With unsupervised learning, we only have unlabeled data. The model learns a representation of the data. Unsupervised learning is frequently used to initialize the parameters of the model when we have a lot of unlabeled data and a small fraction of labeled data. We first train an unsupervised model and, after that, we use the weights of the model to train a supervised model. In reinforcement learning, the model has some input data and a reward depending on the output of the model. The model learns a policy that maximizes the reward. Reinforcement learning has been applied successfully to strategic games such as Go and even classic Atari video games. src]34) What is data augmentation? Can you give some examples? [Data augmentation is a technique for synthesizing new data by modifying existing data in such a way that the target is not changed, or it is changed in a known way. Computer vision is one of fields where data augmentation is very useful. There are many modifications that we can do to images:
src]35) What is Turing test? [The Turing test is a method to test the machine’s ability to match the human level intelligence. A machine is used to challenge the human intelligence that when it passes the test, it is considered as intelligent. Yet a machine could be viewed as intelligent without sufficiently knowing about people to mimic a human. 36) What is Precision?Precision (also called positive predictive value) is the fraction of relevant instances among the retrieved instances 37) What is Recall?Recall (also known as sensitivity) is the fraction of relevant instances that have been retrieved over the total amount of relevant instances.
Recall = true positive / (true positive + false negative) src]38) Define F1-score. [It is the weighted average of precision and recall. It considers both false positive and false negative into account. It is used to measure the model’s performance. src]39) What is cost function? [Cost function is a scalar functions which Quantifies the error factor of the Neural Network. Lower the cost function better the Neural network. Eg: MNIST Data set to classify the image, input image is digit 2 and the Neural network wrongly predicts it to be 3 src]40) List different activation neurons or functions. [
41) Define Learning Rate.Learning rate is a hyper-parameter that controls how much we are adjusting the weights of our network with respect the loss gradient. [src] 42) What is Momentum (w.r.t NN optimization)?Momentum lets the optimization algorithm remembers its last step, and adds some proportion of it to the current step. This way, even if the algorithm is stuck in a flat region, or a small local minimum, it can get out and continue towards the true minimum. [src] 43) What is the difference between Batch Gradient Descent and Stochastic Gradient Descent?Batch gradient descent computes the gradient using the whole dataset. This is great for convex, or relatively smooth error manifolds. In this case, we move somewhat directly towards an optimum solution, either local or global. Additionally, batch gradient descent, given an annealed learning rate, will eventually find the minimum located in it's basin of attraction. Stochastic gradient descent (SGD) computes the gradient using a single sample. SGD works well (Not well, I suppose, but better than batch gradient descent) for error manifolds that have lots of local maxima/minima. In this case, the somewhat noisier gradient calculated using the reduced number of samples tends to jerk the model out of local minima into a region that hopefully is more optimal. [src] 44) Epoch vs. Batch vs. Iteration.
src]45) What is vanishing gradient? [As we add more and more hidden layers, back propagation becomes less and less useful in passing information to the lower layers. In effect, as information is passed back, the gradients begin to vanish and become small relative to the weights of the networks. src]46) What are dropouts? [Dropout is a simple way to prevent a neural network from overfitting. It is the dropping out of some of the units in a neural network. It is similar to the natural reproduction process, where the nature produces offsprings by combining distinct genes (dropping out others) rather than strengthening the co-adapting of them. src]47) Define LSTM. [Long Short Term Memory – are explicitly designed to address the long term dependency problem, by maintaining a state what to remember and what to forget. src]48) List the key components of LSTM. [
|
2023-10-27
2022-08-15
2022-08-17
2022-09-23
2022-08-13
请发表评论