[PDF] 10-601 Machine Learning Midterm Exam





Previous PDF Next PDF



Model Question Papers

28 Nov 2018 Develop an autonomous system using reinforcement learning. 5. Evaluate various machine learning algorithms and build a solution for real-world.





Question Paper Solutions - Dec. 2018/ Jan. 2019 Machine Learning

Machine Learning. 15CS73. Question Paper Solution - Dec. 2018/ Jan 2019. 7. C. consider a football game between two rival teams: Team O and Team 1. Suppose.



SAMPLE QUESTION PAPER 1 CLASS X ARTIFICIAL

Machine Learning + ______ = Artificial Intelligence. (i) TRNUALA GNLAGAUE CPSISEROGN. (ii) UTECPOMR ISOVNI. (iii) EPED RIGLENAN. (iv) URAENL KNWTEOR. 17 



Boosting Ant Colony Optimization via Solution Prediction and

In the rest of this paper we will only use SVM as our ML model. 4.3 Machine learning for combinatorial optimization: a methodological tour d'horizon.



5656-hidden-technical-debt-in-machine-learning-systems.pdf

Most of the use cases described in this paper have talked about the cost of maintaining a single model but mature systems may have dozens or hundreds of models.



QUESTION BANK MALLA REDDY COLLEGE OF ENGINEERING

Describe Train Model using Machine Learning Algorithm Test model. (10M ) e) What is the basic problem being solved by the Bridge pattern? [2] f) What are ...



18AI61 Model Question Paper-1 with effect from 2019-20 (CBCS

(a) What is Machine Leaning? Explain the applications of Machine Learning. 04M. (b) Discuss the any four main challenges of machine learning. 08M. (c). Consider 



Model Question Paper 18AI56 Fifth Semester B.E. Degree

Model Question Paper. 18AI56. Fifth Semester B.E. Degree Examination March 2022. Mathematics for Machine Learning. Time: 3 hours. Max. Marks: 100. Note: Answer 



10-601 Machine Learning Midterm Exam

18-Oct-2012 (1). Page 2 of 16. Page 4. 10-601 Machine Learning. Midterm Exam. October 18 2012. (g) [3 points] Suppose we clustered a set of N data points ...



Questions Bank

Explain the inductive biased hypothesis space and unbiased learner. 6. What are the basic design issues and approaches to machine learning? 7. How is Candidate 



Model Question Papers

11-Dec-2018 applications. Model Question Paper. Total Duration (H:M): 3:00. Course: Machine Learning. Maximum Marks: ...



DEEP LEARNING B.TECH-IT VIII SEM QUESTION BANK Question

Question - What are the applications of Machine Learning There are papers that highlight that deep NN with the right architectures achieve better.



K-means Clustering and Extensions

13-Sept-2018 Important: ALL answers have to be written on the question paper itself (dedicated space). Intro to Machine Learning (CS771A).





Using Machine Learning to Transform Supply Chain Management

machine learning (ML) companies can now This paper highlights some of the key challenges ... model for a marine shipment business



10-601 Machine Learning Midterm Exam

10-Dec-2012 The decision boundary for part a will be linear and part b will be quadratic. Page 2 of 20. Page 4. 10-601 Matchine Learning. Final Exam.



SAMPLE QUESTION PAPER 1 CLASS X ARTIFICIAL

(iv) Self-driving cars use CV and deep learning. 21. Which of the following statements is correct? (1). (a) An AI model is a computer program that 



Question Classification using Machine Learning Approaches

The paper is organized as follows. Section 2 describes about. Naïve Bayes Classifier and the algorithm which has been employed for question categorization using 



[PDF] 10-601 Machine Learning Midterm Exam

18 oct 2012 · This exam has 16 pages make sure you have all pages before you begin means solution be assigned to the same cluster in the Gaussian 





KTU S7 CSE MACHINE LEARNING SOLVED QUESTION PAPERS

29 déc 2019 · KTU S7 CSE MACHINE LEARNING SOLVED QUESTION PAPERS S7 CSE ML Solved Question papers DOWNLOAD ML Solved Questions Whatsapp this post



[PDF] EXAMPLE Machine Learning Exam questions

EXAMPLE Machine Learning (C395) Exam Questions (1) Question: Explain the Solution: Training can be posed as an optimization problem in which the goal



Machine Learning CS467 Question Papers (2015 batch) - KtuQbank

Preview and Download all the question papers of Machine Learning CS467 Question Papers (2015 batch) of branch Computer Science cs asked in the KTU exams 



[PDF] Questions Bank

Questions Bank Subject Name: Machine Learning Subject Code: 15CS73 Sem: VII Module -1 Questions 1 De4fine the following terms: a Learning



Previous year question paper for ML (BE Information Technology

Download free solved previous year question paper for Machine Learning from From our ML question paper bank students can download solved previous year 



[PDF] Deep Learning for Question Answering - UMass CICS

Briefly: deep learning + NLP basics • Factoid QA that capture the meanings of questions? Practical solution: train language model on original



Machine Learning Question With Answers Module 1 - VTUPulse

17CS73 Machine Learning Question With Answers Module 1 18CS71 Candidate elimination FInd s Algorithm Version Space Unbiased Learning VTUPulse com



[PDF] 1924103-machine-learningpdf

SRM Nagar Kattankulathur – 603 203 DEPARTMENT OF INFORMATION TECHNOLOGY QUESTION BANK I SEMESTER – M Tech - Data Science 1924103– MACHINE LEARNING

:

Midterm for CSC421/2516,

Neural Networks and Deep Learning

Winter 2019

Friday, Feb. 15, 6:10-7:40pm

Name:Student number:

This is a closed-book test. It is marked out of 15 marks. Please answer

ALL of the questions. Here is some advice:

The questions are NOT arranged in order of diculty, so you should attempt every question.

Questions that ask you to \brie

y explain" something only require short (1-3 sentence) explanations. Don't write a full page of text. We're just looking for the main idea. None of the questions require long derivations. If you nd yourself plug- ging through lots of equations, consider giving less detail or moving on to the next question.

Many questions have more than one right answer.

CSC421/2516 Winter 2019 Midterm Test

Q1:/ 1

Q2:/ 1

Q3:/ 1

Q4:/ 2

Q5:/ 1

Q6:/ 1

Q7:/ 3

Q8:/ 2

Q9:/ 3

Final mark:/ 15

2

CSC421/2516 Winter 2019 Midterm Test

1.[1pt]In our discussion of language modeling, we used the following model for the

probability of a sentence. p(w1;:::;wT) =p(w1)p(w2jw1)p(wTjw1;:::;wT1) (step 1) p(wtjw1;:::;wt1) =p(wtjwt3;wt2;wt1) (step 2) For each of the two steps, say what assumptions (if any) must be made about the distribution of sentences in order for that step to be valid. (You may assume that all the necessary conditional distributions are well-dened.)

Step 1:

No assumption or c hainrule of probabilit y.

Marking:(+0.5) for correct answer. Answers that mentioned axioms of probability also were given full marks.

Step 2:

Mark ovassumption (of order three).

Marking:(+0.5) for correct answer. Answers that explained Markov assumption in words were also given full marks.

Mean:0.70/1

2.[1pt]Consider the following binary classiciation problem from Lecture 3, which we

showed was impossible for a linear classier to solve.The training set consists of patterns A and B in all possible translations, with wrap-

around. Consider a neural network that consists of a 1D convolution layer with a linear activation function, followed by a linear layer with a logistic output. Can such an architecture perfectly classify all of the training examples? Why or why not? No. Convolution layers are linear, and any composition of linear layers is still linear.

We showed the classes are not linearly separable.

Marking:(+0.5) Correct answer and partial justication. (+0.5) Correct justica- tion. A complete answer includes mention of the whole neural network computing a linear function up to the nal non-linearity and the data being linearly inseparable.

Mean:0.62/1

3

CSC421/2516 Winter 2019 Midterm Test

3.[1pt]Recall thatautograd.numpy.dotdoes some additional work thatnumpy.dot

does not need to do. Brie y describe the additional work it is doing. You may want to refer to the inputs and outputs toautograd.numpy.dot. In addition, autograd.numpy.dot add a node to the computation graph and stores its actual input and output value during the forward computation. Marking:Full marks were given to most students for mentioning the construction of a computation graph. (-0.5) marks for being too vague and just mentioning keywords. (-1) marks o for saying something incorrect.

Mean:0.84/1

4.[2pts]Recall the following plot of the number of stochastic gradient descent (SGD)

iterations required to reach a given loss, as a function of the batch size:. (a)[1pt]For small batch sizes, the number of iterations required to reach the target loss decreases as the batch size increases. Why is that? Larger batch sizes reduces the variance in the gradient estimation of SGD. Hence, larger batch converges faster than smaller batch. Marking:Most mentions of variance or noise being decreased were sucient to get full marks for this question. (-1) for not mentioning anything regarding noise/variance or accuracy of gradient estimate given by SGD. 4

CSC421/2516 Winter 2019 Midterm Test

(b)[1pt]For large batch sizes, the number of iterations does not change much as the batch size is increased. Why is that? As the batch size grows larger, SGD eectively becomes full batch gradient de- scent. Marking:Full marks given for mentioning that large batches approximate full- batch gradient descent, so not much noise to be reduced in gradient estimation. (-1) if answer has no mention of full-batch gradient descent, (-0.5) if answer is vague.

Mean:1.00/2

5

CSC421/2516 Winter 2019 Midterm Test

5.[1pt]Suppose we are doing gradient descent on a quadratic objective:

J() =12

>A We showed that the dynamics of gradient descent with learning ratecould be ana- lyzed in terms of the spectral decompositionA=QQ>, whereQis an orthogonal matrix containing the eigenvectors ofA, and= diag(1;:::;D) is a diagonal matrix containing the eigenvalues ofAin ascending order. t=Q(I)tQ>0: Based on this formula, what is the valueCsuch that the gradient descent iterates diverge for > Cbut converge for < C? Brie y justify your answer. j1maxj<1 2 max

Marking:0:5 if1

max. 0:5 if no derivation is given in the justication.

Mean:0.40/1

6.[1pt]Consider the GloVe cost function, in terms of matricesRand~Rcontaining word

embeddingsfrig;f~rjg

J(R;~R) =X

i;jf(xij)(r>i~rjlogxij)2: (We left out the bias parameters for simplicity.) Show that this cost function is not con- vex, using a similar argument to how we showed that training a multilayer perceptron is not convex.

This is a non-convex cost function.

Solution 1: When we permute the dimensions of the embedding vectors inRand~R jointly, the cost function remains the same. To show convexity does not apply here, we can take the average of these permuted embedding vectors. The resulting embedding vectors will have all the same value for all the dimension, which will almost surely have a higher cost than the learnt embedding vectors. (Note: as an alternative to permutation symmetry, you can simply replaceRwithRand~Rwith~R.) 6

CSC421/2516 Winter 2019 Midterm Test

Solution 2: We can interchangeRand~Rdirectly and the cost function will remain the same. If we average the two embedding matrix R+~R2 , we will have the same word embedding vector for both matrices, which will have higher cost than the original cost function. It is because the highest occurance will always be the inner product of the words with itself. Marking:0:5 for noting the swap-invariance or permutation-invariance. 0:5 for ap- plying the invariance to get the average of solutions.

Mean:0.32/1

7

CSC421/2516 Winter 2019 Midterm Test

7.[3pts]Recall that the softmax function takes in a vector (z1;:::;zD) and returns a

vector (y1;:::;yD). We can express it in the following form: r=X je zjyi=ezir (a)[1pt]ConsiderD= 2, i.e. just two inputs and outputs to the softmax. Draw the computation graph relatingz1,z2,r,y1, andy2.z 1z 2y 1y 2r Marking:(+0.5) for having all nodes. (+0.5) for having all edges. (b)[1pt]Determine the backprop updates for computing thez jwhen given they i. You do not need to justify your answer. (You may give your answer either for

D= 2 or for the more general case.)r=X

iy iez ir 2z j=y jez jr +rez j Marking:(+0.5) for each equation. Common mistakes were missing the partial derivative from yjfor zjor missing the summation for r. (c)[1pt]Write a function to implement the vector-Jacobian product (VJP) for the softmax function based on your answer from part (b). For eciency, it should operate on a mini-batch. The inputs are: a matrixZof sizeNDgiving a batch of input vectors.Nis the batch size andDis the number of dimensions. Each row gives one input vector z= (z1;:::;zD). A matrixY_bargiving the output error signals. It is alsoND. 8

CSC421/2516 Winter 2019 Midterm Test

The output should be the error signalZ_bar. Do not use aforloop. def softmax_vjp(Z, Y_bar):

R = np.sum(np.exp(Z), axis = 1, keepdims=True)

R_bar = -np.sum(Y_bar * np.exp(Z), axis=1, keepdims=True)/R**2

Z_bar = Y_bar * (np.exp(Z)/R) + R_bar * np.exp(Z)

return Z_bar Marking:Full marks were given if the general proecdure is correct, without taking into account the keyword argumentsaxisandkeepdims. Otherwise (-0.5) for every mistake. Most common mistake was using the matrix multiplication operationnp.dotas either element-wise multiplication, or using it instead of np.inner.

Mean:2.23/3

9

CSC421/2516 Winter 2019 Midterm Test

8.[2pts]In this question, you will design a convolutional network to detect vertical

boundaries in an image. The architecture of the network is as shown on the right.

The ReLU activation function is applied to the

rst convolution layer. The output layer uses the linear activation function.

For this question, you may assume either the

standard denition of convolution (which ips and lters) or the version used in conv nets (which skips the ltering step). Conveniently, the same answer works either way.

In order to make the gure printable for the

exam paper, we use white to denote 0 and darker values to denote larger (more positive) values.(a)[1pt]Design two convolution kernels for the rst layer, of size 33. One of them should detect dark/light boundaries, and the other should detect light/dark boundaries. (It doesn't matter which is which.) You don't need to justify your answer.Marking:(+0.5) for a kernel that has a positive gradient in the left-right direc- tion. (+0.5) for a kernel that has a negative gradient in the left-right direction. Answers that satised the above criteria but weren't horizontally symmetric were given partial marks. 10

CSC421/2516 Winter 2019 Midterm Test

(b)[1pt]Design convolution kernels of size 33 for the output layer, which computes

the desired output. You don't need to justify your answer.Marking:(+1) for any two kernels that add the feature maps from the previous

quotesdbs_dbs8.pdfusesText_14
[PDF] machine learning tutorial pdf

[PDF] machine learning with python ppt

[PDF] macintosh

[PDF] macleay valley travel reviews

[PDF] macleay valley travel tasmania

[PDF] macos 10.15 compatibility

[PDF] macos catalina security features

[PDF] macos security guide

[PDF] macos server

[PDF] macos server mojave

[PDF] macos virtualization license

[PDF] macromolecules

[PDF] macron ce soir

[PDF] macros in 8086 microprocessor pdf

[PDF] macros in excel pdf books