NPTEL Introduction to Machine Learning Assignment 1 Answers 2023

In this post, We have provided answers of NPTEL Introduction to Machine Learning Assignment 1. We provided answers here only for reference. Plz, do your assignment at your own knowledge.

NPTEL Introduction To Machine Learning Week 1 Assignment Answer 2023

1. Which of the following is a supervised learning problem ?

  • Grouping related documents from an unannotated corpus.
  • Predicting credit approval based on historical data.
  • Predicting if a new image has cat or dog based on the historical data of other images of cats and dogs, where you are supplied the information about which image is cat or dog.
  • Fingerprint recognition of a particular person used in biometric attendance from the fingerprint data of various other people and that particular person.

2. Which of the following are classification problems?

  • Predict the runs a cricketer will score in a particular match.
  • Predict which team will win a tournament.
  • Predict whether it will rain today.
  • Predict your mood tomorrow.

3. Which of the following is a regression task?

  • Predicting the monthly sales of a cloth store in rupees.
  • Predicting if a user would like to listen to a newly released song or not based on historical data.
  • Predicting the confirmation probability (in fraction) of your train ticket whose current status is waiting list based on historical data.
  • Predicting if a patient has diabetes or not based on historical medical records.
  • Predicting if a customer is satisfied or unsatisfied from the product purchased from ecommerce website using the the reviews he/she wrote for the purchased product.

4. Which of the following is an unsupervised learning task?

  • Group audio files based on language of the speakers.
  • Group applicants to a university based on their nationality.
  • Predict a student’s performance in the final exams.
  • Predict the trajectory of a meteorite.

5. Which of the following is a categorical feature?

  • Number of rooms in a hostel.
  • Gender of a person
  • Your weekly expenditure in rupees.
  • Ethnicity of a p e rson
  • Area (in sq. centimeter) of your laptop screen.
  • The color of the curtains in your room.
  • Number of legs an animal.
  • Minimum RAM requirement (in GB) of a system to play a game like FIFA, DOTA.

6. Which of the following is a reinforcement learning task?

  • Learning to drive a cycle
  • Learning to predict stock prices
  • Learning to play chess
  • Leaning to predict spam labels for e-mails

7. Let X and Y be a uniformly distributed random variable over the interval [0,4][0,4] and [0,6][0,6] respectively. If X and Y are independent events, then compute the probability, P(max(X,Y)>3)

  • None of the above

NPTEL Introduction to Machine Learning Assignment 1 Answers 2023

9. Which of the following statements are true? Check all that apply.

  • A model with more parameters is more prone to overfitting and typically has higher variance.
  • If a learning algorithm is suffering from high bias, only adding more training examples may not improve the test error significantly.
  • When debugging learning algorithms, it is useful to plot a learning curve to understand if there is a high bias or high variance problem.
  • If a neural network has much lower training error than test error, then adding more layers will help bring the test error down because we can fit the test set better.

10. Bias and variance are given by :

  • E[f^(x)]−f(x),E[(E[f^(x)]−f^(x)) 2 ]
  • E[f^(x)]−f(x),E[(E[f^(x)]−f^(x))] 2
  • (E[f^(x)]−f(x))2,E[(E[f^(x)]−f^(x)) 2 ]
  • (E[f^(x)]−f(x))2,E[(E[f^(x)]−f^(x))] 2

NPTEL Introduction to Machine Learning Assignment 1 Answers 2022 [July-Dec]

1. Which of the following are supervised learning problems? (multiple may be correct) a. Learning to drive using a reward signal. b. Predicting disease from blood sample. c. Grouping students in the same class based on similar features. d. Face recognition to unlock your phone.

2. Which of the following are classification problems? (multiple may be correct) a. Predict the runs a cricketer will score in a particular match. b. Predict which team will win a tournament. c. Predict whether it will rain today. d. Predict your mood tomorrow.

Answers will be Uploaded Shortly and it will be Notified on Telegram, So  JOIN NOW

NPTEL Introduction to Machine Learning Assignment 1 Answers 2023

3. Which of the following is a regression task? (multiple options may be correct) a. Predict the price of a house 10 years after it is constructed. b. Predict if a house will be standing 50 years after it is constructed. c. Predict the weight of food wasted in a restaurant during next month. d. Predict the sales of a new Apple product.

4. Which of the following is an unsupervised learning task? (multiple options may be correct) a. Group audio files based on language of the speakers. b. Group applicants to a university based on their nationality. c. Predict a student’s performance in the final exams. d. Predict the trajectory of a meteorite.

5. Given below is your dataset. You are using KNN regression with K=3. What is the prediction for a new input value (3, 2)?

6. Which of the following is a reinforcement learning task? (multiple options may be correct)

7. Find the mean of squared error for the given predictions:

8. Find the mean of 0-1 loss for the given predictions:

👇 For Week 02 Assignment Answers 👇

9. Bias and variance are given by:

10. Which of the following are true about bias and variance? (multiple options may be correct)

For More NPTEL Answers:-  CLICK HERE Join Our Telegram:-  CLICK HERE

Assignment 1
Assignment 2
Assignment 3
Assignment 4
Assignment 5
Assignment 6
Assignment 7
Assignment 8
Assignment 9
Assignment 10
Assignment 11NA
Assignment 12NA

About Introduction to Machine Learning

With the increased availability of data from varied sources there has been increasing attention paid to the various data driven disciplines such as analytics and machine learning. In this course we intend to introduce some of the basic concepts of machine learning from a mathematically well motivated perspective. We will cover the different learning paradigms and some of the more popular algorithms and architectures used in each of these paradigms. 

COURSE LAYOUT

  • Week 0:  Probability Theory, Linear Algebra, Convex Optimization – (Recap)
  • Week 1:  Introduction: Statistical Decision Theory – Regression, Classification, Bias Variance
  • Week 2:  Linear Regression, Multivariate Regression, Subset Selection, Shrinkage Methods, Principal Component Regression, Partial Least squares
  • Week 3:  Linear Classification, Logistic Regression, Linear Discriminant Analysis
  • Week 4:  Perceptron, Support Vector Machines
  • Week 5:  Neural Networks – Introduction, Early Models, Perceptron Learning, Backpropagation, Initialization, Training & Validation, Parameter Estimation – MLE, MAP, Bayesian Estimation
  • Week 6:  Decision Trees, Regression Trees, Stopping Criterion & Pruning loss functions, Categorical Attributes, Multiway Splits, Missing Values, Decision Trees – Instability Evaluation Measures
  • Week 7:  Bootstrapping & Cross Validation, Class Evaluation Measures, ROC curve, MDL, Ensemble Methods – Bagging, Committee Machines and Stacking, Boosting
  • Week 8:  Gradient Boosting, Random Forests, Multi-class Classification, Naive Bayes, Bayesian Networks
  • Week 9:  Undirected Graphical Models, HMM, Variable Elimination, Belief Propagation
  • Week 10:  Partitional Clustering, Hierarchical Clustering, Birch Algorithm, CURE Algorithm, Density-based Clustering
  • Week 11:  Gaussian Mixture Models, Expectation Maximization
  • Week 12:  Learning Theory, Introduction to Reinforcement Learning, Optional videos (RL framework, TD learning, Solution Methods, Applications)

CRITERIA TO GET A CERTIFICATE

Average assignment score = 25% of average of best 8 assignments out of the total 12 assignments given in the course. Exam score = 75% of the proctored certification exam score out of 100

Final score = Average assignment score + Exam score

YOU WILL BE ELIGIBLE FOR A CERTIFICATE ONLY IF AVERAGE ASSIGNMENT SCORE >=10/25 AND EXAM SCORE >= 30/75. If one of the 2 criteria is not met, you will not get the certificate even if the Final score >= 40/100.

NPTEL Introduction to Machine Learning Assignment 1 Answers [Jan – June 2022]

Q1. Which of the following is a supervised learning problem? 

a. Grouping related documents from an unannotated corpus.  b. Predicting credit approval based on historical data  c. Predicting rainfall based on historical data  d. Predicting if a customer is going to return or keep a particular product he/she purchased from e-commerce website based on the historical data about the customer purchases and the particular product.  e. Fingerprint recognition of a particular person used in biometric attendance from the fingerprint data of various other people and that particular person

Answer:- b, c, d , e

Q2. Which of the following is not a classification problem? 

a. Predicting the temperature (in Celsius) of a room from other environmental features (such as atmospheric pressure, humidity etc).  b.Predicting if a cricket player is a batsman or bowler given his playing records.  c. Predicting the price of house (in INR) based on the data consisting prices of other house (in INR) and its features such as area, number of rooms, location etc.  d. Filtering of spam messages  e. Predicting the weather for tomorrow as “hot”, “cold”, or “rainy” based on the historical data wind speed, humidity, temperature, and precipitation.

Answer:- a, c

Q3. Which of the following is a regression task? (multiple options may be correct) 

a. Predicting the monthly sales of a cloth store in rupees.  b. Predicting if a user would like to listen to a newly released song or not based on historical data.  c. Predicting the confirmation probability (in fraction) of your train ticket whose current status is waiting list based on historical data.  d. Predicting if a patient has diabetes or not based on historical medical records.  e. Predicting if a customer is satisfied or unsatisfied from the product purchased from e-commerce website using the the reviews he/she wrote for the purchased product.

Q4. Which of the following is an unsupervised task? 

a. Predicting if a new edible item is sweet or spicy based on the information of the ingredients, their quantities, and labels (sweet or spicy) for many other similar dishes.  b. Grouping related documents from an unannotated corpus.  c. Grouping of hand-written digits from their image.  d. Predicting the time (in days) a PhD student will take to complete his/her thesis to earn a degree based on the historical data such as qualifications, department, institute, research area, and time taken by other scholars to earn the degree.  e. all of the above

Answer:- c, d

Q5. Which of the following is a categorical feature? 

a. Number of rooms in a hostel.  b. Minimum RAM requirement (in GB) of a system to play a game like FIFA, DOTA.  c. Your weekly expenditure in rupees.  d. Ethnicity of a person  e. Area (in sq. centimeter) of your laptop screen.  f. The color of the curtains in your room.

Answer:- d, f

Q6. Let X and Y be a uniformly distributed random variable over the interval [0, 4] and [0, 6] respectively. If X and Y are independent events, then compute the probability, P(max(X,Y)>3

a. 1/6 b. 5/6 c. 2/3 d. 1/2 e. 2/6 f. 5/8 g. None of the above

NOTE:- Answers of  Introduction to Machine Learning Assignment 1 will be uploaded shortly and it will be notified on Telegram, So  JOIN NOW

Q7. Let the trace and determinant of a matrix A[acbd] be 6 and 16 respectively. The eigenvalues of A are

Q8. What happens when your model complexity increases? (multiple options may be correct) 

a. Model Bias decreases  b. Model Bias increases  c. Variance of the model decreases  d. Variance of the model increases

Answer:- a, d

Q9. A new phone, E-Corp X1 has been announced and it is what you’ve been waiting for, all along. You decide to read the reviews before buying it. From past experiences, you’ve figured out that good reviews mean that the product is good 90% of the time and bad reviews mean that it is bad 70% of the time. Upon glancing through the reviews section, you find out that the X1 has been reviewed 1269 times and only 172 of them were bad reviews. What is the probability that, if you order the X1, it is a bad phone? 

a. 0.136  b. 0.160  c. 0.360  d. 0.840  e. 0.773  f. 0.573  g. 0.181

Q10. Which of the following are false about bias and variance of overfitted and underfitted models? (multiple options may be correct) 

a. Underfitted models have high bias.  b. Underfitted models have low bias.  c. Overfitted models have low variance.  d. Overfitted models have high variance.

NPTEL Introduction to Machine Learning Assignment 1 Answers 2022:- In This article, we have provided the answers of Introduction to Machine Learning Assignment 1.

Disclaimer :- We do not claim 100% surety of solutions, these solutions are based on our sole expertise, and by using posting these answers we are simply looking to help students as a reference, so we urge do your assignment on your own.

For More NPTEL Answers:-  CLICK HERE

Join Our Telegram:-  CLICK HERE

Leave a Comment Cancel reply

You must be logged in to post a comment.

Please Enable JavaScript in your Browser to Visit this Site.

Assignments

Jump to: [Homeworks] [Projects] [Quizzes] [Exams]

There will be one homework (HW) for each topical unit of the course. Due about a week after we finish that unit.

These are intended to build your conceptual analysis skills plus your implementation skills in Python.

  • HW0 : Numerical Programming Fundamentals
  • HW1 : Regression, Cross-Validation, and Regularization
  • HW2 : Evaluating Binary Classifiers and Implementing Logistic Regression
  • HW3 : Neural Networks and Stochastic Gradient Descent
  • HW4 : Trees
  • HW5 : Kernel Methods and PCA

After completing each unit, there will be a 20 minute quiz (taken online via gradescope).

Each quiz will be designed to assess your conceptual understanding about each unit.

Probably 10 questions. Most questions will be true/false or multiple choice, with perhaps 1-3 short answer questions.

You can view the conceptual questions in each unit's in-class demos/labs and homework as good practice for the corresponding quiz.

There will be three larger "projects" throughout the semester:

  • Project A: Classifying Images with Feature Transformations
  • Project B: Classifying Sentiment from Text Reviews
  • Project C: Recommendation Systems for Movies

Projects are meant to be open-ended and encourage creativity. They are meant to be case studies of applications of the ML concepts from class to three "real world" use cases: image classification, text classification, and recommendations of movies to users.

Each project will due approximately 4 weeks after being handed out. Start early! Do not wait until the last few days.

Projects will generally be centered around a particular methodology for solving a specific task and involve significant programming (with some combination of developing core methods from scratch or using existing libraries). You will need to consider some conceptual issues, write a program to solve the task, and evaluate your program through experiments to compare the performance of different algorithms and methods.

Your main deliverable will be a short report (2-4 pages), describing your approach and providing several figures/tables to explain your results to the reader.

You’ll be assessed on effort, the sophistication of your technical approach, the clarity of your explanations, the evidence that you present to support your evaluative claims, and the performance of your implementation. A high-performing approach with little explanation will receive little credit, while a careful set of experiments that illuminate why a particular direction turned out to be a dead end may receive close to full credit.

logo

Introduction to Machine Learning

Assignment 1: classifier by hand, assignment 1: classifier by hand #.

In this assignment, you learn step by step how to code a binary classifier by hand!

Don’t worry, it will be guided through.

Introduction: calorimeter showers #

A calorimeter in the context of experimental particle physics is a sub-detector aiming at measuring the energy of incoming particles. At CERN Large Hadron Collider, the giant multipurpose detectors ATLAS and CMS are both equipped with electromagnetic and hadronic calorimeters. The electronic calorimeter, as its name indicates, is measuring the energy of incoming electrons. It is a destructive method: the energetic electron entering the calorimeter will interact with its dense material via the electromagnetic force. It eventually results in the generation of a shower of particles (electromagnetic shower), with a characteristic average depth and width. The depth is along the direction of the incoming particle and the width is the dimension perpendicular to it.

Problem? There can be noisy signals in electromagnetic calorimeters that are generated by hadrons, not electrons.

Your mission is to help physicists by coding a classifier to select electron-showers (signal) from hadron-showers (background).

To this end, you are given a dataset of shower characterists from previous measurements where the incoming particle was known. The main features are the depth and the width.

../_images/a01_showers.png

Visualization of an electron shower (left) and hadron shower (right). From ResearchGate #

Hadron showers are on average longer in the direction of the incoming hadron (depth) and large in the transverse direction (width).

../_images/a01_showers_distribs.png

1. Get the Data #

Download the dataset and put it on your GDrive. Open a new Jupyter-notebook from Google-Colab . To mount your drive:

For the following, import the NumPy and pandas libraries:

1.1 Get and load the data Read the dataset into a dataframe df . What are the columns? Which column stores the labels (targets)?

1.2 How many samples are there?

2. Feature Scaling #

2.1 Explain If the parameters are initialized randomly between 0 and 1 and the data are not zero-centered, what happens to the gradient descent? Explain the behaviour.

2.2 Standardization Create for each input feature an extra column in the dataframe to rescale it to a distribution of zero-mean and unit-variance. To see statistical information on a dataframe, a convenient method is:

We will take \(x_1\) and \(x_2\) in the order of the dataframe’s columns. By searching in the documentation for methods retrieving the mean and standard deviation, complete the following code:

Hint: recall Definition 12 and the equation to scale a feature according to the standardization method.

Check your results by calling df.describe() on the updated dataframe.

3. Data Prep #

Let’s make the dataset ready for the classifier. As seen in class, the hypothesis function in the linear assumption has the dot product \(\sum_{j=0}^n x^{(i)}_j \theta_j = x^{(i)} \theta^{\; T}\) , where by convention \(x_0 = 1\) . With 2 input features, there are two parameters to optimize for each feature and the intercept term \(\theta_0\) . To perform the dot product above, let’s add to the dataframe a column:

3.1 Adding x0 column Add a column x0 to the dataframe df filled entirely with ones.

3.2 Matrix X Create a new dataframe X that contain the x0 column and the columns of the two scaled input features.

3.3 Labels to binary The target column contains alphabetical labels. Create an extra column in your dataframe called y containing 1 if the sample is an electron shower and 0 if it is a hadron one.

Hint: you can first create an extra column full of 0, then apply a filter using the .loc property from DataFrame. It can be in the form:

Read the pandas documentation on .loc .

3.4 Vector y Extract from the dataframe this y column with the binary labels in a separate dataframe y .

4. DataFrames to Numpy #

The inputs are almost ready, yet there are still some steps. As we saw in Lecture 3 in Part Performance Metrics , we need to split the dataset in a training and a testing sets. We will use the very convenient train_test_split method from Sciki-Learn. Then, we convert the resulting dataframes to NumPy arrays. This Python library is the standard in data science. In this assignment, you will manipulate NumPy arrays and build the foundations you need to master the programming aspect of machine learning.

Copy this code into your notebook:

4.1 Shapes Show the dimensions of the four NumPy arrays using the .shape property. Comment the numbers with respect to the notations defined in class. Does it make sense?

4.2 Test size Looking at the shapes, explain what test_size represents.

5. Useful Functions #

We saw a lot of functions in the lectures: hypothesis, logistic, cost, etc. We will code these functions as python functions to make the code more modular and easier to read.

In the following you will work with two dimensional NumPy arrays. Make sure the objects you declare in the code have the correct dimension (number of rows and number of columns).

To help you, look at the examples below. This creates a 2 x 3 matrix:

This is a list:

This makes a 1D NumPy array:

This declares a 2D NumPy array with one row (aka arow vector):

5.1 Linear Sum Write a function computing the linear sum of the features \(X\) with the \(\boldsymbol{\theta}\) parameters for each input sample.

What should be the dimensions of the returned object? Make sure your function returns a 2D array with the correct shape.

5.2 Logistic Function Write a function computing the logistic function:

5.3 Hypothesis Function Using the two functions above, write the hypothesis function \(h_\theta(\boldsymbol{x^{(i)}})\) :

5.4 Partial Derivatives of Cross-Entropy Cost Function In the linear assumption where \(z(\boldsymbol{x^{(i)}}) = \sum_{j=0}^n \theta_j x^{(i)}_j\) , the partial derivatives of the cross-entropy cost function are:

Write a function that takes three column vectors (m \(\times\) 1) and computes the partial derivatives \(\frac{\partial}{\partial \theta_j} J(\theta)\) for a given feature \(j\) :

Hint: perform an array operation to store all the derivatives in a column vector. Then sum over the elements of that vector. At the end your function should return a scalar, i.e. one value.

5.5 Cross-Entropy Cost Function Write a function computing the total cost from the 2D column vectors of predictions and observations:

6. Classifier #

The core of the action.

Luckily a skeleton is provided. You will have to replace the statments # ... by proper code. It will mostly consist of calling the functions you defined in the previous section.

Test your code frequently. To do so, you can assign dummy temporary values for the variables you do not use yet, so that python knows they are defined and your code can run.

If you struggle or cannot finish, summarize in your notebook your trials and investigations.

7. Plot cost versus epochs #

Use the following macro to plot the variation of the total cost vs the iteration number:

Call this macro:

You should get something like this:

../_images/a01_cost_vs_N.png

7.1: Describe the plot; what is the fundamental difference between the two series train and test?

7.2: What would it mean if there would be a bigger gap between the test and training values of the cost?

8. Performance #

We will write our own functions to quantitatively assess the performance of the classifier.

Before counting the true and false predictions, we need… predictions! We already wrote a function h_class outputting a prediction as a continuous variable between 0 and 1 , equivalent to a probabilitiy. The function below is calling h_class and then fills a python list of binary predictions, so either 0 or 1. For the boundary, recall in the lecture that the sigmoid is symmetric around y = 0.5, so we will work with this boundary for now. Copy this to your notebook:

Call the function:

We will work with lists from now on, so flatten the observed test values:

8.1 Accuracy Write a function computing the accuracy of the classifier:

Call your function using the test set and print the result.

Call your function, still with the test set of course, and print the result.

BRAVO! You know now the math behind a binary classifier!

X. BONUS: Decision Boundaries #

This is for advanced programmers and/or your curiosity and/or if you have the time. Bonus points will be given even if you answer with math equations only and not necessarily the associated python code. Of course you if you succeed in getting the python, more bonus points for you!

Goal We want to draw on a scatter plot the lines corresponding to different decision boundaries.

Scroll down at the very end to see where we are heading to.

X.0 Scatter plot The first step is to split the signal and background into two different dataframes. Using the general dataframe df defined at the beginning:

The plotting macro:

To call it:

X.1 Useful functions Recall the logistic function:

Write a function rev_sigmoid that outputs the value \(z = f(\hat{y})\) .

Write a function scale_inputs that scales a list of raw input features, either \(x_1\) or \(x_2\) , according to the standardization procedure.

Write the function unscale_inputs that does the contrary.

X.2 Equation For a given threshold \(\hat{y}\) , write the equation of the line boundary: \(x_2 = f(\boldsymbol{\theta}, x_1, \hat{y})\) .

X.3 Coordinate points To draw a line on a plot in Matplotlib, one needs to provide the coordinates as a set of two data points.

Write a function that compute the coordinates x2_left and x2_right – associated with the values of x1_min and x1_max respectively – of a decision boundary line at a given threshold \(\hat{y}\) . (recall 0.5 is the standard one for logistic regression).

Warrior-level bonus: compute this for several thresholds, i.e. the function returns a list of line properties. Tip: it’s convenient to store the result in a dictionary. For instance you can have keys threshold , x2_left , x2_right .

X.4 Plotting the boundaries (advanced) In the scatter plot code provided, uncomment the boundary section and draw the line(s) using the Matplotlib plot function.

In the very end, this is how it would render:

../_images/a01_scatter_with_boundaries.png

Scatter plot of electron and hadron showers with decision boundary lines for various thresholds. Code will be shown while releasing solutions. #

The higher the threshold, the more the boundary line shifts downwards in the electron-dense area. Why is that the case?

  You are encouraged to work in groups of two, however submissions are individual.

If you have received help from your peers and/or have worked within a group, summarize in the header of your notebook the involved students, how they helped you and the contributions of each group member. This is to ensure fairness in the evaluation.

You can use the internet such as the official pages of relevant libraries, forum websites to debug, etc. However, using an AI such as ChatGPT would be considered cheating (and not good for you anyway to develop your programming skills).

The instructor and tutors are available throughout the week to answer your questions. Write an email with your well-articulated question(s). Put in CC your teammates if any.

Thank you and do not forget to have fun while coding!

  • Computer Science and Engineering
  • NOC:Introduction to Machine Learning(Course sponsored by Aricent) (Video) 
  • Co-ordinated by : IIT Madras
  • Available from : 2016-01-19
  • Intro Video
  • A brief introduction to machine learning
  • Supervised Learning
  • Unsupervised Learning
  • Reinforcement Learning
  • Probability Basics - 1
  • Probability Basics - 2
  • Linear Algebra - 1
  • Linear Algebra - 2
  • Statistical Decision Theory - Regression
  • Statistical Decision Theory - Classification
  • Bias-Variance
  • Linear Regression
  • Multivariate Regression
  • Subset Selection 1
  • Subset Selection 2
  • Shrinkage Methods
  • Principal Components Regression
  • Partial Least Squares
  • Linear Classification
  • Logistic Regression
  • Linear Discriminant Analysis 1
  • Linear Discriminant Analysis 2
  • Linear Discriminant Analysis 3
  • Weka Tutorial
  • Optimization
  • Perceptron Learning
  • SVM - Formulation
  • SVM - Interpretation & Analysis
  • SVMs for Linearly Non Separable Data
  • SVM Kernels
  • SVM - Hinge Loss Formulation
  • Early Models
  • Backpropogation I
  • Backpropogation II
  • Initialization, Training & Validation
  • Maximum Likelihood Estimate
  • Priors & MAP Estimate
  • Bayesian Parameter Estimation
  • Introduction
  • Regression Trees
  • Stopping Criteria & Pruning
  • Loss Functions for Classification
  • Categorical Attributes
  • Multiway Splits
  • Missing Values, Imputation & Surrogate Splits
  • Instability, Smoothness & Repeated Subtrees
  • Evaluation Measures I
  • Bootstrapping & Cross Validation
  • 2 Class Evaluation Measures
  • The ROC Curve
  • Minimum Description Length & Exploratory Analysis
  • Introduction to Hypothesis Testing
  • Basic Concepts
  • Sampling Distributions & the Z Test
  • Student\'s t-test
  • The Two Sample & Paired Sample t-tests
  • Confidence Intervals
  • Bagging, Committee Machines & Stacking
  • Gradient Boosting
  • Random Forest
  • Naive Bayes
  • Bayesian Networks
  • Undirected Graphical Models - Introduction
  • Undirected Graphical Models - Potential Functions
  • Hidden Markov Models
  • Variable Elimination
  • Belief Propagation
  • Partitional Clustering
  • Hierarchical Clustering
  • Threshold Graphs
  • The BIRCH Algorithm
  • The CURE Algorithm
  • Density Based Clustering
  • Gaussian Mixture Models
  • Expectation Maximization
  • Expectation Maximization Continued
  • Spectral Clustering
  • Learning Theory
  • Frequent Itemset Mining
  • The Apriori Property
  • Introduction to Reinforcement Learning
  • RL Framework and TD Learning
  • Solution Methods & Applications
  • Multi-class Classification
  • Watch on YouTube
  • Assignments
  • Download Videos
  • Transcripts
  • Handouts (1)
Module NameDownloadDescriptionDownload Size
Linear Regression Linear Algebra Tutorial192
Sl.No Chapter Name MP4 Download
1A brief introduction to machine learning
2Supervised Learning
3Unsupervised Learning
4Reinforcement Learning
5Probability Basics - 1
6Probability Basics - 2
7Linear Algebra - 1
8Linear Algebra - 2
9Statistical Decision Theory - Regression
10Statistical Decision Theory - Classification
11Bias-Variance
12Linear Regression
13Multivariate Regression
14Subset Selection 1
15Subset Selection 2
16Shrinkage Methods
17Principal Components Regression
18Partial Least Squares
19Linear Classification
20Logistic Regression
21Linear Discriminant Analysis 1
22Linear Discriminant Analysis 2
23Linear Discriminant Analysis 3
24Optimization
25Perceptron Learning
26SVM - Formulation
27SVM - Interpretation & Analysis
28SVMs for Linearly Non Separable Data
29SVM Kernels
30SVM - Hinge Loss Formulation
31Weka Tutorial
32Early Models
33Backpropogation I
34Backpropogation II
35Initialization, Training & Validation
36Maximum Likelihood Estimate
37Priors & MAP Estimate
38Bayesian Parameter Estimation
39Introduction
40Regression Trees
41Stopping Criteria & Pruning
42Loss Functions for Classification
43Categorical Attributes
44Multiway Splits
45Missing Values, Imputation & Surrogate Splits
46Instability, Smoothness & Repeated Subtrees
47Tutorial
48Evaluation Measures I
49Bootstrapping & Cross Validation
502 Class Evaluation Measures
51The ROC Curve
52Minimum Description Length & Exploratory Analysis
53Introduction to Hypothesis Testing
54Basic Concepts
55Sampling Distributions & the Z Test
56Student\'s t-test
57The Two Sample & Paired Sample t-tests
58Confidence Intervals
59Bagging, Committee Machines & Stacking
60Boosting
61Gradient Boosting
62Random Forest
63Naive Bayes
64Bayesian Networks
65Undirected Graphical Models - Introduction
66Undirected Graphical Models - Potential Functions
67Hidden Markov Models
68Variable Elimination
69Belief Propagation
70Partitional Clustering
71Hierarchical Clustering
72Threshold Graphs
73The BIRCH Algorithm
74The CURE Algorithm
75Density Based Clustering
76Gaussian Mixture Models
77Expectation Maximization
78Expectation Maximization Continued
79Spectral Clustering
80Learning Theory
81Frequent Itemset Mining
82The Apriori Property
83Introduction to Reinforcement Learning
84RL Framework and TD Learning
85Solution Methods & Applications
86Multi-class Classification
Sl.No Chapter Name English
1A brief introduction to machine learning
2Supervised Learning
3Unsupervised Learning
4Reinforcement Learning
5Probability Basics - 1
6Probability Basics - 2
7Linear Algebra - 1
8Linear Algebra - 2
9Statistical Decision Theory - Regression
10Statistical Decision Theory - Classification
11Bias-Variance
12Linear Regression
13Multivariate Regression
14Subset Selection 1
15Subset Selection 2
16Shrinkage Methods
17Principal Components Regression
18Partial Least Squares
19Linear Classification
20Logistic Regression
21Linear Discriminant Analysis 1
22Linear Discriminant Analysis 2
23Linear Discriminant Analysis 3
24Optimization
25Perceptron Learning
26SVM - Formulation
27SVM - Interpretation & Analysis
28SVMs for Linearly Non Separable Data
29SVM Kernels
30SVM - Hinge Loss Formulation
31Weka Tutorial
32Early Models
33Backpropogation I
34Backpropogation II
35Initialization, Training & Validation
36Maximum Likelihood Estimate
37Priors & MAP Estimate
38Bayesian Parameter Estimation
39Introduction
40Regression Trees
41Stopping Criteria & Pruning
42Loss Functions for Classification
43Categorical Attributes
44Multiway Splits
45Missing Values, Imputation & Surrogate Splits
46Instability, Smoothness & Repeated Subtrees
47Tutorial
48Evaluation Measures I
49Bootstrapping & Cross Validation
502 Class Evaluation Measures
51The ROC Curve
52Minimum Description Length & Exploratory Analysis
53Introduction to Hypothesis Testing
54Basic Concepts
55Sampling Distributions & the Z Test
56Student\'s t-test
57The Two Sample & Paired Sample t-tests
58Confidence Intervals
59Bagging, Committee Machines & Stacking
60Boosting
61Gradient Boosting
62Random Forest
63Naive Bayes
64Bayesian Networks
65Undirected Graphical Models - Introduction
66Undirected Graphical Models - Potential Functions
67Hidden Markov Models
68Variable Elimination
69Belief Propagation
70Partitional Clustering
71Hierarchical Clustering
72Threshold Graphs
73The BIRCH Algorithm
74The CURE Algorithm
75Density Based Clustering
76Gaussian Mixture Models
77Expectation Maximization
78Expectation Maximization Continued
79Spectral Clustering
80Learning Theory
81Frequent Itemset Mining
82The Apriori Property
83Introduction to Reinforcement Learning
84RL Framework and TD Learning
85Solution Methods & Applications
86Multi-class Classification
Sl.No Language Book link
1English
2BengaliNot Available
3GujaratiNot Available
4HindiNot Available
5KannadaNot Available
6MalayalamNot Available
7MarathiNot Available
8TamilNot Available
9TeluguNot Available

swayam-logo

Introduction to Machine Learning

--> --> --> --> --> --> --> --> --> --> --> --> --> --> --> --> --> -->

Note: This exam date is subjected to change based on seat availability. You can check final exam date on your hall ticket.

Page Visits

Course layout, books and references.

  • The Elements of Statistical Learning, by Trevor Hastie, Robert Tibshirani, Jerome H. Friedman (freely available online)
  • Pattern Recognition and Machine Learning, by Christopher Bishop (optional)

Instructor bio

introduction to machine learning assignment 1

Prof. Balaraman Ravindran

Course certificate.

introduction to machine learning assignment 1

DOWNLOAD APP

introduction to machine learning assignment 1

SWAYAM SUPPORT

Please choose the SWAYAM National Coordinator for support. * :

IMAGES

  1. INTRODUCTION TO MACHINE LEARNING ASSIGNMENT ASSIGNMENT 1 Machine

    introduction to machine learning assignment 1

  2. Week 1 : Assignment 1

    introduction to machine learning assignment 1

  3. Intro to Machine Learning

    introduction to machine learning assignment 1

  4. NPTEL Introduction to Machine Learning Assignment 1 Answers 2024

    introduction to machine learning assignment 1

  5. Assignment 1 CS 7641 Machine Learning

    introduction to machine learning assignment 1

  6. NPTEL INTRODUCTION TO MACHINE LEARNING ASSIGNMENT 1 Answers 2022

    introduction to machine learning assignment 1

COMMENTS

  1. Introduction To Machine Learning Week 1 Assignment 1 Solution

    Introduction To Machine Learning Week 1 Assignment 1 Solution | NPTEL | Swayam | Jul - Dec 2023 TechEd Quest 4.54K subscribers Subscribed 90 8K views 1 year ago #python #nptel #machinelearning # ...

  2. "Introduction to Machine Learning

    In this course we intend to introduce some of the basic concepts of machine learning from a mathematically well motivated perspective.

  3. Assignment 1

    NPTEL assignment introduction to machine learning prof. ravindran which of the following are supervised learning problems? (multiple may be correct) learning to

  4. NPTEL

    Introduction To Machine Learning - IITKGP ABOUT THE COURSE : This course provides a concise introduction to the fundamental concepts in machine learning and popular machine learning algorithms.

  5. applied-machine-learning-in-python/Assignment+1.ipynb at master

    Solutions to the 'Applied Machine Learning In Python' Coursera course exercises - amirkeren/applied-machine-learning-in-python

  6. PDF noc19 cs52 assignment Week 1

    noc19 cs52 assignment Week 1. swayam NPTEL » Introduction to Machine Learning (IITKGP) Announcements Unit 3 - Week 1 About the Course [email protected] Mentor Ask a Question Progress Course outline How to access the portal Week O Assignment O week 1 Lecture 01 : Introduction Lecture 02 : Different Types of Learning Lecture 03 ...

  7. NPTEL Introduction to Machine Learning Assignment 1 Answers 2023

    Answer:- a, d. NPTEL Introduction to Machine Learning Assignment 1 Answers 2022:- In This article, we have provided the answers of Introduction to Machine Learning Assignment 1. Disclaimer :- We do not claim 100% surety of solutions, these solutions are based on our sole expertise, and by using posting these answers we are simply looking to ...

  8. PDF Assignment 1

    Introduction to Machine Learning Prof. B. Ravindran Which of the following tasks can be best solved using clustering. Predicting the amount of rainfall based on various cues

  9. PDF COMP486: Machine Learning Assignment 1: Introduction to Machine Learning

    COMP486: Machine Learning Assignment 1: Introduction to Machine Learning (50 points) In this assignment we will revise some of the most fundamental and important concepts in machine learning that every student in this course should know by heart. The goal is to ensure everything is clear to you before we continue to the rest of the course.

  10. Assignment 1

    Assignment 1 - Introduction to Machine Learning. For this assignment, you will be using the Breast Cancer Wisconsin (Diagnostic) Database to create a classifier that can help diagnose patients. First, read through the description of the dataset (below). :Number of Instances: 569.

  11. Assignments

    Each quiz will be designed to assess your conceptual understanding about each unit. Probably 10 questions. Most questions will be true/false or multiple choice, with perhaps 1-3 short answer questions. You can view the conceptual questions in each unit's in-class demos/labs and homework as good practice for the corresponding quiz.

  12. Introduction To Machine Learning

    Introduction To Machine Learning - Week 1 Answers Solution 2024 | NPTEL | SWAYAMYour Queries : introduction to machine learningnptel introduction to machine...

  13. Assignment 1: Classifier By Hand

    Assignment 1: Classifier By Hand In this assignment, you learn step by step how to code a binary classifier by hand! Don't worry, it will be guided through.

  14. PDF Introduction to Machine Learning

    For this assignment, submit a hard copy of all of your answers and of your code for Question 1.

  15. Assignment 1

    Then Upload it to Assignment 1 in the LMS. This work is licensed under the Creative Commons Attribution 4.0 International license agreement. 50. Introduction to Spark Assignment 2

  16. Introduction to Machine Learning

    NPTEL provides E-learning through online Web and Video courses various streams.

  17. coursera-applied-machine-learning-with-python/Assignment+1.py at master

    # # --- # # Assignment 1 - Introduction to Machine Learning # For this assignment, you will be using the Breast Cancer Wisconsin (Diagnostic) Database to create a classifier that can help diagnose patients.

  18. NPTEL Introduction to Machine Learning

    🔊NPTEL Introduction to Machine Learning - IITKGP Week 1 Quiz Assignment Solutions | July 2022 This course provides a concise introduction to the fundamental concepts in machine learning and ...

  19. Introduction to Machine Learning

    He has nearly two decades of research experience in machine learning and specifically reinforcement learning. Currently his research interests are centered on learning from and through interactions and span the areas of data mining, social network analysis, and reinforcement learning.

  20. Assignment

    Assignment - 1 assignment introduction to machine learning prof. ravindran which of the following is supervised learning problem? grouping related documents

  21. NPTEL Introduction to Machine Learning Week 1 Quiz Assignment Solutions

    🔊NPTEL Introduction to Machine Learning Week 1 Quiz Assignment Solutions | Jan 2022 | IIT MadrasWith the increased availability of data from varied sources ...

  22. Introduction to Machine Learning

    Simple Introduction to Machine Learning. Module 1 • 7 hours to complete. The focus of this module is to introduce the concepts of machine learning with as little mathematics as possible. We will introduce basic concepts in machine learning, including logistic regression, a simple but widely employed machine learning (ML) method.

Course Status : Completed
Course Type : Elective
Duration : 12 weeks
Category :
Credit Points : 3
Undergraduate/Postgraduate
Start Date : 18 Jan 2021
End Date : 09 Apr 2021
Enrollment Ends : 01 Feb 2021
Exam Date : 25 Apr 2021 IST