100% found this document useful (1 vote)
256 views3 pages

Big Data Computing - Assignment 7

1) The document is a summary of an assignment submitted for an NPTEL online course on Big Data Computing. 2) The assignment contains 8 multiple choice questions related to topics like bagging, random forests, gradient boosting, and decision trees. 3) The student's responses to the 8 questions are recorded, with the correct answers highlighted in blue.

Uploaded by

VarshaMega
Copyright
© © All Rights Reserved
We take content rights seriously. If you suspect this is your content, claim it here.
Available Formats
Download as PDF, TXT or read online on Scribd
100% found this document useful (1 vote)
256 views3 pages

Big Data Computing - Assignment 7

1) The document is a summary of an assignment submitted for an NPTEL online course on Big Data Computing. 2) The assignment contains 8 multiple choice questions related to topics like bagging, random forests, gradient boosting, and decision trees. 3) The student's responses to the 8 questions are recorded, with the correct answers highlighted in blue.

Uploaded by

VarshaMega
Copyright
© © All Rights Reserved
We take content rights seriously. If you suspect this is your content, claim it here.
Available Formats
Download as PDF, TXT or read online on Scribd

10/8/21, 9:33 AM Big Data Computing - - Unit 9 - Week-7

Assessment submitted.

([Link]      

([Link]
X

remeshbabu@[Link] 

NPTEL ([Link]
»
Big Data Computing (course)

Course
Thank you for taking the Week -
outline 7:Assignment-7.
How does an
NPTEL online
course work?
Week - 7:Assignment-7
Week-0 Your last recorded submission was on 2021-10-08, 09:33 Due date: 2021-10-13, 23:59 IST.
IST
Week-1
1) Suppose you are using a bagging based algorithm say a Random Forest in model 1 point
building. Which of the following can be true?

Week-2

1. Number of tree should be as large as possible

Week-3
2. You will have interpretability after using Random Forest

Week-4
Only 1

Only 2
Week-5

Both 1 and 2

Week-6
None of the mentioned

2) To apply bagging to regression trees which of the following is/are true in such case 1 point
Week-7 ?

Decision Trees
1. We build the N regression with N bootstrap sample

for Big Data


Analytics
2. We take the average the of N regression tree

(unit? 3. Each tree has a high variance with low bias


unit=67&lesson=68)

1 and 2
Big Data

2 and 3
Predictive
Analytics
1 and 3
(Part-I) (unit?
1, 2 and 3
unit=67&lesson=69)
3) In which of the following scenario a gain ratio is preferred over Information Gain ? 1 point
Big Data
Predictive
When a categorical variable has very small number of category
Analytics
(Part-II) (unit?

Number of categories is the not the reason
unit=67&lesson=70)
When a categorical variable has very large number of category

[Link] 1/3
10/8/21, 9:33 AM Big Data Computing - - Unit 9 - Week-7

Week-7:
None of the mentioned
Assessment submitted.
Lecture
X material (unit? 4) Which of the following is/are true about Random Forest and Gradient Boosting 1 point
unit=67&lesson=71) ensemble methods ?

Feedback for
Week 7 (unit? 1. Both methods can be used for classification task

unit=67&lesson=72) 2. Random Forest is use for classification whereas Gradient Boosting is use for regression
task

Quiz: Week -
3. Random Forest is use for regression whereas Gradient Boosting is use for Classification
7:Assignment-
task

7
(assessment?
4. Both methods can be used for regression task
name=98)

1 and 2
Text Transcripts
2 and 3

2 and 4
Books

1 and 4
  5) Given an attribute table shown below, which stores the basic information of attribute 1 point
a, including the row identifier of instance row_id , values of attribute values (a) and class labels of
instances  c.

   

Which of the following attribute will first provide the pure subset ?


Humidity

[Link] 2/3
10/8/21, 9:33 AM Big Data Computing - - Unit 9 - Week-7


Wind
Assessment submitted.

Outlook
X

None of the mentioned

6) True or False ?
1 point

Bagging provides an averaging over a set of possible datasets, removing noisy and non-stable
parts of models.


True

False

7) Hundreds of trees can be aggregated to form a Random forest model. Which of the 1 point
following is true about any individual tree in Random Forest?

1. Individual tree is built on a subset of the features

2. Individual tree is built on all the features

3. Individual tree is built on a subset of observations

4. Individual tree is built on full set of observations


1 and 3

1 and 4

2 and 3

2 and 4

8) Boosting any algorithm takes into consideration the weak learners. Which of the 1 point
following is the main reason behind using weak learners ?

Reason I-To prevent overfitting

Reason II- To prevent underfitting


Reason I

Reason II

Both the Reasons

None of the Reasons

You may submit any number of times before the due date. The final submission will be
considered for grading.
Submit Answers

[Link] 3/3

You might also like