**NPTEL An Introduction to Artificial Intelligence Assignment 9 Answers 2022:-** All the Answers provided here to help the students as a reference, You must submit your assignment at your own knowledge

## What is An Introduction to Artificial Intelligence?

An Introduction to Artificial Intelligence by **IIT Delhi** course introduces the variety of concepts in the field of artificial intelligence. It discusses the philosophy of AI, and how to model a new problem as an AI problem. It describes a variety of models such as search, logic, Bayes nets, and MDPs, which can be used to model a new problem. It also teaches many first algorithms to solve each formulation. The course prepares a student to take a variety of focused, advanced courses in various subfields of AI.

## CRITERIA TO GET A CERTIFICATE

Average assignment score = 25% of the average of best 8 assignments out of the total 12 assignments given in the course.

Exam score = 75% of the proctored certification exam score out of 100

Final score = Average assignment score + Exam score

**YOU WILL BE ELIGIBLE FOR A CERTIFICATE ONLY IF THE AVERAGE ASSIGNMENT SCORE >=10/25 AND EXAM SCORE >= 30/75. If one of the 2 criteria is not met, you will not get the certificate even if the Final score >= 40/100.**

An Introduction to Artificial Intelligence | Answers |

Assignment 1 | Click Here |

Assignment 2 | Click Here |

Assignment 3 | Click Here |

Assignment 4 | Click Here |

Assignment 5 | Click Here |

Assignment 6 | Click Here |

Assignment 7 | Click Here |

Assignment 8 | Click Here |

Assignment 9 | Click Here |

Assignment 10 | Click Here |

Assignment 11 | Click Here |

Assignment 12 | NA |

## NPTEL An Introduction to Artificial Intelligence Assignment 9 Answers 2022:-

**Q1.** Which of the following is true about the MAP (Maximum a posteriori estimate) estimation learning framework?

a. It is equivalent to Maximum Likelihood learning with infinite data

b. It is equivalent to Maximum Likelihood learning if P(θ) is independent of θ

c. it can be used without having any prior knowledge about the parameters

d. The performance of MAP is better with dense data compared to sparse data

**Answer:- a, d**

**Answers will be Uploaded Shortly and it will be Notified on Telegram, So JOIN NOW**

**Q2.** What facts are true about smoothing?

- Smoothed estimates of probabilities fit the evidence better than un-smoothed estimates.
- The process of smoothing can be viewed as imposing a prior distribution over the set of parameters.
- Smoothing allows us to account for data which wasn’t seen in the evidence.
- Smoothing is a form of regularization which prevents overfitting in Bayesian networks.

**Answer: a, c**

**Q3.** Consider three boolean variables X, Y, and Z. Consider the following data:

There can be multiple Bayesian networks that can be used to model such a universe. Assume that we assume a Bayesian Network as shown below:

If the value of the parameter P(¬z|x,¬y) is m/n such that m and n have no common factors. Then, what is the value of m+n? Assume add-one smoothing.

**Answer: 343.6**

**Q4.** Consider the following Bayesian Network from which we wish to compute P(x|z) using rejection sampling:

**Answer: 86.9**

**Q5.** Assume that we toss a biased coin with heads probability p, 100 times. We get heads 66 times out of 100. If the Maximum Likelihood estimate of the parameter p is m/n where m and n don’t have common factors,

then the value of m+n is?

**Answer: 77**

**👇FOR NEXT WEEK ASSIGNMENT ANSWERS👇**

**Q6.** Now, assume that we had a prior distribution over p as shown below:

**Answer:- 6.5**

**Q7.** Which of the following task(s) are not suited for a goal based agent?

**Answer: b, c**

**Q8.** Which of the following are true ?

- Rejection sampling is very wasteful when the probability of getting the evidence in the samples is very low.
- We perform conditional probability weighting on the samples while doing Gibbs Sampling in MCMC algorithm since we have already fixed the evidence variables.
- We perform random walk while sampling variables in Likelihood Weighting, MCMC with Gibbs sampling, but not in Rejection sampling.
- Likelihood Weighting functions well if we have many evidence wars with some samples having nearly all the total weight

**Answer: a**

**Q9.** Consider the following Bayesian Network:

- P(C|A,B,D,F,E) = α. P(C|A). P(C|B)
- P(C|A,B,D,F,E) = α. P(C|A,B)
- P(C|A,B,D,F,E) = α. P(C|A,B). P(D|C,E)
- P(C|A,B,D,F,E) = α. P(C|A,B,D,E)

**Answer: b, c**

**If there are any changes in answers will notify you on telegram so you can get a 100% score, So Join**

**Q10.** Which of the following options are correct about the environment of Tic Tac Toe?

- Fully observable
- Stochastic
- Continuous
- Static

**Answer: a, c**

**Disclaimer**:- We do not claim 100% surety of solutions, these solutions are based on our sole expertise, and by using posting these answers we are simply looking to help students as a reference, so we urge do your assignment on your own.

**For More NPTEL Answers:- CLICK HERE**

**Join Our Telegram:- CLICK HERE**

**NPTEL An Introduction to Artificial Intelligence Assignment 9 Answers 2022:-** All the Answers provided here to help the students as a reference, You must submit your assignment at your own knowledge