Free CertNexus AIP-210 Exam Actual Questions

The questions for AIP-210 were last updated On Dec 20, 2024

Question No. 1

Which of the following is a type 1 error in statistical hypothesis testing?

Show Answer Hide Answer
Correct Answer: D

A type 1 error in statistical hypothesis testing is when the null hypothesis is true, but is rejected. This means that the test falsely concludes that there is a significant difference or effect when there is none. The probability of making a type 1 error is denoted by alpha, which is also known as the significance level of the test. A type 1 error can be reduced by choosing a smaller alpha value, but this may increase the chance of making a type 2 error, which is when the null hypothesis is false but fails to be rejected. Reference: [Type I and type II errors - Wikipedia], [Type I Error and Type II Error - Statistics How To]


Question No. 2

Which of the following tests should be performed at the production level before deploying a newly retrained model?

Show Answer Hide Answer
Correct Answer: B

Performance testing is a type of testing that should be performed at the production level before deploying a newly retrained model. Performance testing measures how well the model meets the non-functional requirements, such as speed, scalability, reliability, availability, and resource consumption. Performance testing can help identify any bottlenecks or issues that may affect the user experience or satisfaction with the model. Reference: [Performance Testing Tutorial: What is, Types, Metrics & Example], [Performance Testing for Machine Learning Systems | by David Talby | Towards Data Science]


Question No. 3

Which of the following approaches is best if a limited portion of your training data is labeled?

Show Answer Hide Answer
Correct Answer: D

Semi-supervised learning is an approach that is best if a limited portion of your training data is labeled. Semi-supervised learning is a type of machine learning that uses both labeled and unlabeled data to train a model. Semi-supervised learning can leverage the large amount of unlabeled data that is easier and cheaper to obtain and use it to improve the model's performance. Semi-supervised learning can use various techniques, such as self-training, co-training, or generative models, to incorporate unlabeled data into the learning process.


Question No. 4

Which two of the following decrease technical debt in ML systems? (Select two.)

Show Answer Hide Answer
Correct Answer: C, E

Technical debt is a metaphor that describes the implied cost of additional work or rework caused by choosing an easy or quick solution over a better but more complex solution. Technical debt can accumulate in ML systems due to various factors, such as changing requirements, outdated code, poor documentation, or lack of testing. Some of the ways to decrease technical debt in ML systems are:

Documentation readability: Documentation readability refers to how easy it is to understand and use the documentation of an ML system. Documentation readability can help reduce technical debt by providing clear and consistent information about the system's design, functionality, performance, and maintenance. Documentation readability can also facilitate communication and collaboration among different stakeholders, such as developers, testers, users, and managers.

Refactoring: Refactoring is the process of improving the structure and quality of code without changing its functionality. Refactoring can help reduce technical debt by eliminating code smells, such as duplication, complexity, or inconsistency. Refactoring can also enhance the readability, maintainability, and extensibility of code.


Question No. 5

Which of the following is the correct definition of the quality criteria that describes completeness?

Show Answer Hide Answer
Correct Answer: A

Completeness is a quality criterion that describes the degree to which all required measures are known. Completeness can help assess the coverage and availability of data for a given purpose or analysis. Completeness can be measured by comparing the actual number of measures with the expected number of measures, or by identifying and counting any missing, null, or unknown values in the data.