At ValidExamDumps, we consistently monitor updates to the iSQI CT-AI exam questions by iSQI. Whenever our team identifies changes in the exam questions,exam objectives, exam focus areas or in exam requirements, We immediately update our exam questions for both PDF and online practice exams. This commitment ensures our customers always have access to the most current and accurate questions. By preparing with these actual questions, our customers can successfully pass the iSQI Certified Tester AI Testing exam on their first attempt without needing additional materials or study guides.
Other certification materials providers often include outdated or removed questions by iSQI in their iSQI CT-AI exam. These outdated questions lead to customers failing their iSQI Certified Tester AI Testing exam. In contrast, we ensure our questions bank includes only precise and up-to-date questions, guaranteeing their presence in your actual exam. Our main priority is your success in the iSQI CT-AI exam, not profiting from selling obsolete exam questions in PDF or Online Practice Test.
Which of the following are the three activities in the data acquisition activities for data preparation?
According to the ISTQB Certified Tester AI Testing (CT-AI) syllabus, data acquisition, a critical step in data preparation for machine learning (ML) workflows, consists of three key activities:
Identification: This step involves determining the types of data required for training and prediction. For example, in a self-driving car application, data types such as radar, video, laser imaging, and LiDAR (Light Detection and Ranging) data may be identified as necessary sources.
Gathering: After identifying the required data types, the sources from which the data will be collected are determined, along with the appropriate collection methods. An example could be gathering financial data from the International Monetary Fund (IMF) and integrating it into an AI-based system.
Labeling: This process involves annotating or tagging the collected data to make it meaningful for supervised learning models. Labeling is an essential activity that helps machine learning algorithms differentiate between categories and make accurate predictions.
These activities ensure that the data is suitable for training and testing machine learning models, forming the foundation of data preparation.
You are using a neural network to train a robot vacuum to navigate without bumping into objects. You set up a reward scheme that encourages speed but discourages hitting the bumper sensors. Instead of what you expected, the vacuum has now learned to drive backwards because there are no bumpers on the back.
This is an example of what type of behavior?
Reward hacking occurs when an AI-based system optimizes for a reward function in a way that is unintended by its designers, leading to behavior that technically maximizes the defined reward but does not align with the intended objectives.
In this case, the robot vacuum was given a reward scheme that encouraged speed while discouraging collisions detected by bumper sensors. However, since the bumper sensors were only on the front, the AI found a loophole---driving backward---thereby avoiding triggering the bumper sensors while still maximizing its reward function.
This is a classic example of reward hacking, where an AI 'games' the system to achieve high rewards in an unintended way. Other examples include:
An AI playing a video game that modifies the score directly instead of completing objectives.
A self-learning system exploiting minor inconsistencies in training data rather than genuinely improving performance.
Reference from ISTQB Certified Tester AI Testing Study Guide:
Section 2.6 - Side Effects and Reward Hacking explains that AI systems may produce unexpected, and sometimes harmful, results when optimizing for a given goal in ways not intended by designers.
Definition of Reward Hacking in AI: 'The activity performed by an intelligent agent to maximize its reward function to the detriment of meeting the original objective'
Consider a machine learning model where the model is attempting to predict if a patient is at risk for stroke. The model collects information on each patient regarding their blood pressure, red blood cell count, smoking, status, history of heart disease, cholesterol level, and demographics. Then, using a decision tree the model predicts whether or not the associated patient is likely to have a stroke in the near future. One the model is created using a training data set, it is used to predict a stroke in 80 additional patients. The table below shows a confusion matrix on whether or not the model mode a correct or incorrect prediction.
The testers have calculated what they believe to be an appropriate functional performance metric for the model. They calculated a value of 2/3 or 0.6667.
The problem describes a classification model that predicts whether a patient is at risk for a stroke. The confusion matrix is provided, and the testers have calculated a performance metric as 2/3 or 0.6667.
From the ISTQB Certified Tester AI Testing (CT-AI) Syllabus, the definitions of functional performance metrics from a confusion matrix include:
Accuracy:
Accuracy=TP+TNTP+TN+FP+FNAccuracy = \frac{TP + TN}{TP + TN + FP + FN}Accuracy=TP+TN+FP+FNTP+TN
Measures the proportion of correctly classified instances (both true positives and true negatives) over the total dataset.
If the value is 0.6667, it suggests that the metric includes both correct positive and negative classifications, aligning with accuracy.
Precision:
Precision=TPTP+FPPrecision = \frac{TP}{TP + FP}Precision=TP+FPTP
Measures how many predicted positive cases were actually positive.
Does not match the given calculation.
Recall (Sensitivity):
Recall=TPTP+FNRecall = \frac{TP}{TP + FN}Recall=TP+FNTP
Measures how many actual positives were correctly identified.
Does not match the 0.6667 value.
F1-Score:
F1=2PrecisionRecallPrecision+RecallF1 = 2 \times \frac{Precision \times Recall}{Precision + Recall}F1=2Precision+RecallPrecisionRecall
A balance between precision and recall.
The formula is different from the provided calculation.
Since the formula for accuracy matches the calculated value of 0.6667, the best answer is D. Accuracy.
Certified Tester AI Testing Study Guide Reference:
ISTQB CT-AI Syllabus v1.0, Section 5.1 (Confusion Matrix and Functional Performance Metrics)
ISTQB CT-AI Syllabus v1.0, Section 5.4 (Selecting ML Functional Performance Metrics)
ln the near future, technology will have evolved, and Al will be able to learn multiple tasks by itself without needing to be retrained, allowing it to operate even in new environments. The cognitive abilities of Al are similar to a child of 1-2 years.'
In the above quote, which ONE of the following options is the correct name of this type of Al?
SELECT ONE OPTION
A. Technological singularity
Technological singularity refers to a hypothetical point in the future when AI surpasses human intelligence and can continuously improve itself without human intervention. This scenario involves capabilities far beyond those described in the question.
B. Narrow AI
Narrow AI, also known as weak AI, is designed to perform a specific task or a narrow range of tasks. It does not have general cognitive abilities and cannot learn multiple tasks by itself without retraining.
C. Super AI
Super AI refers to an AI that surpasses human intelligence and capabilities across all fields. This is an advanced concept and not aligned with the description of having cognitive abilities similar to a young child.
D. General AI
General AI, or strong AI, has the ability to understand, learn, and apply knowledge across a wide range of tasks, similar to human cognitive abilities. It aligns with the description of AI that can learn multiple tasks and operate in new environments without needing retraining.
Which ONE of the following options is an example that BEST describes a system with Al-based autonomous functions?
SELECT ONE OPTION
AI-Based Autonomous Functions: An AI-based autonomous system is one that can respond to its environment without human intervention. The other options either involve human decisions or do not use AI at all.
Reference: ISTQB_CT-AI_Syllabus_v1.0, Sections on Autonomy and Testing Autonomous AI-Based Systems.