The model flags anomalies in logs and also proposes partitions for input validation tests. Which metrics BEST evaluate these two outcomes together?
You are tasked with applying structured prompting to perform impact analysis on recent code changes. Which of the following improvements would BEST align the prompt with structured prompt engineering best practices for comprehensive impact analysis?
When an organization uses an AI chatbot for testing, what is the PRIMARY LLMOps concern?
An attacker sends extremely long prompts to overflow context so the model leaks snippets from its training data. Which attack vector is this?
Which of the following is NOT a valid form of LLM-driven test data generation?
Your team needs to generate 500 API test cases for a REST API with 50 endpoints. You have documented 10 exemplar test cases that follow your organization's standard format. You want the LLM to generate test cases following the pattern demonstrated in your examples. Which of the following prompting techniques is BEST suited to achieve your goal in this scenario?
Which technique MOST directly reduces hallucinations by grounding the model in project realities?
In the context of software testing, which statements (i—v) about foundation, instruction-tuned, and reasoning LLMs are CORRECT?
i. Foundation LLMs are best suited for broad exploratory ideation when test requirements are underspecified.
ii. Instruction-tuned LLMs are strongest at adhering to fixed test case formats (e.g., Gherkin) from clear prompts.
iii. Reasoning LLMs are strongest at multi-step root-cause analysis across logs, defects, and requirements.
iv. Foundation LLMs are optimal for strict policy compliance and template conformance.
v. Instruction-tuned LLMs can follow stepwise reasoning without any additional training or prompting.