Artificial-intelligence
Question 1 |
A | (a)-(i), (b)-(iv), (c)-(iii), (d)-(ii) |
B | (a)-(iv), (b)-(i), (c)-(ii), (d)-(iii) |
C | (a)-(i), (b)-(iv), (c)-(ii), (d)-(iii) |
D | (a)-(iv), (b)-(ii), (c)-(i), (d)-(iii) |
Question 1 Explanation:
Steepest – accent Hill Climbing→ Considers all moves from current state and selects best move.
Branch – and – bound → Keeps track of all partial paths which can be a candidate for further exploration
Constraint satisfaction → Discover problem state(s) that satisfy a set of constraints
Means – end – analysis → Detects difference between current state and goal state
Branch – and – bound → Keeps track of all partial paths which can be a candidate for further exploration
Constraint satisfaction → Discover problem state(s) that satisfy a set of constraints
Means – end – analysis → Detects difference between current state and goal state
Question 2 |
A | (a)-(iii), (b)-(iv), (c)-(i), (d)-(ii) |
B | (a)-(iii), (b)-(iv), (c)-(ii), (d)-(i) |
C | (a)-(iv), (b)-(iii), (c)-(i), (d)-(ii) |
D | (a)-(iv), (b)-(iii), (c)-(ii), (d)-(i) |
Question 2 Explanation:
Intelligence → Judgemental
Knowledge → Codifiable, endorsed with relevance and purpose
Information → Scattered facts, easily transferable
Data → Contextual, tacit, transfer needs learning
Knowledge → Codifiable, endorsed with relevance and purpose
Information → Scattered facts, easily transferable
Data → Contextual, tacit, transfer needs learning
Question 3 |
Consider a single perceptron with sign activation function. The perceptron is represented by weight vector [0.4 −0.3 0.1]t and a bias θ = 0. If the input vector to the perceptron is X = [0.2 0.6 0.5] then the output of the perceptron is:
A | 1 |
B | 0 |
C | -0.25 |
D | -1 |
Question 4 |
The Sigmoid activation function f(t) is defined as:
A | |
B | |
C | |
D |
Question 5 |
Entropy of a discrete random variable with possible values {x1, x2, ..., xn} and probability density function P(X) is :
The value of b gives the units of entropy. The unit for b=10 is :
The value of b gives the units of entropy. The unit for b=10 is :
A | bits |
B | bann |
C | nats |
D | deca |
Question 6 |
For any binary (n, h) linear code with minimum distance (2t+1) or greater
A | 2t+1 |
B | t+1 |
C | t |
D | t-1 |
Question 7 |
Consider a Takagi - Sugeno - Kang (TSK) Model consisting rules of the form: If Xi is Ai1 and... and xr is Air THEN y = fi(x1, x2,.... xr) = bi0 + bi1x1 + birxr assume, ai is the matching degree of rule i, then the total output of the model is given by:
A | |
B | |
C | |
D |
Question 8 |
Standard planning algorithms assume environment to be __________.
A | Both deterministic and fully observable |
B | Neither deterministic nor fully observable |
C | Deterministic but not fully observable |
D | Not deterministic but fully observable |
Question 8 Explanation:
→ Classical planning environments that are fully observable, deterministic, finite, static and discrete (in time, action, objects and effects).
Question 9 |
Consider the following:
(a) Trapping at local maxima
(b) Reaching a plateau
(c) Traversal along the ridge.
Which of the following option represents shortcomings of the hill climbing algorithm?
(a) Trapping at local maxima
(b) Reaching a plateau
(c) Traversal along the ridge.
Which of the following option represents shortcomings of the hill climbing algorithm?
A | (a) and (b) only |
B | (a) and (c) only |
C | (b) and (c) only |
D | (a), (b) and (c) |
Question 9 Explanation:
Hill climbing limitations:
1. Local Maxima: Hill-climbing algorithm reaching the vicinity a local maximum value, gets drawn towards the peak and gets stuck there, having no other place to go.
2. Ridges: These are sequences of local maxima, making it difficult for the algorithm to navigate.
3. Plateaux: This is a flat state-space region. As there is no uphill to go, algorithm often gets lost in the plateau.
To avoid above problems using 3 standard types of hill climbing algorithm is
1. Stochastic Hill Climbing selects at random from the uphill moves. The probability of selection varies with the steepness of the uphill move.
2. First-Choice Climbing implements the above one by generating successors randomly until a better one is found.
3. Random-restart hill climbing searches from randomly generated initial moves until the goal state is reached.
1. Local Maxima: Hill-climbing algorithm reaching the vicinity a local maximum value, gets drawn towards the peak and gets stuck there, having no other place to go.
2. Ridges: These are sequences of local maxima, making it difficult for the algorithm to navigate.
3. Plateaux: This is a flat state-space region. As there is no uphill to go, algorithm often gets lost in the plateau.
To avoid above problems using 3 standard types of hill climbing algorithm is
1. Stochastic Hill Climbing selects at random from the uphill moves. The probability of selection varies with the steepness of the uphill move.
2. First-Choice Climbing implements the above one by generating successors randomly until a better one is found.
3. Random-restart hill climbing searches from randomly generated initial moves until the goal state is reached.
Question 10 |
According to Dempster-Shafer theory for uncertainty management,
Where Bel(A) denotes Belief of event A.
A | |
B | |
C | |
D |
Question 11 |
Consider the following learning algorithms:
(a) Logistic regression
(b) Back propagation
(c) Linear repression Which of the following option represents classification algorithms?
A | (a) and (b) only |
B | (a) and (c) only |
C | (b) and (c) only |
D | (a), (b) and (c) |
Question 11 Explanation:
The classification learning algorithms are
1. Logistic regression
2. Back propagation
Note: They given spelling mistake in Logistic regression instead of “Logistic repression”.
According to final key, given marks to all.
1. Logistic regression
2. Back propagation
Note: They given spelling mistake in Logistic regression instead of “Logistic repression”.
According to final key, given marks to all.
Question 12 |
Let Wo represents weight between node i at layer k and node j at layer (k – 1) of a given multilayer perceptron. The weight updation using gradient descent method is given by
Where α and E represents learning rate and Error in the output respectively?
A | |
B | |
C | |
D |
Question 13 |
Given below are two statements: one is labelled as Assertion A and the other is labelled as Reason R.
Assertion A: Dendral is an expert system
Reason R: The rationality of an agent is not related to its reaction to the environment.
In the light of the above statements. choose the correct answer from the options given below.
Assertion A: Dendral is an expert system
Reason R: The rationality of an agent is not related to its reaction to the environment.
In the light of the above statements. choose the correct answer from the options given below.
A | Both A and R are true and R is the correct explanation of A
|
B | Both A and R are true, but R is NOT the correct explanation of A
|
C | A is true but R is false
|
D | A is false but R is true |
Question 14 |
Which is not a component of the natural language understanding process?
A | Morphological analysis
|
B | Semantic analysis
|
C | Pragmatic analysis
|
D | Meaning analysis |
Question 14 Explanation:
Meaning analysis is not typically considered a distinct component of the natural language understanding (NLU) process. Instead, it is often encompassed within the broader category of semantic analysis.
The key components of the NLU process include:
Morphological Analysis: This component deals with the analysis of word structure, including breaking words into meaningful units (morphemes), inflections, and word forms.
Semantic Analysis: This is the component responsible for understanding the meaning of words, phrases, and sentences. It involves determining the relationships between words and extracting the intended meaning.
Pragmatic Analysis: Pragmatics focuses on the interpretation of language in context, including factors like speech acts, implicatures, and understanding the intentions and presuppositions of the speaker.
So, "Meaning analysis" is usually encompassed within "Semantic analysis," and all the other components listed are integral parts of the natural language understanding process.
Question 15 |
Which of the following is not a property of a good system for representation of knowledge in a particular domain?
A | Presentation adequacy
|
B | Inferential adequacy
|
C | Inferential efficiency
|
D | Acquisitional efficiency |
Question 15 Explanation:
The property that is not typically considered a property of a good system for the representation of knowledge in a particular domain is "Presentation adequacy."
Presentation adequacy refers to how well the system's knowledge representation can be presented and understood by humans. While it's important to have a representation that can be comprehended by humans, the primary properties often associated with a good knowledge representation system are:
Inferential Adequacy: The system's ability to support reasoning and inference within the domain. It should be able to draw meaningful conclusions and make inferences based on the represented knowledge.
Inferential Efficiency: How efficiently the system can perform reasoning and inference. A good system should allow for efficient processing and deduction of new knowledge from the existing representation.
Acquisitional Efficiency: How efficiently the system can acquire or learn new knowledge and integrate it into the existing representation. This relates to the ease of updating and expanding the knowledge base.
While presentation adequacy is important for human understanding, it's not traditionally considered one of the core properties of a knowledge representation system. Instead, it's often viewed as an interface or display issue, focusing on how well the representation can be communicated to users.
Question 16 |
Which of the following is not a mutation operator in a genetic algorithm?
A.Random resetting
B.Scramble
C.Inversion
D.Difference
Choose the correct answer from the options given below
A.Random resetting
B.Scramble
C.Inversion
D.Difference
Choose the correct answer from the options given below
A | A and B only
|
B | B and D only
|
C | C and D only
|
D | D only |
Question 16 Explanation:
A genetic algorithm typically uses various mutation operators to introduce diversity into the population. Here's an explanation of each of the options:
A. Random Resetting: Random resetting is a mutation operator where one or more genes in an individual's chromosome are randomly changed or reset to new random values. It is a valid mutation operator in genetic algorithms.
B. Scramble: The scramble operator involves shuffling or permuting a subset of genes within a chromosome. It is a valid mutation operator in genetic algorithms.
C. Inversion: The inversion operator reverses the order of a subset of genes within a chromosome. It is a valid mutation operator in genetic algorithms.
D. Difference: "Difference" is not a standard mutation operator in genetic algorithms. While operators like "random resetting," "scramble," and "inversion" are commonly used, "difference" is not a recognized mutation operator in this context.
So, the correct answer is D only because "Difference" is not a mutation operator commonly used in genetic algorithms.
Question 17 |
Given below are two statements:
Statement I: Fuzzifier is a part of a fuzzy system
Statement Ii: Inference engine is a part of fuzzy system
In the ligt of the above statements, choose the most appropriate answer from the options given below
Statement I: Fuzzifier is a part of a fuzzy system
Statement Ii: Inference engine is a part of fuzzy system
In the ligt of the above statements, choose the most appropriate answer from the options given below
A | Both statement I and Statement II are correct
|
B | Both statement I and Statement II are incorrect
|
C | Statement I is correct but Statement II is incorrect
|
D | Statement I is incorrect but Statement II is correct |
Question 17 Explanation:
Statement I correctly identifies that a "fuzzifier" is a component of a fuzzy system. A fuzzifier is responsible for converting crisp (non-fuzzy) inputs into fuzzy sets.
Statement II is also correct because an "inference engine" is a crucial component of a fuzzy system. It's responsible for making decisions and performing reasoning based on fuzzy logic rules and inputs.
Both statements are accurate, and there is no conflict between them.
Question 18 |
Consider the following statements
A. C-fuzzy means cluster is supervised method of learning
B. PCA is used for dimension reduction
C. Apriori is not a supervised technique
D. When a machine learning model becomes so specially tuned to its exact input data that it fails to generalize to other similar data it is called underfitting
Choose the correct answer from the options given below
A. C-fuzzy means cluster is supervised method of learning
B. PCA is used for dimension reduction
C. Apriori is not a supervised technique
D. When a machine learning model becomes so specially tuned to its exact input data that it fails to generalize to other similar data it is called underfitting
Choose the correct answer from the options given below
A | A and B
|
B | B and C
|
C | C and D
|
D | D and A
|
Question 18 Explanation:
Statement B is correct. PCA (Principal Component Analysis) is indeed used for dimension reduction in machine learning and data analysis.
Statement C is also correct. Apriori is a frequent itemset mining algorithm used in association rule learning and is not a supervised technique in machine learning.
Statements A and D are not correct:
Statement A is incorrect. "C-fuzzy" is not a standard term in machine learning, and the statement doesn't accurately describe a supervised method of learning.
Statement D is also incorrect. "Underfitting" is when a model is too simple and fails to capture the underlying patterns in data. It is the opposite of overfitting, which is when a model becomes too specialized to its training data.
Question 19 |
Which of the following is not a solution representation in a genetic algorithm?
A | Binary valued
|
B | Real valued
|
C | Permutation
|
D | Combinations |
Question 19 Explanation:
"Combinations" is not typically a direct solution representation in a genetic algorithm. In genetic algorithms, the common solution representations include:
Binary Valued: Where each gene in an individual is represented as a binary value (0 or 1).
Real Valued: Where each gene in an individual is represented as a real number, often within a specific range.
Permutation: Where the genes represent a permutation or ordering of elements. This is often used for problems like the Traveling Salesman Problem.
"Combinations" as a direct representation is not commonly used in genetic algorithms. Instead, it might be implemented using other representations like binary, real-valued, or permutation depending on the specific problem being solved.
Question 20 |
A 4-input neuron has weights 1,2,3,4. The transfer function is linear with the constant of proportionality being equal to 3. The inputs are 5,7,10,30, respectively, Then the output will be,
A | 120 |
B | 213 |
C | 410 |
D | 507 |
Question 21 |
Which Artificial intelligence technique enables the computers to understand the associations and relationships between objects & Events?
A | Heuristic Processing |
B | Cognitive Science |
C | Relative Symbolism |
D | Pattern matching |
Question 22 |
What does the values of alpha-beta search get updated?
A | Along the path of search |
B | Initial state itself |
C | At the end |
D | None of these |
Question 23 |
The A* algorithm is optimal when,
A | It always finds the solution with the lowest total cost if the heuristic 'h' is admissible. |
B | Always finds the solution with the highest total cost if the heuristic 'h' is admissible. |
C | Finds the solution with the lowest total cost if the heuristic 'h' is not admissible. |
D | It always finds the solution with the highest total cost if the heuristic 'h' is not admissible. |
Question 24 |
Overfitting is expected when we observe that?
A | With training iterations error on training set as well as test set decreases |
B | With training iterations error on training set decreases but test set increases |
C | With training iterations error on training set as well as test set increases |
D | With training iterations training set as well as test error remains constant |
Question 25 |
In a database, a rule is defined as (P1 and P2) or P3? R1(0.8) and R2(0.3),where P1,P2,P3 are premises and R1,R2 are conclusions of rules with certainty factors(CF) 0.8 and 0.3 respectively. If any running program has produced P1,P2,P3 with CF as 0.5,0.8,0.2 respectively, find the CF of results on the basis of premises.
A | CF(R1=0.8),CF(R2=0.3)
|
B | CF(R1=0.40),CF(R2=0.15)
|
C | CF(R1=0.15),CF(R2=0.35)
|
D | CF(R1=0.8),CF(R2=0.35) |
Question 26 |
In a game playing search tree, up to which depth α-β pruning can be applied?
(A) Root (0) level
(B) 6 level
(C) 8 level
(D) Depends on utility value in a breadth first order
(A) Root (0) level
(B) 6 level
(C) 8 level
(D) Depends on utility value in a breadth first order
A | (B) and (C) only |
B | (A) and (B) only |
C | (A),(B) and (C) only |
D | (A) and (D) only |
Question 26 Explanation:
Alpha-beta pruning is a modified version of the minimax algorithm. It is an optimization technique for the minimax algorithm.
Alpha-beta Algorithm:
- Uses Depth first search
- only considers nodes along a single path from root at any time
α = highest-value choice found at any choice point of path for MAX (initially, α = −infinity)
β = lowest-value choice found at any choice point of path for MIN (initially, β = +infinity)
- Pass current values of α and β down to child nodes during search.
- Update values of α and β during search:
- MAX updates α at MAX nodes
- MIN updates β at MIN nodes
When to Prune:
- Prune whenever α ≥ β.
- Prune below a Max node whose alpha value becomes greater than or equal to the beta value of its ancestors.
- Max nodes update alpha based on children’s returned values. - Prune below a Min node whose beta value becomes less than or equal to the alpha value of its ancestors.
- Min nodes update beta based on children’s returned values.
Effectiveness of Alpha-Beta Search:
- Alpha/beta best case is O(b(d/2)) rather than O(bd)
- This is the same as having a branching factor of sqrt(b),
- (sqrt(b))d/ = b(d/2) (i.e., we have effectively gone from b to square root of b)
- In chess go from b ~ 35 to b ~ 6
- permitting much deeper search in the same amount of time
- In practice it is often b(2d/3)
Alpha-beta Algorithm:
- Uses Depth first search
- only considers nodes along a single path from root at any time
α = highest-value choice found at any choice point of path for MAX (initially, α = −infinity)
β = lowest-value choice found at any choice point of path for MIN (initially, β = +infinity)
- Pass current values of α and β down to child nodes during search.
- Update values of α and β during search:
- MAX updates α at MAX nodes
- MIN updates β at MIN nodes
When to Prune:
- Prune whenever α ≥ β.
- Prune below a Max node whose alpha value becomes greater than or equal to the beta value of its ancestors.
- Max nodes update alpha based on children’s returned values. - Prune below a Min node whose beta value becomes less than or equal to the alpha value of its ancestors.
- Min nodes update beta based on children’s returned values.
Effectiveness of Alpha-Beta Search:
- Alpha/beta best case is O(b(d/2)) rather than O(bd)
- This is the same as having a branching factor of sqrt(b),
- (sqrt(b))d/ = b(d/2) (i.e., we have effectively gone from b to square root of b)
- In chess go from b ~ 35 to b ~ 6
- permitting much deeper search in the same amount of time
- In practice it is often b(2d/3)
Question 27 |
Which of the following is NOT true in problem solving in artificial intelligence?
A | Implements heuristic search techniques |
B | Solution steps are not explicit |
C | Knowledge is imprecise |
D | it works on or implements repetition mechanism |
Question 28 |
Match List I with List II
List I List II
A) Branch-and-bound (I) Keeps track of all partial paths which can be can be a candidate for further exploration.
B) Steepest-ascent hill climbing (II) Detects difference between current state and goal state.
C) Constraint satisfaction (III) Discovers problem state(s) that satisfy a set of constraints.
D) Means-end-analysis (IV) Considers all moves from current state and selects the best move.
Choose the correct answer from the options given below:
A | A-I, B-IV, C-III, D-II |
B | A-I, B-II, C-III, D-IV
|
C | A-II, B-I, C-III, D-IV |
D | A-II, B-IV, C-III, D-I |
Question 28 Explanation:
Branch-and-bound→ Keep track of all partial paths which can be a candidate for further exploration.
Steepest-ascent hill climbing → Considers all moves from current state and selects the best move.
Constraint satisfaction → Discovers problem state(s) that satisfy a set of constraints.
Means-end-analysis → Detects difference between current state and goal state.
Steepest-ascent hill climbing → Considers all moves from current state and selects the best move.
Constraint satisfaction → Discovers problem state(s) that satisfy a set of constraints.
Means-end-analysis → Detects difference between current state and goal state.
Question 29 |
If f(x)=x is my friend, and p(x) = x is perfect, then the correct logical translation of the statement "some of my friends are not perfect" is _____.
A | |
B | |
C | |
D |
Question 29 Explanation:
Input:
f(x)=x is my friend
p(x) = x is perfect
So, they are asking about SOME. Finally, outer most parentheses will get SOME.
So, based on this we will eliminate 2 options.
They are given conditions like NOT perfect. So, we get ⌐p(x).
The final condition is ∃x(f(x)∧⌐p(x))
f(x)=x is my friend
p(x) = x is perfect
So, they are asking about SOME. Finally, outer most parentheses will get SOME.
So, based on this we will eliminate 2 options.
They are given conditions like NOT perfect. So, we get ⌐p(x).
The final condition is ∃x(f(x)∧⌐p(x))
Question 30 |
Match List I with List II
Choose the correct answer from the options given below
Choose the correct answer from the options given below
A | A-II, B-IV, C-I, D-III
|
B | A-II, B-III, C-I, D-IV |
C | A-III, B-II, C-IV, D-I |
D | A-III, B-IV, C-II, D-I |
Question 30 Explanation:
Greedy best-first search algorithm always selects the path which appears best at that moment. It is the combination of depth-first search and breadth-first search algorithms.
Time Complexity: The worst case time complexity of Greedy best first search is O(bm).
Space Complexity: The worst case space complexity of Greedy best first search is O(bm). Where, m is the maximum depth of the search space.
Complete: Greedy best-first search is also incomplete, even if the given state space is finite.
Optimal: Greedy best first search algorithm is not optimal.
Note:Refer the corresponding algorithms from standard sources.
Time Complexity: The worst case time complexity of Greedy best first search is O(bm).
Space Complexity: The worst case space complexity of Greedy best first search is O(bm). Where, m is the maximum depth of the search space.
Complete: Greedy best-first search is also incomplete, even if the given state space is finite.
Optimal: Greedy best first search algorithm is not optimal.
Note:Refer the corresponding algorithms from standard sources.
Question 31 |
Which of the following pairs of propositions are not logically equivalent?
A | |
B | |
C | |
D |
Question 31 Explanation:
Question 32 |
Given below are two statements:
If two variables V1and V2 are used for clustering, then consider the following statements for k means clustering with k=3:-
Statement I: If V1and V2 have correlation of 1 the cluster centroid will be in straight line.
Statement II: If V1and V2 have correlation of 0 the cluster centroid will be in straight line.
In the light of the above statements, choose the correct answer from the options given below
A | Both Statement I and Statement II are true |
B | Both Statement I and Statement II are false |
C | Statement I is correct but Statement II is false |
D | Statement I is incorrect but Statement II is true |
Question 32 Explanation:
If the correlation between the variables V1 and V2 is 1, then all the data points will be in a straight line. So, all the three cluster centroids will form a straight line as well.
Question 33 |
Which of the following statements are true?
A) A sentence ∝ entails another sentence ß if ß is true in few words where is true.
B) Forward chaining and backward chaining are very natural reasoning algorithms for knowledge bases in Horn form.
C) Sound inference algorithms derive all sentences that are entailed.
D) Propositional logic does not scale to environments of unbounded size.
Choose the correct answer from the options given below:
A | (A) and (B) only
|
B | (B) and (C) only |
C | (C) and (D) only |
D | (B) and (D) only |
Question 33 Explanation:
Statement A is false : The relationship of entailment between sentence is crucial to our understanding of reasoning. A sentence α entails another sentence β if β is true in all world where α is true. Equivalent definitions include the validity of the sentence α⇒β and the unsatisfiability of sentence α∧¬β.
Statement D is false:Propositional logic does not scale to environments of unbounded size because it lacks the expressive power to deal concisely with time, space and universal patterns of relationships among objects.
Statement B is true:
Refer the below link:
https://www.iiia.csic.es/~puyol/IAGA/Teoria/07-AgentsLogicsII.pdf Statement C is true:
Sound/truth preserving: An inference algorithm that derives only entailed sentences. Soundness is a highly desirable property. (e.g. model checking is a sound procedure when it is applicable.)
Statement D is false:Propositional logic does not scale to environments of unbounded size because it lacks the expressive power to deal concisely with time, space and universal patterns of relationships among objects.
Statement B is true:
Refer the below link:
https://www.iiia.csic.es/~puyol/IAGA/Teoria/07-AgentsLogicsII.pdf Statement C is true:
Sound/truth preserving: An inference algorithm that derives only entailed sentences. Soundness is a highly desirable property. (e.g. model checking is a sound procedure when it is applicable.)
Question 34 |
Which of the following statements are true?
A) Minimax search is breadth-first; it processes all the nodes at a level before moving to a node in the next level.
B) The effectiveness of the alpha-beta pruning is highly dependent on the order in which the states are examined.
C) The alpha-beta search algorithm computes the same optimal moves as the minimax algorithm.
D) Optimal play in games of imperfect information does not require reasoning about the current and future belief states of each player.
Choose the correct answer from the options given below:
A | (A) and (C) only |
B | (A) and (D) only |
C | (B) and (C) only |
D | (C) and (D) only |
Question 34 Explanation:
Minimax is a decision rule used in artificial intelligence, decision theory, game theory, statistics, and philosophy for minimizing the possible loss for a worst case (maximum loss) scenario.
Optimal decision in deterministic, perfect information games
Idea : choose the move resulting in the highest minimax value
Completeness: Yes if the tree is finite
Optimality: Yes, against an optimal opponent.
Time Complexity: O(bm)
Space Complexity: O(bm) – depth first exploration.
Hence Statement (A) is true.
Statement (B):
Alpha Bound of J:
→ The max current The max current val of all MAX ancestors of J of all MAX ancestors of J
→ Exploration of a min node, J, Exploration of a min node, J, is stopped when its value is stopped when its value equals or falls below alpha. equals or falls below alpha.
→ In a min node, we n node, we update beta update beta
Beta Bound of J:
→ The min current The min current val of all MIN ancestors of J of all MIN ancestors of J
→ Exploration of a Exploration of a max node, J ax node, J, is stopped when its stopped when its value equals or exceeds beta equals or exceeds beta
→ In a max node, we update a ax node, we update alpha
Pruning does not affect the final result
Does ordering affect the pruning process?
Best case O(bm/2)
Random (instead of best first search) - O(b3m/4)
Hence statement (B) is false.
Statement C: This statement is true.
Statement D: This statement is false because past exploration information is used from transposition tables.
Optimal decision in deterministic, perfect information games
Idea : choose the move resulting in the highest minimax value
Completeness: Yes if the tree is finite
Optimality: Yes, against an optimal opponent.
Time Complexity: O(bm)
Space Complexity: O(bm) – depth first exploration.
Hence Statement (A) is true.
Statement (B):
Alpha Bound of J:
→ The max current The max current val of all MAX ancestors of J of all MAX ancestors of J
→ Exploration of a min node, J, Exploration of a min node, J, is stopped when its value is stopped when its value equals or falls below alpha. equals or falls below alpha.
→ In a min node, we n node, we update beta update beta
Beta Bound of J:
→ The min current The min current val of all MIN ancestors of J of all MIN ancestors of J
→ Exploration of a Exploration of a max node, J ax node, J, is stopped when its stopped when its value equals or exceeds beta equals or exceeds beta
→ In a max node, we update a ax node, we update alpha
Pruning does not affect the final result
Does ordering affect the pruning process?
Best case O(bm/2)
Random (instead of best first search) - O(b3m/4)
Hence statement (B) is false.
Statement C: This statement is true.
Statement D: This statement is false because past exploration information is used from transposition tables.
Question 35 |
Given below are two statements:
Statement I: A genetic algorithm is a stochastic hill climbing search in which a large population of states is maintained.
Statement II: In a nondeterministic environment, agents can apply AND-OR search to generate containing plans that reach the goal regardless of which outcomes occur during execution.
In the light of the above statements, choose the correct answers from the options given below
A | Both Statement I and Statement II are true |
B | Both Statement I and Statement II are false |
C | Statement I is correct but Statement II is false
|
D | Statement I is incorrect but Statement II is true |
Question 35 Explanation:
In a genetic algorithm, a population of candidate solutions (called individuals, creatures, or phenotypes) to an optimization problem is evolved toward better solutions. Each candidate solution has a set of properties (its chromosomes or genotype) which can be mutated and altered; traditionally, solutions are represented in binary as strings of 0s and 1s, but other encodings are also possible.
In nondeterministic environments, percepts tell the agent which of the possible outcomes has actually occurred.Solutions for nondeterministic problems can contain nested if-then-else statements that create a tree rather than a sequence of actions
In nondeterministic environments, percepts tell the agent which of the possible outcomes has actually occurred.Solutions for nondeterministic problems can contain nested if-then-else statements that create a tree rather than a sequence of actions
Question 36 |
Consider the following argument with premise
A | This is a valid argument. |
B | Steps (C) and (E) are not correct inferences |
C | Steps (D) and (F) are not correct inferences |
D | Step (G) is not a correct inference |
Question 37 |
Consider the statement below.
A person who is radical (R) is electable (E) if he/she is conservative (C), but otherwise is not electable.
Few probable logical assertions of the above sentence are given below.,
Which of the above logical assertions are true?
Choose the correct answer from the options given below:
Which of the above logical assertions are true?
Choose the correct answer from the options given below:
A | (B) only |
B | (C) only |
C | (A) and (C) only |
D | (B) and (D) only |
Question 37 Explanation:
1. (R ∧E) ↔C
This is not equivalent. It says that all (and only) conservatives are radical and electable.
2. R →(E ↔C)
This one is equivalent. if a person is a radical then they are electable if and only if they are conservative.
3. R →((C →E) ∨¬E)
This one is vacuous. It’s equivalent to ¬R ∨ (¬C ∨ E ∨ ¬E), which is true in all interpretations.
4.R ⇒ (E ⇐⇒ C) ≡ R ⇒ ((E ⇒ C) ∧ (C ⇒ E))
≡ ¬R ∨ ((¬E ∨ C) ∧ (¬C ∨ E))
≡ (¬R ∨ ¬E ∨ C) ∧ (¬R ∨ ¬C ∨ E))
This is not equivalent. It says that all (and only) conservatives are radical and electable.
2. R →(E ↔C)
This one is equivalent. if a person is a radical then they are electable if and only if they are conservative.
3. R →((C →E) ∨¬E)
This one is vacuous. It’s equivalent to ¬R ∨ (¬C ∨ E ∨ ¬E), which is true in all interpretations.
4.R ⇒ (E ⇐⇒ C) ≡ R ⇒ ((E ⇒ C) ∧ (C ⇒ E))
≡ ¬R ∨ ((¬E ∨ C) ∧ (¬C ∨ E))
≡ (¬R ∨ ¬E ∨ C) ∧ (¬R ∨ ¬C ∨ E))
Question 38 |
Match the following:
A | a-i, b-ii, c-iii, d-iv |
B | a-i, b-iii, c-ii, d-iv |
C | a-iii, b-ii, c-iv, d-i |
D | a-ii, b-iii, c-i, d-iv |
Question 38 Explanation:
Affiliate Marketing: Vendors ask partners to place logos on partner’s site. If customers click, come to vendors and buy.
Viral Marketing: Spread your brand on the net by word-of-mouth. Receivers will send your information to their friends.
Group Purchasing: Aggregating the demands of small buyers to get a large volume. Then negotiate a price.
Bartering Online: Exchanging surplus products and services with the process administered completely online by an intermediary. Company receives “points” for its contribution.
Viral Marketing: Spread your brand on the net by word-of-mouth. Receivers will send your information to their friends.
Group Purchasing: Aggregating the demands of small buyers to get a large volume. Then negotiate a price.
Bartering Online: Exchanging surplus products and services with the process administered completely online by an intermediary. Company receives “points” for its contribution.
Question 39 |
Match the following :
A | a-i, b-ii, c-iii, d-iv |
B | a-i, b-iii, c-iv, d-ii |
C | a-ii, b-iii, c-iv, d-i |
D | a-ii, b-ii, c-iii, d-iv |
Question 39 Explanation:
Absurd→ Clearly impossible being contrary to some evident truth.
Ambiguous→ Capable of more than one interpretation or meaning.
Axiom→ An assertion that is accepted and used without a proof.
Conjecture→ An opinion preferably based on some experience or wisdom
Ambiguous→ Capable of more than one interpretation or meaning.
Axiom→ An assertion that is accepted and used without a proof.
Conjecture→ An opinion preferably based on some experience or wisdom
Question 40 |
Let P(m, n) be the statement “m divides n” where the Universe of discourse for both the variables is the set of positive integers. Determine the truth values of the following propositions.
(a)∃m ∀n P(m, n)
(b)∀n P(1, n)
(c) ∀m ∀n P(m, n)
(a)∃m ∀n P(m, n)
(b)∀n P(1, n)
(c) ∀m ∀n P(m, n)
A | (a) - True; (b) - True; (c) - False |
B | (a) - True; (b) - False; (c) - False |
C | (a) - False; (b) - False; (c) - False |
D | (a) - True; (b) - True; (c) - True |
Question 40 Explanation:
Given P(m,n) ="m divides n"
Statement-A is ∃m ∀n P(m, n). Here, there exists some positive integer which divides every positive integer. It is true because there is positive integer 1 which divides every positive integer.
Statement-B is ∀n P(1, n). Here, 1 divided every positive integer. It is true.
Statement-C is ∀m ∀n P(m, n). Here, every positive integer divided every positive integer. It is false.
Statement-A is ∃m ∀n P(m, n). Here, there exists some positive integer which divides every positive integer. It is true because there is positive integer 1 which divides every positive integer.
Statement-B is ∀n P(1, n). Here, 1 divided every positive integer. It is true.
Statement-C is ∀m ∀n P(m, n). Here, every positive integer divided every positive integer. It is false.
Question 41 |
An agent can improve its performance by
A | Learning
|
B | Responding
|
C | Observing
|
D | Perceiving
|
Question 41 Explanation:
→ An intelligent agent (IA) is an autonomous entity which observes through sensors and acts upon an environment using actuators (i.e. it is an agent) and directs its activity towards achieving goals (i.e. it is "rational", as defined in economics).
→ Intelligent agents may also learn or use knowledge to achieve their goals. They may be very simple or very complex. A reflex machine, such as a thermostat, is considered an example of an intelligent agent.
→ Intelligent agents may also learn or use knowledge to achieve their goals. They may be very simple or very complex. A reflex machine, such as a thermostat, is considered an example of an intelligent agent.
Question 42 |
Consider following sentences regarding A*, an informed search strategy in Artificial Intelligence (AI).
- (a) A* expands all nodes with f(n)<C*.
(b) A* expands no nodes with f(n)/C*.
(c) Pruning is integral to A*.
Here, C* is the cost of the optimal solution path.
A | Both statement (a) and statement (b) are true. |
B | Both statement (a) and statement (c) are true. |
C | Both statement (b) and statement (c) are true. |
D | All the statements (a), (b) and (c) are true.
|
Question 42 Explanation:
A* search:
→ A* combines the value of the heuristic function h(n)and the cost to reach the node ‘n’, g(n).
→ Evaluation function f(n) = g(n) + h(n) thus estimates the cost of the cheapest solution through ‘n’.
→ A* tries the node with the lowest f(n) value first.
→ This leads to both complete and optimal search algorithm, provided that h(n) satisfies certain conditions.
Optimality of A*:
→ A* expands all nodes ‘n’ for which f(n)
→ However, all nodes n for which f(n) > C* get pruned.
→ It is clear that A* search is complete.
→ A* search is also optimally efficient for any given heuristic function, because any algorithm that does not expand all nodes with f(n)
→ Despite being complete, optimal, and optimally efficient, A* search also has its weaknesses.
→ The number of nodes for which f(n)< C* for most problems is exponential in the length of the solution.
Reference:
http://www.cs.tut.fi/~elomaa/teach/AI-2011-3.pdf
→ A* combines the value of the heuristic function h(n)and the cost to reach the node ‘n’, g(n).
→ Evaluation function f(n) = g(n) + h(n) thus estimates the cost of the cheapest solution through ‘n’.
→ A* tries the node with the lowest f(n) value first.
→ This leads to both complete and optimal search algorithm, provided that h(n) satisfies certain conditions.
Optimality of A*:
→ A* expands all nodes ‘n’ for which f(n)
→ It is clear that A* search is complete.
→ A* search is also optimally efficient for any given heuristic function, because any algorithm that does not expand all nodes with f(n)
→ The number of nodes for which f(n)< C* for most problems is exponential in the length of the solution.
Reference:
http://www.cs.tut.fi/~elomaa/teach/AI-2011-3.pdf
Question 43 |
In Artificial Intelligence (AI), an environment is uncertain if it is
A | Not fully observable and not deterministic |
B | Not fully observable or not deterministic |
C | Fully observable but not deterministic
|
D | Not fully observable but deterministic
|
Question 43 Explanation:
→ Deterministic AI environments are those on which the outcome can be determined based on a specific state. In other words, deterministic environments ignore uncertainty.
→ Most real world AI environments are not deterministic. Instead, they can be classified as stochastic. Self-driving vehicles are a classic example of stochastic AI processes.
→ Most real world AI environments are not deterministic. Instead, they can be classified as stochastic. Self-driving vehicles are a classic example of stochastic AI processes.
Question 44 |
In Artificial Intelligence (AI), a simple reflex agent selects actions on the basis of
A | current percept, completely ignoring rest of the percept history. |
B | rest of the percept history, completely ignoring current percept. |
C | both current percept and complete percept history.
|
D | both current percept and just previous percept.
|
Question 44 Explanation:
→ Simple reflex agents act only on the basis of the current percept, ignoring the rest of the percept history. The agent function is based on the condition-action rule: "if condition, then action".
→ This agent function only succeeds when the environment is fully observable. Some reflex agents can also contain information on their current state which allows them to disregard conditions whose actuators are already triggered.
→ Infinite loops are often unavoidable for simple reflex agents operating in partially observable environments. Note: If the agent can randomize its actions, it may be possible to escape from infinite loops.
→ This agent function only succeeds when the environment is fully observable. Some reflex agents can also contain information on their current state which allows them to disregard conditions whose actuators are already triggered.
→ Infinite loops are often unavoidable for simple reflex agents operating in partially observable environments. Note: If the agent can randomize its actions, it may be possible to escape from infinite loops.
Question 45 |
In heuristic search algorithms in Artificial Intelligence (AI), if a collection of admissible heuristics h1.......hm is available for a problem and none of them dominates any of the others, which should we choose ?
A | h(n) = max{h1 (n), ...., hm(n)} |
B | h(n) = min{h1(n), ...., hm(n)} |
C | h(n) = avg{h1(n), ...., hm(n)} |
D | h(n) = sum{h1(n), ...., hm(n)} |
Question 45 Explanation:
Heuristic Search Strategies:
A key component of an evaluation function is a heuristic function h(n), which estimates the cost of the cheapest path from node ‘n’ to a goal node.
→ In connection of a search problem “heuristics” refers to a certain (but loose) upper or lower bound for the cost of the best solution.
→ Goal states are nevertheless identified: in a corresponding node ‘n’ it is required that h(n)=0
E.g., a certain lower bound bringing no information would be to set h(n) ≅ 0
→ Heuristic functions are the most common form in which additional knowledge is imported to the search algorithm.
Generating admissible heuristics from relaxed problems:
→ To come up with heuristic functions one can study relaxed problems from which some restrictions of the original problem have been removed.
→ The cost of an optimal solution to a relaxed problem is an admissible heuristic for the original problem (does not overestimate).
→ The optimal solution in the original problem is, by definition, also a solution in the relaxed problem.
Example:
→ Heuristic h1 for the 8-puzzle gives perfectly accurate path length for a simplified version of the puzzle, where a tile can move anywhere.
→ Similarly h2 gives an optimal solution to a relaxed 8-puzzle, where tiles can move also to occupied squares.
→ If a collection of admissible heuristics is available for a problem, and none of them dominates any of the others, we can use the composite function.
h(n) = max { h1(n), …, hm(n) }
→ The composite function dominates all of its component functions and is consistent if none of the components overestimates. Reference:
http://www.cs.tut.fi/~elomaa/teach/AI-2011-3.pdf
A key component of an evaluation function is a heuristic function h(n), which estimates the cost of the cheapest path from node ‘n’ to a goal node.
→ In connection of a search problem “heuristics” refers to a certain (but loose) upper or lower bound for the cost of the best solution.
→ Goal states are nevertheless identified: in a corresponding node ‘n’ it is required that h(n)=0
E.g., a certain lower bound bringing no information would be to set h(n) ≅ 0
→ Heuristic functions are the most common form in which additional knowledge is imported to the search algorithm.
Generating admissible heuristics from relaxed problems:
→ To come up with heuristic functions one can study relaxed problems from which some restrictions of the original problem have been removed.
→ The cost of an optimal solution to a relaxed problem is an admissible heuristic for the original problem (does not overestimate).
→ The optimal solution in the original problem is, by definition, also a solution in the relaxed problem.
Example:
→ Heuristic h1 for the 8-puzzle gives perfectly accurate path length for a simplified version of the puzzle, where a tile can move anywhere.
→ Similarly h2 gives an optimal solution to a relaxed 8-puzzle, where tiles can move also to occupied squares.
→ If a collection of admissible heuristics is available for a problem, and none of them dominates any of the others, we can use the composite function.
h(n) = max { h1(n), …, hm(n) }
→ The composite function dominates all of its component functions and is consistent if none of the components overestimates. Reference:
http://www.cs.tut.fi/~elomaa/teach/AI-2011-3.pdf
Question 46 |
In Challenge-Response authentication the claimant
A | Proves that she knows the secret without revealing it |
B | Proves that she doesn’t know the secret |
C | Reveals the secret
|
D | Gives a challenge |
Question 46 Explanation:
→ Challenge-Response authentication is a family of protocols in which one party presents a question ("challenge") and another party must provide a valid answer ("response") to be authenticated.
→ The simplest example of a challenge–response protocol is password authentication, where the challenge is asking for the password and the valid response is the correct password.
→ A more interesting challenge–response technique works as follows. Say, Bob is controlling access to some resource. Alice comes along seeking entry. Bob issues a challenge, perhaps "52w72y". Alice must respond with the one string of characters which "fits" the challenge Bob issued. The "fit" is determined by an algorithm "known" to Bob and Alice. (The correct response might be as simple as "63x83z" (each character of response one more than that of challenge), but in the real world, the "rules" would be much more complex.) Bob issues a different challenge each time, and thus knowing a previous correct response (even if it isn't "hidden" by the means of communication used between Alice and Bob) is of no use.
→ The simplest example of a challenge–response protocol is password authentication, where the challenge is asking for the password and the valid response is the correct password.
→ A more interesting challenge–response technique works as follows. Say, Bob is controlling access to some resource. Alice comes along seeking entry. Bob issues a challenge, perhaps "52w72y". Alice must respond with the one string of characters which "fits" the challenge Bob issued. The "fit" is determined by an algorithm "known" to Bob and Alice. (The correct response might be as simple as "63x83z" (each character of response one more than that of challenge), but in the real world, the "rules" would be much more complex.) Bob issues a different challenge each time, and thus knowing a previous correct response (even if it isn't "hidden" by the means of communication used between Alice and Bob) is of no use.
Question 47 |
Which of the following is an unsupervised neural network?
A | RBS |
B | Hopfield |
C | Back propagation |
D | Kohonen |
E | Incomplete Question |
Question 47 Explanation:
→ A self-organizing map (SOM) or self-organizing feature map (SOFM) is a type of artificial neural network (ANN) that is trained using unsupervised learning to produce a low-dimensional (typically two-dimensional), discretized representation of the input space of the training samples, called a map, and is therefore a method to do dimensionality reduction.
→ Kohonen map or network is self-organizing map
→ Hopfield nets serve as content-addressable ("associative") memory systems with binary threshold nodes. They are guaranteed to converge to a local minimum, but will sometimes converge to a false pattern (wrong local minimum) rather than the stored pattern (expected local minimum). Hopfield networks also provide a model for understanding human memory.
→ Backpropagation is a method used in artificial neural networks to calculate a gradient that is needed in the calculation of the weights to be used in the network
→ Kohonen map or network is self-organizing map
→ Hopfield nets serve as content-addressable ("associative") memory systems with binary threshold nodes. They are guaranteed to converge to a local minimum, but will sometimes converge to a false pattern (wrong local minimum) rather than the stored pattern (expected local minimum). Hopfield networks also provide a model for understanding human memory.
→ Backpropagation is a method used in artificial neural networks to calculate a gradient that is needed in the calculation of the weights to be used in the network
Question 48 |
Hopfield networks are a type of__
A | Gigabit network |
B | Terabyte network |
C | Artificial Neural network |
D | Wireless network |
Question 48 Explanation:
A Hopfield neural network is a type of artificial neural network invented by John Hopfield in 1982. It usually works by first learning a number of binary patterns and then returning the one
that is the most similar to a given input.
What defines a Hopfield network:
It is composed of only one layer of nodes or units each of which is connected to all the others but not itself. It is therefore a feedback network, which means that its outputs are redirected to its inputs. Every unit also acts as an input and an output of the network. Thus the number of nodes, inputs, outputs of the network are equal. Additionally, each one of the neurons in a has a binary state or activation value, usually represented as 1 or -1, which is its
particular output. The state of each node generally converges, meaning that the state of each node becomes fixed after a certain number of updates.
Question 49 |
Self Organizing maps are___
A | A type of statistical tool for data analysis |
B | A type of Artificial Swarm networks |
C | A type of particle Swarm algorithm |
D | None of the above |
Question 49 Explanation:
A self-organizing map or self-organizing feature map is a type of artificial neural network that is trained using unsupervised learning to produce a low-dimensional, discretized representation of the input space of the training samples, called a map, and is therefore a method to do dimensionality reduction.
Self-organizing maps differ from other artificial neural networks as they apply competitive learning as opposed to error-correction learning (such as backpropagation with gradient descent), and in the sense that they use a neighborhood function to preserve the topological properties of the input space.
Question 50 |
Sigmoidal feedforward artificial neural networks with one hidden layer can / are ___
A | Approximation any continuous function |
B | Approximation any disContinuous function |
C | Approximation any continuous function and its derivatives of arbitrary order. |
D | Exact modeling technique |
Question 50 Explanation:
Multilayer perceptron class of networks consists of multiple layers of computational units, usually interconnected in a feed-forward way. Each neuron in one layer has directed connections to the neurons of the subsequent layer. In many applications the units of these networks apply a sigmoid function as an activation function.
A feedforward neural network is an artificial neural network wherein connections between the nodes do not form a cycle. As such, it is different from recurrent neural networks. The feedforward neural network was the first and simplest type of artificial neural network devised. In this network, the information moves in only one direction, forward, from the input nodes, through the hidden nodes (if any) and to the output nodes. There are no cycles or loops in the network.
A feedforward neural network is an artificial neural network wherein connections between the nodes do not form a cycle. As such, it is different from recurrent neural networks. The feedforward neural network was the first and simplest type of artificial neural network devised. In this network, the information moves in only one direction, forward, from the input nodes, through the hidden nodes (if any) and to the output nodes. There are no cycles or loops in the network.
Access subject wise (1000+) question and answers by becoming as a solutions adda PRO SUBSCRIBER with Ad-Free content
Register Now
You have completed
questions
question
Your score is
Correct
Wrong
Partial-Credit
You have not finished your quiz. If you leave this page, your progress will be lost.
Correct Answer
You Selected
Not Attempted
Final Score on Quiz
Attempted Questions Correct
Attempted Questions Wrong
Questions Not Attempted
Total Questions on Quiz
Question Details
Results
Date
Score
Hint
Time allowed
minutes
seconds
Time used
Answer Choice(s) Selected
Question Text
Need more practice!
Keep trying!
Not bad!
Good work!
Perfect!