arXiv:2105.12205v1 [cs.AI] 25 May 2021A New Score for Adaptive Tests in Bayesian and Credal Networks Alessandro Antonucci, Francesca Mangili, Claudio Bonesana, and Giorgia Adorni Istituto Dalle Molle di Studi sull’Intelligenza Artificial e, Lugano, Switzerland {alessandro,francesca,claudio.bonesana,giorgia.adorn i}@idsia.ch Abstract. Atest is adaptive when its sequence andnumber of questions is dynamically tuned on the basis of the estimated skills of t he taker. Graphical models, such as Bayesian networks, are used for ad aptive tests as they allow to model the uncertainty about the questions an d the skills in an explainable fashion, especially when coping with mult iple skills. A better elicitation of the uncertainty in the question/ski lls relations can be achieved by interval probabilities. This turns the mo del into a credalnetwork, thus making more challenging the inferential comp lexity of the queries required to select questions. This is especia lly the case for the information theoretic quantities used as scoresto drive the adaptive mechanism. We present an alternative family of scores, base d on the mode of the posterior probabilities, and hence easier to exp lain. This makes considerably simpler the evaluation in the credal cas e, without significantly affectingthe qualityofthe adaptive process. Numerical tests on synthetic and real-world data are used to support this cla im. Keywords: computer adaptive tests ·information theory ·credal net- works ·Bayesian networks ·index of qualitative variation 1 Introduction A test or an exam can be naturally intended as a measurement proce ss, with the questions acting as sensors measuring the skills of the test taker in a particular discipline. Such measurement is typically imperfect with the skills modelle d as latent variables whose actual values cannot be revealed in a perfec tly reliable way. The role of the questions, whose answers are regarded inste ad as mani- fest variables, is to reduce the uncertainty about the latent skills. Following this perspective, probabilistic models are an obvious framework to desc ribe tests. Consider for instance the example in Figure 1, where a Bayesian netw ork evalu- ates the probability that the test taker knows how to multiply intege rs. In such framework making the test adaptive, i.e., picking a next question on the basis of the currentknowledgelevelofthe testtakerisalsoverynatural. The information gain for the available questions might be used to select the question le ading to the more informative results (e.g., according to Table 1, Q1is more informative thanQ2no matter what the answer is). This might also be done before the answer on the basis of expectations over the possible alternatives .2 Antonucci et al. A critical point when coping with such approaches is to provide a realis tic assessment for the probabilistic parameters associated with the m odelling of the relationsbetweenthe questionsand the skills.Havingto providesha rpnumerical values for these probabilities might be difficult. As the skill is a latent qu antity, complete data are not available for a statistical learning and a direct elicitation should be typically demanded to experts (e.g., a teacher). Yet, it mig ht be not obvious to express such a domain knowledge by single numbers and a m ore robust elicitation, such as a probability interval (e.g., P(Q1= 1|S1= 1)∈ [0.85,0.95]), might add realism and robustness to the modelling process [13]. With such generalized assessments of the parameters a Bayesian n etwork simply becomes a credalnetwork [20]. The counterpart of such increased realism is the higher computational complexity characterizing inference in cr edal networks [19]. This is an issue especially when coping with information theoretic me asures such an information gain, whose computation in credal networks mig ht lead to complex non-linear optimization tasks [17]. The goal of this paper is to investigate the potential of alternative s to the information-theoreticscoresdrivingthequestionselectioninadap tivetestsbased on directed graphical models, no matter whether these are Bayes ian or credal networks. In particular, we consider a family of scores based on th e (expected) mode of the posterior distributions over the skills. We show that, wh en coping withcredalnetworks,thecomputationofthesescorescanbere ducedtoasimpler sequenceoflinearprogrammingtask.Moreover,weshowthatthe sescoresbenefit of better explainability properties, thus allowing for a more transpa rent process in the question selection. P(Q1= 1|S= 1) = 0 .9 P(Q1= 1|S= 0) = 0 .3 P(Q2= 1|S= 1) = 0 .6 P(Q2= 1|S= 0) = 0 .4Knows multiplication ( S) 10×5? (Q1) 13×14?(Q2) Fig.1.A Bayesian network over Boolean variables modelling a simpl e test to evaluate integer multiplication skill with two questions. The paper is organized as follows. A critical discussion about the exis ting work in this area is in Section 2. The necessary background material is reviewed in Section 3. The adaptive testing concepts are introduced in Sectio n 4 and specialized to graphical models in 5. The technical part of the paper is in Section 6, where the new scores are discussed and specialized to the creda l case, while the experiments are in Section 7. Conclusions and outlooks are in Sec tion 8.A New Score for Adaptive Tests in Bayesian and Credal Network s 3 Table 1. Posterior probabilities of the skill after one or two questi ons in the test based on the Bayesian network in Figure 1. A uniform prior over the s kill is considered. Probabilities are regarded as grades and sorted from the low est one. Bounds obtained with a perturbation ǫ=±0.05 of all the input parameters are also reported. Q1Q2P(S= 1|q1,q2)P(S= 1|q1,q2)P(S= 1|q1,q2) 0 0 0 .087 0 .028 0 .187 0− 0.125 0 .052 0 .220 0 1 0 .176 0 .092 0 .256 −0 0 .400 0 .306 0 .506 −1 0 .600 0 .599 0 .603 1 0 0 .667 0 .626 0 .708 1− 0.750 0 .748 0 .757 1 1 0 .818 0 .784 0 .852 2 Related Work Modelling atest asa processrelatinglatent and manifest variablessin ce the clas- sicalitem response theory (IRT), that has been widely used even to implement adaptive sequences [12]. Despite its success related to the easeof implementation and inference, IRT might be inadequate when coping with multiple laten t skills, especially when these are dependent. This moved researcherstow ards the area of probabilistic graphical models [15], as practical tools to implement IR T in more complex setups [2]. Eventually, Bayesian networks have been even tually iden- tified as a suitable formalism to models tests, even behind the IRT fra mework [23], this being especially the case for adaptive models [24] and coach ed solving [10]. In order to cope with latent skills, some authors successfully ad opted EM approaches to these models [21], this also involving the extreme situa tion of no ground truth information about the answers [5]. As an alternative a pproach to the same issue, some authors considered relaxations of the Bayes ian formalism, such as fuzzy models [6] and imprecise probabilities [17]. The latter is the di- rection we consider here, but trying to overcome the computation al limitations of that approach when coping with information-theoretic scores. This has some analogy with the approach in [9], that is focused on the Bayesian case only, but whose score, based on the same-decision problem, appears hard to be extended to the imprecise framework without affecting the computational co mplexity. 3 Background on Bayesian and Credal Networks We denote variables by Latin uppercase letters, while using lowercas e for their generic values, and calligraphic for the set of their possible values. T hus,v∈ V is a possible value of V. Here we only consider discrete variables.1 1IRT uses instead with continuous skills. Yet, when coping pr obabilistic models, hav- ing discrete skill does no prevent evaluations to range over a continuous domain. E.g., see Table 1, where the grade corresponds to a (continuo us) probability.4 Antonucci et al. 3.1 Bayesian Networks A probability mass function (PMF) over Vis denoted as P(V), whileP(v) is the probability assigned to state v. Given a function fofV, its expectation with respect to P(V) isEP(f) :=/summationtext v∈VP(v)f(v). The expectation of −logb[P(V)] is calledentropyand denoted also as H(X).2We setb:=|V|to have the maximum of the entropy, achieved for uniform PMFs, equal to one. Given joint PMF P(U,V), the marginal PMF P(V) is obtained by sum- ming out the other variable, i.e., P(v) =/summationtext u∈UP(u,v). Conditional PMFs such asP(U|v) are similarly obtained by Bayes’s rule, i.e., P(u|v) =P(u,v)/P(v) provided that P(v)>0. Notation P(U|V) :={P(U|v)}v∈Vis used for such con- ditional probability table (CPT). The entropy of a conditional PMF is d efined as in the unconditional case and denoted as H(U|v). The conditional entropy is a weighted average of entropies of the conditional PMFs, i.e., H(U|V) :=/summationtext v∈VH(U|v)P(v). IfP(u,v) =P(u)P(v) for each u∈ Uandv∈ V, variables UandVare independent. Conditional formulations are also considered. We assume the set of variables V:= (V1,...,V r) to be in one-to-one corre- spondence with a directed acyclic graph G. For each V∈V, the parents of V, i.e., the predecessors of VinG, are denoted as Pa V. GraphGtogether with the collection of CPTs {P(V|PaV)}V∈Vprovides a Bayesian network (BN) specifi- cation [15]. Under the Markov condition, i.e., every variable is condition ally in- dependent of its non-descendants non-parents given its parent s, a BN compactly defines a joint PMF P(V) that factorizes as P(v) =/producttext V∈VP(v|paV). Inference, intended as the computation of the posterior PMF of a single (querie d) variables given some evidence about other variables, is in general NP-hard, b ut exact and approximate schemes are available (see [15] for details). 3.2 Credal Sets and Credal Networks A set of PMFs over Vis denoted as K(V) and called credal set (CS). Expec- tations based on CSs are the bounds of the PMF expectations with r espect to the CS. Thus E[f] := inf P(V)∈K(V)E[f] and similarly for the supremum E. Ex- pectations of events are in particular called lower and upper probab ilities and denoted as PandP. Notation K(U|v) is used for a set of conditional CSs, while K(U|V) :={K(U|v)}v∈Vis a credal CPT (CCPT). Analogously to a BN, a credal network (CN) is specified by graph Gtogether with a family of CCPTs {K(V|PaV)}V∈V[11]. A CN defines a joint CS K(V) corresponding to all the joint PMFs induced by BNs whose CPTs are c onsis- tent with the CN CCPTs. For CNs, we intend inference as the comput ation of the lower and upper posterior probabilities. The task generalizes BN inference being therefore NP-hard, see [19] for a deeper characterization . Yet, exact and approximate schemes are also available to practically compute infere nces [4]. 2We set 0·logb0 = 0 to cope with zero probabilities.A New Score for Adaptive Tests in Bayesian and Credal Network s 5 4 Testing Algorithms Atypicaltestaimsatevaluatingthe knowledgelevelofatesttaker σonthe basis of her answers to a number of questions. Let Qdenote a repository of questions availabletothe instructor.The orderand the number ofquestions pickedfrom Q to be asked to σmight not be defined in advance. We call testing algorithm (TA) a procedure taking care of the selection of the sequence of quest ions asked to the test taker, and to decide when the test stops. Algorithm 1 depicts a general TA scheme, with edenoting the array of the answers collected from test taker σ. Algorithm 1 General TA: given the profile σand repository Q, an evaluation based on answers eis returned. 1:e←∅ 2:while not Stopping (e)do 3:Q∗←Pick(Q,e) 4:q∗←Answer(Q∗,σ) 5:e←e∪{Q∗=q∗} 6:Q←Q\{Q∗} 7:end while 8:returnEvaluate (e) Boolean function Stopping decides whether the test should end, this choice being possibly based on the previous answers in e. Trivial stopping rules might be based on the number of questions asked to the test takes ( Stopping (e) = 1 if and only if |e|> n) or on the number of correct answers provided that a maximum number of questions is not exceeded. Function Pickselects instead the question to be asked to the student from the repository Q. A TA is called adaptive when this function takes into account the previous answers e. Trivial non-adaptive strategies might consist in randomly picking an element ofQor following a fixed order. Function Answeris simply collecting (or simulating) the answeroftest taker σtoaparticularquestion Q. In ourassumptions,this answer is not affected by the previous answers to other questions.3 Finally,Evaluate is a function returning the overall judgement of the test (e.g., a numerical grade or a pass/fail Boolean) on the basis of all th e answers collected after the test termination. Trivial examples of such func tions are the percentage of correct answers or a Boolean that is true when a su fficient number of correct answers has been provided. Note also that in our assum ptions the TA isexchangeable , i.e., the stopping rule, the question finder and the evaluation function are invariant with respect to permutations in e[22]. In other words, 3Generalized setupswhere thequalityofthestudentanswer i s affectedbytheprevious answers will be discussed at the end of the paper. This might i nclude a fatigue model negatively affecting the quality of the answers when ma ny questions have been already answered as well as the presence of revealing questions that might improve the quality of other answers [16].6 Antonucci et al. the same next question, the same evaluation and the same stopping decision is produced for any two students, who provided the same list of answ ers in two different orders. A TA is supposed to achieve reliable evaluation of taker σfrom the answers e. As each answer is individually assumed to improve such quality, asking all the questions, no matter the order because of the exchangeability as sumption, is an obvious choice. Yet, this might be impractical (e.g., because of time lim itations) or just provide an unnecessary burden to the test taker. The go al of a good TA is therefore to trade off the evaluation accuracy and the number o f questions.4 5 Adaptive Testing in Bayesian and Credal Networks The general TA setup in Algorithm 1 can be easily specialized to BNs as f ollows. First, we identify the profile σof the test taker with the actual states of a number of latent discrete variables, called skills. LetS={Si}n j=1denote these skill variables, and sσthe actual values of the skills for the taker. Skills are typically ordinal variables, whose states corresponds to increasin g knowledge levels. Questions in Qare still described as manifest variables whose actual values are returned by the answerfunction. This is achieved by a (possibly stochastic) function of the actual profile sσ. This reflects the taker perspective, while the teacher has clearly no access to sσ. As a remark, note that we might often coarsenthe set ofpossible values Qfor eachQ∈Q: for instance, amultiple choicequestionwiththreeoptionsmighthaveasinglerightanswer,t hetwoother answers being indistinguishable from the evaluation point of view.5 A joint PMF over the skills Sand the questions Qis supposed to be avail- able. In particular we assume this to correspond to a BN whose grap h has the questions as leaf nodes. Thus, for each Q∈Q,PaQ⊆Sand we call PaQ thescopeof question Q. Note that this assumption about the graph is simply reflecting a statement about the conditional independence betwe en (the answer to) a question and all the other skills and questions given scope of th e ques- tion. This basically means that the answers to other questions are n ot directly affecting the answer to a particular question, and this naturally follo ws from the exchangeability assumption.6 As the available data are typically incomplete because of the latent na ture of the skills, dedicated learning strategies, such as various form of constrained EM should be considered to train a BN from data. We refer the reade r to the variouscontributionsofPlajnerand Vomlel in this field (e.g.,[21]) for acomplete discussion of that approach. Here we assume the BN quantification available. 4In some generalized setups, other elements such as a serendipity in choice in order to avoid tedious sequences of questions might be also consid ered [7]. 5The case of abstention to an answer and the consequent problem of modelling the incompleteness is a topic we do not consider here for the sake of conciseness. Yet, general approaches based on the ideas in [18] could be easily adopted. 6Moving to other setups would not be really critical because o f the separation prop- erties of observed nodes in Bayesian and credal networks, se e for instance [3,8].A New Score for Adaptive Tests in Bayesian and Credal Network s 7 In such a BN framework, Stopping (e) might be naturally based on an eval- uation of the posterior PMF P(S|e), this being also the case for Evaluate . Re- garding the question selection, Pickmight be similarly based on the (posterior) CPTP(S|Q,e), whose values for the different answers to Qmight be weighted by the marginal P(Q|e). More specifically, entropies and conditional entropies are considered by Algorithm 2, while the evaluation is based on a condit ional expectation for a given utility function. Algorithm 2 Information Theoretic TA in BN over the questions Qand the skillsS: given the student profile sσ, the algorithms returns an evaluation cor- responding to the expectation of an evaluation function fwith respect to the posterior for the skills given the answers e. 1:e=∅ 2:whileH(S|e)> H∗do 3:Q∗←argmax Q∈Q[H(S|e)−H(S|Q,e)] 4:q∗←Answer(Q∗,sσ) 5:e←e∪{Q∗=q∗} 6:Q←Q\{Q∗} 7:end while 8:returnEP(S|e)[f(S)] When no data are available for the BN training, elicitation techniques s hould beconsideredinstead.AsalreadydiscussedCNsmightofferabette rformalismto capture domain knowledge, especially by providing interval-valued pr obabilities instead of sharp values. If this is the case, a CN version of Algorithm 2 can be equivalentlyconsidered.MovingtoCNsisalmostthesame,providedt hatbounds on the entropy are used instead for decisions. Yet, the price of su ch increased realism in the elicitation is the higher complexity characterizinginferen ces based on CNs. The work in [17] offers a critical discussion of those issues, that are only partially addressed by heuristic techniques used there to appr oximate such bounds. In the next section we consider an alternative approach t o cope with CNs and adaptive TAs based on different scores used to select the q uestions. 6 Coping with the Mode Following[25], we can regardthe PMF entropy(and its conditionalver sion)used by Algorithm 2 as an example of index of qualitative variation (IQV). An IQV is just a normalized number that takes value zero for degenerate P MFs, one on uniform ones, being independent on the number of possible states ( and samples for empirical models). The closer to uniform is the PMF, the higher is t he index and vice versa. In order to bypass the computational issues related to its applicat ion with CNs and the explainability limits with both BNs and CNs, we want to consid er8 Antonucci et al. alternative IQVs to replace entropy in Algorithm 2. Wilkox deviation from the mode(DM) appears a sensible option. Given PMF P(V), this corresponds to: M(V) := 1−/summationdisplay v∈Vmaxv′∈VP(v′)−P(v) |V|−1. (1) It is a trivial exercise to check that this is a proper IQV, with the sam e unimodal behaviour of the entropy. In terms of explainability, being a linear fu nction of the modal probability, the numerical value of the DM offers a more tr ansparent interpretation than the entropy. From a computational point of v iew, for both marginaland unconditional PMFs, both the entropy and the DM can be directly obtained from the probabilities of the singletons. The situation is different when computing the bounds of these quant ities with respect to a CS. The bounds of M(V) are obtained from the upper and lower probabilities of the singletons by simple algebra, i.e, M(V) := max P(V)∈K(V)M(V) :=|V|−maxv′∈VP(v′) |V|−1, (2) and analogously with the lower probabilities for M(V). Maximizing entropy requires instead a non-trivial, but convex, optimization. See for ins tance [1] for an iterative procedure to find such maximum when coping with CSs defi ned by probability intervals. The situation is even more critical for the minimiz ation, that has been proved to be NP-hard in [26]. The optimization becomes even more challenging for conditional entr opies, there basically are mixtures of conditional entropies based on impre cise weights. Consequently, in [17], only inner approximation for the upper bound h ave been derived.ThesituationisdifferentforconditionalDMs.Thefollowingr esultoffers a feasible approachin a simplified setup, to be later extended to the g eneralcase. Theorem 1. Under the setup of Section 5, consider a CN with a single skill Sand a single question Q, that is a child of S. LetK(S)andK(Q|S)be the CCPTs of such CN. Let also Q={q1,...,qn}andS={s1,...,sm}. The upper conditional DM, i.e., M(S|Q) :=|S|− max P(S)∈K(S) P(Q|S)∈K(Q|S)/summationdisplay i=1,...,n/bracketleftbigg max j=1,...,mP(sj|qi)/bracketrightbigg P(qi),(3) whose normalizing denominator was omitted for the sake of br evity, is such that: M(S|Q) :=m−max ˆji=1,...,m i=1,...,nΩ(ˆj1,...,ˆjn), (4)A New Score for Adaptive Tests in Bayesian and Credal Network s 9 whereΩ(ˆj1,...,ˆjn)is the solution of the following linear programming task her e below. max/summationdisplay jxij s.t./summationdisplay ijxij= 1 (5) xij≥0 ∀i,j (6) /summationdisplay ixij≥P(sj)∀j (7) /summationdisplay ixij≤P(sj)∀j (8) P(qi|sj)/summationdisplay ixij≤xij ∀i,j (9) P(qi|sj)/summationdisplay ixij≥xij ∀i,j (10) xiˆji≥xij ∀i,j (11) Note that the bounds on the sums over the indexes and on the uni versal quanti- fiers are also omitted for the sake of brevity. Proof. Equation (3)rewrites as: M(S|Q) =m−max P(S)∈K(S) P(Q|S)∈K(Q|S)n/summationdisplay i=1/bracketleftbigg max j=1,...,mP(sj)P(qi|sj)/bracketrightbigg .(12) Let us define the variables of such constrained optimization task as: xij:=P(sj)·P(qi|sj). (13) for each i= 1,...,nandj= 1,...,m. Let us show how the CCPT constraints can be easily reformulated with respect to such new variable s by simply noticing thatxij=P(sj,qi), and hence P(si) =/summationtext ixijandP(qj|si) =xij/(/summationtext kxkj). Consequently, the interval constraints on P(S)corresponds to the linear con- straints in Equations (7)and(8). Similarly, for P(Q|S), we obtain: P(qi|sj)≤xij/summationtext kxkj≤P(qi|sj), (14) that easily gives the linear constraints in Equations (9)and(10). The non- negativity of the probabilities corresponds to Equation (6), while Equation (5) gives the normalization of P(S)and the normalization of P(Q|S)is by con- struction. Equation (12)rewrites therefore as: M(S|Q) =m−max {vij}ij∈Γ/summationdisplay imax jxij, (15)10 Antonucci et al. whereΓdenotes the linear constraints in Equations (5)-(10). If we set ˆji:= argmax jxij, (16) Equation (15)rewrites as M(S|Q) = max {vij}ij∈Γ′/summationdisplay ixiˆji, (17) whereΓ′are the constraints in Γwith the additional (linear) constraints in Equation (11), that are implementing Equation (16). The optimization on the right-hand side of Equation (17)is not a linear programming task, as the values of the indexes ˆjicannot be decided in advance being potentially different for different assignments of the optimization variables consistent with the constraints in Γ. Yet,we might address such optimization as a brute-force task with respect to all the possible assignati on of the indexes ˆji. This is exactly what is done by Equation (4)where all the mnpossible assignations are considered. This proves the thesis. ⊓ ⊔ An analogous result with the linear programming tasks minimizing the sa me objectivefunctionswithexactlythesameconstraintsallowstocom puteM(S|Q). The overall complexity is clearly O(mn) withn:=|Q|. This means quadratic complexity for any test where only the difference between a wrong a nd a right answer is considered from an elicitation perspective, and tractable computations providedthatthenumberofpossibleanswerstothesamequestion wedistinguish is bounded by a small constant. Coping with multiple answers becomes trivial by means of the results in [3], that allows to merge multiple observed c hildren into a single one. Finally, the case of multiple skills might be similarly conside red by using the marginal bounds of the single skills in Equations (7) and (8 ). 7 Experiments In this section we validate the ideas outlined in the previous section in o rder to check whether or not the DM can be used for TAs as a sensible altern ative to information-theoreticscoressuchastheentropy.IntheBNcon text,thisissimply achieved by computing the necessary updated probabilities, while Th eorem 1 is used instead for CNs. 7.1 Single-Skill Experiments on Synthetic Data For a very first validation of our approach, we consider a simple setu p made of a single Boolean skill Sand a repository with 18 Boolean questions based on nine different parametrizations (two questions for parametrization). In such BN, the CPT of a question can be parametrized by two numbers. E.g., in the ex ample in Figure 1, we used the probabilities of correctly answering the ques tion givenA New Score for Adaptive Tests in Bayesian and Credal Network s 11 that the skill is present or not, i.e., P(Q= 1|S= 1) and P(Q= 1|S= 0). A more interpretable parametrization can be obtained as follows: δ:= 1−1 2[P(Q= 1|S= 1)+P(Q= 1|S= 0)], (18) κ:=P(Q= 1|S= 1)−P(Q= 1|S= 0). (19) Note that P(Q= 1|S= 1)> P(Q= 1|S= 0) is an obvious rationality con- straint for questions, otherwise having the skill would make less likely to answer properly to a question. Both parameters are therefore non-neg ative. Parameter δ, corresponding to the (arithmetic) average of the probability for a wrong an- swer over the different skill values, can be regarded as a normalized index of the question difficulty . E.g., in Figure 1, Q1(δ= 0.4) is less difficult than Q2 (δ= 0.5). Parameter κcan be instead regarded as a descriptor of the differ- ence of the conditional PMFs associated with the different skill value s. In the most extreme case κ= 1, the CPT P(Q|S) is diagonal implementing an iden- tity mapping between the skill and the question. We therefore rega rdκas a indicator of the discriminative power of the question. In our tests, for the BN quantification, we consider the nine possible parametrizations corr esponding to (δ,γ)∈[0.4,0.5,0.6]2. ForP(S) we use instead a uniform quantification. For the CN approach we perturb all the BN parameters with ǫ=±0.05, thus obtaining a CN quantification. A group of 1024 simulated students, half of the m having S= 0 and half with S= 1 is used for simulations. The student answers are sampled from the CPT of the asked question on the basis of the stud ent profile. Figure 2 (left) depicts the accuracy of the BN and CN approaches b ased on both the entropy and the DM scores. For credal models, decisions are based on the mid-point between the lower and the upper probability, while lower entropy and conditional entropies are used. We notably see all the adaptive approaches outperforming a non-adaptive, random, choice of the questions. To better in- vestigate the strong overlap between these trajectories, in Figu re 2 (right) we compute the Brier score and we might observe the strong similarity b etween DM and entropy approaches in both the Bayesian and the credal ca se, with the credal approaches slightly outperforming the Bayesian ones. 7.2 Multi-Skill Experiments on Real Data For a validation on real data, we consider an online German language p lacement test (see also [17]). Four different Booleanskills associated with differ ent abilities (vocabulary, communication, listening and reading) are considered and modeled by a chain-shaped graph, for which BN and CN quantification are alre ady avail- able. A repository of 64 Boolean questions, 16 for each skill, with fou r different levels of difficulty and discriminative power, have been used. Experiments have been achieved by means of the CREMA library for c redal networks [14].7The Java code used for the simulations is available together with the Python scripts used to analyze the results and the model spec ifications.8 7github.com/IDSIA/crema 8github.com/IDSIA/adaptive-tests12 Antonucci et al. 0 5 10 15 2000.20.40.60.81 Number of questionsAccuracy 0 5 10 15 2000.10.20.30.40.5 Number of questionsBrier DistanceCredal Entropy Credal Mode Random Bayesian Entropy Bayesian Mode Fig.2.Accuracy (left) and Brier distance (right) of TAs for a singl e-skill BN/CN Performancesare evaluated as for the previous model, the only diff erence be- ing that here the accuracy is aggregated by average over the sep arate accuracies for the four skills. The observed behaviour, depicted in Figure 3, is a nalogous to that of the single skill case: entropy-based and mode-based sc ores are provid- ing similar results, with the credal approach typically leading to more a ccurate evaluations (or evaluations of the same quality with fewer questions ). 0 10 20 30 40 50 6000.20.40.60.81 Number of questionsAggregated AccuracyCredal Entropy Credal Mode Random Bayesian Entropy Bayesian Mode Fig.3.Aggregated Accuracy for a multi-skill TA 8 Outlooks and Conclusions A new score for adaptive testing in Bayesian and credal networks h as been proposed. Our proposal is based on indexes of qualitative variation , being in particular focused on the modal probability for their explainability fe atures. An algorithm to evaluate this quantity in the credal case is derived. Our experi- ments show that moving to these scores does not really affect the q uality of the selection process. Besides a deeper experimental validation, a nec essary future work consists in the derivation of simpler elicitation strategies for th ese model in order to promote their application to real-world testing environme nts.A New Score for Adaptive Tests in Bayesian and Credal Network s 13 References 1. Abellan, J., Moral, S.: Maximum of entropy for credal sets . International Journal of Uncertainty, Fuzziness and Knowledge-Based Systems 11(05), 587–597 (2003) 2. Almond, R.G., Mislevy, R.J.: Graphical models and comput erized adaptive testing. Applied Psychological Measurement 23(3), 223–237 (1999) 3. Antonucci, A., Piatti, A.: Modeling unreliable observat ions in Bayesian networks by credal networks. In: Godo, L., Pugliese, A. (eds.) Scalab le Uncertainty Man- agement, Third International Conference, SUM 2009. Procee dings. Lecture Notes in Computer Science, vol. 5785, pp. 28–39. Springer (2009) 4. Antonucci, A., de Campos, C.P., Huber, D., Zaffalon, M.: Ap proximating credal networkinferences bylinear programming. In:vanderGaag, L.C. (ed.)Proceedings of the 12th European Conference on Symbolic and Quantitativ e Approaches to Reasoning with Uncertainty. Lecture Notes in Artificial Int elligence, vol. 7958, pp. 13–25. Springer, Utrecht, The Netherlands (2013) 5. Bachrach, Y., Graepel, T., Minka, T., Guiver, J.: How to gr ade a test without knowing the answers—a Bayesian graphical model for adaptiv e crowdsourcing and aptitude testing. arXiv preprint arXiv:1206.6386 (2012) 6. Badaracco, M., Mart´ ınez, L.: A fuzzy linguistic algorit hm for adaptive test in intel- ligent tutoring system based on competences. Expert System s with Applications 40(8), 3073–3086 (2013) 7. Badran, M.E.K., Abdo, J.B., Al Jurdi, W., Demerjian, J.: A daptive serendipity for recommender systems: Let it find you. In: ICAART (2). pp. 7 39–745 (2019) 8. Bolt, J.H., De Bock, J., Renooij, S.: Exploiting Bayesian network sensitivity func- tions for inference in credal networks. In: Proceedings of t he Twenty-Second Eu- ropean Conference on Artificial Intelligence (ECAI). vol. 2 85, pp. 646–654. IOS Press (2016) 9. Chen, S.J., Choi, A., Darwiche, A.: Computer adaptive tes ting using the same- decision probability. In: BMA@ UAI. pp. 34–43 (2015) 10. Conati, C., Gertner, A.S., VanLehn, K., Druzdzel, M.J.: On-line student modeling for coached problem solving using Bayesian networks. In: Us er Modeling. pp. 231– 242. Springer (1997) 11. Cozman, F.G.: Credal networks. Artificial intelligence 120(2), 199–233 (2000) 12. Embretson, S.E., Reise, S.P.: Item response theory. Psy chology Press (2013) 13. H´ ajek, A., Smithson, M.: Rationality and indeterminat e probabilities. Synthese 187(1), 33–48 (2012) 14. Huber, D., Caba˜ nas, R., Antonucci, A., Zaffalon, M.: CRE MA: a Java library for credal network inference. In: Jaeger, M., Nielsen, T. (e ds.) Proceedings of the 10th International Conference on Probabilistic Graphical Models (PGM 2020). Proceedings of Machine Learning Research, PMLR, Aalborg, D enmark (2020) 15. Koller, D., Friedman, N.: Probabilistic Graphical Mode ls: Principles and Tech- niques. MIT press (2009) 16. Laitusis, C.C., Morgan, D.L., Bridgeman, B., Zanna, J., Stone, E.: Examination of fatigue effects from extended-time accommodations on the SAT reasoning test. ETS Research Report Series 2007(2), i–13 (2007) 17. Mangili, F., Bonesana, C., Antonucci, A.: Reliable know ledge-based adaptive tests by credal networks. In: Antonucci, A., Cholvy, L., Papini, O . (eds.) Symbolic and QuantitativeApproachestoReasoning with Uncertainty.EC SQARU2017. Lecture Notes in Computer Science, vol. 10369, pp. 282–291. Springe r, Cham (2017)14 Antonucci et al. 18. Marchetti, S., Antonucci, A.: Reliable uncertain evide nce modeling in Bayesian networks by credal networks. In: Brawner, K.W., Rus, V. (eds .) Proceedings of the Thirty-First International Florida Artificial Intelligen ce Research Society Confer- ence (FLAIRS-31). pp. 513–518. AAAI Press, Melbourne, Flor ida, USA (2018) 19. Mau´ a, D.D., De Campos, C.P., Benavoli, A., Antonucci, A .: Probabilistic infer- ence in credal networks: new complexity results. Journal of Artificial Intelligence Research 50, 603–637 (2014) 20. Piatti, A., Antonucci, A., Zaffalon, M.: Building knowle dge-based expert systems by credal networks: a tutorial. In: Baswell, A. (ed.) Advanc es in Mathematics Research, vol. 11, chap. 2. Nova Science Publishers, New Yor k (2010) 21. Plajner, M., Vomlel,J.: Monotonicityinpractice ofada ptivetesting. arXivpreprint arXiv:2009.06981 (2020) 22. Sawatzky, R., Ratner, P.A., Kopec, J.A., Wu, A.D., Zumbo , B.D.: The accuracy of computerized adaptive testing in heterogeneous populat ions: A mixture item- response theory analysis. PloS one 11(3), e0150563 (2016) 23. Vomlel, J.: Bayesian networks in educational testing. I nternational Journal of Un- certainty, Fuzziness and Knowledge-Based Systems 12(supp01), 83–100 (2004) 24. Vomlel, J.: Building adaptive tests using Bayesian netw orks. Kybernetika 40(3), 333–348 (2004) 25. Wilcox, A.R.: Indices of qualitative variation and poli tical measurement. Western Political Quarterly 26(2), 325–343 (1973) 26. Xiang, G., Kosheleva, O., , Klir, G.J.: Estimating infor mation amount under in- terval uncertainty: algorithmic solvability and computat ional complexity. Tech. Rep. 158, Departmental Technical Reports (CS) (2006)