 Methodology article
 Open
 Published:
Application of permanents of square matrices for DNA identification in multiplefatality cases
BMC Geneticsvolume 14, Article number: 72 (2013)
Abstract
Background
DNA profiling is essential for individual identification. In forensic medicine, the likelihood ratio (LR) is commonly used to identify individuals. The LR is calculated by comparing two hypotheses for the sample DNA: that the sample DNA is identical or related to a reference DNA, and that it is randomly sampled from a population. For multiplefatality cases, however, identification should be considered as an assignment problem, and a particular sample and reference pair should therefore be compared with other possibilities conditional on the entire dataset.
Results
We developed a new method to compute the probability via permanents of square matrices of nonnegative entries. As the exact permanent is known as a #Pcomplete problem, we applied the Huber–Law algorithm to approximate the permanents. We performed a computer simulation to evaluate the performance of our method via receiver operating characteristic curve analysis compared with LR under the assumption of a closed incident. Differences between the two methods were well demonstrated when references provided neither obligate alleles nor impossible alleles. The new method exhibited higher sensitivity (0.188 vs. 0.055) at a threshold value of 0.999, at which specificity was 1, and it exhibited higher area under a receiver operating characteristic curve (0.990 vs. 0.959, P = 9.6E15).
Conclusions
Our method therefore offers a solution for a computationally intensive assignment problem and may be a viable alternative to LRbased identification for closedincident multiplefatality cases.
Background
DNA profiling is crucial in the identification of human remains, particularly when other physical clues are absent. Currently, the genetic status of an individual is commonly profiled by short tandem repeat (STR) loci. There are two types of approaches to DNAbased identification of an unidentified body: 1) direct matching, and 2) kinship testing. Direct matching is performed when the reference DNA of a victim can be obtained from his/her belongings (direct reference). When direct reference is not available, indirect reference is obtained from the victim’s relatives, and a probability distribution of genotypes is inferred for the victim. In either approach, identification is determined based on the likelihood ratio (LR) between two hypotheses: that the DNA of an unidentified body is that of a particular missing person (MP) (hypothesis H_{1}), and that it is randomly sampled from a population (hypothesis H_{2}). This approach with use of STRs has been spread since mid1990th as a routine use for individual identification. The first application to mass disaster identification was reported in 1997 [1], and DNA identification with STRs became the major tool in mass disaster identification. Although the methodology has been well studied and established for onetoone identification problem, assignment between many bodies and many families (manytomany identification) has not been studied well. Potential problems of applying the onetoone comparison to manytomany situation are 1) that the confidence in assigning a body to a family is not independent of the confidence in assigning others, 2) that the confidence in assignment is not independent of the number of missing data (such as unrecovered bodies and unreported MPs).
To illustrate the problems, if we assume that the number of victims is two and that the DNAs of two bodies and reference DNAs for two MPs are all available, we can assume with a high degree of confidence that either body 1 or body 2 is MP 1 or MP 2. The conventional LRbased method (LR method) independently compares body 1 to MP 1, body 1 to MP 2, body 2 to MP 1, and body 2 to MP 2. However, these four comparisons are not independent; if the probability of body 1 corresponding to MP 1 is very high, the probability of body 1 corresponding to MP 2 should be very small because only two possibilities are considered for the identity of body 1 and the two probabilities must add up to 1. Another concern with the LR method is as follows. If we consider a LR of 1000 to be sufficiently high to approve the identity of a body, what if the LRs for both of our MPs exceed 1000? In this case, we might decide that the MP with the higher LR is body 1. But what if the LR for MP 1 is 1001, and that for MP 2 is 1002? Should we elect MP 2? When the number of victims is small, for example less than 10, the “best” assignment might be successfully determined by “intelligence”, but this method would fail if the number of victims is as large as 1000 or above. The fact that LR is always independent of the probabilities of other pairs in the data raises another problem in LR: Unavailable data cannot be taken into account. Here, “unavailable data” means unrecovered bodies and MPs with no available reference. As unavailable data are a source of uncertainty, the amount of unavailable data should affect the test result. Note that LR is reduced by considering a prior probability, one divided by the total number of victims, but test results do not differ regardless of whether data are complete or most are absent.
To address these problems, manytomany approaches need to be developed. Lin et al. [2] proposed a global approach that compares the likelihood of assigning all samples obtained from human remains to all reporting families to the likelihood of the null hypothesis where all the samples were assumed to be obtained randomly from a population, which was computationally easy but not realistic. We propose a new method, called the permanent method, in which we compare two hypotheses: H_{1}, that the identity of body i is MP j; and H^{’}_{2}, instead of H_{2}, that the identity of body i is any MP other than MP j. Clearly, this comparison is more realistic in that all possible hypotheses involving victims of the same incident are considered, and importantly, these two hypotheses together comprise all possibilities, unlike the LR method; that is, the probabilities for the two hypotheses must add up to 1. Therefore in the example above, if the probability that body 1 is MP 1 is 0.99, the probability that body 1 is MP 2 is 0.01.
We considered identification in multiplefatality cases as an assignment problem and developed a method to compute the probability that body i corresponds to MP j conditional on the entire dataset. This approach can be summarized as follows. Assume that every recovered body has been genotyped and that a probability distribution of its genotypes is given. Assume that every reported MP has DNA available for genotyping from a “reference” person (either the victim or a close relative), which in turn can be genotyped to obtain a probability distribution of the genotypes. Using this information, we can compute the matrix P, which quantifies the compatibility of the body’s genotype with that of a reference person. The matrix P is then used to compute matrix Q, which computes the probability that body i be assigned to MP j. This computation takes into account the number of victims, the number of “unreported” victims, and the number of “unrecovered” bodies. If the resulting probability exceeds a certain threshold v, then the body is matched to the person. By employing the thresholdbased decision made for each pair of one body and one MP, this approach does not necessarily optimize an assignment as a whole, but uses information that is global.
To handle this problem, data must contain the same number of bodies and MPs both theoretically and technically: Theoretically, the number of bodies (whether or not they were recovered) must be the same as the number of MPs (whether or not they were reported); technically, the algorithm to solve this problem requires a square matrix as described below. Therefore, when some of the bodies are not recovered, then they are assumed to exist, but with a genotype probability distribution that is the same as that of the population. Similarly, when some of the victims are not reported as missing or have no available reference DNA, then they are assumed to exist, but with a genotype probability distribution that is the same as that of the population. The key idea in computing the probability is to find the sum of weights of perfect matching in a bipartite graph, known as the permanent of a square matrix. Exact computation of a permanent is #Pcomplete [3]. A #P class is a set of counting problems that belongs to nondeterministic polynomial time, and a problem is #Pcomplete if and only if it is in #P. Because it has been proved that the permanent is #Pcomplete, exact computation of the permanent is not possible in polynomial time, and algorithms for polynomial time approximation have therefore been proposed [4–7]. We employed the Huber–Law algorithm [6], which is, to the best of our knowledge, currently the fastest algorithm when the matrix is dense. The Huber–Law algorithm is an exact sampling algorithm based on the sequential acceptance/rejection method. The accuracy of estimation depends on the acceptance number obtained from an iteration process, and thus when the probability of acceptance is small, computation time may increase dramatically. What is particularly relevant in the case of DNA identification is that a permanent becomes extremely small as the total number of victims increases. Computation time for an extremely small permanent estimation has not been investigated in previous studies.
In this paper, we first describe the logic of our method. We then present the results of a computer simulation performed to evaluate the performance of our method compared with that of the conventional LRbased method, along with results showing the influence of the presence of unavailable data, using receiver operating characteristic (ROC) curve analysis, and the results of assessment of the accuracy and processing time of the approximation algorithm in the case of DNA identification.
Methods
Notation
Assume that a fatality incident has a total of n_{ w } victims. Assume that their bodies (some of which may not have been recovered) are denoted by B = (b_{ i }), i = 1, 2, ⋯, n_{ w }, and the corresponding MPs (some of whom may not have been reported) are denoted by M = (m_{ j }), j = 1, 2, ⋯, n_{ w }. Because all bodies may not have been recovered, let n_{ b } ≤ n_{ w } be the number of recovered bodies. Similarly, because some victims of the fatality may not have been reported, let n_{ m } ≤ n_{ w } be the number of reported MPs. A victim is considered “reported” if either a direct or an indirect reference is available for genotyping. There are four possible situations: 1) n_{ w } = n_{ b } = n_{ m }, 2) n_{ w } > n_{ b } and n_{ w } = n_{ m }, 3) n_{ w } = n_{ b } and n_{ w } > n_{ m }, and 4) n_{ w } > n_{ b } and n_{ w } > n_{ m }. In cases 2), 3), and 4), unrecovered bodies and/or unreported MPs exist. β = (β_{ i }), i = 1, ⋯, n_{ w } denotes the probability distributions of the genotypes of n_{ w } bodies with the following properties of β_{ i }; 1) the number of elements equals to the number of possible genotypes, 2) each element represents the probability of the genotype being consistent with the genotype of b_{ i }, 3) the sum of all the elements equals 1, and 4) when the genotype can be determined, one element is 1 and the other elements are 0. The elements of β_{ i } take either one of the two possible forms:
where γ_{ b,i } denotes the probability distribution of the genotype of b_{ i }, given that the DNA from the body is available for genotyping, and γ_{*} is the probability distribution of genotypes in a population, whereby genotypes of n_{ w } – n_{ b } unrecovered bodies are assumed to follow γ_{*}.
Similarly, μ = (μ_{ j }), j = 1, ⋯, n_{ w } denotes the probability distributions of the genotypes of n_{ w } MPs, and μ_{ j } has the same properties as β_{ i }. The elements of μ_{ j } are also given as follows:
Here, γ_{ m,j } is the probability distribution of the genotype of m_{ j }, given its reference DNA, and the probability distribution for n_{ w } – n_{ m } unreported MPs is assumed to follow γ_{ * }; β_{ i } and μ_{ j } can be expressed as a vector in which the value of each element is the probability of each genotype.
We define an n_{ w } × n_{ w } matrix P = (p_{ i,j }), where p_{ i,j } is the probability that b_{ i } and m_{ j } have an identical genotype, and is calculated as the inner product of β_{ i } and μ_{ j }. We also define an n_{ w } × n_{ w } matrix Q = (q_{ i,j }), where q_{ i,j } is the probability that b_{ i } and m_{ j } are identical conditional on a matrix P.
To calculate a matrix Q from a matrix P, we consider the permanents of square matrices[8]. We use perm(A) to denote the permanent of an n × n matrix A = (α_{ i,j }). The set of permutations on {1, 2, ⋯, n} is S_{ n } and is needed to define perm(A). Let ${S}_{n}^{\left(i,j\right)}$ denote a subset of S_{ n }, ${S}_{n}^{\left(i,j\right)}=\left\{\sigma \sigma \left(i\right)=j\right\}\in {S}_{n}$, where σ(i) denotes the ith element of σ. We use A_{i,j} to denote the (n – 1) × (n – 1) matrix obtained by removing the ith row and jth column from A.
For the criterion for assignment based on Q, we use v ∈ [0, 1]. The output of the permanent method is either “assigned” or “suspended” for each pair of one body and one MP. A b_{ i } and m_{ j } pair are “assigned” when q_{ i,j } exceeds v, and therefore we can approve the identity of b_{ i } to be m_{ j }. Otherwise, the identity of b_{ i } is “suspended”, i.e., not approved to be m_{ j }. P, Q, and perm(A) are defined in the following sections.
Matrix P
Probability p_{ i,j } is calculated as the inner product of β_{ i } and μ_{ j }. A matrix P consists of four submatrices
Here, p_{*,j}, p_{i,*}, and p_{*,*} indicate the inner products of γ_{*} and γ_{ m,j }, γ_{ b,i } and γ_{ * }, and two γ_{ * }, respectively. Moreover, ∂_{1} is an n_{ b } × n_{ m } matrix; ∂_{2} is an (n_{ w } – n_{ b }) × n_{ m } matrix consisting of identical rows, $\left({p}_{*,1}\cdots {p}_{*,{n}_{m}}\right)$; ∂_{3} is an n_{ b } × (n_{ w } – n_{ m }) matrix consisting of identical columns, ${\left({p}_{1,*}\cdots {p}_{{n}_{b},*}\right)}^{T}$; and ∂_{4} is an (n_{ w } – n_{ b }) × (n_{ w } – n_{ m }) matrix consisting of identical values, p_{*,*}. Each submatrix corresponds to the collection of each of the following four cases:
Case 1. b_{ i } is recovered and m_{ j } is reported.
Case 2. b_{ i } is not recovered and m_{ j } is reported.
Case 3. b_{ i } is recovered and m_{ j } is not reported.
Case 4. b_{ i } is not recovered and m_{ j } is not reported.
Matrix P is required to have at least one nonzero values in each row and column to obtain matrix Q. However, it can happen that all elements in a row or column are zero because mutations and genotyping errors may cause Mendelian errors between truly related individuals. The problem can be overcome by considering the probabilities of these factors; for example, when the genotyping error rate is 0.05, in β_{ i }, the probability of an observed genotype is 0.95, and the total probability of other genotypes is 0.05. By this, any Mendelian errors do not cause zero in matrix P. In our simulation, we did not assume mutations and genotyping errors to simplify the comparison with the LR method, but all elements in the matrix P were nonzero values because we limited the family types (see Methods “Generating families with status patterns”). Not assuming mutations and genotyping errors favors neither our method nor the LR method because mutation influence on only values of matrix P, which is also used in the LR method, but not on computation process from matrix P to matrix Q.
Permanent
In this section, we define the permanent and describe the equations that are important for explaining our method.
The permanent of a matrix A is defined as follows:
Each term expresses the likelihood, or weight, of a permutation. The permanent of A_{–i,–j} can be expressed by ${S}_{n}^{\left(i,j\right)}$ and α_{ i,j } as follows:
The numerator is the sum of weights of permutations in which α_{ i,j } is included. Here, perm(A_{–i,–j}) is given by canceling α_{ i,j } in all terms, and perm(A) can be rewritten by using A_{–i,–j} and α_{ i,j } as follows:
Matrix Q
Assigning each of B to exactly one of M without selecting the same MP more than once is considered to be an assignment problem. If a certain permutation $\sigma \in {S}_{{n}_{w}}$ is adopted, b_{ i } and m_{σ (i)} are paired with each other. That is, if a permutation $\sigma \in {S}_{{n}_{w}}^{\left(i,j\right)}$ is adopted, b_{ i } and m_{ j } are always paired with each other. Here, we refine the definition of q_{ i,j } as the ratio of the sum of the likelihoods of $\sigma \in {S}_{{n}_{w}}^{\left(i,j\right)}$ to the sum of the likelihoods of $\sigma \in {S}_{{n}_{w}}$. By considering Eq. (1) and Eq. (2), we express q_{ i,j } as follows:
A matrix Q is given by calculating q_{ i,j } for all pairs of i and j.
Because both the numerator and the denominator of Eq. (4) include the permanent, we apply the following strategy to minimize estimation error. Let Q' = (q' _{i,j}) denote an n_{ w } × n_{ w } matrix consisting of the numerator of Eq. (4). We avoid using perm(P), instead substituting $\sum}_{i=1}^{{n}_{w}}q{\text{'}}_{i,j$ and $\sum}_{j=1}^{{n}_{w}}q{\text{'}}_{i,j$ for perm(P) by Eq. (3):
Here, $\sum}_{i=1}^{{n}_{w}}{{q}^{\prime}}_{i,j}}={\displaystyle {\sum}_{j=1}^{{n}_{w}}{{q}^{\prime}}_{i,j$ up to an approximation error. We take a smaller value for stringency.
Determining assignment
We determine assignment based on a matrix Q and on v, a threshold value for approving identification. Thus, we can approve b_{ i } and m_{ j } to be identical if q_{ i,j } > v. To avoid confusion, we say that b_{ i } and m_{ j } are paired if they are coupled with each other in a permutation σ, and that b_{ i } and m_{ j } are assigned if q_{ i,j } > v. If q_{ i,j } ≤ v, we say that the decision for this pair is suspended. We assign or suspend all pairs using the identical value of v. Typically, an assignment problem is an optimization problem, in which the solution attempts to find a permutation that optimizes an objective function. For the multifatality victim identification problem considered here, such a global objective function is not necessarily appropriate. Therefore, we can decide to either assign or suspend each pair based on v. This strategy may lead to the result that none of the elements in a row or a column exceed the value of v. In this case, we make no assignment for the corresponding body or MP.
When the value of v is set above 0.5, a unique solution can be derived for the problem: For a given b_{ i }, when q_{ i,j } > v > 0.5, for any ${j}^{\prime}\ne j,{q}_{i,{j}^{\prime}}<1\nu <0.5$ because ${\sum}_{j=1}^{{n}_{w}}{q}_{i,j}}=1$. This can be shown by Eq. (3) as follows. Equation (3) indicates that all column totals and row totals of a matrix Q′, which is an n_{ w } × n_{ w } matrix consisting of the numerator of Eq. (4), are equal to perm(P). Therefore, all column totals and row totals of a matrix Q, which is a matrix Q′ divided by perm(P), are equal to 1. A similar solution can be found for a given m_{ j }.
In other words, a matrix Q is approximately a doubly stochastic matrix, which is defined as a square matrix of nonnegative real numbers, of rows and columns each summed to 1. Therefore, when one element in a column is larger than 0.5, no other elements in the same column can be larger than 0.5. The same can be said in terms of rows. Therefore, v > 0.5 is the condition sufficient to ensure the uniqueness of the solution. Hereafter, we call this method the permanent method.
What value should be used for cutoff depends on situations and criteria conventionally accepted in a society. Typically, in the LR method, LR of 10^{6} is used, which is equivalent to probability of 0.999999. However, the LR method does not tell you the true probability of making a mistake because there are only two possibilities considered in the LR method; 1) the body is obtained from the MP and 2) the body is randomly drawn from the population. On the other hand, the probability given by the permanent method is truly the probability of making a wrong assignment. Therefore, we consider that the permanent method does not need to follow the conventional cutoff. After close discussion with a forensic expert, expecting one mistake in 1000 of the times is stringent enough to discuss the utility of the permanent method.
The permanent method is illustrated by Figure 1.
We summarize the permanent method here so that the advantage of the permanent method would be clear. Assuming that we are interested in the probability that the identity of body 1 (b_{1}) is MP 1 (m_{1}). p_{1,1} is the probability that m_{1} has the genotype of b_{1} given the reference DNA for m_{1}, and relationship between m_{1} and the reference person. Therefore, p_{1,1} is independent of genotypes of other bodies and reference DNAs of other MPs. A likelihood of a possible hypothesis of pairing n_{ w } bodies and n_{ w } MPs is calculated by multiplying elements in matrix P that correspond to the paired bodies and MPs. q_{1,1} is given by the proportion of the sum of the likelihoods of the hypotheses in which b_{1} and m_{1} are paired to the sum of the likelihoods of all possible hypotheses. By using matrix Q instead of matrix P, a probability of b_{1} corresponding to m_{1} influences the probabilities of b_{1} corresponding other MPs. For example, in matrix P, two elements in the same row may have high probability such as p_{1,1} = 0.95, and p_{1,2} = 0.98; p_{1,2} is only slightly higher than p_{1,1}, so we would not determine the identity of b_{1}. Matrix Q is computed by considering all the other elements in matrix P. And it may result in q_{1,1} = 0.0001 and q_{1,2} = 0.9999 depending on other elements in matrix P. What makes this happen? It can happen when, p_{2,1} is much higher than p_{1,1}, for example p_{2,1} = 0.9999999999, and all the other elements in the first column is 0 or very small. In this case, matrix Q reflects this logic: b_{2} is assigned to m_{1}, and therefore, b_{1} must not be m_{1}, and therefore, b_{1} is assigned to m_{2}.
Likelihood ratio and posterior probability
The LR of b_{ i } corresponding to m_{ j } is defined as p_{ i,j } divided by the genotype frequency of b_{ i } in a population,
The numerator of Eq. (6) corresponds to the likelihood that hypothesis H_{1} is true (the DNA of an unidentified body is that of a particular MP), and the denominator corresponds to the likelihood that H_{2} is true (the DNA is randomly sampled from a population). Therefore, the numerator is equal to p_{ i,j } defined in the permanent method. The difference between the two methods becomes apparent after we calculate matrix P: Unlike in the permanent method, in the LR method, the matrix P is divided by the genotype frequency of the corresponding body. For example, p_{1,j}, j = 1,2,…, n_{ w } are divided by the genotype frequency of b_{1}.
In multiplefatality cases, the prior odds are commonly considered as 1/(n_{w} – 1), and according to the Bayesian theorem, the posterior odds are given by the product of LR and the prior odds, r = LR/(n_{w} – 1). We use the posterior odds, r, for test values of the LR method.
Assessment of performance
Probability and odds are interchangeable, and therefore we can use the same assignment criterion, v, to compare q_{ i,j } and r. We can assess the performance of discrimination between identical and nonidentical pairs by using the area under a ROC curve (AUC) [9]. We state that a body and MP pair is identical when the body is that of the MP, and that the pair is nonidentical when the body is that of another MP. A ROC curve plots sensitivity against 1 − specificity, and the AUC is commonly used to evaluate the performance of a measure. Larger values of AUC indicate higher performance.
In our case, sensitivity is defined as the ratio of the number of assigned truly identical pairs to the total number of truly identical pairs, and specificity is defined as the ratio of the number of suspended truly nonidentical pairs to the total number of truly nonidentical pairs. AUC is calculated by summing the areas of trapezoids formed between two adjacent cutoff points (trapezoidal method), and its confidence interval (CI) is estimated by the DeLong method [10]. Two ROC curves are compared by testing their AUCs with a nonparametric method described by DeLong et al. [10].
In the LR method, two or more MPs can be assigned to a body by any cutoff v. In a practical situation, the identity of the body needs to be determined to one, otherwise, it remains undetermined. However, here we consider that, the LR method assigns all pairs that exceeded the cutoff because decision for this case in practice is made with combinatorial information other than DNA, and because the ROC analysis is to assess discriminating performance of an index itself. The same strategy was applied to the permanent method when the cutoff value was set to be v ≤ 0.5.
Simulating a population
We performed a computerbased simulation to assess the performance of the permanent method for DNA profiles of 15 STR loci available in the ABI AmpFISTR Identifiler® PCR Amplification Kit (Applied Biosystem, Foster City, CA, USA). Theoretically, as the number of markers increases, the performance of both the permanent and the LR methods increases. However, both methods assume independence of markers. That is a possible limitation to the number of independent markers, and currently identification by a large number of loci has not been applied to practice. The amelogenin locus was excluded from the study because it is used for sex determination. Although two loci are located within chromosome 2 and another two loci within chromosome 5, we assumed that all 15 loci are independent because the influence of the recombination rate is not an issue when assessing the performance of measures. We used previously reported allele frequencies in a Japanese population [11].
DNA profiles of 12,500 families were simulated for the 15 STR loci. Simulated pedigree trees are shown in Figure 2. Genotypes of founders (I1, I2, I3, I4, and II5) were randomly provided under the assumption of the Hardy–Weinberg equilibrium, and the genotypes of other family members were determined stochastically based on founders’ genotypes without mutations. Each individual has two types of information: One is his/her genotype, and the other concerns the relationship with family members. We describe the simulated 12,500 families as pooled families.
Generating families with status patterns
For each individual, we set one of three statuses, missing, typed, or nottyped, without allowing for multiple MPs in a family. Missing status means that a body is being sought for the person, typed status is assigned when the DNA of the person is available for genotyping as a reference for the MP, and nottyped means that the family members’ DNA is not available. The status pattern of a family indicates which person is assigned to which status. For example, if in a family of father, mother, and a child, the child is missing, and the mother provides her DNA as a reference, while the DNA of the father is unavailable, then the status pattern of this family is that the child is missing, the mother is typed, and the father is nottyped. The simulated genotype of the MP serves as the genotype of an unidentified body. We assume that no error occurs in genotyping, and thus, one element of β_{ i } is 1 with all the others 0. A vector μ_{ j }, which is to be estimated from the reference DNA, was computed by an algorithm and program implemented by the authors [12].
We calculated a matrix P between all typed individuals in each family and each body, not between each typed individual and each body. Throughout, we assumed that only kinship testing is employed, because direct matching has almost absolute discriminatory power such as 1 in 2.5 × 10^{17}[13]. LR may not be sufficiently high to determine the identity when kinship testing is performed using a reference DNA that provides no obligate alleles or impossible alleles. Obligate alleles are those that an individual must carry and are revealed by the genotypes of his/her parents. Impossible alleles are those that cannot be carried without mutations; they are revealed, for example, when a child has an allele that is carried by none of his/her parents. In other words, therefore, these references that were considered in our simulation do not reveal Mendelian error in any genotypes. We used 10 status patterns that provide neither obligatory alleles nor impossible alleles for MPs. Otherwise, both the LR and the permanent method were highly accurate with little difference between them. Columns headed “Family types” in Table 1 list these 10 types of families. The position of each individual in the pedigree tree is expressed as a combination of generation and a number shown below each symbol in Figure 2 (e.g., I2 indicates a person numbered 2 in generation I).
Data for ROC analyses
For comparison with the LR method, we assumed a complete dataset, i.e., n_{ w } = n_{ b } = n_{ m }. Complete data are required to compare two methods under the same condition because the LR method does not take unavailable data into account. We generated 20 datasets, each of which consisted of 20 families randomly drawn from the pooled families. Uniform data were generated by assigning the same status pattern to all 400 families, 20 datasets × 20 families. Because we used 10 types of pedigrees, there were 200 datasets, 10 types × 20 sets, in total. We used the same 400 families for all family types, so that performance could be compared using a set of families derived from the same set of founders. For comparison in a more realistic situation, 20 mixed datasets were generated by assigning status patterns randomly chosen from the 10 types with an equal probability to the same 400 families. The same sets of families were used again for the same reason. To summarize the results of 20 datasets, we pooled 8,000 (20 datasets × 20 bodies × 20 families) values of matrix Q obtained from the 20 datasets, and a ROC curve was drawn using all the values. We describe the AUC obtained from this ROC curve as pooled AUC. Results for uniform datasets are summarized in Table 1 for each family type and shown with mean AUC (standard deviation (SD)), pooled AUC (95% CI), and P values of the DeLong test for the alternative hypothesis that the pooled AUC of the permanent method is greater than that of the LR method. Similarly for the mixed data, the results were summarized with mean AUC (SD), pooled AUC of all 20 datasets, and P values (Table 1).
For assessing the influence of unavailable data, 20 families were randomly drawn from the pooled families. The bodies of MPs of these families were assumed to have been recovered, and we call this data complete part. Appropriate numbers of bodies and/or families were randomly drawn and added to the data. These additional bodies or families did not have identical counterparts, and we call this data additional part. Data for the additional part were drawn 10 times to obtain 10 datasets, each with the same complete part. We simulated three types of situations: 1) n_{ b } < n_{ m } = n_{ w }, 2) n_{ m } < n_{ b } = n_{ w }, and 3) n_{ b }, n_{ m } < n_{ w }. For situation 1, family data were added to the complete part, and body data that were unavailable were completed using γ_{*}. For situation 2, body data were added, and family data that were unavailable were completed via γ_{*}. For situation 3, equal numbers of families and bodies were added, and the remainder were completed with γ_{*}. A matrix Q was calculated using the entire dataset, and only values of the complete part were used to draw the ROC curve. This is because the aim of this simulation was to compare performance using complete data to demonstrate how the amount of unavailable data influences the performance. The reason for this is as follows. The number of identical pairs depends on the size of the complete part, i.e., 20 in our simulation. On the other hand, the number of nonidentical pairs increases according to the size of the incomplete part, because all entries in the incomplete part are incorrect matches. Therefore, the number of nonidentical pairs differs between a complete dataset and an incomplete dataset. This means that the influence of one incorrect assignment on specificity differs between the complete dataset and the incomplete dataset; the influence is weaker in the incomplete dataset because the number of nonidentical pairs is larger in the incomplete dataset than that in the complete dataset. Therefore, the performance would be naturally different when all entries in both complete and incomplete parts are taken into account. However, we are interested in how the performance is changed by the presence of the incomplete part, compared to the performance in the complete dataset. Therefore, we performed a ROC analysis for the complete part in the incomplete dataset. On the other hand, the results of sensitivity and specificity at the threshold v = 0.999 were obtained from all the entries in both the complete and incomplete parts.
Approximation of the permanent of the square matrix
We employed an algorithm to approximate a permanent of a nonnegative square matrix, as described by Huber and Law [6]. Briefly, two parameters, δ and ϵ, define the accuracy of the approximation as follows: For any ϵ ≥ 0 and δ ∈ (0, 1], the permanent of a square matrix with arbitrary nonnegative entries can be approximated within a factor of 1+ δ with probability at least 1–ϵ. That is, δ gives the lower and upper limits of the estimate, and ϵ gives the probability that the estimate falls within the limits. The algorithm is based on a sequential acceptance/rejection procedure, in which the ratio of the number of accepted trials (the acceptance number) to the total number of trials is used to approximate a permanent. The two accuracy parameters define the acceptance number as k = 14δ^{–2}ln(2/ϵ). Because both small δ and ϵ increase k, a large acceptance number means that accuracy is high. The sequential acceptance/rejection iteration process ends when a preset k is achieved. We used δ = 0.5 and ϵ = 0.0001, and thus k = 555, when not specified. We chose these values to maximize accuracy within a practical computational cost range. We verify these values next under “Assessment of approximation accuracy and computation time” and in the Results under “Accuracy parameters of approximation and computation time”.
Assessment of approximation accuracy and computation time
To assess accuracy and computation time in our case, test matrices were generated using values obtained from a matrix P of the mixed datasets used in the ROC analyses. To obtain practical matrices, each matrix contained exactly one value for an identical pair in each row and column, and values for nonidentical pairs in the remaining elements. We prepared matrices of sizes 20 and 30 for assessing accuracy, and matrices of sizes of 10, 20, 30, and 40 for assessing computation time. We calculated the permanents of these matrices 100 times each, with combinations of two accuracy parameters, δ = (0.5, 1) and ϵ = (0.1, 0.01, 0.001, 0.0001, 0.00001). For δ = 0.5 and ϵ = 0.0001, we also simulated a worsteffect scenario, in which approximation errors affect the test result in the worst manner. The worsteffect scenario is defined as that in which, for a body, ${q}_{\mathit{ij}}^{\text{'}}$ for an identical pair is approximated to be equal to the lower limit, and in which ${q}_{\mathit{ij}}^{\text{'}}$ of the most probable pair among nonidentical pairs in the same row for the body in matrix Q′ is approximated to be equal to the upper limit. Because ϵ = 0.0001 is sufficiently low, we consider that no more than one nonidentical pair in each row of matrix Q′ is approximated to be equal to the upper limit. Here, the most probable nonidentical pair is not the result of an approximated permanent, but is obtained if it were possible to exactly compute the permanent. Therefore, in the worsteffect scenario, the probability of the most probable nonidentical pair is overestimated, the probability of the identical pair is underestimated, and thus these errors maximize the influence on specificity and sensitivity. We assessed the performance of the permanent method for a case in which this worst effect of errors occurred to all bodies in a dataset. We assumed that the matrix Q′ of the same mixed dataset used in the ROC analysis was obtained from an exact computation of permanents, and we generated matrix Q′ for the worst case from the assumed exact results, and computed matrix Q with the same method as described in the Methods under “Matrix Q”.
Mean computation time was obtained by approximating three times for each matrix size and for the same combinations of parameters.
Computation environment and software
Genotype simulation, computation of conditional probabilities and LRs, evaluation of performance, and assessment of accuracy and processing time were performed with R v2.13.1 [14]. The computation environment was as follows: Vostro 260S (Dell), 64bit Windows 7 operating system, Intel Core i52400 CPU (3.10 GHz). We computed a probability distribution of genotypes of a MP, μ_{ j }, a matrix P and LR using an algorithm proposed and implemented by the present authors [12] to compute kinship applicable to general forms of pedigrees, including large extended families. ROC analysis was performed with the pROC package [15] in R.
Results
Comparison of permanent method and LRbased method
First, we assessed performance on uniform datasets to compare performance for each family type. Figure 3A–C shows distributions of probabilities obtained with the permanent method and posterior odds obtained with the LR method for uniform datasets of family types 1, 4, and 10 respectively, corresponding to family type in Table 1.
The histograms indicate that the distributions of identical pairs and nonidentical pairs overlap less when family types 1 and 4 were tested with the permanent method as compared with the LR method. Notably, when tested with the permanent method, the distributions of identical pairs and nonidentical pairs show no overlap for family type 1, or sibship tests. Therefore, the two distributions are more clearly separated when the permanent method is used. For family type 10, or betweencousins test, however, the distributions do not differ much between the two methods. Figure 4A–C shows ROC curves of pooled results of 20 uniform datasets for the same three family types. For all the three types, the permanent method significantly outperformed the LR method in terms of AUC (P = 1.6E05, 2.1E13, and 2.5E08 for type 1, 4, and 10, respectively). Figure S1 and S2 show histograms and ROC curves for other family types, and Table S1 lists statistics for each dataset (see Additional file 1: Figure S1, Additional file 2: Figure S2, Additional file 3: Table S1). The permanent method exhibited significantly better performance than the LR method for all family types in terms of AUC. When the threshold value was set to 0.999, a realistic threshold in practice, the permanent method exhibited higher sensitivity than the LR method for all family except type 10, while specificity was 1 with both methods for all family types.
Next, we assessed performance on mixed datasets for comparison in a more practical situation. Counts of simulated family types included in each dataset are listed in Table S2 (see Additional file 3: Table S2). Figure 3D shows the distributions of identical pairs and nonidentical pairs. The overlapping area is smaller when the permanent method is used. Figure 4D shows ROC curves for pooled results of 20 mixed datasets. Table 1 and Table S3 show statistics for the pooled results of all datasets and the result of each dataset, respectively (see Additional file 3: Table S3). The permanent method exhibited significantly better performance (AUC: 0.990 vs. 0.959, P = 9.6E15). This suggests that the permanent method should be available for use in practical situations. In a practical situation, achieving high specificity is considered to be more important than optimizing both sensitivity and specificity. In this paper, we use v = 0.999, which means that, when the probability of a body and missingperson pair exceeds 0.999, the sum of probabilities of the other pairs in the same column or row is 0.001, and thus we consider 0.999 to be a highly reliable threshold (our choice of the threshold value is discussed further in the Discussion). With v = 0.999, the permanent method exhibited higher sensitivity than the LR method (0.188 vs. 0.055), while specificity was 1 with both methods (Table 1, Figure 5A, Figure 5B).
ROC analysis indicated some difference between the two methods in terms of overall performance. We also compared differences in the judgment of each pair. Figure 6 shows the distributions of values of the permanent method and LR method. Dashed lines indicate the level of v = 0.999, with the four regions divided by the dashed lines demonstrating the following judgments:

top right: assigned by both methods,

top left: assigned by LR method and suspended by permanent method,

bottom left: suspended by both methods, and

bottom right: assigned by permanent method but suspended by LR method.
Topright and bottomleft regions indicate that the judgments of the two methods were consistent, while the topleft and bottomright regions indicate that the judgments between the two methods differed. Table 2 shows pair counts classified into each region at threshold values 0.999 and 0.9999. At the level of 0.999, neither method assigned any nonidentical pairs. The permanent method assigned 75 truly identical pairs, among which the LR method failed to assign 53 pairs. The LR method assigned 22 truly identical pairs, all of which were assigned by the permanent method. At the level of 0.9999, the permanent method assigned 43 truly identical pairs, among which the LR method failed to assign 31 pairs. The LR method assigned 12 pairs, all of which were assigned by the permanent method. Figure 5A and Figure 5B plot specificity and sensitivity, respectively, at various values of v. Specificity curves are similar with both methods, and the permanent method and LR method reached the specificity of 1 at 0.993 and 0.952, respectively. On the other hand, sensitivity differed between the two methods, with the permanent method exhibiting a higher sensitivity at any value of v. These results suggest that, at the level of 0.999 or higher, both methods safely discriminate nonidentical pairs, and the permanent method demonstrates higher sensitivity. As shown in Figure 3D, Figure 6 also indicates that the test values of nonidentical pairs are distributed more widely when the permanent method is used.
Demonstration of how the permanent method performs for incomplete data
Here we demonstrate the performance of the permanent method on incomplete data compared with complete data. Table S4 lists the counts of family types drawn by random sampling in each dataset (see Additional file 3: Table S4). Table 3 lists the mean decrease in AUCs resulting from the complete part of incomplete datasets from AUC of the complete dataset in each situation and each number of additional bodies and/or families. Statistics of each dataset are listed in Table S5 (see Additional file 3: Table S5). Sensitivity and specificity at the level of 0.999 were calculated from all the entries in both complete and incomplete parts, and are listed in Table 3. Note that all the complete parts of the incomplete datasets and the complete dataset include the identical set of bodies and reference families, but conditional probabilities resulting from the permanent method differ because of variation in the additional part. Although the number of total victims is limited in our simulation because of computation time constraints, discussed in the next section, as the amount of additionalpart data increased, mean decreases from complete data increased, and sensitivity at v = 0.999 showed a tendency to decrease. Even with 10 additional bodies and/or reference families, which means that about 30% of the data are missing, the permanent method still exhibited equal or higher sensitivity at 0.999 and higher AUC compared with the LR results for the corresponding complete data.
Accuracy parameters of approximation and computation time
We investigated the distribution of the estimates of the permanent in the setting of DNA identification. Figure 7A shows the distribution of estimates obtained from 100 estimates of permanents plotted against the acceptance number, k. Table S6 shows values of δ and ϵ and their corresponding values of k (see Additional file 3: Table S6). Under the condition used to assess performance with the above results (k = 555), the differences between the minimal and the maximal estimates were 0.09 and 0.08 in a log scale for the matrix sizes 20 and 30, respectively, which means that the maximal estimate fell on the minimal estimate multiplied by 10^{0.09} = 1.23 and 10^{0.08} = 1.20, respectively (Figure 7A). When simulating the worst case, we observed a decrease in sensitivity from 0.19, which was the sensitivity obtained from the mixed datasets shown in Table 1, to 0.18 (four pairs assigned in the assumed exact result failed to exceed v = 0.999 in the worst case), and no changes in specificity (1 even in the worst case). Figure S3 shows that the assumed exact test results and the test results in the worst case only differ slightly; test values for identical pairs in the worst case are slightly lower than the corresponding assumed true values, and test values for nonidentical pairs in the worst case are slightly higher than the corresponding assumed true values (see Additional file 4: Figure S3). Figure 7B plots computation time against acceptance number. When the matrix size was 40, computation time increased dramatically, and it took approximately 11,000 seconds = 3 hours to obtain the acceptance number used in the simulation.
Discussion
The individual identification of unidentified bodies is an important issue. Developments in DNAbased techniques have provided strong evidence for personal identification. Despite progress in genotyping, challenging cases are still encountered, such that LR fails to provide strong evidence because kinship between sample and reference DNA itself is not close. If it occurs in a singlefatality case, DNA would not be the firstline evidence. In the case of multiple fatalities, approaches optimized for multiplefatality cases may overcome this problem to some extent. LRbased identification is optimal when comparison is only possible with a random person from a population, i.e., a singlefatality case. In multiplefatality cases, however, comparison can be performed with other DNA obtained from the same case. Currently, identification in multiplefatality cases is based on a conventional LR weighted by prior odds. Although the LR method can take the total number of victims into account via the prior odds, it still compares each pair only with the population where the possible identity of a body can be limited to one of MPs. The new approach we have described in this paper considers the identification problem as an assignment problem and provides the probability for a particular pair of sample and reference DNA conditional on the entire observed data. Because the permanent method considers all assignment hypotheses, it can be reasonably speculated that, given pairs assigned with high confidence, conditional probabilities of nonidentical pairs decrease and those of other identical pairs increase. Thus, we expect the permanent method to be capable of discriminating between identical and nonidentical pairs that cannot be clearly discriminated with LR.
In a separate study, the distribution of combined sibship index (CSI) with 15 STR loci was investigated, and it was found that the distribution of CSI for siblings overlapped with that for random pairs; it was also found that 1.3–1.6% of siblings had CSI less than 1, and 1.4–1.9% of random pairs had CSI greater than 1 [16]. Thus, a sibship testing, or type 1 in our simulation, in multiplefatality cases would be more difficult, where prior odds reduce LR.
The results of our computer simulation show that the performance of the permanent method is significantly higher than that of the LR method. In the case of individual identification, high specificity is required even at the cost of sensitivity, because finding an incorrect identity is a crucial problem. Thus, we also focused on sensitivity and specificity at v = 0.999 and higher, and the permanent method displayed higher performance in this range. However, the choice of threshold value in practice is rather arbitrary depending on the situation and the criteria conventionally accepted in a society. After close discussion with a forensic expert, we decided to use v = 0.999 and v = 0.9999 to discuss our results. For other threshold values, Figure 5A and Figure 5B show sensitivity and specificity according to the values of threshold in our simulation. Uniform datasets were tested to assess performance in terms of difficulty levels for a problem. The three family types, 1, 4, and 10, were expected to represent levels of relatively easy, difficult, and very difficult, respectively. Because type 1 corresponds to a sibship testing, the fact that the permanent method resulted in no overlap between distributions of identical pairs and nonidentical pairs, and that nonidentical pairs showed distributions with larger variance, indicate well the greater discriminating power of the permanent method (Figure 3A).
In this paper, we used only cases in which references provide neither obligate alleles nor impossible alleles; that is, any genotypes are consistent without assuming mutation. This is because the current genotyping system provides a discriminating power strong enough to overcome the theoretical problem in the LR method, and both methods perform highly with almost no difference between them (data not shown). Thus, the power of the permanent method may be maintained up to some extent, but not as far as between cousins (although the ROC analysis indicated that the permanent method performed better in terms of AUC, the sensitivity and specificity at the practical threshold did not differ between the two methods). In practice, it is clear that we should attempt to collect reference DNA from relatives who are as closely related as possible, in order to obtain obligate alleles. If this is not possible, it is important to consider whether references provide impossible alleles. Consider a situation in which the DNA of the parents of a MP is not available but that of all four grandparents is available, but for financial reasons we can only type two persons. Two grandparents on the same side may provide impossible alleles, but two from each side do not. In our simulation, these two situations led to a substantial difference in results; when two grandparents on the same side were used, both the permanent and the LR method were more powerful in determining the identities (data not shown).
Importantly, the permanent method can take uncertainty into account. Two situations lead to uncertainty: the presence of unavailable data, and the number of unknown victims. As simulated in this paper, the former can be considered by the permanent method. Although the latter is not demonstrated, the same strategy can be applied: We would set an expected total number of victims and complete the unavailable data with γ_{*}. As we do not know the exact number of victims, the expected number would be arbitrary to some extent. Judgment in the LR method does not change no matter how much unavailable data exist, whereas assignment by the permanent method becomes more difficult with more unavailable data. Therefore, the test value of the permanent method appears to reflect our assignment confidence better than that of the LR method.
Although the theoretical basis of our approach is reasonable and adequate, the approach is computationally challenging, because the permanent method requires estimation of the permanents n_{ w } × n_{ w } times. Computation time can be the primary cause of limitation. The most efficient algorithm for the exact permanent requires Θ(n2^{n}) arithmetic operation. It has been proven that exact computation of the permanent is a #Pcomplete problem, even for 0,1 matrices, and thus computation in polynomial time is not possible [3]. Since then, researchers have focused on approximation algorithms. Currently, to the best of our knowledge, the approximation approach described by Huber and Law [6] is the fastest algorithm when the problems become dense. Huber and Law achieved O(n^{4} log n) expected running time, though there is no a priori bound on the running time for matrices with arbitrary nonnegative entries. In our study, we implemented the Huber–Law methodology to simulate our approach. In our simulation environment, we experienced a dramatic increase in computation time when the matrix size was 40. Therefore, the permanent method is currently practically applicable to the assignment of relatively small numbers of victims in a closed incident. However, we are hopeful about the possibility of extending the application, even to large open incidents, because of advances in two areas: (1) approximation algorithms and combination theory, and (2) computer performance. On the first point, approximation algorithms for the permanent of a matrix are a vigorously studied field of mathematics. Recent improvements include those of Jerrum et al. (2004) [7], Bezáková et al. (2006) [4], and Huber and Law (2008) [6]; the computational complexities of these methods are O(n^{10} (logn)^{3}), O(n^{7} (logn)^{4}), and O(n^{4} log n), respectively. These dramatic decreases in complexity have been experienced only in the last ten years, and further developments can therefore be expected. Second, we expect that the permanent method will enjoy more extended application in the near future because computer performance is continually improving. For these reasons, we believe that there is merit in our new idea for DNA identification in multiplefatality cases, even with current computational limitations.
Because computation of the permanent method depends on approximation algorithms of a matrix, we must describe the algorithmspecific consideration. The Huber–Law algorithm defines two accuracy parameters δ and ϵ. The acceptance number k, given by δ and ϵ, only influences the computation process. This means that various sets of δ and ϵ that give the same k are mutually interchangeable. However, here we suggest using values of δ and ϵ to better understand the accuracy. Smaller values of δ and ϵ give a more accurate approximation. We verified that δ = 0.5 and ϵ = 0.0001 can give sufficiently accurate test results for our simulation condition, and thus we suggest using δ ≤ 0.5 and ϵ ≤ 0.0001.
To reduce computation time, a Monte Carlo method can be applied where the total number of iterations is set instead of k. However, with a Monte Carlo method, we must take care how many accepted trials are obtained after the iteration. In the case of DNA identification, we found lower probabilities in obtaining one accepted trial as the matrix size increased. In this case, the Monte Carlo method may result in small values of k, and estimates may become highly unreliable, as shown in Figure 7A. Therefore, the Monte Carlo method is not recommended for practical DNA identification problems.
Conclusions
The permanent method provides further evidence for identification in terms of conditional probability. We have shown that this method is capable of detecting identical pairs of low LR and is highly robust in terms of specificity. With two methods used in combination, DNAbased identification may exhibit higher performance. It is also important that the permanent method is capable of weighting the presence of unavailable data, unlike the LR method. Currently, the permanent method is computationally limited to relatively small datasets obtained from closed incidents.
References
 1.
Olaisen B, Stenersen M, Mevåg B: Identification by DNA analysis of the victims of the August 1996 Spitsbergen civil aircraft disaster. Nat Genet. 1997, 15 (4): 402405. 10.1038/ng0497402.
 2.
Lin TH, Myers EW, Xing EP: Interpreting anonymous DNA samples from mass disasters–probabilistic forensic inference using genetic markers. Bioinformatics. 2006, 22 (14): e298e306. 10.1093/bioinformatics/btl200.
 3.
Valiant L: The complexity of computing the permanent. Theor Comput Sci. 1979, 8 (2): 189201. 10.1016/03043975(79)900446.
 4.
Bezáková I, Stefankovič D, Vazirani VV, Vigoda E: Accelerating Simulated Annealing for the Permanent and Combinatorial Counting Problems. Annual ACMSIAM Symposium on Discrete Algorithms (SODA). 2006, Miami: ACM Press, 900907. 2224 January 2006
 5.
Broder AZ: How hard is it to marry at random? (On the approximation of the permanent). Proceedings of the eighteenth annual ACM symposium on Theory of computing: 11/01/1986. 1986, Berkeley: ACM, 5058. 2830 May
 6.
Huber M, Law J: Fast approximation of the permanent for very dense problems. Proceedings of the nineteenth annual ACMSIAM symposium on Discrete algorithms: 01/20/2008. 2008, San Francisco: SIAM, 681689. 2022 January
 7.
Jerrum M, Sinclair A, Vigoda E: A polynomialtime approximation algorithm for the permanent of a matrix with nonnegative entries. J ACM (JACM). 2004, 51 (4): 671697. 10.1145/1008731.1008738.
 8.
Weisstein EW: "Permanent." From MathWorld  A Wolfram Web Resource. http://mathworld.wolfram.com/Permanent.html,
 9.
Swets JA: Measuring the accuracy of diagnostic systems. Science. 1988, 240 (4857): 12851293. 10.1126/science.3287615.
 10.
DeLong ER, DeLong DM, ClarkePearson DL: Comparing the areas under two or more correlated receiver operating characteristic curves: a nonparametric approach. Biometrics. 1988, 44 (3): 837845. 10.2307/2531595.
 11.
Yoshida K, Takahashi K, Kasai K: Allele frequencies of 15 loci using AmpFlSTR Identifiler Kit in Japanese population. J Forensic Sci. 2005, 50 (3): 718719.
 12.
Narahara M, Tamaki K, Yamada R: Application of genetic descent graphs to computation of probabilities for complex pedigrees. DNA Polymorphism. 2012, 20: 278285.
 13.
Butler JM: Forensic DNA testing. Cold Spring Harb Protoc. 2011, 2011 (12): 14381450.
 14.
Team RDC: R: A Language and Environment for Statistical Computing. In. 2011, Vienna, Austria: R Foundation for Statistical Computing
 15.
Robin X, Turck N, Hainard A, Tiberti N, Lisacek F, Sanchez JC, Muller M: pROC: an opensource package for R and S + to analyze and compare ROC curves. BMC Bioinforma. 2011, 12: 7710.1186/147121051277. England
 16.
Pu CE, Linacre A: Systematic evaluation of sensitivity and specificity of sibship determination by using 15 STR loci. J Forensic Leg Med. 2008, 15: 329334. 10.1016/j.jflm.2007.12.018. England
Acknowledgements
We started this study upon the historical disaster, the 2011 earthquake off the Pacific coast of Tōhoku with tsunami. We pray for the victims and the recovery from the destruction.
Author information
Additional information
Competing interests
The authors declare that they have no competing interests.
Authors’ contributions
MN participated in the design of the study, wrote the program, carried out the computersimulation and the statistical analysis, and drafted the manuscript. KT conceived of the study, and supervised the forensic aspect of the study. RY developed the algorithm, wrote the program, designed the study, and helped to draft the manuscript. All authors read and approved the final manuscript.
Electronic supplementary material
Additional file 1: Distributions of conditional probabilities of permanent method and posterior odds of LR method. Distributions of identical pairs and nonidentical pairs are shown in red and blue, respectively. Probabilities obtained with the permanent method are shown as odds. Values are obtained from 20 uniform datasets of family types 2, 3, 5, 6, 7, 8, or 9. Family types are defined in Table 1. (PDF 1 MB)
Additional file 2: ROC curves of pooled results obtained from 20 datasets for each family type. ROC curves of test results for family types 2, 3, 5, 6, 7, 8, and 9 are shown. Discriminant performance was compared between the permanent method (solid line) and the LR method (dashed line). AUC (95% confidence interval (CI)) and P values of the DeLong test are shown. Family types are defined in Table 1. (PDF 1016 KB)
Rights and permissions
About this article
Received
Accepted
Published
DOI
Keywords
 DNA polymorphism
 DNAbased identification
 Multiplefatality cases
 Permanent of square matrix
 Assignment problem