1. Introduction
Let
represent the book amount of the ith item in an account, with the total population book amount denoted as
. At regular intervals, an auditor samples
n line items from the account and to compare against their correct values. Therefore, let
denote the audited amount for the ith line item, and let
denote the error amount. Note that the total book amount is known to the auditor. The fundamental issue is constructing confidence limits for means or totals in finite populations where the underlying distribution is highly skewed and contains a substantial proportion of zero values. This situation is often encountered in statistical applications such as statistical auditing, reliability assessments, and insurance. The most distinctive feature of accounting data is the large proportion of line items without error, while an audit sample may not yield any nonzero error amounts. For data with mostly zero observations, classical interval estimation based on asymptotic normality is not reliable. In auditing practice, auditors are often more interested in obtaining lower or upper confidence limits than in obtaining two-sided confidence intervals. Specifically, independent public accountants are focused on estimating the lower confidence bound for the total audited amount to avoid overestimation and potential legal liability. Stringer and Kaplan (see [
1,
2]) have demonstrated that accounting populations are highly positively skewed, with considerable diversity in the characteristics of error amounts across subsystems. There are several distributions that also exhibit the same form of the distribution observed in accounting populations. These include the Gamma, Log-normal, Weibull, and Beta distributions. The error rates are usually very low, which render many existing statistical procedures inappropriate for estimating and hypothesis testing of error rates and error amounts. There are two main types of audit tests where statistical sampling is advantageous. The first is a compliance test, used to determine the procedural error rate in a population of transactions. The second is a substantive test of details, aimed at evaluating the aggregate monetary error in a stated balance. Inferences about the total error amount are usually made using confidence intervals, which are closely related to hypothesis testing. The decision-making process in auditing is treated as a problem of testing statistical hypotheses about admissibility of the total or the mean accounting errors. This approach allows auditors to control both the significance level (risk of incorrect rejection) and the type II error probability (risk of incorrect acceptance). Substantive tests of details verify the accuracy of recorded monetary values in financial statements, providing direct evidence regarding the accuracy of the total recorded values. Auditors may apply substantive tests extensively or use compliance tests to evaluate the efficiency and effectiveness of internal controls in mitigating material errors. In compliance tests, the variable of interest is the error rate (proportion of transactions for which the internal control operates wrongly). Samples of transactions are used to infer this error rate.
2. A Mixture of Probability Distributions as a Model for Generating Accounting Values
Wywiał in [
3] proposed the following model. Let
U be the population of accounting documents of size
N with given accounting totals. Some of the documents contain errors. In the population
U, there are given accounting totals (values)
for each element
. Let
be the observation of a random vector
. We denote the true book values (without errors) by
,
and let
be the random vector observation
. Vector of accounting values contaminated by errors
will be an observation of the random vector
. Finally, let
, where
if
.
In practice, all values of X are known before auditing process. Observations x of X are treated as a specific auxiliary data. Auditing process leads to observation of values , and , . Let , , . Their values will be denoted , , .
Let an auditor arbitrarily select a sample
s of size
n from
U. Hence,
is the subvector of
X,
. The random vector
is observed in
s where the objects are controlled. After the auditing process, the sample
s is split into two disjoint sub-samples
and
where
. The set
is of size
and the set
is of size
. In the sub-sample
, there are observed accounting amounts without errors. Before the auditing process, we have observations of the following data:
where
After the auditing process, we have observations of the following data:
Values
T,
,
X,
,
,
and
are denoted, respectively, by
t,
,
x,
,
,
and
. In the following work, we assume that
and
.
Let be the expected mean accounting error. Audit purpose is inference on or on the expected total accounting error .
Let
be the probability distribution function of the random variable
Y, whose values are true accounting and
where
is the parameter space. The distribution function of
W is denoted by
, where
. Moreover, let
. We assume that an accounting errors appears with probability
p. We can write
when an accounting error occurs
and
when it does not occur
. According to the well-known total probability theorem we have:
and finally
where
and
is the parameter space. Hence, the probability distribution of the observed accounting amounts is a mixture of the distribution function
) of the true amounts and the distribution function
) of the amounts contaminated by errors. When the random variables
Y and
W are continuous, by differentiating both sides of Equation (
1) we have
Therefore, the probability density of the observed accounting amounts is a mixture of density
) of the true amounts and density
) of the amounts contaminated by errors. Let
R and
Y be independent and
R is the accounting error. Hence
,
,
. The basic moments of the random variable
X are:
2.1. A Mixture of Gamma Probability Distributions as a Model for Generating Accounting Values
The well-known gamma probability distribution we denote by
, where parameters
and
are called scale and shape parameters. The shape of gamma density distribution does not depend on the scale parameter because its skewness and kurtosis coefficients are equal to
and
, respectively. Wywiał in [
3] considered the model based on a mixture of gamma distributions. Let
and
be independent random variables. The advantage of this model is that the density function for the sum of gamma distributions can be determined. Based on the above assumption, the random variable
. Using the previous considerations, we obtain
where
From Formulas (
3) and (
4) we obtain
For more on the use of gamma decomposition to model accounting values, see the articles [
4,
5]. The book amounts are treated as values of a random variable which distribution is a mixture of the distributions of correct amount and the distribution of the true amount contaminated by error. Distributions of correct amount and true amount contaminated by error are right-skewed because small book amounts are more frequent than large book amounts. It is convenient to assume that the book values are additive function of true accounting amounts and accounting errors. Hence, we can expect that the above proposed quite simple model describes accounting data well.
2.2. A Mixture of Poisson Probability Distributions as a Model for Generating Accounting Values
We assume that the variable
(true values) and
(accounting error), where
is a Poisson distribution with probability function
If variables
Y and
R are independent, then the random variable
.
From Formulas (
3) and (
4) we obtain
3. Sequential Tests Based on the Ratio of the Likelihood Function
The set of issues involved in proceeding sequentially in verifying statistical hypotheses is called sequential analysis (see [
6]). Sequential analysis was created by Wald (see [
7]). The sequential approach to statistical inference was the subject of systematic research during the Second World War. This research was concerned with the quality of munitions and was conducted by the Statistical Research Group at Columbia University. Data in sequential analysis are used both to decide when to end an observation (data collection) and to draw actual conclusions (regarding the parameter being estimated or hypotheses being tested). The sequential method is applicable in auditing when assessing the internal control system in compliance testing when verifying hypotheses regarding the proportion of accounting errors. In their doctoral thesis, Byekwaso in [
8] used a polynomial Dirichlet model to estimate and test hypotheses about error rates in accounting populations using Bayesian methods with sequential stopping rules.
3.1. Sequential Ratio Test
Let be the density (probability) function with , where is subject to verification, and is the set of parameters indifferent from the point of view of hypothesis testing against . We propose a hypothesis . Hypothesis is verified by comparing with the Hypothesis , where is some specific value different from , and , . During the Sequential Probability Ratio Test (SPRT), we progressively sample one or more items at each stage of the sequential hypothesis verification. Depending on the predetermined errors of the and , the numbers and satisfying the condition . The elements are drawn into a simple sample. For the first drawn observation , let us determine the value of the ratio . If
then we reject the Hypothesis in favor of the hypothesis .
then we accept the Hypothesis .
then we select the next element to sample.
In sequential proceedings, it is more convenient to operate with logarithms of numbers
A and
B and random variables
If in the sample we have
m elements
then if
then we reject the Hypothesis in favour of the hypothesis ;
then we accept the Hypothesis ;
then we select the next element to sample.
Let
X be a random variable with parameter
and of density
. We consider the following simple hypotheses about the value of the parameter
The verification of the null hypothesis (
14) consists of calculating on the basis of an
n elementary random sample
the value of the statistic:
and comparing it with the designated constants
A and
B. The SPRT test can be extended to versions in which the unknown parameters are replaced by their maximum-likelihood estimators ([
9]) and then the previously described sequential procedure is applied
where
,
.
3.2. The Expected Sample Size
The approximate expected value of the sample size for a sequential ratio test is given by the formula ([
6]):
where:
,
is the OC function of the sequential ratio test. The OC function describes the probability of accepting the Hypothesis
([
10]). The functions
are written with the formula
Function
for a variable of
x of continuous type satisfies the following condition
Function
for a variable of
x of discrete type satisfies the following condition
The expected value
is determined from the formula
We use the above expressions to determine the expected sample size for tests on the mean value of accounting errors. We test the hypothesis
where
denotes the mean value of the accounting errors determined from the mixture of Poisson or gamma distribution.
The expected value
is determined by the formula:
where:
and
The expected value
is determined by the formula:
The expected value
for a mixture of gamma distributions is determined from the formula:
The expected value
is determined from the formula:
where:
and
For fixed parameters a, c, , p expected values i can be calculated by suitable numerical integration methods using, for example, Mathematica.
3.3. Simulation-Based Sample Size Determination
The purpose of the simulation study was to determine the average sample size
and simulation probabilities of accepting and rejecting the Hypothesis
. The determination of the average sample size using the Monte Carlo method was dealt with by Boiroju (see [
11]). The purpose of an audit may be to certify the accuracy of a financial settlement, in which case a situation arises where the auditor focuses more on minimizing risk
rather than
. Therefore,
was estimated under the assumption that the observed data in the sample were generated under the assumption that the Hypothesis
is true.
- 1.
For fixed values of and we determine and . To the variables , j, k are assigned the values , .
- 2.
For we generate a dataset of size from a mixture of Poisson distributions with the following parameters: , . The parameter b is determined from the formula .
- 3.
We draw a sample size , . We divide the population into subsets , , .
- 4.
Estimate the mixture parameters and provided based on the credibility function. For the parameters obtained, we calculate the logarithm of the credibility function .
- 5.
Estimate the mixture parameters and provided based on the credibility function. For the parameters obtained, we calculate the logarithm of the credibility function .
- 6.
Let us calculate the value of the test statistic .
- 7.
Repeat steps 3 to 6 until and .
- 8.
If then , which results in the rejection of the Hypothesis in favour of the Hypothesis .
- 9.
If then , which means that we reject the Hypothesis even though the hypothesis is true.
- 10.
For each i we determine the number of sample elements necessary to decide whether to accept or reject the hypothesis . We stop the algorithm when .
- 11.
Determine the mean sample size , the standard deviation of the simulated sample sizes , the simulated probability of accepting the Hypothesis , simulation probability of rejecting the Hypothesis and simulation probability of terminating a sequential procedure below a threshold .
For a mixture of Poisson distributions, the following hypothesis was tested:
Details of how the algorithm works are shown in
Figure 1.
The function uses the Newton–Raphson algorithm and is included in the R package. Simulations were performed for three mixing values , , for a mixture of Poisson distribution. The corresponding parameters were determined for the mixing parameter .
Table 1,
Table 2,
Table 3,
Table 4,
Table 5 and
Table 6 show the results
obtained from Formulas (
24)–(
27) for a mixture of Poisson distributions. The values of
have been determined in Mathematica. For a mixture of gamma distributions, the following hypothesis was tested:
Parameters were used to generate a mixture of gamma distributions:
,
. For the mixing parameters
,
,
the corresponding parameters were determined
. The results of the simulation studies are presented in
Table 7,
Table 8,
Table 9,
Table 10,
Table 11 and
Table 12.
The average sample size decreases with increasing . The smallest sample sizes are obtained for data from a mixture of Poisson distributions generated with the parameter . For and for the mean sample value is 149 (3.7% of the population size). For this parameter, the arithmetic mean of the contributing error values is the largest. Similarly for the mean sample value is 307 (7.7% of the population size) and for the the mean sample value is 502 (12.6% of the population size).
The sample mean values obtained are lower than the values obtained using Formulas (
24)–(
27). Comparing values
and
in
Table 1,
Table 2,
Table 3,
Table 4,
Table 5 and
Table 6 only makes sense if the value of
is close to unity. This means that only then have almost all iterations of simulated sample sizes needed finished before the population size is exhausted. For a fixed value of
, changing the value of
causes negligible changes in the mean of sample size. The simulation probability
of accepting the Hypothesis
in
Table 1,
Table 2,
Table 3,
Table 4,
Table 5 and
Table 6 is smaller than assumed. Reasons for this may be that the critical values of the Wald test are approximated by (
) or the method of drawing, in which we draw 1%
population elements rather than individual elements, i.e., in our case there is a sequential draw of 40 population elements.
Figure 1,
Figure 2,
Figure 3,
Figure 4,
Figure 5 and
Figure 6 show that the average sample size decreases with increasing
. The smallest sample sizes are obtained for data from a mixture of gamma distributions generated with the parameter
. For
and for
the average sample size is 1345 (33.7% population size). For this parameter, the arithmetic mean of the contributing error values
is the largest. Similarly for
sample mean value is 1848 (46.2% population size), and for
average sample value is 2280 (57% population size). For
,
and for
the average sample value is 2419 (60.5% population size). Similarly for
average sample size is 2954 (73.9% population size), and for
average sample size is 3121 (78% population size).
4. Conclusions
This paper presents the application of a sequential test based on the likelihood ratio function for audit studies. The main objective of the simulation studies was to determine the expected sample size.
Section 3.2 outlines formulas for numerically calculating the expected sample size when testing the hypothesis of average accounting errors using a mixture of Poisson distributions as the underlying model for accounting values. The expected sample sizes for the mixture of Poisson distributions were obtained both analytically and through simulation. For the mixture of gamma distributions, the expected sample sizes values were determined exclusively through simulation.
The values of
obtained in
Table 1,
Table 2,
Table 3,
Table 4,
Table 5 and
Table 6 indicate that this sequential test is practical when the mean error size
significantly exceeds
. Otherwise, the expected sample size
becomes impractically large compared to the typical population size of accounting documents. The simulation studies were conducted only for selected hypothetical and alternative parameters (average audit errors), due to the time-consuming nature of the simulations. It is possible that different parameter settings could yield more favorable outcomes in terms of efficiency.