Bio Mathematical Analysis on the Evaluation Scheme for Large Scale Innovation Competition
 ^{1.} Hubei University of Technology, China
Abstract
Largescale innovation contests are becoming increasingly popular globally, with many innovators and enterprises actively participating. Work allocation optimization and fair judging are two significant concerns in innovation contests. This academic paper aims to recommend an optimized” crossassignment” program by leveraging the optimal objective function method to improve the comparability of scores given by various judges. We conducted databased descriptive statistical analysis and concluded that the twostage and weighted evaluation schemes are more beneficial than the traditional judging scheme. Nonetheless, there are still some shortcomings that require addressing. To enhance fairness, we propose an improved twostage evaluation scheme. In the first stage, we normalize the scores with a normal distribution. In the second stage of the process, we implement a system utilizing the Borda sorting technique to categorize submissions into five distinct groups for judges to evaluate based on their perceptions. We also detail a method for weighting tied scores to determine the final rankings. Testing indicates that this approach yields a Normalized Discounted Cumulative Gain (NDCG) of 0.8667, implying greater fairness and precision in the assessment of submissions.
Keywords
• Evaluation program
• Normalization of a normal distribution
• Borda sorting method
• Grey correlation analysis
• Analysis of variance.
CITATION
Zhao B, Jiang X (2023) Bio Mathematical Analysis on the Evaluation Scheme for Large Scale Innovation Competition. Ann Cardiovasc Dis 7(1): 1035.
INTRODUCTION
Largescale innovationbased competitions are an effective means of fostering science, technology, innovation, and entrepreneurship development. They attract innovators from different fields to bring innovative solutions to society. The judging program largely determines the success of a competition. Its fairness and transparency are essential to attract more talented participants. Therefore, to ensure the sustainability of largescale innovation competitions, it is necessary to conduct research and improve the effectiveness of the judging program. There is a lack of a standardized judging mechanism in these competitions. A twostage (online and onsite judging) or threestage (online judging, onsite judging, and defense judging) process is usually used. The critical aspect of this type of competition is innovation. Innovation refers to the ability to perceive what others do not understand. Evaluation of the same work by different experts can lead to divergent opinions, while innovation leads to novel solutions to problems. It is, therefore, essential to develop an unbiased, impartial, and systematic innovation competition selection scheme to ensure credibility and recognition.
The design and improvement of evaluation programs for large scale innovation competitions have long interested scholars and experts in various disciplines, driven by recent rapid advances in science and technology. In their 2006 coauthored publication, Henry Chesbrough, Wim Vanhaverbeke, and Joel West introduced the concept of open innovation from its inception [1], discussed its implications for competitions, provided a framework for evaluating largescale innovation programs, and explored the impact of open innovation for matches. Innovation is the result of discovery, and societal progress results from innovation. The evaluation of innovative ideas has generated much debate, and there are on going efforts to establish unbiased evaluation methods, including the framework developed by Poetz and Schreier in 2012. Their model emphasizes the importance of involving regular users to achieve diverse innovations [2], while considering feasibility and social impact. You Qinggen [3], has developed a userfriendly evaluation index system for experts, which provides a theoretical contribution to improving the current evaluation indexes. In addition, Changchao and Minglong [4], investigated the feasibility of evaluation using AHP hierarchical analysis, grey cluster analysis, and other algorithms in the “Challenge Cup” startup project. In particular, for large innovation competitions, such as joint national and provincial competitions, existing programs are typically based on those developed for smaller competitors, which are not practical to use and tend to be inconsistent in large rounds, leading participants to question the results.
This paper will, therefore, focus on the following. For innovation competitions of significant scale, we offer a thorough evaluation process. Our approach encompasses suggesting an ideal “crossdistribution” format for the blind judging phase, refining the calculation of the standard scores to augment the impartiality of the verdicts, and producing an appropriate evaluative framework for resolving contentious submissions.
There are both practical and theoretical implications to our research. These models theoretically improve the current evaluation criteria and suggest a methodology and confidence level for the evaluators assessing the projects submitted to the competition. We have advanced relevant theoretical research by identifying, raising, analyzing, and resolving issues and integrating them into a coherent theoretical research framework based on the work of innovation competitions.
DATA
The data utilized in this paper was provided by a largescale competition based on innovation. Technical term abbreviations, when used, were explained. The language used was clear, objective, and valueneutral, with a formal register. Biased phrasing was avoided. It was divided into three copies, each undergoing two stages of judging. Five experts assessed the entries in the first stage and generate draw and standard scores. The structure of the paper followed the conventional academic sections and maintained consistent author and institution formatting. The text was precise, free from grammatical, spelling, and punctuation errors, and presented a logical flow of information with causal connections between statements. In the second stage, another panel of three experts reviewed the entries, generating raw and standard scores and a concordance score. In the initial phase, the mean scores of the five experts were calculated, and the pieces positioned within the top 16% of all teams were admitted to the subsequent evaluation stage. After reassessing the standard scores and making appropriate adjustments to the standard scores of a few works with significant differences, the standard scores of the five experts in the first stage and the standard scores of the three experts in the second stage will be averaged into four scores. The resulting scores will then be ranked based on the final total scores to establish the ranking of the works. The dataset comprises 3,000 teams and 125 experts. Each piece of work was randomly assigned to five experts in the first stage, while three experts were given in the second stage. The experts worked independently without interacting with one another throughout the process.
RELATED WORK AND METHODS
Preliminary work
Data Cleaning: (1) Processing of missing values: For experts who did not rate some of the data points, we remove individual samples with missing values. (2) Processing of outliers: Individual samples with outliers are removed from the data. For example, we remove ratings with significant extreme differences in works that did not participate in the second evaluation or that did not win the prize in the second evaluation. We convert the data types to ensure the consistency of our analysis. (3) Conversion of data types: Convert different styles in the data to the correct data type. For example, “first prize” in the data should be converted to 1, “second prize” to 2, “third prize” to 3 and” did not win” to 0. Consistency should be maintained throughout the processing. (4) Processing for consistency: Only papers that reached the second stage were considered to allow a comparative analysis of the two locations. (5) Preprocessing stage: The raw data were technically processed to extract the ten indicators necessary for the model covered by the dataset to establish the evaluation model for this paper. ? the raw score of the first expert in the first evaluation; ? the raw score of the second expert in the first evaluation; ? the raw score of the third expert in the first evaluation; ? the raw score of the fourth expert in the first evaluation; ? the raw score of the fifth expert in the first evaluation; ?the raw score of the fourth expert in the first evaluation; ? the raw score of the fifth expert in the first evaluation; ? The raw score of the first expert in the second evaluation; ?The raw score of the second expert in the second evaluation; ? The raw score of the third expert in the second evaluation;
Cross distribution: For the case of 3,000 teams and 125 judges, where five experts judge each entry, we propose maximizing comparability between judges’ scores while satisfying the requirement that at least one judge rules each entry. The objective function is formulated to increase the comparability between scores from different reviewers and satisfy the associated constraints. (1) The level of expertise of the 125 expert reviewers will remain the same. (2) Five experts are chosen randomly to assess each submission. (3) At least one expert reviewer evaluates each submission. (4) Each reviewer has the same number of reviews for each submission.
Record the binary variables x[I, j],ith denoting the team number from 13000 and j denoting the expert number from 1125.
The objective function is:
Where yi represents the size of the intersection between the work of the ith team and the work evaluated by different experts. We calculate the size of each meeting between the entries in the set scored by other experts, and finally, we find the average of these intersection sizes in this objective function. We optimize the allocation scheme by maximizing this average to increase comparability with different expert scores [5].
To further explain, the quantitative relationship in the question is scaled down, as shown in [Figure 1 Figure 2]. below,
Figure 1: Collection of 3000 entries with 125 expert judges (left).
Figure 2: Collection of 200 entries with 15 expert judges (right).
from 3000 teams and 125 experts, each entry rated by five experts, to 200 teams and 15 experts, each entry rated by three experts, which allows us to visually observe the intersection between each expert’s sets of entries.
This allows us to see the intersection of each expert’s portfolio, as the scores of entries with significant intersections are more accurate. In contrast, the scores of entries with minor intersections are less convincing. For the reduced set of 200 entries and 15 experts, the number of entries with minor intersections increases, and the number of entries with significant intersections decreases
Finally, to improve the comparability of the scores given by different experts, the graph should be presented in such a way that the middle part of the set moves to the outer circle to produce a certain number of intersections, and the set in the outer ring spreads out to ensure that there is only a suitable number of intersections
In the implementation of this process, the current workload of each expert can be guaranteed for the time being under the same conditions.
Three thousand works are to be evaluated by 125 experts, each piece to be considered by five experts, then each expert to assess an average of 120 jobs, which not only improves the efficiency of the work of the task but also ensures that each expert works for a consistent period. This not only improves the efficiency of the task but also ensures that each expert works for the same amount of time.
Maximize the collection of works: The proposed scheme is to maximize the intersection between each paper and the set of documents reviewed by different experts, i.e., the gain maximization constraint, if we expect to increase the comparability of the grades given by other reviewers [6].
The specific steps are as follows:
Step 1: Number the 3000 works and 125 experts; for example, the works are numbered from1 to3000, and the experts are numbered from 1 to 125. Then, five experts are assigned to each piece for evaluation. Each expert can be selected repeatedly during the assignment because the number of experts is limited.
Step 2: If the experts can be selected repeatedly, resulting in the inconsistency of the number of experts’ selection, we adopt the average distribution of the number of works, and the number of jobs assigned to each expert is consistent at 120.
Step 3: With the above constraint, the intersection size between each team and the set of papers evaluated by each expert is calculated using the builtin Python package.
Step 4: Finally, the maximum value of the intersection size between each entry and the set of entries judged by different experts is chosen as the optimal solution to determine the best crossdistribution scheme.
That is, the constraint to obtain the objective function of the average value of the intersection size between each work and the collection of outcomes evaluated by different experts is 498. To ensure that the number of works selected by each expert is the same, it can satisfy the maximization of the comparability between the grades given by different evaluation experts [7].
Standard scores: A standardized score is calculated for each data point, where that standardized score is shifted with a mean value of 50. Deals with more significant standard deviations will result in a more extensive range of variation in the normal score.
The formula for this standardized score is then:
Where ak is the corresponding score given by an expert, a is the sample mean of the score given by an expert, and is the sample standard deviation of the score given by an expert.
Firstly, we calculated the standard score of each entry for the five expert judges in the first stage. The average of these five standard scores was used to determine the score of the entries in the first stage. These scores are then ranked, and the top 16% of entries are selected for the second judging stage. In the second stage, three experts will assess these entries, giving them three separate scores. These scores are converted into standard scores and then adjusted as necessary. At the end of the second stage, a composite score is given for each entry by adding up these three standard scores. The final overall score is the average of the phase 1 standard scores added to the stage 2 composite score, which ranks all entries that have participated in both stages of judging.
Data analysis and modelling: The means of the same factor at different levels are usually various when dealing with experimental data [8]. The overall change in scores between the two stages can be observed by calculating the difference in means between the two stages. In addition, the difference in mean scores can be used to quantify the change in work performance between the two assessment levels [9]. The specific steps are as follows:
Step 1: Take the standard scores given by the reviewing experts in the first stage and add them together, then divide by the total number of reviewers to get the average score of the first stage.
Step 2: Take the standard scores assessed by the reviewing experts in the second stage and add them up, then divide by the total number of reviewers to get the average score of the second stage.
Step 3: Calculate the mean difference: subtract the second stage’s mean from the first stage’s.
As can be seen from Figure 3]:
Figure 3: Comparison Chart of Mean Differences Between Stage One and Stage Two.
Stage 1 and Stage 2 have overall fluctuations, and works with high scores in Stage 1 have higher results in Stage 2; on the contrary, works with scores of medium level and below in Stage 1 have lower results in Stage 2 than in Stage 1, and most of them have negative mean differences, indicating that there are differences in the results of the evaluation of Stage 1 and Stage 2.
Analysis of variance (ANOVA): An analysis of variance difference (ANOVA) was performed to determine whether there was a significant difference between the two phases, and the hypothesis to be tested was whether the twostage review program would produce more reasonable assessment results than the nostage review program [10].
Specific steps:
Step 1: Test the hypothesis Null hypothesis (H0): No significant difference exists between the twostage and nostage review programs. Alternative hypothesis (H1): A significant difference exists between the twostage and nonstaged judging schemes.
Step 2: Selection of the test method and calculation of the statistics. Paired samples ttest.
Step 3: Interpret the results
From the results of the twostage ttest in Table 1,
Table 1: Table of twostage ttest results.

Mean ± standard deviation 





paired variable 
Pairing 1 
Pairing 2 
Pairwise difference (pair 1  pair 2) 
t 
df 
P 
Cohen's d 
Firststage variance paired with secondstage variance 
33±24 
28±20 
5±4 
6 
1499 
00.000*** 
0.18 
the first stage variance is based on the variables paired with the second stage clash. The significance pvalue is 0.000***, which represents significance at the level of rejection of the original hypothesis, so there is a significant difference between the twostage accreditation program and the nostage accreditation program.
Comparison of twostage review programs: Figure 4,
Figure 4: Comparison of the difference between Stage 1 and Stage 2 variances.
plotted using Python, shows that most of the differences are harmful and that the twostage judging scheme has the following advantages over the nostage judging scheme:
(1) Improved consistency: It is expected that scores in the second stage will show improved consistency, indicating that judges are more consistent in their scoring of entries, i.e., subjective variation between judges in the second stage is reduced, resulting in more centered scores within a given range.
(2) Minor Differences: The data from Stage 1 showed more significant discrepancies, i.e, more considerable differences in scores between entries. The second judging stage minimized these discrepancies, resulting in more uniform scores between works.
(3) More Uniform Judging Criteria: Stage 2 was assessed with greater stringency, leading to more harmonized scores. The reviewers applied more comparable assessment criteria to the entries in the second stage.
(4) Limiting effect: In certain instances, the assessment standards were constrained during the subsequent phase, culminating in more centralized scores.
Finally, the analysis suggests that largescale innovation competitions tend to follow the standard score as a ranking method, which assumes that the academic level of the works evaluated by each reviewer is similar. However, the number of jointly evaluated papers is limited; each reviewer evaluates only a few documents. As a result, the assumptions underlying the standard review scheme may no longer be applicable.
Establishment of scoring weights based on gray correlation analysis: Grey correlation analysis is a powerful method for handling grey systems and is particularly helpful in assessing correlations among various factors and their impact on a specific objective [11]. To obtain respective weights, we conducted a grey correlation analysis on the scores provided by five experts for each entry in the first round, three for each entry in the second round, and the final overall score. The steps involved in this analysis are as follows:
(1) Data Preparation: A table containing each factor should be prepared. Each column represents a single factor, and each row represents a sample.
(2) Data Normalization: Data processing should remain neutral. The data should be normalized for each column and transformed into the range [0, 1].
(3) Calculate the Correlation Coefficient: Calculate the correlation coefficient between each factor and the target by using the selected correlation function. This provides the correlation coefficient of each indicator:
(4)
(4) Determine the weights: The experts’ weights in both evaluations were obtained after normalization using each expert’s grey correlation degree.
The resulting weights assigned to each expert, obtained through grey correlation analysis and related data processing, are shown in Table 2.
Table 2: The weighting of grades in the assessment.

Experts in the first review 
Expert I in the second review 
Expert II in the second review 
Expert III in the second review 
weights 
0.28 
0.242 
0.237 
0.241 
These weighting measures aim to enhance the precise consideration of each expert’s review contribution and thus promote the review process’s accuracy and credibility [12].
Exploring the averagebased judging model: The averagebased judging model calculates the final score of the competition by taking the average of the five experts’ scores in the first judging and adding the raw scores of the three experts in the second judging. The following assumptions are made:
The raw score dataset in the first judging is represented as { 1 2 , , , }, X x x x = n where i x denotes the mark of the ith expert in the first judging.
The raw score dataset in the second review is represented as { 1, 2 ,., }, Y y y y = n where i
Denotes the mark of the ith expert in the second review.
The average score of the five experts in the first review is calculated as follows:
(5)
Where x i n i = = ( 1,2, ) denotes the original score of the expert in the first evaluation, andaverage1 indicates the average score of the expert in the initial assessment.
The final score of the twostage evaluation is calculated as:
(6)
Using software such as Mat lab and Excel, the average of the scores of the five experts in the first review can be calculated. Then, the scores of the entries that went into the second review are added to the original scores of the three experts in the second review to obtain the final score of the competition. The partial scores of the final ranking can be seen in Table 3.
Table 3: Final scores and rankings for the first evaluation scheme.
order of precedence 
Final scores calculated for the initial assessment scheme 
order of precedence 
Final scores calculated for the first evaluation scheme 
1 
80.76 
14 
67.54 
2 
79.44 
15 
67.32 
3 
79.43 
16 
67.16 
4 
77.15 
17 
67.16 
5 
76.12 
18 
67.15 
6 
73.96 
19 
67.02 
7 
73.35 
20 
66.79 
8 
71.69 
21 
66.36 
9 
70.78 
22 
66.11 
10 
70.7 
23 
65.84 
11 
69.93 
24 
65.75 
12 
68.46 
25 
65.46 
13 
67.77 


In judging the competition, the average score method calculates the scores given by several judges. This significantly reduces the disproportionate influence of individual judges on the result and promotes impartiality and fairness. It also allows for combining different perspectives and scores, as judges have additional expertise, experience, and preferences. This leads to a more comprehensive assessment. Finally, averaging the scores of several assessors can reduce the influence of mis judgments or incorrect ratings by individual assessors on the results and increase the accuracy and reliability of the assessment results. In general, the assessment results’ impartiality, comprehensiveness, and authenticity can be promoted, and the results of the competition are made more objective and trustworthy by using the average score assessment.
The weighted averagebased judging model: In the second evaluation method, grey correlation analysis is used to determine the average of the expert scores from the first evaluation and the four weightings of the original scores provided by three additional expert evaluators during the second evaluation. The weighted average technique derives the final score. The following assumptions are made:
The raw score dataset in the first judging is represented as X x x x ={ 1 2 , , , n} , where i x (i=1,2,...,n) denotes the mark of the i th expert in the first judging.
The raw score dataset in the second review is represented as { 1, 2 ,., }, Y y y y = n where i
(i=1,2,...,n) denotes the mark of the ith expert in the second review.
The dataset of weights for each rating is represented as ω= {ω0, ω1,...,ωn}, where ωi denotes the importance of the i+1th rating in the second review.
Therefore, the average score of the first stage review is calculated as follows:
(7)
Where x i n i = = ( 1,2, ) denotes the original score of the experts in the first review, and average1indicates the average score of the experts in the first review.
The final score W is calculated as:
(8)
Using Mat lab and Excel software, the grey correlation analysis was utilized to determine the average of expert scores from the first review and the four weights of raw scores from the three expert reviews in the second review. The final grades were then obtained through the weighted average method. The subscores of the final ranking are shown in Table 4.
Table 4: Final scores and rankings of the second evaluation program.
order of precedence 
Final scores for the second accreditation program 
order of precedence 
Final scores for the second accreditation program 
1 
69.17 
14 
61.94 
2 
68.84 
15 
61.90 
3 
66.97 
16 
61.82 
4 
66.59 
17 
61.34 
5 
66.20 
18 
60.98 
6 
64.94 
19 
60.59 
7 
64.82 
20 
60.34 
8 
64.43 
21 
60.31 
9 
63.57 
22 
60.31 
10 
63.53 
23 
60.15 
11 
62.27 
24 
59.83 
12 
62.27 
25 
59.48 
13 
62.03 


To enhance the impartiality of the judging process, we use a weighted average method for our competitions. The judges with more competence and expertise are given a greater weight to influence the final results of the judging process [13]. Only objective judgments are considered, and subjective assessments are explicitly marked as such, helping to ensure that the judges’ scores align with the contest’s requirements and standards. By assigning a higher weight to the most critical reviewers, the accuracy and reliability of the final judgment can be ensured, as their opinions are more strongly represented. Similarly, weighing the reviewers can consider their respective characteristics, backgrounds, and experience, thus better reflecting the contributions and credibility of the reviewers. This approach allows for adaptive adjustments to specific circumstances, thereby increasing the rationality and reliability of the review results.
We then compare the differential rankings of the two assessment methods to determine their comparative merits, allowing for a fairer, more valid, and more effective assessment.
Comparison of Ranking Differences in Two Evaluation Models: The rankings of the final grades, which were obtained via separate analyses using both the average and weighted average formulas, were compared to the dataset grade rankings. This percentage difference was used to analyze the grade distribution characteristics for each expert, the original grades allocated for each assignment, and the adjusted grades (e.g., standard scores). In addition, the effectiveness of both methods was compared to determine which way is more advantageous [14].
The average formula calculates the final score as:
(9)
Where x i n i = = ( 1,2, ) notes the raw score of the expert in the first evaluation, and i y (i=1,2, . . . , n) denotes the raw score of the expert in the second evaluation.
The weighted average formula calculates the final score as follows:
10
Where ω1 denotes the weight of the ith grade.
The final grade, Z, calculated using the average formula, and the last grade, W, computed using the weighted average formula, are sorted. The sorted results are denoted as { a z z z w w w 1 2 , ,. , ,. n } nd{ 1 2 n } ′ ′ ′ ′ ′ ′ To measure the degree of difference between the rankings in the original dataset and the rankings in another dataset, we calculate the difference ratio one and difference ratio 2. The mathematical formulas for these discrepancy ratios are as follows:
11
12
where n denotes the number of grades in the dataset, rank ( )i x denotes the rank of the ith quality in the original dataset, ( )i z ′ means the rank of the grade i y in the other dataset that corresponds to the ith grade i x , and rank wi ′ denotes the rank of the grade wi ′ in the other dataset that corresponds to the ith grade ( )i x , rank( )i x  rank ( )i z ′  denotes the ranking difference.
Using software such as Mat lab and Excel, the final results obtained from the average and weighted average formulas can be compared with those from the original dataset to calculate the difference ratios. The ranking difference results of the two schemes are shown in Figure 5 and Figure 6.
Figure 5: Difference in ranking for the first evaluation scheme
Figure 6: Difference in ranking for the second evaluation scheme.
Where the vertical coordinate signifies the variance between the ranking order calculated by the averaging and weighted average methods and the original order, 0 indicates no differences, a negative number indicates a descending ranking order and a positive number indicates an ascending ranking order. From the graph provided, it can be inferred that the first scoring scheme shows a percentage difference in ranking of approximately 0.57%. The second scoring scheme offers a difference of roughly 0.28%. It is worth noting that the percentage difference in the order of the second scoring scheme is lower than that of the first. In other words, the percentage difference in charge of the weighted average method is lower than that of the middle way The analysis of the means, extremes, and variances produced by the two methods and the comparison with Table 5,6 shows that
Table 5: Comparison of the parameters of the two review programs.

The first accreditation program 
The second evaluation program 
Mean 
50.63 
50.89 
Extreme Deviation 
34.73 
34.24 
variance (statistics) 
34.92 
33.69 
Table 6: Symbolic representation.
Assumptions 
symbolic representation 
Collection of works 
w = {w1, w2, ..., wn} 
Expert rating scale 
R = {A, B, C, D, E} 
Expert rating score 
S = {5, 4, 3, 2, 1} 
Weights of entries 
ω= {ω1,ω2,ω3} 
Number of works 
n 
Number of experts 
m (assumed to be 3) 
Maximum score of an entry 
Tmax = 15 
Minimum score for an entry 
Tmin = 3 
the extremes and variances of the first evaluation method exceed those of the second. It is understood that minor extreme deviation and variance indicate a narrower range of data values, resulting in more minor differences between data points and more stable data that is less susceptible to outliers and closer to the mean. This indicates better consistency between the data and the scoring criteria.
To sum up, the second model of the scoring scheme appears superior to the first. After analyzing and discussing different models, we have narrowed our focus to whether the subjective views of the experts influence the review results. This has led to the development a twostage scoring process aimed at optimizing the review process and minimizing the influence of human factors on the final results. This has led to the product of a twostage scoring process aimed at optimizing the review process and minimizing the impact of human factors on the final results.
Controlling Subjective Factors Experimental Design: With judges applying their independent criteria, innovation competitions suffer from inconsistent scoring. Lenient judges may award higher scores, while stricter judges award lower [15]. Contributing to this subjectivity is the lack of clearly defined examlike criteria. The study found that the main influences on the subjective rating of the index are emotional factors related to the subject and object of the assessment, the methods, and mechanisms used in the evaluation and that the expert’s ability to judge is not reflected in the actual rating [16]. To improve the evaluation process, a more impartial method should be used. We conducted a hypothesis test to confirm the accuracy of our assumption, and the results confirmed our hypothesis. The first stage of the scoring process uses the normal distribution of judges’ scores for standardization purposes. This method removes the influence of the judges’ subjective opinions, thus ensuring fairer and more objective scores [17]. Therefore, the independent criteria of each judge no longer affect the scores but rather align with the distributional characteristics of standard normal distribution, enhancing consistency and fairness. In addition, a new ranking approach based on the Borda sorting method has been applied to rank the entries in the second round of judging [18].
Establishment of Nonlinear Programming Model: A comprehensive twostage integration method has been proposed [19]. In the first stage of the evaluation, the scoring results of each expert are analyzed employing a normal distribution. Then, the distribution is normalized to ensure that each expert adheres to the same scoring criteria [20]. In the second stage, we have adopted a new categorization approach based on the Borda method [21,22]. The main objective is to categorize the works submitted for the second evaluation into five ABCDE grades: A has 5 points, B has 4 points, C has 3 points, D has 2 points, and E has 1 point. Each work is evaluated by three experts who assign a grade according to the content of the work and the corresponding score. They then sum up the total scores of all the pieces. Each submission is evaluated by three experts who assign grades based on the quality of the content and subsequently provide corresponding scores. The final score is then calculated by summarising the individual scores given by each expert. The maximum number of points that can be obtained is 15, while the minimum is 3. Innovation competitions, such as modeling contests, generally list the first, second, and third prizes without ranking. Therefore, the works need to be sorted for segmentation. Results are divided into different phases during the segmentation process if they score similarly. At this point, the second weighting can be applied [23,24]. Firstly, the weight of the scores from the three expert judges is obtained. Then, the weighting is multiplied by the corresponding scores and added up. By doing so, different total scores are accepted, and the highest score is considered a reasonable and fair ranking. This segmentation makes it easier to identify the winning entries in creative competitions. It also ensures that the judging aligns with the predetermined award standards [25].
For better modeling in the second stage of judging, we have made the following assumptions and limitations:
Limitations
1. Each work is scored by three experts and categorized into five grades: A, B, C, D, and E.
2. The scores of the works are calculated based on the scores corresponding to the grades.
3. The total score of the work is the sum of the scores rated by the three experts.
4. The total score of the work is within the range [T min, T max].
5. The works are ranked according to the total score, with the highest score being the first place.
6. The entries are segmented based on their total scores to determine the first, second, and thirdplace entries. The specific flowchart of the twostage judging program is shown in Figure 7.
Figure 7: Model building flowchart.
In this flowchart:
First, let f(x) represent the probability density value at a given x. represents the mean σ represents the standard deviation, represents the circumference, and exp represents the natural exponential function normalized as follows [26]:
Let Z represent the transformed random variable, X represent the original customarily distributed random variable, μ represent the mean of the original random variable, and present the standard deviation of the original random variable. The normalized average distribution formula is as follows:
Finally, the Borda ranking method is divided into the following five main steps:
Step 1: Calculate the degree of affiliation. ij x can be expressed as the result of the jth evaluation method of the ith review expert. It is a simplistic normalization:
Step 2: Calculate the fuzzy frequency number:
Step 3: Transform the ranking into a score:
Step 4: Calculate the Borda number:
Where the more significant the value of FBi , the higher the rankings are.
RESULTS AND DISCUSSION
To ensure the fairness, impartiality, and scientific validity of the judging process in the innovation category of the competition, we investigate whether the subjective evaluations of the experts affect the judging results. As a result, we have designed a twostage scoring system consisting of standard distribution standardization and a Borda rankingbased sorting method. This approach allows us to optimize the judging process and reduce the potential influence of human factors on the final decision [27]. We implemented a twostage scoring scheme using Mat lab based on competition scoring data. In the first stage, we normalized the scores using the normal distribution. To achieve this, we calculated the normal distribution of each judge’s score and obtained the normalized result. The figure below shows the normal distribution of some judges’ scores and the resulting normalization process.
Four judges (P005, P022, P127, and P230) were randomly selected from a pool of 125 experts. As can be seen in Figure 8,
Figure 8: Normal Distribution of Ratings of Selected Experts in the First Review
their scores met the criteria for a normal distribution [28]. This procedure aimed to establish consistency in the scoring criteria of each expert Figure 8, shows the normalization of the experts’ scores in the first review. This was achieved by standardizing the distribution. Figure 9
Figure 9: Normalization of the Normal Distribution of Selected Expert Ratings in the First Review.
shows the results of the first review, which involved developing ratings and league tables. By standardizing the scoring process, we were able to eliminate scoring bias. By standardizing the scores, we stopped the discrimination by making the scores follow a standard normal distribution. This process ensured that the scoring results would be fair by removing the bias caused by different scoring criteria.
As shown in Figure 10,
Figure 10: Ranking table of scores after normalization in the first review
we have obtained the scores and rankings from the first round of evaluations. By standardizing the scores, we successfully eliminated the bias in the scoring. Bias in scoring can occur due to different judges or scoring criteria, resulting in unfair scoring results. By normalizing the scoresusing a normal distribution, we transformed the scores into a standard normal distribution, eliminating the bias caused by different scoring criteria and ensuring a fairer scoring result [29].
In the second stage of the evaluation, we implemented an improved approach based on the Borda ranking method, as illustrated in Figure 11.
Figure 11: Ranking of scores in the second review.
We obtained the final results of the second round of evaluations through this new ranking method based on the Borda ranking method. Since the competition adopted a twostage evaluation model, this represents the review’s outcome. We can observe the ranking and positions of each team.
To prove the validity of our sorting method and model, NDCG Normalized Discounted Cumulative Gain(NDCG) is used for validation. The value of NDCG ranges from 0 to 1, where 1indicates the best sorting result, and 0 indicates the worst. Higher NDCG values indicate better sort quality and lower NDCG values indicate poorer quality [30].In general, the following general guidelines can be used to evaluate NDCG values:
When 1>NDCG ≥ 0.8, it is an excellent sorting result;
When 0.8>NDCG≥0.6, it is a good sorting result;
When 0.6>NDCG≥0.4, it is a fair sorting result;
When 0.4>NDCG≥0, it is a poor sorting result.
We calculated the NDCG result of the reordering model for the second judging by Mat lab as 0.8667>0.8, which means that our proposed new ordering model will get an excellent ordering result in the judging process.
This paper proposes a more streamlined judging technique that divides the competition judging into two phases and implements different judging methods. First, we applied a standardization technique using a normal distribution to evaluate the scores of each judge. This ensured that the judges had identical scoring criteria, thereby increasing the impartiality of the scoring. The first judging stage was crucial, enabling many outstanding teams to be selected. The second stage of the judging process was introducing a new judging method based on the Borda method. The entries were divided into five levels, with the judges assigning a score to each group based on their judgment. This method ensures that the evaluation process is fair and objective. The final ranking of each entry was determined by the total number of points it received. When multiple entries receive the same score, a weighting process increases the accuracy and fairness of the final ranking results. In addition, our methodology shows practical solid performance. It has been validated by NDCG scores, which confirm the successful control of subjective factors and the production of quality ranking results in the judging scheme. This ensures a fair and accurate judging process. It also guarantees a reasonable ranking of competition entries.
FUTURE WORK AND IMPROVEMENTS
The judging scheme proposed in this study for largescale innovation contests can be applied to innovation contests and various other largescale assessments and evaluations, such as art tests, exams of multiple levels, and elections of public officials, while achieving a more equitable level of accuracy. Its benefits are the enhancement of impartiality, consistency, and objectivity throughout the assessment process. Implementing and designing largescale innovation competitions is a complex process influenced by several factors that impact each other. This paper focuses on controlling subjective factors that influence the scoring of entries and only provides solutions for controversial entries. To ensure the quality and longevity of competitions, further research into additional features for factor analysis is necessary, and a continued focus on improving the judging scheme is imperative.
ACKNOWLEDGMENTS
This work was supported by the Philosophical and Social Sciences Research Project of Hubei Education Department (19Y049), and the Staring Research Foundation for the Ph.D. of Hubei University of Technology (BSQD2019054), Hubei Province, China.
REFERENCES
 Chesbrough HW, Vanhaverbeke W, West J. Open Innovation:Researching a New Paradigm. 2006; 84: 12591262.
 Poetz MK, Schreier M. The Value of Crowdsourcing: Can Users Really Compete with Professionals in Generating New Product Ideas. J Produ Innova Manage. 2012; 29: 245256.
 You Qinggen. Research and Implementation of Expert Selection Model for WISCO Research Project Review. Huazhong University of Science and Technology. 2014; 32: 5660.
 Kim Y, Kim Y, Kim J, Lee S, Kwon S. Boosting on the functional ANOVA decomposition. Statistics & Its Interface. 2009; 2: 361368.
 Kannemann K. The Exact Evaluation of 2way Crossclassifications: An Algorithmic Solution. J Biometrical. 2010; 24: 157169.
 Song Y, Shi X, Hu B, Jin Y. Team Building Algorithm Based on Multi Objective Greedy Strategy for Gain Maximization. High Technol Communi. 2018; 28: 279290.
 Mielik Inen T, Ukkonen E. The Complexity of Maximum Matroid Greedoid Intersection and Weighted greedoid maximization. Discre Appl Mathemat. 2006; 154: 684691.
 Jia Tianli. Hypothesis Analysis Model of the Causes of MeanDifferences. Mathemat Pract Understa. 2005; 64: 212215.
 Daoji S. An analysis of variance test for the extreme value distribution.Tianjin Daxue Xuebao. 1988; 1988: 116121.
 Shapiro SS, Wilk MB. An analysis of variance test for normality(complete samples). Biometrika. 1975; 67: 215216.
 Guohong S, Zhenhui Z. Application of Gray Correlation Analysis in Fault Tree Diagnosis. China Safety Science J. 1999; 30: 15051507.
 Liu G, Yu J. Gray correlation analysis and prediction models of living refuse generation in Shanghai city. Waste Management. 2007; 27: 345351.
 Victor G, Juan DTM. The Eurovision Song Contest: voting rules, biases and rationality. J Cultur Econo. 2022; 47: 247277.
 Looney MA. Evaluating Judge Performance in Sport. J Appli Measure. 2004; 5: 3147.
 Margaret AC, Marjorie CK. A proposed approach for evaluating expert wine judge performance using descriptive statistics. J Wine Res. 1996; 7: 8390.
 Wang Yonglin. Subjective Indicators in Educational Evaluation and Factors Affecting Their Judgment. Educa Sci. 2017; 33: 1419.
 Xu Kai. Exploration and Research on Retrospective Evaluation Selection Method Based on Peer Review Experts. Beijing University of Chinese Medicine. 201; 7: 4750.
 Ryo I, Kazumasa O. Borda Count Method for Fiscal Policy: A Political Economic Analysis. The Institute of Comparative Economic Studies, Hosei University. 2022; 36: 2540.
 Chen Yuan. Research on DecisionMaking Methods for Scientific Fund Project Review and Selection. North eastern University. 2011; 9: 23 26.
 Lee SM, Kim KH, Kim EJ. Selection and Classification of Bacterial Strains Using Standardization and Cluster Analysis. J AnimaL Sci Technol. 2012; 32: 5456.
 Zheng Rui. Ranking Risk with Borda Method. Enterprise Reform and Management. 2015; 32: 154156.
 Orouskhani M, Shi D, Cheng X. A Fuzzy Adaptive Dynamic NSGAII With FuzzyBased Borda Ranking Method and its Application to Multimedia Data Analysis. IEEE Transac on Fuzzy System. 2021; 29: 118128.
 Tang Huanling, Wang Jingdong, Lu Yuchang. A New Weight Allocation Strategy to Reduce Similar Topic Classification Errors. Comp Engineer Applica. 2004; 58: 185188.
 Wang Zeyan, Gu Hongfang, Yi Xiaoxin, Zhang Shenru. A Linear Combination Weighting Method Based on Entropy. Syst Engineer Theory Practi. 2003; 69: 112116.
 Kilgour DM, Grégoire JC, Foley AM. Correction to: Weighted scoring elections: is Borda best?. Social Choice and Welfare. 2021; 58: 12.
 Pierrat L, Samuel K. Standardization of the Logistic Distribution based on Entropy. Int J Performability Eng. 2013; 9: 352354.
 Jian W, Qianqian M, HuChen L. A metaevaluation model on science and technology project review experts using IVIFBWM and MULTIMOORA. Expert Systems With Applications. 2021; 36: 6870.
 Ke Nianri, Zhang Chaomo, Wang Yingwei. Application of Normal Distribution Method in Well Logging Curve Standardization in M Oilfield. J Yangtze University (Natural Science Edition). 2010; 7: 76 78.
 Aroian, Leo A. Handbook of the Normal Distribution. Technometrics.1996; 25:112115.
 Taylor M, Guiver J, Robertson S, Minka T. ABSTRACT SoftRank: Optimising NonSmooth Rank Metrics. Inproceedings. 2008; 7: 77 86.