How the false positive value affects accuracy? T P <mtext>&#xA0;</mtext> ( t r

Azzalictpdv

Azzalictpdv

Answered question

2022-05-07

How the false positive value affects accuracy?
T P   ( t r u e   p o s i t i v e )   =   2739
T N   ( t r u e   n e g a t i v e )   =   103217
F P   ( f a l s e   p o s i t i v e )   =   43423
F N   ( f a l s e   n e g a t i v e )   =   5022
a c c u r a c y = T P + T N T P + T N + F P + F N
In this case the accuracy is 0.68. Can I say that I have low accuracy because the value false positive is high? There is any relathion between false positive and the parameters true positive or true negative?

Answer & Explanation

Calvin Oneill

Calvin Oneill

Beginner2022-05-08Added 20 answers

You might be interested in the false positive rate. This is
F P R = F P / ( T N + F P ) = 43423 / ( 43423 + 103217 ) = 0.3
F N R = F N / ( T P + F N ) = 5022 / ( 2739 + 5022 ) = 0.65
Your false positive rate is actually very low. You are mostly classifying negatives correctly as negative. But your false negative rate is higher at 0.65. You are not detecting enough images that are actually true images. For a more detailed discussion, read on.
The accuracy is an overall summary of the success of your tests. It is sometimes helpful to look at a breakdown of it into sensitivity and specificity. Sensitivity is the true positive rate, i.e. the chance your program predicts a positive given that it's supposed to be positive. Specificity is the true negative rate, i.e. the proportion of negatives correctly predicted as negative.
s e n s i t i v i t y = T P T P + F N = 2739 2739 + 5022 = 0.35
s p e c i f i c i t y = T N T N + F P = 103217 103217 + 43423 = 0.7
Based on these results, it seems that your test is not as good as identifying positives as it is at correctly identifying negatives. One possible problem could be that there is a difference in the prevalence of images from the truth and produced by your computer algorithm. Here a confusion matrix of the data you have collected:
a c t u a l p r e d i c t e d t r u t h a l g o r i t h m t r u t h 2739 43423 a l g o r i t h m 5022 103217
If you compute the accuracies separately for the true images and the algorithm images, you find that the accuracy for the true images is 2739 2739 + 5022 = 0.35 and the accuracy for your algorithmic images is 103217 103217 + 43423 = 0.7. (Notice these are the definitions of sensitivity and specificity, the same above.) The reason that your overall accuracy is high is because you have a much higher prevalence of algorithm-generated images; your algorithm tends to guess conservatively, and the fact that there are more algorithm generated images pushes your overall accuracy up.
A useful one-number summary combining sensitivity and specificity is the F 1 score, defined as
1 β 2 β 2 + 1 1 s e n s i t i v i t y + 1 1 + β 2 1 s p e c i f i c i t y
When β = 1, a common choice, this is the harmonic average of sensitivity and specificity:
1 1 2 ( 1 s e n s i t i v i t y + 1 s p e c i f i c i t y )
The F 1 score of your test is 0.47. The closer to 1, the better. You might try to optimize your algorithm based on the F 1 score.
There is some difference between your F 1 score and accuracy:
F 1 : 0.47
a c c u r a c y : 0.68

Do you have a similar question?

Recalculate according to your conditions!

New Questions in College Statistics

Ask your question.
Get an expert answer.

Let our experts help you. Answer in as fast as 15 minutes.

Didn't find what you were looking for?