Fix our evaluation formula

This commit is contained in:
Ian Cordasco 2016-02-19 15:30:54 -06:00
parent bda3124ffa
commit 0827df3e2a

View file

@ -82,7 +82,7 @@ reports=no
# respectively contain the number of errors / warnings messages and the total
# number of statements analyzed. This is used by the global evaluation report
# (RP0004).
evaluation=100.0 - ((float(5 * error + warning + refactor + convention) / statement) * 100)
evaluation=10.0 - ((float(5 * error + warning + refactor + convention) / statement) * 10)
# Template used to display messages. This is a python new-style format string
# used to format the message information. See doc for all details