[LNT] r237658 - Refactor ComparisonResult to be more self contained

Chris Matthews cmatthews5 at apple.com
Wed Jun 3 07:20:08 PDT 2015


This looks familiar. I wonder if I fixed only locally. Will fix today!

Sent from my iPhone

> On Jun 3, 2015, at 7:16 AM, James Molloy <james at jamesmolloy.co.uk> wrote:
> 
> Hi Chris,
> 
> I just updated our LNT instance and this commit causes the following internal server error:
> 
> ... snip ...
>   File "/home/llvm-test/sandbox/lnt/lnt/server/ui/templates/v4_run.html", line 334, in block "body"
>     {% set cr = request_info.sri.get_geomean_comparison_result(
>   File "/home/llvm-test/sandbox/lnt/lnt/server/reporting/analysis.py", line 286, in get_geomean_comparison_result
>     bigger_is_better=field.bigger_is_better)
>   File "/home/llvm-test/sandbox/lnt/lnt/server/reporting/analysis.py", line 60, in __init__
>     self.stddev = stats.standard_deviation(samples)
>   File "/home/llvm-test/sandbox/lnt/lnt/util/stats.py", line 36, in standard_deviation
>     m = mean(l)
>   File "/home/llvm-test/sandbox/lnt/lnt/util/stats.py", line 15, in mean
>     return sum(l)/len(l)
> TypeError: unsupported operand type(s) for +: 'int' and 'NoneType'
> 
> Could you please take a look?
> 
> Cheers,
> 
> James
> 
> [P.S: if you reply to james.molloy at arm.com perhaps I might get your reply - I still lay the blame at gmail's forwarding eating your messages]
> 
>> On Tue, 19 May 2015 at 03:11 Chris Matthews <cmatthews5 at apple.com> wrote:
>> Author: cmatthews
>> Date: Mon May 18 20:59:20 2015
>> New Revision: 237658
>> 
>> URL: http://llvm.org/viewvc/llvm-project?rev=237658&view=rev
>> Log:
>> Refactor ComparisonResult to be more self contained
>> 
>> Modified:
>>     lnt/trunk/lnt/server/reporting/analysis.py
>>     lnt/trunk/lnt/server/ui/templates/v4_run.html
>>     lnt/trunk/tests/server/reporting/analysis.py
>> 
>> Modified: lnt/trunk/lnt/server/reporting/analysis.py
>> URL: http://llvm.org/viewvc/llvm-project/lnt/trunk/lnt/server/reporting/analysis.py?rev=237658&r1=237657&r2=237658&view=diff
>> ==============================================================================
>> --- lnt/trunk/lnt/server/reporting/analysis.py (original)
>> +++ lnt/trunk/lnt/server/reporting/analysis.py Mon May 18 20:59:20 2015
>> @@ -31,43 +31,67 @@ def calc_geomean(run_values):
>>  class ComparisonResult:
>>      """A ComparisonResult is ultimatly responsible for determining if a test
>>      improves, regresses or does not change, given some new and old data."""
>> -
>> -    def __init__(self,cur_value, prev_value, delta, pct_delta, stddev, MAD,
>> -                 cur_failed, prev_failed, samples, prev_samples, stddev_mean = None,
>> -                 confidence_lv = .05, bigger_is_better = False):
>> -        self.current = cur_value
>> -        self.previous = prev_value
>> -        self.delta = delta
>> -        self.pct_delta = pct_delta
>> -        self.stddev = stddev
>> -        self.MAD = MAD
>> +
>> +    def __init__(self, aggregation_fn,
>> +                 cur_failed, prev_failed, samples, prev_samples,
>> +                 confidence_lv=0.05, bigger_is_better=False):
>> +        self.aggregation_fn = aggregation_fn
>> +        if samples:
>> +            self.current = aggregation_fn(samples)
>> +        else:
>> +            self.current = None
>> +        if prev_samples:
>> +            self.previous = aggregation_fn(prev_samples)
>> +        else:
>> +            self.previous = None
>> +
>> +        # Compute the comparison status for the test value.
>> +        if self.current and self.previous and self.previous != 0:
>> +            self.delta = self.current - self.previous
>> +            self.pct_delta = self.delta / self.previous
>> +        else:
>> +            self.delta = 0
>> +            self.pct_delta = 0.0
>> +
>> +        # If we have multiple values for this run, use that to estimate the
>> +        # distribution.
>> +        if samples and len(samples) > 1:
>> +            self.stddev = stats.standard_deviation(samples)
>> +            self.MAD = stats.median_absolute_deviation(samples)
>> +        else:
>> +            self.stddev = None
>> +            self.MAD = None
>> +
>> +        self.stddev_mean = None  # Only calculate this if needed.
>>          self.failed = cur_failed
>>          self.prev_failed = prev_failed
>>          self.samples = samples
>>          self.prev_samples = prev_samples
>> -        self.stddev_mean = stddev_mean
>> +
>>          self.confidence_lv = confidence_lv
>>          self.bigger_is_better = bigger_is_better
>> 
>> +    @property
>> +    def stddev_mean(self):
>> +        """The mean around stddev for current sampples. Cached after first call.
>> +        """
>> +        if not self.stddev_mean:
>> +            self.stddev_mean = stats.mean(self.samples)
>> +        return self.stddev_mean
>> +
>>      def __repr__(self):
>>          """Print this ComparisonResult's constructor.
>> -
>> +
>>          Handy for generating test cases for comparisons doing odd things."""
>> -        frmt = "{}(" + "{}, " * 11 + ")"
>> -        return frmt.format("ComparisonResult",
>> -                           self.current,
>> -                           self.previous,
>> -                           self.delta,
>> -                           self.pct_delta,
>> -                           self.stddev,
>> -                           self.MAD,
>> -                           self.failed,
>> -                           self.prev_failed,
>> -                           self.samples,
>> -                           self.prev_samples,
>> -                           self.stddev_mean,
>> -                           self.confidence_lv,
>> -                           self.bigger_is_better)
>> +        fmt = "{}(" + "{}, " * 7 + ")"
>> +        return fmt.format(self.__class__.__name__,
>> +                          self.aggregation_fn.__name__,
>> +                          self.failed,
>> +                          self.prev_failed,
>> +                          self.samples,
>> +                          self.prev_samples,
>> +                          self.confidence_lv,
>> +                          bool(self.bigger_is_better))
>> 
>>      def is_result_interesting(self):
>>          """is_result_interesting() -> bool
>> @@ -237,77 +261,27 @@ class RunInfo(object):
>>                        if s[field.index] is not None]
>>          prev_values = [s[field.index] for s in prev_samples
>>                         if s[field.index] is not None]
>> -        if run_values:
>> -            run_value = self.aggregation_fn(run_values)
>> -        else:
>> -            run_value = None
>> -        if prev_values:
>> -            prev_value = self.aggregation_fn(prev_values)
>> -        else:
>> -            prev_value = None
>> -
>> -        # If we have multiple values for this run, use that to estimate the
>> -        # distribution.
>> -        if run_values and len(run_values) > 1:
>> -            stddev = stats.standard_deviation(run_values)
>> -            MAD = stats.median_absolute_deviation(run_values)
>> -            stddev_mean = stats.mean(run_values)
>> -        else:
>> -            stddev = None
>> -            MAD = None
>> -            stddev_mean = None
>> -
>> -        # If we are missing current or comparison values we are done.
>> -        if run_value is None or prev_value is None:
>> -            return ComparisonResult(
>> -                run_value, prev_value, delta=None,
>> -                pct_delta = None, stddev = stddev, MAD = MAD,
>> -                cur_failed = run_failed, prev_failed = prev_failed,
>> -                samples = run_values, prev_samples = prev_values,
>> -                confidence_lv = self.confidence_lv,
>> -                bigger_is_better = field.bigger_is_better)
>> -
>> -        # Compute the comparison status for the test value.
>> -        delta = run_value - prev_value
>> -        if prev_value != 0:
>> -            pct_delta = delta / prev_value
>> -        else:
>> -            pct_delta = 0.0
>> -
>> -        return ComparisonResult(run_value, prev_value, delta,
>> -                                pct_delta, stddev, MAD,
>> -                                run_failed, prev_failed, run_values,
>> -                                prev_values, stddev_mean, self.confidence_lv,
>> -                                bigger_is_better = field.bigger_is_better)
>> -
>> +
>> +        r = ComparisonResult(self.aggregation_fn,
>> +                             run_failed, prev_failed, run_values,
>> +                             prev_values, self.confidence_lv,
>> +                             bigger_is_better=field.bigger_is_better)
>> +        print repr(r)
>> +        return r
>> 
>> -    def get_geomean_comparison_result(self, run, compare_to, field, tests,
>> -                                      comparison_window=[]):
>> +    def get_geomean_comparison_result(self, run, compare_to, field, tests):
>>          if tests:
>>              prev_values,run_values = zip(*[(cr.previous,cr.current) for _,_,cr in tests])
>>          else:
>>              prev_values,run_values = [], []
>> 
>> -        run_geomean = calc_geomean(run_values)
>> -        prev_geomean = calc_geomean(prev_values)
>> -
>> -        if run_geomean and prev_geomean:
>> -            delta = run_geomean - prev_geomean
>> -            if prev_geomean != 0:
>> -                pct_delta = delta / prev_geomean
>> -            else:
>> -                pct_delta = 0.0
>> -        else:
>> -            delta = pct_delta = 0
>> -
>> -        return ComparisonResult(run_geomean, prev_geomean, delta,
>> -                                pct_delta, stddev=None, MAD=None,
>> -                                cur_failed=run_values and not run_geomean,
>> -                                prev_failed=prev_values and not prev_geomean,
>> -                                samples=[run_geomean] if run_geomean else [],
>> -                                prev_samples=[prev_geomean] if prev_geomean else [],
>> +        return ComparisonResult(calc_geomean,
>> +                                cur_failed=bool(run_values),
>> +                                prev_failed=bool(prev_values),
>> +                                samples=run_values,
>> +                                prev_samples=prev_values,
>>                                  confidence_lv=0,
>> -                                bigger_is_better = field.bigger_is_better)
>> +                                bigger_is_better=field.bigger_is_better)
>> 
>>      def _load_samples_for_runs(self, run_ids):
>>          # Find the set of new runs to load.
>> 
>> Modified: lnt/trunk/lnt/server/ui/templates/v4_run.html
>> URL: http://llvm.org/viewvc/llvm-project/lnt/trunk/lnt/server/ui/templates/v4_run.html?rev=237658&r1=237657&r2=237658&view=diff
>> ==============================================================================
>> --- lnt/trunk/lnt/server/ui/templates/v4_run.html (original)
>> +++ lnt/trunk/lnt/server/ui/templates/v4_run.html Mon May 18 20:59:20 2015
>> @@ -332,7 +332,7 @@
>>          </tbody>
>>          <tfoot>
>>            {% set cr = request_info.sri.get_geomean_comparison_result(
>> -                      run, compare_to, field, tests, request_info.comparison_window) %}
>> +                      run, compare_to, field, tests) %}
>>            <td><input type="checkbox" name="mean" value="{{machine.id}}.{{field.index}}"></td>
>>            <td><a href="{{graph_base}}&mean={{machine.id}}.{{field.index}}">Geometric Mean</a></td>
>>            {{ get_cell_value(cr) }}
>> 
>> Modified: lnt/trunk/tests/server/reporting/analysis.py
>> URL: http://llvm.org/viewvc/llvm-project/lnt/trunk/tests/server/reporting/analysis.py?rev=237658&r1=237657&r2=237658&view=diff
>> ==============================================================================
>> --- lnt/trunk/tests/server/reporting/analysis.py (original)
>> +++ lnt/trunk/tests/server/reporting/analysis.py Mon May 18 20:59:20 2015
>> @@ -4,7 +4,7 @@
>>  import unittest
>>  import lnt.util.stats as stats
>>  from lnt.server.reporting.analysis import ComparisonResult, REGRESSED, IMPROVED
>> -from lnt.server.reporting.analysis import UNCHANGED_PASS
>> +from lnt.server.reporting.analysis import UNCHANGED_PASS, UNCHANGED_FAIL
>> 
>> 
>>  class ComparisonResultTest(unittest.TestCase):
>> @@ -13,15 +13,8 @@ class ComparisonResultTest(unittest.Test
>>      def test_comp(self):
>>          """Test a real example."""
>>          curr_samples = [0.0887, 0.0919, 0.0903]
>> -        prev = 0.0858
>> -        cur = min(curr_samples)
>> -        stddev = stats.standard_deviation(curr_samples)
>> -        MAD = stats.median_absolute_deviation(curr_samples)
>> -        stddev_mean = stats.mean(curr_samples)
>> -        uninteresting = ComparisonResult(cur, prev, cur-prev,
>> -                                         (cur-prev)/prev, stddev, MAD,
>> -                                         False, False, curr_samples, [prev],
>> -                                         stddev_mean)
>> +        prev = [0.0858]
>> +        uninteresting = ComparisonResult(min, False, False, curr_samples, prev)
>> 
>>          self.assertFalse(uninteresting.is_result_interesting())
>>          self.assertEquals(uninteresting.get_test_status(), UNCHANGED_PASS)
>> @@ -29,31 +22,36 @@ class ComparisonResultTest(unittest.Test
>> 
>>      def test_slower(self):
>>          """Test getting a simple regression."""
>> -        slower = ComparisonResult(10, 5, 5, 0.5, None, None,
>> -                                  False, False, [10], [5], None)
>> +        slower = ComparisonResult(min,
>> +                                  False, False, [10], [5])
>>          self.assertEquals(slower.get_value_status(), REGRESSED)
>>          self.assertTrue(slower.is_result_interesting())
>> 
>>      def test_faster(self):
>>          """Test getting a simple improvement."""
>> 
>> -        faster = ComparisonResult(5, 10, -5, -0.5, None, None,
>> -                                  False, False, [5], [10], None)
>> +        faster = ComparisonResult(min,
>> +                                  False, False, [5], [10])
>>          self.assertEquals(faster.get_value_status(), IMPROVED)
>>          self.assertTrue(faster.is_result_interesting())
>> 
>>      def test_improved_status(self):
>>          """Test getting a test status improvement."""
>> -        improved = ComparisonResult(None, None, None, None, None, None,
>> -                                    False, True, [5], [10], None)
>> +        improved = ComparisonResult(min,
>> +                                    False, True, [1], None)
>>          self.assertEquals(improved.get_test_status(), IMPROVED)
>> 
>>      def test_regressed_status(self):
>>          """Test getting a test status improvement."""
>> -        improved = ComparisonResult(None, None, None, None, None, None,
>> -                                    True, False, [5], [10], None)
>> +        improved = ComparisonResult(min,
>> +                                    True, False, None, [10])
>>          self.assertEquals(improved.get_test_status(), REGRESSED)
>> 
>> +    def test_keep_on_failing_status(self):
>> +        """Test getting a repeated fail."""
>> +        improved = ComparisonResult(min,
>> +                                    True, True, None, None)
>> +        self.assertEquals(improved.get_test_status(), UNCHANGED_FAIL)
>> 
>>  if __name__ == '__main__':
>>      unittest.main()
>> 
>> 
>> _______________________________________________
>> llvm-commits mailing list
>> llvm-commits at cs.uiuc.edu
>> http://lists.cs.uiuc.edu/mailman/listinfo/llvm-commits
-------------- next part --------------
An HTML attachment was scrubbed...
URL: <http://lists.llvm.org/pipermail/llvm-commits/attachments/20150603/f89244bb/attachment.html>


More information about the llvm-commits mailing list