| import sys |
| import time |
|
|
| import pytest |
|
|
| from kafka.errors import QuotaViolationError |
| from kafka.metrics import DictReporter, MetricConfig, MetricName, Metrics, Quota |
| from kafka.metrics.measurable import AbstractMeasurable |
| from kafka.metrics.stats import (Avg, Count, Max, Min, Percentile, Percentiles, |
| Rate, Total) |
| from kafka.metrics.stats.percentiles import BucketSizing |
| from kafka.metrics.stats.rate import TimeUnit |
|
|
| EPS = 0.000001 |
|
|
|
|
| @pytest.fixture |
| def time_keeper(): |
| return TimeKeeper() |
|
|
|
|
| @pytest.fixture |
| def config(): |
| return MetricConfig() |
|
|
|
|
| @pytest.fixture |
| def reporter(): |
| return DictReporter() |
|
|
|
|
| @pytest.fixture |
| def metrics(request, config, reporter): |
| metrics = Metrics(config, [reporter], enable_expiration=True) |
| yield metrics |
| metrics.close() |
|
|
|
|
| def test_MetricName(): |
| |
| |
|
|
| |
| |
| name1 = MetricName('name', 'group', 'A metric.', {'a': 1, 'b': 2}) |
| name2 = MetricName('name', 'group', 'A description.', {'a': 1, 'b': 2}) |
| assert name1 == name2 |
|
|
| name1 = MetricName('name', 'group', tags={'a': 1, 'b': 2}) |
| name2 = MetricName('name', 'group', tags={'a': 1, 'b': 2}) |
| assert name1 == name2 |
|
|
| name1 = MetricName('foo', 'group') |
| name2 = MetricName('name', 'group') |
| assert name1 != name2 |
|
|
| name1 = MetricName('name', 'foo') |
| name2 = MetricName('name', 'group') |
| assert name1 != name2 |
|
|
| |
| with pytest.raises(Exception): |
| MetricName('', 'group') |
| with pytest.raises(Exception): |
| MetricName('name', None) |
| |
| with pytest.raises(Exception): |
| MetricName('name', 'group', tags=set()) |
|
|
| |
| |
| tags = {'a': 1} |
| name = MetricName('name', 'group', 'description', tags=tags) |
| with pytest.raises(AttributeError): |
| name.name = 'new name' |
| with pytest.raises(AttributeError): |
| name.group = 'new name' |
| with pytest.raises(AttributeError): |
| name.tags = {} |
| |
| name.tags['b'] = 2 |
| assert name.tags == tags |
|
|
|
|
| def test_simple_stats(mocker, time_keeper, config, metrics): |
| mocker.patch('time.time', side_effect=time_keeper.time) |
|
|
| measurable = ConstantMeasurable() |
|
|
| metrics.add_metric(metrics.metric_name('direct.measurable', 'grp1', |
| 'The fraction of time an appender waits for space allocation.'), |
| measurable) |
| sensor = metrics.sensor('test.sensor') |
| sensor.add(metrics.metric_name('test.avg', 'grp1'), Avg()) |
| sensor.add(metrics.metric_name('test.max', 'grp1'), Max()) |
| sensor.add(metrics.metric_name('test.min', 'grp1'), Min()) |
| sensor.add(metrics.metric_name('test.rate', 'grp1'), Rate(TimeUnit.SECONDS)) |
| sensor.add(metrics.metric_name('test.occurences', 'grp1'),Rate(TimeUnit.SECONDS, Count())) |
| sensor.add(metrics.metric_name('test.count', 'grp1'), Count()) |
| percentiles = [Percentile(metrics.metric_name('test.median', 'grp1'), 50.0), |
| Percentile(metrics.metric_name('test.perc99_9', 'grp1'), 99.9)] |
| sensor.add_compound(Percentiles(100, BucketSizing.CONSTANT, 100, -100, |
| percentiles=percentiles)) |
|
|
| sensor2 = metrics.sensor('test.sensor2') |
| sensor2.add(metrics.metric_name('s2.total', 'grp1'), Total()) |
| sensor2.record(5.0) |
|
|
| sum_val = 0 |
| count = 10 |
| for i in range(count): |
| sensor.record(i) |
| sum_val += i |
|
|
| |
| elapsed_secs = (config.time_window_ms * (config.samples - 1)) / 1000.0 |
| assert abs(count / elapsed_secs - |
| metrics.metrics.get(metrics.metric_name('test.occurences', 'grp1')).value()) \ |
| < EPS, 'Occurrences(0...%d) = %f' % (count, count / elapsed_secs) |
|
|
| |
| sleep_time_seconds = 2.0 |
| time_keeper.sleep(sleep_time_seconds) |
| elapsed_secs += sleep_time_seconds |
|
|
| assert abs(5.0 - metrics.metrics.get(metrics.metric_name('s2.total', 'grp1')).value()) \ |
| < EPS, 's2 reflects the constant value' |
| assert abs(4.5 - metrics.metrics.get(metrics.metric_name('test.avg', 'grp1')).value()) \ |
| < EPS, 'Avg(0...9) = 4.5' |
| assert abs((count - 1) - metrics.metrics.get(metrics.metric_name('test.max', 'grp1')).value()) \ |
| < EPS, 'Max(0...9) = 9' |
| assert abs(0.0 - metrics.metrics.get(metrics.metric_name('test.min', 'grp1')).value()) \ |
| < EPS, 'Min(0...9) = 0' |
| assert abs((sum_val / elapsed_secs) - metrics.metrics.get(metrics.metric_name('test.rate', 'grp1')).value()) \ |
| < EPS, 'Rate(0...9) = 1.40625' |
| assert abs((count / elapsed_secs) - metrics.metrics.get(metrics.metric_name('test.occurences', 'grp1')).value()) \ |
| < EPS, 'Occurrences(0...%d) = %f' % (count, count / elapsed_secs) |
| assert abs(count - metrics.metrics.get(metrics.metric_name('test.count', 'grp1')).value()) \ |
| < EPS, 'Count(0...9) = 10' |
|
|
|
|
| def test_hierarchical_sensors(metrics): |
| parent1 = metrics.sensor('test.parent1') |
| parent1.add(metrics.metric_name('test.parent1.count', 'grp1'), Count()) |
| parent2 = metrics.sensor('test.parent2') |
| parent2.add(metrics.metric_name('test.parent2.count', 'grp1'), Count()) |
| child1 = metrics.sensor('test.child1', parents=[parent1, parent2]) |
| child1.add(metrics.metric_name('test.child1.count', 'grp1'), Count()) |
| child2 = metrics.sensor('test.child2', parents=[parent1]) |
| child2.add(metrics.metric_name('test.child2.count', 'grp1'), Count()) |
| grandchild = metrics.sensor('test.grandchild', parents=[child1]) |
| grandchild.add(metrics.metric_name('test.grandchild.count', 'grp1'), Count()) |
|
|
| |
| parent1.record() |
| parent2.record() |
| child1.record() |
| child2.record() |
| grandchild.record() |
|
|
| p1 = parent1.metrics[0].value() |
| p2 = parent2.metrics[0].value() |
| c1 = child1.metrics[0].value() |
| c2 = child2.metrics[0].value() |
| gc = grandchild.metrics[0].value() |
|
|
| |
| assert 1.0 == gc |
| assert 1.0 + gc == c1 |
| assert 1.0 == c2 |
| assert 1.0 + c1 == p2 |
| assert 1.0 + c1 + c2 == p1 |
| assert [child1, child2] == metrics._children_sensors.get(parent1) |
| assert [child1] == metrics._children_sensors.get(parent2) |
| assert metrics._children_sensors.get(grandchild) is None |
|
|
|
|
| def test_bad_sensor_hierarchy(metrics): |
| parent = metrics.sensor('parent') |
| child1 = metrics.sensor('child1', parents=[parent]) |
| child2 = metrics.sensor('child2', parents=[parent]) |
|
|
| with pytest.raises(ValueError): |
| metrics.sensor('gc', parents=[child1, child2]) |
|
|
|
|
| def test_remove_sensor(metrics): |
| size = len(metrics.metrics) |
| parent1 = metrics.sensor('test.parent1') |
| parent1.add(metrics.metric_name('test.parent1.count', 'grp1'), Count()) |
| parent2 = metrics.sensor('test.parent2') |
| parent2.add(metrics.metric_name('test.parent2.count', 'grp1'), Count()) |
| child1 = metrics.sensor('test.child1', parents=[parent1, parent2]) |
| child1.add(metrics.metric_name('test.child1.count', 'grp1'), Count()) |
| child2 = metrics.sensor('test.child2', parents=[parent2]) |
| child2.add(metrics.metric_name('test.child2.count', 'grp1'), Count()) |
| grandchild1 = metrics.sensor('test.gchild2', parents=[child2]) |
| grandchild1.add(metrics.metric_name('test.gchild2.count', 'grp1'), Count()) |
|
|
| sensor = metrics.get_sensor('test.parent1') |
| assert sensor is not None |
| metrics.remove_sensor('test.parent1') |
| assert metrics.get_sensor('test.parent1') is None |
| assert metrics.metrics.get(metrics.metric_name('test.parent1.count', 'grp1')) is None |
| assert metrics.get_sensor('test.child1') is None |
| assert metrics._children_sensors.get(sensor) is None |
| assert metrics.metrics.get(metrics.metric_name('test.child1.count', 'grp1')) is None |
|
|
| sensor = metrics.get_sensor('test.gchild2') |
| assert sensor is not None |
| metrics.remove_sensor('test.gchild2') |
| assert metrics.get_sensor('test.gchild2') is None |
| assert metrics._children_sensors.get(sensor) is None |
| assert metrics.metrics.get(metrics.metric_name('test.gchild2.count', 'grp1')) is None |
|
|
| sensor = metrics.get_sensor('test.child2') |
| assert sensor is not None |
| metrics.remove_sensor('test.child2') |
| assert metrics.get_sensor('test.child2') is None |
| assert metrics._children_sensors.get(sensor) is None |
| assert metrics.metrics.get(metrics.metric_name('test.child2.count', 'grp1')) is None |
|
|
| sensor = metrics.get_sensor('test.parent2') |
| assert sensor is not None |
| metrics.remove_sensor('test.parent2') |
| assert metrics.get_sensor('test.parent2') is None |
| assert metrics._children_sensors.get(sensor) is None |
| assert metrics.metrics.get(metrics.metric_name('test.parent2.count', 'grp1')) is None |
|
|
| assert size == len(metrics.metrics) |
|
|
|
|
| def test_remove_inactive_metrics(mocker, time_keeper, metrics): |
| mocker.patch('time.time', side_effect=time_keeper.time) |
|
|
| s1 = metrics.sensor('test.s1', None, 1) |
| s1.add(metrics.metric_name('test.s1.count', 'grp1'), Count()) |
|
|
| s2 = metrics.sensor('test.s2', None, 3) |
| s2.add(metrics.metric_name('test.s2.count', 'grp1'), Count()) |
|
|
| purger = Metrics.ExpireSensorTask |
| purger.run(metrics) |
| assert metrics.get_sensor('test.s1') is not None, \ |
| 'Sensor test.s1 must be present' |
| assert metrics.metrics.get(metrics.metric_name('test.s1.count', 'grp1')) is not None, \ |
| 'MetricName test.s1.count must be present' |
| assert metrics.get_sensor('test.s2') is not None, \ |
| 'Sensor test.s2 must be present' |
| assert metrics.metrics.get(metrics.metric_name('test.s2.count', 'grp1')) is not None, \ |
| 'MetricName test.s2.count must be present' |
|
|
| time_keeper.sleep(1.001) |
| purger.run(metrics) |
| assert metrics.get_sensor('test.s1') is None, \ |
| 'Sensor test.s1 should have been purged' |
| assert metrics.metrics.get(metrics.metric_name('test.s1.count', 'grp1')) is None, \ |
| 'MetricName test.s1.count should have been purged' |
| assert metrics.get_sensor('test.s2') is not None, \ |
| 'Sensor test.s2 must be present' |
| assert metrics.metrics.get(metrics.metric_name('test.s2.count', 'grp1')) is not None, \ |
| 'MetricName test.s2.count must be present' |
|
|
| |
| |
| s2.record() |
|
|
| time_keeper.sleep(2) |
| purger.run(metrics) |
| assert metrics.get_sensor('test.s2') is not None, \ |
| 'Sensor test.s2 must be present' |
| assert metrics.metrics.get(metrics.metric_name('test.s2.count', 'grp1')) is not None, \ |
| 'MetricName test.s2.count must be present' |
|
|
| |
| time_keeper.sleep(1) |
| purger.run(metrics) |
| assert metrics.get_sensor('test.s1') is None, \ |
| 'Sensor test.s2 should have been purged' |
| assert metrics.metrics.get(metrics.metric_name('test.s1.count', 'grp1')) is None, \ |
| 'MetricName test.s2.count should have been purged' |
|
|
| |
| s1 = metrics.sensor('test.s1', None, 1) |
| s1.add(metrics.metric_name('test.s1.count', 'grp1'), Count()) |
| assert metrics.get_sensor('test.s1') is not None, \ |
| 'Sensor test.s1 must be present' |
| assert metrics.metrics.get(metrics.metric_name('test.s1.count', 'grp1')) is not None, \ |
| 'MetricName test.s1.count must be present' |
|
|
|
|
| def test_remove_metric(metrics): |
| size = len(metrics.metrics) |
| metrics.add_metric(metrics.metric_name('test1', 'grp1'), Count()) |
| metrics.add_metric(metrics.metric_name('test2', 'grp1'), Count()) |
|
|
| assert metrics.remove_metric(metrics.metric_name('test1', 'grp1')) is not None |
| assert metrics.metrics.get(metrics.metric_name('test1', 'grp1')) is None |
| assert metrics.metrics.get(metrics.metric_name('test2', 'grp1')) is not None |
|
|
| assert metrics.remove_metric(metrics.metric_name('test2', 'grp1')) is not None |
| assert metrics.metrics.get(metrics.metric_name('test2', 'grp1')) is None |
|
|
| assert size == len(metrics.metrics) |
|
|
|
|
| def test_event_windowing(mocker, time_keeper): |
| mocker.patch('time.time', side_effect=time_keeper.time) |
|
|
| count = Count() |
| config = MetricConfig(event_window=1, samples=2) |
| count.record(config, 1.0, time_keeper.ms()) |
| count.record(config, 1.0, time_keeper.ms()) |
| assert 2.0 == count.measure(config, time_keeper.ms()) |
| count.record(config, 1.0, time_keeper.ms()) |
| assert 2.0 == count.measure(config, time_keeper.ms()) |
|
|
|
|
| def test_time_windowing(mocker, time_keeper): |
| mocker.patch('time.time', side_effect=time_keeper.time) |
|
|
| count = Count() |
| config = MetricConfig(time_window_ms=1, samples=2) |
| count.record(config, 1.0, time_keeper.ms()) |
| time_keeper.sleep(.001) |
| count.record(config, 1.0, time_keeper.ms()) |
| assert 2.0 == count.measure(config, time_keeper.ms()) |
| time_keeper.sleep(.001) |
| count.record(config, 1.0, time_keeper.ms()) |
| assert 2.0 == count.measure(config, time_keeper.ms()) |
|
|
|
|
| def test_old_data_has_no_effect(mocker, time_keeper): |
| mocker.patch('time.time', side_effect=time_keeper.time) |
|
|
| max_stat = Max() |
| min_stat = Min() |
| avg_stat = Avg() |
| count_stat = Count() |
| window_ms = 100 |
| samples = 2 |
| config = MetricConfig(time_window_ms=window_ms, samples=samples) |
| max_stat.record(config, 50, time_keeper.ms()) |
| min_stat.record(config, 50, time_keeper.ms()) |
| avg_stat.record(config, 50, time_keeper.ms()) |
| count_stat.record(config, 50, time_keeper.ms()) |
|
|
| time_keeper.sleep(samples * window_ms / 1000.0) |
| assert float('-inf') == max_stat.measure(config, time_keeper.ms()) |
| assert float(sys.maxsize) == min_stat.measure(config, time_keeper.ms()) |
| assert 0.0 == avg_stat.measure(config, time_keeper.ms()) |
| assert 0 == count_stat.measure(config, time_keeper.ms()) |
|
|
|
|
| def test_duplicate_MetricName(metrics): |
| metrics.sensor('test').add(metrics.metric_name('test', 'grp1'), Avg()) |
| with pytest.raises(ValueError): |
| metrics.sensor('test2').add(metrics.metric_name('test', 'grp1'), Total()) |
|
|
|
|
| def test_Quotas(metrics): |
| sensor = metrics.sensor('test') |
| sensor.add(metrics.metric_name('test1.total', 'grp1'), Total(), |
| MetricConfig(quota=Quota.upper_bound(5.0))) |
| sensor.add(metrics.metric_name('test2.total', 'grp1'), Total(), |
| MetricConfig(quota=Quota.lower_bound(0.0))) |
| sensor.record(5.0) |
| with pytest.raises(QuotaViolationError): |
| sensor.record(1.0) |
|
|
| assert abs(6.0 - metrics.metrics.get(metrics.metric_name('test1.total', 'grp1')).value()) \ |
| < EPS |
|
|
| sensor.record(-6.0) |
| with pytest.raises(QuotaViolationError): |
| sensor.record(-1.0) |
|
|
|
|
| def test_Quotas_equality(): |
| quota1 = Quota.upper_bound(10.5) |
| quota2 = Quota.lower_bound(10.5) |
| assert quota1 != quota2, 'Quota with different upper values should not be equal' |
|
|
| quota3 = Quota.lower_bound(10.5) |
| assert quota2 == quota3, 'Quota with same upper and bound values should be equal' |
|
|
|
|
| def test_Percentiles(metrics): |
| buckets = 100 |
| _percentiles = [ |
| Percentile(metrics.metric_name('test.p25', 'grp1'), 25), |
| Percentile(metrics.metric_name('test.p50', 'grp1'), 50), |
| Percentile(metrics.metric_name('test.p75', 'grp1'), 75), |
| ] |
| percs = Percentiles(4 * buckets, BucketSizing.CONSTANT, 100.0, 0.0, |
| percentiles=_percentiles) |
| config = MetricConfig(event_window=50, samples=2) |
| sensor = metrics.sensor('test', config) |
| sensor.add_compound(percs) |
| p25 = metrics.metrics.get(metrics.metric_name('test.p25', 'grp1')) |
| p50 = metrics.metrics.get(metrics.metric_name('test.p50', 'grp1')) |
| p75 = metrics.metrics.get(metrics.metric_name('test.p75', 'grp1')) |
|
|
| |
| for i in range(buckets): |
| sensor.record(i) |
|
|
| assert abs(p25.value() - 25) < 1.0 |
| assert abs(p50.value() - 50) < 1.0 |
| assert abs(p75.value() - 75) < 1.0 |
|
|
| for i in range(buckets): |
| sensor.record(0.0) |
|
|
| assert p25.value() < 1.0 |
| assert p50.value() < 1.0 |
| assert p75.value() < 1.0 |
|
|
| def test_rate_windowing(mocker, time_keeper, metrics): |
| mocker.patch('time.time', side_effect=time_keeper.time) |
|
|
| |
| config = MetricConfig(samples=3) |
| sensor = metrics.sensor('test.sensor', config) |
| sensor.add(metrics.metric_name('test.rate', 'grp1'), Rate(TimeUnit.SECONDS)) |
|
|
| sum_val = 0 |
| count = config.samples - 1 |
| |
| for i in range(count): |
| sensor.record(100) |
| sum_val += 100 |
| time_keeper.sleep(config.time_window_ms / 1000.0) |
|
|
| |
| time_keeper.sleep(config.time_window_ms / 2.0 / 1000.0) |
|
|
| |
| elapsed_secs = (config.time_window_ms * (config.samples - 1) + config.time_window_ms / 2.0) / 1000.0 |
|
|
| kafka_metric = metrics.metrics.get(metrics.metric_name('test.rate', 'grp1')) |
| assert abs((sum_val / elapsed_secs) - kafka_metric.value()) < EPS, \ |
| 'Rate(0...2) = 2.666' |
| assert abs(elapsed_secs - (kafka_metric.measurable.window_size(config, time.time() * 1000) / 1000.0)) \ |
| < EPS, 'Elapsed Time = 75 seconds' |
|
|
|
|
| def test_reporter(metrics): |
| reporter = DictReporter() |
| foo_reporter = DictReporter(prefix='foo') |
| metrics.add_reporter(reporter) |
| metrics.add_reporter(foo_reporter) |
| sensor = metrics.sensor('kafka.requests') |
| sensor.add(metrics.metric_name('pack.bean1.avg', 'grp1'), Avg()) |
| sensor.add(metrics.metric_name('pack.bean2.total', 'grp2'), Total()) |
| sensor2 = metrics.sensor('kafka.blah') |
| sensor2.add(metrics.metric_name('pack.bean1.some', 'grp1'), Total()) |
| sensor2.add(metrics.metric_name('pack.bean2.some', 'grp1', |
| tags={'a': 42, 'b': 'bar'}), Total()) |
|
|
| |
| expected = { |
| 'kafka-metrics-count': {'count': 5.0}, |
| 'grp2': {'pack.bean2.total': 0.0}, |
| 'grp1': {'pack.bean1.avg': 0.0, 'pack.bean1.some': 0.0}, |
| 'grp1.a=42,b=bar': {'pack.bean2.some': 0.0}, |
| } |
| assert expected == reporter.snapshot() |
|
|
| for key in list(expected.keys()): |
| metrics = expected.pop(key) |
| expected['foo.%s' % (key,)] = metrics |
| assert expected == foo_reporter.snapshot() |
|
|
|
|
| class ConstantMeasurable(AbstractMeasurable): |
| _value = 0.0 |
|
|
| def measure(self, config, now): |
| return self._value |
|
|
|
|
| class TimeKeeper(object): |
| """ |
| A clock that you can manually advance by calling sleep |
| """ |
| def __init__(self, auto_tick_ms=0): |
| self._millis = time.time() * 1000 |
| self._auto_tick_ms = auto_tick_ms |
|
|
| def time(self): |
| return self.ms() / 1000.0 |
|
|
| def ms(self): |
| self.sleep(self._auto_tick_ms) |
| return self._millis |
|
|
| def sleep(self, seconds): |
| self._millis += (seconds * 1000) |
|
|