| 123456789101112131415161718192021222324252627282930313233343536373839404142434445464748495051525354555657585960616263646566676869707172737475767778798081828384858687888990919293949596979899100101102103104105106107108109110111112113114115116117118119120121122123124125126127128129130131132133134135136137138139140141142143144145146147148149150151152153154155156157158159160161162163164165166167168169170171172173174175176177178179180181182183184185186187188189190191192193194195196197198199200201202203204205206207208209210211212213214215216217218219220221222223224225226227228229230231232233234235236237238239240241242243244245246247248249250251252253254255256257258259260261262263264265266267268269270271272273274275276277278279280281282283284285286287288 | //  Copyright (c) 2011-present, Facebook, Inc.  All rights reserved.//  This source code is licensed under both the GPLv2 (found in the//  COPYING file in the root directory) and Apache 2.0 License//  (found in the LICENSE.Apache file in the root directory).//// Copyright (c) 2011 The LevelDB Authors. All rights reserved.// Use of this source code is governed by a BSD-style license that can be// found in the LICENSE file. See the AUTHORS file for names of contributors.#include "monitoring/histogram.h"#include <stdio.h>#include <cassert>#include <cinttypes>#include <cmath>#include "port/port.h"#include "util/cast_util.h"namespace ROCKSDB_NAMESPACE {HistogramBucketMapper::HistogramBucketMapper() {  // If you change this, you also need to change  // size of array buckets_ in HistogramImpl  bucketValues_ = {1, 2};  valueIndexMap_ = {{1, 0}, {2, 1}};  double bucket_val = static_cast<double>(bucketValues_.back());  while ((bucket_val = 1.5 * bucket_val) <= static_cast<double>(port::kMaxUint64)) {    bucketValues_.push_back(static_cast<uint64_t>(bucket_val));    // Extracts two most significant digits to make histogram buckets more    // human-readable. E.g., 172 becomes 170.    uint64_t pow_of_ten = 1;    while (bucketValues_.back() / 10 > 10) {      bucketValues_.back() /= 10;      pow_of_ten *= 10;    }    bucketValues_.back() *= pow_of_ten;    valueIndexMap_[bucketValues_.back()] = bucketValues_.size() - 1;  }  maxBucketValue_ = bucketValues_.back();  minBucketValue_ = bucketValues_.front();}size_t HistogramBucketMapper::IndexForValue(const uint64_t value) const {  if (value >= maxBucketValue_) {    return bucketValues_.size() - 1;  } else if ( value >= minBucketValue_ ) {    std::map<uint64_t, uint64_t>::const_iterator lowerBound =      valueIndexMap_.lower_bound(value);    if (lowerBound != valueIndexMap_.end()) {      return static_cast<size_t>(lowerBound->second);    } else {      return 0;    }  } else {    return 0;  }}namespace {  const HistogramBucketMapper bucketMapper;}HistogramStat::HistogramStat()  : num_buckets_(bucketMapper.BucketCount()) {  assert(num_buckets_ == sizeof(buckets_) / sizeof(*buckets_));  Clear();}void HistogramStat::Clear() {  min_.store(bucketMapper.LastValue(), std::memory_order_relaxed);  max_.store(0, std::memory_order_relaxed);  num_.store(0, std::memory_order_relaxed);  sum_.store(0, std::memory_order_relaxed);  sum_squares_.store(0, std::memory_order_relaxed);  for (unsigned int b = 0; b < num_buckets_; b++) {    buckets_[b].store(0, std::memory_order_relaxed);  }};bool HistogramStat::Empty() const { return num() == 0; }void HistogramStat::Add(uint64_t value) {  // This function is designed to be lock free, as it's in the critical path  // of any operation. Each individual value is atomic and the order of updates  // by concurrent threads is tolerable.  const size_t index = bucketMapper.IndexForValue(value);  assert(index < num_buckets_);  buckets_[index].store(buckets_[index].load(std::memory_order_relaxed) + 1,                        std::memory_order_relaxed);  uint64_t old_min = min();  if (value < old_min) {    min_.store(value, std::memory_order_relaxed);  }  uint64_t old_max = max();  if (value > old_max) {    max_.store(value, std::memory_order_relaxed);  }  num_.store(num_.load(std::memory_order_relaxed) + 1,             std::memory_order_relaxed);  sum_.store(sum_.load(std::memory_order_relaxed) + value,             std::memory_order_relaxed);  sum_squares_.store(      sum_squares_.load(std::memory_order_relaxed) + value * value,      std::memory_order_relaxed);}void HistogramStat::Merge(const HistogramStat& other) {  // This function needs to be performned with the outer lock acquired  // However, atomic operation on every member is still need, since Add()  // requires no lock and value update can still happen concurrently  uint64_t old_min = min();  uint64_t other_min = other.min();  while (other_min < old_min &&         !min_.compare_exchange_weak(old_min, other_min)) {}  uint64_t old_max = max();  uint64_t other_max = other.max();  while (other_max > old_max &&         !max_.compare_exchange_weak(old_max, other_max)) {}  num_.fetch_add(other.num(), std::memory_order_relaxed);  sum_.fetch_add(other.sum(), std::memory_order_relaxed);  sum_squares_.fetch_add(other.sum_squares(), std::memory_order_relaxed);  for (unsigned int b = 0; b < num_buckets_; b++) {    buckets_[b].fetch_add(other.bucket_at(b), std::memory_order_relaxed);  }}double HistogramStat::Median() const {  return Percentile(50.0);}double HistogramStat::Percentile(double p) const {  double threshold = num() * (p / 100.0);  uint64_t cumulative_sum = 0;  for (unsigned int b = 0; b < num_buckets_; b++) {    uint64_t bucket_value = bucket_at(b);    cumulative_sum += bucket_value;    if (cumulative_sum >= threshold) {      // Scale linearly within this bucket      uint64_t left_point = (b == 0) ? 0 : bucketMapper.BucketLimit(b-1);      uint64_t right_point = bucketMapper.BucketLimit(b);      uint64_t left_sum = cumulative_sum - bucket_value;      uint64_t right_sum = cumulative_sum;      double pos = 0;      uint64_t right_left_diff = right_sum - left_sum;      if (right_left_diff != 0) {       pos = (threshold - left_sum) / right_left_diff;      }      double r = left_point + (right_point - left_point) * pos;      uint64_t cur_min = min();      uint64_t cur_max = max();      if (r < cur_min) r = static_cast<double>(cur_min);      if (r > cur_max) r = static_cast<double>(cur_max);      return r;    }  }  return static_cast<double>(max());}double HistogramStat::Average() const {  uint64_t cur_num = num();  uint64_t cur_sum = sum();  if (cur_num == 0) return 0;  return static_cast<double>(cur_sum) / static_cast<double>(cur_num);}double HistogramStat::StandardDeviation() const {  uint64_t cur_num = num();  uint64_t cur_sum = sum();  uint64_t cur_sum_squares = sum_squares();  if (cur_num == 0) return 0;  double variance =      static_cast<double>(cur_sum_squares * cur_num - cur_sum * cur_sum) /      static_cast<double>(cur_num * cur_num);  return std::sqrt(variance);}std::string HistogramStat::ToString() const {  uint64_t cur_num = num();  std::string r;  char buf[1650];  snprintf(buf, sizeof(buf),           "Count: %" PRIu64 " Average: %.4f  StdDev: %.2f\n",           cur_num, Average(), StandardDeviation());  r.append(buf);  snprintf(buf, sizeof(buf),           "Min: %" PRIu64 "  Median: %.4f  Max: %" PRIu64 "\n",           (cur_num == 0 ? 0 : min()), Median(), (cur_num == 0 ? 0 : max()));  r.append(buf);  snprintf(buf, sizeof(buf),           "Percentiles: "           "P50: %.2f P75: %.2f P99: %.2f P99.9: %.2f P99.99: %.2f\n",           Percentile(50), Percentile(75), Percentile(99), Percentile(99.9),           Percentile(99.99));  r.append(buf);  r.append("------------------------------------------------------\n");  if (cur_num == 0) return r;   // all buckets are empty  const double mult = 100.0 / cur_num;  uint64_t cumulative_sum = 0;  for (unsigned int b = 0; b < num_buckets_; b++) {    uint64_t bucket_value = bucket_at(b);    if (bucket_value <= 0.0) continue;    cumulative_sum += bucket_value;    snprintf(buf, sizeof(buf),             "%c %7" PRIu64 ", %7" PRIu64 " ] %8" PRIu64 " %7.3f%% %7.3f%% ",             (b == 0) ? '[' : '(',             (b == 0) ? 0 : bucketMapper.BucketLimit(b-1),  // left              bucketMapper.BucketLimit(b),  // right              bucket_value,                   // count             (mult * bucket_value),           // percentage             (mult * cumulative_sum));       // cumulative percentage    r.append(buf);    // Add hash marks based on percentage; 20 marks for 100%.    size_t marks = static_cast<size_t>(mult * bucket_value / 5 + 0.5);    r.append(marks, '#');    r.push_back('\n');  }  return r;}void HistogramStat::Data(HistogramData * const data) const {  assert(data);  data->median = Median();  data->percentile95 = Percentile(95);  data->percentile99 = Percentile(99);  data->max = static_cast<double>(max());  data->average = Average();  data->standard_deviation = StandardDeviation();  data->count = num();  data->sum = sum();  data->min = static_cast<double>(min());}void HistogramImpl::Clear() {  std::lock_guard<std::mutex> lock(mutex_);  stats_.Clear();}bool HistogramImpl::Empty() const {  return stats_.Empty();}void HistogramImpl::Add(uint64_t value) {  stats_.Add(value);}void HistogramImpl::Merge(const Histogram& other) {  if (strcmp(Name(), other.Name()) == 0) {    Merge(        *static_cast_with_check<const HistogramImpl, const Histogram>(&other));  }}void HistogramImpl::Merge(const HistogramImpl& other) {  std::lock_guard<std::mutex> lock(mutex_);  stats_.Merge(other.stats_);}double HistogramImpl::Median() const {  return stats_.Median();}double HistogramImpl::Percentile(double p) const {  return stats_.Percentile(p);}double HistogramImpl::Average() const {  return stats_.Average();}double HistogramImpl::StandardDeviation() const { return stats_.StandardDeviation();}std::string HistogramImpl::ToString() const {  return stats_.ToString();}void HistogramImpl::Data(HistogramData * const data) const {  stats_.Data(data);}}  // namespace ROCKSDB_NAMESPACE
 |