2010-09-28 12:35:53 +02:00
|
|
|
// { dg-options "-std=gnu++0x" }
|
|
|
|
|
2011-03-01 01:40:53 +01:00
|
|
|
// Use smaller statistics when running on simulators, so it takes less time.
|
|
|
|
// { dg-options "-std=gnu++0x -DSAMPLES=10000" { target simulator } }
|
|
|
|
|
|
|
|
// Copyright (C) 2010, 2011 Free Software Foundation, Inc.
|
2010-09-28 12:35:53 +02:00
|
|
|
//
|
|
|
|
// This file is part of the GNU ISO C++ Library. This library is free
|
|
|
|
// software; you can redistribute it and/or modify it under the
|
|
|
|
// terms of the GNU General Public License as published by the
|
|
|
|
// Free Software Foundation; either version 3, or (at your option)
|
|
|
|
// any later version.
|
|
|
|
//
|
|
|
|
// This library is distributed in the hope that it will be useful,
|
|
|
|
// but WITHOUT ANY WARRANTY; without even the implied warranty of
|
|
|
|
// MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the
|
|
|
|
// GNU General Public License for more details.
|
|
|
|
//
|
|
|
|
// You should have received a copy of the GNU General Public License
|
|
|
|
// along with this library; see the file COPYING3. If not see
|
|
|
|
// <http://www.gnu.org/licenses/>.
|
|
|
|
|
|
|
|
// This file uses the chi^2 test to measure the quality of a hash
|
|
|
|
// function, by computing the uniformity with which it distributes a set
|
|
|
|
// of N strings into k buckets (where k is significantly greater than N).
|
|
|
|
//
|
|
|
|
// Each bucket has B[i] strings in it. The expected value of each bucket
|
|
|
|
// for a uniform distribution is z = N/k, so
|
|
|
|
// chi^2 = Sum_i (B[i] - z)^2 / z.
|
|
|
|
//
|
|
|
|
// We check whether chi^2 is small enough to be consistent with the
|
|
|
|
// hypothesis of a uniform distribution. If F(chi^2, k-1) is close to
|
|
|
|
// 0 (where F is the cumulative probability distribution), we can
|
|
|
|
// reject that hypothesis. So we don't want F to be too small, which
|
|
|
|
// for large k, means we want chi^2 to be not too much larger than k.
|
|
|
|
//
|
|
|
|
// We use the chi^2 test for several sets of strings. Any non-horrible
|
|
|
|
// hash function should do well with purely random strings. A really
|
|
|
|
// good hash function will also do well with more structured sets,
|
|
|
|
// including ones where the strings differ by only a few bits.
|
|
|
|
|
|
|
|
#include <algorithm>
|
|
|
|
#include <cstdlib>
|
|
|
|
#include <cstdio>
|
|
|
|
#include <fstream>
|
|
|
|
#include <functional>
|
|
|
|
#include <iostream>
|
|
|
|
#include <iterator>
|
|
|
|
#include <string>
|
|
|
|
#include <unordered_set>
|
|
|
|
#include <vector>
|
|
|
|
#include <testsuite_hooks.h>
|
|
|
|
|
|
|
|
#ifndef SAMPLES
|
|
|
|
#define SAMPLES 300000
|
|
|
|
#endif
|
|
|
|
|
|
|
|
template <typename Container>
|
|
|
|
double
|
|
|
|
chi2_hash(const Container& c, long buckets)
|
|
|
|
{
|
|
|
|
std::vector<int> counts(buckets);
|
|
|
|
std::hash<std::string> hasher;
|
|
|
|
double elements = 0;
|
|
|
|
for (auto i = c.begin(); i != c.end(); ++i)
|
|
|
|
{
|
|
|
|
++counts[hasher(*i) % buckets];
|
|
|
|
++elements;
|
|
|
|
}
|
|
|
|
|
|
|
|
const double z = elements / buckets;
|
|
|
|
double sum = 0;
|
|
|
|
for (long i = 0; i < buckets; ++i)
|
|
|
|
{
|
|
|
|
double delta = counts[i] - z;
|
|
|
|
sum += delta*delta;
|
|
|
|
}
|
|
|
|
return sum/z;
|
|
|
|
}
|
|
|
|
|
|
|
|
// Tests chi^2 for a distribution of uniformly generated random strings.
|
|
|
|
void
|
|
|
|
test_uniform_random()
|
|
|
|
{
|
|
|
|
bool test __attribute__((unused)) = true;
|
|
|
|
std::srand(137);
|
|
|
|
std::unordered_set<std::string> set;
|
|
|
|
std::string s;
|
|
|
|
const unsigned long N = SAMPLES;
|
|
|
|
const unsigned long k = N/100;
|
|
|
|
const unsigned int len = 25;
|
|
|
|
while (set.size() < N)
|
|
|
|
{
|
|
|
|
s.clear();
|
2010-11-01 01:08:58 +01:00
|
|
|
for (unsigned int i = 0; i < len; ++i)
|
|
|
|
s.push_back(rand() % 128);
|
2010-09-28 12:35:53 +02:00
|
|
|
set.insert(s);
|
|
|
|
}
|
|
|
|
|
|
|
|
double chi2 = chi2_hash(set, k);
|
|
|
|
VERIFY( chi2 < k*1.1 );
|
|
|
|
}
|
|
|
|
|
|
|
|
// Tests chi^2 for a distribution of strings that differ from each
|
|
|
|
// other by only a few bits. We start with an arbitrary base string, and
|
|
|
|
// flip three random bits for each member of the set.
|
|
|
|
void
|
|
|
|
test_bit_flip_set()
|
|
|
|
{
|
|
|
|
bool test __attribute__((unused)) = true;
|
|
|
|
const unsigned long N = SAMPLES;
|
|
|
|
const unsigned long k = N/100;
|
|
|
|
const unsigned int len = 67;
|
|
|
|
const unsigned int bitlen = len * 8;
|
|
|
|
const unsigned int bits_to_flip = 3;
|
|
|
|
const char base[len+1] = "abcdefghijklmnopqrstuvwxyz"
|
|
|
|
"ABCDEFGHIJKLMNOPQRSTUVWXYZ"
|
|
|
|
"0123456789!@#$%";
|
|
|
|
|
|
|
|
std::unordered_set<std::string> set;
|
|
|
|
while (set.size() < N)
|
|
|
|
{
|
|
|
|
std::string s(base, base+len);
|
2010-11-01 01:08:58 +01:00
|
|
|
for (unsigned int i = 0; i < bits_to_flip; ++i)
|
2010-09-28 12:35:53 +02:00
|
|
|
{
|
|
|
|
int bit = rand() % bitlen;
|
|
|
|
s[bit/8] ^= (1 << (bit%8));
|
|
|
|
}
|
|
|
|
set.insert(s);
|
|
|
|
}
|
|
|
|
|
|
|
|
double chi2 = chi2_hash(set, k);
|
|
|
|
VERIFY( chi2 < k*1.1 );
|
|
|
|
}
|
|
|
|
|
|
|
|
// Tests chi^2 of a set of strings that all have a similar pattern,
|
|
|
|
// intended to mimic some sort of ID string.
|
|
|
|
void
|
|
|
|
test_numeric_pattern_set()
|
|
|
|
{
|
|
|
|
bool test __attribute__((unused)) = true;
|
|
|
|
const unsigned long N = SAMPLES;
|
|
|
|
const unsigned long k = N/100;
|
|
|
|
std::vector<std::string> set;
|
|
|
|
for (unsigned long i = 0; i < N; ++i)
|
|
|
|
{
|
|
|
|
long i1 = i % 100000;
|
|
|
|
long i2 = i / 100000;
|
|
|
|
char buf[16];
|
|
|
|
std::sprintf(buf, "XX-%05lu-%05lu", i1, i2);
|
|
|
|
set.push_back(buf);
|
|
|
|
}
|
|
|
|
|
|
|
|
double chi2 = chi2_hash(set, k);
|
|
|
|
VERIFY( chi2 < k*1.1 );
|
|
|
|
}
|
|
|
|
|
|
|
|
// Tests chi^2 for a set of strings that all consist of '1' and '0'.
|
|
|
|
void
|
|
|
|
test_bit_string_set()
|
|
|
|
{
|
|
|
|
bool test __attribute__((unused)) = true;
|
|
|
|
const unsigned long N = SAMPLES;
|
|
|
|
const unsigned long k = N/100;
|
|
|
|
std::vector<std::string> set;
|
|
|
|
std::string s;
|
|
|
|
for (unsigned long i = 0; i < N; ++i)
|
|
|
|
{
|
|
|
|
s.clear();
|
2010-11-01 01:08:58 +01:00
|
|
|
for (unsigned int j = 0; j < sizeof(unsigned long) * 8; ++j)
|
2010-09-28 12:35:53 +02:00
|
|
|
{
|
|
|
|
const bool bit = (1UL << j) & i;
|
|
|
|
s.push_back(bit ? '1' : '0');
|
|
|
|
}
|
|
|
|
set.push_back(s);
|
|
|
|
}
|
|
|
|
|
|
|
|
double chi2 = chi2_hash(set, k);
|
|
|
|
VERIFY( chi2 < k*1.1 );
|
|
|
|
}
|
|
|
|
|
|
|
|
// Tests chi^2 for a set of words taken from a document written in English.
|
|
|
|
void
|
|
|
|
test_document_words()
|
|
|
|
{
|
2011-04-19 05:59:16 +02:00
|
|
|
// That file is 187587 single-word lines. To avoid a timeout, just skip
|
|
|
|
// this part, which would take up to 95% of the program runtime (with
|
|
|
|
// SAMPLES == 10000), if we're not supposed to run anywhere that long.
|
|
|
|
#if SAMPLES >= 100000
|
2010-09-28 12:35:53 +02:00
|
|
|
bool test __attribute__((unused)) = true;
|
|
|
|
const std::string f_name = "thirty_years_among_the_dead_preproc.txt";
|
|
|
|
std::ifstream in(f_name);
|
|
|
|
VERIFY( in.is_open() );
|
|
|
|
std::vector<std::string> words;
|
|
|
|
words.assign(std::istream_iterator<std::string>(in),
|
|
|
|
std::istream_iterator<std::string>());
|
|
|
|
VERIFY( words.size() > 100000 );
|
|
|
|
std::sort(words.begin(), words.end());
|
|
|
|
auto it = std::unique(words.begin(), words.end());
|
|
|
|
words.erase(it, words.end());
|
|
|
|
VERIFY( words.size() > 5000 );
|
|
|
|
|
|
|
|
const unsigned long k = words.size() / 20;
|
|
|
|
double chi2 = chi2_hash(words, k);
|
|
|
|
VERIFY( chi2 < k*1.1 );
|
2011-04-19 05:59:16 +02:00
|
|
|
#endif
|
2010-09-28 12:35:53 +02:00
|
|
|
}
|
|
|
|
|
|
|
|
int
|
|
|
|
main()
|
|
|
|
{
|
|
|
|
test_uniform_random();
|
|
|
|
test_bit_flip_set();
|
|
|
|
test_numeric_pattern_set();
|
|
|
|
test_bit_string_set();
|
|
|
|
test_document_words();
|
|
|
|
return 0;
|
|
|
|
}
|