2019-08-22 14:34:58 +01:00
|
|
|
package stats
|
|
|
|
|
|
|
|
import (
|
|
|
|
"fmt"
|
|
|
|
"net"
|
|
|
|
"os"
|
|
|
|
"sync/atomic"
|
|
|
|
"testing"
|
|
|
|
|
|
|
|
"github.com/stretchr/testify/assert"
|
|
|
|
)
|
|
|
|
|
2019-09-10 15:59:10 +01:00
|
|
|
func UIntArrayEquals(a []uint64, b []uint64) bool {
|
2019-08-22 14:34:58 +01:00
|
|
|
if len(a) != len(b) {
|
|
|
|
return false
|
|
|
|
}
|
|
|
|
|
|
|
|
for i := range a {
|
|
|
|
if a[i] != b[i] {
|
|
|
|
return false
|
|
|
|
}
|
|
|
|
}
|
|
|
|
|
|
|
|
return true
|
|
|
|
}
|
|
|
|
|
|
|
|
func TestStats(t *testing.T) {
|
2019-09-16 14:14:52 +01:00
|
|
|
conf := Config{
|
|
|
|
Filename: "./stats.db",
|
|
|
|
LimitDays: 1,
|
|
|
|
}
|
2019-09-25 13:36:09 +01:00
|
|
|
s, _ := createObject(conf)
|
2019-08-22 14:34:58 +01:00
|
|
|
|
|
|
|
e := Entry{}
|
|
|
|
|
|
|
|
e.Domain = "domain"
|
|
|
|
e.Client = net.ParseIP("127.0.0.1")
|
|
|
|
e.Result = RFiltered
|
|
|
|
e.Time = 123456
|
|
|
|
s.Update(e)
|
|
|
|
|
|
|
|
e.Domain = "domain"
|
|
|
|
e.Client = net.ParseIP("127.0.0.1")
|
|
|
|
e.Result = RNotFiltered
|
|
|
|
e.Time = 123456
|
|
|
|
s.Update(e)
|
|
|
|
|
2019-09-25 13:36:09 +01:00
|
|
|
d := s.getData(Hours)
|
2019-09-10 15:59:10 +01:00
|
|
|
a := []uint64{0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 2}
|
|
|
|
assert.True(t, UIntArrayEquals(d["dns_queries"].([]uint64), a))
|
2019-08-22 14:34:58 +01:00
|
|
|
|
2019-09-10 15:59:10 +01:00
|
|
|
a = []uint64{0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 1}
|
|
|
|
assert.True(t, UIntArrayEquals(d["blocked_filtering"].([]uint64), a))
|
2019-08-22 14:34:58 +01:00
|
|
|
|
2019-09-10 15:59:10 +01:00
|
|
|
a = []uint64{0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0}
|
|
|
|
assert.True(t, UIntArrayEquals(d["replaced_safebrowsing"].([]uint64), a))
|
2019-08-22 14:34:58 +01:00
|
|
|
|
2019-09-10 15:59:10 +01:00
|
|
|
a = []uint64{0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0}
|
|
|
|
assert.True(t, UIntArrayEquals(d["replaced_parental"].([]uint64), a))
|
2019-08-22 14:34:58 +01:00
|
|
|
|
2019-09-10 15:59:10 +01:00
|
|
|
m := d["top_queried_domains"].([]map[string]uint64)
|
2019-08-22 14:34:58 +01:00
|
|
|
assert.True(t, m[0]["domain"] == 1)
|
|
|
|
|
2019-09-10 15:59:10 +01:00
|
|
|
m = d["top_blocked_domains"].([]map[string]uint64)
|
2019-08-22 14:34:58 +01:00
|
|
|
assert.True(t, m[0]["domain"] == 1)
|
|
|
|
|
2019-09-10 15:59:10 +01:00
|
|
|
m = d["top_clients"].([]map[string]uint64)
|
2019-08-22 14:34:58 +01:00
|
|
|
assert.True(t, m[0]["127.0.0.1"] == 2)
|
|
|
|
|
2019-09-10 15:59:10 +01:00
|
|
|
assert.True(t, d["num_dns_queries"].(uint64) == 2)
|
|
|
|
assert.True(t, d["num_blocked_filtering"].(uint64) == 1)
|
|
|
|
assert.True(t, d["num_replaced_safebrowsing"].(uint64) == 0)
|
|
|
|
assert.True(t, d["num_replaced_safesearch"].(uint64) == 0)
|
|
|
|
assert.True(t, d["num_replaced_parental"].(uint64) == 0)
|
2019-08-22 14:34:58 +01:00
|
|
|
assert.True(t, d["avg_processing_time"].(float64) == 0.123456)
|
|
|
|
|
2019-10-07 13:56:33 +01:00
|
|
|
topClients := s.GetTopClientsIP(2)
|
|
|
|
assert.True(t, topClients[0] == "127.0.0.1")
|
|
|
|
|
2019-09-25 13:36:09 +01:00
|
|
|
s.clear()
|
2019-08-22 14:34:58 +01:00
|
|
|
s.Close()
|
2019-09-16 14:14:52 +01:00
|
|
|
os.Remove(conf.Filename)
|
2019-08-22 14:34:58 +01:00
|
|
|
}
|
|
|
|
|
|
|
|
func TestLargeNumbers(t *testing.T) {
|
|
|
|
var hour int32
|
|
|
|
hour = 1
|
2019-09-10 15:59:10 +01:00
|
|
|
newID := func() uint32 {
|
2019-08-22 14:34:58 +01:00
|
|
|
// use "atomic" to make Go race detector happy
|
2019-09-10 15:59:10 +01:00
|
|
|
return uint32(atomic.LoadInt32(&hour))
|
2019-08-22 14:34:58 +01:00
|
|
|
}
|
|
|
|
|
|
|
|
// log.SetLevel(log.DEBUG)
|
2019-09-16 14:14:52 +01:00
|
|
|
conf := Config{
|
|
|
|
Filename: "./stats.db",
|
|
|
|
LimitDays: 1,
|
|
|
|
UnitID: newID,
|
|
|
|
}
|
|
|
|
os.Remove(conf.Filename)
|
2019-09-25 13:36:09 +01:00
|
|
|
s, _ := createObject(conf)
|
2019-08-22 14:34:58 +01:00
|
|
|
e := Entry{}
|
|
|
|
|
|
|
|
n := 1000 // number of distinct clients and domains every hour
|
|
|
|
for h := 0; h != 12; h++ {
|
|
|
|
if h != 0 {
|
|
|
|
atomic.AddInt32(&hour, 1)
|
|
|
|
}
|
|
|
|
for i := 0; i != n; i++ {
|
|
|
|
e.Domain = fmt.Sprintf("domain%d", i)
|
|
|
|
e.Client = net.ParseIP("127.0.0.1")
|
|
|
|
e.Client[2] = byte((i & 0xff00) >> 8)
|
|
|
|
e.Client[3] = byte(i & 0xff)
|
|
|
|
e.Result = RNotFiltered
|
|
|
|
e.Time = 123456
|
|
|
|
s.Update(e)
|
|
|
|
}
|
|
|
|
}
|
|
|
|
|
2019-09-25 13:36:09 +01:00
|
|
|
d := s.getData(Hours)
|
2019-09-10 15:59:10 +01:00
|
|
|
assert.True(t, d["num_dns_queries"].(uint64) == uint64(int(hour)*n))
|
2019-08-22 14:34:58 +01:00
|
|
|
|
|
|
|
s.Close()
|
2019-09-16 14:14:52 +01:00
|
|
|
os.Remove(conf.Filename)
|
2019-08-22 14:34:58 +01:00
|
|
|
}
|
2019-10-09 15:17:57 +01:00
|
|
|
|
|
|
|
// this code is a chunk copied from getData() that generates aggregate data per day
|
|
|
|
func aggregateDataPerDay(firstID uint32) int {
|
|
|
|
firstDayID := (firstID + 24 - 1) / 24 * 24 // align_ceil(24)
|
|
|
|
a := []uint64{}
|
|
|
|
var sum uint64
|
|
|
|
id := firstDayID
|
|
|
|
nextDayID := firstDayID + 24
|
|
|
|
for i := firstDayID - firstID; int(i) != 720; i++ {
|
|
|
|
sum++
|
|
|
|
if id == nextDayID {
|
|
|
|
a = append(a, sum)
|
|
|
|
sum = 0
|
|
|
|
nextDayID += 24
|
|
|
|
}
|
|
|
|
id++
|
|
|
|
}
|
|
|
|
if id <= nextDayID {
|
|
|
|
a = append(a, sum)
|
|
|
|
}
|
|
|
|
return len(a)
|
|
|
|
}
|
|
|
|
func TestAggregateDataPerTimeUnit(t *testing.T) {
|
|
|
|
for i := 0; i != 25; i++ {
|
|
|
|
alen := aggregateDataPerDay(uint32(i))
|
|
|
|
assert.True(t, alen == 30, "i=%d", i)
|
|
|
|
}
|
|
|
|
}
|