AdGuardHome/internal/querylog/http.go

195 lines
5.0 KiB
Go
Raw Normal View History

package querylog
import (
"encoding/json"
"fmt"
"net/http"
"net/url"
"strconv"
"time"
"github.com/AdguardTeam/AdGuardHome/internal/util"
"github.com/AdguardTeam/golibs/jsonutil"
"github.com/AdguardTeam/golibs/log"
)
type qlogConfig struct {
Enabled bool `json:"enabled"`
Interval uint32 `json:"interval"`
AnonymizeClientIP bool `json:"anonymize_client_ip"`
}
// Register web handlers
func (l *queryLog) initWeb() {
l.conf.HTTPRegister(http.MethodGet, "/control/querylog", l.handleQueryLog)
l.conf.HTTPRegister(http.MethodGet, "/control/querylog_info", l.handleQueryLogInfo)
l.conf.HTTPRegister(http.MethodPost, "/control/querylog_clear", l.handleQueryLogClear)
l.conf.HTTPRegister(http.MethodPost, "/control/querylog_config", l.handleQueryLogConfig)
}
func httpError(r *http.Request, w http.ResponseWriter, code int, format string, args ...interface{}) {
text := fmt.Sprintf(format, args...)
log.Info("QueryLog: %s %s: %s", r.Method, r.URL, text)
http.Error(w, text, code)
}
func (l *queryLog) handleQueryLog(w http.ResponseWriter, r *http.Request) {
params, err := l.parseSearchParams(r)
if err != nil {
httpError(r, w, http.StatusBadRequest, "failed to parse params: %s", err)
return
}
// search for the log entries
entries, oldest := l.search(params)
// convert log entries to JSON
data := l.entriesToJSON(entries, oldest)
jsonVal, err := json.Marshal(data)
if err != nil {
httpError(r, w, http.StatusInternalServerError, "Couldn't marshal data into json: %s", err)
return
}
w.Header().Set("Content-Type", "application/json")
_, err = w.Write(jsonVal)
if err != nil {
httpError(r, w, http.StatusInternalServerError, "Unable to write response json: %s", err)
}
}
func (l *queryLog) handleQueryLogClear(_ http.ResponseWriter, _ *http.Request) {
l.clear()
}
// Get configuration
func (l *queryLog) handleQueryLogInfo(w http.ResponseWriter, r *http.Request) {
resp := qlogConfig{}
resp.Enabled = l.conf.Enabled
resp.Interval = l.conf.Interval
2020-03-03 17:21:53 +00:00
resp.AnonymizeClientIP = l.conf.AnonymizeClientIP
jsonVal, err := json.Marshal(resp)
if err != nil {
httpError(r, w, http.StatusInternalServerError, "json encode: %s", err)
return
}
w.Header().Set("Content-Type", "application/json")
_, err = w.Write(jsonVal)
if err != nil {
httpError(r, w, http.StatusInternalServerError, "http write: %s", err)
}
}
// Set configuration
func (l *queryLog) handleQueryLogConfig(w http.ResponseWriter, r *http.Request) {
d := qlogConfig{}
req, err := jsonutil.DecodeObject(&d, r.Body)
if err != nil {
httpError(r, w, http.StatusBadRequest, "%s", err)
return
}
if req.Exists("interval") && !checkInterval(d.Interval) {
httpError(r, w, http.StatusBadRequest, "Unsupported interval")
return
}
l.lock.Lock()
// copy data, modify it, then activate. Other threads (readers) don't need to use this lock.
conf := *l.conf
if req.Exists("enabled") {
conf.Enabled = d.Enabled
}
if req.Exists("interval") {
conf.Interval = d.Interval
}
2020-03-03 17:21:53 +00:00
if req.Exists("anonymize_client_ip") {
conf.AnonymizeClientIP = d.AnonymizeClientIP
}
l.conf = &conf
l.lock.Unlock()
l.conf.ConfigModified()
}
// "value" -> value, return TRUE
func getDoubleQuotesEnclosedValue(s *string) bool {
t := *s
if len(t) >= 2 && t[0] == '"' && t[len(t)-1] == '"' {
*s = t[1 : len(t)-1]
return true
}
return false
}
// parseSearchCriteria - parses "searchCriteria" from the specified query parameter
func (l *queryLog) parseSearchCriteria(q url.Values, name string, ct criteriaType) (bool, searchCriteria, error) {
val := q.Get(name)
if len(val) == 0 {
return false, searchCriteria{}, nil
}
c := searchCriteria{
criteriaType: ct,
value: val,
}
if getDoubleQuotesEnclosedValue(&c.value) {
c.strict = true
}
if ct == ctFilteringStatus && !util.ContainsString(filteringStatusValues, c.value) {
return false, c, fmt.Errorf("invalid value %s", c.value)
}
return true, c, nil
}
// parseSearchParams - parses "searchParams" from the HTTP request's query string
func (l *queryLog) parseSearchParams(r *http.Request) (*searchParams, error) {
p := newSearchParams()
var err error
q := r.URL.Query()
olderThan := q.Get("older_than")
if len(olderThan) != 0 {
p.olderThan, err = time.Parse(time.RFC3339Nano, olderThan)
if err != nil {
return nil, err
}
}
if limit, err := strconv.ParseInt(q.Get("limit"), 10, 64); err == nil {
p.limit = int(limit)
}
if offset, err := strconv.ParseInt(q.Get("offset"), 10, 64); err == nil {
p.offset = int(offset)
Pull request 743: + client: Query Logs Infinite Scroll Merge in DNS/adguard-home from feature/infinite_scroll_query_logs to master Squashed commit of the following: commit 4407ef2e7c055066257da791fbd65e6b0a495729 Merge: 40b74522 0a4781be Author: ArtemBaskal <a.baskal@adguard.com> Date: Tue Sep 1 16:20:23 2020 +0300 Merge branch 'master' into feature/infinite_scroll_query_logs commit 40b745225112cf8d664220ed8f484b0aa16e997c Author: ArtemBaskal <a.baskal@adguard.com> Date: Tue Sep 1 15:46:27 2020 +0300 Remove dynamic translation of toasts commit f08fa7b8c6a243f6b10e924aebccc183ce7814fd Author: ArtemBaskal <a.baskal@adguard.com> Date: Tue Sep 1 13:59:53 2020 +0300 Remove renderLimitIdx, update isEntireLog commit 0f1b02616faaa5759c0a3f6d8257117fa22094d9 Author: ArtemBaskal <a.baskal@adguard.com> Date: Tue Sep 1 11:11:14 2020 +0300 Rename variables commit 0928570c689c1fa704af775382620d68893e7c1c Author: ArtemBaskal <a.baskal@adguard.com> Date: Tue Sep 1 11:06:50 2020 +0300 Make query logs short polling function more expressive commit 9e773cbd6c287a1c799fa2680f3462508462ea7a Author: ArtemBaskal <a.baskal@adguard.com> Date: Tue Sep 1 11:06:19 2020 +0300 Fix Toast translation interface commit f9c57033e5adc5788954cf086b2f114dd8938bcb Author: ArtemBaskal <a.baskal@adguard.com> Date: Mon Aug 31 17:01:36 2020 +0300 Do not hide loader commit b86ba48613437f5559a748ad9aa4cf79d15db082 Author: ArtemBaskal <a.baskal@adguard.com> Date: Mon Aug 31 16:56:34 2020 +0300 Add dynamic translation for all toasts commit b9d1d9b447ca90a3c179e503fa5d4abd3516321e Author: ArtemBaskal <a.baskal@adguard.com> Date: Mon Aug 31 16:39:29 2020 +0300 Prevent getting query logs recursion if query is not changed commit e25189749f7912648cca4503cfa8d0ad898c4bb6 Author: ArtemBaskal <a.baskal@adguard.com> Date: Mon Aug 31 10:13:20 2020 +0300 Decrease page limit to 20 commit 8b248ac5276899de838abf2dc9a69e47599cfc12 Author: ArtemBaskal <a.baskal@adguard.com> Date: Fri Aug 28 18:47:12 2020 +0300 Return checkFilteredLogs commit bf2d65c4a3dca0da6b15f632ae11042b7c8e2045 Author: ArtemBaskal <a.baskal@adguard.com> Date: Fri Aug 28 18:33:51 2020 +0300 Review changes commit 01b5250f9d9136a1f334086d3e2f00d1a928b37b Author: ArtemBaskal <a.baskal@adguard.com> Date: Fri Aug 28 15:29:59 2020 +0300 Remove checkFilteredLogs commit 25b364c41e6a1489d930c8b3b39b1ab43723f29d Merge: 1dc66034 2c666cbd Author: Andrey Meshkov <am@adguard.com> Date: Fri Aug 28 14:28:47 2020 +0300 Merge branch 'feature/infinite_scroll_query_logs' of ssh://bit.adguard.com:7999/dns/adguard-home into feature/infinite_scroll_query_logs commit 1dc6603421cde9847e792bfe77ff6546e53fbc2a Author: Andrey Meshkov <am@adguard.com> Date: Fri Aug 28 14:28:01 2020 +0300 disregard maxFileScanEntries only if offset is set commit bad741ed7f1dccf6959d43d000b8c0150f526f9e Author: Andrey Meshkov <am@adguard.com> Date: Fri Aug 28 11:57:45 2020 +0300 Fix search behavior when limit is specified commit 2c666cbdde465cf17434126830dd99ceedfc4cbc Author: ArtemBaskal <a.baskal@adguard.com> Date: Thu Aug 27 18:50:28 2020 +0300 Hide table ref loader during data loading commit 8b4f7fe642ef9e87a979813dcdbd7817d64c27f9 Author: ArtemBaskal <a.baskal@adguard.com> Date: Thu Aug 27 18:43:24 2020 +0300 Repair search commit 26fae1ae01a789999b8a2181d60b35663a20460a Author: ArtemBaskal <a.baskal@adguard.com> Date: Thu Aug 27 17:59:27 2020 +0300 Resetting initial render index, change loader position on search commit e2c97ae1a288438267eef9aec71b979319674a71 Author: ArtemBaskal <a.baskal@adguard.com> Date: Thu Aug 27 16:02:03 2020 +0300 Change isScrolledIntoView ... and 32 more commits
2020-09-01 14:30:30 +01:00
// If we don't use "olderThan" and use offset/limit instead, we should change the default behavior
// and scan all log records until we found enough log entries
p.maxFileScanEntries = 0
}
paramNames := map[string]criteriaType{
"search": ctDomainOrClient,
"response_status": ctFilteringStatus,
}
for k, v := range paramNames {
ok, c, err := l.parseSearchCriteria(q, k, v)
if err != nil {
return nil, err
}
if ok {
p.searchCriteria = append(p.searchCriteria, c)
}
}
return p, nil
}