mirror of
https://github.com/AdguardTeam/AdGuardHome.git
synced 2024-11-25 22:45:46 +03:00
7c35d208b1
Updates #1273. Squashed commit of the following: commit 55b78153b1b775c855e759011141bbbe6d4b962c Author: Artem Baskal <a.baskal@adguard.com> Date: Fri Apr 2 16:55:39 2021 +0300 Update client_info in case of null commit 5c80c1438ed9d961af11617831b704d6ae15cc34 Author: Ainar Garipov <A.Garipov@AdGuard.COM> Date: Fri Apr 2 16:24:14 2021 +0300 querylog: always set client_info commit b48efd64d757cc0bcf5b34de22fdd0b0464d98a6 Merge: 4ed7eab523c9f528
Author: Ainar Garipov <A.Garipov@AdGuard.COM> Date: Fri Apr 2 16:22:08 2021 +0300 Merge branch 'master' into 1273-querylog-client-name commit 4ed7eab52b6b5b0c0ddb5aa5a3225a62d1f9265b Merge: dbf990eb70d4c70e
Author: Ainar Garipov <A.Garipov@AdGuard.COM> Date: Fri Apr 2 12:57:17 2021 +0300 Merge branch 'master' into 1273-querylog-client-name commit dbf990eb881116754554270e7b691b5db8e9ee34 Author: Ainar Garipov <A.Garipov@AdGuard.COM> Date: Fri Apr 2 12:56:13 2021 +0300 home: imp names commit c2cfdef494ca26fff62b9fa008f1b389d9d4d46b Author: Artem Baskal <a.baskal@adguard.com> Date: Thu Apr 1 19:26:04 2021 +0300 Rename to whois commit e3cc4a68ee576770b1922680155308e33bed31e8 Author: Ainar Garipov <A.Garipov@AdGuard.COM> Date: Thu Apr 1 19:03:42 2021 +0300 home: imp whois more commit 3b8ef8691c298aff35946b35923ef2e5b1f9bbbe Author: Ainar Garipov <A.Garipov@AdGuard.COM> Date: Thu Apr 1 18:51:14 2021 +0300 home: imp whois resp commit fb97e0d74976723a512d6ff4c69e830fe59c8df8 Author: Artem Baskal <a.baskal@adguard.com> Date: Thu Apr 1 18:00:03 2021 +0300 Fix client_info ids prop types commit 298005189e372651ceff453e88aca19ee925a138 Author: Artem Baskal <a.baskal@adguard.com> Date: Thu Apr 1 17:58:14 2021 +0300 Adapt changes on client commit aa1769f64197d865478a66271da483babfc5dfd0 Author: Ainar Garipov <A.Garipov@AdGuard.COM> Date: Thu Apr 1 17:18:36 2021 +0300 all: add more fields to querylog client commit 4b2a2dbd380ec410f3068d15ea16430912e03e33 Merge: cda92c3f2e4e2f62
Author: Ainar Garipov <A.Garipov@AdGuard.COM> Date: Thu Apr 1 16:57:26 2021 +0300 Merge branch 'master' into 1273-querylog-client-name commit cda92c3f0331cbac252f3163d31457f716bc7f2c Author: Ainar Garipov <A.Garipov@AdGuard.COM> Date: Mon Mar 29 18:03:51 2021 +0300 querylog: fix windows tests commit 5a56f0a32608869ed93a38f18f63ea3a20f7bde2 Merge: 627e4958e710ce11
Author: Ainar Garipov <A.Garipov@AdGuard.COM> Date: Mon Mar 29 17:45:53 2021 +0300 Merge branch 'master' into 1273-querylog-client-name commit 627e495828e82d44cc77aa393536479f23cc68b7 Author: Ainar Garipov <A.Garipov@AdGuard.COM> Date: Mon Mar 29 17:44:49 2021 +0300 querylog: add tests, imp code, docs commit 6dec468a2f0c29357875ff99458e0e8f8e580e6d Author: Ainar Garipov <A.Garipov@AdGuard.COM> Date: Fri Mar 26 16:10:47 2021 +0300 querylog: search clients by name, enrich http resp
207 lines
5.2 KiB
Go
207 lines
5.2 KiB
Go
package querylog
|
|
|
|
import (
|
|
"encoding/json"
|
|
"fmt"
|
|
"net/http"
|
|
"net/url"
|
|
"strconv"
|
|
"time"
|
|
|
|
"github.com/AdguardTeam/golibs/jsonutil"
|
|
"github.com/AdguardTeam/golibs/log"
|
|
)
|
|
|
|
type qlogConfig struct {
|
|
Enabled bool `json:"enabled"`
|
|
Interval uint32 `json:"interval"`
|
|
AnonymizeClientIP bool `json:"anonymize_client_ip"`
|
|
}
|
|
|
|
// Register web handlers
|
|
func (l *queryLog) initWeb() {
|
|
l.conf.HTTPRegister(http.MethodGet, "/control/querylog", l.handleQueryLog)
|
|
l.conf.HTTPRegister(http.MethodGet, "/control/querylog_info", l.handleQueryLogInfo)
|
|
l.conf.HTTPRegister(http.MethodPost, "/control/querylog_clear", l.handleQueryLogClear)
|
|
l.conf.HTTPRegister(http.MethodPost, "/control/querylog_config", l.handleQueryLogConfig)
|
|
}
|
|
|
|
func httpError(r *http.Request, w http.ResponseWriter, code int, format string, args ...interface{}) {
|
|
text := fmt.Sprintf(format, args...)
|
|
|
|
log.Info("QueryLog: %s %s: %s", r.Method, r.URL, text)
|
|
|
|
http.Error(w, text, code)
|
|
}
|
|
|
|
func (l *queryLog) handleQueryLog(w http.ResponseWriter, r *http.Request) {
|
|
params, err := l.parseSearchParams(r)
|
|
if err != nil {
|
|
httpError(r, w, http.StatusBadRequest, "failed to parse params: %s", err)
|
|
return
|
|
}
|
|
|
|
// search for the log entries
|
|
entries, oldest := l.search(params)
|
|
|
|
// convert log entries to JSON
|
|
data := l.entriesToJSON(entries, oldest)
|
|
|
|
jsonVal, err := json.Marshal(data)
|
|
if err != nil {
|
|
httpError(r, w, http.StatusInternalServerError, "Couldn't marshal data into json: %s", err)
|
|
return
|
|
}
|
|
|
|
w.Header().Set("Content-Type", "application/json")
|
|
_, err = w.Write(jsonVal)
|
|
if err != nil {
|
|
httpError(r, w, http.StatusInternalServerError, "Unable to write response json: %s", err)
|
|
}
|
|
}
|
|
|
|
func (l *queryLog) handleQueryLogClear(_ http.ResponseWriter, _ *http.Request) {
|
|
l.clear()
|
|
}
|
|
|
|
// Get configuration
|
|
func (l *queryLog) handleQueryLogInfo(w http.ResponseWriter, r *http.Request) {
|
|
resp := qlogConfig{}
|
|
resp.Enabled = l.conf.Enabled
|
|
resp.Interval = l.conf.RotationIvl
|
|
resp.AnonymizeClientIP = l.conf.AnonymizeClientIP
|
|
|
|
jsonVal, err := json.Marshal(resp)
|
|
if err != nil {
|
|
httpError(r, w, http.StatusInternalServerError, "json encode: %s", err)
|
|
return
|
|
}
|
|
w.Header().Set("Content-Type", "application/json")
|
|
_, err = w.Write(jsonVal)
|
|
if err != nil {
|
|
httpError(r, w, http.StatusInternalServerError, "http write: %s", err)
|
|
}
|
|
}
|
|
|
|
// Set configuration
|
|
func (l *queryLog) handleQueryLogConfig(w http.ResponseWriter, r *http.Request) {
|
|
d := qlogConfig{}
|
|
req, err := jsonutil.DecodeObject(&d, r.Body)
|
|
if err != nil {
|
|
httpError(r, w, http.StatusBadRequest, "%s", err)
|
|
return
|
|
}
|
|
|
|
if req.Exists("interval") && !checkInterval(d.Interval) {
|
|
httpError(r, w, http.StatusBadRequest, "Unsupported interval")
|
|
return
|
|
}
|
|
|
|
l.lock.Lock()
|
|
// copy data, modify it, then activate. Other threads (readers) don't need to use this lock.
|
|
conf := *l.conf
|
|
if req.Exists("enabled") {
|
|
conf.Enabled = d.Enabled
|
|
}
|
|
if req.Exists("interval") {
|
|
conf.RotationIvl = d.Interval
|
|
}
|
|
if req.Exists("anonymize_client_ip") {
|
|
conf.AnonymizeClientIP = d.AnonymizeClientIP
|
|
}
|
|
l.conf = &conf
|
|
l.lock.Unlock()
|
|
|
|
l.conf.ConfigModified()
|
|
}
|
|
|
|
// "value" -> value, return TRUE
|
|
func getDoubleQuotesEnclosedValue(s *string) bool {
|
|
t := *s
|
|
if len(t) >= 2 && t[0] == '"' && t[len(t)-1] == '"' {
|
|
*s = t[1 : len(t)-1]
|
|
return true
|
|
}
|
|
return false
|
|
}
|
|
|
|
// inStr checks if string is in the slice of strings.
|
|
func inStr(strs []string, str string) (ok bool) {
|
|
for _, s := range strs {
|
|
if s == str {
|
|
return true
|
|
}
|
|
}
|
|
|
|
return false
|
|
}
|
|
|
|
// parseSearchCriteria - parses "searchCriteria" from the specified query parameter
|
|
func (l *queryLog) parseSearchCriteria(q url.Values, name string, ct criteriaType) (bool, searchCriteria, error) {
|
|
val := q.Get(name)
|
|
if len(val) == 0 {
|
|
return false, searchCriteria{}, nil
|
|
}
|
|
|
|
c := searchCriteria{
|
|
criteriaType: ct,
|
|
value: val,
|
|
}
|
|
if getDoubleQuotesEnclosedValue(&c.value) {
|
|
c.strict = true
|
|
}
|
|
|
|
if ct == ctFilteringStatus && !inStr(filteringStatusValues, c.value) {
|
|
return false, c, fmt.Errorf("invalid value %s", c.value)
|
|
}
|
|
|
|
return true, c, nil
|
|
}
|
|
|
|
// parseSearchParams - parses "searchParams" from the HTTP request's query string
|
|
func (l *queryLog) parseSearchParams(r *http.Request) (p *searchParams, err error) {
|
|
p = newSearchParams()
|
|
|
|
q := r.URL.Query()
|
|
olderThan := q.Get("older_than")
|
|
if len(olderThan) != 0 {
|
|
p.olderThan, err = time.Parse(time.RFC3339Nano, olderThan)
|
|
if err != nil {
|
|
return nil, err
|
|
}
|
|
}
|
|
|
|
var limit64 int64
|
|
if limit64, err = strconv.ParseInt(q.Get("limit"), 10, 64); err == nil {
|
|
p.limit = int(limit64)
|
|
}
|
|
|
|
var offset64 int64
|
|
if offset64, err = strconv.ParseInt(q.Get("offset"), 10, 64); err == nil {
|
|
p.offset = int(offset64)
|
|
|
|
// If we don't use "olderThan" and use offset/limit instead, we should change the default behavior
|
|
// and scan all log records until we found enough log entries
|
|
p.maxFileScanEntries = 0
|
|
}
|
|
|
|
paramNames := map[string]criteriaType{
|
|
"search": ctDomainOrClient,
|
|
"response_status": ctFilteringStatus,
|
|
}
|
|
|
|
for k, v := range paramNames {
|
|
var ok bool
|
|
var c searchCriteria
|
|
ok, c, err = l.parseSearchCriteria(q, k, v)
|
|
if err != nil {
|
|
return nil, err
|
|
}
|
|
|
|
if ok {
|
|
p.searchCriteria = append(p.searchCriteria, c)
|
|
}
|
|
}
|
|
|
|
return p, nil
|
|
}
|