mirror of
https://github.com/safing/portmaster
synced 2025-04-10 22:19:11 +00:00
* Move portbase into monorepo * Add new simple module mgr * [WIP] Switch to new simple module mgr * Add StateMgr and more worker variants * [WIP] Switch more modules * [WIP] Switch more modules * [WIP] swtich more modules * [WIP] switch all SPN modules * [WIP] switch all service modules * [WIP] Convert all workers to the new module system * [WIP] add new task system to module manager * [WIP] Add second take for scheduling workers * [WIP] Add FIXME for bugs in new scheduler * [WIP] Add minor improvements to scheduler * [WIP] Add new worker scheduler * [WIP] Fix more bug related to new module system * [WIP] Fix start handing of the new module system * [WIP] Improve startup process * [WIP] Fix minor issues * [WIP] Fix missing subsystem in settings * [WIP] Initialize managers in constructor * [WIP] Move module event initialization to constrictors * [WIP] Fix setting for enabling and disabling the SPN module * [WIP] Move API registeration into module construction * [WIP] Update states mgr for all modules * [WIP] Add CmdLine operation support * Add state helper methods to module group and instance * Add notification and module status handling to status package * Fix starting issues * Remove pilot widget and update security lock to new status data * Remove debug logs * Improve http server shutdown * Add workaround for cleanly shutting down firewall+netquery * Improve logging * Add syncing states with notifications for new module system * Improve starting, stopping, shutdown; resolve FIXMEs/TODOs * [WIP] Fix most unit tests * Review new module system and fix minor issues * Push shutdown and restart events again via API * Set sleep mode via interface * Update example/template module * [WIP] Fix spn/cabin unit test * Remove deprecated UI elements * Make log output more similar for the logging transition phase * Switch spn hub and observer cmds to new module system * Fix log sources * Make worker mgr less error prone * Fix tests and minor issues * Fix observation hub * Improve shutdown and restart handling * Split up big connection.go source file * Move varint and dsd packages to structures repo * Improve expansion test * Fix linter warnings * Fix interception module on windows * Fix linter errors --------- Co-authored-by: Vladimir Stoilov <vladimir@safing.io>
456 lines
12 KiB
Go
456 lines
12 KiB
Go
package resolver
|
|
|
|
import (
|
|
"context"
|
|
"errors"
|
|
"fmt"
|
|
"net"
|
|
"strings"
|
|
"sync"
|
|
"time"
|
|
|
|
"github.com/miekg/dns"
|
|
|
|
"github.com/safing/portmaster/base/log"
|
|
"github.com/safing/portmaster/service/mgr"
|
|
"github.com/safing/portmaster/service/netenv"
|
|
"github.com/safing/portmaster/service/network/netutils"
|
|
)
|
|
|
|
// DNS Classes.
|
|
const (
|
|
DNSClassMulticast = dns.ClassINET | 1<<15
|
|
)
|
|
|
|
var (
|
|
multicast4Conn *net.UDPConn
|
|
multicast6Conn *net.UDPConn
|
|
unicast4Conn *net.UDPConn
|
|
unicast6Conn *net.UDPConn
|
|
|
|
questions = make(map[uint16]*savedQuestion)
|
|
questionsLock sync.Mutex
|
|
|
|
mDNSResolver = &Resolver{
|
|
ConfigURL: ServerSourceMDNS,
|
|
Info: &ResolverInfo{
|
|
Type: ServerTypeMDNS,
|
|
Source: ServerSourceMDNS,
|
|
IPScope: netutils.SiteLocal,
|
|
},
|
|
Conn: &mDNSResolverConn{},
|
|
}
|
|
mDNSResolvers = []*Resolver{mDNSResolver}
|
|
)
|
|
|
|
type mDNSResolverConn struct{}
|
|
|
|
func (mrc *mDNSResolverConn) Query(ctx context.Context, q *Query) (*RRCache, error) {
|
|
return queryMulticastDNS(ctx, q)
|
|
}
|
|
|
|
func (mrc *mDNSResolverConn) ReportFailure() {}
|
|
|
|
func (mrc *mDNSResolverConn) IsFailing() bool {
|
|
return false
|
|
}
|
|
|
|
func (mrc *mDNSResolverConn) ResetFailure() {}
|
|
|
|
func (mrc *mDNSResolverConn) ForceReconnect(_ context.Context) {}
|
|
|
|
type savedQuestion struct {
|
|
question dns.Question
|
|
expires time.Time
|
|
response chan *RRCache
|
|
}
|
|
|
|
func indexOfRR(entry *dns.RR_Header, list *[]dns.RR) int {
|
|
for k, v := range *list {
|
|
if entry.Name == v.Header().Name && entry.Rrtype == v.Header().Rrtype {
|
|
return k
|
|
}
|
|
}
|
|
return -1
|
|
}
|
|
|
|
//nolint:gocyclo,gocognit // TODO: make simpler
|
|
func listenToMDNS(wc *mgr.WorkerCtx) error {
|
|
var err error
|
|
messages := make(chan *dns.Msg, 32)
|
|
|
|
// TODO: init and start every listener in its own service worker
|
|
// this will make the more resilient and actually able to restart
|
|
|
|
multicast4Conn, err = net.ListenMulticastUDP("udp4", nil, &net.UDPAddr{IP: net.IPv4(224, 0, 0, 251), Port: 5353})
|
|
if err != nil {
|
|
// TODO: retry after some time
|
|
log.Warningf("intel(mdns): failed to create udp4 listen multicast socket: %s", err)
|
|
} else {
|
|
module.mgr.Go("mdns udp4 multicast listener", func(wc *mgr.WorkerCtx) error {
|
|
return listenForDNSPackets(wc.Ctx(), multicast4Conn, messages)
|
|
})
|
|
defer func() {
|
|
_ = multicast4Conn.Close()
|
|
}()
|
|
}
|
|
|
|
unicast4Conn, err = net.ListenUDP("udp4", &net.UDPAddr{IP: net.IPv4zero, Port: 0})
|
|
if err != nil {
|
|
// TODO: retry after some time
|
|
log.Warningf("intel(mdns): failed to create udp4 listen socket: %s", err)
|
|
} else {
|
|
module.mgr.Go("mdns udp4 unicast listener", func(wc *mgr.WorkerCtx) error {
|
|
return listenForDNSPackets(wc.Ctx(), unicast4Conn, messages)
|
|
})
|
|
defer func() {
|
|
_ = unicast4Conn.Close()
|
|
}()
|
|
}
|
|
|
|
if netenv.IPv6Enabled() {
|
|
multicast6Conn, err = net.ListenMulticastUDP("udp6", nil, &net.UDPAddr{IP: net.IP([]byte{0xff, 0x02, 0x00, 0x00, 0x00, 0x00, 0x00, 0x00, 0x00, 0x00, 0x00, 0x00, 0x00, 0x00, 0x00, 0xfb}), Port: 5353})
|
|
if err != nil {
|
|
// TODO: retry after some time
|
|
log.Warningf("intel(mdns): failed to create udp6 listen multicast socket: %s", err)
|
|
} else {
|
|
module.mgr.Go("mdns udp6 multicast listener", func(wc *mgr.WorkerCtx) error {
|
|
return listenForDNSPackets(wc.Ctx(), multicast6Conn, messages)
|
|
})
|
|
defer func() {
|
|
_ = multicast6Conn.Close()
|
|
}()
|
|
}
|
|
|
|
unicast6Conn, err = net.ListenUDP("udp6", &net.UDPAddr{IP: net.IPv6zero, Port: 0})
|
|
if err != nil {
|
|
// TODO: retry after some time
|
|
log.Warningf("intel(mdns): failed to create udp6 listen socket: %s", err)
|
|
} else {
|
|
module.mgr.Go("mdns udp6 unicast listener", func(wc *mgr.WorkerCtx) error {
|
|
return listenForDNSPackets(wc.Ctx(), unicast6Conn, messages)
|
|
})
|
|
defer func() {
|
|
_ = unicast6Conn.Close()
|
|
}()
|
|
}
|
|
} else {
|
|
log.Warningf("resolver: no IPv6 stack detected, disabling IPv6 mDNS resolver")
|
|
}
|
|
|
|
// start message handler
|
|
module.mgr.Go("mdns message handler", func(wc *mgr.WorkerCtx) error {
|
|
return handleMDNSMessages(wc.Ctx(), messages)
|
|
})
|
|
|
|
// wait for shutdown
|
|
<-wc.Done()
|
|
return nil
|
|
}
|
|
|
|
func handleMDNSMessages(ctx context.Context, messages chan *dns.Msg) error { //nolint:maintidx // TODO: Improve.
|
|
for {
|
|
select {
|
|
case <-ctx.Done():
|
|
return nil
|
|
case message := <-messages:
|
|
// log.Tracef("resolver: got net mdns message: %s", message)
|
|
|
|
var err error
|
|
var question *dns.Question
|
|
var saveFullRequest bool
|
|
scavengedRecords := make(map[string]dns.RR)
|
|
var rrCache *RRCache
|
|
|
|
// save every received response
|
|
// if previous save was less than 2 seconds ago, add to response, else replace
|
|
// pick out A and AAAA records and save separately
|
|
|
|
// continue if not response
|
|
if !message.Response {
|
|
// log.Tracef("resolver: mdns message has no response, ignoring")
|
|
continue
|
|
}
|
|
|
|
// continue if rcode is not success
|
|
if message.Rcode != dns.RcodeSuccess {
|
|
// log.Tracef("resolver: mdns message has error, ignoring")
|
|
continue
|
|
}
|
|
|
|
// continue if answer section is empty
|
|
if len(message.Answer) == 0 {
|
|
// log.Tracef("resolver: mdns message has no answers, ignoring")
|
|
continue
|
|
}
|
|
|
|
// return saved question
|
|
questionsLock.Lock()
|
|
savedQ := questions[message.MsgHdr.Id]
|
|
questionsLock.Unlock()
|
|
|
|
// get question, some servers do not reply with question
|
|
if len(message.Question) > 0 {
|
|
question = &message.Question[0]
|
|
// if questions do not match, disregard saved question
|
|
if savedQ != nil && message.Question[0].String() != savedQ.question.String() {
|
|
savedQ = nil
|
|
}
|
|
} else if savedQ != nil {
|
|
question = &savedQ.question
|
|
}
|
|
|
|
if question != nil {
|
|
// continue if class is not INTERNET
|
|
if question.Qclass != dns.ClassINET && question.Qclass != DNSClassMulticast {
|
|
continue
|
|
}
|
|
// mark request to be saved
|
|
saveFullRequest = true
|
|
}
|
|
|
|
// get entry from database
|
|
if saveFullRequest {
|
|
// get from database
|
|
rrCache, err = GetRRCache(question.Name, dns.Type(question.Qtype))
|
|
// if we have no cached entry, or it has been updated more than two seconds ago, or if it expired:
|
|
// create new and do not append
|
|
if err != nil || rrCache.Modified < time.Now().Add(-2*time.Second).Unix() || rrCache.Expired() {
|
|
rrCache = &RRCache{
|
|
Domain: question.Name,
|
|
Question: dns.Type(question.Qtype),
|
|
RCode: dns.RcodeSuccess,
|
|
Resolver: mDNSResolver.Info.Copy(),
|
|
}
|
|
}
|
|
}
|
|
|
|
// add all entries to RRCache
|
|
for _, entry := range message.Answer {
|
|
if domainInScope(entry.Header().Name, multicastDomains) {
|
|
if saveFullRequest {
|
|
k := indexOfRR(entry.Header(), &rrCache.Answer)
|
|
if k == -1 {
|
|
rrCache.Answer = append(rrCache.Answer, entry)
|
|
} else {
|
|
rrCache.Answer[k] = entry
|
|
}
|
|
}
|
|
switch entry.(type) {
|
|
case *dns.A:
|
|
scavengedRecords[fmt.Sprintf("%sA", entry.Header().Name)] = entry
|
|
case *dns.AAAA:
|
|
scavengedRecords[fmt.Sprintf("%sAAAA", entry.Header().Name)] = entry
|
|
case *dns.PTR:
|
|
if !strings.HasPrefix(entry.Header().Name, "_") {
|
|
scavengedRecords[fmt.Sprintf("%sPTR", entry.Header().Name)] = entry
|
|
}
|
|
}
|
|
}
|
|
}
|
|
for _, entry := range message.Ns {
|
|
if domainInScope(entry.Header().Name, multicastDomains) {
|
|
if saveFullRequest {
|
|
k := indexOfRR(entry.Header(), &rrCache.Ns)
|
|
if k == -1 {
|
|
rrCache.Ns = append(rrCache.Ns, entry)
|
|
} else {
|
|
rrCache.Ns[k] = entry
|
|
}
|
|
}
|
|
switch entry.(type) {
|
|
case *dns.A:
|
|
scavengedRecords[fmt.Sprintf("%sA", entry.Header().Name)] = entry
|
|
case *dns.AAAA:
|
|
scavengedRecords[fmt.Sprintf("%sAAAA", entry.Header().Name)] = entry
|
|
case *dns.PTR:
|
|
if !strings.HasPrefix(entry.Header().Name, "_") {
|
|
scavengedRecords[fmt.Sprintf("%sPTR", entry.Header().Name)] = entry
|
|
}
|
|
}
|
|
}
|
|
}
|
|
for _, entry := range message.Extra {
|
|
if domainInScope(entry.Header().Name, multicastDomains) {
|
|
if saveFullRequest {
|
|
k := indexOfRR(entry.Header(), &rrCache.Extra)
|
|
if k == -1 {
|
|
rrCache.Extra = append(rrCache.Extra, entry)
|
|
} else {
|
|
rrCache.Extra[k] = entry
|
|
}
|
|
}
|
|
switch entry.(type) {
|
|
case *dns.A:
|
|
scavengedRecords[fmt.Sprintf("%sA", entry.Header().Name)] = entry
|
|
case *dns.AAAA:
|
|
scavengedRecords[fmt.Sprintf("%sAAAA", entry.Header().Name)] = entry
|
|
case *dns.PTR:
|
|
if !strings.HasPrefix(entry.Header().Name, "_") {
|
|
scavengedRecords[fmt.Sprintf("%sPTR", entry.Header().Name)] = entry
|
|
}
|
|
}
|
|
}
|
|
}
|
|
|
|
var questionID string
|
|
if saveFullRequest {
|
|
rrCache.Clean(minMDnsTTL)
|
|
err := rrCache.Save()
|
|
if err != nil {
|
|
log.Warningf("resolver: failed to cache RR %s: %s", rrCache.Domain, err)
|
|
}
|
|
|
|
// return finished response
|
|
if savedQ != nil {
|
|
select {
|
|
case savedQ.response <- rrCache:
|
|
default:
|
|
}
|
|
}
|
|
|
|
questionID = fmt.Sprintf("%s%s", question.Name, dns.Type(question.Qtype).String())
|
|
}
|
|
|
|
for k, v := range scavengedRecords {
|
|
if saveFullRequest && k == questionID {
|
|
continue
|
|
}
|
|
rrCache = &RRCache{
|
|
Domain: v.Header().Name,
|
|
Question: dns.Type(v.Header().Class),
|
|
RCode: dns.RcodeSuccess,
|
|
Answer: []dns.RR{v},
|
|
Resolver: mDNSResolver.Info.Copy(),
|
|
}
|
|
rrCache.Clean(minMDnsTTL)
|
|
err := rrCache.Save()
|
|
if err != nil {
|
|
log.Warningf("resolver: failed to cache RR %s: %s", rrCache.Domain, err)
|
|
}
|
|
// log.Tracef("resolver: mdns scavenged %s", k)
|
|
}
|
|
|
|
}
|
|
|
|
cleanSavedQuestions()
|
|
}
|
|
}
|
|
|
|
func listenForDNSPackets(ctx context.Context, conn *net.UDPConn, messages chan *dns.Msg) error {
|
|
buf := make([]byte, 65536)
|
|
for {
|
|
n, err := conn.Read(buf)
|
|
if err != nil {
|
|
if module.mgr.IsDone() {
|
|
return nil
|
|
}
|
|
log.Debugf("resolver: failed to read packet: %s", err)
|
|
return err
|
|
}
|
|
message := new(dns.Msg)
|
|
if err = message.Unpack(buf[:n]); err != nil {
|
|
log.Debugf("resolver: failed to unpack message: %s", err)
|
|
continue
|
|
}
|
|
select {
|
|
case messages <- message:
|
|
case <-ctx.Done():
|
|
return nil
|
|
}
|
|
}
|
|
}
|
|
|
|
func queryMulticastDNS(ctx context.Context, q *Query) (*RRCache, error) {
|
|
// check for active connections
|
|
if unicast4Conn == nil && unicast6Conn == nil {
|
|
return nil, errors.New("unicast mdns connections not initialized")
|
|
}
|
|
|
|
// trace log
|
|
log.Tracer(ctx).Trace("resolver: resolving with mDNS")
|
|
|
|
// create query
|
|
dnsQuery := new(dns.Msg)
|
|
dnsQuery.SetQuestion(q.FQDN, uint16(q.QType))
|
|
// request unicast response
|
|
// q.Question[0].Qclass |= 1 << 15
|
|
dnsQuery.RecursionDesired = false
|
|
|
|
// create response channel
|
|
response := make(chan *RRCache)
|
|
|
|
// save question
|
|
questionsLock.Lock()
|
|
defer questionsLock.Unlock()
|
|
questions[dnsQuery.MsgHdr.Id] = &savedQuestion{
|
|
question: dnsQuery.Question[0],
|
|
expires: time.Now().Add(10 * time.Second),
|
|
response: response,
|
|
}
|
|
|
|
// pack qeury
|
|
buf, err := dnsQuery.Pack()
|
|
if err != nil {
|
|
return nil, fmt.Errorf("failed to pack query: %w", err)
|
|
}
|
|
|
|
// send queries
|
|
if unicast4Conn != nil && uint16(q.QType) != dns.TypeAAAA {
|
|
err = unicast4Conn.SetWriteDeadline(time.Now().Add(1 * time.Second))
|
|
if err != nil {
|
|
return nil, fmt.Errorf("failed to configure query (set timout): %w", err)
|
|
}
|
|
|
|
_, err = unicast4Conn.WriteToUDP(buf, &net.UDPAddr{IP: net.IPv4(224, 0, 0, 251), Port: 5353})
|
|
if err != nil {
|
|
return nil, fmt.Errorf("failed to send query: %w", err)
|
|
}
|
|
}
|
|
if unicast6Conn != nil && uint16(q.QType) != dns.TypeA {
|
|
err = unicast6Conn.SetWriteDeadline(time.Now().Add(1 * time.Second))
|
|
if err != nil {
|
|
return nil, fmt.Errorf("failed to configure query (set timout): %w", err)
|
|
}
|
|
|
|
_, err = unicast6Conn.WriteToUDP(buf, &net.UDPAddr{IP: net.IP([]byte{0xff, 0x02, 0x00, 0x00, 0x00, 0x00, 0x00, 0x00, 0x00, 0x00, 0x00, 0x00, 0x00, 0x00, 0x00, 0xfb}), Port: 5353})
|
|
if err != nil {
|
|
return nil, fmt.Errorf("failed to send query: %w", err)
|
|
}
|
|
}
|
|
|
|
// wait for response or timeout
|
|
select {
|
|
case rrCache := <-response:
|
|
if rrCache != nil {
|
|
return rrCache, nil
|
|
}
|
|
case <-time.After(1 * time.Second):
|
|
// check cache again
|
|
rrCache, err := GetRRCache(q.FQDN, q.QType)
|
|
if err == nil {
|
|
return rrCache, nil
|
|
}
|
|
case <-ctx.Done():
|
|
return nil, ctx.Err()
|
|
}
|
|
|
|
// Respond with NXDomain.
|
|
return &RRCache{
|
|
Domain: q.FQDN,
|
|
Question: q.QType,
|
|
RCode: dns.RcodeNameError,
|
|
Resolver: mDNSResolver.Info.Copy(),
|
|
}, nil
|
|
}
|
|
|
|
func cleanSavedQuestions() {
|
|
questionsLock.Lock()
|
|
defer questionsLock.Unlock()
|
|
now := time.Now()
|
|
for msgID, savedQuestion := range questions {
|
|
if now.After(savedQuestion.expires) {
|
|
delete(questions, msgID)
|
|
}
|
|
}
|
|
}
|