mirror of
https://github.com/safing/portmaster
synced 2025-04-10 14:09:11 +00:00
* Move portbase into monorepo * Add new simple module mgr * [WIP] Switch to new simple module mgr * Add StateMgr and more worker variants * [WIP] Switch more modules * [WIP] Switch more modules * [WIP] swtich more modules * [WIP] switch all SPN modules * [WIP] switch all service modules * [WIP] Convert all workers to the new module system * [WIP] add new task system to module manager * [WIP] Add second take for scheduling workers * [WIP] Add FIXME for bugs in new scheduler * [WIP] Add minor improvements to scheduler * [WIP] Add new worker scheduler * [WIP] Fix more bug related to new module system * [WIP] Fix start handing of the new module system * [WIP] Improve startup process * [WIP] Fix minor issues * [WIP] Fix missing subsystem in settings * [WIP] Initialize managers in constructor * [WIP] Move module event initialization to constrictors * [WIP] Fix setting for enabling and disabling the SPN module * [WIP] Move API registeration into module construction * [WIP] Update states mgr for all modules * [WIP] Add CmdLine operation support * Add state helper methods to module group and instance * Add notification and module status handling to status package * Fix starting issues * Remove pilot widget and update security lock to new status data * Remove debug logs * Improve http server shutdown * Add workaround for cleanly shutting down firewall+netquery * Improve logging * Add syncing states with notifications for new module system * Improve starting, stopping, shutdown; resolve FIXMEs/TODOs * [WIP] Fix most unit tests * Review new module system and fix minor issues * Push shutdown and restart events again via API * Set sleep mode via interface * Update example/template module * [WIP] Fix spn/cabin unit test * Remove deprecated UI elements * Make log output more similar for the logging transition phase * Switch spn hub and observer cmds to new module system * Fix log sources * Make worker mgr less error prone * Fix tests and minor issues * Fix observation hub * Improve shutdown and restart handling * Split up big connection.go source file * Move varint and dsd packages to structures repo * Improve expansion test * Fix linter warnings * Fix interception module on windows * Fix linter errors --------- Co-authored-by: Vladimir Stoilov <vladimir@safing.io>
594 lines
16 KiB
Go
594 lines
16 KiB
Go
package resolver
|
|
|
|
import (
|
|
"context"
|
|
"errors"
|
|
"fmt"
|
|
"net"
|
|
"strings"
|
|
"sync"
|
|
"time"
|
|
|
|
"github.com/miekg/dns"
|
|
"golang.org/x/net/publicsuffix"
|
|
|
|
"github.com/safing/portmaster/base/database"
|
|
"github.com/safing/portmaster/base/log"
|
|
"github.com/safing/portmaster/service/mgr"
|
|
"github.com/safing/portmaster/service/netenv"
|
|
)
|
|
|
|
// Errors.
|
|
var (
|
|
// Basic Errors.
|
|
|
|
// ErrNotFound is a basic error that will match all "not found" errors.
|
|
ErrNotFound = errors.New("record could not be found")
|
|
// ErrBlocked is basic error that will match all "blocked" errors.
|
|
ErrBlocked = errors.New("query was blocked")
|
|
// ErrLocalhost is returned to *.localhost queries.
|
|
ErrLocalhost = errors.New("query for localhost")
|
|
// ErrTimeout is returned when a query times out.
|
|
ErrTimeout = errors.New("query timed out")
|
|
// ErrOffline is returned when no network connection is detected.
|
|
ErrOffline = errors.New("device is offine")
|
|
// ErrFailure is returned when the type of failure is unclear.
|
|
ErrFailure = errors.New("query failed")
|
|
// ErrContinue is returned when the resolver has no answer, and the next resolver should be asked.
|
|
ErrContinue = errors.New("resolver has no answer")
|
|
// ErrShuttingDown is returned when the resolver is shutting down.
|
|
ErrShuttingDown = errors.New("resolver is shutting down")
|
|
|
|
// Detailed Errors.
|
|
|
|
// ErrTestDomainsDisabled wraps ErrBlocked.
|
|
ErrTestDomainsDisabled = fmt.Errorf("%w: test domains disabled", ErrBlocked)
|
|
// ErrSpecialDomainsDisabled wraps ErrBlocked.
|
|
ErrSpecialDomainsDisabled = fmt.Errorf("%w: special domains disabled", ErrBlocked)
|
|
// ErrInvalid wraps ErrNotFound.
|
|
ErrInvalid = fmt.Errorf("%w: invalid request", ErrNotFound)
|
|
// ErrNoCompliance wraps ErrBlocked and is returned when no resolvers were able to comply with the current settings.
|
|
ErrNoCompliance = fmt.Errorf("%w: no compliant resolvers for this query", ErrBlocked)
|
|
)
|
|
|
|
const (
|
|
minTTL = 60 // 1 Minute
|
|
refreshTTL = minTTL / 2
|
|
minMDnsTTL = 60 // 1 Minute
|
|
maxTTL = 24 * 60 * 60 // 24 hours
|
|
)
|
|
|
|
var (
|
|
dupReqMap = make(map[string]*dedupeStatus)
|
|
dupReqLock sync.Mutex
|
|
)
|
|
|
|
type dedupeStatus struct {
|
|
completed chan struct{}
|
|
waitUntil time.Time
|
|
superseded bool
|
|
}
|
|
|
|
// BlockedUpstreamError is returned when a DNS request
|
|
// has been blocked by the upstream server.
|
|
type BlockedUpstreamError struct {
|
|
ResolverName string
|
|
}
|
|
|
|
func (blocked *BlockedUpstreamError) Error() string {
|
|
return fmt.Sprintf("%s by upstream DNS resolver %s", ErrBlocked, blocked.ResolverName)
|
|
}
|
|
|
|
// Unwrap implements errors.Unwrapper.
|
|
func (blocked *BlockedUpstreamError) Unwrap() error {
|
|
return ErrBlocked
|
|
}
|
|
|
|
// Query describes a dns query.
|
|
type Query struct {
|
|
FQDN string
|
|
QType dns.Type
|
|
NoCaching bool
|
|
IgnoreFailing bool
|
|
LocalResolversOnly bool
|
|
|
|
// ICANNSpace signifies if the domain is within ICANN managed domain space.
|
|
ICANNSpace bool
|
|
// Domain root is the effective TLD +1.
|
|
DomainRoot string
|
|
|
|
// internal
|
|
dotPrefixedFQDN string
|
|
}
|
|
|
|
// ID returns the ID of the query consisting of the domain and question type.
|
|
func (q *Query) ID() string {
|
|
return q.FQDN + q.QType.String()
|
|
}
|
|
|
|
// InitPublicSuffixData initializes the public suffix data.
|
|
func (q *Query) InitPublicSuffixData() {
|
|
// Get public suffix and derive if domain is in ICANN space.
|
|
suffix, icann := publicsuffix.PublicSuffix(strings.TrimSuffix(q.FQDN, "."))
|
|
if icann || strings.Contains(suffix, ".") {
|
|
q.ICANNSpace = true
|
|
}
|
|
// Override some cases.
|
|
switch suffix {
|
|
case "example":
|
|
q.ICANNSpace = true // Defined by ICANN.
|
|
case "invalid":
|
|
q.ICANNSpace = true // Defined by ICANN.
|
|
case "local":
|
|
q.ICANNSpace = true // Defined by ICANN.
|
|
case "localhost":
|
|
q.ICANNSpace = true // Defined by ICANN.
|
|
case "onion":
|
|
q.ICANNSpace = false // Defined by ICANN, but special.
|
|
case "test":
|
|
q.ICANNSpace = true // Defined by ICANN.
|
|
}
|
|
// Add suffix to adhere to FQDN format.
|
|
suffix += "."
|
|
|
|
switch {
|
|
case len(q.FQDN) == len(suffix):
|
|
// We are at or below the domain root, reset.
|
|
q.DomainRoot = ""
|
|
case len(q.FQDN) > len(suffix):
|
|
domainRootStart := strings.LastIndex(q.FQDN[:len(q.FQDN)-len(suffix)-1], ".") + 1
|
|
q.DomainRoot = q.FQDN[domainRootStart:]
|
|
}
|
|
}
|
|
|
|
// check runs sanity checks and does some initialization. Returns whether the query passed the basic checks.
|
|
func (q *Query) check() (ok bool) {
|
|
if q.FQDN == "" {
|
|
return false
|
|
}
|
|
|
|
// init
|
|
q.FQDN = dns.Fqdn(q.FQDN)
|
|
if q.FQDN == "." {
|
|
q.dotPrefixedFQDN = q.FQDN
|
|
} else {
|
|
q.dotPrefixedFQDN = "." + q.FQDN
|
|
}
|
|
|
|
return true
|
|
}
|
|
|
|
// Resolve resolves the given query for a domain and type and returns a RRCache object or nil, if the query failed.
|
|
func Resolve(ctx context.Context, q *Query) (rrCache *RRCache, err error) {
|
|
// sanity check
|
|
if q == nil || !q.check() {
|
|
return nil, ErrInvalid
|
|
}
|
|
|
|
// log
|
|
// try adding a context tracer
|
|
ctx, tracer := log.AddTracer(ctx)
|
|
defer tracer.Submit()
|
|
log.Tracer(ctx).Tracef("resolver: resolving %s%s", q.FQDN, q.QType)
|
|
|
|
// check query compliance
|
|
if err = q.checkCompliance(); err != nil {
|
|
return nil, err
|
|
}
|
|
|
|
// check the cache
|
|
if !q.NoCaching {
|
|
rrCache = checkCache(ctx, q)
|
|
if rrCache != nil {
|
|
switch {
|
|
case !rrCache.Expired():
|
|
// Return non-expired cached entry immediately.
|
|
return rrCache, nil
|
|
case useStaleCache():
|
|
// Return expired cache if we should use stale cache entries,
|
|
// but start an async query instead.
|
|
log.Tracer(ctx).Tracef(
|
|
"resolver: using stale cache entry that expired %s ago",
|
|
time.Since(time.Unix(rrCache.Expires, 0)).Round(time.Second),
|
|
)
|
|
startAsyncQuery(ctx, q, rrCache)
|
|
return rrCache, nil
|
|
}
|
|
}
|
|
|
|
// dedupe!
|
|
markRequestFinished := deduplicateRequest(ctx, q)
|
|
if markRequestFinished == nil {
|
|
// we waited for another request, recheck the cache!
|
|
rrCache = checkCache(ctx, q)
|
|
if rrCache != nil && (!rrCache.Expired() || useStaleCache()) {
|
|
// Return non-expired or expired entry if we should use stale cache entries.
|
|
// There just was a request, so do not trigger an async query.
|
|
return rrCache, nil
|
|
}
|
|
log.Tracer(ctx).Debugf("resolver: waited for another %s%s query, but cache missed!", q.FQDN, q.QType)
|
|
// if cache is still empty or non-compliant, go ahead and just query
|
|
} else {
|
|
// we are the first!
|
|
defer markRequestFinished()
|
|
}
|
|
}
|
|
|
|
return resolveAndCache(ctx, q, rrCache)
|
|
}
|
|
|
|
func checkCache(ctx context.Context, q *Query) *RRCache {
|
|
// Never ask cache for connectivity domains.
|
|
if netenv.IsConnectivityDomain(q.FQDN) {
|
|
return nil
|
|
}
|
|
|
|
// Get data from cache.
|
|
rrCache, err := GetRRCache(q.FQDN, q.QType)
|
|
// Return if entry is not in cache.
|
|
if err != nil {
|
|
if !errors.Is(err, database.ErrNotFound) {
|
|
log.Tracer(ctx).Warningf("resolver: getting RRCache %s%s from database failed: %s", q.FQDN, q.QType.String(), err)
|
|
}
|
|
return nil
|
|
}
|
|
|
|
// Get the resolver that the rrCache was resolved with.
|
|
resolver := getActiveResolverByIDWithLocking(rrCache.Resolver.ID())
|
|
if resolver == nil {
|
|
log.Tracer(ctx).Debugf("resolver: ignoring RRCache %s%s because source server %q has been removed", q.FQDN, q.QType.String(), rrCache.Resolver.ID())
|
|
return nil
|
|
}
|
|
|
|
// Check compliance of the resolver, return if non-compliant.
|
|
err = resolver.checkCompliance(ctx, q)
|
|
if err != nil {
|
|
log.Tracer(ctx).Debugf("resolver: cached entry for %s%s does not comply to query parameters: %s", q.FQDN, q.QType.String(), err)
|
|
return nil
|
|
}
|
|
|
|
switch {
|
|
case shouldResetCache(q):
|
|
// Check if we want to reset the cache for this entry.
|
|
err := ResetCachedRecord(q.FQDN, q.QType.String())
|
|
switch {
|
|
case err == nil:
|
|
log.Tracer(ctx).Infof("resolver: cache for %s%s was reset", q.FQDN, q.QType)
|
|
case errors.Is(err, database.ErrNotFound):
|
|
log.Tracer(ctx).Tracef("resolver: cache for %s%s was already reset (is empty)", q.FQDN, q.QType)
|
|
default:
|
|
log.Tracer(ctx).Warningf("resolver: failed to reset cache for %s%s: %s", q.FQDN, q.QType, err)
|
|
}
|
|
return nil
|
|
|
|
case rrCache.Expired():
|
|
// Check if the cache has already expired.
|
|
// We still return the cache, if it isn't NXDomain, as it will be used if the
|
|
// new query fails.
|
|
if rrCache.RCode == dns.RcodeSuccess {
|
|
return rrCache
|
|
}
|
|
return nil
|
|
|
|
case rrCache.ExpiresSoon():
|
|
// Check if the cache will expire soon and start an async request.
|
|
startAsyncQuery(ctx, q, rrCache)
|
|
return rrCache
|
|
|
|
default:
|
|
// Return still valid cache entry.
|
|
log.Tracer(ctx).Tracef(
|
|
"resolver: using cached RR (expires in %s)",
|
|
time.Until(time.Unix(rrCache.Expires, 0)).Round(time.Second),
|
|
)
|
|
return rrCache
|
|
}
|
|
}
|
|
|
|
func startAsyncQuery(ctx context.Context, q *Query, currentRRCache *RRCache) {
|
|
// Check if an async query was already started.
|
|
if currentRRCache.RequestingNew {
|
|
return
|
|
}
|
|
|
|
// Set flag and log that we are refreshing this entry.
|
|
currentRRCache.RequestingNew = true
|
|
if currentRRCache.Expired() {
|
|
log.Tracer(ctx).Tracef(
|
|
"resolver: cache for %s has expired %s ago, refreshing async now",
|
|
q.ID(),
|
|
time.Since(time.Unix(currentRRCache.Expires, 0)).Round(time.Second),
|
|
)
|
|
} else {
|
|
log.Tracer(ctx).Tracef(
|
|
"resolver: cache for %s will expire in %s, refreshing async now",
|
|
q.ID(),
|
|
time.Until(time.Unix(currentRRCache.Expires, 0)).Round(time.Second),
|
|
)
|
|
}
|
|
|
|
// resolve async
|
|
module.mgr.Go("resolve async", func(wc *mgr.WorkerCtx) error {
|
|
tracingCtx, tracer := log.AddTracer(wc.Ctx())
|
|
defer tracer.Submit()
|
|
tracer.Tracef("resolver: resolving %s async", q.ID())
|
|
_, err := resolveAndCache(tracingCtx, q, nil)
|
|
if err != nil {
|
|
tracer.Warningf("resolver: async query for %s failed: %s", q.ID(), err)
|
|
} else {
|
|
tracer.Infof("resolver: async query for %s succeeded", q.ID())
|
|
}
|
|
return nil
|
|
})
|
|
}
|
|
|
|
func deduplicateRequest(ctx context.Context, q *Query) (finishRequest func()) {
|
|
// create identifier key
|
|
dupKey := q.ID()
|
|
|
|
// restart here if waiting timed out
|
|
retry:
|
|
|
|
dupReqLock.Lock()
|
|
|
|
// get duplicate request waitgroup
|
|
status, requestActive := dupReqMap[dupKey]
|
|
|
|
// check if the request ist active
|
|
if requestActive {
|
|
// someone else is already on it!
|
|
if time.Now().Before(status.waitUntil) {
|
|
dupReqLock.Unlock()
|
|
|
|
// log that we are waiting
|
|
log.Tracer(ctx).Tracef("resolver: waiting for duplicate query for %s to complete", dupKey)
|
|
// wait
|
|
select {
|
|
case <-status.completed:
|
|
// done!
|
|
return nil
|
|
case <-time.After(maxRequestTimeout):
|
|
// something went wrong with the query, retry
|
|
goto retry
|
|
case <-ctx.Done():
|
|
return nil
|
|
}
|
|
} else {
|
|
// but that someone is taking too long
|
|
status.superseded = true
|
|
}
|
|
}
|
|
|
|
// we are currently the only one doing a request for this
|
|
|
|
// create new status
|
|
status = &dedupeStatus{
|
|
completed: make(chan struct{}),
|
|
waitUntil: time.Now().Add(maxRequestTimeout),
|
|
}
|
|
// add to registry
|
|
dupReqMap[dupKey] = status
|
|
|
|
dupReqLock.Unlock()
|
|
|
|
// return function to mark request as finished
|
|
return func() {
|
|
dupReqLock.Lock()
|
|
defer dupReqLock.Unlock()
|
|
// mark request as done
|
|
close(status.completed)
|
|
// delete from registry
|
|
if !status.superseded {
|
|
delete(dupReqMap, dupKey)
|
|
}
|
|
}
|
|
}
|
|
|
|
func resolveAndCache(ctx context.Context, q *Query, oldCache *RRCache) (rrCache *RRCache, err error) { //nolint:gocognit,gocyclo
|
|
// get resolvers
|
|
resolvers, primarySource, tryAll := GetResolversInScope(ctx, q)
|
|
if len(resolvers) == 0 {
|
|
return nil, ErrNoCompliance
|
|
}
|
|
|
|
// check if we are online
|
|
if netenv.GetOnlineStatus() == netenv.StatusOffline && primarySource != ServerSourceEnv {
|
|
if q.FQDN != netenv.DNSTestDomain && !netenv.IsConnectivityDomain(q.FQDN) {
|
|
// we are offline and this is not an online check query
|
|
return oldCache, ErrOffline
|
|
}
|
|
log.Tracer(ctx).Debugf("resolver: allowing online status test domain %s to resolve even though offline", q.FQDN)
|
|
}
|
|
|
|
// Report when all configured resolvers are failing.
|
|
var failureReported bool
|
|
defer func() {
|
|
if failureReported &&
|
|
netenv.Online() &&
|
|
primarySource == ServerSourceConfigured &&
|
|
allConfiguredResolversAreFailing() {
|
|
notifyAboutFailingResolvers()
|
|
}
|
|
}()
|
|
|
|
// start resolving
|
|
for _, resolver := range resolvers {
|
|
if module.mgr.IsDone() {
|
|
return nil, ErrShuttingDown
|
|
}
|
|
|
|
// Skip failing resolvers.
|
|
if resolver.Conn.IsFailing() {
|
|
log.Tracer(ctx).Tracef("resolver: skipping resolver %s, because it is failing", resolver)
|
|
continue
|
|
}
|
|
|
|
// Skip unreachable link-local resolvers.
|
|
if resolver.LinkLocalUnavailable {
|
|
log.Tracer(ctx).Tracef("resolver: skipping resolver %s, because it is link-local and not available", resolver)
|
|
continue
|
|
}
|
|
|
|
// resolve
|
|
log.Tracer(ctx).Tracef("resolver: sending query for %s to %s", q.ID(), resolver.Info.ID())
|
|
rrCache, err = resolver.Conn.Query(ctx, q)
|
|
if err != nil {
|
|
switch {
|
|
case errors.Is(err, ErrNotFound):
|
|
// NXDomain, or similar
|
|
if tryAll {
|
|
continue
|
|
}
|
|
return nil, err
|
|
case errors.Is(err, ErrBlocked):
|
|
// some resolvers might also block
|
|
return nil, err
|
|
case netenv.GetOnlineStatus() == netenv.StatusOffline &&
|
|
q.FQDN != netenv.DNSTestDomain &&
|
|
!netenv.IsConnectivityDomain(q.FQDN):
|
|
// we are offline and this is not an online check query
|
|
return oldCache, ErrOffline
|
|
case errors.Is(err, ErrContinue):
|
|
continue
|
|
case errors.Is(err, ErrTimeout):
|
|
resolver.Conn.ReportFailure()
|
|
failureReported = true
|
|
log.Tracer(ctx).Debugf("resolver: query to %s timed out", resolver.Info.ID())
|
|
continue
|
|
case errors.Is(err, context.Canceled):
|
|
return nil, err
|
|
case errors.Is(err, context.DeadlineExceeded):
|
|
return nil, err
|
|
case errors.Is(err, ErrShuttingDown):
|
|
return nil, err
|
|
default:
|
|
resolver.Conn.ReportFailure()
|
|
failureReported = true
|
|
log.Tracer(ctx).Warningf("resolver: query to %s failed: %s", resolver.Info.ID(), err)
|
|
continue
|
|
}
|
|
}
|
|
if rrCache == nil {
|
|
// Defensive: This should normally not happen.
|
|
continue
|
|
}
|
|
// Check if request succeeded and whether we should try another resolver.
|
|
if rrCache.RCode != dns.RcodeSuccess && tryAll {
|
|
continue
|
|
}
|
|
|
|
// Report a successful connection.
|
|
resolver.Conn.ResetFailure()
|
|
// Reset failing resolvers notification, if querying in global scope.
|
|
if primarySource == ServerSourceConfigured && !allConfiguredResolversAreFailing() {
|
|
resetFailingResolversNotification()
|
|
}
|
|
|
|
break
|
|
}
|
|
|
|
// Validate return values.
|
|
if err == nil && rrCache == nil {
|
|
err = ErrNotFound
|
|
}
|
|
|
|
// Handle error.
|
|
if err != nil {
|
|
// Check if we can return an older cache instead of the error.
|
|
if oldCache != nil {
|
|
oldCache.IsBackup = true
|
|
log.Tracer(ctx).Debugf("resolver: serving backup cache of %s because query failed: %s", q.ID(), err)
|
|
return oldCache, nil
|
|
}
|
|
|
|
return nil, fmt.Errorf("all %d query-compliant resolvers failed, last error: %w", len(resolvers), err)
|
|
}
|
|
|
|
// Adjust TTLs.
|
|
rrCache.Clean(minTTL)
|
|
|
|
// Save the new entry if cache is enabled and the record may be cached.
|
|
if !q.NoCaching && rrCache.Cacheable() {
|
|
err = rrCache.Save()
|
|
if err != nil {
|
|
log.Tracer(ctx).Warningf("resolver: failed to cache RR for %s: %s", q.ID(), err)
|
|
}
|
|
}
|
|
|
|
return rrCache, nil
|
|
}
|
|
|
|
var (
|
|
cacheResetLock sync.Mutex
|
|
cacheResetID string
|
|
cacheResetSeenCnt int
|
|
)
|
|
|
|
func shouldResetCache(q *Query) (reset bool) {
|
|
cacheResetLock.Lock()
|
|
defer cacheResetLock.Unlock()
|
|
|
|
// reset to new domain
|
|
qID := q.ID()
|
|
if qID != cacheResetID {
|
|
cacheResetID = qID
|
|
cacheResetSeenCnt = 1
|
|
return false
|
|
}
|
|
|
|
// increase and check if threshold is reached
|
|
cacheResetSeenCnt++
|
|
if cacheResetSeenCnt >= 3 { // 3 to trigger reset
|
|
cacheResetSeenCnt = -7 // 10 for follow-up resets
|
|
return true
|
|
}
|
|
|
|
return false
|
|
}
|
|
|
|
// testConnectivity test if resolving a query succeeds and returns whether the
|
|
// query itself succeeded, separate from interpreting the result.
|
|
// Provide a resolver to use or automatically select one if nil.
|
|
func testConnectivity(ctx context.Context, fdqn string, resolver *Resolver) (ips []net.IP, ok bool, err error) {
|
|
q := &Query{
|
|
FQDN: fdqn,
|
|
QType: dns.Type(dns.TypeA),
|
|
NoCaching: true,
|
|
}
|
|
if !q.check() {
|
|
return nil, false, ErrInvalid
|
|
}
|
|
|
|
// Resolve with given resolver or auto-select.
|
|
var rrCache *RRCache
|
|
if resolver != nil {
|
|
rrCache, err = resolver.Conn.Query(ctx, q)
|
|
} else {
|
|
rrCache, err = resolveAndCache(ctx, q, nil)
|
|
}
|
|
|
|
// Enhance results.
|
|
switch {
|
|
case err == nil:
|
|
switch rrCache.RCode {
|
|
case dns.RcodeNameError:
|
|
return nil, true, ErrNotFound
|
|
case dns.RcodeRefused:
|
|
return nil, true, errors.New("refused")
|
|
default:
|
|
ips := rrCache.ExportAllARecords()
|
|
if len(ips) > 0 {
|
|
return ips, true, nil
|
|
}
|
|
return nil, true, ErrNotFound
|
|
}
|
|
case errors.Is(err, ErrNotFound):
|
|
return nil, true, err
|
|
case errors.Is(err, ErrNoCompliance):
|
|
return nil, true, err
|
|
case errors.Is(err, ErrBlocked):
|
|
return nil, true, err
|
|
default:
|
|
return nil, false, err
|
|
}
|
|
}
|