mirror of
https://github.com/gravitl/netmaker.git
synced 2026-04-23 00:17:10 +08:00
292af315dd
* feat(go): add user schema; * feat(go): migrate to user schema; * feat(go): add audit fields; * feat(go): remove unused fields from the network model; * feat(go): add network schema; * feat(go): migrate to network schema; * refactor(go): add comment to clarify migration logic; * fix(go): test failures; * fix(go): test failures; * feat(go): change membership table to store memberships at all scopes; * feat(go): add schema for access grants; * feat(go): remove nameservers from new networks table; ensure db passed for schema functions; * feat(go): set max conns for sqlite to 1; * fix(go): issues updating user account status; * NM-236: streamline operations in HA mode * NM-236: only master pod should subscribe to updates from clients * refactor(go): remove converters and access grants; * refactor(go): add json tags in schema models; * refactor(go): rename file to migrate_v1_6_0.go; * refactor(go): add user groups and user roles tables; use schema tables; * refactor(go): inline get and list from schema package; * refactor(go): inline get network and list users from schema package; * fix(go): staticcheck issues; * fix(go): remove test not in use; fix test case; * fix(go): validate network; * fix(go): resolve static checks; * fix(go): new models errors; * fix(go): test errors; * fix(go): handle no records; * fix(go): add validations for user object; * fix(go): set correct extclient status; * fix(go): test error; * feat(go): make schema the base package; * feat(go): add host schema; * feat(go): use schema host everywhere; * feat(go): inline get host, list hosts and delete host; * feat(go): use non-ptr value; * feat(go): use save to upsert all fields; * feat(go): use save to upsert all fields; * feat(go): save turn endpoint as string; * feat(go): check for gorm error record not found; * fix(go): test failures; * fix(go): update all network fields; * fix(go): update all network fields; * feat(go): add paginated list networks api; * feat(go): add paginated list users api; * feat(go): add paginated list hosts api; * feat(go): add pagination to list groups api; * fix(go): comment; * fix(go): implement marshal and unmarshal text for custom types; * fix(go): implement marshal and unmarshal json for custom types; * fix(go): just use the old model for unmarshalling; * fix(go): implement marshal and unmarshal json for custom types; * NM-271:Import swap: compress/gzip replaced with github.com/klauspost/compress/gzip (2-4x faster, wire-compatible output). Added sync import. Two sync.Pool variables (gzipWriterPool, bufferPool): reuse gzip.Writer and bytes.Buffer across calls instead of allocating fresh ones per publish. compressPayload rewritten: pulls writer + buffer from pools, resets them, compresses at gzip.BestSpeed (level 1), copies the result out of the pooled buffer, and returns both objects to the pools. * feat(go): remove paginated list networks api; * feat(go): use custom paginated response object; * NM-271: Improve server scalability under high host count - Replace stdlib compress/gzip with klauspost/compress at BestSpeed and pool gzip writers and buffers via sync.Pool to eliminate compression as the dominant CPU hotspot. - Debounce peer update broadcasts with a 500ms resettable window capped at 3s max-wait, coalescing rapid-fire PublishPeerUpdate calls into a single broadcast cycle. - Cache HostPeerInfo (batch-refreshed by debounce worker) and HostPeerUpdate (stored as side-effect of each publish) so the pull API and peer_info API serve from pre-computed maps instead of triggering expensive per-host computations under thundering herd conditions. - Warm both caches synchronously at startup before the first publish cycle so early pull requests are served instantly. - Bound concurrent MQTT publishes to 5 via semaphore to prevent broker TCP buffer overflows that caused broken pipe disconnects. - Remove manual Disconnect+SetupMQTT from ConnectionLostHandler and rely on the paho client's built-in AutoReconnect; add a 5s retry wait in publish() to ride out brief reconnection windows. * NM-271: Reduce server CPU contention under high concurrent load - Cache ServerSettings with atomic.Value to eliminate repeated DB reads on every pull request (was 32+ goroutines blocked on read lock) - Batch UpdateNodeCheckin writes in memory, flush every 30s to reduce per-checkin write lock contention (was 88+ goroutines blocked) - Enable SQLite WAL mode + busy_timeout and remove global dbMutex; let SQLite handle concurrency natively (reads no longer block writes) - Move ResetFailedOverPeer/ResetAutoRelayedPeer to async in pull() handler since results don't affect the cached response - Skip no-op UpsertNode writes in failover/relay reset functions (early return when node has no failover/relay state) - Remove CheckHostPorts from hostUpdateFallback hot path - Switch to pure-Go SQLite driver (glebarez/sqlite), set CGO_ENABLED=0 * fix(go): ensure default values for page and per_page are used when not passed; * fix(go): rename v1.6.0 to v1.5.1; * fix(go): check for gorm.ErrRecordNotFound instead of database.IsEmptyRecord; * fix(go): use host id, not pending host id; * NM-271: Revert pure-Go SQLite and FIPS disable to verify impact Revert to CGO-based mattn/go-sqlite3 driver and re-enable FIPS to isolate whether these changes are still needed now that the global dbMutex has been removed and WAL mode is enabled. Keep WAL mode pragma with mattn-compatible DSN format. * feat(go): add filters to paginated apis; * feat(go): add filters to paginated apis; * feat(go): remove check for max username length; * feat(go): add filters to count as well; * feat(go): use library to check email address validity; * feat(go): ignore pagination if params not passed; * fix(go): pagination issues; * fix(go): check exists before using; * fix(go): remove debug log; * NM-271: rm debug logs * NM-271: check if caching is enabled * NM-271: add server sync mq topic for HA mode * NM-271: fix build * NM-271: push metrics in batch to exproter over api * NM-271: use basic auth for exporter metrics api * fix(go): use gorm err record not found; * NM-271: Add monitoring stack on demand * NM-271: -m arg for install script should only add monitoring stack * fix(go): use gorm err record not found; * NM-271: update docker compose file for prometheus * NM-271: update docker compose file for prometheus * fix(go): use user principal name when creating pending user; * fix(go): use schema package for consts; * NM-236: rm duplicate network hook * NM-271: add server topic to reset idp hooks on master node * fix(go): prevent disabling superadmin user; Co-authored-by: tenki-reviewer[bot] <262613592+tenki-reviewer[bot]@users.noreply.github.com> * fix(go): swap is admin and is superadmin; Co-authored-by: tenki-reviewer[bot] <262613592+tenki-reviewer[bot]@users.noreply.github.com> * fix(go): remove dead code block; https://github.com/gravitl/netmaker/pull/3910#discussion_r2928837937 * fix(go): incorrect message when trying to disable self; https://github.com/gravitl/netmaker/pull/3910#discussion_r2928837934 * NM-271: fix stale peers on reset_failovered pull and add HTTP timeout to metrics exporter Run the failover/relay reset synchronously in the pull handler so the response reflects post-reset topology instead of serving stale cached peers. Add a 30s timeout to the metrics exporter HTTP client to prevent PushAllMetricsToExporter from blocking the Keepalive loop. * NM-271: fix gzip pool corruption, MQTT topic mismatch, stale settings cache, and reduce redundant DB fetches - Only return gzip.Writer to pool after successful Close to prevent silently malformed MQTT payloads from a previously errored writer. - Fix serversync subscription to exact topic match since syncType is now in the message payload, not the topic path. - Prevent zero-value ServerSettings from being cached indefinitely when the DB record is missing or unmarshal fails on startup. - Return fetched hosts/nodes from RefreshHostPeerInfoCache so warmPeerCaches reuses them instead of querying the DB twice. - Compute fresh HostPeerUpdate on reset_failovered pull instead of serving stale cache, and store result back for subsequent requests. * NM-271: fix gzip writer pool leak, log checkin flush errors, and fix master pod ordinal parsing - Reset gzip.Writer to io.Discard before returning to pool so errored writers are never leaked or silently reused with corrupt state. - Track and log failed DB inserts in FlushNodeCheckins so operators have visibility when check-in timestamps are lost. - Parse StatefulSet pod ordinal as integer instead of using HasSuffix to prevent netmaker-10 from being misidentified as master pod. * NM-271: simplify masterpod logic * fix(go): use correct header; Co-authored-by: tenki-reviewer[bot] <262613592+tenki-reviewer[bot]@users.noreply.github.com> * fix(go): return after error response; Co-authored-by: tenki-reviewer[bot] <262613592+tenki-reviewer[bot]@users.noreply.github.com> * fix(go): use correct order of params; https://github.com/gravitl/netmaker/pull/3910#discussion_r2929593036 * fix(go): set default values for page and page size; use v2 instead of /list; * NM-271: use host name * Update mq/serversync.go Co-authored-by: tenki-reviewer[bot] <262613592+tenki-reviewer[bot]@users.noreply.github.com> * NM-271: fix duplicate serversynce case * NM-271: streamline gw updates * Update logic/auth.go Co-authored-by: tenki-reviewer[bot] <262613592+tenki-reviewer[bot]@users.noreply.github.com> * Update schema/user_roles.go Co-authored-by: tenki-reviewer[bot] <262613592+tenki-reviewer[bot]@users.noreply.github.com> * fix(go): syntax error; * fix(go): set default values when page and per_page are not passed or 0; * fix(go): use uuid.parse instead of uuid.must parse; * fix(go): review errors; * fix(go): review errors; * Update controllers/user.go Co-authored-by: tenki-reviewer[bot] <262613592+tenki-reviewer[bot]@users.noreply.github.com> * Update controllers/user.go Co-authored-by: tenki-reviewer[bot] <262613592+tenki-reviewer[bot]@users.noreply.github.com> * NM-163: fix errors: * Update db/types/options.go Co-authored-by: tenki-reviewer[bot] <262613592+tenki-reviewer[bot]@users.noreply.github.com> * fix(go): persist return user in event; * Update db/types/options.go Co-authored-by: tenki-reviewer[bot] <262613592+tenki-reviewer[bot]@users.noreply.github.com> * NM-271: signal pull on ip changes * NM-163: duplicate lines of code * NM-163: fix(go): fix missing return and filter parsing in user controller - Add missing return after error response in updateUserAccountStatus to prevent double-response and spurious ext-client side-effects - Use switch statements in listUsers to skip unrecognized account_status and mfa_status filter values * NM-271: signal pull req on node ip change * fix(go): check for both min and max page size; * NM-271: refresh node object before update * fix(go): enclose transfer superadmin in transaction; * fix(go): review errors; * fix(go): remove free tier checks; * fix(go): review fixes; * NM-271: streamline ip pool ops * NM-271: fix tests, set max idle conns * NM-271: fix(go): fix data races in settings cache and peer update worker - Use pointer type in atomic.Value for serverSettingsCache to avoid replacing the variable non-atomically in InvalidateServerSettingsCache - Swap peerUpdateReplace flag before draining the channel to prevent a concurrent replacePeers=true from being consumed by the wrong cycle --------- Co-authored-by: VishalDalwadi <dalwadivishal26@gmail.com> Co-authored-by: Vishal Dalwadi <51291657+VishalDalwadi@users.noreply.github.com> Co-authored-by: tenki-reviewer[bot] <262613592+tenki-reviewer[bot]@users.noreply.github.com>
577 lines
18 KiB
Go
577 lines
18 KiB
Go
package controller
|
|
|
|
import (
|
|
"context"
|
|
"encoding/json"
|
|
"errors"
|
|
"fmt"
|
|
"net/http"
|
|
"slices"
|
|
|
|
"github.com/google/uuid"
|
|
"github.com/gorilla/mux"
|
|
"github.com/gravitl/netmaker/db"
|
|
"github.com/gravitl/netmaker/logger"
|
|
"github.com/gravitl/netmaker/logic"
|
|
"github.com/gravitl/netmaker/models"
|
|
"github.com/gravitl/netmaker/mq"
|
|
"github.com/gravitl/netmaker/schema"
|
|
"github.com/gravitl/netmaker/servercfg"
|
|
"golang.org/x/exp/slog"
|
|
)
|
|
|
|
func gwHandlers(r *mux.Router) {
|
|
r.HandleFunc("/api/nodes/{network}/{nodeid}/gateway", logic.SecurityCheck(true, http.HandlerFunc(createGateway))).Methods(http.MethodPost)
|
|
r.HandleFunc("/api/nodes/{network}/{nodeid}/gateway", logic.SecurityCheck(true, http.HandlerFunc(deleteGateway))).Methods(http.MethodDelete)
|
|
r.HandleFunc("/api/nodes/{network}/{nodeid}/gateway/assign", logic.SecurityCheck(true, http.HandlerFunc(assignGw))).Methods(http.MethodPost)
|
|
r.HandleFunc("/api/nodes/{network}/{nodeid}/gateway/unassign", logic.SecurityCheck(true, http.HandlerFunc(unassignGw))).Methods(http.MethodPost)
|
|
// old relay handlers
|
|
r.HandleFunc("/api/nodes/{network}/{nodeid}/createrelay", logic.SecurityCheck(true, http.HandlerFunc(createGateway))).Methods(http.MethodPost)
|
|
r.HandleFunc("/api/nodes/{network}/{nodeid}/deleterelay", logic.SecurityCheck(true, http.HandlerFunc(deleteGateway))).Methods(http.MethodDelete)
|
|
}
|
|
|
|
// @Summary Create a gateway
|
|
// @Router /api/nodes/{network}/{nodeid}/gateway [post]
|
|
// @Tags Gateways
|
|
// @Security oauth
|
|
// @Accept json
|
|
// @Produce json
|
|
// @Param network path string true "Network ID"
|
|
// @Param nodeid path string true "Node ID"
|
|
// @Param body body models.CreateGwReq true "Gateway request"
|
|
// @Success 200 {object} models.ApiNode
|
|
// @Failure 500 {object} models.ErrorResponse
|
|
func createGateway(w http.ResponseWriter, r *http.Request) {
|
|
w.Header().Set("Content-Type", "application/json")
|
|
var params = mux.Vars(r)
|
|
nodeid := params["nodeid"]
|
|
netid := params["network"]
|
|
node, err := logic.ValidateParams(nodeid, netid)
|
|
if err != nil {
|
|
logic.ReturnErrorResponse(w, r, logic.FormatError(err, "badrequest"))
|
|
return
|
|
}
|
|
host := &schema.Host{
|
|
ID: node.HostID,
|
|
}
|
|
err = host.Get(r.Context())
|
|
if err != nil {
|
|
logic.ReturnErrorResponse(w, r, logic.FormatError(err, "badrequest"))
|
|
return
|
|
}
|
|
var req models.CreateGwReq
|
|
err = json.NewDecoder(r.Body).Decode(&req)
|
|
if err != nil {
|
|
logic.ReturnErrorResponse(w, r, logic.FormatError(err, "badrequest"))
|
|
return
|
|
}
|
|
if req.IsInternetGateway && len(req.InetNodeClientIDs) > 0 {
|
|
err = logic.ValidateInetGwReq(node, req.InetNodeReq, false)
|
|
if err != nil {
|
|
logic.ReturnErrorResponse(w, r, logic.FormatError(err, "badrequest"))
|
|
return
|
|
}
|
|
}
|
|
|
|
node, err = logic.CreateIngressGateway(netid, nodeid, req.IngressRequest)
|
|
if err != nil {
|
|
logger.Log(0, r.Header.Get("user"),
|
|
fmt.Sprintf("failed to create gateway on node [%s] on network [%s]: %v",
|
|
nodeid, netid, err))
|
|
logic.ReturnErrorResponse(w, r, logic.FormatError(err, "internal"))
|
|
return
|
|
}
|
|
req.RelayRequest.NetID = netid
|
|
req.RelayRequest.NodeID = nodeid
|
|
_, relayNode, err := logic.CreateRelay(req.RelayRequest)
|
|
if err != nil {
|
|
logger.Log(
|
|
0,
|
|
r.Header.Get("user"),
|
|
fmt.Sprintf(
|
|
"failed to create relay on node [%s] on network [%s]: %v",
|
|
req.RelayRequest.NodeID,
|
|
req.RelayRequest.NetID,
|
|
err,
|
|
),
|
|
)
|
|
logic.ReturnErrorResponse(w, r, logic.FormatError(err, "internal"))
|
|
return
|
|
}
|
|
if req.IsInternetGateway {
|
|
if host.DNS != "yes" {
|
|
host.DNS = "yes"
|
|
logic.UpsertHost(host)
|
|
}
|
|
}
|
|
|
|
if len(req.InetNodeClientIDs) > 0 {
|
|
logic.SetInternetGw(&node, req.InetNodeReq)
|
|
if node.IsGw && node.IngressDNS == "" {
|
|
node.IngressDNS = "1.1.1.1"
|
|
}
|
|
logic.UpsertNode(&node)
|
|
}
|
|
logger.Log(
|
|
1,
|
|
r.Header.Get("user"),
|
|
"created gw node",
|
|
req.RelayRequest.NodeID,
|
|
"on network",
|
|
req.RelayRequest.NetID,
|
|
)
|
|
logic.GetNodeStatus(&relayNode, false)
|
|
apiNode := relayNode.ConvertToAPINode()
|
|
logic.LogEvent(&models.Event{
|
|
Action: schema.Create,
|
|
Source: models.Subject{
|
|
ID: r.Header.Get("user"),
|
|
Name: r.Header.Get("user"),
|
|
Type: schema.UserSub,
|
|
},
|
|
TriggeredBy: r.Header.Get("user"),
|
|
Target: models.Subject{
|
|
ID: node.ID.String(),
|
|
Name: host.Name,
|
|
Type: schema.GatewaySub,
|
|
},
|
|
Origin: schema.Dashboard,
|
|
})
|
|
host.IsStaticPort = true
|
|
logic.UpsertHost(host)
|
|
w.WriteHeader(http.StatusOK)
|
|
json.NewEncoder(w).Encode(apiNode)
|
|
go func() {
|
|
for _, relayedNodeID := range relayNode.RelayedNodes {
|
|
relayedNode, err := logic.GetNodeByID(relayedNodeID)
|
|
if err == nil {
|
|
if relayedNode.FailedOverBy != uuid.Nil {
|
|
logic.ResetFailedOverPeer(&relayedNode)
|
|
}
|
|
if len(relayedNode.AutoRelayedPeers) > 0 {
|
|
logic.ResetAutoRelayedPeer(&relayedNode)
|
|
}
|
|
}
|
|
}
|
|
logic.ResetAutoRelayedPeer(&node)
|
|
if err := mq.NodeUpdate(&node); err != nil {
|
|
slog.Error("error publishing node update to node", "node", node.ID, "error", err)
|
|
}
|
|
mq.PublishPeerUpdate(false)
|
|
}()
|
|
|
|
}
|
|
|
|
// @Summary Delete a gateway
|
|
// @Router /api/nodes/{network}/{nodeid}/gateway [delete]
|
|
// @Tags Gateways
|
|
// @Security oauth
|
|
// @Produce json
|
|
// @Param network path string true "Network ID"
|
|
// @Param nodeid path string true "Node ID"
|
|
// @Success 200 {object} models.ApiNode
|
|
// @Failure 500 {object} models.ErrorResponse
|
|
func deleteGateway(w http.ResponseWriter, r *http.Request) {
|
|
w.Header().Set("Content-Type", "application/json")
|
|
var params = mux.Vars(r)
|
|
nodeid := params["nodeid"]
|
|
netid := params["network"]
|
|
node, err := logic.ValidateParams(nodeid, netid)
|
|
if err != nil {
|
|
logic.ReturnErrorResponse(w, r, logic.FormatError(err, "badrequest"))
|
|
return
|
|
}
|
|
node, removedClients, err := logic.DeleteIngressGateway(nodeid)
|
|
if err != nil {
|
|
logger.Log(0, r.Header.Get("user"),
|
|
fmt.Sprintf("failed to delete ingress gateway on node [%s] on network [%s]: %v",
|
|
nodeid, netid, err))
|
|
logic.ReturnErrorResponse(w, r, logic.FormatError(err, "internal"))
|
|
return
|
|
}
|
|
|
|
updateNodes, node, err := logic.DeleteRelay(netid, nodeid)
|
|
if err != nil {
|
|
logger.Log(0, r.Header.Get("user"), "error decoding request body: ", err.Error())
|
|
logic.ReturnErrorResponse(w, r, logic.FormatError(err, "internal"))
|
|
return
|
|
}
|
|
node, err = logic.GetNodeByID(node.ID.String())
|
|
if err != nil {
|
|
logger.Log(0, r.Header.Get("user"), "failed to get node", err.Error())
|
|
logic.ReturnErrorResponse(w, r, logic.FormatError(err, "internal"))
|
|
return
|
|
}
|
|
host := &schema.Host{
|
|
ID: node.HostID,
|
|
}
|
|
err = host.Get(r.Context())
|
|
if err != nil {
|
|
logic.ReturnErrorResponse(w, r, logic.FormatError(err, "badrequest"))
|
|
return
|
|
}
|
|
logic.UnsetInternetGw(&node)
|
|
node.IsGw = false
|
|
if node.IsAutoRelay {
|
|
logic.ResetAutoRelay(&node)
|
|
}
|
|
node.IsAutoRelay = false
|
|
logic.UpsertNode(&node)
|
|
logger.Log(1, r.Header.Get("user"), "deleted gw", nodeid, "on network", netid)
|
|
|
|
go func() {
|
|
host := &schema.Host{
|
|
ID: node.HostID,
|
|
}
|
|
err = host.Get(db.WithContext(context.TODO()))
|
|
if err == nil {
|
|
allNodes, err := logic.GetAllNodes()
|
|
if err != nil {
|
|
return
|
|
}
|
|
|
|
for _, relayedNode := range updateNodes {
|
|
err = mq.NodeUpdate(&relayedNode)
|
|
if err != nil {
|
|
logger.Log(
|
|
1,
|
|
"relayed node update ",
|
|
relayedNode.ID.String(),
|
|
"on network",
|
|
relayedNode.Network,
|
|
": ",
|
|
err.Error(),
|
|
)
|
|
|
|
}
|
|
h := &schema.Host{
|
|
ID: relayedNode.HostID,
|
|
}
|
|
err = h.Get(db.WithContext(context.TODO()))
|
|
if err == nil {
|
|
if h.OS == models.OS_Types.IoT {
|
|
nodes, err := logic.GetAllNodes()
|
|
if err != nil {
|
|
return
|
|
}
|
|
node.IsRelay = true // for iot update to recognise that it has to delete relay peer
|
|
if err = mq.PublishSingleHostPeerUpdate(h, nodes, &node, nil, false, nil); err != nil {
|
|
logger.Log(1, "failed to publish peer update to host", h.ID.String(), ": ", err.Error())
|
|
}
|
|
}
|
|
}
|
|
}
|
|
if len(removedClients) > 0 {
|
|
if err := mq.PublishSingleHostPeerUpdate(host, allNodes, nil, removedClients[:], false, nil); err != nil {
|
|
slog.Error("publishSingleHostUpdate", "host", host.Name, "error", err)
|
|
}
|
|
}
|
|
mq.PublishPeerUpdate(false)
|
|
if err := mq.NodeUpdate(&node); err != nil {
|
|
slog.Error(
|
|
"error publishing node update to node",
|
|
"node",
|
|
node.ID,
|
|
"error",
|
|
err,
|
|
)
|
|
}
|
|
if servercfg.IsDNSMode() {
|
|
logic.SetDNS()
|
|
}
|
|
|
|
}
|
|
|
|
}()
|
|
logic.LogEvent(&models.Event{
|
|
Action: schema.Delete,
|
|
Source: models.Subject{
|
|
ID: r.Header.Get("user"),
|
|
Name: r.Header.Get("user"),
|
|
Type: schema.UserSub,
|
|
},
|
|
TriggeredBy: r.Header.Get("user"),
|
|
Target: models.Subject{
|
|
ID: node.ID.String(),
|
|
Name: host.Name,
|
|
Type: schema.GatewaySub,
|
|
},
|
|
Origin: schema.Dashboard,
|
|
Diff: models.Diff{
|
|
Old: node,
|
|
New: node,
|
|
},
|
|
})
|
|
logic.GetNodeStatus(&node, false)
|
|
apiNode := node.ConvertToAPINode()
|
|
logger.Log(1, r.Header.Get("user"), "deleted ingress gateway", nodeid)
|
|
w.WriteHeader(http.StatusOK)
|
|
json.NewEncoder(w).Encode(apiNode)
|
|
}
|
|
|
|
// @Summary Assign a node to a gateway
|
|
// @Router /api/nodes/{network}/{nodeid}/gateway/assign [post]
|
|
// @Tags Gateways
|
|
// @Security oauth
|
|
// @Produce json
|
|
// @Param network path string true "Network ID"
|
|
// @Param nodeid path string true "Client node ID to assign to gateway"
|
|
// @Param gw_id query string true "Gateway node ID"
|
|
// @Param auto_assign_gw query bool false "Enable auto-assign gateway (Pro only)"
|
|
// @Success 200 {object} models.ApiNode
|
|
// @Failure 400 {object} models.ErrorResponse
|
|
// @Failure 500 {object} models.ErrorResponse
|
|
func assignGw(w http.ResponseWriter, r *http.Request) {
|
|
w.Header().Set("Content-Type", "application/json")
|
|
var params = mux.Vars(r)
|
|
nodeid := params["nodeid"]
|
|
netid := params["network"]
|
|
gwid := r.URL.Query().Get("gw_id")
|
|
autoAssignGw := r.URL.Query().Get("auto_assign_gw") == "true"
|
|
// Validate client node
|
|
node, err := logic.ValidateParams(nodeid, netid)
|
|
if err != nil {
|
|
logic.ReturnErrorResponse(w, r, logic.FormatError(err, "badrequest"))
|
|
return
|
|
}
|
|
if !servercfg.IsPro {
|
|
autoAssignGw = false
|
|
}
|
|
if autoAssignGw {
|
|
|
|
if node.RelayedBy != "" {
|
|
gatewayNode, err := logic.GetNodeByID(node.RelayedBy)
|
|
if err == nil {
|
|
// check if gw gateway Node has the relayed Node
|
|
if !slices.Contains(gatewayNode.RelayedNodes, node.ID.String()) {
|
|
gatewayNode.RelayedNodes = append(gatewayNode.RelayedNodes, node.ID.String())
|
|
}
|
|
newNodes := gatewayNode.RelayedNodes
|
|
newNodes = logic.RemoveAllFromSlice(newNodes, node.ID.String())
|
|
logic.UpdateRelayNodes(gatewayNode.ID.String(), gatewayNode.RelayedNodes, newNodes)
|
|
// Unassign client nodes (set their InternetGwID to empty)
|
|
if node.InternetGwID != "" {
|
|
node.InternetGwID = ""
|
|
gatewayNode.InetNodeReq.InetNodeClientIDs = logic.RemoveAllFromSlice(gatewayNode.InetNodeReq.InetNodeClientIDs, node.ID.String())
|
|
logic.UpsertNode(&gatewayNode)
|
|
}
|
|
} else {
|
|
node.RelayedBy = ""
|
|
node.InternetGwID = ""
|
|
}
|
|
node, _ = logic.GetNodeByID(node.ID.String())
|
|
}
|
|
node.AutoAssignGateway = true
|
|
logic.UpsertNode(&node)
|
|
logic.GetNodeStatus(&node, false)
|
|
go func() {
|
|
if node.FailedOverBy != uuid.Nil {
|
|
logic.ResetFailedOverPeer(&node)
|
|
}
|
|
if len(node.AutoRelayedPeers) > 0 {
|
|
logic.ResetAutoRelayedPeer(&node)
|
|
}
|
|
if err := mq.NodeUpdate(&node); err != nil {
|
|
slog.Error("error publishing node update to node", "node", node.ID, "error", err)
|
|
}
|
|
mq.PublishPeerUpdate(false)
|
|
}()
|
|
logic.ReturnSuccessResponseWithJson(w, r, node.ConvertToAPINode(), "auto assigned gateway")
|
|
return
|
|
}
|
|
if node.RelayedBy != "" {
|
|
logic.ReturnErrorResponse(w, r, logic.FormatError(errors.New("node is already using a gw"), "badrequest"))
|
|
return
|
|
}
|
|
// Validate gateway node
|
|
gatewayNode, err := logic.ValidateParams(gwid, netid)
|
|
if err != nil {
|
|
logic.ReturnErrorResponse(w, r, logic.FormatError(err, "badrequest"))
|
|
return
|
|
}
|
|
|
|
// Check if node is a gateway
|
|
if !gatewayNode.IsGw {
|
|
logic.ReturnErrorResponse(w, r, logic.FormatError(fmt.Errorf("node %s is not a gateway", nodeid), "badrequest"))
|
|
return
|
|
}
|
|
newNodes := []string{node.ID.String()}
|
|
newNodes = append(newNodes, gatewayNode.RelayedNodes...)
|
|
newNodes = logic.UniqueStrings(newNodes)
|
|
logic.UpdateRelayNodes(gatewayNode.ID.String(), gatewayNode.RelayedNodes, newNodes)
|
|
|
|
node, err = logic.GetNodeByID(node.ID.String())
|
|
if err != nil {
|
|
logic.ReturnErrorResponse(w, r, logic.FormatError(err, "internal"))
|
|
return
|
|
}
|
|
host := &schema.Host{
|
|
ID: node.HostID,
|
|
}
|
|
err = host.Get(r.Context())
|
|
if err != nil {
|
|
logic.ReturnErrorResponse(w, r, logic.FormatError(err, "badrequest"))
|
|
return
|
|
}
|
|
|
|
logger.Log(1, r.Header.Get("user"),
|
|
fmt.Sprintf("assigned nodes to gateway [%s] on network [%s]",
|
|
nodeid, netid))
|
|
|
|
logic.LogEvent(&models.Event{
|
|
Action: schema.GatewayAssign,
|
|
Source: models.Subject{
|
|
ID: r.Header.Get("user"),
|
|
Name: r.Header.Get("user"),
|
|
Type: schema.UserSub,
|
|
},
|
|
TriggeredBy: r.Header.Get("user"),
|
|
Target: models.Subject{
|
|
ID: node.ID.String(),
|
|
Name: host.Name,
|
|
Type: schema.GatewaySub,
|
|
},
|
|
Origin: schema.Dashboard,
|
|
})
|
|
|
|
logic.GetNodeStatus(&node, false)
|
|
apiNode := node.ConvertToAPINode()
|
|
|
|
go func() {
|
|
|
|
if node.FailedOverBy != uuid.Nil {
|
|
logic.ResetFailedOverPeer(&node)
|
|
}
|
|
if len(node.AutoRelayedPeers) > 0 {
|
|
logic.ResetAutoRelayedPeer(&node)
|
|
}
|
|
if err := mq.NodeUpdate(&node); err != nil {
|
|
slog.Error("error publishing node update to node", "node", node.ID, "error", err)
|
|
}
|
|
mq.PublishPeerUpdate(false)
|
|
}()
|
|
|
|
logic.ReturnSuccessResponseWithJson(w, r, apiNode, "assigned gateway")
|
|
}
|
|
|
|
// @Summary Unassign client nodes from a gateway
|
|
// @Router /api/nodes/{network}/{nodeid}/gateway/unassign [post]
|
|
// @Tags Gateways
|
|
// @Security oauth
|
|
// @Produce json
|
|
// @Param network path string true "Network ID"
|
|
// @Param nodeid path string true "Node ID"
|
|
// @Success 200 {object} models.ApiNode
|
|
// @Failure 400 {object} models.ErrorResponse
|
|
// @Failure 500 {object} models.ErrorResponse
|
|
func unassignGw(w http.ResponseWriter, r *http.Request) {
|
|
w.Header().Set("Content-Type", "application/json")
|
|
var params = mux.Vars(r)
|
|
nodeid := params["nodeid"]
|
|
netid := params["network"]
|
|
// Validate gateway node
|
|
node, err := logic.ValidateParams(nodeid, netid)
|
|
if err != nil {
|
|
logic.ReturnErrorResponse(w, r, logic.FormatError(err, "badrequest"))
|
|
return
|
|
}
|
|
host := &schema.Host{
|
|
ID: node.HostID,
|
|
}
|
|
err = host.Get(r.Context())
|
|
if err != nil {
|
|
logic.ReturnErrorResponse(w, r, logic.FormatError(err, "badrequest"))
|
|
return
|
|
}
|
|
gwid := node.RelayedBy
|
|
if node.AutoAssignGateway && gwid == "" {
|
|
node.AutoAssignGateway = false
|
|
logic.UpsertNode(&node)
|
|
go func() {
|
|
if err := mq.NodeUpdate(&node); err != nil {
|
|
slog.Error("error publishing node update to node", "node", node.ID, "error", err)
|
|
}
|
|
mq.PublishPeerUpdate(false)
|
|
}()
|
|
logic.ReturnSuccessResponseWithJson(w, r, node.ConvertToAPINode(), "unassigned gateway")
|
|
return
|
|
}
|
|
|
|
// Validate gateway node
|
|
gatewayNode, err := logic.ValidateParams(gwid, netid)
|
|
if err != nil {
|
|
logic.ReturnErrorResponse(w, r, logic.FormatError(err, "badrequest"))
|
|
return
|
|
}
|
|
|
|
// Unassign client nodes (set their InternetGwID to empty)
|
|
hadInternetGwID := node.InternetGwID != ""
|
|
if hadInternetGwID {
|
|
node.InternetGwID = ""
|
|
}
|
|
node.AutoAssignGateway = false
|
|
// Fetch gateway node fresh to avoid race conditions with concurrent updates
|
|
freshGatewayNode, err := logic.GetNodeByID(gatewayNode.ID.String())
|
|
if err != nil {
|
|
logic.ReturnErrorResponse(w, r, logic.FormatError(err, "internal"))
|
|
return
|
|
}
|
|
// Update InetNodeReq on fresh gateway node if needed
|
|
if hadInternetGwID {
|
|
freshGatewayNode.InetNodeReq.InetNodeClientIDs = logic.RemoveAllFromSlice(freshGatewayNode.InetNodeReq.InetNodeClientIDs, node.ID.String())
|
|
}
|
|
// Unset Relayed node - ensure node is properly removed from gateway's RelayedNodes
|
|
// Use fresh gateway node's RelayedNodes to avoid race conditions
|
|
oldNodes := freshGatewayNode.RelayedNodes
|
|
if !slices.Contains(oldNodes, node.ID.String()) {
|
|
// If node is not in the list but has RelayedBy set, add it temporarily to oldNodes
|
|
// so UpdateRelayNodes can properly unset it
|
|
oldNodes = append(oldNodes, node.ID.String())
|
|
}
|
|
newNodes := logic.RemoveAllFromSlice(oldNodes, node.ID.String())
|
|
// Now update RelayedNodes - this will fetch and save the gateway node with updated RelayedNodes
|
|
logic.UpdateRelayNodes(gatewayNode.ID.String(), oldNodes, newNodes)
|
|
// Update InetNodeReq after UpdateRelayNodes to ensure it's not overwritten by stale data
|
|
if hadInternetGwID {
|
|
finalGatewayNode, err := logic.GetNodeByID(gatewayNode.ID.String())
|
|
if err == nil {
|
|
finalGatewayNode.InetNodeReq.InetNodeClientIDs = logic.RemoveAllFromSlice(finalGatewayNode.InetNodeReq.InetNodeClientIDs, node.ID.String())
|
|
logic.UpsertNode(&finalGatewayNode)
|
|
}
|
|
}
|
|
// Explicitly clear RelayedBy and IsRelayed to ensure complete unassignment
|
|
node.RelayedBy = ""
|
|
node.IsRelayed = false
|
|
logic.UpsertNode(&node)
|
|
node, _ = logic.GetNodeByID(node.ID.String())
|
|
logger.Log(1, r.Header.Get("user"),
|
|
fmt.Sprintf("unassigned client nodes from gateway [%s] on network [%s]",
|
|
nodeid, netid))
|
|
|
|
logic.LogEvent(&models.Event{
|
|
Action: schema.GatewayUnAssign,
|
|
Source: models.Subject{
|
|
ID: r.Header.Get("user"),
|
|
Name: r.Header.Get("user"),
|
|
Type: schema.UserSub,
|
|
},
|
|
TriggeredBy: r.Header.Get("user"),
|
|
Target: models.Subject{
|
|
ID: node.ID.String(),
|
|
Name: host.Name,
|
|
Type: schema.GatewaySub,
|
|
},
|
|
Origin: schema.Dashboard,
|
|
})
|
|
|
|
logic.GetNodeStatus(&node, false)
|
|
|
|
go func() {
|
|
if err := mq.NodeUpdate(&node); err != nil {
|
|
slog.Error("error publishing node update to node", "node", node.ID, "error", err)
|
|
}
|
|
mq.PublishPeerUpdate(false)
|
|
}()
|
|
logic.ReturnSuccessResponseWithJson(w, r, node.ConvertToAPINode(), "unassigned gateway")
|
|
}
|