Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

Add link nodes task #1241

Merged
merged 1 commit into from
Jun 20, 2023
Merged
Show file tree
Hide file tree
Changes from all commits
Commits
File filter

Filter by extension

Filter by extension

Conversations
Failed to load comments.
Loading
Jump to
Jump to file
Failed to load files.
Loading
Diff view
Diff view
112 changes: 49 additions & 63 deletions deepfence_server/ingesters/agent.go
Original file line number Diff line number Diff line change
Expand Up @@ -664,18 +664,6 @@ func (nc *neo4jIngester) PushToDBSeq(batches ReportIngestionData, session neo4j.
}
defer tx.Close()

if _, err := tx.Run(`
UNWIND $batch as row
MERGE (n:Node{node_id:row.node_id})
MERGE (cp:CloudProvider{node_id:row.cloud_provider})
MERGE (cr:CloudRegion{node_id:row.cloud_region})
MERGE (cp) -[:HOSTS]-> (cr)
MERGE (cr) -[:HOSTS]-> (n)
SET cp.active = true, cp.pseudo = false, cr.active = true`,
map[string]interface{}{"batch": batches.Host_batch}); err != nil {
return err
}

if _, err := tx.Run(`
UNWIND $batch as row
MERGE (n:ContainerImage{node_id:row.node_id})
Expand All @@ -687,24 +675,6 @@ func (nc *neo4jIngester) PushToDBSeq(batches ReportIngestionData, session neo4j.
return err
}

if _, err := tx.Run(`
UNWIND $batch as row
MERGE (n:Pod{node_id:row.node_id})
SET n+= row, n.updated_at = TIMESTAMP(), n.active = true`,
map[string]interface{}{"batch": batches.Pod_batch}); err != nil {
return err
}

if _, err := tx.Run(`
UNWIND $batch as row
MERGE (n:KubernetesCluster{node_id:row.node_id})
MERGE (cp:CloudProvider{node_id:row.cloud_provider})
MERGE (cp) -[:HOSTS]-> (n)
SET n+= row, n.updated_at = TIMESTAMP(), n.active = true, n.node_type = 'cluster', cp.active = true, cp.pseudo = false`,
map[string]interface{}{"batch": batches.Kubernetes_cluster_batch}); err != nil {
return err
}

if _, err := tx.Run(`
UNWIND $batch as row
MATCH (n:Node{node_id: row.source})
Expand All @@ -716,36 +686,15 @@ func (nc *neo4jIngester) PushToDBSeq(batches ReportIngestionData, session neo4j.
return err
}

if _, err := tx.Run(`
UNWIND $batch as row
MATCH (n:KubernetesCluster{node_id: row.source})
WITH n, row
UNWIND row.destinations as dest
MATCH (m:Node{node_id: dest})
MERGE (n)-[:INSTANCIATE]->(m)`,
map[string]interface{}{"batch": batches.Kubernetes_cluster_edge_batch}); err != nil {
return err
}

if _, err := tx.Run(`
UNWIND $batch as row
MATCH (n:KubernetesCluster{node_id: row.source})
WITH n, row
UNWIND row.destinations as dest
MATCH (m:Pod{node_id: dest})
MERGE (n)-[:HOSTS]->(m)`,
map[string]interface{}{"batch": batches.Pod_edges_batch}); err != nil {
return err
}

if _, err := tx.Run(`
UNWIND $batch as row
MATCH (n:Node{node_id: row.source})
WITH n, row
UNWIND row.destinations as dest
MERGE (m:Pod{node_id: dest})
MERGE (n)-[:HOSTS]->(m)`,
map[string]interface{}{"batch": batches.Pod_host_edges_batch}); err != nil {
MATCH (m:Node{node_id: row.destination})
MERGE (n)-[r:CONNECTS]->(m)
WITH n, r, m, row.pids as rpids
UNWIND rpids as pids
SET r.left_pids = coalesce(r.left_pids, []) + pids.left,
r.right_pids = coalesce(r.right_pids, []) + pids.right`, map[string]interface{}{"batch": batches.Endpoint_edges_batch}); err != nil {
return err
}

Expand Down Expand Up @@ -831,15 +780,52 @@ func (nc *neo4jIngester) PushToDB(batches ReportIngestionData, session neo4j.Ses
return err
}

if _, err := tx.Run(`
UNWIND $batch as row
MERGE (n:KubernetesCluster{node_id:row.node_id})
SET n+= row, n.updated_at = TIMESTAMP(), n.active = true, n.node_type = 'cluster'`,
map[string]interface{}{"batch": batches.Kubernetes_cluster_batch}); err != nil {
return err
}

if _, err := tx.Run(`
UNWIND $batch as row
MERGE (n:Pod{node_id:row.node_id})
SET n+= row, n.updated_at = TIMESTAMP(), n.active = true`,
map[string]interface{}{"batch": batches.Pod_batch}); err != nil {
return err
}

if _, err := tx.Run(`
UNWIND $batch as row
MATCH (n:KubernetesCluster{node_id: row.source})
WITH n, row
UNWIND row.destinations as dest
MATCH (m:Pod{node_id: dest})
MERGE (n)-[:HOSTS]->(m)`,
map[string]interface{}{"batch": batches.Pod_edges_batch}); err != nil {
return err
}

if _, err := tx.Run(`
UNWIND $batch as row
MATCH (n:Node{node_id: row.source})
MATCH (m:Node{node_id: row.destination})
MERGE (n)-[r:CONNECTS]->(m)
WITH n, r, m, row.pids as rpids
UNWIND rpids as pids
SET r.left_pids = coalesce(r.left_pids, []) + pids.left,
r.right_pids = coalesce(r.right_pids, []) + pids.right`, map[string]interface{}{"batch": batches.Endpoint_edges_batch}); err != nil {
WITH n, row
UNWIND row.destinations as dest
MERGE (m:Pod{node_id: dest})
MERGE (n)-[:HOSTS]->(m)`,
map[string]interface{}{"batch": batches.Pod_host_edges_batch}); err != nil {
return err
}

if _, err := tx.Run(`
UNWIND $batch as row
MATCH (n:KubernetesCluster{node_id: row.source})
WITH n, row
UNWIND row.destinations as dest
MATCH (m:Node{node_id: dest})
MERGE (n)-[:INSTANCIATE]->(m)`,
map[string]interface{}{"batch": batches.Kubernetes_cluster_edge_batch}); err != nil {
return err
}

Expand Down
56 changes: 55 additions & 1 deletion deepfence_worker/cronjobs/neo4j.go
Original file line number Diff line number Diff line change
Expand Up @@ -100,7 +100,7 @@ func CleanUpDB(msg *message.Message) error {
if _, err = session.Run(`
MATCH (n:Node)
WHERE n.updated_at < TIMESTAMP()-$time_ms
AND NOT n.node_id IN ["in-the-internet", "out-the-internet"]
AND NOT n.node_id IN ["in-the-internet", "out-the-internet"]
AND n.agent_running=true
AND n.active = true
WITH n LIMIT 10000
Expand Down Expand Up @@ -475,6 +475,60 @@ func LinkCloudResources(msg *message.Message) error {
return nil
}

var linkNodesRunning = atomic.Bool{}

func LinkNodes(msg *message.Message) error {

if linkNodesRunning.Swap(true) {
return nil
}
defer linkNodesRunning.Store(false)

log.Info().Msgf("Link Nodes Starting")
defer log.Info().Msgf("Link Nodes Done")
namespace := msg.Metadata.Get(directory.NamespaceKey)
ctx := directory.NewContextWithNameSpace(directory.NamespaceID(namespace))

nc, err := directory.Neo4jClient(ctx)
if err != nil {
return err
}

session := nc.NewSession(neo4j.SessionConfig{AccessMode: neo4j.AccessModeWrite})
defer session.Close()

txConfig := neo4j.WithTxTimeout(30 * time.Second)

start := time.Now()

if _, err = session.Run(`
MATCH (n:Node)
WHERE not (n) <-[:HOSTS]- (:CloudRegion)
WITH n LIMIT 50000
MERGE (cp:CloudProvider{node_id: n.cloud_provider})
MERGE (cr:CloudRegion{node_id: n.cloud_region})
MERGE (cp) -[:HOSTS]-> (cr)
MERGE (cr) -[:HOSTS]-> (n)
SET cp.active = true, cr.active = true, cp.pseudo = false`,
map[string]interface{}{}, txConfig); err != nil {
return err
}

if _, err := session.Run(`
MATCH (n:KubernetesCluster)
WHERE not (n) <-[:HOSTS]- (:CloudProvider)
MERGE (cp:CloudProvider{node_id:n.cloud_provider})
MERGE (cp) -[:HOSTS]-> (n)
SET cp.active = true, cp.pseudo = false`,
map[string]interface{}{}); err != nil {
return err
}

log.Debug().Msgf("Link Nodes task took: %v", time.Since(start))

return nil
}

func RetryScansDB(msg *message.Message) error {
log.Info().Msgf("Retry scan DB Starting")
defer log.Info().Msgf("Retry scan DB Done")
Expand Down
4 changes: 4 additions & 0 deletions deepfence_worker/cronscheduler/scheduler.go
Original file line number Diff line number Diff line change
Expand Up @@ -175,6 +175,10 @@ func (s *Scheduler) addJobs() error {
if err != nil {
return err
}
_, err = s.cron.AddFunc("@every 30s", s.enqueueTask(sdkUtils.LinkNodesTask))
if err != nil {
return err
}
return nil
}

Expand Down
10 changes: 1 addition & 9 deletions deepfence_worker/ingesters/cloud_resource.go
Original file line number Diff line number Diff line change
Expand Up @@ -148,16 +148,8 @@ func CommitFuncCloudResource(ns string, cs []CloudResource) error {
if len(hosts) > 0 {
_, err = tx.Run(`
UNWIND $batch as row
WITH row
OPTIONAL MATCH (n:Node{node_id:row.node_id})
WITH n, row as row
WHERE n IS NULL or n.active=false
MERGE (m:Node{node_id:row.node_id})
MERGE (cp:CloudProvider{node_id:row.cloud_provider})
MERGE (cr:CloudRegion{node_id:row.cloud_region})
MERGE (cp) -[:HOSTS]-> (cr)
MERGE (cr) -[:HOSTS]-> (m)
SET m+=row, m.updated_at = TIMESTAMP(), cp.active = true, cp.pseudo = false, cr.active = true`,
SET m+=row, m.updated_at = TIMESTAMP()`,
map[string]interface{}{"batch": hosts},
)
}
Expand Down
2 changes: 2 additions & 0 deletions deepfence_worker/worker.go
Original file line number Diff line number Diff line change
Expand Up @@ -215,6 +215,8 @@ func startWorker(wml watermill.LoggerAdapter, cfg config) error {

worker.AddNoPublisherHandler(utils.LinkCloudResourceTask, cronjobs.LinkCloudResources)

worker.AddNoPublisherHandler(utils.LinkNodesTask, cronjobs.LinkNodes)

log.Info().Msg("Starting the consumer")
if err = worker.Run(context.Background()); err != nil {
cancel()
Expand Down
2 changes: 1 addition & 1 deletion golang_deepfence_sdk