Files
crawler_v2/cmd/crawler/main.go
2025-06-09 16:36:27 +02:00

164 lines
3.7 KiB
Go

package main
import (
"context"
"fmt"
"github/pippellia-btc/crawler/pkg/config"
"github/pippellia-btc/crawler/pkg/graph"
"github/pippellia-btc/crawler/pkg/pipe"
"github/pippellia-btc/crawler/pkg/redb"
"github/pippellia-btc/crawler/pkg/walks"
"log"
"os"
"os/signal"
"runtime"
"sync"
"syscall"
"time"
"github.com/nbd-wtf/go-nostr"
"github.com/redis/go-redis/v9"
"github.com/vertex-lab/relay/pkg/eventstore"
)
func main() {
ctx, cancel := context.WithCancel(context.Background())
defer cancel()
go handleSignals(cancel)
config, err := config.Load()
if err != nil {
panic(err)
}
events := make(chan *nostr.Event, config.EventsCapacity)
pubkeys := make(chan string, config.PubkeysCapacity)
store, err := eventstore.New(config.SQLiteURL)
if err != nil {
panic(err)
}
db := redb.New(&redis.Options{Addr: config.RedisAddress})
count, err := db.NodeCount(ctx)
if err != nil {
panic(err)
}
if count == 0 {
if len(config.InitPubkeys) == 0 {
panic("init pubkeys are empty: impossible to initialize")
}
log.Println("initialize from empty database...")
nodes := make([]graph.ID, len(config.InitPubkeys))
for i, pk := range config.InitPubkeys {
nodes[i], err = db.AddNode(ctx, pk)
if err != nil {
panic(err)
}
pubkeys <- pk // add to queue
}
walks, err := walks.Generate(ctx, db, nodes...)
if err != nil {
panic(err)
}
if err := db.AddWalks(ctx, walks...); err != nil {
panic(err)
}
log.Printf("correctly added %d init pubkeys", len(config.InitPubkeys))
}
var producers sync.WaitGroup
var consumers sync.WaitGroup
producers.Add(3)
go func() {
defer producers.Done()
pipe.Firehose(ctx, config.Firehose, db, enqueue(events))
}()
go func() {
defer producers.Done()
pipe.Fetcher(ctx, config.Fetcher, pubkeys, enqueue(events))
}()
go func() {
defer producers.Done()
pipe.Arbiter(ctx, config.Arbiter, db, enqueue(pubkeys))
close(pubkeys) // Arbiter is the only pubkey sender
}()
consumers.Add(1)
go func() {
defer consumers.Done()
pipe.Engine(ctx, config.Engine, store, db, events)
}()
producers.Wait()
close(events)
consumers.Wait()
}
// handleSignals listens for OS signals and triggers context cancellation.
func handleSignals(cancel context.CancelFunc) {
signals := make(chan os.Signal, 1)
signal.Notify(signals, syscall.SIGINT, syscall.SIGTERM)
<-signals
log.Println(" Signal received. Shutting down...")
cancel()
}
// enqueue things into the specified channel or return an error if full.
func enqueue[T any](queue chan T) func(t T) error {
return func(t T) error {
select {
case queue <- t:
return nil
default:
return fmt.Errorf("channel is full, dropping %v", t)
}
}
}
func printStats(ctx context.Context, events chan *nostr.Event, pubkeys chan string) {
filename := "stats.log"
file, err := os.OpenFile(filename, os.O_CREATE|os.O_WRONLY|os.O_APPEND, 0644)
if err != nil {
panic(fmt.Errorf("failed to open log file %s: %w", filename, err))
}
defer file.Close()
log := log.New(file, "stats: ", log.LstdFlags)
ticker := time.NewTicker(10 * time.Second)
defer ticker.Stop()
for {
select {
case <-ctx.Done():
return
case <-ticker.C:
goroutines := runtime.NumGoroutine()
memStats := new(runtime.MemStats)
runtime.ReadMemStats(memStats)
log.Println("---------------------------------------")
log.Printf("events queue: %d/%d\n", len(events), cap(events))
log.Printf("pubkeys queue: %d/%d\n", len(pubkeys), cap(pubkeys))
log.Printf("walks tracker: %v\n", pipe.WalksTracker.Load())
log.Printf("goroutines: %d\n", goroutines)
log.Printf("memory usage: %.2f MB\n", float64(memStats.Alloc)/(1024*1024))
log.Println("---------------------------------------")
}
}
}