@@ -2,6 +2,7 @@ package crawl
22
33import (
44 "context"
5+ "fmt"
56 "math"
67 "time"
78
@@ -10,7 +11,6 @@ import (
1011 "github.com/libp2p/go-libp2p/core/network"
1112 "github.com/libp2p/go-libp2p/core/peer"
1213 rcmgr "github.com/libp2p/go-libp2p/p2p/host/resource-manager"
13- "github.com/pkg/errors"
1414 log "github.com/sirupsen/logrus"
1515
1616 "github.com/dennis-tra/nebula-crawler/pkg/config"
@@ -95,13 +95,13 @@ func NewScheduler(conf *config.Crawl, dbc db.Client) (*Scheduler, error) {
9595 limiter := rcmgr .NewFixedLimiter (rcmgr .InfiniteLimits )
9696 rm , err := rcmgr .NewResourceManager (limiter )
9797 if err != nil {
98- return nil , errors . Wrap ( err , "new resource manager" )
98+ return nil , fmt . Errorf ( "new resource manager: %w" , err )
9999 }
100100
101101 // Initialize a single libp2p node that's shared between all crawlers.
102102 h , err := libp2p .New (libp2p .NoListenAddrs , libp2p .ResourceManager (rm ), libp2p .UserAgent ("nebula-crawler/" + conf .Root .Version ()))
103103 if err != nil {
104- return nil , errors . Wrap ( err , "new libp2p host" )
104+ return nil , fmt . Errorf ( "new libp2p host: %w" , err )
105105 }
106106
107107 s := & Scheduler {
@@ -219,12 +219,12 @@ func (s *Scheduler) CrawlNetwork(ctx context.Context, bootstrap []peer.AddrInfo)
219219
220220 // Persist the crawl results
221221 if err := s .updateCrawl (context .Background (), crawlerCtx , len (s .inCrawlQueue ) == 0 ); err != nil {
222- return errors . Wrap ( err , "persist crawl" )
222+ return fmt . Errorf ( "persist crawl: %w" , err )
223223 }
224224
225225 // Persist associated crawl properties
226226 if err := s .persistCrawlProperties (context .Background ()); err != nil {
227- return errors . Wrap ( err , "persist crawl properties" )
227+ return fmt . Errorf ( "persist crawl properties: %w" , err )
228228 }
229229
230230 // persist all neighbor information
@@ -265,7 +265,7 @@ func (s *Scheduler) initCrawl(ctx context.Context) error {
265265 log .Infoln ("Initializing crawl..." )
266266 crawl , err := s .dbc .InitCrawl (ctx )
267267 if err != nil {
268- return errors . Wrap ( err , "creating crawl in db" )
268+ return fmt . Errorf ( "creating crawl in db: %w" , err )
269269 }
270270 s .crawl = crawl
271271 s .crawlStart = crawl .StartedAt
@@ -283,7 +283,7 @@ func (s *Scheduler) startCrawlers(ctx context.Context) ([]*Crawler, context.Cont
283283 c , err := NewCrawler (s .host , s .config )
284284 if err != nil {
285285 crawlerCancel ()
286- return nil , nil , nil , errors . Wrap ( err , "new crawler" )
286+ return nil , nil , nil , fmt . Errorf ( "new crawler: %w" , err )
287287 }
288288 crawlers = append (crawlers , c )
289289 go c .StartCrawling (crawlerCtx , s .crawlQueue , s .crawlResultsQueue )
@@ -306,7 +306,7 @@ func (s *Scheduler) startPersisters(ctx context.Context) ([]*Persister, context.
306306 p , err := NewPersister (s .dbc , s .config , s .crawl )
307307 if err != nil {
308308 persistersCancel ()
309- return nil , nil , errors . Wrap ( err , "new persister" )
309+ return nil , nil , fmt . Errorf ( "new persister: %w" , err )
310310 }
311311 persisters = append (persisters , p )
312312 go p .StartPersisting (persistersCtx , s .persistQueue , s .persistResultsQueue )
0 commit comments