Documentation ¶
Index ¶
- func SetLogLevel(s string)
- type Crawler
- func (c *Crawler) AddSeeds(seeds []string, force ...bool) (err error)
- func (c *Crawler) Crawl() (err error)
- func (c *Crawler) Dump() (allKeys []string, err error)
- func (c *Crawler) DumpMap() (m map[string]string, err error)
- func (c *Crawler) Flush() (err error)
- func (c *Crawler) Init(config ...Settings) (err error)
- func (c *Crawler) Redo() (err error)
- type Settings
Constants ¶
This section is empty.
Variables ¶
This section is empty.
Functions ¶
func SetLogLevel ¶
func SetLogLevel(s string)
Types ¶
type Crawler ¶
type Crawler struct { // Instance options RedisURL string RedisPort string MaxNumberConnections int MaxNumberWorkers int MaximumNumberOfErrors int TimeIntervalToPrintStats int Debug bool Info bool UseProxy bool UserAgent string Cookie string EraseDB bool MaxQueueSize int // Public options Settings Settings // contains filtered or unexported fields }
Crawler is the crawler instance
func (*Crawler) Crawl ¶
Crawl initiates the pool of connections and begins scraping URLs according to the todo list
Click to show internal directories.
Click to hide internal directories.