Documentation ¶
Index ¶
- Variables
- func PageScreenShot(url string, opts ...core.ConfigOpt) (string, error)
- func StartCrawler(url string, opts ...core.ConfigOpt) (chan core.ReqInfo, error)
- func StartCrawlerV2(url string, opts ...core.ConfigOpt) error
- type Crawler
- func (crawler *Crawler) GetChannel() chan core.ReqInfo
- func (crawler *Crawler) Monitor() error
- func (crawler *Crawler) PageScreenShot() (string, error)
- func (crawler *Crawler) SetBlackList(blackRegStr string)
- func (crawler *Crawler) SetChromeWS(wsAddress string)
- func (crawler *Crawler) SetConcurrent(concurrent int)
- func (crawler *Crawler) SetCookie(domain, k, v string)
- func (crawler *Crawler) SetCookies(domain string, value map[string]string)
- func (crawler *Crawler) SetDangerUrlCheck()
- func (crawler *Crawler) SetExtraHeaders(headers ...string)
- func (crawler *Crawler) SetFormFill(key, value string)
- func (crawler *Crawler) SetFullTimeout(timeout int)
- func (crawler *Crawler) SetHeader(key, value string)
- func (crawler *Crawler) SetHeaders(kv map[string]string)
- func (crawler *Crawler) SetMaxDepth(depth int)
- func (crawler *Crawler) SetMaxUrl(maxUrl int)
- func (crawler *Crawler) SetOnRequest(f func(core.ReqInfo))
- func (crawler *Crawler) SetProxy(proxyAddr string, proxyInfo ...string)
- func (crawler *Crawler) SetScanRange(scanRange int)
- func (crawler *Crawler) SetScanRepeatLevel(scanRepeat int)
- func (crawler *Crawler) SetTags(tagsPath string)
- func (crawler *Crawler) SetTimeout(timeout int)
- func (crawler *Crawler) SetUrlFromProxy(ifYes bool)
- func (crawler *Crawler) SetWhiteList(whiteRegStr string)
- func (crawler *Crawler) Start() error
- func (crawler *Crawler) StartV2() error
- func (crawler *Crawler) StartVRemote() error
Constants ¶
This section is empty.
Variables ¶
View Source
var CrawlerXExports = map[string]interface{}{ "StartCrawler": StartCrawler, "StartCrawlerV2": StartCrawlerV2, "PageScreenShot": PageScreenShot, "proxy": core.WithProxy, "maxUrl": core.WithMaxUrl, "whiteList": core.WithWhiteList, "blackList": core.WithBlackList, "timeout": core.WithTimeout, "maxDepth": core.WithMaxDepth, "formFill": core.WithFormFill, "header": core.WithHeader, "headers": core.WithHeaders, "concurrent": core.WithConcurrent, "cookie": core.WithCookie, "cookies": core.WithCookies, "scanRange": core.WithScanRange, "scanRepeat": core.WithScanRepeat, "checkDanger": core.WithCheckDanger, "tags": core.WithTags, "fullTimeout": core.WithFullCrawlerTimeout, "chromeWS": core.WithChromeWS, "remote": core.WithGetUrlRemote, "extraHeaders": core.WithExtraHeaders, "HighRepeatLevel": detect.HighLevel, "MediumRepeatLevel": detect.MediumLevel, "LowRepeatLevel": detect.LowLevel, "UnLimitRepeat": detect.UnLimit, "AllDomainScan": detect.AllDomain, "SubMenuScan": detect.SubMenu, "TargetUrlScan": detect.TargetUrl, }
Functions ¶
func StartCrawler ¶
Types ¶
type Crawler ¶
type Crawler struct {
// contains filtered or unexported fields
}
func CreateCrawler ¶
func (*Crawler) GetChannel ¶
func (*Crawler) PageScreenShot ¶
func (*Crawler) SetBlackList ¶
func (*Crawler) SetChromeWS ¶
func (*Crawler) SetConcurrent ¶
func (*Crawler) SetCookies ¶
func (*Crawler) SetDangerUrlCheck ¶
func (crawler *Crawler) SetDangerUrlCheck()
func (*Crawler) SetExtraHeaders ¶
func (*Crawler) SetFormFill ¶
func (*Crawler) SetFullTimeout ¶
func (*Crawler) SetHeaders ¶
func (*Crawler) SetMaxDepth ¶
func (*Crawler) SetOnRequest ¶
func (*Crawler) SetScanRange ¶
func (*Crawler) SetScanRepeatLevel ¶
func (*Crawler) SetTimeout ¶
func (*Crawler) SetUrlFromProxy ¶
func (*Crawler) SetWhiteList ¶
func (*Crawler) StartVRemote ¶
Directories ¶
Path | Synopsis |
---|---|
Package config https://github.com/unknwon/goconfig
|
Package config https://github.com/unknwon/goconfig |
Package newcrawlerx @Author bcy2007 2023/3/7 14:25
|
Package newcrawlerx @Author bcy2007 2023/3/7 14:25 |
cmd
Package cmd @Author bcy2007 2023/3/23 10:50
|
Package cmd @Author bcy2007 2023/3/23 10:50 |
Click to show internal directories.
Click to hide internal directories.