import requests import re import json import scrapy from scrapy.linkextractors import LinkExtractor from scrapy.spiders import CrawlSpider, Rule

class WhcSpider(CrawlSpider): name = ‘whc’ allowed_domains = [‘ygdy8.com’] start_urls = [‘http://ygdy8.com/’] ####### 遍历全站list,url rules = ( Rule(LinkExtractor(allow=r’index.html’,deny=‘game’),follow=True), Rule(LinkExtractor(allow=r’/list_\d+_\d+.html’), follow=True), Rule(LinkExtractor(allow=r’/\d+/\d+.html’), callback=‘parse_item’),

)

def parse_item(self, response):

# print(response.url)

#### 获取下载地址

ftp_url=re.findall('ftp',response.text)

print(ftp_url)

#### 写入本地文件

with open('result.txt','a',encoding='utf-8') as f:

f.write(json.dumps(ftp_url,ensure_ascii=False))

推荐链接

评论可见,请评论后查看内容,谢谢!!!
 您阅读本篇文章共花了: