scrapy instantiation

start

from scrapy.cmdline import execute
execute(['scrapy', 'crawl', 'jokespider'])

  

items.py

import scrapy

class JokejiItem(scrapy.Item):
    title=scrapy.Field()
    url=scrapy.Field()

class ListItem(scrapy.Item):
    title=scrapy.Field()
    url=scrapy.Field()

  

spider.py

from scrapy.linkextractors import LinkExtractor
from scrapy.spiders import CrawlSpider, Rule
from jokeji.items import JokejiItem,ListItem

class JokespiderSpider(CrawlSpider):
    name = 'jokespider'
    allowed_domains = ['zizi.cn']
    start_urls = ['http://www.zizi.cn']

    rules = [
        Rule(LinkExtractor(allow=r'/list\w+.htm'), callback='parse_list', follow=True),
        Rule(LinkExtractor(allow=r'/jokehtml/\w+/\d+\.htm',deny=(r'/list')), callback='parse_item', follow=True),
    ]

    def parse_item(self, response):
        item=JokejiItem()
        item['title']='from content'
        return item

    def parse_list(self,response):
        item=ListItem()
        item['url']="from list........"+response.url
        return item

  

pipelines.py

class JokejiPipeline(object):
    def process_item(self, item, spider):
        print(item)

  

猜你喜欢

转载自www.cnblogs.com/pythonClub/p/9841509.html