Warning: file_get_contents(/data/phpspider/zhask/data//catemap/2/python/349.json): failed to open stream: No such file or directory in /data/phpspider/zhask/libs/function.php on line 167

Warning: Invalid argument supplied for foreach() in /data/phpspider/zhask/libs/tag.function.php on line 1116

Notice: Undefined index: in /data/phpspider/zhask/libs/function.php on line 180

Warning: array_chunk() expects parameter 1 to be array, null given in /data/phpspider/zhask/libs/function.php on line 181

Warning: file_get_contents(/data/phpspider/zhask/data//catemap/3/xpath/2.json): failed to open stream: No such file or directory in /data/phpspider/zhask/libs/function.php on line 167

Warning: Invalid argument supplied for foreach() in /data/phpspider/zhask/libs/tag.function.php on line 1116

Notice: Undefined index: in /data/phpspider/zhask/libs/function.php on line 180

Warning: array_chunk() expects parameter 1 to be array, null given in /data/phpspider/zhask/libs/function.php on line 181
Python 如何从不同的URL获取xpath,由start\u requests方法返回_Python_Xpath_Scrapy_Web Crawler - Fatal编程技术网

Python 如何从不同的URL获取xpath,由start\u requests方法返回

Python 如何从不同的URL获取xpath,由start\u requests方法返回,python,xpath,scrapy,web-crawler,Python,Xpath,Scrapy,Web Crawler,这是我的密码: import scrapy from scrapy.spider import BaseSpider from scrapy.selector import Selector import MySQLdb class AmazonSpider(BaseSpider): name = "amazon" allowed_domains = ["amazon.com"] start_urls = [] def parse(self, respons

这是我的密码:

import scrapy
from scrapy.spider import BaseSpider
from scrapy.selector import Selector
import MySQLdb


class AmazonSpider(BaseSpider):
    name = "amazon"
    allowed_domains = ["amazon.com"]
    start_urls = []

    def parse(self, response):
        print self.start_urls

    def start_requests(self):
        conn = MySQLdb.connect(user='root',passwd='root',db='mydb',host='localhost')
        cursor = conn.cursor()
        cursor.execute(
            'SELECT url FROM products;'
            )
        rows = cursor.fetchall()
        for row in rows:
            yield self.make_requests_from_url(row[0])
        conn.close()
如何获取
start\u请求
函数返回的URL的xpath


注意:URL属于不同的域,不尽相同。

yield
使
start\u请求成为生成器。使用
for
循环获取从中返回的每个结果

像这样:

...
my_spider = AmazonSpider()
for my_url in my_spider.start_requests():
    print 'we get URL: %s' % str(my_url)
...