scrapy.utils.iterators.csviter

Here are the examples of the python api scrapy.utils.iterators.csviter taken from open source projects. By voting up you can indicate which examples are most useful and appropriate.

1 Examples 7

3 Source : feed.py
with MIT License
from autofelix

    def parse_rows(self, response):
        """Receives a response and a dict (representing each row) with a key for
        each provided (or detected) header of the CSV file.  This spider also
        gives the opportunity to override adapt_response and
        process_results methods for pre and post-processing purposes.
        """

        for row in csviter(response, self.delimiter, self.headers, self.quotechar):
            ret = iterate_spider_output(self.parse_row(response, row))
            for result_item in self.process_results(response, ret):
                yield result_item

    def parse(self, response):