Scrape Airbnb Data - Cloud Based Scraping

Wednesday, September 28, 2016 7:40 AM

(Download my extraction task of this tutorial HERE just in case you need it.)


Octoparse's cloud servers enable you to speed up your scraping job when the amount of data you plan to collect is very huge.

First you need to configure extraction rule for scraping. 

In this tutorial, I will take Airbnb for example to show you how to scrape data with Cloud Extraction(Cloud based scraping)


(Note: Cloud Extraction is not available for the Free Edition. For more information about different editions, you could click HERE.)


Step 1. Choose “Advanced Mode” ➜ Click “Start”➜ Complete basic information. ➜ Click “Next”.


Step 2. Enter the target URL of Airbnb in the built-in browser. ➜ Click “Go” icon to open the webpage.

( URL of the example: https://www.airbnb.com/s/New-York--NY?source=ds&s_tag=f57NxPuT )


Step 3. Click the pagination link. Click “Expand the selection area” until “Loop click in the element” appears. ➜ Choose “Loop click in the element” to turn the page.

(Note: If you want to extract information from every page of search result, you need to add a page navigation action.)


Step 4. Move your cursor over the section with similar layout, where you would extract data.

Click the first highlighted link ➜ Create a list of sections with similar layout. Click “Create a list of items” (sections with similar layout). ➜ “Add current item to the list”.

Then the first highlighted link has been added to the list.  ➜ Click “Continue to edit the list”.

Click the second highlighted link ➜ Click “Add current item to the list” again. Now we get all the links with similar layout.

Then click “Finish Creating List” ➜ Click “loop” to process the list for extracting the elements in each page.


Step 5. Extract the title of the first section ➜ Click the title ➜ Select “Extract text”. Other contents can be extracted in the same way.


Step 6. All the content will be selected in Data Fields. ➜ Click the “Field Name” to modify.


Step 7. Drag the second “Loop Item” before “Click to paginate” action in the Workflow Designer so that we can grab all the elements of sections from multiple pages.


Step 8. Click “Next” ➜ Click “Next” ➜ Click “Cloud Extraction”. You could see your task status on the task list.


Step 9. You could also schedule your cloud extraction settings to meet your special need. For example, to extract the data once a week on Monday, you should click “Weekly”, “Monday”, “0:00” ➜ Click “Start”. This is extremely important before Cloud Extraction automatically extracts all the selected data on your chosen time.


Step 10. You could also click “Cloud Extraction” to run the task immediately without having to wait for the setting time.


Step 11. The data extracted will be shown in "Data Extracted" pane. Click “View Data” button to view data. You then could export the results to Excel file, databases or other formats and save the file to your computer.





Author: The Octoparse Team




Download Octoparse Today



For more information about Octoparse, please click here.

Sign up today!



Author's Picks


Octoparse Smart Mode -- Get Data in Seconds

Get Started with Octoparse in 2 Minutes

Collect Data from LinkedIn

Collect Data from Amazon

Collect Data from eBay

Top 30 Free Web Scraping Software

30 Free Web Scraping Software

Collect Data from Amazon

Top 30 Free Web Scraping Software

- See more at: http://www.octoparse.com/tutorial/pagination-scrape-data-from-websites-with-query-strings-2/#sthash.gDCJJmOQ.dpuf
We use cookies to enhance your browsing experience. Read about how we use cookies and how you can control them by clicking cookie settings. If you continue to use this site, you consent to our use of cookies.
Accept decline