Java,

Serverless,

Web scraping,

(6 min) ● Kevin Sahin, ● 04 September, 2019

Serverless is a term referring to the execution of code inside ephemeral containers (Function As A Service, or FaaS). It is a hot topic in 2019, after the “micro-service” hype, here come the “nano-services”!

Cloud functions can be triggered by different things such as:

An HTTP call to a REST API

A job in a message queue

A log

IOT event

Cloud functions are a really good fit with web scraping tasks for many reasons. Web Scraping is I/O bound, most of the time is spent waiting for HTTP responses, so we don’t need high-end CPU servers. Cloud functions are cheap (first 1M request is free, then $0.20 per million requests) and easy to set up. Cloud functions are a good fit for parallel scraping, we can create hundreds or thousands of function at the same time for large-scale scraping.

In this introduction, we are going to see how to deploy a slightly modified version of the Craigslist scraper we made on a previous blogpost on AWS Lambda using the serverless framework.

Prerequisites

We are going to use the Serverless framework to build and deploy our project to AWS lambda. Serverless CLI is able to generate lots of boilerplate code in different languages and deploy the code to different cloud providers, like AWS, Google Cloud or Azure.

An AWS account

Node and npm

Serverless CLI and Setup your AWS credentials

Java 8

Maven

Architecture

We will build an API using API Gateway with a single endpoint /items/{query} binded on a lambda function that will respond to us with a JSON array with all items (on the first result page) for this query.

Here is a simple diagram for this architecture:

Create the Maven project

Serverless is able to generate projects in lots of different languages: Java, Python, NodeJS, Scala… We are going to use one of these templates to generate a maven project:

serverless create --template aws-java-maven --name items-api -p aws-java-scraper

You can now open this Maven project in your favorite IDE.

Configuration

The first thing to do is to change the serverless.yml config to implement an API gateway route and bind it to the handleRequest method in the Handler.java class.

service: craigslist-scraper-api provider: name: aws runtime: java8 timeout: 30 package: artifact: target/hello-dev.jar functions: getCraigsListItems: handler: com.serverless.Handler events: - http: path: /items/{searchQuery} method: get

I also added a timeout to 30 seconds. The default timeout with the serverless framework is 6 seconds. Since we're running Java code the Lambda cold start can take several seconds. And then we will make an HTTP request to Craigslist website, so 30 seconds seems good.

Tired of getting blocked while scraping the web? Our API handles headless browsers and rotates proxies for you. Try Now for free Documentation

Function code

Now we can modify the Handler.class. The function logic is easy. First, we retrieve the path parameter called “searchQuery”. Then we create a CraigsListScraper and call the scrape() method with this searchQuery. It will return a List<Item> representing all the items on the first Craigslist's result page.

We then use the ApiGatewayResponse class that was generated by the Serverless framework to return a JSON array containing every item.

You can find the rest of the code in this repository, with the CraigsListScraper and Item class.

@Override public ApiGatewayResponse handleRequest ( Map < String , Object > input , Context context ) { LOG . info ( "received: {}" , input ) ; try { Map < String , String > pathParameters = ( Map < String , String > ) input . get ( "pathParameters" ) ; String query = pathParameters . get ( "searchQuery" ) ; CraigsListScraper scraper = new CraigsListScraper ( ) ; List < Item > items = scraper . scrape ( query ) ; return ApiGatewayResponse . builder ( ) . setStatusCode ( 200 ) . setObjectBody ( items ) . setHeaders ( Collections . singletonMap ( "X-Powered-By" , "AWS Lambda & serverless" ) ) . build ( ) ; } catch ( Exception e ) { LOG . error ( "Error : " + e ) ; Response responseBody = new Response ( "Error while processing URL: " , input ) ; return ApiGatewayResponse . builder ( ) . setStatusCode ( 500 ) . setObjectBody ( responseBody ) . setHeaders ( Collections . singletonMap ( "X-Powered-By" , "AWS Lambda & Serverless" ) ) . build ( ) ; } }

We can now build the project:

mvn clean install

And deploy it to AWS:

serverless deploy Serverless: Packaging service... Serverless: Creating Stack... Serverless: Checking Stack create progress... ..... Serverless: Stack create finished... Serverless: Uploading CloudFormation file to S3... Serverless: Uploading artifacts... Serverless: Uploading service .zip file to S3 (13.35 MB)... Serverless: Validating template... Serverless: Updating Stack... Serverless: Checking Stack update progress... ................................. Serverless: Stack update finished... Service Information service: items-api stage: dev region: us-east-1 stack: items-api-dev api keys: None endpoints: GET - https://tmulioizdf.execute-api.us-east-1.amazonaws.com/dev/items/{searchQuery} functions: getCraigsListItems: items-api-dev-getCraigsListItems

You can then test your function using curl or your web browser with the URL given in the deployment logs ( serverless info will also show this information.)

Here is a query to look for “macBook pro” :

curl https://tmulioizdf.execute-api.us-east-1.amazonaws.com/dev/items/macBook%20pro | json_reformat 1 ↵ % Total % Received % Xferd Average Speed Time Time Time Current Dload Upload Total Spent Left Speed 100 19834 100 19834 0 0 7623 0 0:00:02 0:00:02 --:--:-- 7622 [ { "title" : "2010 15\" Macbook pro 3.06ghz 8gb 320gb osx maverick" , "price" : 325, "url" : "https://sfbay.craigslist.org/eby/sys/d/macbook-pro-306ghz-8gb-320gb/6680853189.html" } , { "title" : "Apple MacBook Pro A1502 13.3\" Late 2013 2.6GHz i5 8 GB 500GB + Extras" , "price" : 875, "url" : "https://sfbay.craigslist.org/pen/sys/d/apple-macbook-pro-alateghz-i5/6688755497.html" } , { "title" : "Apple MacBook Pro Charger USB-C (Latest Model) w/ Box - Like New!" , "price" : 50, "url" : "https://sfbay.craigslist.org/pen/sys/d/apple-macbook-pro-charger-usb/6686902986.html" } , { "title" : "MacBook Pro 13\" C2D 4GB memory 500GB HDD" , "price" : 250, "url" : "https://sfbay.craigslist.org/eby/sys/d/macbook-pro-13-c2d-4gb-memory/6688682499.html" } , { "title" : "Macbook Pro 2011 13\"" , "price" : 475, "url" : "https://sfbay.craigslist.org/eby/sys/d/macbook-pro/6675556875.html" } , { "title" : "Trackpad Touchpad Mouse with Cable and Screws for Apple MacBook Pro" , "price" : 39, "url" : "https://sfbay.craigslist.org/pen/sys/d/trackpad-touchpad-mouse-with/6682812027.html" } , { "title" : "Macbook Pro 13\" i5 very clean, excellent shape! 4GB RAM, 500GB HDD" , "price" : 359, "url" : "https://sfbay.craigslist.org/sfc/sys/d/macbook-pro-13-i5-very-clean/6686879047.html" } , ...

Note that the first invocation will be slow, it took 7 seconds for me. The next invocations will be much quicker.

Go further

This was just a little example, here are some ideas to improve this :

Better error handling

Protect the API with an API Key (really easy to implement with API Gateway)

Save the items to a DynamoDB database

Send the search query to an SQS queue, and trigger the lambda execution with the queue instead of an HTTP request

Send a notification with SNS if an Item is less than a certain price point.

If you like web scraping and are tired taking care of proxies, JS rendering and captchas, you can check our new web scraping API, the first 1000 API calls are on us.

This is the end of this tutorial.

I hope you enjoyed the post. Don't hesitate to experiment with Lambda and other cloud providers, it's really fun, easy, and can drastically reduce your infrastructure costs, especially for web-scraping or asynchronous related tasks.

If you are more into Python, have a look at our 👉 python web scraping tutorial