You don’t need to know coding to scrape LinkedIn jobs at scale!!
If you’ve found your way to this article, chances are you’re not a developer but eager to learn how to scrape a website and seamlessly store its JSON data into Airtable.
What is Airtable – It’s a digital platform that acts like a supercharged spreadsheet, making it easy to organize and store data.
Process of saving the data
Once you sign up on Airtable you will see this on your screen.
You have to click on the “Start from scratch” button. Then a screen will appear where Airtable will ask you to name the project. You can use any name here. I am using Demo-Project.
After that, you can click “Skip setup” and then on Confirm.
This will directly take you to your project page.
Just click on Add extension and let it launch.
After clicking on Add extension you will be redirected to your project page on Airtable. Now, click on Add extension again.
Now, you will be asked to create an account on Data Fetcher.
Once you sign up this box will open on your screen
Click on “Create your first request” and proceed ahead.
Then click on Custom under the Application tab. It will open a box that might appear similar to POSTMAN or Insomnia.
Now, we can start scraping. I would request you go through the documentation of LinkedIn Jobs API. This will give you an idea about the API. How it works and what data it needs to return the response. Below I am attaching the image from the documentation, wherein you can see what inputs you need to give and what they are.
For this tutorial, we are going to focus on this API.
This API will return a list of jobs based on the parameters you pass. Now, let’s say I want to find jobs in Python for geoid 101473624(you can find it on LinkedIn URL) and I need all the data on the first page. Then in this case the API URL will look like this:
Once you have this link, you need to place it in the Data Fetcher Box. (please refer image below)
But before that, in “http://api.scrapingdog.com/linkedinjobs?page=1&geoid=101473624&field=python&api_key=Your-API-Key”, see the last part “Your-API-Key”, here you have to paste your own API key.
You can find your API key on the dashboard of Scrapingdog. (please refer image)
Now, paste the API link inside the Data Fetcher box.
Once done click on Save and Run from the bottom right. After that a tab will appear with the name “Response field Mapping”, you need to click Save & Run again here. (please refer screenshot below)
Data Fetcher will request data from scrapingdog’s API
The JSON result from Scrapingdog API has been saved inside the table. You can find this data inside your Airtable project too.
If you want to download this data in a CSV file then click on “Grid view” on the top left and then click “download CSV”.
This will download all of your results in CSV format.
In this journey, we’ve ventured into the world of web scraping and data automation without the need for coding skills. By harnessing the power of Scrapingdog’s LinkedIn Jobs API, we’ve unlocked a treasure trove of JSON data from LinkedIn, and then seamlessly funneled it into Airtable for organized storage and analysis. What seemed like a complex task has now become accessible to non-coders, empowering them to gather valuable insights, monitor job trends, or curate data for their projects.
Of course, you can use any API to store data inside Airtable using the Data Fetcher extension. Data Fetcher works really great by creating a bridge between the data collection process and non-coders.
If you want to learn more about this extension then you should refer to their documentation. For more such tutorials keep visiting our Blog section. We will be releasing more such content so keep your heads up.