Hello all,
In data scraping activity, I am pulling URL but it is pulling URL and Data of the first page only, very strangely it is happening with the field on which URL is checked otherwise it is pulling data perfectly. If I uncheck the URL of any field then it starts pulling data of all pages.
It is going to all the pages, Pulling all data rows but not the field in which URL is also checked. that too till UR is checked, as I uncheck URL, it starts pulling all rows.
here is a screen shot.
After choosing url we would have been shown with a preview of the table we are going to get
Did that have those url
And ensure once that this option is chosen as yes once after having the preview
I hope we are missing some part of step in data scrapping no worries let’s try to solve this
Cheers @Alok_Shrivastava
I am sharing all the screenshots. One set is a process with “Deiatls field checked URL” and another is “Address field checked URL”. The outcomes in excel as well.
Details was associated with URL, 2nd page is having blank field whereas Addresses are coming flawlessly.
2.Address was associated with URL, 2nd page is having blank field whereas Details are coming flawlessly.
One possibility would be that the selector for the table that you’re scraping on page 2 may not be the same as that what you modeled using Page 1.
Can you please do the following?
Navigate manually to Page 2 or Page 3, then run the Data Scraping wizard and compare the XML it generates with the one that you originally generated with Page 1.
It may lead to an answer.
I did post a response to another forum on an issue I resolved similar to yours. I’ll find it and post it here.
I am pretty sure you will notice some differences. If the differences are simple, you might just fix it with wild card characters. If not, you might need to perform separate extractions and merge them, whichever works best.
Hi @Alok_Shrivastava,
I am able to scrape the data from three page individually.Please check attached file. Data2.zip (17.3 KB)
I have followed the below steps:
open the website and enter the required input to search data for B1 commercial
Yes, if we load page individually and run the code it will work. But it should work automatically for all the pages. No matter there are 3 pages or 100 pages, It should crawl through all the pages and capture all the data. It is what I need. There are 130+ postcodes. It is not possible to load each page of every postcode and capture the data. I do not want to do this repetitive job that is why the robot is required.
kindly try to pull the data and URL in a single process. Where robot should press the Next button and capture all the data with URL till the last page.