Hi Iām trying to scrap from a list of urls however when I set the automation to run in the background it will not open up multiple windows. Iām following this guide almost exactly without any luck: https://www.youtube.com/watch?v=c9IDz8E_nek
Please advise the best way to troubleshoot the issue. My playbooks will not run to completion.
This is actually how the āScrape data in the backgroundā action is designed to behave as it allows you to still continue working in other various applications while it scrape behind the scenes.
If the issue is something different, please let us know. Asa friendly reminder, itās always helpful to share the automation you are working with to assist on our end to troubleshoot for you.
Iāve yet to get a successful output from this automation. I need to deep dive back into it today and see where I am hung up but perhaps it would be helpful to zoom later to go over the issue? I will follow up with this email with a few screenshots as well as soon as I get caught up on a few things.
Iām unable to run it on my end to further troubleshooting because you need to share the first google sheet called āD2 Full Listā publicly with edit rights.
Here is the D2 Full List with edit rights, let me know if you have any issues accessing it. Iāll make those updates based on your suggestions as well.
It appears to be working on my end - believe the issue was the name of the āTest Listā tab header for the URLs. Let me know if you have any questions.
One other thing Iām noticing, when I run the playbook for the first time it seems to be adding in this step to the playbook as Iām not initially setting this up. Is that correct behaviour of the system?
On Tue, Oct 31, 2023 at 12:05 PM Will Shoemaker will@popacta.com wrote:
Hey Jess,
Here is the D2 Full List with edit rights, let me know if you have any issues accessing it. Iāll make those updates based on your suggestions as well.
The Test Full List is returning no results still. I did set up another playbook today using the same sequence which will not run to completion. Link below. Perhaps I need to change my output to something besides a googlesheet?
Yes I am available to zoom any time today, I can send you a meeting invite if you let me know a time that works well for you. Thank you for all of your help with this!
Thats okay just let me know when you have some availability, Iām still working through some issues with the playbook now that I straightened out the version control issues with the google doc.
Iāve noticed the different URLs have slightly different formats, so Iām checking with the team on whether one scraper template can be used across the different sites and Iām also checking why you havenāt been having any luck getting some output.
Are you able to share a quick recording on what kind of error you get when running the playbook please?
Let me know what you find out about the url formats, they all come from the same hosting platform, Learfield so I hope that they are consistent enough to be able to scrape in the full bundle.
Here is a video of running the playbook. As you can see I run it and it begins to scrap the first url then gets hung up there, it never completes the playbook providing a .csv output. However, one time I was able to get it to scrape 3 sites from a google doc and spit the output into the same google doc, I was never able to recreate it though. Let me know if I can send you any more videos or if you would like to zoom today to go over the issues.
Checking in on this. I saw you or possibly Jess created a playbook that used scraping the full table then extracting only the text in email format as a possible work around? I havenāt had a chance to play with that approach yet but is this probably the best way to achieve what weāre trying to do here?
Could I confirm that youād only like to scrape the email address from these websites ? Iāve given it a test and a basic scraper template built off one of the websites wonāt work on the other websites since the table columns are not in the same order across the sites. Weāll need to make some tweaks to the CSS selectors to match the different ordering of the columns across the different sites.
With regards to the bug I was referring too, that is affecting background scraping and weāre rolling out a fix for it soon. For now, scraping in the active tab is still possible.
Yes just email is fine. If we could pull the name column as well that would be helpful but not necessary. Let me me know if you have any other questions.
First and foremost, I want to extend our sincerest apologies for all the inconvenience you may have experienced due to the recent bug in our software. We understand how frustrating such issues can be, and we appreciate your patience and understanding as we worked diligently to address it.
I am pleased to inform you that our development team has successfully identified and resolved the bug that was causing the issues you reported. The fix is now available with a new version of Bardeen (2.37.2) which addresses the issue an issue with background scraping in the previous version. You can check your current version of Bardeen in Google Chrome by following the steps in my screenshot below (the version number will appear next to the Bardeen extension) as you might still be on the old one:
Here is a quick guide on how to update your version of Bardeen: How to Update Bardeen
Thank you for your patience, and if you need assistance, email me directly lucy@bardeen.ai.
Customer Support - bardeen.ai
Explore | @bardeenai | Bardeen Community
Customer Support - bardeen.ai
Explore | @bardeenai | Bardeen Community