Email ✉️  & Phone ☎️ Extractor avatar

Email ✉️ & Phone ☎️ Extractor

Try for free

7 days trial then $30.00/month - No credit card required now

View all Actors
Email ✉️  & Phone ☎️ Extractor

Email ✉️ & Phone ☎️ Extractor

anchor/email-phone-extractor
Try for free

7 days trial then $30.00/month - No credit card required now

Extract emails, phone numbers, and other useful contact information (twitter, linkedIn...) from any list of websites you provide. Best tool for contact lead generation. Export data in structured formats and dominate your outreach game. Capture your leads almost for free, fast, and without limits.

ME

is it possible to exclude some website for extract?

Closed

Meser opened this issue
2 months ago

hey I have a list of websites for some of them I don't want to extract details can I ask to ignore some of the websites so it will skip them? you closed my issue but you didn't gave me the right answer - I want to write websites I want to exclude, not websites I want to include is it possible? thanks

anchor avatar

guillim (anchor)

2 months ago

There is a way to exclude some websites yes, using the pseudo urls with a negative regex. If you are not familiar with regex, then no there is no other way to do it, sorry.

You could still create two different task, one that can follow links for some websites, and the other Task with websites you don’t want any follow links.

Does it make sense?

Sorry if I closed too early the previous issue. You can reopen issues any time, it’s done for that.

Let me know if I can close this one once you are ok with my answers !

Have a good day, and enjoy the Olympics games :)

ME

Meser

2 months ago

Thank you very much for the detailed explanation! Can you give me an example of a situation where I ask not to scan a certain website? Let's say for example that I don't want to scan the following domain - https://en.batumiexpert.com/ and the following domain - https://hayatestate.com/

Can you give me an example that I can use so that I can understand better? Thanks!

anchor avatar

guillim (anchor)

2 months ago

Sure. Using pseudo-Urls you can write something that works I believe.

Here is a small website helping you to write the proper regex https://regexr.com. In your case, you want to "match" anything but some websites (https://en.batumiexpert.com/ and https://hayatestate.com/) which leads to something like this :

^(?!.(batumiexpert|hayatestate/)).$

Note that in this regex, I only look after "batumiexpert" and "hayatestate" without the .com to make it simpler. But feel free to play around with regexr to try new things and add other websites. To add oter websites "|" is equivalent to OR in coding languages.

Hope it helps,

Let me know if you have any questions, otherwise I will close the issue :)

anchor avatar

guillim (anchor)

a month ago

I see you have mulitple questions. Let's finish this issue "is it possible to exclude some website for extract?" first, if you may. From what I see on your results, it seems to work because I do not see any url like "https://en.batumiexpert.com" or "https://hayatestate.com" in the results.

You may think it's because the crawler does not click on links. To make sure of that, you may want to increase to "total number of pages" to something like 1000 and see what it does.

Feel free to open another issue about the "null" problem you mention so that we can discuss this issue in a dedicated topic ! it will be much easier for you and me :)

anchor avatar

guillim (anchor)

22 days ago

I am closing this issue since it's been 25 days. Feel free to reopen it if necessary

Developer
Maintained by Community
Actor metrics
  • 262 monthly users
  • 30 stars
  • 99.2% runs succeeded
  • 6.9 hours response time
  • Created in Oct 2021
  • Modified 9 days ago