|
find email addresses
|
. | Register
online now Please e-mail us if you have any questions. Content
What types of e-mail marketing are there?
Option 1. Download web pages onto your hard drive and then extract the e-mail addresses using our program, search.exe.
Online search and scanning for e-mail addresses and hyperlinks using Link Extractor .
You can use the search engines in these two ways:
Scan newsgroups with Newsgroup Explorer (nntp.exe). |
You can use Search.exe to extract e-mail addresses from a selected directory on your computer containing Web pages or other types of files that may contain e-mail addresses. Files other than Web pages that might contain e-mail addresses include word processing documents, Outlook Express files, address books, files from your newsgroup reader programs and more. Search.exe will help you find e-mail addresses that you didn't even know you had.If you already have a set of Web pages downloaded on your local disk you can use Search.exe to extract the e-mail addresses.
1) Start the program by launching the Search.exe file. If you are unsure how to do this, simply double-click the file name from within Windows Explorer.
2) Select the required directory
3) Click the button "Domain List" to define the list of the countries
4) Use the "Except" button to exclude addresses (For example: abuse, spam, postmaster, etc ...). This will permit you remove addresses that contain these strings of text.
5) Save the "Except" list as a text file.
At extracting you can make file masks on researched files. For example *.txt, *.html. This determines what types of files that Search.exeextract from. To begin extraction, press the Search button. This button will change to a button with the label "Stop". The program will indicate when it has completed searching the files, or you may stop the program at any time by pressing the "Stop" button.
When the program completes, it will create a text file that contains the list of e-mail addresses extracted. For example:
Vista@aha.ru
Mega@cityline.ru
etc,
etc
Or URLs, for example:
http://www.vista.ru
http://www.aha.ru
You can then import the text file directly into Vista NetMail! if you have that
product.
While Search.exe allows you to search for addresses on a local hard drive, Webscape allows you to search directly on the Internet.When you first use Webscape, you need to have Netscape Navigator running. Thereafter, you can run Webscape alone.
The interface of the program is similar to the interface of Search.exe.
But here instead of specifying a file mask and searching your hard drive, you specify a list of URLs and search the Web directly.
You can:
The options in the above picture affect the appearance of the interface.
You can define the keywords to include documents for processing and also define keywords to exclude documents from processing.
The buttons "Save, Except and Delete " - work exactly like those in Search.exe. (mentioned previously).
You can change the maximum time in seconds spent scanning a URL by adjusting the value of 'Skip after' (shown below).
You can set up the options to define your search exactly.
You can select which domains and countries you wish to process in the same way as with Search.exe (mentioned previously).
The Working Directory - Here you select the local directory where you wish to save the search results and the temporary files used in the search.
The visited URL - Enter the name of the file
containing the list of previously searched URLs. The default is
webscan.txt.
Choose a different file name for
different projects. During major projects, this file can get extremely large so
ensure that you change this file name occasionally to keep it manageable!
The excluded URL - Enter the name of the file containing the list of excluded URLs. The default is webexcl.txt.
The keywords exception file - Enter the name of the file containing the keywords to exclude certain URLs and e-mail addresses. The default is webexcp.txt.
URL swap count - this number defines the period
of additional URLs that will be scanned before adding them to the list for
further scanning. For example if the number 2000 is entered, after 2000
URLs are found and displayed in the window, they will be transferred to the
temporary file for further scanning.
Link Extractor
3.
Linkextr.exe
How the program works
Main menu
After downloading the program the main menu will appear on your screen.
The main menu consists of the following options:
Before running the program it is advisable to adjust the general settings.
To do this, launch the program and choose Default Options.
The first thing to do is decide which directory (new path) you will use to save project files and the path to the directory for saving files copied (downloaded) from the Internet.
Download files - this option is used to download files onto your hard drive.
Unless this option is highlighted the system will only download a list of scanned hyperlinks into a special file.
Enter the proxy server properties (if you use a proxy server).
Then choose any other options you would like to use in downloading and searching for hyperlinks.
Let's take a look at the
various options available.
Follow new links / URL - to follow hyperlinks
automatically - this option allows you to automatically extract other websites
linked to the one you are scanning.
Copy subdirectory structure from website - to copy the structure of a subdirectory from the website you wish to download. If this option is highlighted your hard drive will be able to create directories like the ones on the website you are downloading.
Extract local link - to search for local hyperlinks. This option allows you to search for local links on the website you are scanning, i.e. links that refer to other documents on the website.
Extract E-mail from website - this option will allow you to find and save all e-mail addresses shown in the website.
Stay within initial domain list. - A very convenient option that allows you to extract (not download) hyperlinks (websites) not included in the original list of addresses. Here you should decide whether you need to download other websites referred to from one you are downloading. Using this option you will only download the files you order. In this case the sites linked to the one you are investigating will also be downloaded.
For example, you only need to download a list of (URL) addresses
Internet-soft.com/demo
Esalesbiz.com/web
…
and you don't need to download other domains linked to the original list of domains (e.g. http://vista.ru).
Links level limit - number of downloading levels - shows the number of steps involved in the hyperlinks.
An example will help to illustrate this option.
Let's assume there is a hyperlink from one site to another. There is a
link from the second link to the third, etc.
As
you can see, a number of hyperlinks must be followed to get from one site to
another. This option gives you the greatest possible number of hyperlink
steps. Each step enables you to make some hyperlinks with a number of
other websites. So if you have selected only one level, you will only be
able to copy the websites (let's call them XI websites) to which there is a link
on the website you are downloading (scanning), and not the sites with hyperlinks
from XI websites.
The following chart shows how the links level limit works.
Number of connections - In this item you enter the number of simultaneous connections.
As a rule 3 - 10 connections are made. The optimal number of connections will depend on the number of lines you have and the connection speed of your provider.
Save results automatically - To save your results automatically every N of minutes.
This option shows how frequently your interim search results are to be saved.
Time out for one connection - This option gives the maximum amount of time in seconds during which each document (one connection) is downloaded.
At the end of this time the program starts downloading the next document.
Number of retries - The number of attempts made to download each document.
This option shows the number of attempts to download the same file if the provider connection or website link is broken off. The program will make as many attempts to download as you specify.
Swap URL count - The number of addresses added to the swap list of tasks (tree of downloadable addresses).
Does not visit twice already scanned site - This option allows you not to scan the addresses which have already been searched previously.
Apply domainname.com = www.domainname.com
In some sites the hyperlinks to other sites contain no original www symbols and when the same documents are downloaded they may be inscribed twice in different directories. This option is designed to deal with this anomaly in Internet sites. If you highlight this option INTERNET.SOFT.COM and WWW.INTERNET.SOFT.COM will be treated as synonymous addresses. The address is automatically prefixed as www in this type of search.
Expand the nodes parents to make the node visible - This convenience option is intended to graphically represent the tree of websites scanned.
In this way the option shows the current branches of the site being downloaded and enables the program to graphically depict the locations where sites are downloaded.
Identify browser as - This option shows how the program will be identified when the website is downloaded by a remote server.
For example, when you download a page using Internet Explorer 5.0, the remote server performs this operations and writes the contents of the server as a protocol. The Extractor program does the same thing when you visit a website.
We would like to draw your attention to the
following:
Since the worldwide web contains a
huge number of pages great data processing power may be needed as well as a
large amount of disk space on your computer to download links and
websites. A few hours of work by the program may take up many gigabytes on
your hard disk.
File Type Filter: Limiting the types and sizes of files
You can use this option to specify the types of files you want to download and limit their size.
This is important, for example, when you only want to download text documents without banners, pictures or archive files.
In this case, check the option beside html, htm, txt and shtml, etc. files.
You can use these menu options to limit the size of files to be downloaded. If you have selected "Load all file sizes", files of all sizes will be downloaded. Otherwise you will only get the sizes (specified in bytes) you have selected.
URL / Domain Filter: Limitations by names of directories, domain names and files.
You can make limitations by entering certain words in domains. Let's say you're downloading files only from www.yahoo.com. You would only enter yahoo as the filter word.
The filter can be used separately:
- to adjust the word content in a domain name;
- to expand the domain;
- to adjust the contents of a certain word in a directory
name;
- to modify any given word in the file
name.
The filter can be used to include and exclude. If you have entered words into the exclude filter, this means that if the URL contains any of these words, the corresponding files will not be downloaded. If you opt for the include filter, this means that only the names containing the properties specified in the word filter will be downloaded.
Domains: Limitations by domain type.
This option enables you to make limitations by type and country of the domain.
To do this click on the requested domain type.
This is all you have to do for the main program settings.
When you exit the menu window you save by default the data you have entered and you can proceed to download websites and e-mail addresses.
E-mail Extracting
Now we can start a search project. The default properties you have entered will automatically be called up when you start a new project. These properties can be altered and saved for a later time for each separate project.
The "search and download website" concept is at the heart of this system. The term "project" therefore refers to the total number of options that define which sites and properties are to be downloaded.
To start a new project, press "New".
An interface to define search and download (website)
criteria will appear on your screen.
Then enter in the left window the list of websites (URL) which you would like to download. By pressing the Load button you can download the list from a text file.
By pressing "Options" you enter the specific properties for your search project. Here you should check your directories and make sure you have enough disk space to download the required websites.
Search properties are entered the same way as default properties.
After you have assigned the search parameters, close this window and save the project by pressing "save as". Then give the project file any name you choose.
For example, you could call the site yahoo.pro.
Then proceed to download the data by pressing the "Download/Extract" button.
The properties that are most often changed are "download files" and "number of connections". These properties are conveniently located at the top of the search window toolbar to avoid having to exit this environment and enter data in another window.
After downloading data we recommend pressing "save as" or "save".
In this way you will be able to reuse these properties, if needed, at a later time.
To access existing projects, press "open" and choose the name of the project.
You can re-download websites by pressing the "download" button or by continuing the download from the last site you have visited. To do this select the proper file on the "download tree" and press "Resume".
Extract e-mails (Search e-mails)
Using this option you make an e-mail address search
for each page that is scanned. The result of this search will appear on
the toolbar and you can save it in the registered version of the program.
The unregistered version of the program will not allow you to download more than
50 addresses and does not allow you to save search results. Aside from
addresses the file can save the URL of pages where you found e-mail
addresses. The file will save only unique e-mail addresses.
Search Engine Scanner
The main menu of the program contains three items
Previously, you read how it was necessary to set parameters of the search in the "Preferences" menu choice. Engine.exe works the same way as Webscape here.You can choose the menu option "Search Condition" to set the parameters of search. Here, you can choose the search engines used for the creation of your e-mail database.
At the same time, you may use some other programs (Webscape.exe, Engine.exe) to accelerate your search of e-mail addresses. However, the programs should be started from different directories on the hard disk of your computer.
If you simultaneously run several instances of the program, processing many search engines, you will increase the time of your search for e-mail addresses. Since the program will make a list of URLs for a further searching, this will also require significant time for processing.After you provide the parameters of a search, press the "Search" button. As the program finds pages with e-mail addresses and other URLs, they will appear in the program window, similar to the program Webscape. Once again, let us remind you that Netscape Navigator it is not required.
After the search, the results may be saved to a text file. This is the same technique used by the program Webscape.
One of the program parameters is "The visited URL" file. During operation, this file can grow up to very a large size (ten megabyte or more). Therefore it is best to periodically change the name of this file for different searches or to monitor is size.
The program will allow you to scan newsgroup conferences on keywords and compilation the list of matching articles.Before you perform a search, it is necessary to define the list of newsgroup conferences that you have access to through your ISP. If you are unaware of when newsgroups your ISP offers, we suggest using a standard newsgroup client such as Microsoft Outlook Express or Forte's Free Agent News Readers to determine what newsgroups are available.
The main window of the program looks like:
The order of operation is as follows:
1) Press the button "Get Newsgroup list and search articles".Next Page >>
2) Use the "Keywords" button to set the list of keywords.
3) Enter the list of conferences (newsgroup host names) for scanning.
4) Press the button labeled "Save As" to save the results once the search is complete.