Monday, 18 November 2013

Data scraping tool for non-coding journalists launches

A tool which helps non-coding journalists scrape data from websites has launched in public beta today.

Import.io lets you extract data from any website into a spreadsheet simply by mousing over a few rows of information.

Until now import.io, which we reported on back in April, has been available in private developer preview and has been Windows only. It is now also available for Mac and is open to all.

Although import.io plans to charge for some services at a later date, there will always be a free option.

The London-based start-up is trying to solve the problem of the fact that there is "lots of data on the web, but it's difficult to get at", Andrew Fogg, founder of import.io, said in a webinar last week.

Those with the know-how can write a scraper or use an API to get at data, Fogg said. "But imagine if you could turn any website into a spreadsheet or API."

Uses for journalists

Journalists can find stories in data. For example, if I wanted to do a story on the type of journalism jobs being advertised and the salaries offered, I could research this by looking at various websites which advertise journalism jobs.

If I were to gather the data from four different jobs boards and enter the information manually into a spreadsheet it would take would take hours if not days; if I were to write a screen scraper for each of the sites it would require knowledge and would probably take a couple of hours. Using import.io I can create a single dataset from multiple sources in a few minutes.

I can then search and sort the dataset and find out different facts, such as how many unpaid internships are advertised, or how many editors are currently being sought.

How it works

When you download the import.io application you see a web browser. This browser allows you to enter a URL for any site you want to scrape data from.

To take the example of the jobs board, this is structured data, with the job role, description and salaries displayed.

The first step is to set up 'connectors' and to do this you need to teach the system where the data is on the page. This is done by hitting a 'record' button on the right of the browser window and mousing over a few examples, in this case advertised jobs. You then click 'train rows'.

It takes between two and five examples to teach import.io where all of the rows are, Fogg explained in the webinar.

The next step is to declare the type of data and add column names. For example there may be columns for 'job title', 'job description' and 'salary'. Data is then extracted into the table below the browser window.

Data from different websites can then be "mixed" into a single searchable database.

In the example used in the webinar, Fogg demonstrated how import.io could take data relating to rucksacks for sale on a shopping website. The tool can learn the "extraction pattern", Fogg explained, and apply that to to another product. So rather than mousing over the different rows of sleeping bags advertised, for example, import.io was automatically able to detect where the price and product details were on the page as it had learnt the structure from how the rucksacks were organised. The really smart bit is that the data from all products can then be automatically scraped and pulled into the spreadsheet. You can then search 'shoes' and find the data has already been pulled into your database.

When a site changes its code a screen scraper would become ineffective. Import.io has a "resilience to change", Fogg said. It runs tests twice a day and users get notified of any changes and can retrain a connector.

It is worth noting that a site that has been scraped will be able to detect that import.io has extracted the data as it will appear in the source site's web logs.

Case studies

A few organisations have already used import.io for data extraction. Fogg outlined three.

    British Red Cross

The British Red Cross wanted to create an iPhone app with data from the NHS Choices website. The NHS wanted the charity to use the data but the health site does not have an API.

By using import.io, data was scraped from the NHS site. The app is now in the iTunes store and users can use it to enter a postcode to find hospital information based on the data from the NHS site.

"It allowed them to build an API for a website where there wasn't one," Fogg said.

    Hewlett Packard

Fogg explained that Hewlett Packard wanted to monitor the prices of its laptops on retailers' websites.

They used import.io to scrape the data from the various sites and were able monitor the prices at which the laptops were being sold in real-time.

    Recruitment site

A US recruitment firm wanted to set up a system so that when any job vacancy appeared on a competitor's website, they could extract the details and push that into their Salesforce software. The initial solution was to write scrapers, Fogg said, but this was costly and in the end they gave up. Instead they used import.io to scrape the sites and collate the data.


Source: http://www.journalism.co.uk/news/data-scraping-tool-for-non-coding-journalists-launches/s2/a554002/

Sunday, 17 November 2013

ScraperWiki lets anyone scrape Twitter data without coding

The Obama administration’s open data mandate announced on Thursday was made all the better by the unveiling of the new ScraperWiki service on Friday. If you’re not familiar with ScraperWiki, it’s a web-scraping service that has been around for a while but has primarily focused on users with some coding chops or data journalists willing to pay to have someone scrape data sets for them. Its new service, though, currently in beta, also makes it possible for anyone to scrape Twitter to create a custom data set without having to write a single line of code.

Taken alone, ScraperWiki isn’t that big of a deal, but it’s part of a huge revolution that has been called the democratization of data. More data is becoming available all the time — whether from the government, corportations or even our own lives — only it’s not of much use unless you’re able to do something with it. ScraperWiki is now one of a growing list of tools dedicated to helping everyone, not just expert data analysts or coders, analyze — and, in its case, generate — the data that matters to them.

After noticing a particularly large numbers of tweets in my stream about flight delays yesterday, I thought I’d test out ScraperWiki’s new Twitter search function by gathering a bunch of tweets directed to @United. The results — from 1,697 tweets dating back to May 3 — are pretty fun to play with, if not that surprising. (Also, I have no idea how far back the tweet search will go or how long it will take using the free account, which is limited to 30 minutes of compute time a day. I just stopped at some point so I could start digging in.)

First things first, I ran my query. Here’s what the data looks like viewed in a table in the ScraperWiki app.

Next, it’s a matter of analyzing it. ScraperWiki lets you view it in a table (like above), export it to Excel or query it using SQL, and will also summarize it for you. This being Twitter data, the natural thing to do seemed to be analyzing it for sentiment. One simple way to do this right inside the ScraperWiki table is to search for a particular term that might suggest joy or anger. I chose a certain four-letter word that begins with f.

Surprisingly, I only found eight instances. Here’s my favorite: “Your Customer Service is better than a hooker. I paid a bunch of money and you’re still…” (You probably get the idea.)

But if you read my “data for dummies” post from January, you know that we mere mortals have tools at our disposal for dealing with text data in a more refined way. IBM’s Many Eyes service won’t let me score tweets for sentiment, but I can get a pretty good idea overall by looking at how words are used. For this job, though, a simple word cloud won’t work, even after filtering out common words, @united and other obvious terms. Think of how “thanks” can be used sarcastically and you can see why.

Using the customized word tree, you can see that “thanks” sometimes means “thanks.” Other times, not so much. I know it’s easy to dwell on the negative, but consider this: “worst” had 28 hits while “best” had 15. One of those was referring to Tito’s vodka and at least three were referring to skyline views. (Click here to access it and search by whatever word you want.)

Here’s a phrase net filtering the results by phrases where the word “for” connects two words.

Anyhow, this was just a fast, simple and fairly crude example of what ScraperWiki now allows users to do, and how that resulting data can be combined with other tools to analyze and visualize it. Obviously, it’s more powerful if you can code, but new tools are supposedly on the way (remember, this is just a beta version) that should make it easier to scrape data from even more sources.

In the long term, though, services like ScraperWiki should become a lot more valuable as tools for helping us generate and analyze data rather than just believe what we’re told. Want to improve your small business, put your life in context or perhaps just write the best book report your teacher has ever seen? It’s getting easier every day.


Source: http://gigaom.com/2013/05/10/scraperwiki-lets-anyone-scrape-twitter-data-without-coding/

Friday, 15 November 2013

What is data scraping and how can I stop it?

Data scraping (also called web scraping) is the process of extracting information from websites. Data scraping focuses on transforming unstructured website content (usually HTML) into structured data which can be stored in a database or spreadsheet.

The way data is scraped from a website is similar to that used by search bots – human web browsing is simulated by using programs (bots) which extract (scrape) the data from a website.

Unfortunately, there is no efficient way to fully protect your website from data scraping. This is so because data scraping programs (also called data scrapers or web scrapers) obtain the same information as your regular web visitors.

Even if you block the IP address of a data scraper, this will not prevent it from accessing your website. Most data scraping bots use large IP address pools and automatically switch the IP address in case one IP gets blocked. And if you block too many IPs, you will most probably block many of your legitimate visitors.

One of the best ways to protect globally accessible data on a website is through copyright protection. This way you can legally protect the intellectual ownership of your website content.

Another way to protect your site content is to password protect it. This way your website data will be available only to people who can authenticate with the correct username and password.


Source: http://kb.siteground.com/what_is_data_scraping_and_how_can_i_stop_it/

Thursday, 14 November 2013

What you need to know about web scraping: How to understand, identify, and sometimes stop

This is a gust article by Rami Essaid, co-founder and CEO of Distil Networks.

Here’s the thing about web scraping in the travel industry: everyone knows it exists but few know the details.

Details like how does web scraping happen and how will I know? Is web scraping just part of doing business online, or can it be stopped? And lastly, if web scraping can be stopped, should it always be stopped?

These questions and the challenge of web scraping are relevant to every player in the travel industry. Travel suppliers, OTAs and meta search sites are all being scraped. We have the data to prove it; over 30% of travel industry website visitors are web scrapers.

Google Analytics, and most other analytics tools do not automatically remove web scraper traffic, also called “bot” traffic, from your reports – so how would you know this non-human and potentially harmful traffic exists? You have to look for it.

This is a good time to note that I am CEO of a bot-blocking company called Distil Networks, and we serve the travel industry as well as digital publishers and eCommerce sites to protect against web scraping and data theft – we’re on a mission to make the web more secure.

So I am admittedly biased, but will do my best to provide an educational account of what we’ve learned to be true about web scraping in travel – and why this is an issue every travel company should at the very least be knowledgeable about.

Overall, I see an alarming lack of awareness around the prevalence of web scraping and bots in travel, and I see confusion around what to do about it. As we talk this through I’ll explain what these “bots” are, how to find them and how to manage them to better protect and leverage your travel business.

What are bots, web scrapers and site indexers? Which are good and which are bad?

The jargon around web scraping is confusing – bots, web scrapers, data extractors, price scrapers, site indexers and more – what’s the difference? Allow me to quickly clarify.

–> Bots: This is a general term that refers to non-human traffic, or robot traffic that is computer generated. Bots are essentially a line of code or a program that is created to perform specific tasks on a large scale.  Bots can include web scrapers, site indexers and fraud bots. Bots can be good or bad.

–> Web Scraper: (web harvesting or web data extraction) is a computer software technique of extracting information from websites (source, Wikipedia). Web scrapers are usually bad.

If your travel website is being scraped, it is most likely your competitors are collecting competitive intelligence on your prices. Some companies are even built to scrape and report on competitive price as a service. This is difficult to prove, but based on a recent Distil Networks study, prices seem to be main target.You can see more details of the study and infographic here.

One case study is Ryanair. They have been particularly unhappy about web scraping and won a lawsuit against a German company in 2008, incorporated Captcha in 2011 to stop new scrapers, and when Captcha wasn’t totally effective and Cheaptickets was still scraping, they took to the courts once again.

So Ryanair is doing what seems to be a consistent job of fending off web scrapers – at least after the scraping is performed. Unfortunately, the amount of time and energy that goes into identifying and stopping web scraping after the fact is very high, and usually this means the damage has been done.

This type of web scraping is bad because:

    Your competition is likely collecting your price data for competitive intelligence.
    Other travel companies are collecting your flights for resale without your consent.
    Identifying this type of web scraping requires a lot of time and energy, and stopping them generally requires a lot more.

Web scrapers are sometimes good

Sometimes a web scraper is a potential partner in disguise.

Meta search sites like Hipmunk sometimes get their start by scraping travel site data. Once they have enough data and enough traffic to be valuable they go to suppliers and OTAs with a partnership agreement. I’m naming Hipmunk because the Company is one of the few to fess up to site scraping, and one of the few who claim to have quickly stopped scraping when asked.

I’d wager that Hipmunk and others use(d) web scraping because it’s easy, and getting a decision maker at a major travel supplier on the phone is not easy, and finding legitimate channels to acquire supplier data is most definitely not easy.

I’m not saying you should allow this type of site scraping – you shouldn’t. But you should acknowledge the opportunity and create a proper channel for data sharing. And when you send your cease and desist notices to tell scrapers to stop their dirty work, also consider including a note for potential partners and indicate proper channels to request data access.

–> Site Indexer: Good.

Google, Bing and other search sites send site indexer bots all over the web to scour and prioritize content. You want to ensure your strategy includes site indexer access. Bing has long indexed travel suppliers and provided inventory links directly in search results, and recently Google has followed suit.

–> Fraud Bot: Always bad.

Fraud bots look for vulnerabilities and take advantage of your systems; these are the pesky and expensive hackers that game websites by falsely filling in forms, clicking ads, and looking for other vulnerabilities on your site. Reviews sections are a common attack vector for these types of bots.

How to identify and block bad bots and web scrapers

Now that you know the difference between good and bad web scrapers and bots, how do you identify them and how do you stop the bad ones? The first thing to do is incorporate bot-identification into your website security program. There are a number of ways to do this.

In-house

When building an in house solution, it is important to understand that fighting off bots is an arms race. Every day web scraping technology evolves and new bots are written. To have an effective solution, you need a dynamic strategy that is always adapting.

When considering in-house solutions, here are a few common tactics:

    CAPTCHAs – Completely Automated Public Turing Tests to Tell Computers and Humans Apart (CAPTCHA), exist to ensure that user input has not been generated by a computer. This has been the most common method deployed because it is simple to integrate and can be effective, at least at first. The problem is that Captcha’s can be beaten with a little workand more importantly, they are a nuisance to end usersthat can lead to a loss of business.

    Rate Limiting- Advanced scraping utilities are very adept at mimicking normal browsing behavior but most hastily written scripts are not. Bots will follow links and make web requests at a much more frequent, and consistent, rate than normal human users. Limiting IP’s that make several requests per second would be able to catch basic bot behavior.
    IP Blacklists - Subscribing to lists of known botnets & anonymous proxies and uploading them to your firewall access control list will give you a baseline of protection. A good number of scrapers employ botnets and Tor nodes to hide their true location and identity. Always maintain an active blacklist that contains the IP addresses of known scrapers and botnets as well as Tor nodes.

    Add-on Modules – Many companies already own hardware that offers some layer of security. Now, many of those hardware providers are also offering additional modules to try and combat bot attacks. As many companies move more of their services off premise, leveraging cloud hosting and CDN providers, the market share for this type of solution is shrinking.

    It is also important to note that these types of solutions are a good baseline but should not be expected to stop all bots. After all, this is not the core competency of the hardware you are buying, but a mere plugin.

Some example providers are:

    Impreva SecureSphere- Imperva offers Web Application Firewalls, or WAF’s. This is an appliance that applies a set of rules to an HTTP connection. Generally, these rules cover common attacks such as Cross-site Scripting (XSS) and SQL Injection. By customizing the rules to your application, many attacks can be identified and blocked. The effort to perform this customization can be significant and needs to be maintained as the application is modified.

    F5 – ASM – F5 offers many modules on their BigIP load balancers, one of which is the ASM. This module adds WAF functionality directly into the load balancer. Additionally, F5 has added policy-based web application security protection.

Software-as-a-service

There are website security software options that include, and sometimes specialize in web scraping protection. This type of solution, from my perspective, is the most effective path.

The SaaS model allows someone else to manage the problem for you and respond with more efficiency even as new threats evolve.  Again, I’m admittedly biased as I co-founded Distil Networks.

When shopping for a SaaS solution to protect against web scraping, you should consider some of the following factors:

    Does the provider update new threats and rules in real time?
    How does the solution block suspected non-human visitors?
    Which types of proactive blocking techniques, such as code injections, does the provider deploy?
    Which of the reactive techniques, such as rate limiting, are used?
    Does the solution look at all of your traffic or a snapshot?
    Can the solution block bots before they reach your infrastructure – and your data?
    What kind of latency does this solution introduce?

I hope you now have a clearer understanding of web scraping and why it has become so prevalent in travel, and even more important, what you should do to protect and leverage these occurrences.

NB: This is a gust article by Rami Essaid, co-founder and CEO of Distil Networks.

NB2: Locked binder image courtesy Shutterstock.


Source: http://www.tnooz.com/article/what-you-need-to-know-about-web-scraping-how-to-understand-identify-and-sometimes-stop/

Tuesday, 12 November 2013

WP Web Scraper

An easy to implement professional web scraper for WordPress. This can be used to display realtime data from any websites directly into your posts, pages or sidebar. Use this to include realtime stock quotes, cricket or soccer scores or any other generic content. The scraper is an extension of WP_HTTP class for scraping and uses phpQuery or xpath for parsing HTML. Features include:

    Can be easily implemented using the button in the post / page editor.
    Configurable caching of scraped data. Cache timeout in minutes can be defined in minutes for every scrap.
    Configurable Useragent7 for your scraper can be set for every scrap.
    Scrap output can be displayed thru custom template tag, shortcode in page, post and sidebar (through a text widget).
    Other configurable settings like timeout, disabling shortcode etc.
    Error handling - Silent fail, error display, custom error message or display expired cache.
    Clear or replace a regex pattern from the scrap before output.
    Option to pass post arguments to a URL to be scraped.
    Dynamic conversion of scrap to specified character encoding (using incov) to scrap data from a site using different charset.
    Create scrap pages on the fly using dynamic generation of URLs to scrap or post arguments based on your page's get or post arguments.
    Callback function to parse the scraped data.

For demos and support, visit the WP Web Scraper project page. Comments appreciated.

Tags: curl, html, import, page, phpquery, Post, Realtime, sidebar, stock market, web scraping, xpath   



Source: http://wordpress.org/plugins/wp-web-scrapper/

Monday, 11 November 2013

Yellow Page Scraping- How Use Full

In short, technology has changed the world and really changed in the YP industry.

Thanks to the World Wide Web! Now anyone, anywhere, can always access the online YP. Is just a click away, you find the information you need, your city or across the world the best instructors can be mechanical or robotic care hospital locations.

Since its inception in a while with the organization is already one of the most visited sites online Yp. url legitimized their favorite food court / cinema / travel agent / doctor / editors / auto part store / game center provides updated information / settings / layouts / restaurant / hotel / pub / to eat and what not! Hyderabad and around each service can be found here.

Some commentators pointed out that the YP advertising is very expensive compared to that of search engines. So, for a meaningful and ROI of search traffic, as part of a good marketing campaign using the YP? My answer is yes, can and should the YP, local search engine optimization to make to be used as an important component. YP can be used for SEO, and here is some information about how to approach it.

First, why would you use the YP for natural search optimization?
Well, YP, local search companies in the search engines themselves, enjoy the great location. "There / Organic SEO is a fair amount of traffic driven by search engines."
YP sites high page rank and keywords is usually small compared to the sites "trade" types of searches will be able to get. For example, the Google Maps One Box results sites, the highest rank of the following discoveries:

"San Francisco Accountants"," Italian restaurant in Seattle"," Garden Supplies, Atlanta"," Miami Grocery Store".

IYPs provide several portal sites, news, books and many areas of the print YP to find the URL for the big promotion of direct navigation.

However, the questions point to the example I show up, will score well in many cases, companies must adapt their listings in the YP sites. Users click through to the IYP sites, and if no business listings, businesses will not be references.

Now, the question of cost:
Sterling Folx various discussions on the blog point out, YP advertising is expensive, especially in categories that are very popular. My "natural search engine optimization" hat wearing, I suggest that for the first time by optimizing the YP sites is not always cost money. Many of these sites on your company, including a URL to add additional information at no cost, is possible. So, all you can before you pay anything.

As with the major search engines, placement is very important that your entry you want to display to the top of the page. (You can reasonably expect. The YP and look at places where users most of the "Heat Map" of some sort, like search engines, and the sweet spot is close to listing the top of the paginate consider the alternative names may want to improve their rankings.

Biz online directory listings of the evaluation are in order of ranking, though, and this is one area where you could improve any price. Family and friends and others who are positively disposed toward you rate each of these sites ask. Some very satisfied / streaming client? Give them a voucher for future visits and beg them to you to judge on these online sites.

Now, the YP of hyperlinks to sites that help improve your site's Page Rank? In short, it depends on:
(1) If your biz displays information IYP pages are spider and ranking;

(2) link to your site / usable crawl (click on the link NO Followed coated or tracking code, it may not be search engine friendly) is.


Source: http://goarticles.com/article/Yellow-Page-Scraping-How-Use-Full-It-Is/5072397/

Sunday, 10 November 2013

Simple method of Data Scrapping

There are so many tools available on the Internet are scraping data. With these tools, without stress, you can download a large amount of data. The last decade, the Internet revolution as an information center was the world. You can get any information on the Internet. However, if you want to work with specific information, you must find other sites. Download all the information on the website that interests you, then you must copy the information in the document header. Everything seems to work a bit "more difficult. With scraping tools, your time, save money and can reduce manual labor.

Tools for extracting Web data to extract data from HTML pages and Web sites to compare data. Each day, there are many sites are hosted on the Internet. You can not see all the sites the same day. These data mining tools, you can view all pages on the Internet. If you use a wide range of applications, the scraping tool is also useful for you.

Software tools for data retrieval for structured data that is used on the Internet. There are so many Internet search engines to help you find a site for a particular problem would be. Various sites, the data appears in different styles. The expert scraped help you compare the different sites and structures for recording updated data.

And the web crawler software tool is used to index the Web pages on the Internet, moving data to the Internet from your hard drive. With this work, you can surf the Internet much faster than they are connected. It is time to use the tip of the device is important if you try to download data from the Internet. It will take considerable time to download. However, the device with faster Internet rate. There you can download all the corporate data of the person is another tool called e-mail extractor. The tribute, you can easily target your e-mail client. Each time your product is able to send targeted advertisements to customers. The customer database to find the best equipment.

Scraping and data extraction can be used in any organization, corporation, or any company which is a data set targeted customer industry, company, or anything that is available on the net as some data, such as e-ID mail data, site name, search term or what is available on the web. In most cases, data scraping and data mining services, not a product of industry, are marketed and used for example to reach targeted customers as a marketing company, if company X, the city has a restaurant in California, the software relationship that the city's restaurants in California and use that information for marketing your product to market-type restaurant company can extract the data.

MLM and marketing network using data mining and data services to each potential customer for a new client by extracting the data, and call customer service, postcard, e-mail marketing, and thus produce large networks to send large groups of construction companies and their products.

However, there are tolls are scraping on the Internet. And some sites have reliable information about these tools. By paying a nominal amount to download these tools.


Source: http://goarticles.com/article/Simple-method-of-Data-Scrapping/4692026/

Thursday, 24 October 2013

Simple Answer to a Frequently Asked Question, รข€˜What Is Screen Scrapingรข€™?

Undoubtedly, data extraction today has become a laborious task and thus calls the demand for latest technology to accomplish the job. With the support of web screen scraping services, the job to drag out required data and information has become simple and easy. Now the questions arises รขEUR~what is screen scrapingรขEUR(TM)? Well, it is a specially designed program that has proved to be of great help for the purpose of extraction of data, images and heavy files as well. This software helps individuals to download the specific data in the desired format. This service is like a boon for many websites.

There lies a tough competition in the market today. Business entrepreneurs are trying hard to get beneficial outcome in their business growth. With the support and help of scraping services, business owners are extracting the information of many internet users in their website and this readily helps them to grow their business. One big advantage of this program is that it can develop tons of datas in less time. In business scenario, it is time that matters a lot. So, businesses today are making use of this service to get the data available in no time.

Benefits of Screen Scraping

Fast Scraping: One greatest advantage of using this software is that it saves your time and labor. It lessens the chances of making you wait for long hours to provide you data. Also, the quick scraping tools offer you latest data.

Presentable: Scraping programs also offers data in readable format which could be used in a hassle free manner. The service providers can provide data in database file or spreadsheet or any other format as desired by the user. Data which cannot be read is of no use. Presentation means a lot.

As screen scarping is a software, it is made. In its development involves a group of experts that possess great knowledge in the field. They are basically programmers who have gained great expertise in the domain and are efficient to load innumerable dataรขEUR(TM)s from different websites in very little time.

Today, the market is swarming up with various service providers offering screen scraping services. Explore different websites and select one that excites you the most. Going online would not only save your time but also reduce the difficulty of going out in the sweltering sun. Get the details of the firm and contact their service providers to get the data extracted for your business. Furthermore, if you are concerned about the charges, do not worry as the facilities can be availed at realistic rates.

Henceforth, give your business a new turn with the best screen scraping service providers.



Source: http://goarticles.com/article/Simple-Answer-to-a-Frequently-Asked-Question-What-Is-Screen-Scraping/7872372/

Tuesday, 22 October 2013

Screen Scraper Software

Applications for Monitoring Competitor Pricing by using screen scraping.

In a world with seamless integration of internet information, more and more web data extraction services can be found providing reliable ways to monitor competitive pricing for your business. In addition to streamlining content, these companies gather resourceful information. Which is of course a vital asset for any company or private group's use. Not only for collecting and refining web content, you can also make use of gathered information in an organized form for purposes of intelligence, study, and storage for future use. Finding this form of web extraction service for you can take some seriously contemplated decision making, if you don't know where to look. But, with this article you will hopefully find that deciding which one best suites your need doesn't have to be headache inducing in the end.

The first name that comes to mind for monitoring competitor pricing would have to be Mozenda. Being the highest rated on sites like theeasybee.com, they have become a optimal solution for web content scraping of this nature. Mozenda offers a extremely easy, and organized approach with it's carefully crafted user interface. Collecting detailed marketing and research data could not be made simpler than they have made it. Dedicated to the search of online content for projects like competitive pricing, lead generation, or scientific research, you will find that Mozenda has been designed to fit all of your web extraction needs. But this is only a mere glimpse of what it has to offer. Mozenda converts your collected web data into many useful formats like CSV, TSV, XML, and RSS just to name a few. Also, for those new to web extraction, they even offer to set up your first project free of charge. But, I doubt you would even need that with all of the resources made available to you. They have a section on their page offering instructional videos that show you how to set up your very own projects extremely quick, and easily. In addition to the already impressive capabilities of Mozenda's software, they offer many sub services in order to get your job done correctly as well. Giving you more time to actually use the information collected in your projects any manner you like.

At a not too distant second is Kapow Technologies. Proudly claiming to deliver business solutions involving web data in only a fraction of the time as their competitors in software development. They also boast the ability to achieve the same end results in only a fraction of cost as well. Having gained much acclaim with their partnership with IBM in order to create a Web 2.0 Expo application for the IPhone in less than three hours, they definitely have the expertise to carry out the much simpler project ideas like these. One major attraction to their applications are it's abilities to extract with absolutely no coding, through it's exclusive point-and-click develop technology. They are a unique enterprise, capable of wrapping any existing web content or API with this lossless technique.

To see which applications and services work best for you, it is highly suggested that you take advantage of the free trial downloads that are made available on these sites. Most come with a two week test period, which allows more than enough time to figure out which one is best suited for your optimal business performance. Monitoring your competitor's pricing has been made a extremely easy task with all of the accessible options. Luckily, tedious and time-consuming methods are completely a thing of the past.



Source: http://goarticles.com/article/Screen-Scraper-Software/3623340/

Monday, 21 October 2013

Information About Craigslist Scraping Tools

Information is one amongst the foremost vital assets to a business.Whatever trade the business relies in, while not the crucialinformation that helps it to operate, it'll be left to die.However, you are doing not ought to hunt round the net or through pilesof resources so as to urge the data that you just would like. Instead,you can merely take the data that you just have already got and use itto your advantage.

With info being thus promptly accessible for big corporations, itmay be not possible to guess what precisely a corporation can would like this muchdata and data from. completely different jobs together with everything frommedical records analysis, to selling uses net hand tool technology inorder to compile info, analyze it and so use it for his or her ownpurposes.

Another reason that a corporation could utilize an internet hand tool is fordetection of changes. for instance, if you entered into a contract witha company to confirm that their net link stayed on your online page forsix months, they may use an internet hand tool to form certain that you just do notback out. this fashion they additionally don't ought to manually check yourwebsite a day to confirm that the link remains there. This savesthem from wasting their valuable labor prices.

Finally you'll be able to use an internet hand tool to urge all of the info concerning acompany that you just would like. whether or not you wish to seek out out what differentwebsites ar speech concerning your company, otherwise you merely need to seek out allof the data a few bound topic, employing a net hand tool is asimple, fast and simple answer.

There ar many various corporations that give you with the abilityto scrape the net for info. one amongst the businesses to lookat is Mozenda. Mozenda permits you to setup custom programs that scrapethe net for all differing types of knowledge, relying upon the exactneeds that your company has. Another net scraping company that ispopular is thirty Digits net Extractor. they assist you to extract theinformation that you just would like from a spread of internet sites and webapplications. you'll be able to use any type of alternative services to urge all ofyour information scraped from the online.

Web information scraping could be a growing business. There ar such a lot of industriesand businesses that use the data they get from net datascraping to accomplish quite bit. whether or not you would like to scrape information inorder to seek out personal info, past histories, compile databasesof factual info or another use it's terribly real and potential todo so! but, so as to use an internet hand tool effectively you mustmake certain to use a real company.

don't come with any company off thestreet, check that to visualize them against others within the trade. Ifworst involves worse, check drive many completely different corporations. Thenstick with the online hand tool that best meets your wants. check that thatyou let the online hand tool work for you, after all, the net is apowerful tool in your business!



Source: http://goarticles.com/article/Information-About-Craigslist-Scraping-Tools/7507586/

Saturday, 19 October 2013

Craigslist Scraping Data Extraction Tools

It is Associate in Nursing ever developing company that is serving the folks. The craigslist may be a net services company. it's one among st the leading issues in its category. the realm of operation has mature to over forty five countries round the world. This websites may be a specialist in that includes sales promotions.

all types of ads square measure displayed here starting from paid ads and free ads.

Ads of jobs, services, personal sales and lots of a lot of square measure displayed here. Even discussion forums square measure gift here in order that folks will discuss what they like. Their major supply of sales come back from the paid ads associated with jobs. it's thought to be the simplest web site without charge sales promotions on-line.

many folks take into account this because the best for looking jobs, service sand lots of a lot of. there's no marvel that it's stratified at the 33th spot within the whole world. within the u. s. of America it's thought-about because the seventh best web site overall Web Data Extraction Software, Scripts.
And the most astonishing reality is that it manages this whole business by to a small degree variety of staff. There square measure solely regarding thirty staff in it. there's no surprise it's should for those staff to be terribly economical. The success depends upon the co - ordination of those folks. folks will build cash by finance during this business.

If one trains himself and provides his commitment he will undoubtedly become extremely roaring. except for this it's crucial to settle on a tool for posting ads effectively. someone WHO posts several ads on Craigslist is aware of the work load and time it takes. however this stress and cargo are often overcome by employing a sensible Craigslist Posting tool. particularly if the posting tool is all automatic in posting ads it's another advantage. however it's not a straightforward task to zero in on one software package and shopping for it.

as a result of the quantity of software on the market within the net is very large Web Scraper Download.

You can have a headache in selecting one. however those efforts square measure worthwhile as a result of Craigslist is among the simplest which may communicate your ads to the whole world. it's Associate in Nursing economic and a good thanks to develop your business. There square measure lots of craigslist posting tools on the market that is absolutely automatic.

one among st the simplest ways that to choose a tool is to research the options and it should have the automated posting options. And conjointly each product offers a free trial for victimization it. when victimization the trial we are able to decide a tool and die. By these facilities it's simple for analyzing the merchandise.


Source: http://goarticles.com/article/Craigslist-Scraping-Data-Extraction-Tools/7529228/

Wednesday, 16 October 2013

The Manifold Advantages Of Investing In An Efficient Web Scraping Service

Bitrake is an extremely professional and effective online data mining service that would enable you to combine content from several webpages in a very quick and convenient method and deliver the content in any structure you may desire in the most accurate manner. Web scraping may be referred as web harvesting or data scraping a website and is the special method of extracting and assembling details from various websites with the help from web scraping tool along with web scrapping software. It is also connected to web indexing that indexes details on the online web scraper utilizing bot (web scrapping tool). The dissimilarity is that web scraping is actually focused on obtaining unstructured details from diverse resources into a planned arrangement that can be utilized and saved, for instance a database or worksheet.

Frequent services that utilize online web scraper are price-comparison sites or diverse kinds of mash-up websites. The most fundamental method for obtaining details from diverse resources is individual copy-paste. Never web scraping theless, the objective with Bitrake is to create an effective software to the last element. Other methods comprise DOM parsing, upright aggregation platforms and even HTML parses. Web scraping might be in opposition to the conditions of usage of some sites. The enforceability of the terms is uncertain.

While complete replication of original content will in numerous cases is prohibited, in the United States, court ruled in Feist Publications v Rural Telephone Service that replication details is permissible. Bitrate service allows you to obtain specific details from the net without technical information; you just need to send the explanation of your explicit requirements by email and Bitrate will set everything up for you. The latest self-service is formatted through your preferred web browser and formation needs only necessary facts of either Ruby or Javascript. The main constituent of this web scraping tool is a thoughtfully made crawler that is very quick and simple to arrange.

The web scraping software permits the users to identify domains, crawling tempo, filters and preparation making it extremely flexible. Every web page brought by the crawler is effectively processed by a draft that is accountable for extracting and arranging the essential content. Data scraping a website is configured with UI, and in the full-featured package this will be easily completed by Bitrake. However, Bitrake has two vital capabilities, which are:

- Data mining from sites to a planned custom-format (web scraping tool)

- Real-time assessment details on the internet.



Source: http://goarticles.com/article/The-Manifold-Advantages-Of-Investing-In-An-Efficient-Web-Scraping-Service/5509184/

Tuesday, 15 October 2013

Understanding Web Scraping

It is evident that the invention of the internet is one of the greatest inventions of life. This is so because it allows quick recovery of information from large databases. Though the internet has its own negative aspects, its advantages outweigh the demerits f using it. It is therefore the objective of every researcher to understand the concept of web scraping and learn the basics of collecting accurate data from the internet. The following are some of the skills researchers need to know and keep them abreast of:

Understanding File Extensions in Web Scraping

In web scraping the first step to know is file extensions. For instance a site ending with dot-com is either a sales or commercial site. With the involvement of sales activity in such a website, there is a possibility that the data contained therein is inaccurate. Sites that may be ending with dot-gov are sites owned by various governments. The information found on such websites is accurate since they are reviewed by professionals regularly. Sites ending with dot-org are sites owned by non-governmental organizations that are not after making profit. There is a greater probability that the information contained is not accurate. Sites ending with dot-edu are owned by educational institutions. The information found on such sites is sourced by professionals and is of high quality. In case you have no understanding concerning a particular website it is important that get more information from expert data mining services.

Search Engine Limitations in Web Scraping

After understanding the file extensions, the next step is to understand search engine limitations applied to web scraping. These include process such as file extension, filtering or any other parameters. The following are some of the restrictions that need to typed after your search term: for instance if you key in รข€ล“financeรข€ and then click รข€ล“searchรข€ all sites will be listed from the dot-com directory that contain the word finance on its website. If you key in รข€ล“finance site.gov,รข€ of course with the quotation marks, only the government sites that have the word finance will be listed. The same applies to other sites with different file extensions.

Advanced Parameters in Web Scraping

When performing web scraping it is important to understand more skills beyond the file extension. Therefore there is a need to understand particular search terms. For instance if you key in รข€ล“software company in Indiaรข€ without the quotation marks, the search engines will display thousands of websites having รข€ล“softwareรข€, รข€ล“companyรข€ and India in their search terms. If you key in รข€ล“Software Company in Indiaรข€ with the quotation marks, the search engines will only display sites that contain the exact phrase รข€ล“software company in Indiaรข€ within their text.

This article forms the basis of web scraping. Collection of data needs to be carried out by experts and high quality tools. This is to ensure that the quality and accuracy of the data scraped is of high standards. The information extracted from that data has wide applications in business operations including decision making and predictive analytics.


Source: http://goarticles.com/article/Understanding-Web-Scraping/6771732/

Friday, 11 October 2013

How Can You Scrape Data From Amazon

Article Summary:

Amazon.com is a huge site that advertises and sells vast range of products. And hence to extract information of a particular product or large number of products belonging to the same category or myriad categories, professional marketing companies nowadays prefer using Amazon product scraper.

Article Body:

you need to be aware of the exact tools that are used. Amazon.com is a huge site that advertises and sells vast range of products. And hence to extract information of a particular product or large number of products belonging to the same category or myriad categories, professional marketing companies nowadays prefer using Amazon product scraper.

How does this scraper help?

Amazon product scraper is a great help as it captures all the details of the product/products such as product name, model no., its description, selling details and shipping price. It comes with a One Screen Dashboard which makes possible for you to view all the information on single screen. This dashboard also reveals all the extracted keywords, records and elapses. This provision also helps in easy controlling and operation. This scraper is also known as product extractor and rightly so as it crawls through the whole site and extracts Asin, model no, title, description, URL and other relevant details in a readable, clean CSV format which can be easily opened in excel and viewed.

This scraper is compatible with almost all types of computer systems such as Windows Vista, Windows XP, Windows 98, Net Framework 2.0, and Windows 7. It also comes with multiple channel criteria which enables the user to run multiple proxies at one time and search for multiple keywords.

How to scrape data from Amazon?

With the help of this software you can search for hundreds of targeted products with deep-scan technology in matter of few minutes. It provides you with the facility to scrape and search Amazonรƒ¢รข‚¬รข„¢s US API for particular products via 16 search parameters and present them in readable and clean CSV format that can be opened in excel.

For a professional who intends to design a price comparison or Amazon niche site extracting all the details of product and images is quite time consuming and frustrating. But with the help of this automated software you can easily scrape data from Amazon within matter of few minutes.

Ways to scrape data from Amazon

รข€¢ You can scrape data by browsing the product catalog for up to 3 sub-categories
รข€¢ Usually Amazonรƒ¢รข‚¬รข„¢s most Gifted/ top 10 Bestseller/ most hot and wished new listings of product is searched for.
รข€¢ You can also search the site and scrape information by Keywords/Title/Manufacture
รข€¢ Other search can include ASIN or ISBN numbers.


Source: http://goarticles.com/article/How-Can-You-Scrape-Data-From-Amazon/7210828/

Thursday, 10 October 2013

Web Scraping and Financial Matters

Many marketers value the process of harvesting data on the financial sector. They are also conversant with the challenges concerning the collection and processing of the data. Web scraping techniques and technologies are used for tracking and recognizing patterns that are found within the data. This is quite useful to businesses as it shifts through the layers of data, remove unrelated data and only leave the data that has meaningful relationships. This enables companies anticipate rather than just reacting to the customer and financial needs. Web scraping in combination with other complementary technologies and sound business processes, it can be used in reinforcing and redefining financial analysis.

Objectives of web scraping

The following are some of the web scraping services objectives that are covered in this article:

1. Discus show the customization of data and data mining tools may be developed for financial data analysis.

2. What is the usage pattern, in terms of purpose and the categories for the need for financial analysis?

3. Is the development of a tool for financial analysis through web scraping techniques possible?

Web scraping can be regarded as the procedure of extracting or harvesting knowledge for the large quantities of data. It is also known as Knowledge Discovery in Database (KDD). This implies that web scraping involves data collection, data management, database creation and the analysis of data and its understanding.

The following are some of the steps that are involved in web scraping service:

1. Data cleaning. This is the process of removing nose and the inconsistent data. This process is important as it only ensures that only important data should be integrated. This process saves time that will be consumed in the next processes.

2. Data integration. This is the processes of combining multiple sources of information. This process is quite important as it ensure that there is sufficient data for selection purposes.

3. Data selection. This is retrieving of data from databases that are relevant from the data in question.

4. Data transformation. It is the process of consolidating or transforming data into forms, which are appropriate for scraping by performing aggregation operations and summary.
5. Data mining. This is the process where intelligent methods are used in extracting data patterns.

6. Pattern evaluation. It is the identification of the patterns that are quite interesting and ones that represent knowledge and the interesting measures.

7. Knowledge presentation. It is the process where knowledge representation techniques and visualization are used in representing extracted data to the user.

Data Warehouse

Data warehouse may be defined as a store where information that has been mined from different sources, and stored under a unified schema and it resides at a single site.

Majority of banks and financial institutions offer a wide variety of baking services that include checking account balances, savings, customer and business transactions. Other services that may be offered by such companies include investment and credit services. Stock and insurance services may also be offered.

Through web scraping services it is possible for companies to gather data from financial and banking sectors, which may be relatively reliable, high quality and complete. Such data is quite important is it facilitates the analysis and the decision making of a company.



Source: http://goarticles.com/article/Web-Scraping-and-Financial-Matters/6771760/

Wednesday, 9 October 2013

Ultimate Scraping Three Common Methods For Web Data Extraction

So what's the best way to data extraction? It really is dependent upon what your needs are, and what resources you have you can use. Here are some of the pros and cons of the various options, as well as suggestions on once you might use each an individual:

Raw regular expressions in addition to code

<em>Advantages: </em>

- If you're already informed about regular expressions and some form of programming language, this may be a quick solution.

- Regular expressions allow for a fair amount of "fuzziness" inside the matching such that minor changes towards the content won't break them all.

- You likely don't should try to learn any new languages or perhaps tools (again, assuming you're already informed about regular expressions and a new programming language).

- Regular expressions are supported in most of modern programming languages. Daylights, even VBScript has a daily expression engine. It's also nice for the reason that various regular expression implementations don't vary too significantly within their syntax.

<em>Disadvantages: </em>

- They are definitely complex for those that don't have plenty of experience with them. Figuring out regular expressions isn't want going from Perl for you to Java. It's more enjoy going from Perl to make sure you XSLT, where you really have to wrap your mind around an entirely different way of viewing the condition.

- They're often confusing to evaluate. Take a look through a number of the regular expressions people have manufactured to match something as simple as an email address and you'll see what i mean.

- If the content you're endeavoring to match changes (e. h., they change the internet page by adding a brand-new "font" tag) you'll likely must update your regular expressions to take into account the change.

- The data discovery component to the process (traversing various web pages to go to the page containing the data you want) will still should be handled, and can get fairly complex region deal with cookies and additionally such.

<em>When to make use approach: </em> You'll most in all likelihood use straight regular expressions in screen-scraping when you experience a small job you intend to get done quickly. Especially if you now know regular expressions, there's no sense in stepping into other tools if all you decide to do is pull some news headlines off a site.

Ontologies as well as artificial intelligence

<em>Advantages: </em>

- You create the software once and it can awfully extract the data from any page while in the content domain you're looking for.

- The data model is mostly built in. For case in point, if you're extracting data files about cars from online sites the extraction engine now knows what the help to make, model, and price are generally, so it can easily map it to existing data structures (e. gary the gadget guy., insert the data throughout the correct locations in ones own database).

- There is certainly relatively little long-term preservation required. As web sites change you likely might want to do very little to all your extraction engine as a way to account for the transformations.

<em>Disadvantages: </em>

- It's relatively complex for making and work with this engine. The level of expertise needed to even understand an removal engine that uses man-made intelligence and ontologies is noticeably higher than what must deal with regular words and phrases. Professionals Implement Key Search engine optimization Metric Techniques


Source: http://goarticles.com/article/Ultimate-Scraping-Three-Common-Methods-For-Web-Data-Extraction/5123576/

Monday, 7 October 2013

Challenges in Effective Web Data Mining

Data collection and web data mining are critical processes for many companies and the marketing companies today. The techniques usually used include search engines, topic-based searches and directories. Web data mining is necessary for any business that wants to create data warehouses by harvesting data from the internet. This is so because high-quality and intelligent information may not be harvested from the internet easily. Such information is critical as it enables you to get desired results and the business intelligence in demand.

Keyword-based searches are important in marketing of company products. They are usually affected by the following factors:
รข€¢ Irrelevant pages. The use of common and general keywords on the search engines yields millions of web pages. Some of thesepages may be irrelevant and may not be of help to the user.

รข€¢ Ambiguous results.This is usually caused by multi-variant or similar keyword semantics. A name would be an animal, movie or even a sport accessory. This results in web pages that are different what you are actually searching for.

รข€¢ Possibility of missing some web pages.There is a great possibility of missing the most relevant information that is contained on web pages that are not indexed on a given keyword.

One of the factors that prohibit the usage of web data mining is the effectiveness of search engine crawlers. This is widely evidenced by lack of access of the entire web due to search engine crawlers and bot.This can be attributed partly tobandwidth limitations. It is important to understand that there are thousands of databases on the internet that can deliver well-maintained information, high quality and are not easily accessed by crawlers.

In web data mining it is important to understand that majority of search engines have limited choices or alternatives for keyword query combination. For instance, yahoo and Google offer option like phrase and even the exact matches that may limit even the search results. It is usually demands more efforts and even time and thereby get the most important and relevant information.The human behavior and the alternatives usually change of time.This therefore implies that web pages need to be updated frequently and there by reflect on the emerging trends. It is important to realize that there is a limited space for web data mining. This is so because the information that currently exists is heavily relied on keyword-based indices. This does not apply for the real data.

It is important to realize that web data mining is an important tool for any business. It is therefore important to embrace this technology to solve data crisis problems. There are several limitations and many challenges which may have resulted in the quest of effectively and efficiently in rediscovering the use of web resources. However, irrespective of the challenges of web data mining, this technology is an effective tool that can be employed in many technological and scientific fields. It is therefore paramount to embrace this technology and use it fully in order to realize your corporate goals.


Source: http://goarticles.com/article/Challenges-in-Effective-Web-Data-Mining/6771744/

Friday, 4 October 2013

Web Screen Scrape With a Software Program

Which software do you use for data mining? How much time does it take in mining required data and is it able to present in a customized format? Extracting data from the web is

a tedious job, if done manually but the moment you use an application or program, web screen scrape job becomes easy.

Using an application would certainly make data mining an easy affair but the problem is that which application to choose. Availability of a number of software programs makes

it difficult to choose one but you have to select a program because you canรขEUR(TM)t keep mining data manually. Start your search for a data mining software program with

determining your needs. First note down the time a program takes to completing a project.

Quick scraping

The software shouldnรขEUR(TM)t take much time and if it does then thereรขEUR(TM)s no use of investing in the software. A software program that needs time for data mining would

only save your labor and not time. Keep this factor in mind as you canรขEUR(TM)t keep waiting for hours for the software to provide you data. Another reason behind choosing a

quick software program is that you a quick scraping tool would provide you latest data.

Presentation

Extracted data should be presented in readable format that you could use in a hassle free manner. For instance the web screen scrape program should be able to provide data in

spreadsheet or database file or in any other format as desired by the user. Data thatรขEUR(TM)s difficult to read is good for nothing. Presentation matters most. If you

arenรขEUR(TM)t able to understand the data then how could you use in future.

Coded program

Invest in web screen scrape program coded for your project and not for everyone. It should be dedicated to you and not made for public. There are groups that provide coded

programs for data mining. They charge a fee for programming but the job they do worth a fee. Look for a reliable group and get the software program that could make your data

mining job a lot easier.

Whether you are looking for contact details of your targeted audiences or you want to keep a close watch on social media, you need web screen scrape service that would save

your time and labor. If youรขEUR(TM)re using a software program for data mining then you should make sure that the program works according to your wishes.


Source: http://goarticles.com/article/Web-Screen-Scrape-With-a-Software-Program/7763109/

Thursday, 3 October 2013

Web Screen Scrape: Quick and Affordable Data Mining Service

Getting contact details of people living in a certain area or practicing a certain profession isnรขEUR(TM)t a difficult job as you could get the data from websites. You can even get the data in short time so that you could take advantage of it. Web screen scrape service could make data mining a breeze for you.

Extracting data from websites is a tedious job but there isnรขEUR(TM)t any need to mine the data manually as you could get it electronically. The data could be extracted from websites and presented in a readable format like spreadsheet and data file that you could store for future use. The data would be accurate and since you would get the data in short time, you could rely on the information. If your business relies on the data then you should consider using this service.

How much this data extraction service would cost? It wonรขEUR(TM)t cost a fortune. It isnรขEUR(TM)t expensive. Service charge is determined on the number of hours put in data mining. You can locate a service provider and ask him to give quote for his services. If youรขEUR(TM)re satisfied with the service and the charge, you could assign the data mining work to the person.

ThereรขEUR(TM)s hardly any business that doesnรขEUR(TM)t need data. For instance some businesses look for competitor pricing to set their price index. These companies employ a team for data mining. Similarly you can find businesses downloading online directories to get contact details of their targeted customers. Employing people for data mining is a convenient way to get online data but the process is lengthy and frustrating. On the other hand, service is quick and affordable.

You need specific data; you can get it without spending countless hours in downloading data from websites. All you need to do to get the data is contact a credible web screen scrape service provider and assign the data mining job to him. The service provider would present the data in the desired format and in the expected time. As far as budget of the project is concerned, you can negotiate the price with the service provider.

Web screen scrape service is a boon for websites. This service is quite beneficial for websites that rely on data like tour and travel, marketing and PR companies. If you need online data then you should consider hiring this service instead of wasting time on data mining.



Source: http://goarticles.com/article/Web-Screen-Scrape-Quick-and-Affordable-Data-Mining-Service/7783303/

Wednesday, 2 October 2013

Why to Go With a Web Screen Scraping Program?

There is a tough competition in the market, nowadays. Business owners are trying to get the best and beneficial result in their business growth. At present, there are different

kinds of businesses available online. With the support of their specific websites, business owners are promoting their products as well as services online. Currently, most of the

people are internet users and in order to get their contact details, websites owners are availing the benefits of software that can help them to get the desired data in a very short

time. Websites are now extracting relevant data of internet users with the support of web screen scraping software, these days. Undoubtedly, data collection from websites is a

time consuming and laborious job and thus one need to have a dedicated team to do so. However today, with the support of website screen scraping program, it has become so

easy to extract required data from websites as it was never before.

Screen scraping is really a beneficial program that can help people to download the desired data in an appropriate format. Therefore, it would be great for people to select a

screen scraping program instead of going with data mining team. There is no denying to this fact that this software would make your job much easier than before. There are a

number of benefits of using this software for the people in different ways. First of all, this program enables you to save lots of your precious time and to get your particular

project done in a very short time. If there is need to collect contact details of targeted audiences from some specific websites then it can easily be done with the support of this

program.

The best thing about this software is that it would help your data mining team to get rid of the tedious job of data mining from different websites. software will not only make your

data mining team free from the tedious job but also make you able to utilize them in some other productive projects of your company. With the support of this software, you will

surely experience great improvement in your teamรขEUR(TM)s productivity. This program will surely make you able to get the data in the same format you are looking for. It will

allow you to get the required data in suitable format. So, what are you waiting for? Leave all your data extracting problems on this software and enjoy its benefits!



Source: http://goarticles.com/article/Why-to-Go-With-a-Web-Screen-Scraping-Program/7803789/

Tuesday, 1 October 2013

Microsys A1 Website Scraper Review

The A1 scraper by Microsys is a program that is mainly used to scrape websites to extract data in large quantities for later use in webservices. The scraper works to extract text, URLs etc., using multiple Regexes and saving the output into a CSV file. This tool is can be compared with other web harvesting and web scraping services.
How it works
This scraper program works as follows:
Scan mode

    Go to the ScanWebsite tab and enter the site’s URL into the Path subtab.
    Press the ‘Start scan‘ button to cause the crawler to find text, links and other data on this website and cache them.

Important: URLs that you scrape data from have to pass filters defined in both analysis filters and output filters. The defining of those filters can be set at the Analysis filters and Output filters subtabs respectively. They must be set at the website analysis stage (mode).
Extract mode

    Go to the Scraper Options tab
    Enter the Regex(es) into the Regex input area.
    Define the name and path of the output CSV file.
    The scraper automatically finds and extracts the data according to Regex patterns.

The result will be stored in one CSV file for all the given URLs.

There is a need to mention that the set of regular expressions will be run against all the pages scraped.
Some more scraper features

Using the scraper as a website crawler also affords:

    URL filtering.
    Adjustment of the speed of crawling according to service needs rather than server load.

If  you need to extract data from a complex website, just disable Easy mode: out press the  button. A1 Scraper’s full tutorial is available here.
Conclusion

The A1 Scraper is good for mass gathering of URLs, text, etc., with multiple conditions set. However this scraping tool is designed for using only Regex expressions, which can increase the parsing process time greatly.



Source: http://extract-web-data.com/microsys-a1-website-scraper-review/

Saturday, 28 September 2013

Visual Web Ripper: Using External Input Data Sources

Sometimes it is necessary to use external data sources to provide parameters for the scraping process. For example, you have a database with a bunch of ASINs and you need to scrape all product information for each one of them. As far as Visual Web Ripper is concerned, an input data source can be used to provide a list of input values to a data extraction project. A data extraction project will be run once for each row of input values.

An input data source is normally used in one of these scenarios:

    To provide a list of input values for a web form
    To provide a list of start URLs
    To provide input values for Fixed Value elements
    To provide input values for scripts

Visual Web Ripper supports the following input data sources:

    SQL Server Database
    MySQL Database
    OleDB Database
    CSV File
    Script (A script can be used to provide data from almost any data source)

To see it in action you can download a sample project that uses an input CSV file with Amazon ASIN codes to generate Amazon start URLs and extract some product data. Place both the project file and the input CSV file in the default Visual Web Ripper project folder (My Documents\Visual Web Ripper\Projects).

For further information please look at the manual topic, explaining how to use an input data source to generate start URLs.


Source: http://extract-web-data.com/visual-web-ripper-using-external-input-data-sources/

Thursday, 26 September 2013

Using External Input Data in Off-the-shelf Web Scrapers

There is a question I’ve wanted to shed some light upon for a long time already: “What if I need to scrape several URL’s based on data in some external database?“.

For example, recently one of our visitors asked a very good question (thanks, Ed):

    “I have a large list of amazon.com asin. I would like to scrape 10 or so fields for each asin. Is there any web scraping software available that can read each asin from a database and form the destination url to be scraped like http://www.amazon.com/gp/product/{asin} and scrape the data?”

This question impelled me to investigate this matter. I contacted several web scraper developers, and they kindly provided me with detailed answers that allowed me to bring the following summary to your attention:
Visual Web Ripper

An input data source can be used to provide a list of input values to a data extraction project. A data extraction project will be run once for each row of input values. You can find the additional information here.
Web Content Extractor

You can use the -at”filename” command line option to add new URLs from TXT or CSV file:

    WCExtractor.exe projectfile -at”filename” -s

projectfile: the file name of the project (*.wcepr) to open.
filename – the file name of the CSV or TXT file that contains URLs separated by newlines.
-s – starts the extraction process

You can find some options and examples here.
Mozenda

Since Mozenda is cloud-based, the external data needs to be loaded up into the user’s Mozenda account. That data can then be easily used as part of the data extracting process. You can construct URLs, search for strings that match your inputs, or carry through several data fields from an input collection and add data to it as part of your output. The easiest way to get input data from an external source is to use the API to populate data into a Mozenda collection (in the user’s account). You can also input data in the Mozenda web console by importing a .csv file or importing one through our agent building tool.

Once the data is loaded into the cloud, you simply initiate building a Mozenda web agent and refer to that Data list. By using the Load page action and the variable from the inputs, you can construct a URL like http://www.amazon.com/gp/product/%asin%.
Helium Scraper

Here is a video showing how to do this with Helium Scraper:


The video shows how to use the input data as URLs and as search terms. There are many other ways you could use this data, way too many to fit in a video. Also, if you know SQL, you could run a query to get the data directly from an external MS Access database like
SELECT * FROM [MyTable] IN "C:\MyDatabase.mdb"

Note that the database needs to be a “.mdb” file.
WebSundew Data Extractor
Basically this allows using input data from external data sources. This may be CSV, Excel file or a Database (MySQL, MSSQL, etc). Here you can see how to do this in the case of an external file, but you can do it with a database in a similar way (you just need to write an SQL script that returns the necessary data).
In addition to passing URLs from the external sources you can pass other input parameters as well (input fields, for example).
Screen Scraper

Screen Scraper is really designed to be interoperable with all sorts of databases. We have composed a separate article where you can find a tutorial and a sample project about scraping Amazon products based on a list of their ASINs.


Source: http://extract-web-data.com/using-external-input-data-in-off-the-shelf-web-scrapers/

Wednesday, 25 September 2013

How to scrape Yellow Pages with ScreenScraper Chrome Extension

Recently I was asked to help with the job of scraping company information from the Yellow Pages website using the ScreenScraper Chrome Extension. After working with this simple scraper, I decided to create a tutorial on how to use this Google Chrome Extension for scraping pages similar to this one. Hopefully, it will be useful to many of you.
1. Install the Chrome Extension

You can get the extension here. After installation you should see a small monitor icon in the top right corner of your Chrome browser.
2. Open the source page

Let’s open the page from which you want to scrape the company information:



3. Determine the parent element (row)

The first thing you need to do for the scraping is to determine which HTML element will be the parent element. A parent element is the smallest HTML element that contains all the information items you need to scrape (in our case they are Company Name, Company Address and Contact Phone).  To some extent a parent element defines a data row in the resulting table.

To determine it, open Google Chrome Developer Tools (by pressing Ctrl+Shift+I), click the magnifying class (at the bottom of the window) and select the parent element on the page. I selected this one:


As soon as you have selected it, look into the developer tools window and you will see the HTML code related to this element:


As is seen from the highlighted HTML line, you can easily define a parent element by its class: listingInfoAndLogo.

5. Determine the information elements (columns)

After you have learned how to determine the parent element, it should be easy to specify the information elements that contain the information you want to scrape (they represent columns in the resultant table).

Just do this in the same way that you did it for the parent element -  by selecting it on the page:

As you can see, the company name is defined by businessName class.
6. Tune the ScreenScraper itself

After all the data elements you want to scrape are found, open the ScreenScraper by clicking the small monitor icon in the top-right corner of your browser. Then do the following:

    Enter the parent element class name (listingInfoAndLogo in our case) into the Selector field, preceding it with a dot (*see below for why)
    Click the Add Column button
    Enter a field’s name (any) into the Field text box
    Enter the information item class into the Selector text box, preceding it with a dot
    Repeat steps 2-4 for each information item element you want to be scraped


After you enter all these definitions you should see the preview of the scraped data at the bottom of the extension’s windowThat’s it! I hope the tutorial is clear enough. But if not, feel free to write your comments below and I’ll give additional explanations.


Source: http://extract-web-data.com/how-to-scrape-yellow-pages-with-screenscraper-chrome-extension/

Tuesday, 24 September 2013

Data Mining: The AdWords Problem Review

This post is a continuation of the previous post on Advertising on the Web and Data mining. Here we conclude by reviewing some basic algorithms for placing ads on the web. The adwords problem is to be solved with association graphs algorithm in Data Mining.
The AdWords Problem

With Adwords we consider Google’s example. Google’s policy for placing ads is not on the bid value but in the total amount expected to be received for the display of each ad. The value of an ad was taken to be the product of the bid and the click-through rate (the click-through rate is taken for each ad, based on the statistics of displays of an ad).
Match graphs

After we concluded that the ad placement challenge is to be done with on-line algorithms of greedy or generalized balance types (previous post), we need to consider also the match graphs for bid-query match.

This problem involves two sets of nodes and a set of edges between members of the two sets. The first set is the incoming search queries while the second set is the related ads from various sources. The goal is to ๏ฌnd a maximal matching — as large a set of edges as possible that includes no node more than once. The On-Line Solution to the Matching Problem would be a greedy algorithm for ๏ฌnding a match in a bipartite graph. It is done by ordering the edges in certain way. The competitive ratio in this case should be no less than 1/2. That is the on-line greedy matching algorithm finds matches at least half as many nodes as the best off-line algorithm does.
Basic Adwords workflow

Of course, the decision regarding which ads to show must be made on-line. So, there should be only on-line algorithms considered for solving the adwords problem. The input data and algorithm results are as follows.
Input:
1. A set of bids by advertisers for search queries.
2. A click-through rate (CTR) for each advertiser-query pair
3. A budget for each advertiser. We shall assume budgets are for a
month, although any unit of time could be used.
4. A limit on the number of ads to be displayed with each search query.

• Output – Respond to each search query with a set of advertisers of those characteristics:
1. The size of the set is no larger than the limit on the number of ads per query.
2. Each advertiser has a bid on the search query.
3. Each advertiser has enough budget left to pay for the ad if it is
clicked upon.


Implementing the Adwords algorithm

Basically we now have an idea of how ads are selected to go with the answer to a search query. But we have not addressed the problem of ๏ฌnding the bids that have been made on a given query.
Finding bids for search queries

Which algorithm finds the bids that have been made on a given query? For a simple case we can represent a query by the list of its words, in sorted order. Bids are stored in a hash table or similar structure, with a hash key equal to the sorted list of words. A search query can then be matched against bids by a straightforward lookup in the table.

The simplest version of this implementation serves in situations where the bids are on exactly the set of words in the search query (in reality, search engines all offer to advertisers a ‘broad matching’ feature). It also does not consider the historical frequency of queries or statistics.

The task might not just be limited to the same set of words. For example, Google also matches adwords bids to emails. There, the match criterion is not based on the equality of sets but on the inclusiveness of the bid’s set in the related document (email). How could you compare the millions of words from bids with the millions of words in the stream of regular emails?
Matching Algorithm for Documents and Bids

The Matching Algorithm for Documents and Bids has been developed for this purpose. Briefly, it works by processing the words of the document, rarest-๏ฌrst. Word sets whose ๏ฌrst word is the current word are copied to a temporary hash table, with the second word as the key. Sets already in the temporary hash table are examined to see if the word that is their key matches the current word, and, if so, they are rehashed using their next word. Sets whose last word is matched are copied to the output. For more details on this algorithm you might read more from the same book, paragraph 8.5.3.
Summary

The data mining algorithms for ad placement in response to a search query or larger documents (emails) are on-line greedy or generalized balance algorithms on the match graphs. The complex query/bid matching process is to be done with hashing sets-of-words tables with the consequent matching of rare to unrare word order in order to find the total match of the bid set in the document.



Source: http://extract-web-data.com/data-mining-the-adwords-problem-review/

Monday, 23 September 2013

How to Outsource Data Entry Work Effectively

In today's world it is a well known fact that many businesses now outsource data entry work. All businesses are concerned with the running costs of their business as well as keeping clients and staff happy. One of the ways to achieve all of these goals is to use outsourcing techniques, which are growing in strength each year.

Outsourcing is now a staple part of business life. Whether you are a large conglomerate or a small office based business, there are aspects of your business which are already outsourced. For example, you may likely have a contract with a cleaner to clean your office or gardener to tidy up that hedge.

It is true to say that many larger businesses have the time, resources and money to invest in employing their own in-house own data entry specialists. However, mid-sized and smaller companies need to be able to operate at the same level as the large companies, but with less money, time and resources. This is where they can benefit from outsourcing this kind of work.

If you want to outsource data entry work, you need to firstly analyze how much it is going to aid your business. Is it necessary for your data entry work to be outsourced? You need to have a solid idea of your future business plans and work out where the data entry outsourcing fits into the plan. You need to do a lot of research and communicate with prospective outsourcing companies or individuals. Do not be afraid to ask questions; it is your business at stake should anything go wrong.

By outsourcing your data compilation work, you are taking care of many business related issues. Many data entry specialists either work as independent freelancers or may be part of a company specializing in outsourced data entry. This results in lower costs for your business; you are likely to receive a quote from an outsourcing company that is very competitive. If the work is an ad-hoc project, you may find that a freelance data entry worker is the cheapest option.

As the years have shown, outsourcing has proved a viable and advantageous option for many businesses. Whether it is employing a call center supervisor or a data specialist, your lower core competences can be dealt with by outside help. This leaves you to concentrate on the core competences that are of higher importance to the business and allow you to use your valuable time wisely.

Outsourcing is also a lot cheaper than employing in-house staff. The companies that offer to outsource entry of data have skilled workers, who can increase productivity whilst keeping your costs to a minimum. There is also the advantage of focusing your in-house staff; if you outsource data entry work it will allow more interesting, less-time consuming and important projects to be enjoyed by your own staff.

New technology is also emerging each year in the business world. By employing companies to outsource data entry projects you can eliminate some of the risk, save some time and some money. Many outsourcing companies have the latest technology in order for them to keep producing world-class results for their clients.




Source: http://ezinearticles.com/?How-to-Outsource-Data-Entry-Work-Effectively&id=2449297

Friday, 20 September 2013

Data Entry in Outsourcing Businesses

The process in, which a business house engages another company to do a particular type of work instead of using its own employees to do the same work, is called outsourcing. This is basically practiced so that the company can concentrate more on the core function. The cheap cost of outsourcing work is also another reason.

Outsourcing companies are often referred as "business to business" companies. Their business is dependent on the service provided by them to other business houses. Nowadays, every company is engaged in outsourcing. When a sole proprietor gives responsibility to another to buy supplies for the office, then automatically this process becomes outsourcing. In a real sense, it is almost impossible to do everything by yourself. You have to become dependent on those who are skilled in certain fields.

Data entry is one of the oldest and well known as the most common outsourcing activities that have been widely accepted across the globe for a long period of time. Still today, the demand is sky rocketing and the scope of data entry companies are just expanding.

All companies value their data very much. In order to generate good business, you need to deal with your data efficiently. Thus, companies related to BTB activities take care of the data handling very seriously. The employees are trained and prepared for all sorts of detailed oriented work. The services vary from back office support for a banking institute, calculation of medical bills, maintaining payroll functions etc. Banks generally outsource the work of the business class customers. Lock box payment is one of such example.

There are plenty of companies in the market of outsourcing who are engaged in providing in different kinds of services to the clients all across the globe. Many companies, which are earlier engaged into hard core data entry operations, are now exploring the area of medical billing, research work, project work for various universities, marketing job, news agencies, trade and several types of insurance organizations.

You can help your company to grow and reach a tremendous height once you get accustomed to take the advantages from various available data entry work. The service providers take an extra step to make sure that the work those are being delivered are of high quality and fulfill all the requirements as asked by the clients. Accuracy and punctually are the keywords to survive in the outsourcing market. Companies prefer outsourcing as the cost is always lower than the company would require spending on salaries if the same work was done by their own employees. Outsourcing is a very lucrative option for many business houses as it gives you the freedom to concentrate on your core business process and even you end up saving a good sum of money by outsourcing data entry work.



Source: http://ezinearticles.com/?Data-Entry-in-Outsourcing-Businesses&id=2021508

Thursday, 19 September 2013

Outsourcing Data Entry Services

Data or raw information is the backbone of any industry or business organization. However, raw data is seldom useful in its pure form. For it to be of any use, data has to be recorded properly and organized in a particular manner. Only then can data be processed. That is why it is important to ensure accurate data entry. But because of the unwieldy nature of data, feeding data is a repetitive and cumbersome job and it requires heavy investment, both in terms of time and energy from staff. At the same time, it does not require a high level of technical expertise. Due to these factors, data entry can safely be outsourced, enabling companies to devote their time and energy on tasks that enhance their core competence.

Many companies, big and small, are therefore enhancing their productivity by outsourcing the endless monotonous tasks that tend to cut down the organization's productivity. In times to come, outsourcing these services will become the norm and the volume of work that is outsourced will multiply. The main reason for these kinds of development is the Internet. Web based customer service and instant client support has made it possible for service providers to act as one stop business process outsourcing partners to parent companies that require support.

Data entry services are not all alike. Different clients have different demands. While some clients may require recording information coupled with document management and research, others may require additional services like form processing or litigation support. Data entry itself could be from various sources. For instances, sometimes information may need to be typed out from existing documents while at other times, data needs to be extracted from images or scanned documents. To rise up to these challenges, service providers who offer these services must have the expertise and the software to ensure rapid and accurate data entry. That is why it is important to choose your service provider with a lot of care.

Before hiring your outsourcing partner, you need to ask yourself the following questions.

* What kind of reputation does the company enjoy? Do they have sufficient years of experience? What kind of history and background does the company enjoy?

* Do they have a local management arm that you can liaise with on a regular basis?

* Do the service personnel understand your requirements and can they handle them effectively?

* What are the steps taken by the company to ensure that there is absolutely no compromise in confidentiality and security while dealing with vital confidential data?

* Is there a guarantee in place?

* What about client references?

The answers to these questions will help you identify the right partner for outsourcing your data entry service requirements.




Source: http://ezinearticles.com/?Outsourcing-Data-Entry-Services&id=3568373

Tuesday, 17 September 2013

Data Mining and Financial Data Analysis

Introduction:

Most marketers understand the value of collecting financial data, but also realize the challenges of leveraging this knowledge to create intelligent, proactive pathways back to the customer. Data mining - technologies and techniques for recognizing and tracking patterns within data - helps businesses sift through layers of seemingly unrelated data for meaningful relationships, where they can anticipate, rather than simply react to, customer needs as well as financial need. In this accessible introduction, we provides a business and technological overview of data mining and outlines how, along with sound business processes and complementary technologies, data mining can reinforce and redefine for financial analysis.

Objective:

1. The main objective of mining techniques is to discuss how customized data mining tools should be developed for financial data analysis.

2. Usage pattern, in terms of the purpose can be categories as per the need for financial analysis.

3. Develop a tool for financial analysis through data mining techniques.

Data mining:

Data mining is the procedure for extracting or mining knowledge for the large quantity of data or we can say data mining is "knowledge mining for data" or also we can say Knowledge Discovery in Database (KDD). Means data mining is : data collection , database creation, data management, data analysis and understanding.

There are some steps in the process of knowledge discovery in database, such as

1. Data cleaning. (To remove nose and inconsistent data)

2. Data integration. (Where multiple data source may be combined.)

3. Data selection. (Where data relevant to the analysis task are retrieved from the database.)

4. Data transformation. (Where data are transformed or consolidated into forms appropriate for mining by performing summary or aggregation operations, for instance)

5. Data mining. (An essential process where intelligent methods are applied in order to extract data patterns.)

6. Pattern evaluation. (To identify the truly interesting patterns representing knowledge based on some interesting measures.)

7. Knowledge presentation.(Where visualization and knowledge representation techniques are used to present the mined knowledge to the user.)

Data Warehouse:

A data warehouse is a repository of information collected from multiple sources, stored under a unified schema and which usually resides at a single site.

Text:

Most of the banks and financial institutions offer a wide verity of banking services such as checking, savings, business and individual customer transactions, credit and investment services like mutual funds etc. Some also offer insurance services and stock investment services.

There are different types of analysis available, but in this case we want to give one analysis known as "Evolution Analysis".

Data evolution analysis is used for the object whose behavior changes over time. Although this may include characterization, discrimination, association, classification, or clustering of time related data, means we can say this evolution analysis is done through the time series data analysis, sequence or periodicity pattern matching and similarity based data analysis.

Data collect from banking and financial sectors are often relatively complete, reliable and high quality, which gives the facility for analysis and data mining. Here we discuss few cases such as,

Eg, 1. Suppose we have stock market data of the last few years available. And we would like to invest in shares of best companies. A data mining study of stock exchange data may identify stock evolution regularities for overall stocks and for the stocks of particular companies. Such regularities may help predict future trends in stock market prices, contributing our decision making regarding stock investments.

Eg, 2. One may like to view the debt and revenue change by month, by region and by other factors along with minimum, maximum, total, average, and other statistical information. Data ware houses, give the facility for comparative analysis and outlier analysis all are play important roles in financial data analysis and mining.

Eg, 3. Loan payment prediction and customer credit analysis are critical to the business of the bank. There are many factors can strongly influence loan payment performance and customer credit rating. Data mining may help identify important factors and eliminate irrelevant one.

Factors related to the risk of loan payments like term of the loan, debt ratio, payment to income ratio, credit history and many more. The banks than decide whose profile shows relatively low risks according to the critical factor analysis.

We can perform the task faster and create a more sophisticated presentation with financial analysis software. These products condense complex data analyses into easy-to-understand graphic presentations. And there's a bonus: Such software can vault our practice to a more advanced business consulting level and help we attract new clients.

To help us find a program that best fits our needs-and our budget-we examined some of the leading packages that represent, by vendors' estimates, more than 90% of the market. Although all the packages are marketed as financial analysis software, they don't all perform every function needed for full-spectrum analyses. It should allow us to provide a unique service to clients.

The Products:

ACCPAC CFO (Comprehensive Financial Optimizer) is designed for small and medium-size enterprises and can help make business-planning decisions by modeling the impact of various options. This is accomplished by demonstrating the what-if outcomes of small changes. A roll forward feature prepares budgets or forecast reports in minutes. The program also generates a financial scorecard of key financial information and indicators.

Customized Financial Analysis by BizBench provides financial benchmarking to determine how a company compares to others in its industry by using the Risk Management Association (RMA) database. It also highlights key ratios that need improvement and year-to-year trend analysis. A unique function, Back Calculation, calculates the profit targets or the appropriate asset base to support existing sales and profitability. Its DuPont Model Analysis demonstrates how each ratio affects return on equity.

Financial Analysis CS reviews and compares a client's financial position with business peers or industry standards. It also can compare multiple locations of a single business to determine which are most profitable. Users who subscribe to the RMA option can integrate with Financial Analysis CS, which then lets them provide aggregated financial indicators of peers or industry standards, showing clients how their businesses compare.

iLumen regularly collects a client's financial information to provide ongoing analysis. It also provides benchmarking information, comparing the client's financial performance with industry peers. The system is Web-based and can monitor a client's performance on a monthly, quarterly and annual basis. The network can upload a trial balance file directly from any accounting software program and provide charts, graphs and ratios that demonstrate a company's performance for the period. Analysis tools are viewed through customized dashboards.

PlanGuru by New Horizon Technologies can generate client-ready integrated balance sheets, income statements and cash-flow statements. The program includes tools for analyzing data, making projections, forecasting and budgeting. It also supports multiple resulting scenarios. The system can calculate up to 21 financial ratios as well as the breakeven point. PlanGuru uses a spreadsheet-style interface and wizards that guide users through data entry. It can import from Excel, QuickBooks, Peachtree and plain text files. It comes in professional and consultant editions. An add-on, called the Business Analyzer, calculates benchmarks.

ProfitCents by Sageworks is Web-based, so it requires no software or updates. It integrates with QuickBooks, CCH, Caseware, Creative Solutions and Best Software applications. It also provides a wide variety of businesses analyses for nonprofits and sole proprietorships. The company offers free consulting, training and customer support. It's also available in Spanish.

ProfitSystem fx Profit Driver by CCH Tax and Accounting provides a wide range of financial diagnostics and analytics. It provides data in spreadsheet form and can calculate benchmarking against industry standards. The program can track up to 40 periods.



Source: http://ezinearticles.com/?Data-Mining-and-Financial-Data-Analysis&id=2752017