Chapter 3 Process of Web Usage Mining

Similar documents
Web Mining in E-Commerce: Pattern Discovery, Issues and Applications

Pattern Classification based on Web Usage Mining using Neural Network Technique

The influence of caching on web usage mining

Nitin Cyriac et al, Int.J.Computer Technology & Applications,Vol 5 (1), WEB PERSONALIZATION

Overview of Web Mining Techniques and its Application towards Web

CHAPTER - 3 PREPROCESSING OF WEB USAGE DATA FOR LOG ANALYSIS

International Journal of Software and Web Sciences (IJSWS)

A Review Paper on Web Usage Mining and Pattern Discovery

Survey Paper on Web Usage Mining for Web Personalization

A Survey on Web Personalization of Web Usage Mining

Web Usage Mining: A Research Area in Web Mining

Web Data mining-a Research area in Web usage mining

EFFECTIVELY USER PATTERN DISCOVER AND CLASSIFICATION FROM WEB LOG DATABASE

An Effective method for Web Log Preprocessing and Page Access Frequency using Web Usage Mining

Effectively Capturing User Navigation Paths in the Web Using Web Server Logs

Mining Web Data. Lijun Zhang

emetrics Study Llew Mason, Zijian Zheng, Ron Kohavi, Brian Frasca Blue Martini Software {lmason, zijian, ronnyk,

International Journal of Scientific & Engineering Research, Volume 4, Issue 11, November ISSN

Discovering Paths Traversed by Visitors in Web Server Access Logs

UNIT-V WEB MINING. 3/18/2012 Prof. Asha Ambhaikar, RCET Bhilai.

Analytics: measuring web site success. MBA 563 Week 4

Site Activity. Help Documentation

Chapter 2 BACKGROUND OF WEB MINING

Web Mining. Data Mining and Text Mining (UIC Politecnico di Milano) Daniele Loiacono

Web Mining. Data Mining and Text Mining (UIC Politecnico di Milano) Daniele Loiacono

Web Mining. Data Mining and Text Mining (UIC Politecnico di Milano) Daniele Loiacono

DATA MINING II - 1DL460. Spring 2014"

WEB USAGE MINING: ANALYSIS DENSITY-BASED SPATIAL CLUSTERING OF APPLICATIONS WITH NOISE ALGORITHM

INTRODUCTION. Chapter GENERAL

Pre-processing of Web Logs for Mining World Wide Web Browsing Patterns

Chapter 12: Web Usage Mining

Web Crawlers Detection. Yomna ElRashidy

Inferring User Search for Feedback Sessions

and the Forensic Science CC Spring 2007 Prof. Nehru

CLASSIFICATION OF WEB LOG DATA TO IDENTIFY INTERESTED USERS USING DECISION TREES

ANALYTICS DATA To Make Better Content Marketing Decisions

DATA MINING - 1DL105, 1DL111

Overture Advertiser Workbook. Chapter 4: Tracking Your Results

Measurement and evaluation: Web analytics and data mining. MGMT 230 Week 10

Analytics: measuring web site success. MBA 563 Week 3

SEQUENTIAL PATTERN MINING FROM WEB LOG DATA

Web, HTTP and Web Caching

This tutorial has been prepared for computer science graduates to help them understand the basic-to-advanced concepts related to data mining.

DISCOVERING USER IDENTIFICATION MINING TECHNIQUE FOR PREPROCESSED WEB LOG DATA

Mining Web Data. Lijun Zhang

Data warehousing and Phases used in Internet Mining Jitender Ahlawat 1, Joni Birla 2, Mohit Yadav 3

International Journal of Advance Engineering and Research Development. Survey of Web Usage Mining Techniques for Web-based Recommendations

Web Engineering (CC 552)

Web Mining Team 11 Professor Anita Wasilewska CSE 634 : Data Mining Concepts and Techniques

Lecture 9a: Sessions and Cookies

Google Analytics. Gain insight into your users. How To Digital Guide 1

Google Analytics. powerful simplicity, practical insight

Web Usage Mining. Overview Session 1. This material is inspired from the WWW 16 tutorial entitled Analyzing Sequential User Behavior on the Web

Study on Personalized Recommendation Model of Internet Advertisement

6 TOOLS FOR A COMPLETE MARKETING WORKFLOW

Site Audit Boeing

Association Rule Mining among web pages for Discovering Usage Patterns in Web Log Data L.Mohan 1

12 Web Usage Mining. With Bamshad Mobasher and Olfa Nasraoui

CONVERSION TRACKING PIXEL GUIDE

Web Architecture Review Sheet

Improving the prediction of next page request by a web user using Page Rank algorithm

Site Audit Virgin Galactic

Information Retrieval Spring Web retrieval

Google Analytics: Part 3

A PRAGMATIC ALGORITHMIC APPROACH AND PROPOSAL FOR WEB MINING

Design of a Social Networking Analysis and Information Logger Tool

CS 161 Computer Security

Site Audit SpaceX

GOOGLE ANALYTICS 101 INCREASE TRAFFIC AND PROFITS WITH GOOGLE ANALYTICS

If you re a Facebook marketer, you re likely always looking for ways to

Adaptive and Personalized System for Semantic Web Mining

Do-It-Yourself Guide for Advertisers

Openfolio, LLC Privacy Policy

Improved Data Preparation Technique in Web Usage Mining

Vantage Ultimate 2.2 Quick Start Tutorial

Search Engines. Information Retrieval in Practice

Google Analytics 101

Unit 4 The Web. Computer Concepts Unit Contents. 4 Web Overview. 4 Section A: Web Basics. 4 Evolution

Setup Google Analytics

AN OVERVIEW OF SEARCHING AND DISCOVERING WEB BASED INFORMATION RESOURCES

Cross-Browser Functional Testing Best Practices

USER INTEREST LEVEL BASED PREPROCESSING ALGORITHMS USING WEB USAGE MINING

User Session Identification Using Enhanced Href Method

3. WWW and HTTP. Fig.3.1 Architecture of WWW

Configuring Caching Services

Detects Potential Problems. Customizable Data Columns. Support for International Characters

Features of a proxy server: - Nowadays, by using TCP/IP within local area networks, the relaying role that the proxy

Document version Introduction to Webtrekk Analytics

Intro to Analytics Learning Web Analytics

ForeScout Open Integration Module: Data Exchange Plugin

Furl Furled Furling. Social on-line book marking for the masses. Jim Wenzloff Blog:

The Ultimate Guide for Content Marketers. by SEMrush

INTRODUCTION. In this summary version, we present some of the key figures and charts.

A SURVEY- WEB MINING TOOLS AND TECHNIQUE

Smart Bulk SMS & Voice SMS Marketing Script with 2-Way Messaging. Quick-Start Manual

Web Crawlers Detection. Yomna ElRashidy

WEB SECURITY WORKSHOP TEXSAW Presented by Solomon Boyd and Jiayang Wang

The Application of Web Usage Mining In E-commerce Security

Getting the most from your websites SEO. A seven point guide to understanding SEO and how to maximise results

Network Discovery Policies

Transcription:

Chapter 3 Process of Web Usage Mining 3.1 Introduction Users interact frequently with different web sites and can access plenty of information on WWW. The World Wide Web is growing continuously and huge amount of data is generated due to users numerous interactions with web sites. Web Usage Mining is the application of data mining techniques to discover the useful and interesting patterns from web usage data. It supports to know frequently accessed pages, predict user navigation, improve web site structure etc. The web usage data consists of the data from web server logs, browser logs, proxy server logs and user profiles. In Web Usage Mining, data mining techniques are applied to pre-processed web log data in order to find interesting and useful patterns. Visitors browsing behavior is recorded into web server log. By analyzing log files, different questions can be answered such as: What pages are being accessed frequently? From what search engine are visitors coming? Which browser and operating systems are most commonly used by visitors? What are the most recent visits per page? Who is visiting which page? E-commerce has provided effective way of doing business by electronic transactions through internet. For Web Usage Mining in e-commerce, the data of customer profile, inventory and demographic information from other relational databases are integrated with web usage data and visitors behavior patterns can be discovered by applying data mining techniques such as Association Rules, Sequential Analysis, Clustering and Classification. Web Usage Mining can help e-commerce companies to improve 28

the web site, attract visitors, to provide personalized and adaptive service to regular user, identify potential customers for e-commerce, supporting business intelligence and marketing decisions etc. 3.2 Description of Web Usage Mining Process The Web Usage Mining is the application of data mining technique to discover the useful patterns from web usage data. It can discover the user access patterns by mining log files and associated data of particular web site. Figure3.1 shows the process of Web Usage Mining consisting steps Data Collection, Pre-processing, Pattern Discovery and Pattern Analysis. Figure 3.1: The Process of Web Usage Mining The web log files on the web server are major source of data for Web Usage Mining. When user requests resources of web server, each request is recorded in the web log 29

file on web server. As a consequence, users browsing behavior is recorded into the web log file. In Web Usage Mining, data can be collected from server log files that include web server access logs and application server logs. Data is also obtained from site files and operational databases. The data collected in web log file is incomplete and not suitable for mining directly. Pre-processing is necessary to convert the data into suitable form for pattern discovery. Pre-processing can provide accurate, concise data for data mining. Data pre-processing, includes data cleaning, user identification, user sessions identification, path completion and data integration [45]. To discover the novel, potentially useful and interesting information, several methods and data mining algorithms are applied such as Path Analysis, Association Rules, Sequential Patterns, Clustering, Classification etc. Pattern analysis techniques are used to highlight overall patterns in data and to filter out uninteresting patterns. The techniques like Knowledge Query Mechanism, OLAP and visualization are used for pattern analysis. 3.3 Web Server Log The user navigates the pages of the website to access the information. The user agent or browser sends requests to the website server whenever new pages or resources are accessed by the user. The server receives and handles the request and data for the requested resource is sent back to the browser. The information related to that resource requests are recorded on the server in a file. This file is called an internet server log, or a web log. The log files keep a lot of information about each user s access to the web server. Each line or entry in the web log represents a request for a resource. The different web server support different log format. The exact content of an entry varies from log format to log format. Nearly all the server log file contains information such as Visitor s IP address, access date and time, URL of page accessed, the status code returned by the server, bytes transferred to the user etc. Many different web log formats like NCSA Common, Microsoft IIS Log, NCSA Combined etc. are in use. Almost every internet server platform has its own unique format, and some formats can in addition be customized by the server administrator. Figure 3.2 represents the sample web server log entries [51]. 30

Figure 3.2: Sample Web Server Log Entries Web servers can be configured to write different fields into the log file in different formats. The most common fields used by web servers are the followings: IP Address, Login Name, User Name, Timestamp, Request, Status, Bytes, Referrer, User agent. The explanation of fields included in sample web log entries of Figure 3.2 is as under. 1) IP Address: IP address of the remote host that made the request. If DNS host name is available then it will be substituted in place of IP address. 2) Login Name: Login name of the user making the HTTP request. If there is no value, a dash (-) is substituted instead of the login name. 3) User Name: The name of the authenticated user who access the server by making the HTTP request, anonymous users are represented by a dash ( ) 4) Timestamp: Time-stamp indicates date and time of the visit with the web server s time zone. 5) Request: It indicates HTTP request. It consists of three pieces of information: the HTTP method, the requested resource and HTTP protocol version. 31

6) Status Code: The success or failure of the http request is indicated by HTTP status code. For example, "200" is success, 404 for File Not Found, 500 for Internal Server Error. 7) Bytes transferred: It indicates number of bytes transferred to the user as part of the HTTP request. 8) Referrer: The URL of the page the visitor was on when he clicked to come to current page. 9) User agent: It represents the software used by the visitor to access the site. It could be a web browser, web robot or link checker etc. It also indicates the operating system used by the client. MSIE 7.0 means Internet Explorer 7. Windows NT 6.0 indicates Windows Vista. Windows NT6.1 indicates Windows7. A visitor request for web page can generate different web log records. The first record documents the request of user to particular web page and remaining records are created for requests made to obtain images on the requested page. The quantity of information stored in the log files is very large. The raw web log data is usually diverse and incomplete and difficult to be used directly for further pattern mining. The web log data can be preprocessed in order to obtain session information for all users and data reduction techniques can be applied to web log data in order to obtain raw sets for data mining. Before it can be used for anything at all, data must be extracted from the log file and analyzed. 3.4 Data Pre-Processing The data collected from web server log is often incomplete and creates uncertainty. The data pre-processing is necessary in order to clean, correct and complete input data and to mine the knowledge effectively. Pre-processing can provide accurate and concise data for data mining. Data pre-processing task includes data cleaning, user identification, user sessions identification, path completion and data integration [46]. In data pre-processing the server sessions are identified which represents the 32

information such as who accessed the website, what pages were requested and how long each page was viewed. 3.4.1 Parsing The purpose of the parsing is to read the log file and extract data that is relevant to finding traversal patterns. The parser does not modify the data in any way, other than to recombine some closely related data that are split up into smaller parts by some log formats. Some formats are self-documenting in that they start with a format description header. This can be used to auto-detect the log format. But many formats do not have such a header, and the syntax of those that do vary from format to format. Therefore sometimes user has to specify the format of the log before running the parser. All log formats that we are aware of can be described by constructing a text string containing field tags separated by one or more characters and it is called Log format description string as shown in Figure 3.3. We are discussing one approach here for parsing the web log. The field tags represent the various data fields in an entry in the log and are specified by <*>, where * is replaced by the name of the tag. For example <uri>, <host> and <day> are field tags. The separating characters can be whitespace, a comma and whitespace, a tabulator, whitespace and a square bracket, or any other combination of characters. Since we are only interested in parts of the information present in a web log, we create a <skip> tag to be used for fields that are of no interest to us, telling the parser not to bother interpreting the contents of this field. Figure 3.3: Log Format Description String For convenience we can define some of the log formats internally in the parser. So the user can specify one of these by simply feeding the parser the name of the format 33

instead of having to construct the entire format description string himself. The parser knows how to interpret the contents of the field represented by each individual tag. By combining tags into a format description string the parser will know how to parse the entire log. It does this by first splitting the description string into tokens of separators and tags. Every odd numbered token is a separator, and every even numbered token is a field tag. Then it moves through each log entry extracting whatever data is between two separators and depending on the corresponding field tag, either interprets and stores the data or discards it. We want to extract the pieces of information such as IPaddress or host name of the user, time of the request, requested URI, returned server status code from log file. Some of them are spread over multiple fields, most notably the time of the request, but sometimes also the requested URI. When interpreting, the parser recombines those fields into a single piece. If the parser is successful we end up with a list of web log entries where each element in the list corresponds to an entry in the web log file, only with irrelevant data removed. 3.4.2 Cleaning The items which are not related for usage analysis must be removed from the log files. When user requests to particular page from web server, various log entries are recorded. If page contains the images, videos, scripts, flash animations etc then resource requests for them will also be added in the log file. The objective of Web Usage Mining is to find users behaviour. So the entries for these resource requests do not make sense and must be removed from log file. Elimination of irrelevant items can be done by checking the suffix of the URL, which signifies in what format the kinds of files are. For example, the entries from log file with URL suffix jpg, gif, css, js, mov, avi, swf etc can be removed [49]. The list of suffixes can be changed depending on the type of site being analysed. It also involves removing requests where the server returned an error code. Erroneous log records are generated when access requests fail for various reasons and can be identified by HTTP status code. The Resulting status code may have different value, for example 200 is success, 404 File Not Found, 500 Internal Server Error. The entry can be removed if 34

server status code is not 200 something. We can consult at RFC 2616 for complete list [48]. In this way other types of requests can also be eliminated. Cleaning the data significantly reduces the number of log entries we have to work with and it is essential for finding usable access patterns. 3.4.3 User Identification and Session Generation After data cleaning, unique users must be identified. To identify the users, one simple method is to use login information, if users log in before using the web-site or system. Another approach is to use cookies for identifying the visitors of a web-site by storing a unique ID. However, these two methods are not general enough because they depend on the application domain and the quality of the source data. We can use a more general method to identify user. A new IP indicates a new user. The same IP but different user agent means a new user. The user agents are said to be different if it represents different web browsers or operating systems in terms of type and version. The list of log entries is sorted by the combination of IP addresses or host name of the user and the user agent [50]. The result is a list where all entries generated by the same user are clustered together and stored as separate log entry lists. Table 3.1: A List of Cleaned and Clustered Log Entries 35

Table 3.2: Generated Sessions Identification of the user sessions is very important because it constitutes a basic processing unit for discovery of interesting, prominent access patterns and so it largely affects the quality of pattern discovery result. The construction of the user activities in sessions for a particular Web site contains a correct mapping of activities to distinct web users. A long sequence of visits by the users has to be broken into user sessions. A user session is a set of pages visited by the same user within the duration of one particular visit to a web site. Each user session consists of only the pages visited by a user in a row. We consider web log data as a sequence of distinct web pages, where subsequence, such as user sessions can be observed by unusually long gaps between consecutive requests. A session is a collection of log entries generated by the page navigation of a single user within a certain time frame. User sessions are identified on the basis of the user identification. Because of the stateless connections between the client and the web server, Web log records have no visit session designations. Therefore, to reconstruct visit sessions from Web log records, we use the common timeout method, according to which a visit session is considered terminated when the time elapsed between page requests exceeds a 36

specified limit. A set of pages visited by a specific user is considered as a single user session if the pages are requested at a time interval not larger than a specified time period. Generally a time period of between 20 and 30 minutes is considered optimal. When the time interval exceeds the specified time it takes a person to read a web page or the user has left the browser to do other things. The timeout method is fairly effective and accurate for identifying visit sessions. According to Spiliopoulou, more than 90% of genuine visit sessions can be identified by the timeout method [46]. After generating the sessions by sorting the log entries on IP addresses and user agent, we check whether any sessions can be further split up by checking how much time has passed between two consecutive log entries. If the time is above a specific threshold value, the session is split into two sessions. Table 3.1 shows a list of cleaned and clustered log entries for sample and Table 3.2 shows sessions generated from it by considering timeout 30 minutes. The whole session set generated are used as raw data for pattern discovery phase. 3.4.4 Path Completion Due to existence of page cashing technology and proxy servers, some important accesses are not recorded into web server log. It results into missing access references to those pages that have been cashed. Users access patterns are not reflected accurately by incomplete access log. Path completion step is carried out to acquire the missing reference. The referrer information in server logs and knowledge of site topology can be used for effective path completion. If a page request is made that is not directly linked to the last page a user requested, the referrer log can be checked to see what page the request came from. If the page is in the user s recent request history, the assumption is that the user backtracked with the back button available on most browsers, calling up cached versions of the pages until a new page was requested. The site topology can also be used for the same, if referrer information is not clear. If more than one page in the user s history contains a link to the requested page, the page closest to the previously requested page is assumed as a source of the 37

new request [50]. In this way, the missing page references are found and added to the user sessions. 3.4.5 Data Integration In e-commerce, the operational databases contain user data, product, purchases etc. The user data provides demographic information about users and consists of registration and customer profile information. These data must be integrated with preprocessed data while applying Web Usage Mining in e-commerce in order to improve business intelligence [52]. 3.5 Pattern Discovery After identifying user sessions, the various techniques of web usage pattern discovery are applied in order to detect interesting and useful patterns. There are several kinds of access pattern mining that can be performed depending on the needs of the analyst. Some of pattern discovery techniques are discussed below. 3.5.1 Path Analysis Graph models are most commonly used for Path Analysis. In the graph models, a graph represents some relation defined on Web pages and each tree of the graph represents a web site. Each node in the tree represents a web page and edges between trees represent the links between web sites and the edges between nodes inside a same tree represent links between documents at a web site. When path analysis is used on the site as a whole, this information can offer valuable insights about navigational problems. Most graphs are involved in determining frequent traversal patterns and more frequently visited paths in a web site. For Example: What paths do users traversal before they go to a particular URL? Examples of information that can be discovered through path analysis are: 38

69% of clients who accessed /company/products/ order.asp by starting at /company and proceeding through /company/whatsnew.html, and /company/ products/sample.html ; 58% of clients left the site after four or less page references. The first rule tells us that 69% of visitors decided to make a purchase after seeing the sample of the products. The second rule indicates an attrition rate for the site. Since many users don t browse further than four pages into the site, it is tactful to ensure that most important information for example product sample, is contained within four pages of the common site entry points. 3.5.2 Association Rules Association rule generation can be used to relate pages that are most often referenced together in a single session. It predicts the association and correlation among set of items where the presence of one set of items in a transaction implies with a certain degree of confidence the presence of other items. That is, it can discover the correlations between pages that are most often referenced together in a single server session/user session. It can provide the information: What are the set of pages frequently accessed together by web users? What page will be fetched next? What are paths frequently accessed by web users? Implement association rules to on-line shopper can generally find out his/her spending habits on some related products [53]. For example, if a transaction of an on-line shopper consists of a set of items, while each item has a separate URL. Then the shopper s buying pattern will be recorded in the log file, and the knowledge mined from which, can be the form like the following: 35% of clients who accessed the web page with URL /company/products/bread.html, also accessed /company/ products/milk.htm. 39

50% of clients who accessed /company/announcements/ special.html, placed an online order in /company/ products/products1.html Aside from being applicable for business and marketing applications, the presence or absence of such rules can help Web designers to restructure their Web site. The association rules may also serve as a heuristic to fetch documents in order to reduce user-perceived latency when loading a page from a remote site. 3.5.3 Sequential Patterns Sequential patterns discovery is to find the inter-transaction patterns such that the presence of a set of items is followed by another item in the time-stamp ordered transaction set [54]. Web log files can record a set of transactions in time sequence. Using sequential pattern discovery, useful user trends can be discovered, predictions concerning visit pattern can be made, website navigation can be improved and adopt web site contents to individual client requirements or to provide clients with automatic recommendations that best suit customer profiles. By using this approach, Web marketers can predict future visit patterns which will be helpful in placing advertisements aimed at certain user groups. The sequential patterns can be discovered as the following form: 50% of client who bought items in /pcworld/computers/, also placed an order online in /pcworld/accessories/ within 15 days 3.5.4 Clustering Clustering is a technique to group together a set of items having similar characteristics. In the Web Usage domain, there are two kinds of interesting clusters that can be discovered: usage clusters and page clusters. Clustering of users tends to establish groups of users exhibiting similar browsing patterns. Such knowledge is 40

especially useful for inferring user demographics in order to perform market segmentation in E-commerce applications or provide personalized Web content to the users. On the other hand, clustering of pages will discover groups of pages having related content. This information is useful for Internet search engines and Web assistance providers. Clustering identifies visitors who share common characteristics. After you get the customers or visitors profiles, you can specify how many clusters to identify within a group of profiles, and then try to find the set of clusters that best represents the most profiles [54]. Besides information from Web log files, customer profiles often need to be obtained from an on-line survey form when the transaction occurs. For example, you may be asked to answer the questions like age, gender, email account, mailing address, hobbies, etc. Those data will be stored in the company s customer profile database, and will be used for future data mining purpose. An example of clustering could be: 50% of clients who applied discover platinum card in /discovercard/customerservice/newcard, were in the 25-30 age group, with annual income between $40,000 50,000. Clustering of client information can be used on the development and execution of future marketing strategies, online or offline, such as automated mailing campaign. 3.5.5 Decision Trees A decision tree is essentially a flow chart of questions or data points that ultimately leads to a decision [55]. For example, a car-buying decision tree might start by asking whether you want a 2008 or 2009 model year car, then ask what type of car, then ask whether you prefer power or economy, and so on. Ultimately it can determine what might be the best car for you. Decision trees systems are incorporated in productselection systems offered by many vendors. They are great for situations in which a visitor comes to a Web site with a particular need. But once the decision has been made, the answers to the questions contribute little to targeting or personalization of 41

that visitor in the future. Figure 3.4 represents the decision tree for purchase consideration as an example. Figure 3.4: The Decision Tree for Purchase Consideration 3.5.6 Filtering Simple reporting needs require only simple analysis systems. However, as the company s Web becomes more integrated with the other functionality of the company, for example, customer service, human resources, marketing activity, analysis need to rapidly expand. For example, the company launches a marketing campaign. Print and television ads now are designed to drive consumers to a Web site, rather than to call an 800 number or to visit a store. Consequently, tracking online marketing campaign results is no longer a minor issue but a major marketing concern. Often it is difficult to predict which variables are critical until considerable information has been captured and analysed. Consequently, a Web traffic analysis system should allow precise filtering and grouping information even after the data has been collected. Systems that force a company to predict which variables are important before capturing the data can lead to poor decisions because the data will be skewed toward the expected outcome. Filtering information allows a manager to answer specific questions about the site. For example, filters can be used to calculate how many visitors a site received this week from Microsoft. In this example, a filter is set for this week, and for visitors that have the word Microsoft in their domain 42

name e.g.proxy12.microsoft.com. This could be compared to overall traffic to determine what percentage of visitor s work for Microsoft. 3.5.7 Grouping Users usually can draw higher-level conclusions by grouping similar information. For example, grouping all Netscape browsers together and all Microsoft browsers together will show which browser is more popular on the site, regardless of minor versions. Similarly, grouping all referring URLs containing the word Yahoo shows how many visitors came from a Yahoo server. For example: http://search.yahoo.com/bin/search?p=web+miners 3.5.8 Dynamic Site Analysis Traditional Web sites were usually static HTML pages, often hand-crafted by Webmasters. Today, a number of companies, including Microsoft, make systems that allow an HTML file to be dynamically created around a database. This offers advantages like, included centralized storage, flexibility, and version control. But it also presents problems for some Web traffic analysis because the simple URLs normally seen on Web sites may be replaced by very long lines of parameters and cryptic ID numbers. In such systems, query strings typically are used to add critical data to the end of a URL usually delimited with a?. For example, the following referring URL is from google Search: https://www.google.co.in/#q=purchase+computer+online By looking at the data after the? we see that this visitor searched for purchase computer online on google before coming to our site. This information is encoded with a query parameter and separates each search keyword with the + character. In this example, purchase, "computer" and online each is referred to as parameter 43

values. By looking at this information, companies can tell what the visitor is looking for. This information can be used for altering a Web site to ensure that information visitors are looking for is readily available, and for purchasing keywords from search engines. 3.5.9 Cookies Cookies usually are randomly assigned IDs that a Web server gives to a Web browser the first time that the browser connects to a Web site. On subsequent visits, the Web browser sends the same ID back to the Web server, effectively telling the Web site that a specific user has returned. Cookies are independent of IP addresses, and work well on sites with a substantial number of visitors from ISPs. Authenticated usernames even more accurately identify individuals, but they require each user to enter a unique username and password, something that most Web sites are unwilling to mandate. Cookies benefit Web site developers by more easily identifying individual visitors, which results in a greater understanding of how the site is used. Cookies also benefit visitors by allowing Web sites to recognize repeat visits. For example, Amazon.com uses cookies to enable their one-click ordering system. Since Amazon already has your mailing address and credit card on file, you don t reenter this information, making the transaction faster and easier. The cookie does not contain this mailing or credit card information; that information typically was collected when the visitor entered it into a form on the Web site. The cookie merely confirms that the same computer is back during the next site visit. If a Web site uses cookies, information will appear in the cookie field of the log file, and can be used by Web traffic analysis software to do a better job of tracking repeat visitors. Unfortunately, cookies remain a misunderstood and controversial topic. A cookie is not an executable program, so it can t format your hard drive or steal private information. Modern browsers have the ability to turn cookie processing on or off, so users who chose not to accept them are accommodated. 44

3.6 Pattern Analysis The final step in Web Usage Mining process is Pattern analysis. Using pattern discovery, the pattern set is found and then pattern analysis is performed to select the interesting patterns and to filter out uninteresting patterns. The patterns are analyzed using techniques such as Data & Knowledge Querying, OLAP techniques and Usability analysis. Data & Knowledge querying mechanism uses a tool like SQL. In OLAP techniques, the result of pattern discovery is loaded into data cube and then OLAP operations are performed. After this, to interpret the results, visualization techniques such as graphing patterns or assigning colors to different values are used to often highlight overall patterns or trends in the data [47]. The content and structure information can be used to filter out patterns containing pages of a certain usage type, content type, or pages that match a certain hyperlink structure. The result of pattern analysis helps e-commerce in following ways: Improving the design of e-commerce web site according to user s browsing behavior on site in order to better serve the needs of users. Personalizing web sites according to individual s interest and make dynamically changing particular web site for visitor. Developing a security system that can detect the intrusion and to restrict the user s access to certain online contents. Understanding customers need and retaining them by providing customized products, improving satisfaction with help of tracking browsing behavior in e- commerce. Evaluating the effectiveness of advertising by analyzing large number of consumer behavior patterns. 45

3.7 Conclusion Web usage mining is the technique to find useful and interesting information from web usage data. It is useful in e-commerce to improve structure of web site, personalization, identifying customer behavior, Business Intelligence. The web usage data includes data from web server logs, proxy server logs, browser logs etc. In this chapter, we have described the process of web usage mining which consist of steps: Data Collection, Pre-Processing, Pattern Discovery and Pattern Analysis. In data preprocessing step, sequence of tasks needs to be performed such as: Parsing log file, Cleaning, User identification and session generation, Path Completion. In e- commerce, the operational databases store user, product, purchase data. These data must be integrated with pre-processed data while applying web usage mining in e- commerce in order to improve business intelligence. After identifying user sessions, different techniques of web usage pattern discovery are applied such as path analysis, association rules, sequential patterns, clustering, decision trees, dynamic site analysis etc. The pattern analysis is performed to select the interesting patterns and to filter out uninteresting patterns. The patterns are analyzed using techniques such as Data & Knowledge Querying, OLAP techniques and Usability analysis. 46