Graph Structure Over Time

Size: px
Start display at page:

Download "Graph Structure Over Time"

Transcription

1 Graph Structure Over Time Observing how time alters the structure of the IEEE data set Priti Kumar Computer Science Rensselaer Polytechnic Institute Troy, NY Abstract This paper examines the IEEE data set and looks at how different time stamps of the graph differ in terms of structure. It also looks at how subsets of the graph differ from each other. I. INTRODUCTION A social network is a very special type of graph that represents information of how people interact and relate to one another. Because social networks are based on people, the structural properties of such a graph differ greatly from the anticipated structure of a graph that is generated randomly. Human actors in this graph have more logical decisions on who will be connected to them. This paper explores the properties of a graph based on a data set of papers in the IEEE database from the years of 1990 to The paper also examines how different subdivisions of the graph relate to each other. Since the properties are not what would be expected of a random graph, looking at portions of the graph allows us to delve deeper into the reasons for why this might happen and what leads to it. The graph is looked at as it evolves by adding new papers, and by looking at only the new papers added. The consistency of certain unusual properties is observed, in relation to the graph of the entire data set and graphs of other divisions of the data set. These questions of the structure of the graph and how constant they are allow for a deeper understanding of what a social network is. The IEEE data set depicts an even more specialized graph than a general social network. It shows a network specifically based on work done and papers published. The way in which authors connect is significantly different from how nodes of a random graph would connect, and from how people connect in other areas of their lives. This data set provides a specific case that makes these questions even more interesting. II. ORGANIZING THE DATA SET The first step to answering these questions was to put all the information in the data set in a graph. This would provide a better method of analyzing the data. The given data was set up by looking at the paper and when it was written. The given information was the identification number of the paper, the year in which it was written, and the authors who collaborated on it. The information given covered papers written between the years 1990 and To organize the data, all information was represented in a graph. Each author represented a node on the graph. Two nodes are connected if the authors ever worked on a paper together. That means that this is an unweighted graph. The reasoning for why authors are only required to write a single paper together will be discussed in later sections. The division of the graph into subsets is a crucial part of this paper. The overall data set is divided into years. This information is then looked at in two ways. The first of these is a cumulative graph. This is where each year s data includes the data from that year and all previous years. So, when discussing the 1997 cumulative graph, the graph contains all papers from 1990 up to and including This data set is used to look at how additional information changes the structure of the graph. The other method of analysis is on a year-by-year basis. This is where each year s data only contains papers from that year. When referring to the 1997 separated graph, this means that the graph has only the information from the year These two methods are necessary to be able to observe the trends as they relate to the growing graph, and the incoming information being added. To properly understand the developments in the graph, information on what new data is being added is necessary. This

2 organizational structure for the graph provides us with that. III. INCOMING INFORMATION When observing the effects of time on the graph, it is crucial to look at how the information that is being added at each time step relates to the graph and the information from other time steps. Changes in the inputs can help explain how and why certain properties of the overall graph hold true. Some trends in graph of the entire data set will be directly influenced by the incoming information, such as the overall size of the graph. The data set is organized by paper, not by author as the graph is. The first aspect to analyze, then, becomes the number of papers written by year. Looking at the two graph sets, it is clear that the cumulative set will continually increase. Each year adds more papers, and all these new papers will be added to the graph that already contains all previous papers. The separated data set, however, has no specific reasoning to actually show an increase. However, as shown in Chart 1 below, there is a steady rise in the number of papers added each year. The number of papers written in the year 1990 is 1,367. By the year 2000, there are 4,534 papers written, showing a 330% increase in the number of papers. This trend continues, with 8,435 papers from the year This shows a 617% rise from the year 1990, and a 186% increase from the year The number of papers added does appear to reach a small peak at year 2008, where the number of papers is 9,477, but there is not enough information to confirm whether or not the slight decrease that follows is a continuing trend or an anomaly. The year 2007 has 8,789 papers and the year 2009 has 8,810 papers, which shows that 2008 is a peak for that local area. The general trend, however, does indicate a rising number of papers per year. Chart 1: Number of papers by year Chart 2: Number of authors per year In addition to the number of papers written per year, the number of authors added to the data set per year has to be looked at. If the number of papers increases, but the number of total authors remains relatively constant, that would mean that the same authors are continually writing papers. In Chart 2, we see that the number of authors in a specific year shows a trend similar to that of the number of papers per year that are added. For the separated data set, the number of authors added correlates directly to the number of papers added. As with the number of papers added, the number of authors increases significantly until In 1990, the graph starts with 3,331 authors represented in the 1,367 papers. In the year 2000, 11,325 authors wrote papers that were added to the data set. This is an increase of around 340% from the year By 2010, the data set holds 23,827 authors that wrote papers to be added, a 210% rise from 2000 and 715% increase from As shown with the number of papers, the numbers rise until After this there is a slight decrease in the number of authors added. Again, we see the possibility of a new trend, but have no data to properly analyze this effect. As to the cumulative graph, the number of authors increases, but is not just the sum of the authors for each year. Authors that have papers in more than one year are only represented in the cumulative graph once. Still, the rate of increase is similar to the rate of increase of the number of papers. Using this, the average number of authors per paper can be examined. A higher number of authors per paper implies that there are either papers with large numbers of authors, authors that write fewer papers per year, or a combination of the two. A lower number implies that there authors are publishing papers alone or that authors are writing more papers per year. An interesting distinction occurs with this property of the graph in the two representations of the data. The data

3 set that shows an increase in the average for the separated years. This change is not very large, going from in 1990 to in However, the cumulative graph shows a decrease in the average. The final value for the average is in 2010, down from in This information means that new authors are being added to the graph and that authors do get added to the data set at a significant rate. IV. PERCENT OF CONNECTIONS To look at the IEEE data set fully, the amount of connections compared to the number of possible connections should be examined. This gives us a general idea of how hard it is to actually connect to another node in the graph. The calculation for this was done by taking the total number of edges formed in the graph, and dividing that by the number of possible edges. The number of possible edges is the number of nodes multiplied by the number of nodes minus one. This allows for every node to connect to every other node. The quotient of the division is the percent of connections that are actually formed, and for this purpose, also the probability of connections forming. When examining the data set, it is clear that the percent of connections actually formed is fairly low. The value starts at for the separated data set of This value increases for the following few years, but the peak is still low at in The final value of this for the year 2010 is A similar trend to the number of papers and the number of authors per year is shown. Again it is seen that the year 2008 causes a difference in the data. This time, however, the year brings a low to the graph, with a percentage of The value of the previous year is , and the value for the next year is Because the value for 2009 is greater than the value for 2007, this shows more evidence that is possible for 2008 to be an anomaly, but that there is still a slight increase caused. Still, it could easily by a small anomaly, but more data would be needed to confirm anything. The percent of connections decreases so significantly over the years shows that many more authors were either working alone or with the same group of people instead of new people. The cumulative graph gives a slightly different story than the separated graphs. The start value is the same of for the year However, since new nodes are being added to the graph at the same time of old nodes being present, the overall probability decreases significantly compared to the decreased n the separate data set graph. There is no peak in the year 1991, as there was in the separated graph. The percentage of connections just continually declines. The rate of decrease also shows a great deal, as it starts off as being fairly high. The change rate then decreases, as the graph appears to basically level off. Again, there is a lack of data to fully support this idea. The cumulative graph shows much lower values than the separated years graph does. The start value for both is the same, as they both include only the 1990 papers of the data set. The rate of decrease for the cumulative graph is much greater, however. This is intuitively on par, though. Since more nodes are being added, but not all the new nodes are connecting to any old nodes, and not all the old nodes are connecting to any new nodes. While papers are added, the increase in potential connections is too great to overcome the gap and maintain a higher percent of connections formed. This shows that many of the newer nodes are not making as Chart 3: Percent of Connections formed in separate years data set Chart 4: Percent of Connections formed in cumulative years data set

4 many connections to older nodes, as well as not having as many potential connections realized in their own years. By the year 2010, the percent of connections in the cumulative graph is , much lower than the separate year s value of V. DEGREE DISTRIBUTION Another crucial part of the data is the degree distribution of the data. The degree of a node is the values of edges that go from a node to a different node in the graph. Because this is an unweighted, undirected graph, the degree is simply the number of edges that the node has. Another option for this graph would have been to use each paper written by an author as an edge. If an author wrote three papers with the same co-author, the weight of that edge would be three. However, that information did not seem valuable for the question discussed in this paper. The degree distribution shows how authors collaborate with their peers. With this, it can be seen whether people to tend with larger groups of people or smaller. Degree distributions that show higher degrees being more represented than lower degrees would show that most authors prefer to work with a large variety of people. The shape of the curve for a degree distribution graph would also tell us a significant amount. This would tell us what general trend this follows and how rapidly extreme differences are. Part of the question for this is comparing the actual structure of the graph to what should be expected of a random graph with some similar attributes. To examine this, an Erdos-Renyi graph was constructed to match each time stamp of the graph. The random graph was created by entering the number of nodes, n, and the probability of connecting, p, to a simulator. This p value is calculated by using the probability of connecting for the given time step as discussed earlier. The simulator created n nodes. Each node attempted to connect to all other nodes. Based on probability p, the node would connect. This means that there was p chance of connecting, and (1-p) chance of not connecting. A random number between zero and one was generated to try this connection. The random graph gave a baseline expectation of what the degree distribution should look like. We see that the expected graphs show more of a bell curve that has a peak towards the middle. Graphs were generated for each time step. Each of the generated graphs showed a similar bell shape curve for the expected degree distribution. This says that most nodes will have a degree somewhere around the middle of the distribution. For a randomly generated graph, this is the instinctive trend that the graph should follow. There is no specific reason to not connect to other nodes. If the probability matches up, connect. All the generated graphs show the trend, so it can be safely assumed that this is the expected trend for any graph of this type to follow. Chart 5 shows an example of what the graph looks like based on the data of This means the number of nodes for 1993 as a singleton year was used (4,264 authors with a probability of connecting). Chart 6 shows an example of the graph based on 1993 as a cumulative data set (11,262 authors and chance of connecting). While the expected graph shows that nodes connect with a decent number of other nodes in the graph, the actual data set does not show this same idea at all. A significant number of nodes have a low degree of only one or two. This means that all papers they have worked on are with the same group of people. Comparing this to the expected graph, the results are extremely confusing. Why should the difference be so great? However, thinking of this as a social network answers that to some degree. Remembering that this data set is based on papers written together answers another portion of that question. Because this is a social network, it can t be expected for decisions to be made randomly. The random graph allows any node to connect to any other node. A graph based on real people under real circumstance would not create connections like that. Writing a paper is a commitment of time and work. In order for two authors to actually agree to work on a paper together, many preconditions must be met. For two people to collaborate on a paper there is usually some form of existing connection between the two. Very often, it will be people working at the same university or research institution. This limits the possible connections a great deal. Two authors having no previous connection is rare in this case. Another limitation could be physical distance. Depending on the authors, they might not want to work with a collaborator they could not meet in person. A further limiter of connections is the field in which authors are working. While some papers will span multiple fields, many papers require special knowledge of an area. All these explain why the results shown in Charts 7 and 8 are very different from the expected results shown in Charts 5 and 6. Chart 7 shows the singleton year of 1993 s degree distribution. Chart 8 shows the degree distribution of the cumulative data set from

5 Number of nodes Number of nodes Number of nodes Number of nodes Expected Distribution Separated Degree Degree Chart 5 Chart Expected Distribution Cumulative Actual Distribution Separated Degree Actual Distribution Cumulative Degree Chart 7 Chart 8

6 VI. CLUSTREING COEFFICIENT Along with degree distribution, the clustering coefficient is an important statistic about the data set. The clustering coefficient is a method of analyzing how the graph is structured in terms of what nodes are connected to what other nodes. The coefficient looks at a node s neighbors, and sees if the neighbors are connected. If two neighbors are connected, a triangle is formed with the original node and the two connected neighbors. This value is calculated for all the nodes in the graph. The total is then divided by the number of nodes in the graph. This produces a number between zero and one. A complete graph would have a coefficient of one. This statistic shows us whether authors work within groups or with people who do not know each other. Random graphs were used to determine the anticipated clustering coefficient for a graph with this probability of connecting and number of nodes. The Eros-Renyi model for a random graph was used once again, and the expected clustering coefficient was determined by looking at the clustering coefficient for that graph. The expected clustering coefficient correlates directly to the probability of connecting discussed earlier. If the probability entered is , the expected clustering coefficient is the same. This means that with a random graph, the change of forming a triangle of the sort mentioned in the previous paragraph is related to the probability of connecting two nodes in the first place. As with several other properties examined, the value of the clustering coefficient increases with time in this data set. This is related to the idea that the more authors are being added in relation to the number of papers being added. If there is a higher number of authors per paper, the clustering coefficient is going to rise. The charts below (Charts 9 and 10) give clear Chart 9: Clustering Coefficient by Separate Years Chart 10: Clustering Coefficient by Cumulative Years indicators that the clustering coefficient is almost constantly rising (there is a slight anomaly, but is corrected quickly). Chart 9 shows the clustering coefficient for each year. Chart 10 shows the same information for the cumulative data set. While the clustering coefficient of the separated graph rises by 0.19, the cumulative graph rises by only This is a factor of new nodes being added, and old nodes no longer being active. Because of this effect, the clustering coefficient will not rise at the same rate as the separated years data set. In the random graph, the nodes are always active, with no consideration of the time involved. The separated data set also has this, since each year is stand-alone. The anomaly on this graph is the year 1992, which shows a decrease. However, this proves to be only an anomaly, not a trend, because the coefficient increases again after that year. The actual values of the clustering coefficient are wildly different from the expected values. As with the degree distribution, the expected values set up a scenario where random connection greatly changes the situation. The observed values are much, much higher than the anticipated values. The table below (Table 1) shows how these numbers vary. The first column simply lists the years represented. The second column gives the expected clustering coefficient. This is calculated by using the probability of connecting as discussed earlier. For the purpose of this comparison, the actual probability is used. The last column shows the actual clustering coefficient for the graph that represents the years discussed. It is shown that the actual values are higher than the expected values. The actual coefficient for the year 1990 is more than six hundred times the expected value. This gap rises as the probability of connecting decreases, but the clustering coefficient increases.

7 Years Expected Actual Coefficient Included Separated Cumulative E E E E E E E E E E E Table 1 The differences in this information versus the randomized graphs go back to what was discussed with the degree distribution. There are many more limitations in this data set than are represented here and can be shown by a randomized graph. If authors actually are working with others in their work place, it can be assumed that two collaborators of a single author are much more likely to connect than in a randomized graph. Field of work also plays a big role, as authors will look for collaborators in a smaller set than the entire graph. That set will most likely contain previous collaborators, and their collaborators. Because of this, more triangles will be formed, increasing the clustering coefficient. Another aspect of this idea is going through a collaborator to find a new partner. An author looking to write another paper might ask a connection for a recommendation on whom to work with. It is likely that your co-author on one paper will be able to mention another author or co-worker that is participating in a similar field. This effect produces more triangles, and is based on the idea of trust. Having a good experience with a co-author will increase the likelihood of working with them again, but also the likelihood of trusting their opinions on others. When looking for a partner, factors come into play that negates the idea of randomization. Aside from non-random factors that play a big role, there is also the issue of authors who only collaborate on one paper. These authors will increase the clustering coefficient of the entire graph. Each of these authors will have a clustering coefficient of one. This is because all their co-authors have also worked with each other. For example, suppose author A worked on only one paper in the data set, and worked with authors B, C, and D on that paper. There are three possible triangles for A to be included in: A-B-C, A-B-D, and A-C-D. All of these triangles have been formed, because all of these authors are connected through the same paper that let A create connections with all of them. Table 2 shows the percentage of authors with only one paper for the data set. This is shown for the singleton data sets as well as the combined data sets. In addition, the percentage of authors with two papers is shown for comparison s sake. As seen in the table, the percentage of authors with only one paper in their name is incredibly high. The lowest value for this is 83% for a singleton year and 68% for the cumulative data set. Most of these add to the clustering coefficient, although a few did not work with enough collaborators to contribute to the coefficient. The percent of authors with only two papers is also fairly high, and their clustering coefficients are also likely to be high, adding to the overall clustering coefficient.

8 Years Included Number of Authors Number of Authors with One Paper Percent of Authors with One Paper Number of Authors with Two Papers Percent of Authors with Two Papers Separated Cumulative Table 2

9 VII. CONCLUSIONS The IEEE data set shows clear distinctions from a random data set of this size. The degree distributions shown in this graph resemble the projected degree distributions in almost no way. There is a much higher number of nodes with small degrees than the randomized graph would predict. The randomized graph depicts a situation in which most nodes have a higher number of connections. The reality is that situations arise to limit the number of authors an author can work with. The clustering coefficient also demonstrates the effects that non-randomization has on the graph. The significantly higher values of the coefficient for the actual graph than the anticipated values for the randomized graph show that there is clearly social networking occurring in this data set. It is shown that there must be some reasoning for choosing co-authors beyond just randomization. A large part of this is that most authors added to the data set have only a single paper. This unique aspect does not occur in the randomized graph and changes the results by extreme amounts. There are several reasons for the data to be skewed from the estimated graphs, most of which lie in the fact that choosing a co-author is not random, while the connections of nodes in the random graph have no constraints on the connections besides chance. Coauthors are chosen based on a variety of reasons. Location is crucial to meeting potential co-authors, and can cause more triangles between authors. Also, papers actually take time to write. This seems obvious, but also provides insight as to why the degree distribution shows lower numbers of connections. An author will most likely not be able to write 20 papers in a year, which limits how many authors they can actually work with within a given year. The amount of effort it takes to actually form a connection in the IEEE data set is much greater than the effort required for the random data set. In addition, choosing a random co-author can lead to very negative results (such as wasted time, a frustrating experience, etc.). Adding another edge in the random graph has no negative repercussions. The more surprising result was the change in statistics that were represented in the singleton years data set. Most of the changes are presented because of the change in the number of papers added. That statistic, however, seems to increase on a year by year basis with no reasoning. This might be due to things that can t be represented on a graph. Things like availability of resources on a general scale, increase in the popularity of IEEE, increase in the fields that IEEE caters to, and general trends that are not represented would help cause this. The increase in the number of authors added can be seen as a direct function of the increase in the number of papers added. It is more likely that new authors will be added, as there is only a certain amount of work any author can do in a year. The IEEE data set shows clear trends as a factor of time. The increase in the number of papers added does have some effect on the other trends, such as the increase in the number of authors as a factor of time. This is also a reason for the percent of connections formed to be limited. It is much more likely that one author will work with a specific group of nodes and work within that than to explore the graph and work with as many people as possible. That idea leads to the high clustering coefficient and the high degree distribution. All these factors are inter-related, and can stem from the nature of writing a paper as a timeconsuming task that takes much more consideration than a random connection. VIII. FUTURE WORK The work done in this paper is a stepping stone to many other possibilities. Analyzing the graph in terms of structure and as a relation to time provides a higher platform for discovering other questions and some of the data needed to explore it. The next step would be to cluster all the authors and determine how the graph is actually clustered. The high clustering coefficient implies that there will be many very small clusters. This is instead of the possibility of few, but large clusters. The clustering found for this would be based on what algorithms and conditions are used. A seemingly meaningless clustering of only a few people might be very meaningful in this data set, though. The clusters would be based on strong bonds, or common neighbors. The data and algorithms would have to be implemented and tinkered with to determine the specifics. After determining the clusters, it would be interesting to look at physical location and affiliations of the authors in a cluster. It is possible that most clusters form from within a single affiliated university or other form of research establishment. This would require a great deal more of information, though. The results might show what factors authors actually take into account when writing a paper and choosing a coauthor. This data set has many more possibilities, but these two would provide very interesting information especially if structured in the correct manor.

How App Ratings and Reviews Impact Rank on Google Play and the App Store

How App Ratings and Reviews Impact Rank on Google Play and the App Store APP STORE OPTIMIZATION MASTERCLASS How App Ratings and Reviews Impact Rank on Google Play and the App Store BIG APPS GET BIG RATINGS 13,927 AVERAGE NUMBER OF RATINGS FOR TOP-RATED IOS APPS 196,833 AVERAGE

More information

Marketing Performance in Executive perspective on the strategy and effectiveness of marketing

Marketing Performance in Executive perspective on the strategy and effectiveness of  marketing Executive perspective on the strategy and effectiveness of email marketing Table of Contents Introduction 2 The state of email marketing performance 3 Successfully achieving objectives 4 Most important

More information

Chapter 5. Track Geometry Data Analysis

Chapter 5. Track Geometry Data Analysis Chapter Track Geometry Data Analysis This chapter explains how and why the data collected for the track geometry was manipulated. The results of these studies in the time and frequency domain are addressed.

More information

1 Homophily and assortative mixing

1 Homophily and assortative mixing 1 Homophily and assortative mixing Networks, and particularly social networks, often exhibit a property called homophily or assortative mixing, which simply means that the attributes of vertices correlate

More information

Predicting Messaging Response Time in a Long Distance Relationship

Predicting Messaging Response Time in a Long Distance Relationship Predicting Messaging Response Time in a Long Distance Relationship Meng-Chen Shieh m3shieh@ucsd.edu I. Introduction The key to any successful relationship is communication, especially during times when

More information

Question #1: 1. The assigned readings use the phrase "Database Approach." In your own words, what is the essence of a database approach?

Question #1: 1. The assigned readings use the phrase Database Approach. In your own words, what is the essence of a database approach? Question #1: 1. The assigned readings use the phrase "Database Approach." In your own words, what is the essence of a database approach? Respondant: I understand the essence of "database approach" to be

More information

Combinatorics Prof. Dr. L. Sunil Chandran Department of Computer Science and Automation Indian Institute of Science, Bangalore

Combinatorics Prof. Dr. L. Sunil Chandran Department of Computer Science and Automation Indian Institute of Science, Bangalore Combinatorics Prof. Dr. L. Sunil Chandran Department of Computer Science and Automation Indian Institute of Science, Bangalore Lecture - 5 Elementary concepts and basic counting principles So, welcome

More information

3 Graphical Displays of Data

3 Graphical Displays of Data 3 Graphical Displays of Data Reading: SW Chapter 2, Sections 1-6 Summarizing and Displaying Qualitative Data The data below are from a study of thyroid cancer, using NMTR data. The investigators looked

More information

CS224W: Social and Information Network Analysis Project Report: Edge Detection in Review Networks

CS224W: Social and Information Network Analysis Project Report: Edge Detection in Review Networks CS224W: Social and Information Network Analysis Project Report: Edge Detection in Review Networks Archana Sulebele, Usha Prabhu, William Yang (Group 29) Keywords: Link Prediction, Review Networks, Adamic/Adar,

More information

8: Statistics. Populations and Samples. Histograms and Frequency Polygons. Page 1 of 10

8: Statistics. Populations and Samples. Histograms and Frequency Polygons. Page 1 of 10 8: Statistics Statistics: Method of collecting, organizing, analyzing, and interpreting data, as well as drawing conclusions based on the data. Methodology is divided into two main areas. Descriptive Statistics:

More information

Descriptive Statistics Descriptive statistics & pictorial representations of experimental data.

Descriptive Statistics Descriptive statistics & pictorial representations of experimental data. Psychology 312: Lecture 7 Descriptive Statistics Slide #1 Descriptive Statistics Descriptive statistics & pictorial representations of experimental data. In this lecture we will discuss descriptive statistics.

More information

Chapter 2 Basic Structure of High-Dimensional Spaces

Chapter 2 Basic Structure of High-Dimensional Spaces Chapter 2 Basic Structure of High-Dimensional Spaces Data is naturally represented geometrically by associating each record with a point in the space spanned by the attributes. This idea, although simple,

More information

Hardware versus software

Hardware versus software Logic 1 Hardware versus software 2 In hardware such as chip design or architecture, designs are usually proven to be correct using proof tools In software, a program is very rarely proved correct Why?

More information

Chapter 10. Fundamental Network Algorithms. M. E. J. Newman. May 6, M. E. J. Newman Chapter 10 May 6, / 33

Chapter 10. Fundamental Network Algorithms. M. E. J. Newman. May 6, M. E. J. Newman Chapter 10 May 6, / 33 Chapter 10 Fundamental Network Algorithms M. E. J. Newman May 6, 2015 M. E. J. Newman Chapter 10 May 6, 2015 1 / 33 Table of Contents 1 Algorithms for Degrees and Degree Distributions Degree-Degree Correlation

More information

The future of UC&C on mobile

The future of UC&C on mobile SURVEY REPORT The future of UC&C on mobile Published by 2018 Introduction The future of UC&C on mobile report gives us insight into how operators and manufacturers around the world rate their unified communication

More information

CICS insights from IT professionals revealed

CICS insights from IT professionals revealed CICS insights from IT professionals revealed A CICS survey analysis report from: IBM, CICS, and z/os are registered trademarks of International Business Machines Corporation in the United States, other

More information

DESIGNING RESPONSIVE DASHBOARDS. Best Practices for Building Responsive Analytic Applications

DESIGNING RESPONSIVE DASHBOARDS. Best Practices for Building Responsive Analytic Applications DESIGNING RESPONSIVE DASHBOARDS Best Practices for Building Responsive Analytic Applications CONTENTS Responsive Design and Embedded Analytics...1 6 Principles and Best Practices... 2 Tools of the Trade...

More information

Yunfeng Zhang 1, Huan Wang 2, Jie Zhu 1 1 Computer Science & Engineering Department, North China Institute of Aerospace

Yunfeng Zhang 1, Huan Wang 2, Jie Zhu 1 1 Computer Science & Engineering Department, North China Institute of Aerospace [Type text] [Type text] [Type text] ISSN : 0974-7435 Volume 10 Issue 20 BioTechnology 2014 An Indian Journal FULL PAPER BTAIJ, 10(20), 2014 [12526-12531] Exploration on the data mining system construction

More information

3 Graphical Displays of Data

3 Graphical Displays of Data 3 Graphical Displays of Data Reading: SW Chapter 2, Sections 1-6 Summarizing and Displaying Qualitative Data The data below are from a study of thyroid cancer, using NMTR data. The investigators looked

More information

MAT 110 WORKSHOP. Updated Fall 2018

MAT 110 WORKSHOP. Updated Fall 2018 MAT 110 WORKSHOP Updated Fall 2018 UNIT 3: STATISTICS Introduction Choosing a Sample Simple Random Sample: a set of individuals from the population chosen in a way that every individual has an equal chance

More information

Model-Based Design for Large High Integrity Systems: A Discussion Regarding Model Architecture

Model-Based Design for Large High Integrity Systems: A Discussion Regarding Model Architecture Model-Based Design for Large High Integrity Systems: A Discussion Regarding Model Architecture By Mike Anthony and Jon Friedman MathWorks Inc, Natick, MA, 01760 INTRODUCTION From complex controls problems

More information

Research. Eurex NTA Timings 06 June Dennis Lohfert.

Research. Eurex NTA Timings 06 June Dennis Lohfert. Research Eurex NTA Timings 06 June 2013 Dennis Lohfert www.ion.fm 1 Introduction Eurex introduced a new trading platform that represents a radical departure from its previous platform based on OpenVMS

More information

Your Name: Section: INTRODUCTION TO STATISTICAL REASONING Computer Lab #4 Scatterplots and Regression

Your Name: Section: INTRODUCTION TO STATISTICAL REASONING Computer Lab #4 Scatterplots and Regression Your Name: Section: 36-201 INTRODUCTION TO STATISTICAL REASONING Computer Lab #4 Scatterplots and Regression Objectives: 1. To learn how to interpret scatterplots. Specifically you will investigate, using

More information

CSC 2515 Introduction to Machine Learning Assignment 2

CSC 2515 Introduction to Machine Learning Assignment 2 CSC 2515 Introduction to Machine Learning Assignment 2 Zhongtian Qiu(1002274530) Problem 1 See attached scan files for question 1. 2. Neural Network 2.1 Examine the statistics and plots of training error

More information

Whitepaper US SEO Ranking Factors 2012

Whitepaper US SEO Ranking Factors 2012 Whitepaper US SEO Ranking Factors 2012 Authors: Marcus Tober, Sebastian Weber Searchmetrics Inc. 1115 Broadway 12th Floor, Room 1213 New York, NY 10010 Phone: 1 866-411-9494 E-Mail: sales-us@searchmetrics.com

More information

Character Recognition

Character Recognition Character Recognition 5.1 INTRODUCTION Recognition is one of the important steps in image processing. There are different methods such as Histogram method, Hough transformation, Neural computing approaches

More information

You will begin by exploring the locations of the long term care facilities in Massachusetts using descriptive statistics.

You will begin by exploring the locations of the long term care facilities in Massachusetts using descriptive statistics. Getting Started 1. Create a folder on the desktop and call it your last name. 2. Copy and paste the data you will need to your folder from the folder specified by the instructor. Exercise 1: Explore the

More information

Robustness of Centrality Measures for Small-World Networks Containing Systematic Error

Robustness of Centrality Measures for Small-World Networks Containing Systematic Error Robustness of Centrality Measures for Small-World Networks Containing Systematic Error Amanda Lannie Analytical Systems Branch, Air Force Research Laboratory, NY, USA Abstract Social network analysis is

More information

BBS654 Data Mining. Pinar Duygulu. Slides are adapted from Nazli Ikizler

BBS654 Data Mining. Pinar Duygulu. Slides are adapted from Nazli Ikizler BBS654 Data Mining Pinar Duygulu Slides are adapted from Nazli Ikizler 1 Classification Classification systems: Supervised learning Make a rational prediction given evidence There are several methods for

More information

DOWNLOAD PDF BIG IDEAS MATH VERTICAL SHRINK OF A PARABOLA

DOWNLOAD PDF BIG IDEAS MATH VERTICAL SHRINK OF A PARABOLA Chapter 1 : BioMath: Transformation of Graphs Use the results in part (a) to identify the vertex of the parabola. c. Find a vertical line on your graph paper so that when you fold the paper, the left portion

More information

Chapter 2: Understanding Data Distributions with Tables and Graphs

Chapter 2: Understanding Data Distributions with Tables and Graphs Test Bank Chapter 2: Understanding Data with Tables and Graphs Multiple Choice 1. Which of the following would best depict nominal level data? a. pie chart b. line graph c. histogram d. polygon Ans: A

More information

Chapter 2 - Graphical Summaries of Data

Chapter 2 - Graphical Summaries of Data Chapter 2 - Graphical Summaries of Data Data recorded in the sequence in which they are collected and before they are processed or ranked are called raw data. Raw data is often difficult to make sense

More information

Chapter 2: The Normal Distribution

Chapter 2: The Normal Distribution Chapter 2: The Normal Distribution 2.1 Density Curves and the Normal Distributions 2.2 Standard Normal Calculations 1 2 Histogram for Strength of Yarn Bobbins 15.60 16.10 16.60 17.10 17.60 18.10 18.60

More information

Further Maths Notes. Common Mistakes. Read the bold words in the exam! Always check data entry. Write equations in terms of variables

Further Maths Notes. Common Mistakes. Read the bold words in the exam! Always check data entry. Write equations in terms of variables Further Maths Notes Common Mistakes Read the bold words in the exam! Always check data entry Remember to interpret data with the multipliers specified (e.g. in thousands) Write equations in terms of variables

More information

Design and Analysis of Algorithms Prof. Madhavan Mukund Chennai Mathematical Institute. Module 02 Lecture - 45 Memoization

Design and Analysis of Algorithms Prof. Madhavan Mukund Chennai Mathematical Institute. Module 02 Lecture - 45 Memoization Design and Analysis of Algorithms Prof. Madhavan Mukund Chennai Mathematical Institute Module 02 Lecture - 45 Memoization Let us continue our discussion of inductive definitions. (Refer Slide Time: 00:05)

More information

Building Better Parametric Cost Models

Building Better Parametric Cost Models Building Better Parametric Cost Models Based on the PMI PMBOK Guide Fourth Edition 37 IPDI has been reviewed and approved as a provider of project management training by the Project Management Institute

More information

4 KEY FACTORS FOR DATA QUALITY ON A DATA LAKE (OR: HOW TO AVOID THE DATA SWAMP) JOSH HERRITZ MIOSOFT CORPORATION MIOsoft Corporation.

4 KEY FACTORS FOR DATA QUALITY ON A DATA LAKE (OR: HOW TO AVOID THE DATA SWAMP) JOSH HERRITZ MIOSOFT CORPORATION MIOsoft Corporation. 4 KEY FACTORS FOR DATA QUALITY ON A DATA LAKE (OR: HOW TO AVOID THE DATA SWAMP) JOSH HERRITZ MIOSOFT CORPORATION The trends in digital business promise that the future holds an unprecedented volume, variety,

More information

Snowflake Numbers. A look at the Collatz Conjecture in a recreational manner. By Sir Charles W. Shults III

Snowflake Numbers. A look at the Collatz Conjecture in a recreational manner. By Sir Charles W. Shults III Snowflake Numbers A look at the Collatz Conjecture in a recreational manner By Sir Charles W. Shults III For many people, mathematics is something dry and of little interest. I ran across the mention of

More information

Week - 03 Lecture - 18 Recursion. For the last lecture of this week, we will look at recursive functions. (Refer Slide Time: 00:05)

Week - 03 Lecture - 18 Recursion. For the last lecture of this week, we will look at recursive functions. (Refer Slide Time: 00:05) Programming, Data Structures and Algorithms in Python Prof. Madhavan Mukund Department of Computer Science and Engineering Indian Institute of Technology, Madras Week - 03 Lecture - 18 Recursion For the

More information

Whitepaper Spain SEO Ranking Factors 2012

Whitepaper Spain SEO Ranking Factors 2012 Whitepaper Spain SEO Ranking Factors 2012 Authors: Marcus Tober, Sebastian Weber Searchmetrics GmbH Greifswalder Straße 212 10405 Berlin Phone: +49-30-3229535-0 Fax: +49-30-3229535-99 E-Mail: info@searchmetrics.com

More information

CS435 Introduction to Big Data Spring 2018 Colorado State University. 3/21/2018 Week 10-B Sangmi Lee Pallickara. FAQs. Collaborative filtering

CS435 Introduction to Big Data Spring 2018 Colorado State University. 3/21/2018 Week 10-B Sangmi Lee Pallickara. FAQs. Collaborative filtering W10.B.0.0 CS435 Introduction to Big Data W10.B.1 FAQs Term project 5:00PM March 29, 2018 PA2 Recitation: Friday PART 1. LARGE SCALE DATA AALYTICS 4. RECOMMEDATIO SYSTEMS 5. EVALUATIO AD VALIDATIO TECHIQUES

More information

Influence Maximization in Location-Based Social Networks Ivan Suarez, Sudarshan Seshadri, Patrick Cho CS224W Final Project Report

Influence Maximization in Location-Based Social Networks Ivan Suarez, Sudarshan Seshadri, Patrick Cho CS224W Final Project Report Influence Maximization in Location-Based Social Networks Ivan Suarez, Sudarshan Seshadri, Patrick Cho CS224W Final Project Report Abstract The goal of influence maximization has led to research into different

More information

Raw Data is data before it has been arranged in a useful manner or analyzed using statistical techniques.

Raw Data is data before it has been arranged in a useful manner or analyzed using statistical techniques. Section 2.1 - Introduction Graphs are commonly used to organize, summarize, and analyze collections of data. Using a graph to visually present a data set makes it easy to comprehend and to describe the

More information

Frequency Distributions

Frequency Distributions Displaying Data Frequency Distributions After collecting data, the first task for a researcher is to organize and summarize the data so that it is possible to get a general overview of the results. Remember,

More information

Math 340 Fall 2014, Victor Matveev. Binary system, round-off errors, loss of significance, and double precision accuracy.

Math 340 Fall 2014, Victor Matveev. Binary system, round-off errors, loss of significance, and double precision accuracy. Math 340 Fall 2014, Victor Matveev Binary system, round-off errors, loss of significance, and double precision accuracy. 1. Bits and the binary number system A bit is one digit in a binary representation

More information

Survey: Users Share Their Storage Performance Needs. Jim Handy, Objective Analysis Thomas Coughlin, PhD, Coughlin Associates

Survey: Users Share Their Storage Performance Needs. Jim Handy, Objective Analysis Thomas Coughlin, PhD, Coughlin Associates Survey: Users Share Their Storage Performance Needs Jim Handy, Objective Analysis Thomas Coughlin, PhD, Coughlin Associates Table of Contents The Problem... 1 Application Classes... 1 IOPS Needs... 2 Capacity

More information

A Virtual Laboratory for Study of Algorithms

A Virtual Laboratory for Study of Algorithms A Virtual Laboratory for Study of Algorithms Thomas E. O'Neil and Scott Kerlin Computer Science Department University of North Dakota Grand Forks, ND 58202-9015 oneil@cs.und.edu Abstract Empirical studies

More information

1 More configuration model

1 More configuration model 1 More configuration model In the last lecture, we explored the definition of the configuration model, a simple method for drawing networks from the ensemble, and derived some of its mathematical properties.

More information

UNIT 1A EXPLORING UNIVARIATE DATA

UNIT 1A EXPLORING UNIVARIATE DATA A.P. STATISTICS E. Villarreal Lincoln HS Math Department UNIT 1A EXPLORING UNIVARIATE DATA LESSON 1: TYPES OF DATA Here is a list of important terms that we must understand as we begin our study of statistics

More information

Vocabulary. 5-number summary Rule. Area principle. Bar chart. Boxplot. Categorical data condition. Categorical variable.

Vocabulary. 5-number summary Rule. Area principle. Bar chart. Boxplot. Categorical data condition. Categorical variable. 5-number summary 68-95-99.7 Rule Area principle Bar chart Bimodal Boxplot Case Categorical data Categorical variable Center Changing center and spread Conditional distribution Context Contingency table

More information

D-Optimal Designs. Chapter 888. Introduction. D-Optimal Design Overview

D-Optimal Designs. Chapter 888. Introduction. D-Optimal Design Overview Chapter 888 Introduction This procedure generates D-optimal designs for multi-factor experiments with both quantitative and qualitative factors. The factors can have a mixed number of levels. For example,

More information

x y

x y 10. LECTURE 10 Objectives I understand the difficulty in finding an appropriate function for a data set in general. In some cases, I can define a function type that may fit a data set well. Last time,

More information

Descriptive Statistics, Standard Deviation and Standard Error

Descriptive Statistics, Standard Deviation and Standard Error AP Biology Calculations: Descriptive Statistics, Standard Deviation and Standard Error SBI4UP The Scientific Method & Experimental Design Scientific method is used to explore observations and answer questions.

More information

The Transition to Networked Storage

The Transition to Networked Storage The Transition to Networked Storage Jim Metzler Ashton, Metzler & Associates Table of Contents 1.0 Executive Summary... 3 2.0 The Emergence of the Storage Area Network... 3 3.0 The Link Between Business

More information

The Attraction of Complexity

The Attraction of Complexity The Attraction of Complexity Carlo Bottiglieri December 10, 2017 1 Introduction How is complexity distributed through a codebase? Does this distribution present similarities across different projects?

More information

Univariate Statistics Summary

Univariate Statistics Summary Further Maths Univariate Statistics Summary Types of Data Data can be classified as categorical or numerical. Categorical data are observations or records that are arranged according to category. For example:

More information

4. Write sets of directions for how to check for direct variation. How to check for direct variation by analyzing the graph :

4. Write sets of directions for how to check for direct variation. How to check for direct variation by analyzing the graph : Name Direct Variations There are many relationships that two variables can have. One of these relationships is called a direct variation. Use the description and example of direct variation to help you

More information

The Effect of Inverse Document Frequency Weights on Indexed Sequence Retrieval. Kevin C. O'Kane. Department of Computer Science

The Effect of Inverse Document Frequency Weights on Indexed Sequence Retrieval. Kevin C. O'Kane. Department of Computer Science The Effect of Inverse Document Frequency Weights on Indexed Sequence Retrieval Kevin C. O'Kane Department of Computer Science The University of Northern Iowa Cedar Falls, Iowa okane@cs.uni.edu http://www.cs.uni.edu/~okane

More information

Rise-Time Enhancement Techniques for Resistive Array Infrared Scene Projectors

Rise-Time Enhancement Techniques for Resistive Array Infrared Scene Projectors Rise-Time Enhancement Techniques for Resistive Array Infrared Scene Projectors Greg Franks a, Joe LaVeigne a, Kevin Sparkman a, Jim Oleson a a Santa Barbara Infrared, Inc., 30 S. Calle Cesar Chavez, #D,

More information

HOUR 10. Creating Cool Diagrams

HOUR 10. Creating Cool Diagrams HOUR 10 Creating Cool Diagrams In addition to tables for text and charts for numbers, presenters frequently want to show their audience a visual depiction of a complex idea to explain it more comprehensively.

More information

Chapter 1.6 Systems Development Life Cycle.

Chapter 1.6 Systems Development Life Cycle. Chapter 1.6 Systems Development Life Cycle. 1.6 (a) Problem Definition It is important from the outset to ensure that when a computer system is being designed all those that are involved are agreed about

More information

Chapter 3 Analyzing Normal Quantitative Data

Chapter 3 Analyzing Normal Quantitative Data Chapter 3 Analyzing Normal Quantitative Data Introduction: In chapters 1 and 2, we focused on analyzing categorical data and exploring relationships between categorical data sets. We will now be doing

More information

Equations and Functions, Variables and Expressions

Equations and Functions, Variables and Expressions Equations and Functions, Variables and Expressions Equations and functions are ubiquitous components of mathematical language. Success in mathematics beyond basic arithmetic depends on having a solid working

More information

Table of Contents (As covered from textbook)

Table of Contents (As covered from textbook) Table of Contents (As covered from textbook) Ch 1 Data and Decisions Ch 2 Displaying and Describing Categorical Data Ch 3 Displaying and Describing Quantitative Data Ch 4 Correlation and Linear Regression

More information

A SERVICE ORGANIZATION S GUIDE SOC 1, 2, & 3 REPORTS

A SERVICE ORGANIZATION S GUIDE SOC 1, 2, & 3 REPORTS A SERVICE ORGANIZATION S GUIDE SOC 1, 2, & 3 REPORTS Introduction If you re a growing service organization, whether a technology provider, financial services corporation, healthcare company, or professional

More information

Averages and Variation

Averages and Variation Averages and Variation 3 Copyright Cengage Learning. All rights reserved. 3.1-1 Section 3.1 Measures of Central Tendency: Mode, Median, and Mean Copyright Cengage Learning. All rights reserved. 3.1-2 Focus

More information

Optimize Online Testing for Site Optimization: 101. White Paper. White Paper Webtrends 2014 Webtrends, Inc. All Rights Reserved

Optimize Online Testing for Site Optimization: 101. White Paper. White Paper Webtrends 2014 Webtrends, Inc. All Rights Reserved Optimize Online Testing for Site Optimization: 101 White Paper Overview Understanding the differences between A/B and multivariate testing helps marketers select the proper method for reaching optimization

More information

Math 120 Introduction to Statistics Mr. Toner s Lecture Notes 3.1 Measures of Central Tendency

Math 120 Introduction to Statistics Mr. Toner s Lecture Notes 3.1 Measures of Central Tendency Math 1 Introduction to Statistics Mr. Toner s Lecture Notes 3.1 Measures of Central Tendency lowest value + highest value midrange The word average: is very ambiguous and can actually refer to the mean,

More information

Downloaded from

Downloaded from UNIT 2 WHAT IS STATISTICS? Researchers deal with a large amount of data and have to draw dependable conclusions on the basis of data collected for the purpose. Statistics help the researchers in making

More information

Predicting User Ratings Using Status Models on Amazon.com

Predicting User Ratings Using Status Models on Amazon.com Predicting User Ratings Using Status Models on Amazon.com Borui Wang Stanford University borui@stanford.edu Guan (Bell) Wang Stanford University guanw@stanford.edu Group 19 Zhemin Li Stanford University

More information

Two-dimensional Totalistic Code 52

Two-dimensional Totalistic Code 52 Two-dimensional Totalistic Code 52 Todd Rowland Senior Research Associate, Wolfram Research, Inc. 100 Trade Center Drive, Champaign, IL The totalistic two-dimensional cellular automaton code 52 is capable

More information

Six Sigma in the datacenter drives a zero-defects culture

Six Sigma in the datacenter drives a zero-defects culture Six Sigma in the datacenter drives a zero-defects culture Situation Like many IT organizations, Microsoft IT wants to keep its global infrastructure available at all times. Scope, scale, and an environment

More information

Response Network Emerging from Simple Perturbation

Response Network Emerging from Simple Perturbation Journal of the Korean Physical Society, Vol 44, No 3, March 2004, pp 628 632 Response Network Emerging from Simple Perturbation S-W Son, D-H Kim, Y-Y Ahn and H Jeong Department of Physics, Korea Advanced

More information

BIO 360: Vertebrate Physiology Lab 9: Graphing in Excel. Lab 9: Graphing: how, why, when, and what does it mean? Due 3/26

BIO 360: Vertebrate Physiology Lab 9: Graphing in Excel. Lab 9: Graphing: how, why, when, and what does it mean? Due 3/26 Lab 9: Graphing: how, why, when, and what does it mean? Due 3/26 INTRODUCTION Graphs are one of the most important aspects of data analysis and presentation of your of data. They are visual representations

More information

Direct Variations DIRECT AND INVERSE VARIATIONS 19. Name

Direct Variations DIRECT AND INVERSE VARIATIONS 19. Name DIRECT AND INVERSE VARIATIONS 19 Direct Variations Name Of the many relationships that two variables can have, one category is called a direct variation. Use the description and example of direct variation

More information

Erdös-Rényi Graphs, Part 2

Erdös-Rényi Graphs, Part 2 Graphs and Networks Lecture 3 Erdös-Rényi Graphs, Part 2 Daniel A. Spielman September 5, 2013 3.1 Disclaimer These notes are not necessarily an accurate representation of what happened in class. They are

More information

A Generating Function Approach to Analyze Random Graphs

A Generating Function Approach to Analyze Random Graphs A Generating Function Approach to Analyze Random Graphs Presented by - Vilas Veeraraghavan Advisor - Dr. Steven Weber Department of Electrical and Computer Engineering Drexel University April 8, 2005 Presentation

More information

Overview. Frequency Distributions. Chapter 2 Summarizing & Graphing Data. Descriptive Statistics. Inferential Statistics. Frequency Distribution

Overview. Frequency Distributions. Chapter 2 Summarizing & Graphing Data. Descriptive Statistics. Inferential Statistics. Frequency Distribution Chapter 2 Summarizing & Graphing Data Slide 1 Overview Descriptive Statistics Slide 2 A) Overview B) Frequency Distributions C) Visualizing Data summarize or describe the important characteristics of a

More information

Whitepaper Italy SEO Ranking Factors 2012

Whitepaper Italy SEO Ranking Factors 2012 Whitepaper Italy SEO Ranking Factors 2012 Authors: Marcus Tober, Sebastian Weber Searchmetrics GmbH Greifswalder Straße 212 10405 Berlin Phone: +49-30-3229535-0 Fax: +49-30-3229535-99 E-Mail: info@searchmetrics.com

More information

Order from Chaos. University of Nebraska-Lincoln Discrete Mathematics Seminar

Order from Chaos. University of Nebraska-Lincoln Discrete Mathematics Seminar Order from Chaos University of Nebraska-Lincoln Discrete Mathematics Seminar Austin Mohr Department of Mathematics Nebraska Wesleyan University February 8, 20 The (, )-Puzzle Start by drawing six dots

More information

Slide Copyright 2005 Pearson Education, Inc. SEVENTH EDITION and EXPANDED SEVENTH EDITION. Chapter 13. Statistics Sampling Techniques

Slide Copyright 2005 Pearson Education, Inc. SEVENTH EDITION and EXPANDED SEVENTH EDITION. Chapter 13. Statistics Sampling Techniques SEVENTH EDITION and EXPANDED SEVENTH EDITION Slide - Chapter Statistics. Sampling Techniques Statistics Statistics is the art and science of gathering, analyzing, and making inferences from numerical information

More information

An Interesting Way to Combine Numbers

An Interesting Way to Combine Numbers An Interesting Way to Combine Numbers Joshua Zucker and Tom Davis October 12, 2016 Abstract This exercise can be used for middle school students and older. The original problem seems almost impossibly

More information

STA 570 Spring Lecture 5 Tuesday, Feb 1

STA 570 Spring Lecture 5 Tuesday, Feb 1 STA 570 Spring 2011 Lecture 5 Tuesday, Feb 1 Descriptive Statistics Summarizing Univariate Data o Standard Deviation, Empirical Rule, IQR o Boxplots Summarizing Bivariate Data o Contingency Tables o Row

More information

The remarketing report

The remarketing report The remarketing report Benchmark data and analysis on connecting web behavior to email marketing An Experian Cheetahmail white paper The remarketing report: Benchmark data and analysis on connecting web

More information

Hashing. Hashing Procedures

Hashing. Hashing Procedures Hashing Hashing Procedures Let us denote the set of all possible key values (i.e., the universe of keys) used in a dictionary application by U. Suppose an application requires a dictionary in which elements

More information

2.3 Algorithms Using Map-Reduce

2.3 Algorithms Using Map-Reduce 28 CHAPTER 2. MAP-REDUCE AND THE NEW SOFTWARE STACK one becomes available. The Master must also inform each Reduce task that the location of its input from that Map task has changed. Dealing with a failure

More information

Table of Contents Chapter 3. Creating Your Own Pages...

Table of Contents Chapter 3. Creating Your Own Pages... Table of Contents Chapter 3. Creating Your Own Pages... 1 Welcome to Facebook Pages... 1 Pages from a Marketing Perspective... 2 Viral Marketing with Pages... 0 Page Authenticity... 0 Finding Pages...

More information

Stable Statistics of the Blogograph

Stable Statistics of the Blogograph Stable Statistics of the Blogograph Mark Goldberg, Malik Magdon-Ismail, Stephen Kelley, Konstantin Mertsalov Rensselaer Polytechnic Institute Department of Computer Science Abstract. The primary focus

More information

WHITE PAPER. LTE in Mining. Will it provide the predictability, capacity and speed you need for your mine?

WHITE PAPER. LTE in Mining. Will it provide the predictability, capacity and speed you need for your mine? WHITE PAPER LTE in Mining Will it provide the predictability, capacity and speed you need for your mine? Table of Contents I. Executive Overview 3 II. Methods of LTE Deployment Today 4 III. Availability

More information

Multi-threading technology and the challenges of meeting performance and power consumption demands for mobile applications

Multi-threading technology and the challenges of meeting performance and power consumption demands for mobile applications Multi-threading technology and the challenges of meeting performance and power consumption demands for mobile applications September 2013 Navigating between ever-higher performance targets and strict limits

More information

Demonstrating Lorenz Wealth Distribution and Increasing Gini Coefficient with the Iterating (Koch Snowflake) Fractal Attractor.

Demonstrating Lorenz Wealth Distribution and Increasing Gini Coefficient with the Iterating (Koch Snowflake) Fractal Attractor. Demonstrating Lorenz Wealth Distribution and Increasing Gini Coefficient with the Iterating (Koch Snowflake) Fractal Attractor. First published: May 17th, 2015. Blair D. Macdonald Abstract The Koch snowflake

More information

What s New in Spotfire DXP 1.1. Spotfire Product Management January 2007

What s New in Spotfire DXP 1.1. Spotfire Product Management January 2007 What s New in Spotfire DXP 1.1 Spotfire Product Management January 2007 Spotfire DXP Version 1.1 This document highlights the new capabilities planned for release in version 1.1 of Spotfire DXP. In this

More information

Q1) Describe business intelligence system development phases? (6 marks)

Q1) Describe business intelligence system development phases? (6 marks) BUISINESS ANALYTICS AND INTELLIGENCE SOLVED QUESTIONS Q1) Describe business intelligence system development phases? (6 marks) The 4 phases of BI system development are as follow: Analysis phase Design

More information

Annual Public Safety PSAP Survey results

Annual Public Safety PSAP Survey results Annual Public Safety PSAP Survey results 573 PSAP Employees Surveyed March - July 2016 2016 Stratus Technologies. All Rights Reserved. 1 Stratus PSAP Survey March 2016 From March 2016 through July of 2016,

More information

Ecommerce Site Search. A Guide to Evaluating Site Search Solutions

Ecommerce Site Search. A Guide to Evaluating Site Search Solutions Ecommerce Site Search A Guide to Evaluating Site Search Solutions Contents 03 / Introduction 13 / CHAPTER 4: Tips for a Successful Selection Process 04 / CHAPTER 1: The Value of Site Search 16 / Conclusion

More information

(Refer Slide Time: 01:00)

(Refer Slide Time: 01:00) Advanced Operations Research Prof. G. Srinivasan Department of Management Studies Indian Institute of Technology, Madras Lecture minus 26 Heuristics for TSP In this lecture, we continue our discussion

More information

Allstate Insurance Claims Severity: A Machine Learning Approach

Allstate Insurance Claims Severity: A Machine Learning Approach Allstate Insurance Claims Severity: A Machine Learning Approach Rajeeva Gaur SUNet ID: rajeevag Jeff Pickelman SUNet ID: pattern Hongyi Wang SUNet ID: hongyiw I. INTRODUCTION The insurance industry has

More information

Using GitHub to Share with SparkFun a

Using GitHub to Share with SparkFun a Using GitHub to Share with SparkFun a learn.sparkfun.com tutorial Available online at: http://sfe.io/t52 Contents Introduction Gitting Started Forking a Repository Committing, Pushing and Pulling Syncing

More information

Example for calculation of clustering coefficient Node N 1 has 8 neighbors (red arrows) There are 12 connectivities among neighbors (blue arrows)

Example for calculation of clustering coefficient Node N 1 has 8 neighbors (red arrows) There are 12 connectivities among neighbors (blue arrows) Example for calculation of clustering coefficient Node N 1 has 8 neighbors (red arrows) There are 12 connectivities among neighbors (blue arrows) Average clustering coefficient of a graph Overall measure

More information

Introduction to the Active Everywhere Database

Introduction to the Active Everywhere Database Introduction to the Active Everywhere Database INTRODUCTION For almost half a century, the relational database management system (RDBMS) has been the dominant model for database management. This more than

More information