Using Packet Timing Information in Website Fingerprinting

Size: px
Start display at page:

Download "Using Packet Timing Information in Website Fingerprinting"

Transcription

1 Rochester Institute of Technology RIT Scholar Works Theses Thesis/Dissertation Collections Using Packet Timing Information in Website Fingerprinting Mohammad Saidur Rahman Follow this and additional works at: Recommended Citation Rahman, Mohammad Saidur, "Using Packet Timing Information in Website Fingerprinting" (2018). Thesis. Rochester Institute of Technology. Accessed from This Thesis is brought to you for free and open access by the Thesis/Dissertation Collections at RIT Scholar Works. It has been accepted for inclusion in Theses by an authorized administrator of RIT Scholar Works. For more information, please contact

2 Using Packet Timing Information in Website Fingerprinting by Mohammad Saidur Rahman Committe Members Dr. Matthew Wright Bill Stackpole Dr. Leon Reznik In partial fulfillment of the requirements for the degree of Master of Science in Computing Security Rochester Institute of Technology B. Thomas Golisano College of Computing & Information Sciences Department of Computing Security August 21, 2018

3 Rochester Institute of Technology B. Thomas Golisano College of Computing and Information Sciences Master of Science in Computing Security Thesis Approval Form Student Name: Mohammad Saidur Rahman Thesis Title: Using Packet Timing Information in Website Fingerprinting Thesis Committee Name Signature Date Dr. Matthew Wright Committee Chair Bill Stackpole Committee Member Dr. Leon Reznik Committee Member

4 To my Parents and All the Teachers Who Made me Who I am Today!

5 Acknowledgments I would like to express my gratitude to all the teachers in my life who have taught, inspired, and supported me till now. I am and always will be grateful to them. I owe all of my professional accomplishments to them. Firstly, I owe my life time service to make my parents and teachers proud. Only being thankful to my parents is not enough who taught me being a better human being with value sense. Secondly, I am profoundly grateful to my adviser Professor Dr. Matthew Wright for his continuous inspiration, encouragement, feedback, and invaluable advice throughout the course of my thesis work. He is one of the great examples in my life as a great professor and as a great human being. I look forward to learning more and more in the next phases of my life. I would also like to express my gratitude to my committee members Professor Bill Stackpole, and Professor Dr. Leon Reznik for their continuous feedback, suggestions, and comments on my thesis work. Next, I would like to express my heartiest thankfulness to those who taught, inspired, encouraged, and supported me to pursue my higher education. The whole list would be far too long. I include the few who are directly related to this endeavor from the MIS department of University of Dhaka: Md. Ariful Islam, Dr. Hasibur Rashid, Dr. K. M. Salah Uddin, and Dr. Md. Rakibul Hoque. The last but not the least, I want to express my thankfulness to my two great mentors, and friends from the Center for Cybersecurity: Payap Sirinam and Mohsen Imani for their continuous mentoring, support, care, and love. I would also like to thank Kantha Girish Gangadhara for supporting me in my project and being a great friend. iv

6 Abstract Using Packet Timing Information in Website Fingerprinting Mohammad Saidur Rahman Rochester Institute of Technology Supervisor: Dr. Matthew Wright Website Fingerprinting (WF) enables an eavesdropper to discover what sites the user is visiting despite the use of a VPN or even the Tor anonymity system. Recent WF attacks on Tor have reached high enough accuracy (up to 98%) to prompt Tor to consider adopting defenses based on packet padding. Defenses such as Walkie-Talkie mainly remove features related to bursts of traffic without affecting packet timing. This was reasonable given that previous research on WF attacks ignored or deemphasized the use of packet timing information. In this thesis, we examine the extent to which packet timing can be used to facilitate WF attacks. In our experiment, we gained up to 61% accuracy on our unprotected dataset, 54% on our WTF-PAD dataset, and 43% on our Walkie-Talkie dataset using only timing-based features in an SVM classifier. Using a convolutional neural network (CNN), we got 88% accuracy on our unprotected dataset, and 76% and 47% accuracy on our WTF-PAD and Walkie-Talkie dataset respectively. We intend to investigate further to develop an effective and robust WF attack using packet timing. v

7 Table of Contents Acknowledgments Abstract List of Figures List of Tables iv v viii x Chapter 1. Introduction Research Agenda Thesis Organization Chapter 2. Background Tor Network WF Attack Model Machine Learning k-nearest Neighbor (k-nn) Support Vector Machines (SVM) Deep Learning Convolutional Neural Network (CNN) Chapter 3. Related Work WF Attack using Traditional Machine Learning WF Attack using k NN WF Attack using RDF WF Attack using CUMUL Deep Learning in WF Attack WF Attack by SDAE Automated WF vi

8 3.2.3 Deep Fingerprinting Attack WF Defenses WTF-PAD Walkie-Talkie Chapter 4. Packet Timing in Website Fingerprinting Data Collection Data Collection Process Datasets Experimental Design Feature Selection and Extraction Tuning the Number of Bins Hyperparameter Tuning in CNN Chapter 5. Evaluation & Results Experimental Environment Results RQ1: Packet Timing Features RQ2: Classification Value of Timing Features RQ3: WF Attack with Timing Features WF Attack with Defended Tor Traffic Discussion Chapter 6. Conclusion & Future Work 37 References 38 vii

9 List of Figures 2.1 Tor Network Website Fingerprinting Attack A Visualization of Bursts five outgoing bursts interspersed with five incoming bursts Architecture of CNN Operation of the convolution layer Median Packet Time (MED) Variance of the Packet Times Time from the First Packet to the Last Packet (Interval) Interval between the medians of B 1 and B 2 (IMD) Interval between the start of B 1 and the start of B Interval between the end of B 1 and the start of B 2 (IBD) Interval between the start of B 1 and the start of B 2, where both B 1 and B 2 are incoming (to the client) Interval between the start of B 1 and the start of B 2, where both B 1 and B 2 are outgoing (from the client) Processing Features Undefended Tor Traffic: Attack Accuracy with CNN viii

10 5.2 Defended Tor Traffic with WTF-PAD Defense: Attack Accuracy with CNN Defended Tor Traffic with W-T Defense: Attack Accuracy with CNN ix

11 List of Tables 4.1 Datasets Tuning Number of Bins Hyperparameters Tuning of CNN Model for Undefended and WTF-PAD Dataset Hyperparameters Tuning of CNN Model for Walkie-Talkie Dataset Undefended Tor: Attack accuracy Undefended Tor: Attack accuracy in CNN Defended Tor: Attack accuracy x

12 Chapter 1 Introduction We share a lot of our personal information either willingly or unwillingly on the Internet. Since the United States Congress has upheld the rights to ISP providers to sell our browsing history [8], concern for the privacy of this personal information is increasing day by day. Unfortunately, using a VPN or other basic protections cannot effectively help Internet users to protect their privacy [7]. For example, Using an anonymity system such as Tor is a more effective solution to keep users online activities anonymous while browsing the Internet. Tor is one of the most popular privacy-enhancing technologies, with more than two million users each day. Tor is, however, vulnerable to traffic analysis attacks. Traffic analysis enables an attacker analyze the network traffic stream to deduce information about the client s activities and communication. Website fingerprinting (WF) is one such attack, and it has received significant attention by both researchers [9, 19] and Tor developers [18]. In a WF attack, a passive local eavesdropper collects network traffic passing 1

13 between the client and entry node. From the collected traffic, the attacker then extracts various features and feeds them into a machine learning classifier trained to identify which website the client is visiting. Prior work has shown that this kind of attack is very effective, reaching over 90% accuracy [9, 15, 23, 19, 21]. The Tor Project has given much attention to building defenses against WF attacks [17]. The state-of-the-art attacks emphasize bursts, sequences of packets in a single direction. Because of this, defenses primarily seek to obscure burst patterns. This still leaves the timing of packets as a largely untapped and unprotected resource for features for WF attacks. Prior work on WF attacks discounted timing information [9]. This is because timing characteristics change on each visit to the site, which makes it hard to extract consistent patterns. In this thesis, we investigate a novel WF attack using packet timing information. Since individual packet times are generally unreliable as features, we identify more robust timing-related features from the data. We create new representations of timing information that capture meaningful features in a reliable way. We show the effectiveness of our features using traditional machine learning and deep learning algorithms. Using our new timing features with k-nearest neighbor (k-nn) and support vector machine (SVM) classifiers, we get 51% and 61% attack accuracy, respectively. Using the CNN deep-learning classifier, we get 88% accuracy. These preliminary results indicate that timing 2

14 information is useful for WF attack, and it is possible to develop an effective WF attack using timing information. Additionally, our timing features show higher accuracy than that of using packet direction features against Walkie- Talkie (W-T) defense as well. 1.1 Research Agenda From our literature review, we find some limitations in the prior work on WF. Prior work on attacks deemphasized or ignored the value of packet timing information. This is because timing information varies from instance to instance, which makes it hard to find any consistent pattern to develop an attack. In this thesis, we have used packet timing information for WF attack development. As we cannot use raw timing information because of the larger variability between instances, we need to find a new representation of the timing information that we can feed into the machine learning classifier. The research questions (RQ) we investigated in this research are as follows: RQ1: How can we represent packet timing information in a way that robustly reveals patterns over different instances of a site? RQ2: How useful are those new representations for developing WF attack? RQ3: Do the new representations also provide classification value when WF defenses are in place? 3

15 1.2 Thesis Organization This thesis is organized as follows: Chapter 2 provides background of Tor, website fingerprinting, machine learning, and deep learning. Chapter 3 presents related work of both website fingerprinting attacks and defenses. The motivation for using packet timing information, the way we select timing features, and the feature extraction process are discussed in Chapter 4. We explain our evaluation of the experiments and the results in Chapter 5. We then conclude this thesis and discuss future work in Chapter 6. 4

16 Chapter 2 Background In this chapter, first we briefly describe the Tor network in Section 2.1. We provide background on the website fingerprinting (WF) attack model in Section 2.2. In Section 2.3 and Section 2.4, we provide background on machine learning and deep learning respectively, as they are used in WF attacks. 2.1 Tor Network Figure 2.1: Tor Network. Tor keeps the user activities anonymous to ensure privacy of the Tor client. At the time of browsing the Internet on Tor, the user can protect her privacy 5

17 by separating her online activities from her identity. Tor is one of the most popular low-latency anonymity systems with more than two million users each day. Tor ensures the anonymity of the users activities with the help of three nodes that constitute a Tor circuit: entry or guard node, middle node, and exit (see Figure 2.1). For example, the network traffic from the Tor client first goes to the guard node, then guard node passes those traffic to the middle node, then middle node to the exit node, and finally exit node to the web server. Each node is only aware of the previous and next node in the circuit. For example, the guard only knows about the identity of the client and the middle, while the exit only knows about middle and the destination of the client. 2.2 WF Attack Model Figure 2.2: Website Fingerprinting Attack. WF attack enables a local passive eavesdropper to identify the destination of a Tor client s connection by analyzing the network traffic passing from the client to the guard (see Figure 2.2). The term local means that the attacker is sitting 6

18 locally between the client and the guard node, and thus he knows the identity of the client. For example, the client and the attacker are in same wireless network, and the attacker has the capacity to sniff the wireless traffic. The term passive means that the attacker does not modify, insert, or delete any packets, rather, she just observes the traffic. In this threat model, we assume that an attacker has the capacity to monitor the traffic between a client and guard node, and/or she controls the guard node. Figure 2.3: A Visualization of Bursts five outgoing bursts interspersed with five incoming bursts. A WF attacker can collect traffic traces from several websites of her interest such as cnn.com, twitter.com, and so on. These websites are called monitored websites. She can extract statistical data as features from the collected network traffic, such as the total number of incoming packets, the total number of outgoing packets, the timing of each packet, and so on. A particularly important class of features is related to bursts sequences of consecutive incoming and outgoing packets (see Figure 2.3). She can use the features to train a machine learning classifier such as k-nearest neighbors (k-nn) and 7

19 support vector machines (SVM). Using the trained classifier, she can identify the client s destination websites. 2.3 Machine Learning In our experiments, we employ two widely used machine learning (ML) techniques: k-nearest neighbors (k-nn) and support vector machines (SVM). In the following sections, we provide brief explanation of this two techniques k-nearest Neighbor (k-nn) The k-nn algorithm can be used for both classification and regression. In website fingerprinting (WF), we use k-nn as a classifier. As k-nn is a supervised ML algorithm, we feed the data along with the associated label to train the classifier. In our case, the instances of the websites are the data, and the name of the websites are the labels. We train the k-nn classifier with training examples (x, y) where x is the feature and y is the target label or class. The objective is to learn the function g : X Y to find the relationship between x and y so that g(x) can predict the correct label of an unseen x. For every data points, the algorithm measures a similarity index between the k nearest neighbors and the data point x. A popular similarity matrix among researchers is Euclidean distance. 8

20 2.3.2 Support Vector Machines (SVM) Support vector machines (SVM) is one of the popular traditional machine learning techniques for solving classification problems. The basic idea is to plot the data points (features) in a n-dimensional space, where n is the number of features. weight or value of each feature is the coordinate value in the space. The Classification is done by separating the features by finding hyper-planes that divide data points in each class [26]. We want our features to be as farther as possible from the hyper-planes. Hyperplanes can easily be useful to classify the linear features. Classification becomes challenging when the data points are non-linear. For example, network traffic data used in WF attacks have non-linear features. To address this, we apply a technique called kernel trick. The kernel trick transforms low-dimensional data into high-dimensional data. In the higher dimension, the data points become linearly separable [14]. 2.4 Deep Learning Deep learning is a type of machine learning that uses neural network method to train the model to classify more accurately than the traditional machine learning algorithm such as k-nn and SVM. There are different models of deep learning such as stacked denoising autoencoder (SDAE), long short-term memory (LSTM), and convolutional neural network (CNN), and so on. In our experiment, we mainly use convolutional neural network (CNN) as it shows 9

21 better performance than SDAE and LSTM [3, 19, 21]. Hence, we are going to give a brief explanation of CNN Convolutional Neural Network (CNN) Figure 2.4: Architecture of CNN. The convolutional neural network (ConvNet or CNN) has been proven to be a very powerful deep learning model in the area of computer vision to classify images. There are four major operations of a CNN model. We can see a visualization of a CNN architecture 1 from Figure 2.4. Convolution Activation Function Sub-sampling or pooling Classification or fully connected 1 Adapted from Saleh and Ausif [4]. 10

22 Convolution Layer Figure 2.5: Operation of the convolution layer. We feed our data points or raw features into a matrix. We map features from the original feature matrix by convolution. For example, as shown in Figure 2.5 the original input matrix is 5x5, we convolute by 3x3 matrix 2. This 3x3 matrix is called the filter of the convolution. The convolution is performed by moving the 3x3 matrix by 1 pixel over the 5x5 matrix and taking the multiplication value of the 3x3 matrix. Moving by 1 pixel is called stride 1. If we move by 2 pixel, then the stride would be 2. We get the convolved features by scanning the whole image by this 3x3 matrix. Activation Function Activation function is used to introduce non-linear property in a neural network. This function converts the input values to the output values. For example, before getting the final convolved features in a CNN, we apply activation function to introduce non-linearity into the convolved features. It is a very important part of a CNN as a non-linear operation in a otherwise linear system. 2 Adapted from [2] 11

23 The activation function should be selected based on the type of dataset an activation function that works for image classification may not be suitable for WF attacks. There are several widely used activation functions such as sigmoid, rectified linear unit (ReLU), exponential linear unit (ELU), and hyperbolic tangent (tanh). In our CNN models, we use the ReLU, ELU, and tanh activation functions. Sub-sampling or Pooling Layer In this layer, we sub-sample or down-sample the convolved features. This layer reduces the dimensions of each mapped feature. But it keeps the most important features. There are some variations of pooling: max, average, sum, and so on. In our model, we use max pooling. We do that by taking the maximum of the convolved feature. Fully Connected Layer The classification of a CNN is done by the fully connected (FC) layer. Every neuron is connected to each other in the fully connected layer. The convolutional and pooling layers give us high-level features. FC layer does the actual classification by softmax regression. Softmax is a regression model for multiclass classification. The sum of the probabilities of this layer is one. 12

24 Chapter 3 Related Work Tor is the most popular anonymity systems to protect user privacy online. However, an attacker can deanonymize the activity of a Tor client by the website fingerprinting (WF) attack. Indeed, to combat this attack, Tor has deployed WF defenses [17], though these are not considered effective. In prior works, WF attacks and defenses are investigated in two settings: closeworld (CW) and open-world (OW) [5, 6, 9, 10, 23, 24]. The assumption of the close world is that the Tor client is visiting one or more websites from a small set of websites. Only in OW setting, the attacker trains his classifier with only these websites. On the other hand, in an OW setting, the client is assumed to visit any website including the monitored websites. But it is not realistically possible for an attacker to train his classifier with all the possible websites. So, the attacker trains the classifier with a set of sites only the monitored website. The state-of-the-art WF attack reaches up to 98% accuracy in the closed-world setting, and it achieves 95.7% true positive rate (TPR) with 0.7% of false positive rate (FPR) in the open-world setting. 13

25 In this chapter, we discuss WF attacks using traditional machine learning and using deep learning in more detail in Section 3.1 and Section 3.2, respectively. We discuss WF defenses in Section WF Attack using Traditional Machine Learning In 2009, Herrmann et al. [10] published a WF attack using IP packet size for their classifier, but it does not work on Tor, which has fixed-sized packets. Panchenko et al. designed a new attack adding more features: packet volume, packet direction, and timing [16]. They used support vector machines (SVM) for classification and achieved 55% accuracy. In 2012, SVM was again used by Cai et al. who proposed a new attack based on a new representation of the classification instances [5]. Their SVM was based on the Damerau-Levenshtein edit distance and using SVM kernel trick to pre-compute distances between the traces. This same attack was improved by Wang and Goldberg [24] achieving 91% accuracy. In the rest of the section, we are going to give a brief overview of three more effective and efficient WF attacks WF Attack using k NN In 2014, Wang et al. proposed a new attack using a k-nearest neighbor (k-nn) classifier on a large feature set with weight adjustment [23]. They modified the typical k-nn classifier with a weighted distance function. In the general k-nn, Euclidean distance is usually used. There are two phases of the attack: the weight-learning phase and the classification phase. The weights that are 14

26 learned in the first phase are used for classification in the second phase. This attack was the first to use a diverse set of features (bursts, packet ordering, concentration of the packets, number of incoming and outgoing packets, and so on) from the traffic information in a WF attack. In a closed-world setting of 100 websites, they achieved over 90% accuracy WF Attack using RDF Hayes et al. use a novel feature extraction and selection method: they use random decision forests (RDF) to rank features [9]. The classification is performed by the K-NN classifier using the ranked features of the RDF. This attack also achieved over 90% accuracy in close-world scenario WF Attack using CUMUL In 2016, Panchenko et al. proposed a new attack improving features based on packet size, packet ordering, and packet direction [15]. Their attack uses SVM to classify the websites. Their significant contribution is their new feature set. In addition, they collected a new data set that is more realistic. Most work collected the traffic of the 100 most popular websites listed by Alexa known as the Alexa Top 100 [1]. Panchenko et al. collected their data based on trends in Twitter, Google, and random Google search results of web pages. They also collected data from websites censored in China.They were able to achieve 92% accuracy in close-world scenario. 15

27 3.2 Deep Learning in WF Attack WF Attack by SDAE Abe and Goto is the first to explore the effectiveness of deep learning (DL) effectiveness in traffic analysis [3]. They used a Stacked Denoising Autoencoder (SDAE) model as their deep learning model, with a simple input data representation based on incoming and outgoing packet traces. They got 88% accuracy using deep learning without any manual selection of packet features. They used small datasets, which is a reason for their lower attack accuracy. However, they only considered one type of data representation that completely omit timing of the packets Automated WF Rimmer et al. [19] investigated automated feature engineering in WF attacks. They studied three deep learning models: Stacked Denoising Autoencoder (SDAE), Convolutional Neural Network (CNN), and Long-Short Term Memory (LSTM). They also collected large datasets suitable for deep learning models, with 900 websites and 2500 traces for each site. They achieved 96% accuracy in a closed-world setting Deep Fingerprinting Attack Sirinam et al. [21] extensively investigated the use of deep learning in website fingerprinting. Their attack could outperform all the previous attack accu- 16

28 racy reaching over 98% attack accuracy. They evaluated their deep learning model by 100 classes containing 1000 traces each. They developed a powerful convolutional neural network (CNN) model for their attack. Their attack achieved up to 90% accuracy against WTF-PAD, one of the state-of-the-art defenses. They showed the effectiveness of their attack even though the defense is in place. Their attack also achieved 49.7% accuracy against Walkie-Talkie defense. 3.3 WF Defenses WF defenses aim to reduce the effectiveness of these attacks. For defenses against WF attacks, we are interested in the state-of-the-art defenses that are claimed to be highly efficient and have low overheads in latency and bandwidth for users. Tor implemented a HTTP pipeline based defense [17] to combat WF attacks. The purpose of this defense is to change the order of the requests when the number of requests exceeds the depth of the pipeline. This objective is achieved by randomizing the maximum number of requests in a pipeline. The bandwidth overhead of this defense is zero. Recently, Tor has updated this defense mechanism [18] because of the emergence of newly developed attacks. However, neither version of Tor s defenses could reduce the attack accuracy of newly developed attacks [3, 5, 19, 21, 23, 24]. In the next section, we are going to discuss more about the two state-of-the-art defenses, WTF-PAD [13] and Walkie-Talkie [25]. 17

29 3.3.1 WTF-PAD Website Traffic Fingerprinting Protection with Adaptive Defense (WTF-PAD) [13] applies the adaptive padding technique to WF defense in Tor [20]. Adaptive padding works by sending data packets with padding of a certain distribution and with no delay. By doing so, adaptive padding does not incur any latency overhead. However, it does incur 54% bandwidth overhead Walkie-Talkie Wang and Goldberg [25] developed Walkie-Talkie (W-T) defense. Their defense is based on the concept of half-duplex communication and burst molding. The default communication mechanism of Tor is full duplex. In full duplex communication mode, we simultaneously send outgoing packets and receive incoming packets. W-T defense requires the communication of Tor browser to be half-duplex. In half-duplex communication, we send request of packets and wait to receive all the requested packets, after receiving all the packets, we send another request of packets. The client and the guard have to use half-duplex communication mode to implement this defense. This half duplex mode forms a sequence of outgoing and incoming bursts. The client molds the bursts of the traces of sensitive website and non-sensitive websites together. The idea is to make two websites look exactly the same to the attacker. Their defense has 31% bandwidth overhead and 34% latency overhead. 18

30 Chapter 4 Packet Timing in Website Fingerprinting In this chapter, we discuss the motivation for using packet timing information, the way we select timing features, and the feature extraction process. we discuss data collection process in Section and our datasets in Section We discuss our experimental design in Section 4.2. The process of feature selection and extraction is discussed in Section We discuss hyperparameters tuning of our CNN models in Section Data Collection Data Collection Process As we have used the datasets provided by [21], we are going to provide a brief overview of the data collection process that Sirinam et al. used. Firstly, the top Alexa 100 websites are selected as the monitored sites. Homepage of each site is visited 1,250 times and the generated traffic of each visit is dumped in terms of each visit using tcpdump. We call each visit of a site as an instance 19

31 of that site. The data collection is done based on the batch method provided by Wang and Goldberg. The batch method is useful for managing long and short term time variance. A batch contains 25 visits of a site at a time. After that, the crawler visits another site 25 times and this process continues. Batching helps to collect the traffic over time that helps to avoid the IP to be banned. In addition, it helps to capture traffic with variation. The websites are visited using tor-browser-crawler [12]. This crawler is more realistic in the sense that it helps to emulate the real behavior of a Tor client. Hence, this crawler is more realistic to collect Tor traffic than the traditional crawling tools such as wget or curl for our purpose as it has the necessary settings for that. At the end of crawling, the sanity check of the datasets is performed. At first, the instances that do not have more than 50 packets and that do not contain any incoming and outgoing packets are discarded because they do not contain enough information for WF attack. After this, 95 valid sites and 1000 instances of each site are selected as the monitored undefended datasets. For the WTF-PAD defense, the defense is applied to the datasets in simulator to make defended monitored datasets. 20

32 4.1.2 Datasets We use three datasets in our experiments, as shown in Table-4.1. These datasets have been used for the previous work [21, 25] in WF research. We use both defended and undefended Tor traffic. Defended Tor traffic means that a defense mechanism such as WTF-PAD or Walkie-Talkie (W-T) is applied to those traffic. Undefended Tor traffic means the traffic with the default settings of Tor. For undefended Tor traffic and Tor traffic defended with WTF- PAD [13], we use the dataset generated by Sirinam et at. [21]. The Walkie- Talkie defense requires changes to the underlying browser. At the time we performed our experiments, the dataset from Sirinam et al. did not include W-T data. We thus use the smaller dataset used by Wang et al. [25]. Table 4.1: Datasets. Instances in Dataset Source Classes each class Total Undefended Sirinam et al ,000 WTF-PAD Sirinam et al ,000 Walkie-Talkie Wang et al , Experimental Design In this section, we explain our process of selecting and extracting timing features as well as the process of tuning hyper-parameters for our CNN models. We have used the K-NN model of Wang et al. [23] and SVM model of 21

33 Panchenko et al. [15] to best compare our results. Using non-timing based features, they achieve over 90% accuracy. Our intuition was to compare the results using only packet timing information Feature Selection and Extraction We developed several timing-based features that would be more robust from instance to instance of each website than relying on specific packet timings. Since prior work on attacks relies heavily on bursts, we base our timing features on burst-level characteristics. Three of our features are focused on the timing of packets inside a single burst. Median Packet Time (MED): Figure 4.1: Median Packet Time (MED). We can see the extraction process of median packet time within a burst from Figure 4.1. For example, we take the time sequence of the first burst: [0.0, 0.10, 0.20], and take the median of these three values. 22

34 Variance of the Packet Times: Figure 4.2: Variance of the Packet Times. We can see the extraction process of variance packet time within a burst from Figure 4.2. For example, we take the time sequence of the first burst: [0.0, 0.10, 0.20], and take the variance of these three values. Time from the First Packet to the Last Packet (Interval): Figure 4.3: Time from the First Packet to the Last Packet (Interval). We can see the extraction process of interval of the packet times within a burst from Figure 4.3. For example, we take the time sequence of the first burst: [0.0, 0.20], and take the interval of these two values. 23

35 The other five features consider two consecutive bursts B 1 and B 2. Interval between the medians of B 1 and B 2 (IMD): Figure 4.4: Interval between the medians of B 1 and B 2 (IMD). We can see the extraction process of the interval between the medians of B 1 and B 2 (IMD) from Figure 4.4. For example, we take the median of the burst one and the burst two: [0.10, 0.50], and take the interval of these two values. Interval between the Start of B 1 and the Start of B 2 : Figure 4.5: Interval between the start of B 1 and the start of B 2. We can see the extraction process of the interval between the start of B 1 and 24

36 the start of B 2 from Figure 4.5. For example, we take the time of the first packet of the burst one and the time of the first packet of burst two: [0.00, 0.40], and take the interval of these two values. Interval between the End of B 1 and the Start of B 2 (IBD): Figure 4.6: Interval between the end of B 1 and the start of B 2 (IBD). We can see the extraction process of the interval between the end of B 1 and the start of B 2 (IBD) from Figure 4.6. For example, we take the time of the last packet of the burst one and the time of the second packet of burst two: [0.20, 0.40], and take the interval of these two values. Interval between the Start of B 1 and the Start of B 2, where both B 1 and B 2 are Incoming (to the client): We can see the extraction process of the interval between the start of B 1 and the start of B 2, where both B 1 and B 2 are incoming (to the client) from Figure 4.7. For example, we take the time of the first packet of the first incoming 25

37 Figure 4.7: Interval between the start of B 1 and the start of B 2, where both B 1 and B 2 are incoming (to the client). burst and the time of the first packet of the second incoming burst: [0.40, 0.75], and take the interval of these two values. Interval between the Start of B 1 and the Start of B 2, where both B 1 and B 2 are Outgoing (from the client): Figure 4.8: Interval between the start of B 1 and the start of B 2, where both B 1 and B 2 are outgoing (from the client). 26

38 We can see the extraction process of the interval between the start of B 1 and the start of B 2, where both B 1 and B 2 are outgoing (from the client) from Figure 4.8. For example, we take the time of the first packet of the first outgoing burst and the time of the first packet of the second outgoing burst: [0.00, 0.65], and take the interval of these two values. Figure 4.9: Processing Features. To create features that would be robust to different instances, we further process the extracted features (see Figure 4.9). Using the data from all the instances over all websites, we create a histogram of b equal-sized bins for each feature. Then, for each instance, we extract the features from raw data, put them into their respective bins, and take the length of each bin. Finally, we normalize the length of each bin and feed those normalized values into the classifiers Tuning the Number of Bins We tune the number of bins for each dataset. We started with b = 20. We explore the variation in accuracy with the change of the bin sizes. We think that number of features a bin can contain impact the classification accuracy. 27

39 For example, if we increase the number of bins, every bin will contain less features but are more fine-grained. On the other hand, if we decrease the number of bins, every bin will contain more features but are less fine-grained. So we have to make a trade-off between the number of bins and the number of features a bin contain. Taking that into account, we explore different number of bins to find the best one based on the accuracy. We present our search space of bin size in Table 4.2. We could not go beyond b = 25 for Walkie- Talkie defense because the highest number of packets of that dataset is less than 200. Hence, the more we increase bin size, the lesser information the bin contain. We finally select b = 20 for undefended and WTF-PAD defense and b = 10 for Walkie-Talkie defense. Table 4.2: Tuning Number of Bins. Dataset Number of Bins Search Final Bin Size Undefended 5, 10, 15, 20, 25, 30, 35, 40, 45, WTF-PAD 5, 10, 15, 20, 25, 30, 35, 40, 45, Walkie-Talkie 5, 10, 15, 20, Hyperparameter Tuning in CNN Different types of datasets require different hyper-parameters in the CNN model to make the model robust. Usually deep learning models require a lot of data for training. We have sufficient data to use a deep CNN model in the undefended and WTF-PAD datasets. However, for the Walkie-Talkie de- 28

40 fense we have a relatively small dataset. This means we need to use a shallow CNN model to avoid overfitting. We provide an overview of the parameters of our models in Table 4.3 and Table 4.4. Table 4.3: Hyperparameters Tuning of CNN Model for Undefended and WTF- PAD Dataset. Final Parameters Hyperparameters Search Range Undefended WTF-PAD Input Dimension [ ] Optimization Function [Adam, Adamax, SGD] Adamax Number of Filters Layer 1 [Conv1, Conv2] Layer 2 [Conv3, Conv4] Layer 3 [Conv5, Conv6] Layer 4 [Conv7, Conv8] [ ] [ ] [ ] [ ] [32, 32] [64, 64] [128, 128] [256, 256] Adamax [32, 32] [64, 64] [128, 128] [256, 256] Filter Sizes [ ] [8, 8, 8, 8 ] [8, 8, 8, 8 ] Pooling Max Max Max Learning Rate [ ] Training Epochs [ ] Mini-batch Size [ ] Activation Functions [Sigmoid, Tanh, ReLU, ELU] ELU, ReLU ReLU, Tanh Number of FC Layers [1... 4] 2 2 Hidden units (each FCs) [ ] [512, 512] [512, 512] Dropout [Pooling, FC1, FC2] [ ] [0.1, 0.7, 0.5] [0.1, 0.5, 0.5] To further avoid overfitting, we add a kernel regularizer, dropout [22], and batch-normalization in our models. For the kernel regularizer, we adopt l 2 regularization, also called weight decay. Kernel regularization is usually implemented in each convolutional layer. Dropout is implemented in the fully connected (FC) layers. Dropout randomly disconnects some of the units in the 29

41 Table 4.4: Hyperparameters Tuning of CNN Model for Walkie-Talkie Dataset. Final Parameters Hyperparameters Search Range Walkie-Talkie Input Dimension [ ] 80 Optimization Function [Adam, Adamax, SGD] Number of Filters Layer 1 [Conv1, Conv2] Layer 2 [Conv3, Conv4] Layer 3 [Conv5, Conv6] [ ] [ ] [ ] Adamax [16, 16] [32, 32] [64, 64] Filter Sizes [ ] [8, 8, 5, 5 ] Pooling Max Max Learning Rate [ ] Training Epochs [ ] 100 Mini-batch Size [ ] 64 Activation Functions [Sigmoid, Tanh, ReLU, ELU] Number of FC Layers [1... 4] 1 Hidden units (each FCs) [ ] 512 Dropout [FC1] [ ] 0.5 ReLU, Tanh FC layers and make the model more robust for testing. We add batch normalization [11] in both the convolutional layer and the FC layer. We use dropout in both of our CNN models. We use kernel regularizer in the CNN model for the Walkie-Talkie dataset and batch normalization in the CNN model for the undefended and WTD-PAD datasets. 30

42 Chapter 5 Evaluation & Results In this chapter, we discuss the evaluation of our timing features and present results based on the three research questions we investigated. First, we provide a brief overview of our experimental environment in Section 5.1. Secondly, we address our three research questions and the experimental results in Section 5.2. Then we discuss our results in Section Experimental Environment We use a PC with a Core i processor with 8 cores and 32GB RAM, plus a NVIDIA Quadro K1200 GPU with 4GB GPU memory. For compiling deep learning models to run on a GPU, we use Cuda Version 7.5. We develop our model in Python using the Keras deep learning framework as the front end and the Tensorflow deep learning framework as the back end. 31

43 5.2 Results RQ1: Packet Timing Features RQ1: How can we represent packet timing information in a way that robustly reveals patterns over different instances of a site? To respond this question, we selected and extracted eight timing features in a five steps process (see Figure 4.9). We extracted our timing features from the burst-level characteristics. Among the eight features, three features are based on the timing of a single burst. The other five features are based on the two consecutive bursts B 1 and B 2. Refer to Section for the details of our timing features selection and extraction process RQ2: Classification Value of Timing Features RQ2: How useful are those new representations for developing WF attack? We explored the use of all eight features, both separately and together, in both k-nn, and SVM for different settings of the number of bins b in each histogram. Based on our initial findings, we found that the most effective combination was the three features MED, IMD, and IBD with b = 20 bins, so we present results with these features. In our initial phase experiments, we investigated and experimented our timing features with K NN and SVM classifiers. As shown in Table-5.1, MED, IMD, and IBD all provide some classification value on their own, where random 32

44 Table 5.1: Undefended Tor: Attack accuracy. Features Classifier MED IMD IBD Combined k-nn 41% 20% 18% 51% SVM 56% 24% 29% 61% guessing would only reach 1% accuracy in this multi-class scenario. Together, these three features can get 61% accuracy with SVM on undefended Tor traffic. Based on the findings of the initial experiments, we extended our experiments on timing features to the deep learning model. We adopted one of the most powerful deep learning classifiers: Convolutional Neural Network (CNN). Table 5.2: Undefended Tor: Attack accuracy in CNN. Classifier Combined Features CNN 88% model accuracy model loss 0.8 train validation train validation accuracy 0.4 loss epoch (a) Model Accuracy epoch (b) Model Loss Figure 5.1: Undefended Tor Traffic: Attack Accuracy with CNN. Using the CNN classifier, we attain 88% testing accuracy (see Table- 5.2) with 33

45 undefended Tor traffic. We attain up to 98% of training accuracy and up to 92% of validation accuracy at the 50th epoch (see Figure-5.1). We performed our experiments with different structures of CNN model to understand the variation of our results. Finally, we selected the model that gave us the best accuracy. Refer to Table 4.3 for an overview of the search space of our model. While our results are not as good as the state-of-the-art classifiers, it is an encouraging result given that we are not using any of the features about bursts cited in prior work as being important for effective classification RQ3: WF Attack with Timing Features RQ3: Do the new representations also provide classification value when WF defenses are in place? WF Attack with Defended Tor Traffic Table 5.3: Defended Tor: Attack accuracy. Classifier WTF-PAD W-T SVM 54% 43% CNN 76% 47% We experimented with the combined feature set against the two defended datasets, using WTF-PAD and Walkie-Talkie (W-T), as reported in Table Against WTF-PAD, we attain 54% accuracy using SVM and 76% accuracy 34

46 (a) Model Accuracy (b) Model Loss Figure 5.2: Defended Tor Traffic with WTF-PAD Defense: Attack Accuracy with CNN. using CNN. Our CNN model gave us up to 96% of training accuracy and 78% of validation accuracy (see Figure- 5.2). 0.5 train validation model accuracy 4.5 train validation model loss accuracy 0.3 loss epoch epoch (a) Model Accuracy (b) Model Loss Figure 5.3: Defended Tor Traffic with W-T Defense: Attack Accuracy with CNN. Against W-T, we get 43% accuracy using SVM and 47% accuracy using CNN. Note that 47% is higher than any accuracy results reported on W-T to date 35

47 and approaching the 50% theoretical maximum accuracy claimed for W-T [25]. That maximum accuracy depends on not using timing data, so our approach might exceed 50% accuracy when both timing and burst data are considered. 5.3 Discussion Our findings show that packet timing information can be an important candidate for feature for the WF attack. Using timing features provides reasonable attack accuracy against the state-of-the-art defenses as well. However, processing the timing features is a cumbersome task. First, it involves multiple steps to extract the fine-grained timing features. Secondly, tuning the number of bins is also challenging because it requires a trail and error process. No one specific number of bins can work for every dataset. The whole process requires multiple iterations from the start to the end. Another challenging task is hyperparameter tuning to find the best CNN model for our datasets. Finally, we have two datasets that are large enough for our deep learning models, but Walkie-Talkie (W-T) dataset is not large enough. A larger W-T dataset could give us better understanding of the accuracy using our timing features. However, we still get higher accuracy than the accuracies for the other attacks shown in the W-T paper [25] using only timing features. 36

48 Chapter 6 Conclusion & Future Work In this thesis, we propose a novel WF attack based on extracting features from packet timing information. The main purpose of our research is to find useful timing features that can be used in WF attacks. We selected and extracted eight new timing features from packet timestamps in a five-step process. We show the effectiveness of our features in both traditional ML and deep learning models. We find that our features are robust over multiple noisy instances and provide useful classification value against both undefended and defended Tor traffic. Using timing features in our CNN model, we achieve 88% accuracy. We achieve 47% accuracy against W-T defense as well. In future work, we will investigate more to improve this attack and improve our model as well. We will investigate the performance of timing and direction features together. We plan to collect a larger W-T dataset to evaluate further the effectiveness of our timing features. We will also expand our evaluation to Onion Services and the open-world setting. In the future, our goal is to build a defense against this attack, as state-of-the-art defenses do not focus on obscuring the pattern of timing information. 37

49 References [1] Alexa, [2] Feature extraction using convolution, edu/wiki/index.php/feature extraction using convolution. [3] K. Abe and S. Goto, Fingerprinting attack on Tor anonymity using deep learning, Proceedings of the Asia-Pacific Advanced Network, vol. 42, pp , [4] S. Albelwi and A. Mahmood, A framework for designing the architectures of deep convolutional neural networks, Entropy, vol. 19, no. 6, [5] X. Cai, X. C. Zhang, B. Joshi, and R. Johnson, Touching from a distance: Website fingerprinting attacks and defenses, in Proceedings of the 2012 ACM Conference on Computer and Communications Security. ACM, [6] K. P. Dyer, S. E. Coull, T. Ristenpart, and T. Shrimpton, Peek-a-boo, I still see you: Why efficient traffic analysis countermeasures fail, in 2012 IEEE Symposium on Security and Privacy. IEEE, [7] L. Finley, VPNs won t save you from congress Internet privacy giveaway, 38

50 vpns-wont-save-congress-internet-privacy-giveaway/, 2017, accessed: [8] B. Fung, What to expect now that Internet providers can collect and sell your Web browser history, news/the-switch/wp/2017/03/29/what-to-expect-now-that-internet\ -providers-can-collect-and-sell-your-web-browser-history/?utm term=.5c52ff09c2be, 2017, accessed: [9] J. Hayes and G. Danezis, k-fingerprinting: A robust scalable website fingerprinting technique, in USENIX Security Symposium, 2016, pp [10] D. Herrmann, R. Wendolsky, and H. Federrath, Website fingerprinting: Attacking popular privacy enhancing technologies with the multinomial naïve-bayes classifier, in Proceedings of the 2009 ACM Workshop on Cloud Computing Security. ACM, [11] S. Ioffe and C. Szegedy, Batch normalization: Accelerating deep network training by reducing internal covariate shift, in International Conference on Machine Learning, 2015, pp [12] M. Juarez, S. Afroz, G. Acar, C. Diaz, and R. Greenstadt, A critical evaluation of website fingerprinting attacks, in Proceedings of the 2014 ACM Conference on Computer and Communications Security. ACM,

Automated Website Fingerprinting through Deep Learning

Automated Website Fingerprinting through Deep Learning Automated Website Fingerprinting through Deep Learning Vera Rimmer 1, Davy Preuveneers 1, Marc Juarez 2, Tom Van Goethem 1 and Wouter Joosen 1 NDSS 2018 Feb 19th (San Diego, USA) 1 2 Website Fingerprinting

More information

arxiv: v5 [cs.cr] 20 Aug 2018

arxiv: v5 [cs.cr] 20 Aug 2018 Deep Fingerprinting: Undermining Website Fingerprinting Defenses with Deep Learning arxiv:1801.02265v5 [cs.cr] 20 Aug 2018 ABSTRACT Payap Sirinam Rochester Institute of Technology Rochester, New York payap.sirinam@mail.rit.edu

More information

DynaFlow: An Efficient Website Fingerprinting Defense Based on Dynamically-Adjusting Flows

DynaFlow: An Efficient Website Fingerprinting Defense Based on Dynamically-Adjusting Flows DynaFlow: An Efficient Website Fingerprinting Defense Based on Dynamically-Adjusting Flows ABSTRACT David Lu MIT PRIMES davidboxboro@gmail.com Albert Kwon MIT kwonal@mit.edu Website fingerprinting attacks

More information

k-fingerprinting: A Robust Scalable Website Fingerprinting Technique

k-fingerprinting: A Robust Scalable Website Fingerprinting Technique k-fingerprinting: A Robust Scalable Website Fingerprinting Technique Jamie Hayes and George Danezis, University College London https://www.usenix.org/conference/usenixsecurity16/technical-sessions/presentation/hayes

More information

Circuit Fingerprinting Attack: Passive Deanonymization of Tor Hidden Services

Circuit Fingerprinting Attack: Passive Deanonymization of Tor Hidden Services Circuit Fingerprinting Attack: Passive Deanonymization of Tor Hidden Services Albert Kwon 1 Mashael Saad Al-Sabah 123 David Lazar 1 Marc Dacier 2 Srinivas Devadas 1 1 CSAIL/MIT 2 Qatar Computing Research

More information

Index. Umberto Michelucci 2018 U. Michelucci, Applied Deep Learning,

Index. Umberto Michelucci 2018 U. Michelucci, Applied Deep Learning, A Acquisition function, 298, 301 Adam optimizer, 175 178 Anaconda navigator conda command, 3 Create button, 5 download and install, 1 installing packages, 8 Jupyter Notebook, 11 13 left navigation pane,

More information

k-fingerprinting: a Robust Scalable Website Fingerprinting Technique

k-fingerprinting: a Robust Scalable Website Fingerprinting Technique k-fingerprinting: a Robust Scalable Website Fingerprinting Technique Jamie Hayes University College London j.hayes@cs.ucl.ac.uk George Danezis University College London g.danezis@ucl.ac.uk Abstract Website

More information

Deep Learning with Tensorflow AlexNet

Deep Learning with Tensorflow   AlexNet Machine Learning and Computer Vision Group Deep Learning with Tensorflow http://cvml.ist.ac.at/courses/dlwt_w17/ AlexNet Krizhevsky, Alex, Ilya Sutskever, and Geoffrey E. Hinton, "Imagenet classification

More information

arxiv: v3 [cs.cr] 19 Feb 2016

arxiv: v3 [cs.cr] 19 Feb 2016 k-fingerprinting: a Robust Scalable Website Fingerprinting Technique arxiv:1509.00789v3 [cs.cr] 19 Feb 2016 Jamie Hayes University College London j.hayes@cs.ucl.ac.uk Abstract Website fingerprinting enables

More information

arxiv: v3 [cs.cr] 19 Jul 2016

arxiv: v3 [cs.cr] 19 Jul 2016 Toward an Efficient Website Fingerprinting Defense Marc Juarez 1, Mohsen Imani 2, Mike Perry 3, Claudia Diaz 1, and Matthew Wright 2 arxiv:1512.00524v3 [cs.cr] 19 Jul 2016 1 KU Leuven, ESAT/COSIC and iminds,

More information

Deep Learning for Computer Vision II

Deep Learning for Computer Vision II IIIT Hyderabad Deep Learning for Computer Vision II C. V. Jawahar Paradigm Shift Feature Extraction (SIFT, HoG, ) Part Models / Encoding Classifier Sparrow Feature Learning Classifier Sparrow L 1 L 2 L

More information

Accelerating Convolutional Neural Nets. Yunming Zhang

Accelerating Convolutional Neural Nets. Yunming Zhang Accelerating Convolutional Neural Nets Yunming Zhang Focus Convolutional Neural Nets is the state of the art in classifying the images The models take days to train Difficult for the programmers to tune

More information

Perceptron: This is convolution!

Perceptron: This is convolution! Perceptron: This is convolution! v v v Shared weights v Filter = local perceptron. Also called kernel. By pooling responses at different locations, we gain robustness to the exact spatial location of image

More information

Facial Expression Classification with Random Filters Feature Extraction

Facial Expression Classification with Random Filters Feature Extraction Facial Expression Classification with Random Filters Feature Extraction Mengye Ren Facial Monkey mren@cs.toronto.edu Zhi Hao Luo It s Me lzh@cs.toronto.edu I. ABSTRACT In our work, we attempted to tackle

More information

Effective Attacks and Provable Defenses for Website Fingerprinting

Effective Attacks and Provable Defenses for Website Fingerprinting Effective Attacks and Provable Defenses for Website Fingerprinting Tao Wang 1 Xiang Cai 2 Rishab Nithyanand 2 Rob Johnson 2 Ian Goldberg 1 1 University of Waterloo 2 Stony Brook University {t55wang,iang}@cs.uwaterloo.ca

More information

Channel Locality Block: A Variant of Squeeze-and-Excitation

Channel Locality Block: A Variant of Squeeze-and-Excitation Channel Locality Block: A Variant of Squeeze-and-Excitation 1 st Huayu Li Northern Arizona University Flagstaff, United State Northern Arizona University hl459@nau.edu arxiv:1901.01493v1 [cs.lg] 6 Jan

More information

Deep Learning For Video Classification. Presented by Natalie Carlebach & Gil Sharon

Deep Learning For Video Classification. Presented by Natalie Carlebach & Gil Sharon Deep Learning For Video Classification Presented by Natalie Carlebach & Gil Sharon Overview Of Presentation Motivation Challenges of video classification Common datasets 4 different methods presented in

More information

An Adaptive Algorithm for Range Queries in Differential Privacy

An Adaptive Algorithm for Range Queries in Differential Privacy Rochester Institute of Technology RIT Scholar Works Theses Thesis/Dissertation Collections 6-2016 An Adaptive Algorithm for Range Queries in Differential Privacy Asma Alnemari Follow this and additional

More information

COMP 551 Applied Machine Learning Lecture 16: Deep Learning

COMP 551 Applied Machine Learning Lecture 16: Deep Learning COMP 551 Applied Machine Learning Lecture 16: Deep Learning Instructor: Ryan Lowe (ryan.lowe@cs.mcgill.ca) Slides mostly by: Class web page: www.cs.mcgill.ca/~hvanho2/comp551 Unless otherwise noted, all

More information

Machine Learning 13. week

Machine Learning 13. week Machine Learning 13. week Deep Learning Convolutional Neural Network Recurrent Neural Network 1 Why Deep Learning is so Popular? 1. Increase in the amount of data Thanks to the Internet, huge amount of

More information

Know your data - many types of networks

Know your data - many types of networks Architectures Know your data - many types of networks Fixed length representation Variable length representation Online video sequences, or samples of different sizes Images Specific architectures for

More information

Classifying Depositional Environments in Satellite Images

Classifying Depositional Environments in Satellite Images Classifying Depositional Environments in Satellite Images Alex Miltenberger and Rayan Kanfar Department of Geophysics School of Earth, Energy, and Environmental Sciences Stanford University 1 Introduction

More information

Deep Learning. Practical introduction with Keras JORDI TORRES 27/05/2018. Chapter 3 JORDI TORRES

Deep Learning. Practical introduction with Keras JORDI TORRES 27/05/2018. Chapter 3 JORDI TORRES Deep Learning Practical introduction with Keras Chapter 3 27/05/2018 Neuron A neural network is formed by neurons connected to each other; in turn, each connection of one neural network is associated

More information

Convolution Neural Networks for Chinese Handwriting Recognition

Convolution Neural Networks for Chinese Handwriting Recognition Convolution Neural Networks for Chinese Handwriting Recognition Xu Chen Stanford University 450 Serra Mall, Stanford, CA 94305 xchen91@stanford.edu Abstract Convolutional neural networks have been proven

More information

Dynamic Routing Between Capsules

Dynamic Routing Between Capsules Report Explainable Machine Learning Dynamic Routing Between Capsules Author: Michael Dorkenwald Supervisor: Dr. Ullrich Köthe 28. Juni 2018 Inhaltsverzeichnis 1 Introduction 2 2 Motivation 2 3 CapusleNet

More information

A Deep Learning Approach to Vehicle Speed Estimation

A Deep Learning Approach to Vehicle Speed Estimation A Deep Learning Approach to Vehicle Speed Estimation Benjamin Penchas bpenchas@stanford.edu Tobin Bell tbell@stanford.edu Marco Monteiro marcorm@stanford.edu ABSTRACT Given car dashboard video footage,

More information

NVIDIA FOR DEEP LEARNING. Bill Veenhuis

NVIDIA FOR DEEP LEARNING. Bill Veenhuis NVIDIA FOR DEEP LEARNING Bill Veenhuis bveenhuis@nvidia.com Nvidia is the world s leading ai platform ONE ARCHITECTURE CUDA 2 GPU: Perfect Companion for Accelerating Apps & A.I. CPU GPU 3 Intro to AI AGENDA

More information

Exploiting Data-Usage Statistics for Website Fingerprinting Attacks on Android

Exploiting Data-Usage Statistics for Website Fingerprinting Attacks on Android S C I E N C E P A S S I O N T E C H N O L O G Y Exploiting Data-Usage Statistics for Website Fingerprinting Attacks on Android Raphael Spreitzer, Simone Griesmayr, Thomas Korak, and Stefan Mangard IAIK,

More information

LSTM: An Image Classification Model Based on Fashion-MNIST Dataset

LSTM: An Image Classification Model Based on Fashion-MNIST Dataset LSTM: An Image Classification Model Based on Fashion-MNIST Dataset Kexin Zhang, Research School of Computer Science, Australian National University Kexin Zhang, U6342657@anu.edu.au Abstract. The application

More information

Deep Learning and Its Applications

Deep Learning and Its Applications Convolutional Neural Network and Its Application in Image Recognition Oct 28, 2016 Outline 1 A Motivating Example 2 The Convolutional Neural Network (CNN) Model 3 Training the CNN Model 4 Issues and Recent

More information

Deep Face Recognition. Nathan Sun

Deep Face Recognition. Nathan Sun Deep Face Recognition Nathan Sun Why Facial Recognition? Picture ID or video tracking Higher Security for Facial Recognition Software Immensely useful to police in tracking suspects Your face will be an

More information

Keras: Handwritten Digit Recognition using MNIST Dataset

Keras: Handwritten Digit Recognition using MNIST Dataset Keras: Handwritten Digit Recognition using MNIST Dataset IIT PATNA February 9, 2017 1 / 24 OUTLINE 1 Introduction Keras: Deep Learning library for Theano and TensorFlow 2 Installing Keras Installation

More information

COMP9444 Neural Networks and Deep Learning 7. Image Processing. COMP9444 c Alan Blair, 2017

COMP9444 Neural Networks and Deep Learning 7. Image Processing. COMP9444 c Alan Blair, 2017 COMP9444 Neural Networks and Deep Learning 7. Image Processing COMP9444 17s2 Image Processing 1 Outline Image Datasets and Tasks Convolution in Detail AlexNet Weight Initialization Batch Normalization

More information

On Realistically Attacking Tor with Website Fingerprinting

On Realistically Attacking Tor with Website Fingerprinting Proceedings on Privacy Enhancing Technologies ; 2016 (4):21 36 Tao Wang* and Ian Goldberg On Realistically Attacking Tor with Website Fingerprinting Abstract: Website fingerprinting allows a local, passive

More information

DECISION TREES & RANDOM FORESTS X CONVOLUTIONAL NEURAL NETWORKS

DECISION TREES & RANDOM FORESTS X CONVOLUTIONAL NEURAL NETWORKS DECISION TREES & RANDOM FORESTS X CONVOLUTIONAL NEURAL NETWORKS Deep Neural Decision Forests Microsoft Research Cambridge UK, ICCV 2015 Decision Forests, Convolutional Networks and the Models in-between

More information

Machine Learning. Deep Learning. Eric Xing (and Pengtao Xie) , Fall Lecture 8, October 6, Eric CMU,

Machine Learning. Deep Learning. Eric Xing (and Pengtao Xie) , Fall Lecture 8, October 6, Eric CMU, Machine Learning 10-701, Fall 2015 Deep Learning Eric Xing (and Pengtao Xie) Lecture 8, October 6, 2015 Eric Xing @ CMU, 2015 1 A perennial challenge in computer vision: feature engineering SIFT Spin image

More information

Deep Learning. Vladimir Golkov Technical University of Munich Computer Vision Group

Deep Learning. Vladimir Golkov Technical University of Munich Computer Vision Group Deep Learning Vladimir Golkov Technical University of Munich Computer Vision Group 1D Input, 1D Output target input 2 2D Input, 1D Output: Data Distribution Complexity Imagine many dimensions (data occupies

More information

Se Eun Oh*, Shuai Li, and Nicholas Hopper Fingerprinting Keywords in Search Queries over Tor

Se Eun Oh*, Shuai Li, and Nicholas Hopper Fingerprinting Keywords in Search Queries over Tor Proceedings on Privacy Enhancing Technologies ; 2017 (4):251 270 Se Eun Oh*, Shuai Li, and Nicholas Hopper Fingerprinting Keywords in Search Queries over Tor Abstract: Search engine queries contain a great

More information

Report: Privacy-Preserving Classification on Deep Neural Network

Report: Privacy-Preserving Classification on Deep Neural Network Report: Privacy-Preserving Classification on Deep Neural Network Janno Veeorg Supervised by Helger Lipmaa and Raul Vicente Zafra May 25, 2017 1 Introduction In this report we consider following task: how

More information

A Quick Guide on Training a neural network using Keras.

A Quick Guide on Training a neural network using Keras. A Quick Guide on Training a neural network using Keras. TensorFlow and Keras Keras Open source High level, less flexible Easy to learn Perfect for quick implementations Starts by François Chollet from

More information

Machine Learning. MGS Lecture 3: Deep Learning

Machine Learning. MGS Lecture 3: Deep Learning Dr Michel F. Valstar http://cs.nott.ac.uk/~mfv/ Machine Learning MGS Lecture 3: Deep Learning Dr Michel F. Valstar http://cs.nott.ac.uk/~mfv/ WHAT IS DEEP LEARNING? Shallow network: Only one hidden layer

More information

Deep Learning for Embedded Security Evaluation

Deep Learning for Embedded Security Evaluation Deep Learning for Embedded Security Evaluation Emmanuel Prouff 1 1 Laboratoire de Sécurité des Composants, ANSSI, France April 2018, CISCO April 2018, CISCO E. Prouff 1/22 Contents 1. Context and Motivation

More information

ConvolutionalNN's... ConvNet's... deep learnig

ConvolutionalNN's... ConvNet's... deep learnig Deep Learning ConvolutionalNN's... ConvNet's... deep learnig Markus Thaler, TG208 tham@zhaw.ch www.zhaw.ch/~tham Martin Weisenhorn, TB427 weie@zhaw.ch 20.08.2018 1 Neural Networks Classification: up to

More information

POINT CLOUD DEEP LEARNING

POINT CLOUD DEEP LEARNING POINT CLOUD DEEP LEARNING Innfarn Yoo, 3/29/28 / 57 Introduction AGENDA Previous Work Method Result Conclusion 2 / 57 INTRODUCTION 3 / 57 2D OBJECT CLASSIFICATION Deep Learning for 2D Object Classification

More information

Keras: Handwritten Digit Recognition using MNIST Dataset

Keras: Handwritten Digit Recognition using MNIST Dataset Keras: Handwritten Digit Recognition using MNIST Dataset IIT PATNA January 31, 2018 1 / 30 OUTLINE 1 Keras: Introduction 2 Installing Keras 3 Keras: Building, Testing, Improving A Simple Network 2 / 30

More information

Kaggle Data Science Bowl 2017 Technical Report

Kaggle Data Science Bowl 2017 Technical Report Kaggle Data Science Bowl 2017 Technical Report qfpxfd Team May 11, 2017 1 Team Members Table 1: Team members Name E-Mail University Jia Ding dingjia@pku.edu.cn Peking University, Beijing, China Aoxue Li

More information

Study of Residual Networks for Image Recognition

Study of Residual Networks for Image Recognition Study of Residual Networks for Image Recognition Mohammad Sadegh Ebrahimi Stanford University sadegh@stanford.edu Hossein Karkeh Abadi Stanford University hosseink@stanford.edu Abstract Deep neural networks

More information

Machine Learning With Python. Bin Chen Nov. 7, 2017 Research Computing Center

Machine Learning With Python. Bin Chen Nov. 7, 2017 Research Computing Center Machine Learning With Python Bin Chen Nov. 7, 2017 Research Computing Center Outline Introduction to Machine Learning (ML) Introduction to Neural Network (NN) Introduction to Deep Learning NN Introduction

More information

Convolutional Neural Networks

Convolutional Neural Networks Lecturer: Barnabas Poczos Introduction to Machine Learning (Lecture Notes) Convolutional Neural Networks Disclaimer: These notes have not been subjected to the usual scrutiny reserved for formal publications.

More information

Anonymity With Tor. The Onion Router. July 21, Technische Universität München

Anonymity With Tor. The Onion Router. July 21, Technische Universität München The Onion Router Nathan S. Evans Christian Grothoff Technische Universität München July 21, 2011 Overview What is Tor? Motivation Background Material How Tor Works Hidden Services Attacks Specific Attack

More information

An Introduction to Deep Learning with RapidMiner. Philipp Schlunder - RapidMiner Research

An Introduction to Deep Learning with RapidMiner. Philipp Schlunder - RapidMiner Research An Introduction to Deep Learning with RapidMiner Philipp Schlunder - RapidMiner Research What s in store for today?. Things to know before getting started 2. What s Deep Learning anyway? 3. How to use

More information

Fusion of Mini-Deep Nets

Fusion of Mini-Deep Nets Rochester Institute of Technology RIT Scholar Works Theses Thesis/Dissertation Collections 5-2016 Fusion of Mini-Deep Nets Sai Prasad Nooka spn8235@rit.edu Follow this and additional works at: http://scholarworks.rit.edu/theses

More information

Convolutional Layer Pooling Layer Fully Connected Layer Regularization

Convolutional Layer Pooling Layer Fully Connected Layer Regularization Semi-Parallel Deep Neural Networks (SPDNN), Convergence and Generalization Shabab Bazrafkan, Peter Corcoran Center for Cognitive, Connected & Computational Imaging, College of Engineering & Informatics,

More information

Comparing Dropout Nets to Sum-Product Networks for Predicting Molecular Activity

Comparing Dropout Nets to Sum-Product Networks for Predicting Molecular Activity 000 001 002 003 004 005 006 007 008 009 010 011 012 013 014 015 016 017 018 019 020 021 022 023 024 025 026 027 028 029 030 031 032 033 034 035 036 037 038 039 040 041 042 043 044 045 046 047 048 049 050

More information

Deep Learning. Visualizing and Understanding Convolutional Networks. Christopher Funk. Pennsylvania State University.

Deep Learning. Visualizing and Understanding Convolutional Networks. Christopher Funk. Pennsylvania State University. Visualizing and Understanding Convolutional Networks Christopher Pennsylvania State University February 23, 2015 Some Slide Information taken from Pierre Sermanet (Google) presentation on and Computer

More information

Link Prediction for Social Network

Link Prediction for Social Network Link Prediction for Social Network Ning Lin Computer Science and Engineering University of California, San Diego Email: nil016@eng.ucsd.edu Abstract Friendship recommendation has become an important issue

More information

Avoiding The Man on the Wire: Improving Tor s Security with Trust-Aware Path Selection

Avoiding The Man on the Wire: Improving Tor s Security with Trust-Aware Path Selection Avoiding The Man on the Wire: Improving Tor s Security with Trust-Aware Path Selection Aaron Johnson Rob Jansen Aaron D. Jaggard Joan Feigenbaum Paul Syverson (U.S. Naval Research Laboratory) (U.S. Naval

More information

Neural Network Optimization and Tuning / Spring 2018 / Recitation 3

Neural Network Optimization and Tuning / Spring 2018 / Recitation 3 Neural Network Optimization and Tuning 11-785 / Spring 2018 / Recitation 3 1 Logistics You will work through a Jupyter notebook that contains sample and starter code with explanations and comments throughout.

More information

Transfer Learning. Style Transfer in Deep Learning

Transfer Learning. Style Transfer in Deep Learning Transfer Learning & Style Transfer in Deep Learning 4-DEC-2016 Gal Barzilai, Ram Machlev Deep Learning Seminar School of Electrical Engineering Tel Aviv University Part 1: Transfer Learning in Deep Learning

More information

Face Recognition A Deep Learning Approach

Face Recognition A Deep Learning Approach Face Recognition A Deep Learning Approach Lihi Shiloh Tal Perl Deep Learning Seminar 2 Outline What about Cat recognition? Classical face recognition Modern face recognition DeepFace FaceNet Comparison

More information

Deep Learning for Computer Vision with MATLAB By Jon Cherrie

Deep Learning for Computer Vision with MATLAB By Jon Cherrie Deep Learning for Computer Vision with MATLAB By Jon Cherrie 2015 The MathWorks, Inc. 1 Deep learning is getting a lot of attention "Dahl and his colleagues won $22,000 with a deeplearning system. 'We

More information

ECE 5470 Classification, Machine Learning, and Neural Network Review

ECE 5470 Classification, Machine Learning, and Neural Network Review ECE 5470 Classification, Machine Learning, and Neural Network Review Due December 1. Solution set Instructions: These questions are to be answered on this document which should be submitted to blackboard

More information

Lecture 20: Neural Networks for NLP. Zubin Pahuja

Lecture 20: Neural Networks for NLP. Zubin Pahuja Lecture 20: Neural Networks for NLP Zubin Pahuja zpahuja2@illinois.edu courses.engr.illinois.edu/cs447 CS447: Natural Language Processing 1 Today s Lecture Feed-forward neural networks as classifiers simple

More information

Deep Tracking: Biologically Inspired Tracking with Deep Convolutional Networks

Deep Tracking: Biologically Inspired Tracking with Deep Convolutional Networks Deep Tracking: Biologically Inspired Tracking with Deep Convolutional Networks Si Chen The George Washington University sichen@gwmail.gwu.edu Meera Hahn Emory University mhahn7@emory.edu Mentor: Afshin

More information

Comparing Website Fingerprinting Attacks and Defenses

Comparing Website Fingerprinting Attacks and Defenses Comparing Website Fingerprinting Attacks and Defenses Tao Wang Cheriton School of Computer Science University of Waterloo Ian Goldberg Cheriton School of Computer Science University of Waterloo Abstract

More information

CS229 Final Project: Predicting Expected Response Times

CS229 Final Project: Predicting Expected  Response Times CS229 Final Project: Predicting Expected Email Response Times Laura Cruz-Albrecht (lcruzalb), Kevin Khieu (kkhieu) December 15, 2017 1 Introduction Each day, countless emails are sent out, yet the time

More information

CAP 6412 Advanced Computer Vision

CAP 6412 Advanced Computer Vision CAP 6412 Advanced Computer Vision http://www.cs.ucf.edu/~bgong/cap6412.html Boqing Gong April 21st, 2016 Today Administrivia Free parameters in an approach, model, or algorithm? Egocentric videos by Aisha

More information

CPSC 340: Machine Learning and Data Mining. Deep Learning Fall 2016

CPSC 340: Machine Learning and Data Mining. Deep Learning Fall 2016 CPSC 340: Machine Learning and Data Mining Deep Learning Fall 2016 Assignment 5: Due Friday. Assignment 6: Due next Friday. Final: Admin December 12 (8:30am HEBB 100) Covers Assignments 1-6. Final from

More information

Measuring Information Leakage in Website Fingerprinting Attacks

Measuring Information Leakage in Website Fingerprinting Attacks Measuring Information Leakage in Website Fingerprinting Attacks Shuai Li University of Minnesota Minneapolis, USA shuai@cs.umn.edu Huajun Guo University of Minnesota Minneapolis, USA guoxx66@umn.edu Nicholas

More information

Convolutional Neural Networks. Computer Vision Jia-Bin Huang, Virginia Tech

Convolutional Neural Networks. Computer Vision Jia-Bin Huang, Virginia Tech Convolutional Neural Networks Computer Vision Jia-Bin Huang, Virginia Tech Today s class Overview Convolutional Neural Network (CNN) Training CNN Understanding and Visualizing CNN Image Categorization:

More information

CS489/698: Intro to ML

CS489/698: Intro to ML CS489/698: Intro to ML Lecture 14: Training of Deep NNs Instructor: Sun Sun 1 Outline Activation functions Regularization Gradient-based optimization 2 Examples of activation functions 3 5/28/18 Sun Sun

More information

SEMANTIC COMPUTING. Lecture 8: Introduction to Deep Learning. TU Dresden, 7 December Dagmar Gromann International Center For Computational Logic

SEMANTIC COMPUTING. Lecture 8: Introduction to Deep Learning. TU Dresden, 7 December Dagmar Gromann International Center For Computational Logic SEMANTIC COMPUTING Lecture 8: Introduction to Deep Learning Dagmar Gromann International Center For Computational Logic TU Dresden, 7 December 2018 Overview Introduction Deep Learning General Neural Networks

More information

MoonRiver: Deep Neural Network in C++

MoonRiver: Deep Neural Network in C++ MoonRiver: Deep Neural Network in C++ Chung-Yi Weng Computer Science & Engineering University of Washington chungyi@cs.washington.edu Abstract Artificial intelligence resurges with its dramatic improvement

More information

Vulnerability of machine learning models to adversarial examples

Vulnerability of machine learning models to adversarial examples Vulnerability of machine learning models to adversarial examples Petra Vidnerová Institute of Computer Science The Czech Academy of Sciences Hora Informaticae 1 Outline Introduction Works on adversarial

More information

Accelerometer Gesture Recognition

Accelerometer Gesture Recognition Accelerometer Gesture Recognition Michael Xie xie@cs.stanford.edu David Pan napdivad@stanford.edu December 12, 2014 Abstract Our goal is to make gesture-based input for smartphones and smartwatches accurate

More information

Predicting Messaging Response Time in a Long Distance Relationship

Predicting Messaging Response Time in a Long Distance Relationship Predicting Messaging Response Time in a Long Distance Relationship Meng-Chen Shieh m3shieh@ucsd.edu I. Introduction The key to any successful relationship is communication, especially during times when

More information

Hello Edge: Keyword Spotting on Microcontrollers

Hello Edge: Keyword Spotting on Microcontrollers Hello Edge: Keyword Spotting on Microcontrollers Yundong Zhang, Naveen Suda, Liangzhen Lai and Vikas Chandra ARM Research, Stanford University arxiv.org, 2017 Presented by Mohammad Mofrad University of

More information

Inception Network Overview. David White CS793

Inception Network Overview. David White CS793 Inception Network Overview David White CS793 So, Leonardo DiCaprio dreams about dreaming... https://m.media-amazon.com/images/m/mv5bmjaxmzy3njcxnf5bml5banbnxkftztcwnti5otm0mw@@._v1_sy1000_cr0,0,675,1 000_AL_.jpg

More information

Research on Pruning Convolutional Neural Network, Autoencoder and Capsule Network

Research on Pruning Convolutional Neural Network, Autoencoder and Capsule Network Research on Pruning Convolutional Neural Network, Autoencoder and Capsule Network Tianyu Wang Australia National University, Colledge of Engineering and Computer Science u@anu.edu.au Abstract. Some tasks,

More information

Lecture 2 Notes. Outline. Neural Networks. The Big Idea. Architecture. Instructors: Parth Shah, Riju Pahwa

Lecture 2 Notes. Outline. Neural Networks. The Big Idea. Architecture. Instructors: Parth Shah, Riju Pahwa Instructors: Parth Shah, Riju Pahwa Lecture 2 Notes Outline 1. Neural Networks The Big Idea Architecture SGD and Backpropagation 2. Convolutional Neural Networks Intuition Architecture 3. Recurrent Neural

More information

Quo Vadis, Action Recognition? A New Model and the Kinetics Dataset. By Joa õ Carreira and Andrew Zisserman Presenter: Zhisheng Huang 03/02/2018

Quo Vadis, Action Recognition? A New Model and the Kinetics Dataset. By Joa õ Carreira and Andrew Zisserman Presenter: Zhisheng Huang 03/02/2018 Quo Vadis, Action Recognition? A New Model and the Kinetics Dataset By Joa õ Carreira and Andrew Zisserman Presenter: Zhisheng Huang 03/02/2018 Outline: Introduction Action classification architectures

More information

CMU Lecture 18: Deep learning and Vision: Convolutional neural networks. Teacher: Gianni A. Di Caro

CMU Lecture 18: Deep learning and Vision: Convolutional neural networks. Teacher: Gianni A. Di Caro CMU 15-781 Lecture 18: Deep learning and Vision: Convolutional neural networks Teacher: Gianni A. Di Caro DEEP, SHALLOW, CONNECTED, SPARSE? Fully connected multi-layer feed-forward perceptrons: More powerful

More information

Character Recognition from Google Street View Images

Character Recognition from Google Street View Images Character Recognition from Google Street View Images Indian Institute of Technology Course Project Report CS365A By Ritesh Kumar (11602) and Srikant Singh (12729) Under the guidance of Professor Amitabha

More information

CS 6501: Deep Learning for Computer Graphics. Training Neural Networks II. Connelly Barnes

CS 6501: Deep Learning for Computer Graphics. Training Neural Networks II. Connelly Barnes CS 6501: Deep Learning for Computer Graphics Training Neural Networks II Connelly Barnes Overview Preprocessing Initialization Vanishing/exploding gradients problem Batch normalization Dropout Additional

More information

Detecting Network Intrusions

Detecting Network Intrusions Detecting Network Intrusions Naveen Krishnamurthi, Kevin Miller Stanford University, Computer Science {naveenk1, kmiller4}@stanford.edu Abstract The purpose of this project is to create a predictive model

More information

Machine Learning and Algorithms for Data Mining Practical 2: Neural Networks

Machine Learning and Algorithms for Data Mining Practical 2: Neural Networks CST Part III/MPhil in Advanced Computer Science 2016-2017 Machine Learning and Algorithms for Data Mining Practical 2: Neural Networks Demonstrators: Petar Veličković, Duo Wang Lecturers: Mateja Jamnik,

More information

AUTOMATED STUDENT S ATTENDANCE ENTERING SYSTEM BY ELIMINATING FORGE SIGNATURES

AUTOMATED STUDENT S ATTENDANCE ENTERING SYSTEM BY ELIMINATING FORGE SIGNATURES AUTOMATED STUDENT S ATTENDANCE ENTERING SYSTEM BY ELIMINATING FORGE SIGNATURES K. P. M. L. P. Weerasinghe 149235H Faculty of Information Technology University of Moratuwa June 2017 AUTOMATED STUDENT S

More information

Weighted Convolutional Neural Network. Ensemble.

Weighted Convolutional Neural Network. Ensemble. Weighted Convolutional Neural Network Ensemble Xavier Frazão and Luís A. Alexandre Dept. of Informatics, Univ. Beira Interior and Instituto de Telecomunicações Covilhã, Portugal xavierfrazao@gmail.com

More information

Data Visualization and Classification of Artificially Created Images

Data Visualization and Classification of Artificially Created Images Central Washington University ScholarWorks@CWU All Master's Theses Master's Theses Winter 2018 Data Visualization and Classification of Artificially Created Images Dmytro Dovhalets dovhaletsd@gmail.com

More information

Lecture #11: The Perceptron

Lecture #11: The Perceptron Lecture #11: The Perceptron Mat Kallada STAT2450 - Introduction to Data Mining Outline for Today Welcome back! Assignment 3 The Perceptron Learning Method Perceptron Learning Rule Assignment 3 Will be

More information

Outlier detection using autoencoders

Outlier detection using autoencoders Outlier detection using autoencoders August 19, 2016 Author: Olga Lyudchik Supervisors: Dr. Jean-Roch Vlimant Dr. Maurizio Pierini CERN Non Member State Summer Student Report 2016 Abstract Outlier detection

More information

End-to-end Driving Controls Predictions from Images

End-to-end Driving Controls Predictions from Images End-to-end Driving Controls Predictions from Images Jan Felix Heyse and Maxime Bouton CS229 Final Project Report December 17, 2016 Abstract Autonomous driving is a promising technology to improve transportation

More information

Deep neural networks II

Deep neural networks II Deep neural networks II May 31 st, 2018 Yong Jae Lee UC Davis Many slides from Rob Fergus, Svetlana Lazebnik, Jia-Bin Huang, Derek Hoiem, Adriana Kovashka, Why (convolutional) neural networks? State of

More information

Machine Learning Methods for Ship Detection in Satellite Images

Machine Learning Methods for Ship Detection in Satellite Images Machine Learning Methods for Ship Detection in Satellite Images Yifan Li yil150@ucsd.edu Huadong Zhang huz095@ucsd.edu Qianfeng Guo qig020@ucsd.edu Xiaoshi Li xil758@ucsd.edu Abstract In this project,

More information

arxiv: v2 [cs.cr] 20 Sep 2017

arxiv: v2 [cs.cr] 20 Sep 2017 How Unique is Your.onion? An Analysis of the Fingerprintability of Tor Onion Services Rebekah Overdorf Drexel University Philadelphia, Pennsylvania rebekah.overdorf@drexel.edu Marc Juarez ESAT-COSIC and

More information

A FRAMEWORK OF EXTRACTING MULTI-SCALE FEATURES USING MULTIPLE CONVOLUTIONAL NEURAL NETWORKS. Kuan-Chuan Peng and Tsuhan Chen

A FRAMEWORK OF EXTRACTING MULTI-SCALE FEATURES USING MULTIPLE CONVOLUTIONAL NEURAL NETWORKS. Kuan-Chuan Peng and Tsuhan Chen A FRAMEWORK OF EXTRACTING MULTI-SCALE FEATURES USING MULTIPLE CONVOLUTIONAL NEURAL NETWORKS Kuan-Chuan Peng and Tsuhan Chen School of Electrical and Computer Engineering, Cornell University, Ithaca, NY

More information

Deep Learning Applications

Deep Learning Applications October 20, 2017 Overview Supervised Learning Feedforward neural network Convolution neural network Recurrent neural network Recursive neural network (Recursive neural tensor network) Unsupervised Learning

More information

Improving the way neural networks learn Srikumar Ramalingam School of Computing University of Utah

Improving the way neural networks learn Srikumar Ramalingam School of Computing University of Utah Improving the way neural networks learn Srikumar Ramalingam School of Computing University of Utah Reference Most of the slides are taken from the third chapter of the online book by Michael Nielson: neuralnetworksanddeeplearning.com

More information

Sentiment Analysis for Amazon Reviews

Sentiment Analysis for Amazon Reviews Sentiment Analysis for Amazon Reviews Wanliang Tan wanliang@stanford.edu Xinyu Wang xwang7@stanford.edu Xinyu Xu xinyu17@stanford.edu Abstract Sentiment analysis of product reviews, an application problem,

More information

Classification Algorithms in Data Mining

Classification Algorithms in Data Mining August 9th, 2016 Suhas Mallesh Yash Thakkar Ashok Choudhary CIS660 Data Mining and Big Data Processing -Dr. Sunnie S. Chung Classification Algorithms in Data Mining Deciding on the classification algorithms

More information