_id
stringlengths
40
40
title
stringlengths
8
300
text
stringlengths
0
10k
0a283fb395343cd26984425306ca24c85b09ccdb
Automatic Indexing Based on Bayesian Inference Networks
In this paper, a Bayesian inference network model for automatic indexing with index terms (descriptors) from a prescribed vocabulary is presented. It requires an indexing dictionary with rules mapping terms of the respective subject field onto descriptors and inverted lists for terms occuring in a set of documents of the subject field and descriptors manually assigned to these documents. The indexing dictionary can be derived automatically from a set of manually indexed documents. An application of the network model is described, followed by an indexing example and some experimental results about the indexing performance of the network model.
1ac5b0628ff249c388ff5ca934a9ccbec577cbd7
Beyond Market Baskets: Generalizing Association Rules to Correlations
One of the most well-studied problems in data mining is mining for association rules in market basket data. Association rules, whose significance is measured via support and confidence, are intended to identify rules of the type, “A customer purchasing item A often also purchases item B.” Motivated by the goal of generalizing beyond market baskets and the association rules used with them, we develop the notion of mining rules that identify correlations (generalizing associations), and we consider both the absence and presence of items as a basis for generating rules. We propose measuring significance of associations via the chi-squared test for correlation from classical statistics. This leads to a measure that is upward closed in the itemset lattice, enabling us to reduce the mining problem to the search for a border between correlated and uncorrelated itemsets in the lattice. We develop pruning strategies and devise an efficient algorithm for the resulting problem. We demonstrate its effectiveness by testing it on census data and finding term dependence in a corpus of text documents, as well as on synthetic data.
16afeecd0f4dbccdd8e281e0e7e443bd08681da1
Measuring the Impact of Network Performance on Cloud-Based Speech Recognition Applications
Cloud-based speech recognition systems enhance Web surfing, transportation, health care, etc. For example, using voice commands helps drivers search the Internet without affecting traffic safety risks. User frustration with network traffic problems can affect the usability of these applications. The performance of these type of applications should be robust in difficult network conditions. We evaluate the performance of several client-server speech recognition applications, under various network conditions. We measure transcription delay and accuracy of each application under different packet loss and jitter values. Results of our study show that performance of client-server speech recognition systems is affected by jitter and packet loss; which commonly occur in WiFi and cellular networks.
41f3b1aebde4c342211185d2b5e339a60ceff9e2
Friction modeling in linear chemical-mechanical planarization
In this article, we develop an analytical model of the relationship between the wafer/pad friction and process configuration. We also provide experimental validation of this model for in situ process monitoring. CMP thus demonstrates that the knowledge and methodologies developed for friction modeling and control can be used to advance the understanding, monitoring, and control of semiconductor manufacturing processes. Meanwhile, relevant issues and challenges in real-time monitoring of CMP are presented as sources of future development.
b972f638b7c4ed22e1bcb573520bb232ea88cda5
Efficient, Safe, and Probably Approximately Complete Learning of Action Models
In this paper we explore the theoretical boundaries of planning in a setting where no model of the agent’s actions is given. Instead of an action model, a set of successfully executed plans are given and the task is to generate a plan that is safe, i.e., guaranteed to achieve the goal without failing. To this end, we show how to learn a conservative model of the world in which actions are guaranteed to be applicable. This conservative model is then given to an off-the-shelf classical planner, resulting in a plan that is guaranteed to achieve the goal. However, this reduction from a model-free planning to a model-based planning is not complete: in some cases a plan will not be found even when such exists. We analyze the relation between the number of observed plans and the likelihood that our conservative approach will indeed fail to solve a solvable problem. Our analysis show that the number of trajectories needed scales gracefully.
e38b9f339e858c8ac95679737a0852d21c48d89c
Pressure characteristics at the stump/socket interface in transtibial amputees using an adaptive prosthetic foot.
BACKGROUND The technological advances that have been made in developing highly functional prostheses are promising for very active patients but we do not yet know whether they cause an increase in biomechanical load along with possibly negative consequences for pressure conditions in the socket. Therefore, this study monitored the socket pressure at specific locations of the stump when using a microprocessor-controlled adaptive prosthetic ankle under different walking conditions. METHODS Twelve unilateral transtibial amputees between 43 and 59 years of age were provided with the Proprio-Foot (Ossur) and underwent an instrumented 3D gait analysis in level, stair, and incline walking, including synchronous data capturing of socket pressure. Peak pressures and pressure time integrals (PTI) at three different locations were compared for five walking conditions with and without using the device's ankle adaptation mode. FINDINGS Highest peak pressures of 2.4 k Pa/kg were found for incline ascent at the calf muscle as compared to 2.1 k Pa/kg in level walking with large inter-individual variance. In stair ascent a strong correlation was found between maximum knee moment and socket pressure. The most significant pressure changes relative to level walking were seen in ramp descent anteriorly towards the stump end, with PTI values being almost twice as high as those in level walking. Adapting the angle of the prosthesis on stairs and ramps modified the pressure data such that they were closer to those in level walking. INTERPRETATION Pressure at the stump depends on the knee moments involved in each walking condition. Adapting the prosthetic ankle angle is a valuable means of modifying joint kinetics and thereby the pressure distribution at the stump. However, large inter-individual differences in local pressures underline the importance of individual socket fitting.
6c237c3638eefe1eb39212f801cd857bedc004ee
Exploiting Electronic Health Records to Mine Drug Effects on Laboratory Test Results
The proliferation of Electronic Health Records (EHRs) challenges data miners to discover potential and previously unknown patterns from a large collection of medical data. One of the tasks that we address in this paper is to reveal previously unknown effects of drugs on laboratory test results. We propose a method that leverages drug information to find a meaningful list of drugs that have an effect on the laboratory result. We formulate the problem as a convex non smooth function and develop a proximal gradient method to optimize it. The model has been evaluated on two important use cases: lowering low-density lipoproteins and glycated hemoglobin test results. The experimental results provide evidence that the proposed method is more accurate than the state-of-the-art method, rediscover drugs that are known to lower the levels of laboratory test results, and most importantly, discover additional potential drugs that may also lower these levels.
ebae5af27aafd39358a46c83c1409885773254dd
A Survey of Vehicular Ad hoc Networks Routing Protocols
In recent years, the aspect of vehicular ad hoc network (VANET) is becoming an interesting research area; VANET is a mobile ad hoc network considered as a special case of mobile ad hoc network (MANET). Similar to MANET, VANET is characterized as autonomous and self-configured wireless network. However, VANET has very dynamic topology, large and variable network size, and constrained mobility; these characteristics led to the need for efficient routing and resource saving VANET protocols, to fit with different VANET environments. These differences render traditional MANET’s protocols unsuitable for VANET. The aim of this work is to give a survey of the VANETs routing mechanisms, this paper gives an overview of Vehicular ad hoc networks (VANETs) and the existing VANET routing protocols; mainly it focused on vehicle to vehicle (V2V) communication and protocols. The paper also represents the general outlines and goals of VANETs, investigates different routing schemes that have been developed for VANETs, as well as providing classifications of VANET routing protocols (focusing on two classification forms), and gives summarized comparisons between different classes in the context of their methodologies used, strengths, and limitations of each class scheme compared to other classes. Finally, it extracts the current trends and the challenges for efficient routing mechanisms in VANETs.
d012519a924e41aa7ff49d9b6be58033bd60fd9c
Predicting hospital admission at emergency department triage using machine learning
OBJECTIVE To predict hospital admission at the time of ED triage using patient history in addition to information collected at triage. METHODS This retrospective study included all adult ED visits between March 2014 and July 2017 from one academic and two community emergency rooms that resulted in either admission or discharge. A total of 972 variables were extracted per patient visit. Samples were randomly partitioned into training (80%), validation (10%), and test (10%) sets. We trained a series of nine binary classifiers using logistic regression (LR), gradient boosting (XGBoost), and deep neural networks (DNN) on three dataset types: one using only triage information, one using only patient history, and one using the full set of variables. Next, we tested the potential benefit of additional training samples by training models on increasing fractions of our data. Lastly, variables of importance were identified using information gain as a metric to create a low-dimensional model. RESULTS A total of 560,486 patient visits were included in the study, with an overall admission risk of 29.7%. Models trained on triage information yielded a test AUC of 0.87 for LR (95% CI 0.86-0.87), 0.87 for XGBoost (95% CI 0.87-0.88) and 0.87 for DNN (95% CI 0.87-0.88). Models trained on patient history yielded an AUC of 0.86 for LR (95% CI 0.86-0.87), 0.87 for XGBoost (95% CI 0.87-0.87) and 0.87 for DNN (95% CI 0.87-0.88). Models trained on the full set of variables yielded an AUC of 0.91 for LR (95% CI 0.91-0.91), 0.92 for XGBoost (95% CI 0.92-0.93) and 0.92 for DNN (95% CI 0.92-0.92). All algorithms reached maximum performance at 50% of the training set or less. A low-dimensional XGBoost model built on ESI level, outpatient medication counts, demographics, and hospital usage statistics yielded an AUC of 0.91 (95% CI 0.91-0.91). CONCLUSION Machine learning can robustly predict hospital admission using triage information and patient history. The addition of historical information improves predictive performance significantly compared to using triage information alone, highlighting the need to incorporate these variables into prediction models.
89a523135fc9cb3b0eb6cade2d1eab1b17ea42f4
Stress sensitivity of fault seismicity : a comparison between limited-offset oblique and major strike-slip faults
We present a new three-dimensional inventory of the southern San Francisco Bay area faults and use it to calculate stress applied principally by the 1989 M=7.1 Loma Prieta earthquake, and to compare fault seismicity rates before and after 1989. The major high-angle right-lateral faults exhibit a different response to the stress change than do minor oblique (rightlateral/thrust) faults. Seismicity on oblique-slip faults in the southern Santa Clara Valley thrust belt increased where the faults were unclamped. The strong dependence of seismicity change on normal stress change implies a high coefficient of static friction. In contrast, we observe that faults with significant offset (> 50-100 km) behave differently; microseismicity on the Hayward fault diminished where right-lateral shear stress was reduced, and where it was unclamped by the Loma Prieta earthquake. We observe a similar response on the San Andreas fault zone in southern California after the Landers earthquake sequence. Additionally, the offshore San Gregorio fault shows a seismicity rate increase where right-lateral/oblique shear stress was increased by the Loma Prieta earthquake despite also being clamped by it. These responses are consistent with either a low coefficient of static friction or high pore fluid pressures within the fault zones. We can explain the different behavior of the two styles of faults if those with large cumulative offset become impermeable through gouge buildup; coseismically pressurized pore fluids could be trapped and negate imposed normal stress changes, whereas in more limited offset faults fluids could rapidly escape. The difference in behavior between minor and major faults may explain why frictional failure criteria that apply intermediate coefficients of static friction can be effective in describing the broad distributions of aftershocks that follow large earthquakes, since many of these events occur both inside and outside major fault zones.
5d6959c6d37ed3cc910cd436865a4c2a73284c7c
Photoplethysmography and its detailed pulse waveform analysis for arterial stiffness
Arterial stiffness index is one of the biomechanical indices of vascular healthiness. These indexes are based on detailed pulse waveform analysis which is presented here. After photoplethysmographyic (PPG) pulse wave measurement, we decompose the pulse waveform for the estimation and determination of arterial elasticity. Firstly, it is electro-optically measured PPG signal and by electromechanical film (EMFi) measured signal that are analyzed and investigated by dividing each wave into five logarithmic normal function components. For both the PPG and EMFi waveform we can find very easily a good fit between the original and overlapped and summed wave components. Each wave component is assumed to resemble certain phenomenon in the arteries and certain indexes can be calculated for example based on the mutual timing of the components. Several studies have demonstrated that these kinds of indexes calculated based on actual biomechanical processed can predict future cardiovascular events. Many dynamic factors, e.g., arterial stiffness, depend on fixed structural features of the arterial wall. For more accurate description, arterial stiffness is estimated based on pulse wave decomposition analysis in the radial measured by EMFi and PPG method and tibial arterial walls measured by PPG method parallelly. Elucidation of the precise relationship between endothelial function and arterial stiffness can be done through biomechanics. However, arterial wall elasticity awaits still further biomechanical studies with clinical relations and the influence of arterial flexibility, resistance and ageing inside of the radial pulse waveform.
cb0c85c4eb75016a7098ca0c452e13812b9c95e9
Iterated learning and the evolution of language
Iterated learning describes the process whereby an individual learns their behaviour by exposure to another individual's behaviour, who themselves learnt it in the same way. It can be seen as a key mechanism of cultural evolution. We review various methods for understanding how behaviour is shaped by the iterated learning process: computational agent-based simulations; mathematical modelling; and laboratory experiments in humans and non-human animals. We show how this framework has been used to explain the origins of structure in language, and argue that cultural evolution must be considered alongside biological evolution in explanations of language origins.
dc53c638f58bf3982c5a6ed82002d56c955763c2
Effective and efficient correlation analysis with application to Market Basket Analysis and Network Community Detection
Finding the most interesting correlations among items is essential for problems in many commercial, medical, and scientific domains. For example, what kinds of items should be recommended with regard to what has been purchased by a customer? How to arrange the store shelf in order to increase sales? How to partition the whole social network into several communities for successful advertising campaigns? Which set of individuals on a social network should we target to convince in order to trigger a large cascade of further adoptions? When conducting correlation analysis, traditional methods have both effectiveness and efficiency problems, which will be addressed in this dissertation. Here, we explore the effectiveness problem in three ways. First, we expand the set of desirable properties and study the property satisfaction for different correlation measures. Second, we study different techniques to adjust original correlation measure, and propose two new correlation measures: the Simplified χ with Continuity Correction and the Simplified χ with Support. Third, we study the upper and lower bounds of different measures and categorize them by the bound differences. Combining with the above three directions, we provide guidelines for users to choose the proper measure according to their situations. With the proper correlation measure, we start to solve the efficiency problem for a large dataset. Here, we propose a fully-correlated itemset (FCI) framework to decouple the correlation measure from the need for efficient search. By wrapping the desired measure in our FCI framework, we take advantage of the desired measure’s superiority in evaluating itemsets, eliminate itemsets with irrelevant items, and achieve good computational performance. In addition, we identify a 1-dimensional monotone property of the upper bound of any good correlation measure, and different 2-dimensional
afad9773e74db1927cff4c284dee8afbc4fb849d
Circularly polarized array antenna with corporate-feed network and series-feed elements
In this paper, corporate-feed circularly polarized microstrip array antennas are studied. The antenna element is a series-feed slot-coupled structure. Series feeding causes sequential rotation effect at the element level. Antenna elements are then used to form the subarray by applying sequential rotation to their feeding. Arrays having 4, 16, and 64 elements were made. The maximum achieved gains are 15.3, 21, and 25.4 dBic, respectively. All arrays have less than 15 dB return loss and 3 dB axial ratio from 10 to 13 GHz. The patterns are all quite symmetrical.
5a131856df045cf27a2d5056cea2d2401e2d81b2
When Social Bots Attack: Modeling Susceptibility of Users in Online Social Networks
Social bots are automatic or semi-automatic computer programs that mimic humans and/or human behavior in online social networks. Social bots can attack users (targets) in online social networks to pursue a variety of latent goals, such as to spread information or to influence targets. Without a deep understanding of the nature of such attacks or the susceptibility of users, the potential of social media as an instrument for facilitating discourse or democratic processes is in jeopardy. In this paper, we study data from the Social Bot Challenge 2011 an experiment conducted by the WebEcologyProject during 2011 in which three teams implemented a number of social bots that aimed to influence user behavior on Twitter. Using this data, we aim to develop models to (i) identify susceptible users among a set of targets and (ii) predict users’ level of susceptibility. We explore the predictiveness of three different groups of features (network, behavioral and linguistic features) for these tasks. Our results suggest that susceptible users tend to use Twitter for a conversational purpose and tend to be more open and social since they communicate with many different users, use more social words and show more affection than non-susceptible users.
618c47bc44c3b6fc27067211214afccce6f7cd2c
Speed Estimation and Abnormality Detection from Surveillance Cameras
Motivated by the increasing industry trends towards autonomous driving, vehicles, and transportation we focus on developing a traffic analysis framework for the automatic exploitation of a large pool of available data relative to traffic applications. We propose a cooperative detection and tracking algorithm for the retrieval of vehicle trajectories in video surveillance footage based on deep CNN features that is ultimately used for two separate traffic analysis modalities: (a) vehicle speed estimation based on a state of the art fully automatic camera calibration algorithm and (b) the detection of possibly abnormal events in the scene using robust optical flow descriptors of the detected vehicles and Fisher vector representations of spatiotemporal visual volumes. Finally we measure the performance of our proposed methods in the NVIDIA AI CITY challenge evaluation dataset.
2230381a078241c4385bb9c20b385e8f0da70b9b
Blind detection of photomontage using higher order statistics
We investigate the prospect of using bicoherence features for blind image splicing detection. Image splicing is an essential operation for digital photomontaging, which in turn is a technique for creating image forgery. We examine the properties of bicoherence features on a data set, which contains image blocks of diverse image properties. We then demonstrate the limitation of the baseline bicoherence features for image splicing detection. Our investigation has led to two suggestions for improving the performance of bicoherence features, i.e., estimating the bicoherence features of the authentic counterpart and incorporating features that characterize the variance of the feature performance. The features derived from the suggestions are evaluated with support vector machine (SVM) classification and is shown to improve the image splicing detection accuracy from 62% to about 70%.
1f0396c08e8485358b3d1e13f451ed12ecfc1a77
Building a Question-Answering Chatbot using Forum Data in the Semantic Space
We build a conversational agent which knowledge base is an online forum for parents of autistic children. We collect about 35,000 threads totalling some 600,000 replies, and label 1% of them for usefulness using Amazon Mechanical Turk. We train a Random Forest Classifier using sent2vec features to label the remaining thread replies. Then, we use word2vec to match user queries conceptually with a thread, and then a reply with a predefined context window.
ade7178613e4db90d6a551cb372aebae4c4fa0bf
Time series forecasting of cyber attack intensity
Cyber attacks occur on a near daily basis and are becoming exponentially more common. While some research aims to detect the characteristics of an attack, little focus has been given to patterns of attacks in general. This paper aims to exploit temporal correlations between the number of attacks per day in order to predict future intensity of cyber incidents. Through analysis of attack data collected from Hackmageddon, correlation was found among reported attack volume in consecutive days. This paper presents a forecasting system that aims to predict the number of cyber attacks on a given day based only on a set of historical attack count data. Our system conducts ARIMA time series forecasting on all previously collected incidents to predict the expected number of attacks on a future date. Our tool is able to use only a subset of data relevant to a specific attack method. Prediction models are dynamically updated over time as new data is collected to improve accuracy. Our system outperforms naive forecasting methods by 14.1% when predicting attacks of any type, and up to 21.2% when forecasting attacks of a specific type. Our system also produces a model which more accurately predicts future cyber attack intensity behavior.
c906c9b2daddf67ebd949c71fc707d697065c6a0
Semantics-Aware Machine Learning for Function Recognition in Binary Code
Function recognition in program binaries serves as the foundation for many binary instrumentation and analysis tasks. However, as binaries are usually stripped before distribution, function information is indeed absent in most binaries. By far, identifying functions in stripped binaries remains a challenge. Recent research work proposes to recognize functions in binary code through machine learning techniques. The recognition model, including typical function entry point patterns, is automatically constructed through learning. However, we observed that as previous work only leverages syntax-level features to train the model, binary obfuscation techniques can undermine the pre-learned models in real-world usage scenarios. In this paper, we propose FID, a semantics-based method to recognize functions in stripped binaries. We leverage symbolic execution to generate semantic information and learn the function recognition model through well-performing machine learning techniques.FID extracts semantic information from binary code and, therefore, is effectively adapted to different compilers and optimizations. Moreover, we also demonstrate that FID has high recognition accuracy on binaries transformed by widely-used obfuscation techniques. We evaluate FID with over four thousand test cases. Our evaluation shows that FID is comparable with previous work on normal binaries and it notably outperforms existing tools on obfuscated code.
df6b604d1352d4bd81604730f9000d7a29574384
eyond virtual museums : Experiencing immersive virtual reality in eal museums
Contemporary museums are much more than places devoted to the placement and the exhibition of collections and artworks; indeed, they are nowadays considered as a privileged means for communication and play a central role in making culture accessible to the mass audience. One of the keys to approach the general public is the use of new technologies and novel interaction paradigms. These means, which bring with them an undeniable appeal, allow curators to modulate the cultural proposal by structuring different courses for different user profiles. Immersive Virtual reality (VR) is probably one of the most appealing and potentially effective technologies to serve this purpose; nevertheless, it is still quite uncommon to find immersive installations in museums. Starting from our 10 years’ experience in this topic, and following an in-depth survey about these technologies and their use in cultural contexts, we propose a classification of VR installations, specifically oriented to cultural heritage applications, based on their features in terms of interaction and immersion. On the basis of this classification, aiming to provide a tool for framing VR systems which would hopefully suggest indications related to costs, usability and quality of the sensorial experience, we analyze a series of live examples of which we point out strengths and weak points. We then summarize the current state and the very next future, identifying the major issues that prevent these technologies from being actually widespread, and outline proposals for a more se of pervasive and effective u
9aade3d26996ce7ef6d657130464504b8d812534
Face Alignment With Deep Regression
In this paper, we present a deep regression approach for face alignment. The deep regressor is a neural network that consists of a global layer and multistage local layers. The global layer estimates the initial face shape from the whole image, while the following local layers iteratively update the shape with local image observations. Combining standard derivations and numerical approximations, we make all layers able to backpropagate error differentials, so that we can apply the standard backpropagation to jointly learn the parameters from all layers. We show that the resulting deep regressor gradually and evenly approaches the true facial landmarks stage by stage, avoiding the tendency that often occurs in the cascaded regression methods and deteriorates the overall performance: yielding early stage regressors with high alignment accuracy gains but later stage regressors with low alignment accuracy gains. Experimental results on standard benchmarks demonstrate that our approach brings significant improvements over previous cascaded regression algorithms.
f85ccab7173e543f2bfd4c7a81fb14e147695740
A method to infer emotions from facial Action Units
We present a robust method to map detected facial Action Units (AUs) to six basic emotions. Automatic AU recognition is prone to errors due to illumination, tracking failures and occlusions. Hence, traditional rule based methods to map AUs to emotions are very sensitive to false positives and misses among the AUs. In our method, a set of chosen AUs are mapped to the six basic emotions using a learned statistical relationship and a suitable matching technique. Relationships between the AUs and emotions are captured as template strings comprising the most discriminative AUs for each emotion. The template strings are computed using a concept called discriminative power. The Longest Common Subsequence (LCS) distance, an approach for approximate string matching, is applied to calculate the closeness of a test string of AUs with the template strings, and hence infer the underlying emotions. LCS is found to be efficient in handling practical issues like erroneous AU detection and helps to reduce false predictions. The proposed method is tested with various databases like CK+, ISL, FACS, JAFFE, MindReading and many real-world video frames. We compare our performance with rule based techniques, and show clear improvement on both benchmark databases and real-world datasets.
7539293eaadec85917bcfcf4ecc53e7bdd41c227
Using phrases and document metadata to improve topic modeling of clinical reports
Probabilistic topic models provide an unsupervised method for analyzing unstructured text, which have the potential to be integrated into clinical automatic summarization systems. Clinical documents are accompanied by metadata in a patient's medical history and frequently contains multiword concepts that can be valuable for accurately interpreting the included text. While existing methods have attempted to address these problems individually, we present a unified model for free-text clinical documents that integrates contextual patient- and document-level data, and discovers multi-word concepts. In the proposed model, phrases are represented by chained n-grams and a Dirichlet hyper-parameter is weighted by both document-level and patient-level context. This method and three other Latent Dirichlet allocation models were fit to a large collection of clinical reports. Examples of resulting topics demonstrate the results of the new model and the quality of the representations are evaluated using empirical log likelihood. The proposed model was able to create informative prior probabilities based on patient and document information, and captured phrases that represented various clinical concepts. The representation using the proposed model had a significantly higher empirical log likelihood than the compared methods. Integrating document metadata and capturing phrases in clinical text greatly improves the topic representation of clinical documents. The resulting clinically informative topics may effectively serve as the basis for an automatic summarization system for clinical reports.
291c3f4393987f67cded328e984dbae84af643cb
Faster Dynamic Programming for Markov Decision Processes
OF THESIS FASTER DYNAMIC PROGRAMMING FOR MARKOV DECISION PROCESSES Markov decision processes (MDPs) are a general framework used by Artificial Intelligence (AI) researchers to model decision theoretic planning problems. Solving real world MDPs has been a major and challenging research topic in the AI literature. This paper discusses two main groups of approaches in solving MDPs. The first group of approaches combines the strategies of heuristic search and dynamic programming to expedite the convergence process. The second makes use of graphical structures in MDPs to decrease the effort of classic dynamic programming algorithms. Two new algorithms proposed by the author, MBLAO* and TVI, are described here.
370063c5491147d88d57bbcd865eb5004484c1eb
A Review of Technical Approaches to Realizing Near-Field Communication Mobile Payments
This article describes and compares four approaches to storing payment keys and executing payment applications on mobile phones via near-field communication at the point of sale. Even though the comparison hinges on security--specifically, how well the keys and payment application are protected against misuse--other criteria such as hardware requirements, availability, management complexity, and performance are also identified and discussed.
21e480ad39c52d8e770810f8319750a34f8bc091
Exploiting geographic dependencies for real estate appraisal: a mutual perspective of ranking and clustering
It is traditionally a challenge for home buyers to understand, compare and contrast the investment values of real estates. While a number of estate appraisal methods have been developed to value real property, the performances of these methods have been limited by the traditional data sources for estate appraisal. However, with the development of new ways of collecting estate-related mobile data, there is a potential to leverage geographic dependencies of estates for enhancing estate appraisal. Indeed, the geographic dependencies of the value of an estate can be from the characteristics of its own neighborhood (individual), the values of its nearby estates (peer), and the prosperity of the affiliated latent business area (zone). To this end, in this paper, we propose a geographic method, named ClusRanking, for estate appraisal by leveraging the mutual enforcement of ranking and clustering power. ClusRanking is able to exploit geographic individual, peer, and zone dependencies in a probabilistic ranking model. Specifically, we first extract the geographic utility of estates from geography data, estimate the neighborhood popularity of estates by mining taxicab trajectory data, and model the influence of latent business areas via ClusRanking. Also, we use a linear model to fuse these three influential factors and predict estate investment values. Moreover, we simultaneously consider individual, peer and zone dependencies, and derive an estate-specific ranking likelihood as the objective function. Finally, we conduct a comprehensive evaluation with real-world estate related data, and the experimental results demonstrate the effectiveness of our method.
044a9cb24e2863c6bcaaf39b7a210fbb11b381e9
A Low-Bandwidth Network File System
Users rarely consider running network file systems over slow or wide-area networks, as the performance would be unacceptable and the bandwidth consumption too high. Nonetheless, efficient remote file access would often be desirable over such networks---particularly when high latency makes remote login sessions unresponsive. Rather than run interactive programs such as editors remotely, users could run the programs locally and manipulate remote files through the file system. To do so, however, would require a network file system that consumes less bandwidth than most current file systems.This paper presents LBFS, a network file system designed for low-bandwidth networks. LBFS exploits similarities between files or versions of the same file to save bandwidth. It avoids sending data over the network when the same data can already be found in the server's file system or the client's cache. Using this technique in conjunction with conventional compression and caching, LBFS consumes over an order of magnitude less bandwidth than traditional network file systems on common workloads.
964b6997a9c7852deff71d34894dbdc38d34fbdf
Algorithms for Delta Compression and Remote File Synchronization
Delta compression and remote file synchronization techniques are concerned with efficient file transfer over a slow communication link in the case where the receiving party already has a similar file (or files). This problem arises naturally, e.g., when distributing updated versions of software over a network or synchronizing personal files between different accounts and devices. More generally, the problem is becoming increasingly common in many networkbased applications where files and content are widely replicated, frequently modified, and cut and reassembled in different contexts and packagings. In this chapter, we survey techniques, software tools, and applications for delta compression, remote file synchronization, and closely related problems. We first focus on delta compression, where the sender knows all the similar files that are held by the receiver. In the second part, we survey work on the related, but in many ways quite different, problem of remote file synchronization, where the sender does not have a copy of the files held by the receiver. Work supported by NSF CAREER Award NSF CCR-0093400 and by Intel Corporation.
148ec401da7d5859a9488c0f9a34200de71cc824
Leases: An Efficient Fault-Tolerant Mechanism for Distributed File Cache Consistency
Caching introduces the overhead and complexity of ensuring consistency, reducing some of its performance benefits. In a distributed system, caching must deal with the additional complications of communication and host failures. Leases are proposed as a time-based mechanism that provides efficient consistent access to cached data in distributed systems. Non-Byzantine failures affect performance, not correctness, with their effect minimized by short leases. An analytic model and an evaluation for file access in the V system show that leases of short duration provide good performance. The impact of leases on performance grows more significant in systems of larger scale and higher processor performance.
41f1abe566060e53ad93d8cfa8c39ac582256868
Implementing Fault-Tolerant Services Using the State Machine Approach: A Tutorial
The state machine approach is a general method for implementing fault-tolerant services in distributed systems. This paper reviews the approach and describes protocols for two different failure models—Byzantine and fail stop. Systems reconfiguration techniques for removing faulty components and integrating repaired components are also discussed.
46f766c11df69808453e14c900bcb3f4e081fcae
Copy Detection Mechanisms for Digital Documents
In a digital library system, documents are available in digital form and therefore are more easily copied and their copyrights are more easily violated. This is a very serious problem, as it discourages owners of valuable information from sharing it with authorized users. There are two main philosophies for addressing this problem: prevention and detection. The former actually makes unauthorized use of documents difficult or impossible while the latter makes it easier to discover such activity.In this paper we propose a system for registering documents and then detecting copies, either complete copies or partial copies. We describe algorithms for such detection, and metrics required for evaluating detection mechanisms (covering accuracy, efficiency, and security). We also describe a working prototype, called COPS, describe implementation issues, and present experimental results that suggest the proper settings for copy detection parameters.
6db68f27bcb7c9c001bb0c144c1d0ac5d69a3f3a
Formal Analysis of Graphical Security Models
Copyright and moral rights for the publications made accessible in the public portal are retained by the authors and/or other copyright owners and it is a condition of accessing publications that users recognise and abide by the legal requirements associated with these rights. • Users may download and print one copy of any publication from the public portal for the purpose of private study or research. • You may not further distribute the material or use it for any profit-making activity or commercial gain • You may freely distribute the URL identifying the publication in the public portal ? If you believe that this document breaches copyright please contact us providing details, and we will remove access to the work immediately and investigate your claim. May not music be described as the mathematics of the sense, mathematics as music of the reason? The musician feels mathematics, the mathematician thinks music: music the dream, mathematics the working life. Summary The increasing usage of computer-based systems in almost every aspect of our daily life makes more and more dangerous the threat posed by potential attackers , and more and more rewarding a successful attack. Moreover, the complexity of these systems is also increasing, including physical devices, software components and human actors interacting with each other to form so-called socio-technical systems. The importance of socio-technical systems to modern societies requires verifying their security properties formally, while their inherent complexity makes manual analyses impracticable. Graphical models for security offer an unrivalled opportunity to describe socio-technical systems, for they allow to represent different aspects like human behaviour, computation and physical phenomena in an abstract yet uniform manner. Moreover, these models can be assigned a formal semantics, thereby allowing formal verification of their properties. Finally, their appealing graphical notations enable to communicate security concerns in an understandable way also to non-experts, often in charge of the decision making. This dissertation argues that automated techniques can be developed on graphical security models to evaluate qualitative and quantitative security properties of socio-technical systems and to synthesise optimal attack and defence strategies. In support to this claim we develop analysis techniques for widely-used graphical security models such as attack trees and attack-defence trees. Our analyses cope with the optimisation of multiple parameters of an attack and defence scenario. Improving on the literature, in case of conflicting parameters such as probability and cost we compute the set of optimal solutions …
e88eec15946dd19bdcf69db882f204386e05ff48
Robust techniques for background subtraction in urban traffic video
Identifying moving objects from a video sequence is a fundamental and critical task in many computer-vision applications. A common approach is to perform background subtraction, which identifies moving objects from the portion of a video frame that differs significantly from a background model. There are many challenges in developing a good background subtraction algorithm. First, it must be robust against changes in illumination. Second, it should avoid detecting non-stationary background objects such as swinging leaves, rain, snow, and shadow cast by moving objects. Finally, its internal background model should react quickly to changes in background such as starting and stopping of vehicles. In this paper, we compare various background subtraction algorithms for detecting moving vehicles and pedestrians in urban traffic video sequences. We consider approaches varying from simple techniques such as frame differencing and adaptive median filtering, to more sophisticated probabilistic modeling techniques. While complicated techniques often produce superior performance, our experiments show that simple techniques such as adaptive median filtering can produce good results with much lower computational complexity.
2232416778783616736149c870a69beb13cda743
Face Recognition in a Meeting Room
In this paper, weinvestigaterecognition of humanfaces in a meetingroom. The major challenges of identifying humanfacesin this environmentincludelow quality of input images,poor illumination,unrestrictedheadposesand continuouslychangingfacial expressionsandocclusion.In order to addresstheseproblemswe proposea novel algorithm, DynamicSpaceWarping (DSW).Thebasic idea of the algorithm is to combinelocal features under certain spatial constraints. We compare DSWwith the eigenface approachondatacollectedfromvariousmeetings.Wehave testedboth front and profile face imagesand imageswith two stagesof occlusion.Theexperimentalresultsindicate thattheDSWapproachoutperformstheeigenfaceapproach in bothcases.
74c24d7454a2408f766e4d9e507a0e9c3d80312f
A Provably-Secure ECC-Based Authentication Scheme for Wireless Sensor Networks
A smart-card-based user authentication scheme for wireless sensor networks (in short, a SUA-WSN scheme) is designed to restrict access to the sensor data only to users who are in possession of both a smart card and the corresponding password. While a significant number of SUA-WSN schemes have been suggested in recent years, their intended security properties lack formal definitions and proofs in a widely-accepted model. One consequence is that SUA-WSN schemes insecure against various attacks have proliferated. In this paper, we devise a security model for the analysis of SUA-WSN schemes by extending the widely-accepted model of Bellare, Pointcheval and Rogaway (2000). Our model provides formal definitions of authenticated key exchange and user anonymity while capturing side-channel attacks, as well as other common attacks. We also propose a new SUA-WSN scheme based on elliptic curve cryptography (ECC), and prove its security properties in our extended model. To the best of our knowledge, our proposed scheme is the first SUA-WSN scheme that provably achieves both authenticated key exchange and user anonymity. Our scheme is also computationally competitive with other ECC-based (non-provably secure) schemes.
853860b6472b2c883be4085a3460042fc8b1af3e
Laplacian Auto-Encoders: An explicit learning of nonlinear data manifold
A key factor contributing to the success of many auto-encoders based deep learning techniques is the implicit consideration of the underlying data manifold in their training criteria. In this paper, we aim to make this consideration more explicit by training auto-encoders completely from the manifold learning perspective. We propose a novel unsupervised manifold learning method termed Laplacian Auto-Encoders (LAEs). Starting from a general regularized function learning framework, LAE regularizes training of autoencoders so that the learned encoding function has the locality-preserving property for data points on the manifold. By exploiting the analog relation between the graph Laplacian and the Laplace–Beltrami operator on the continuous manifold, we derive discrete approximations of the firstand higher-order auto-encoder regularizers that can be applied in practical scenarios, where only data points sampled from the distribution on the manifold are available. Our proposed LAE has potentially better generalization capability, due to its explicit respect of the underlying data manifold. Extensive experiments on benchmark visual classification datasets show that LAE consistently outperforms alternative auto-encoders recently proposed in deep learning literature, especially when training samples are relatively scarce. & 2015 Elsevier B.V. All rights reserved.
81fd1a1f72963ce16ddebacea82e71dab3d6992a
Interactive surface design with interlocking elements
We present an interactive tool for designing physical surfaces made from flexible interlocking quadrilateral elements of a single size and shape. With the element shape fixed, the design task becomes one of finding a discrete structure---i.e., element connectivity and binary orientations---that leads to a desired geometry. In order to address this challenging problem of combinatorial geometry, we propose a forward modeling tool that allows the user to interactively explore the space of feasible designs. Paralleling principles from conventional modeling software, our approach leverages a library of base shapes that can be instantiated, combined, and extended using two fundamental operations: merging and extrusion. In order to assist the user in building the designs, we furthermore propose a method to automatically generate assembly instructions. We demonstrate the versatility of our method by creating a diverse set of digital and physical examples that can serve as personalized lamps or decorative items.
90a5393b72b85ec21fae9a108ed5dd3e99837701
The Role of Instructional Design in Persuasion: A Comics Approach for Improving Cybersecurity
Although computer security technologies are the first line of defence to secure users, their success is dependent on individuals’ behaviour. It is therefore necessary to persuade users to practice good computer security. Our interview analysis of users’ conceptualization of security password guessing attacks, antivirus protection, and mobile online privacy shows that poor understanding of security threats influences users’ motivation and ability to practice safe behaviours. We designed and developed an online interactive comic series called Secure Comics based on instructional design principles to address this problem. An eye-tracking experiment suggests that the graphical and interactive components of the comics direct users’ attention and facilitate comprehension of the information. In our evaluations of Secure Comics, results from several user studies show that the comics improve understanding and motivate positive changes in security management behaviour. We discuss the implication of the findings to better understand the role of instructional design and persuasion in education technology.
4c9f9f228390d1370e0df91d2565a2559444431d
SimRT: an automated framework to support regression testing for data races
Concurrent programs are prone to various classes of difficult-to-detect faults, of which data races are particularly prevalent. Prior work has attempted to increase the cost-effectiveness of approaches for testing for data races by employing race detection techniques, but to date, no work has considered cost-effective approaches for re-testing for races as programs evolve. In this paper we present SimRT, an automated regression testing framework for use in detecting races introduced by code modifications. SimRT employs a regression test selection technique, focused on sets of program elements related to race detection, to reduce the number of test cases that must be run on a changed program to detect races that occur due to code modifications, and it employs a test case prioritization technique to improve the rate at which such races are detected. Our empirical study of SimRT reveals that it is more efficient and effective for revealing races than other approaches, and that its constituent test selection and prioritization components each contribute to its performance.
9da15b932df57a8f959471ebc977d620efb18cc1
PicToSeek: combining color and shape invariant features for image retrieval
We aim at combining color and shape invariants for indexing and retrieving images. To this end, color models are proposed independent of the object geometry, object pose, and illumination. From these color models, color invariant edges are derived from which shape invariant features are computed. Computational methods are described to combine the color and shape invariants into a unified high-dimensional invariant feature set for discriminatory object retrieval. Experiments have been conducted on a database consisting of 500 images taken from multicolored man-made objects in real world scenes. From the theoretical and experimental results it is concluded that object retrieval based on composite color and shape invariant features provides excellent retrieval accuracy. Object retrieval based on color invariants provides very high retrieval accuracy whereas object retrieval based entirely on shape invariants yields poor discriminative power. Furthermore, the image retrieval scheme is highly robust to partial occlusion, object clutter and a change in the object's pose. Finally, the image retrieval scheme is integrated into the PicToSeek system on-line at http://www.wins.uva.nl/research/isis/PicToSeek/ for searching images on the World Wide Web.
3923c0deee252ba10562a4378fc2bbc4885282b3
Fake Colorized Image Detection
Image forensics aims to detect the manipulation of digital images. Currently, splicing detection, copy-move detection, and image retouching detection are attracting significant attention from researchers. However, image editing techniques develop over time. An emerging image editing technique is colorization, in which grayscale images are colorized with realistic colors. Unfortunately, this technique may also be intentionally applied to certain images to confound object recognition algorithms. To the best of our knowledge, no forensic technique has yet been invented to identify whether an image is colorized. We observed that, compared with natural images, colorized images, which are generated by three state-of-the-art methods, possess statistical differences for the hue and saturation channels. Besides, we also observe statistical inconsistencies in the dark and bright channels, because the colorization process will inevitably affect the dark and bright channel values. Based on our observations, i.e., potential traces in the hue, saturation, dark, and bright channels, we propose two simple yet effective detection methods for fake colorized images: Histogram-based fake colorized image detection and feature encoding-based fake colorized image detection. Experimental results demonstrate that both proposed methods exhibit a decent performance against multiple state-of-the-art colorization approaches.
63db36cb0b5c8dad17a0c02ab95fde805d585513
Assessing patient suitability for short-term cognitive therapy with an interpersonal focus
In the current study, the development and initial validation of the Suitability for Short-Term Cognitive Therapy (SSCT) interview procedure is reported. The SSCT is an interview and rating procedure designed to evaluate the potential appropriateness of patients for short-term cognitive therapy with an interpersonal focus. It consists of a 1-hour, semistructured interview, focused on eliciting information from the patient relevant to nine selection criteria. The procedures involved in the development of this scale are described in detail, and preliminary evidence suggesting that the selection criteria can be rated reliably is presented. In addition, data indicating that scores on the SSCT scale predict the outcome of short-term cognitive therapy on multiple dependent measures, including both therapist and patient perspectives, are reported. It is concluded that the SSCT is a potentially useful scale for identifying patients who may be suitable, or unsuitable, for the type of short-term cognitive therapy administered in the present study.
2dc18f661b400033abd1086b917c451d3358aef2
Visible Machine Learning for Biomedicine
A major ambition of artificial intelligence lies in translating patient data to successful therapies. Machine learning models face particular challenges in biomedicine, however, including handling of extreme data heterogeneity and lack of mechanistic insight into predictions. Here, we argue for "visible" approaches that guide model structure with experimental biology.
09e941ab733b2c6c26261cb85f00f145d9063b0b
Automated Text Summarization In SUMMARIST
SUMMARIST is an attempt to create a robust automated text summarization system, based on the ‘equation’: summarization = topic identification + interpretation + generation. Each of these stages contains several independent modules, many of them trained on large corpora of text. We describe the system’s architecture and provide details of some of its modules.
03ce7c63eea901962dfae539b3ca6c77d65c5c38
Spike-Based Synaptic Plasticity in Silicon: Design, Implementation, Application, and Challenges
The ability to carry out signal processing, classification, recognition, and computation in artificial spiking neural networks (SNNs) is mediated by their synapses. In particular, through activity-dependent alteration of their efficacies, synapses play a fundamental role in learning. The mathematical prescriptions under which synapses modify their weights are termed synaptic plasticity rules. These learning rules can be based on abstract computational neuroscience models or on detailed biophysical ones. As these rules are being proposed and developed by experimental and computational neuroscientists, engineers strive to design and implement them in silicon and en masse in order to employ them in complex real-world applications. In this paper, we describe analog very large-scale integration (VLSI) circuit implementations of multiple synaptic plasticity rules, ranging from phenomenological ones (e.g., based on spike timing, mean firing rates, or both) to biophysically realistic ones (e.g., calcium-dependent models). We discuss the application domains, weaknesses, and strengths of various representative approaches proposed in the literature, and provide insight into the challenges that engineers face when designing and implementing synaptic plasticity rules in VLSI technology for utilizing them in real-world applications.
bc9f5d844ea6b989feb989cc6c8fc34f721a6b06
Networks versus markets in international trade
I propose a network/search view of international trade in differentiated products. I present evidence that supports the view that proximity and common language/colonial ties are more important for differentiated products than for products traded on organized exchanges in matching international buyers and sellers, and that search barriers to trade are higher for differentiated than for homogeneous products. I also discuss alternative explanations for the findings.
3973e14770350ed54ba1272aa3e19b4d21f5dad3
Obstacle Detection and Tracking for the Urban Challenge
This paper describes the obstacle detection and tracking algorithms developed for Boss, which is Carnegie Mellon University 's winning entry in the 2007 DARPA Urban Challenge. We describe the tracking subsystem and show how it functions in the context of the larger perception system. The tracking subsystem gives the robot the ability to understand complex scenarios of urban driving to safely operate in the proximity of other vehicles. The tracking system fuses sensor data from more than a dozen sensors with additional information about the environment to generate a coherent situational model. A novel multiple-model approach is used to track the objects based on the quality of the sensor data. Finally, the architecture of the tracking subsystem explicitly abstracts each of the levels of processing. The subsystem can easily be extended by adding new sensors and validation algorithms.
6a694487451957937adddbd682d3851fabd45626
Question answering passage retrieval using dependency relations
State-of-the-art question answering (QA) systems employ term-density ranking to retrieve answer passages. Such methods often retrieve incorrect passages as relationships among question terms are not considered. Previous studies attempted to address this problem by matching dependency relations between questions and answers. They used strict matching, which fails when semantically equivalent relationships are phrased differently. We propose fuzzy relation matching based on statistical models. We present two methods for learning relation mapping scores from past QA pairs: one based on mutual information and the other on expectation maximization. Experimental results show that our method significantly outperforms state-of-the-art density-based passage retrieval methods by up to 78% in mean reciprocal rank. Relation matching also brings about a 50% improvement in a system enhanced by query expansion.
a3676ecae39afd35b1f7075fc630e28cfbb5a188
Nitro: Hardware-Based System Call Tracing for Virtual Machines
Virtual machine introspection (VMI) describes the method of monitoring and analyzing the state of a virtual machine from the hypervisor level. This lends itself well to security applications, though the hardware virtualization support from Intel and AMD was not designed with VMI in mind. This results in many challenges for developers of hardware-supported VMI systems. This paper describes the design and implementation of our prototype framework, Nitro, for system call tracing and monitoring. Since Nitro is a purely VMI-based system, it remains isolated from attacks originating within the guest operating system and is not directly visible from within the guest. Nitro is extremely flexible as it supports all three system call mechanisms provided by the Intel x86 architecture and has been proven to work in Windows, Linux, 32-bit, and 64-bit environments. The high performance of our system allows for real-time capturing and dissemination of data without hindering usability. This is supported by extensive testing with various guest operating systems. In addition, Nitro is resistant to circumvention attempts due to a construction called hardware rooting. Finally, Nitro surpasses similar systems in both performance and functionality.
4b31ec67990a5fa81e7c1cf9fa2dbebcb91ded59
Adapting Naive Bayes to Domain Adaptation for Sentiment Analysis
In the community of sentiment analysis, supervised learning techniques have been shown to perform very well. When transferred to another domain, however, a supervised sentiment classifier often performs extremely bad. This is so-called domain-transfer problem. In this work, we attempt to attack this problem by making the maximum use of both the old-domain data and the unlabeled new-domain data. To leverage knowledge from the old-domain data, we proposed an effective measure, i.e., Frequently Co-occurring Entropy (FCE), to pick out generalizable features that occur frequently in both domains and have similar occurring probability. To gain knowledge from the newdomain data, we proposed Adapted Naïve Bayes (ANB), a weighted transfer version of Naive Bayes Classifier. The experimental results indicate that proposed approach could improve the performance of base classifier dramatically, and even provide much better performance than the transfer-learning baseline, i.e. the Naïve Bayes Transfer Classifier (NTBC).
654952a9cc4f3526dda8adf220a50a27a5c91449
DendroPy: a Python library for phylogenetic computing
UNLABELLED DendroPy is a cross-platform library for the Python programming language that provides for object-oriented reading, writing, simulation and manipulation of phylogenetic data, with an emphasis on phylogenetic tree operations. DendroPy uses a splits-hash mapping to perform rapid calculations of tree distances, similarities and shape under various metrics. It contains rich simulation routines to generate trees under a number of different phylogenetic and coalescent models. DendroPy's data simulation and manipulation facilities, in conjunction with its support of a broad range of phylogenetic data formats (NEXUS, Newick, PHYLIP, FASTA, NeXML, etc.), allow it to serve a useful role in various phyloinformatics and phylogeographic pipelines. AVAILABILITY The stable release of the library is available for download and automated installation through the Python Package Index site (http://pypi.python.org/pypi/DendroPy), while the active development source code repository is available to the public from GitHub (http://github.com/jeetsukumaran/DendroPy).
6a69cab99a68869b2f6361c6a3004657e2deeae4
Ground plane segmentation for mobile robot visual navigation
We describe a method of mobile robot monocular visual navigation, which uses multiple visual cues to detect and segment the ground plane in the robot’s field of view. Corner points are tracked through an image sequence and grouped into coplanar regions using a method which we call an H-based tracker. The H-based tracker employs planar homographys and is initialised by 5-point planar projective invariants. This allows us to detect ground plane patches and the colour within such patches is subsequently modelled. These patches are grown by colour classification to give a ground plane segmentation, which is then used as an input to a new variant of the artificial potential field algorithm.
13c48b8c10022b4b2262c5d12f255e21f566cecc
Practical design considerations for a LLC multi-resonant DC-DC converter in battery charging applications
In this paper, resonant tank design procedure and practical design considerations are presented for a high performance LLC multi-resonant dc-dc converter in a two-stage smart battery charger for neighborhood electric vehicle applications. The multi-resonant converter has been analyzed and its performance characteristics are presented. It eliminates both low and high frequency current ripple on the battery, thus maximizing battery life without penalizing the volume of the charger. Simulation and experimental results are presented for a prototype unit converting 390 V from the input dc link to an output voltage range of 48 V to 72 V dc at 650 W. The prototype achieves a peak efficiency of 96%.
27229aff757b797d0cae7bead5a236431b253b91
Predictive State Temporal Difference Learning
We propose a new approach to value function approximation which combines linear temporal difference reinforcement learning with subspace identification. In practical applications, reinforcement learning (RL) is complicated by the fact that state is either high-dimensional or partially observable. Therefore, RL methods are designed to work with features of state rather than state itself, and the success or failure of learning is often determined by the suitability of the selected features. By comparison, subspace identification (SSID) methods are designed to select a feature set which preserves as much information as possible about state. In this paper we connect the two approaches, looking at the problem of reinforcement learning with a large set of features, each of which may only be marginally useful for value function approximation. We introduce a new algorithm for this situation, called Predictive State Temporal Difference (PSTD) learning. As in SSID for predictive state representations, PSTD finds a linear compression operator that projects a large set of features down to a small set that preserves the maximum amount of predictive information. As in RL, PSTD then uses a Bellman recursion to estimate a value function. We discuss the connection between PSTD and prior approaches in RL and SSID. We prove that PSTD is statistically consistent, perform several experiments that illustrate its properties, and demonstrate its potential on a difficult optimal stopping problem.
21ff1d20dd7b3e6b1ea02036c0176d200ec5626d
Loss Max-Pooling for Semantic Image Segmentation
We introduce a novel loss max-pooling concept for handling imbalanced training data distributions, applicable as alternative loss layer in the context of deep neural networks for semantic image segmentation. Most real-world semantic segmentation datasets exhibit long tail distributions with few object categories comprising the majority of data and consequently biasing the classifiers towards them. Our method adaptively re-weights the contributions of each pixel based on their observed losses, targeting under-performing classification results as often encountered for under-represented object classes. Our approach goes beyond conventional cost-sensitive learning attempts through adaptive considerations that allow us to indirectly address both, inter- and intra-class imbalances. We provide a theoretical justification of our approach, complementary to experimental analyses on benchmark datasets. In our experiments on the Cityscapes and Pascal VOC 2012 segmentation datasets we find consistently improved results, demonstrating the efficacy of our approach.
87982ff47c0614cf40204970208312abe943641f
Comparing and evaluating the sentiment on newspaper articles: A preliminary experiment
Recent years have brought a symbolic growth in the volume of research in Sentiment Analysis, mostly on highly subjective text types like movie or product reviews. The main difference these texts have with news articles is that their target is apparently defined and unique across the text. Thence while dealing with news articles, we performed three subtasks namely identifying the target; separation of good and bad news content from the good and bad sentiment expressed on the target and analysis of clearly marked opinion that is expressed explicitly, not needing interpretation or the use of world knowledge. On concluding these tasks, we present our work on mining opinions about three different Indian political parties during elections in the year 2009. We built a Corpus of 689 opinion-rich instances from three different English dailies namely The Hindu, Times of India and Economic Times extracted from 02/ 01/ 2009 to 05/ 01/ 2009 (MM/ DD/ YY). In which (a) we tested the relative suitability of various sentiment analysis methods (both machine learning and lexical based) and (b) we attempted to separate positive or negative opinion from good or bad news. Evaluation includes comparison of three sentiment analysis methods (two machine learning based and one lexical based) and analyzing the choice of certain words used in political text which influence the Sentiments of public in polls. This preliminary experiment will benefit in predicting and forecasting the winning party in forthcoming Indian elections 2014.
2538e3eb24d26f31482c479d95d2e26c0e79b990
Natural Language Processing (almost) from Scratch
We propose a unified neural network architecture and learning algorithm that can be applied to various natural language processing tasks including: part-of-speech tagging, chunking, named entity recognition, and semantic role labeling. This versatility is achieved by trying to avoid task-specific engineering and therefore disregarding a lot of prior knowledge. Instead of exploiting man-made input features carefully optimized for each task, our system learns internal representations on the basis of vast amounts of mostly unlabeled training data. This work is then used as a basis for building a freely available tagging system with good performance and minimal computational requirements.
317deb87586baa4ee7c7b5dfc603ebed94d1da07
Deep Learning for Efficient Discriminative Parsing
We propose a new fast purely discriminative algorithm for natural language parsing, based on a “deep” recurrent convolutional graph transformer network (GTN). Assuming a decomposition of a parse tree into a stack of “levels”, the network predicts a level of the tree taking into account predictions of previous levels. Using only few basic text features which leverage word representations from Collobert and Weston (2008), we show similar performance (in F1 score) to existing pure discriminative parsers and existing “benchmark” parsers (like Collins parser, probabilistic context-free grammars based), with a huge speed advantage.
0354210007fbe92385acf407549b5cacb41b5835
Distributed and overlapping representations of faces and objects in ventral temporal cortex.
The functional architecture of the object vision pathway in the human brain was investigated using functional magnetic resonance imaging to measure patterns of response in ventral temporal cortex while subjects viewed faces, cats, five categories of man-made objects, and nonsense pictures. A distinct pattern of response was found for each stimulus category. The distinctiveness of the response to a given category was not due simply to the regions that responded maximally to that category, because the category being viewed also could be identified on the basis of the pattern of response when those regions were excluded from the analysis. Patterns of response that discriminated among all categories were found even within cortical regions that responded maximally to only one category. These results indicate that the representations of faces and objects in ventral temporal cortex are widely distributed and overlapping.
04cc04457e09e17897f9256c86b45b92d70a401f
A latent factor model for highly multi-relational data
Many data such as social networks, movie preferences or knowledge bases are multi-relational, in that they describe multiple relations between entities. While there is a large body of work focused on modeling these data, modeling these multiple types of relations jointly remains challenging. Further, existing approaches tend to breakdown when the number of these types grows. In this paper, we propose a method for modeling large multi-relational datasets, with possibly thousands of relations. Our model is based on a bilinear structure, which captures various orders of interaction of the data, and also shares sparse latent factors across different relations. We illustrate the performance of our approach on standard tensor-factorization datasets where we attain, or outperform, state-of-the-art results. Finally, a NLP application demonstrates our scalability and the ability of our model to learn efficient and semantically meaningful verb representations.
052b1d8ce63b07fec3de9dbb583772d860b7c769
Learning representations by back-propagating errors
We describe a new learning procedure, back-propagation, for networks of neurone-like units. The procedure repeatedly adjusts the weights of the connections in the network so as to minimize a measure of the difference between the actual output vector of the net and the desired output vector. As a result of the weight adjustments, internal ‘hidden’ units which are not part of the input or output come to represent important features of the task domain, and the regularities in the task are captured by the interactions of these units. The ability to create useful new features distinguishes back-propagation from earlier, simpler methods such as the perceptron-convergence procedure1.
50a6b2b84a9d11ed168ce6380ff17e76136cdfe7
A memory insensitive technique for large model simplification
In this paper we propose three simple, but significant improvements to the OoCS (Out-of-Core Simplification) algorithm of Lindstrom [20] which increase the quality of approximations and extend the applicability of the algorithm to an even larger class of compute systems.The original OoCS algorithm has memory complexity that depends on the size of the output mesh, but no dependency on the size of the input mesh. That is, it can be used to simplify meshes of arbitrarily large size, but the complexity of the output mesh is limited by the amount of memory available. Our first contribution is a version of OoCS that removes the dependency of having enough memory to hold (even) the simplified mesh. With our new algorithm, the whole process is made essentially independent of the available memory on the host computer. Our new technique uses disk instead of main memory, but it is carefully designed to avoid costly random accesses.Our two other contributions improve the quality of the approximations generated by OoCS. We propose a scheme for preserving surface boundaries which does not use connectivity information, and a scheme for constraining the position of the "representative vertex" of a grid cell to an optimal position inside the cell.
22630a79f1c50603c1356f6ac9dc8524a18d4061
SecondNet: a data center network virtualization architecture with bandwidth guarantees
In this paper, we propose virtual data center (VDC) as the unit of resource allocation for multiple tenants in the cloud. VDCs are more desirable than physical data centers because the resources allocated to VDCs can be rapidly adjusted as tenants' needs change. To enable the VDC abstraction, we design a data center network virtualization architecture called SecondNet. SecondNet achieves scalability by distributing all the virtual-to-physical mapping, routing, and bandwidth reservation state in server hypervisors. Its port-switching based source routing (PSSR) further makes SecondNet applicable to arbitrary network topologies using commodity servers and switches. SecondNet introduces a centralized VDC allocation algorithm for bandwidth guaranteed virtual to physical mapping. Simulations demonstrate that our VDC allocation achieves high network utilization and low time complexity. Our implementation and experiments show that we can build SecondNet on top of various network topologies, and SecondNet provides bandwidth guarantee and elasticity, as designed.
5af83b56353c5fba0518c203d192ffb6375cd986
A deep multiple instance model to predict prostate cancer metastasis from nuclear morphology
We consider the problem of identifying the patients who are diagnosed with highgrade prostate cancer using the histopathology of tumor in a prostate needle biopsy and are at a very high risk of lethal cancer progression. We hypothesize that the morphology of tumor cell nuclei in digital images from the biopsy can be used to predict tumor aggressiveness and posit the presence of metastasis as a surrogate for disease specific mortality. For this purpose, we apply a compositional multiinstance learning approach which encodes images of nuclei through a convolutional neural network, then predicts the presence of metastasis from sets of encoded nuclei. Through experiments on prostate needle biopsies (PNBX) from a patient cohort with known presence (M1 stage, n = 85) or absence (M0 stage, n = 86) of metastatic disease, we obtained an average area under the receiver operating characteristic curve of 0.71± 0.08 for predicting metastatic cases. These results support our hypothesis that information related to metastatic capacity of prostate cancer cells can be obtained through analysis of nuclei and establish a baseline for future research aimed at predicting the risk of future metastatic disease at a time when it might be preventable.
d055b799c521b28bd4d6bf2fc905819d8e88207c
Design of a dual circular polarization microstrip patch array antenna
Design of a microstrip array antenna to achieve dual circular polarization is proposed in this paper. The proposed antenna is a 2×2 array antenna where each patch element is circularly polarized. The feed network has microstrip lines, cross slot lines and air-bridges. The array antenna can excite both right-hand circular polarization (RHCP) and left-hand circular polarization (LHCP) without using any 90° hybrid circuit or PIN diode. “Both-sided MIC Technology” is used to design the feed network as it provides flexibility to place several types of transmission lines on both sides of the dielectric substrate. The design frequency of the proposed array antenna is 10 GHz. The simulated return loss exhibits an impedance bandwidth of greater than 5% and the 3-dB axial ratio bandwidths for both RHCP and LHCP are approximately 1.39%. The structure and the basic operation along with the simulation results of the proposed dual circularly polarized array antenna are demonstrated in this paper.
2b72dd0d33e0892436394ef7642c6b517a1c71fd
Matching Visual Saliency to Confidence in Plots of Uncertain Data
Conveying data uncertainty in visualizations is crucial for preventing viewers from drawing conclusions based on untrustworthy data points. This paper proposes a methodology for efficiently generating density plots of uncertain multivariate data sets that draws viewers to preattentively identify values of high certainty while not calling attention to uncertain values. We demonstrate how to augment scatter plots and parallel coordinates plots to incorporate statistically modeled uncertainty and show how to integrate them with existing multivariate analysis techniques, including outlier detection and interactive brushing. Computing high quality density plots can be expensive for large data sets, so we also describe a probabilistic plotting technique that summarizes the data without requiring explicit density plot computation. These techniques have been useful for identifying brain tumors in multivariate magnetic resonance spectroscopy data and we describe how to extend them to visualize ensemble data sets.
7d9facefffc720079d837aa421ab79d4856e2c88
Lightweight, High-Force Gripper Inspired by Chuck Clamping Devices
In this letter, we present a novel gripper, whose design was inspired by chuck clamping devices, for transferring heavy objects and assembling parts precisely in industrial applications. The developed gripper is lightweight (0.9 kg), can manipulate heavy payloads (over 23 kgf), and can automatically align its position and posture via a grasping motion. A fingertip design criterion is presented for the position alignment, while a control strategy is presented for the posture alignment. With one actuator, this gripper realized the above features. This letter describes the mathematical analyses and experiments used to validate these key metrics.
29f07c86886af63f9bf43d089373ac1f7a95ea0e
A Multiarmed Bandit Incentive Mechanism for Crowdsourcing Demand Response in Smart Grids
Demand response is a critical part of renewable integration and energy cost reduction goals across the world. Motivated by the need to reduce costs arising from electricity shortage and renewable energy fluctuations, we propose a novel multiarmed bandit mechanism for demand response (MAB-MDR) which makes monetary offers to strategic consumers who have unknown response characteristics, to incetivize reduction in demand. Our work is inspired by a novel connection we make to crowdsourcing mechanisms. The proposed mechanism incorporates realistic features of the demand response problem including time varying and quadratic cost function. The mechanism marries auctions, that allow users to report their preferences, with online algorithms, that allow distribution companies to learn user-specific parameters. We show that MAB-MDR is dominant strategy incentive compatible, individually rational, and achieves sublinear regret. Such mechanisms can be effectively deployed in smart grids using new information and control architecture innovations and lead to welcome savings in energy costs.
8e7fdb9d3fc0fef1f82f126072fc675e01ce5873
Clarifying Hypotheses by Sketching Data
Discussions between data analysts and colleagues or clients with no statistical background are difficult, as the analyst often has to teach and explain their statistical and domain knowledge. We investigate work practices of data analysts who collaborate with non-experts, and report findings regarding types of analysis, collaboration and availability of data. Based on these, we have created a tool to enhance collaboration between data analysts and their clients in the initial stages of the analytical process. Sketching time series data allows analysts to discuss expectations for later analysis. We propose function composition rather than freehand sketching, in order to structure the analyst-client conversation by independently expressing expected features in the data. We evaluate the usability of our prototype through two small studies, and report on user feedback for future iterations.
0567283bc9affd475eae7cebaae658692a64d5a4
Intelligent Widgets for Intuitive Interaction and Coordination in Smart Home Environments
The intelligent home environment is a well-established example of the Ambient Intelligence application domain. A variety of sensors and actuators can be used to have the home environment adapt towards changing circumstances and user preferences. However, the complexity of how these intelligent home automation systems operate is often beyond the comprehension of non-technical users, and adding new technology to an existing infrastructure is often a burden. In this paper, we present a home automation framework designed based on smart widgets with a model driven methodology that raises the level of abstraction to configure home automation equipment. It aims to simplify user-level home automation management by mapping high-level home automation concepts onto a low-level composition and configuration of the automation building blocks with a reverse mapping to simplify the integration of new equipment into existing home automation systems. Experiments have shown that the mappings we proposed are sufficient to represent household appliances to the end user in a simple way and that new mappings can easily be added to our framework.
a79f43246bed540084ca2d1fcf99a68c69820747
A Hybrid Approach to Detect and Localize Texts in Natural Scene Images
Text detection and localization in natural scene images is important for content-based image analysis. This problem is challenging due to the complex background, the non-uniform illumination, the variations of text font, size and line orientation. In this paper, we present a hybrid approach to robustly detect and localize texts in natural scene images. A text region detector is designed to estimate the text existing confidence and scale information in image pyramid, which help segment candidate text components by local binarization. To efficiently filter out the non-text components, a conditional random field (CRF) model considering unary component properties and binary contextual component relationships with supervised parameter learning is proposed. Finally, text components are grouped into text lines/words with a learning-based energy minimization method. Since all the three stages are learning-based, there are very few parameters requiring manual tuning. Experimental results evaluated on the ICDAR 2005 competition dataset show that our approach yields higher precision and recall performance compared with state-of-the-art methods. We also evaluated our approach on a multilingual image dataset with promising results.
1b9de2d1e74fbe49bf852fa495f63c31bb038a31
A Pneumatic-Driven Haptic Glove with Force and Tactile Feedback
The advent of Oculus Rift indicates the start of a booming era of virtual reality. In order to increase the immersive feeling of interaction with the virtual world, haptic devices allow us to touch and manipulate virtual objects in an intuitive way. In this paper, we introduce a portable and low-cost haptic glove that provides both force and tactile feedback using a direct-control pneumatic concept. To produce force feedback, two inlet ports of a double acting pneumatic cylinder are opened and closed via solenoid DC valves through Pulse-width modulation (PWM) technique. For tactile feedback, an air bladder is actuated using a diaphragm pump via PWM operated solenoid valve. Experiments on a single finger prototype validated that the glove can provide force and tactile feedback with sufficient moving range of the finger joints. The maximum continuous force is 9 Newton and the response time is less than 400ms. The glove is light weighted and easy to be mounted on the index finger. The proposed glove could be potentially used for virtual reality grasping scenarios and for teleoperation of a robotic hand for handling hazardous objects.
c6504fbbfcf32854e0bd35eb70539cafbecf332f
Client-side rate adaptation scheme for HTTP adaptive streaming based on playout buffer model
HTTP Adaptive Streaming (HAS) is an adaptive bitrate streaming technique which is able to adapt to the network conditions using conventional HTTP web servers. An HAS player periodically requests pre-encoded video chunks by sending an HTTP GET message. When the downloading a video chunk is finished, the player estimates the network bandwidth by calculating the goodput and adjusts the video quality based on its estimates. However, the bandwidth estimation in application layer is pretty inaccurate due to its architectural limitation. We show that inaccurate bandwidth estimation in rate adaptation may incur serious rate oscillations which is poor quality-of-experience for users. In this paper, we propose a buffer-based rate adaptation scheme which eliminates the bandwidth estimation step in rate adaptation to provide a smooth playback of HTTP-based streaming. We evaluate the performance of the HAS player implemented in the ns-3 network simulator. Our simulation results show that the proposed scheme significantly improves the stability by replacing bandwidth estimation with buffer occupancy estimation.
03dc771ebf5b7bc3ccf8c4689d918924da524fe4
Approximating dynamic global illumination in image space
Physically plausible illumination at real-time framerates is often achieved using approximations. One popular example is ambient occlusion (AO), for which very simple and efficient implementations are used extensively in production. Recent methods approximate AO between nearby geometry in screen space (SSAO). The key observation described in this paper is, that screen-space occlusion methods can be used to compute many more types of effects than just occlusion, such as directional shadows and indirect color bleeding. The proposed generalization has only a small overhead compared to classic SSAO, approximates direct and one-bounce light transport in screen space, can be combined with other methods that simulate transport for macro structures and is visually equivalent to SSAO in the worst case without introducing new artifacts. Since our method works in screen space, it does not depend on the geometric complexity. Plausible directional occlusion and indirect lighting effects can be displayed for large and fully dynamic scenes at real-time frame rates.
1b656883bed80fdec1d109ae04873540720610fa
Development and validation of the childhood narcissism scale.
In this article, we describe the development and validation of a short (10 item) but comprehensive self-report measure of childhood narcissism. The Childhood Narcissism Scale (CNS) is a 1-dimensional measure of stable individual differences in childhood narcissism with strong internal consistency reliability (Studies 1-4). The CNS is virtually unrelated to conventional measures of self-esteem but is positively related to self-appraised superiority, social evaluative concern and self-esteem contingency, agentic interpersonal goals, and emotional extremity (Study 5). Furthermore, the CNS is negatively related to empathic concern and positively related to aggression following ego threat (Study 6). These results suggest that childhood narcissism has similar psychological and interpersonal correlates as adult narcissism. The CNS provides researchers a convenient tool for measuring narcissism in children and young adolescents with strong preliminary psychometric characteristics.
92c1f538613ff4923a8fa3407a16bed4aed361ac
Progressive Analytics: A Computation Paradigm for Exploratory Data Analysis
Exploring data requires a fast feedback loop from the analyst to the system, with a latency below about 10 seconds because of human cognitive limitations. When data becomes large or analysis becomes complex, sequential computations can no longer be completed in a few seconds and data exploration is severely hampered. This article describes a novel computation paradigm called Progressive Computation for Data Analysis or more concisely Progressive Analytics, that brings at the programming language level a low-latency guarantee by performing computations in a progressive fashion. Moving this progressive computation at the language level relieves the programmer of exploratory data analysis systems from implementing the whole analytics pipeline in a progressive way from scratch, streamlining the implementation of scalable exploratory data analysis systems. This article describes the new paradigm through a prototype implementation called ProgressiVis, and explains the requirements it implies through examples.
bb6508fb4457f09b5e146254220247bc4ea7b71c
Multiclass Classification of Driver Perceived Workload Using Long Short-Term Memory based Recurrent Neural Network
Human sensing enables intelligent vehicles to provide driver-adaptive support by classifying perceived workload into multiple levels. Objective of this study is to classify driver workload associated with traffic complexity into five levels. We conducted driving experiments in systematically varied traffic complexity levels in a simulator. We recorded driver physiological signals including electrocardiography, electrodermal activity, and electroencephalography. In addition, we integrated driver performance and subjective workload measures. Deep learning based models outperform statistical machine learning methods when dealing with dynamic time-series data with variable sequence lengths. We show that our long short-term memory based recurrent neural network model can classify driver perceived-workload into five classes with an accuracy of 74.5%. Since perceived workload differ between individual drivers for the same traffic situation, our results further highlight the significance of including driver characteristics such as driving style and workload sensitivity to achieve higher classification accuracy.
c8e7c2a9201eb6217e62266c9d8c061b5394866e
Data mining for censored time-to-event data: a Bayesian network model for predicting cardiovascular risk from electronic health record data
Models for predicting the risk of cardiovascular (CV) events based on individual patient characteristics are important tools for managing patient care. Most current and commonly used risk prediction models have been built from carefully selected epidemiological cohorts. However, the homogeneity and limited size of such cohorts restrict the predictive power and generalizability of these risk models to other populations. Electronic health data (EHD) from large health care systems provide access to data on large, heterogeneous, and contemporaneous patient populations. The unique features and challenges of EHD, including missing risk factor information, non-linear relationships between risk factors and CV event outcomes, and differing effects from different patient subgroups, demand novel machine learning approaches to risk model development. In this paper, we present a machine learning approach based on Bayesian networks trained on EHD to predict the probability of having a CV event within 5 years. In such data, event status may be unknown for some individuals, as the event time is right-censored due to disenrollment and incomplete follow-up. Since many traditional data mining methods are not well-suited for such data, we describe how to modify both modeling and assessment techniques to account for censored observation times. We show that our approach can lead to better predictive performance than the Cox proportional hazards model (i.e., a regression-based approach commonly used for censored, time-to-event data) or a Bayesian network with ad hoc approaches to right-censoring. Our techniques are motivated by and illustrated on data from a large US Midwestern health care system.
fc5a530ea80a3295d0872b85c3991a4d81336a61
Voice Activated Virtual Assistants Personality Perceptions and Desires- Comparing Personality Evaluation Frameworks
Currently, Voice Activated Virtual Assistants and Artificial Intelligence technologies are not just about performance or the functionalities they can carry out, it is also about the associated personality. This empirical multi-country study explores the personality perceptions of current VAVA users regarding these technologies. Since this is a rather unexplored territory for research, this study has identified two well-established personality evaluation methodologies, Aaker’s traits approach and Jung’s archetypes, to investigate current perceived personality and future desired personality of the four main Voice Activated Virtual Assistants: Siri, Google Assistant, Cortana and Alexa. Following are a summary of results by each methodology, and an analysis of the commonalities found between the two methodologies.
f64d18d4bad30ea544aa828eacfa83208f2b7815
Conceptualizing Context for Pervasive Advertising
Profile-driven personalization based on socio-demographics is currently regarded as the most convenient base for successful personalized advertising. However, signs point to the dormant power of context recognition: Advertising systems that can adapt to the situational context of a consumer will rapidly gain importance. While technologies that can sense the environment are increasingly advanced, questions such as what to sense and how to adapt to a consumer’s context are largely unanswered. In this chapter, we analyze the purchase context of a retail outlet and conceptualize it such that adaptive pervasive advertising applications really deliver on their potential: showing the right message at the right time to the right recipient. full version published as: Bauer, Christine & Spiekermann, Sarah (2011). Conceptualizing Context for Pervasive Advertising. In Müller, Jörg, Alt, Florian, & Michelis, Daniel (Eds.), Pervasive Advertising (pp. 159-183). London: Springer.
2e92ddcf2e7a9d6c27875ec442637e13753f21a2
Self-Soldering Connectors for Modular Robots
The connection mechanism between neighboring modules is the most critical subsystem of each module in a modular robot. Here, we describe a strong, lightweight, and solid-state connection method based on heating a low melting point alloy to form reversible soldered connections. No external manipulation is required for forming or breaking connections between adjacent connectors, making this method suitable for reconfigurable systems such as self-reconfiguring modular robots. Energy is only consumed when switching connectivity, and the ability to transfer power and signal through the connector is inherent to the method. Soldering connectors have no moving parts, are orders of magnitude lighter than other connectors, and are readily mass manufacturable. The mechanical strength of the connector is measured as 173 N, which is enough to support many robot modules, and hundreds of connection cycles are performed before failure.
459fbc416eb9a55920645c741b1e4cce95f39786
The Numerics of GANs
In this paper, we analyze the numerics of common algorithms for training Generative Adversarial Networks (GANs). Using the formalism of smooth two-player games we analyze the associated gradient vector field of GAN training objectives. Our findings suggest that the convergence of current algorithms suffers due to two factors: i) presence of eigenvalues of the Jacobian of the gradient vector field with zero real-part, and ii) eigenvalues with big imaginary part. Using these findings, we design a new algorithm that overcomes some of these limitations and has better convergence properties. Experimentally, we demonstrate its superiority on training common GAN architectures and show convergence on GAN architectures that are known to be notoriously hard to train.
40c8e3894314581d0241162374602d68a6d1f38c
Culture and Institutions
A growing body of empirical work measuring different types of cultural traits has shown that culture matters for a variety of economic outcomes. This paper focuses on one specific aspect of the relevance of culture: its relationship to institutions. We review work with a theoretical, empirical, and historical bent to assess the presence of a two-way causal effect between culture and institutions. 1 We thank Benjamin Friedman and Andrei Shleifer for useful conversations and Janet Currie, Steven Durlauf, and six anonymous referees for excellent comments.
6b894324281bd4b0251549c0e40802d6ca3d0b8f
Challenges and prospects of lithium-sulfur batteries.
Electrical energy storage is one of the most critical needs of 21st century society. Applications that depend on electrical energy storage include portable electronics, electric vehicles, and devices for renewable energy storage from solar and wind. Lithium-ion (Li-ion) batteries have the highest energy density among the rechargeable battery chemistries. As a result, Li-ion batteries have proven successful in the portable electronics market and will play a significant role in large-scale energy storage. Over the past two decades, Li-ion batteries based on insertion cathodes have reached a cathode capacity of ∼250 mA h g(-1) and an energy density of ∼800 W h kg(-1), which do not meet the requirement of ∼500 km between charges for all-electric vehicles. With a goal of increasing energy density, researchers are pursuing alternative cathode materials such as sulfur and O2 that can offer capacities that exceed those of conventional insertion cathodes, such as LiCoO2 and LiMn2O4, by an order of magnitude (>1500 mA h g(-1)). Sulfur, one of the most abundant elements on earth, is an electrochemically active material that can accept up to two electrons per atom at ∼2.1 V vs Li/Li(+). As a result, sulfur cathode materials have a high theoretical capacity of 1675 mA h g(-1), and lithium-sulfur (Li-S) batteries have a theoretical energy density of ∼2600 W h kg(-1). Unlike conventional insertion cathode materials, sulfur undergoes a series of compositional and structural changes during cycling, which involve soluble polysulfides and insoluble sulfides. As a result, researchers have struggled with the maintenance of a stable electrode structure, full utilization of the active material, and sufficient cycle life with good system efficiency. Although researchers have made significant progress on rechargeable Li-S batteries in the last decade, these cycle life and efficiency problems prevent their use in commercial cells. To overcome these persistent problems, researchers will need new sulfur composite cathodes with favorable properties and performance and new Li-S cell configurations. In this Account, we first focus on the development of novel composite cathode materials including sulfur-carbon and sulfur-polymer composites, describing the design principles, structure and properties, and electrochemical performances of these new materials. We then cover new cell configurations with carbon interlayers and Li/dissolved polysulfide cells, emphasizing the potential of these approaches to advance capacity retention and system efficiency. Finally, we provide a brief survey of efficient electrolytes. The Account summarizes improvements that could bring Li-S technology closer to mass commercialization.
e8691980eeb827b10cdfb4cc402b3f43f020bc6a
Segmentation Guided Attention Networks for Visual Question Answering
In this paper we propose to solve the problem of Visual Question Answering by using a novel segmentation guided attention based network which we call SegAttendNet. We use image segmentation maps, generated by a Fully Convolutional Deep Neural Network to refine our attention maps and use these refined attention maps to make the model focus on the relevant parts of the image to answer a question. The refined attention maps are used by the LSTM network to learn to produce the answer. We presently train our model on the visual7W dataset and do a category wise evaluation of the 7 question categories. We achieve state of the art results on this dataset and beat the previous benchmark on this dataset by a 1.5% margin improving the question answering accuracy from 54.1% to 55.6% and demonstrate improvements in each of the question categories. We also visualize our generated attention maps and note their improvement over the attention maps generated by the previous best approach.
07f3f736d90125cb2b04e7408782af411c67dd5a
Convolutional Neural Network Architectures for Matching Natural Language Sentences
Semantic matching is of central importance to many natural language tasks [2, 28]. A successful matching algorithm needs to adequately model the internal structures of language objects and the interaction between them. As a step toward this goal, we propose convolutional neural network models for matching two sentences, by adapting the convolutional strategy in vision and speech. The proposed models not only nicely represent the hierarchical structures of sentences with their layerby-layer composition and pooling, but also capture the rich matching patterns at different levels. Our models are rather generic, requiring no prior knowledge on language, and can hence be applied to matching tasks of different nature and in different languages. The empirical study on a variety of matching tasks demonstrates the efficacy of the proposed model on a variety of matching tasks and its superiority to competitor models.
0af737eae02032e66e035dfed7f853ccb095d6f5
ABCNN: Attention-Based Convolutional Neural Network for Modeling Sentence Pairs
How to model a pair of sentences is a critical issue in many NLP tasks such as answer selection (AS), paraphrase identification (PI) and textual entailment (TE). Most prior work (i) deals with one individual task by fine-tuning a specific system; (ii) models each sentence’s representation separately, rarely considering the impact of the other sentence; or (iii) relies fully on manually designed, task-specific linguistic features. This work presents a general Attention Based Convolutional Neural Network (ABCNN) for modeling a pair of sentences. We make three contributions. (i) The ABCNN can be applied to a wide variety of tasks that require modeling of sentence pairs. (ii) We propose three attention schemes that integrate mutual influence between sentences into CNNs; thus, the representation of each sentence takes into consideration its counterpart. These interdependent sentence pair representations are more powerful than isolated sentence representations. (iii) ABCNNs achieve state-of-the-art performance on AS, PI and TE tasks. We release code at: https://github.com/yinwenpeng/Answer_Selection.
1c059493904b2244d2280b8b4c0c7d3ca115be73
node2vec: Scalable Feature Learning for Networks
Prediction tasks over nodes and edges in networks require careful effort in engineering features used by learning algorithms. Recent research in the broader field of representation learning has led to significant progress in automating prediction by learning the features themselves. However, present feature learning approaches are not expressive enough to capture the diversity of connectivity patterns observed in networks. Here we propose node2vec, an algorithmic framework for learning continuous feature representations for nodes in networks. In node2vec, we learn a mapping of nodes to a low-dimensional space of features that maximizes the likelihood of preserving network neighborhoods of nodes. We define a flexible notion of a node's network neighborhood and design a biased random walk procedure, which efficiently explores diverse neighborhoods. Our algorithm generalizes prior work which is based on rigid notions of network neighborhoods, and we argue that the added flexibility in exploring neighborhoods is the key to learning richer representations. We demonstrate the efficacy of node2vec over existing state-of-the-art techniques on multi-label classification and link prediction in several real-world networks from diverse domains. Taken together, our work represents a new way for efficiently learning state-of-the-art task-independent representations in complex networks.
468b9055950c428b17f0bf2ff63fe48a6cb6c998
A Neural Attention Model for Abstractive Sentence Summarization
Summarization based on text extraction is inherently limited, but generation-style abstractive methods have proven challenging to build. In this work, we propose a fully data-driven approach to abstractive sentence summarization. Our method utilizes a local attention-based model that generates each word of the summary conditioned on the input sentence. While the model is structurally simple, it can easily be trained end-to-end and scales to a large amount of training data. The model shows significant performance gains on the DUC-2004 shared task compared with several strong baselines.
81eb0a1ea90a6f6d5e7f14cb3397a4ee0f77824a
Question/Answer Matching for CQA System via Combining Lexical and Sequential Information
Community-based Question Answering (CQA) has become popular in knowledge sharing sites since it allows users to get answers to complex, detailed, and personal questions directly from other users. Large archives of historical questions and associated answers have been accumulated. Retrieving relevant historical answers that best match a question is an essential component of a CQA service. Most state of the art approaches are based on bag-of-words models, which have been proven successful in a range of text matching tasks, but are insufficient for capturing the important word sequence information in short text matching. In this paper, a new architecture is proposed to more effectively model the complicated matching relations between questions and answers. It utilises a similarity matrix which contains both lexical and sequential information. Afterwards the information is put into a deep architecture to find potentially suitable answers. The experimental study shows its potential in improving matching accuracy of question and answer.
81ff60a35e57e150875cfdde735fe69d19e9fdc4
Development of attentional networks in childhood
Recent research in attention has involved three networks of anatomical areas that carry out the functions of orienting, alerting and executive control (including conflict monitoring). There have been extensive cognitive and neuroimaging studies of these networks in adults. We developed an integrated Attention Network Test (ANT) to measure the efficiency of the three networks with adults. We have now adapted this test to study the development of these networks during childhood. The test is a child-friendly version of the flanker task with alerting and orienting cues. We studied the development of the attentional networks in a cross-sectional experiment with four age groups ranging from 6 through 9 (Experiment 1). In a second experiment, we compared children (age 10 years) and adult performance in both child and adults versions of the ANT. Reaction time and accuracy improved at each age interval and positive values were found for the average efficiency of each of the networks. Alertness showed evidence of change up to and beyond age 10, while conflict scores appear stable after age seven and orienting scores do not change in the age range studied. A final experiment with forty 7-year-old children suggested that children like adults showed independence between the three networks under some conditions.
6c1cabe3f5980cbc50d290c2ed60b9aca624eab8
Mathematical modelling of infectious diseases.
INTRODUCTION Mathematical models allow us to extrapolate from current information about the state and progress of an outbreak, to predict the future and, most importantly, to quantify the uncertainty in these predictions. Here, we illustrate these principles in relation to the current H1N1 epidemic. SOURCES OF DATA Many sources of data are used in mathematical modelling, with some forms of model requiring vastly more data than others. However, a good estimation of the number of cases is vitally important. AREAS OF AGREEMENT Mathematical models, and the statistical tools that underpin them, are now a fundamental element in planning control and mitigation measures against any future epidemic of an infectious disease. Well-parameterized mathematical models allow us to test a variety of possible control strategies in computer simulations before applying them in reality. AREAS OF CONTROVERSY The interaction between modellers and public-health practitioners and the level of detail needed for models to be of use. GROWING POINTS The need for stronger statistical links between models and data. AREAS TIMELY FOR DEVELOPING RESEARCH Greater appreciation by the medical community of the uses and limitations of models and a greater appreciation by modellers of the constraints on public-health resources.
5e22c4362df3b0accbe04517c41848a2b229efd1
Predicting sports events from past results Towards effective betting on football matches
A system for predicting the results of football matches that beats the bookmakers’ odds is presented. The predictions for the matches are based on previous results of the teams involved.
de93c4f886bdf55bfc1bcaefad648d5996ed3302
Modern Intrusion Detection, Data Mining, and Degrees of Attack Guilt
This chapter examines the state of modern intrusion detection, with a particular emphasis on the emerging approach of data mining. The discussion paralleIs two important aspects of intrusion detection: general detection strategy (misuse detection versus anomaly detection) and data source (individual hosts versus network trafik). Misuse detection attempts to match known patterns of intrusion , while anomaly detection searches for deviations from normal behavior . Between the two approaches, only anomaly detection has the ability to detect unknown attacks. A particularly promising approach to anomaly detection combines association mining with other forms of machine learning such as classification. Moreover, the data source that an intrusion detection system employs significantly impacts the types of attacks it can detect. There is a tradeoff in the level of detailed information available verD. Barbará et al. (ed .), Applications of Data Mining in Computer Security © Kluwer Academic Publishers 2002 s
df25eaf576f55c09bb460d67134646fcb422b2ac
AGA: Attribute-Guided Augmentation
We consider the problem of data augmentation, i.e., generating artificial samples to extend a given corpus of training data. Specifically, we propose attributed-guided augmentation (AGA) which learns a mapping that allows to synthesize data such that an attribute of a synthesized sample is at a desired value or strength. This is particularly interesting in situations where little data with no attribute annotation is available for learning, but we have access to a large external corpus of heavily annotated samples. While prior works primarily augment in the space of images, we propose to perform augmentation in feature space instead. We implement our approach as a deep encoder-decoder architecture that learns the synthesis function in an end-to-end manner. We demonstrate the utility of our approach on the problems of (1) one-shot object recognition in a transfer-learning setting where we have no prior knowledge of the new classes, as well as (2) object-based one-shot scene recognition. As external data, we leverage 3D depth and pose information from the SUN RGB-D dataset. Our experiments show that attribute-guided augmentation of high-level CNN features considerably improves one-shot recognition performance on both problems.
b52abc5f401a6dec62d650f5a2a500f469b9a7c0
A Case Study on Barriers and Enhancements of the PET Bottle-to-Bottle Recycling Systems in Germany and Sweden
Problem: The demand of beverages in PET bottles is constantly increasing. In this context, environmental, technological and regulatory aspects set a stronger focus on recycling. Generally, the reuse of recycled material from post-consumer PET bottles in bottle-to-bottle applications is seen as least environmentally harmful. However, closedloop systems are not widely implemented in Europe. Previous research mainly focuses on open-loop recycling systems and generally lacks discussion about the current German and Swedish systems and their challenges. Furthermore, previous studies lack theoretical and practical enhancements for bottle-to-bottle recycling from a managerial perspective. Purpose: The purpose of this study is to compare the PET bottle recycling systems in Germany and Sweden, analyse the main barriers and develop enhancements for closedloop systems. Method: This qualitative study employs a case study strategy about the two cases of Germany and Sweden. In total, 14 semi-structured interviews are conducted with respondents from different industry sectors within the PET bottle recycling systems. The empirical data is categorised and then analysed by pattern matching with the developed theoretical framework. Conclusion: Due to the theoretical and practical commitment to closed-loop recycling, the Swedish PET bottle recycling system outperforms the Germany system. In Germany, bottle-to-bottle recycling is currently performed on a smaller scale without a unified system. The main barriers for bottle-to-bottle recycling are distinguished into (1) quality and material factors, (2) regulatory and legal factors, (3) economic and market factors and (4) factors influenced by consumers. The enhancements for the systems are (1) quality and material factors, (2) regulatory and legal factors, (3) recollection factors and (4) expanding factors. Lastly, the authors provide further recommendations, which are (1) a recycling content symbol on bottle labels, (2) a council for bottle quality in Germany, (3) a quality seal for the holistic systems, (4) a reduction of transportation in Sweden and (5) an increase of consumer awareness on PET bottle consumption.
9e00005045a23f3f6b2c9fca094930f8ce42f9f6
Managing Portfolios of Development Projects in a Complex Environment How the UN assign priorities to Programs at the Country
2ec2f8cd6cf1a393acbc7881b8c81a78269cf5f7
Learning Image Embeddings using Convolutional Neural Networks for Improved Multi-Modal Semantics
We construct multi-modal concept representations by concatenating a skip-gram linguistic representation vector with a visual concept representation vector computed using the feature extraction layers of a deep convolutional neural network (CNN) trained on a large labeled object recognition dataset. This transfer learning approach brings a clear performance gain over features based on the traditional bag-of-visual-word approach. Experimental results are reported on the WordSim353 and MEN semantic relatedness evaluation tasks. We use visual features computed using either ImageNet or ESP Game images.
94d1a665c0c7fbd017c9f3c50d35992e1c0c1ed0
Molecular and Morphological Characterization of Aphelenchoides fuchsi sp. n. (Nematoda: Aphelenchoididae) Isolated from Pinus eldarica in Western Iran.
Aphelenchoides fuchsi sp. n. is described and illustrated from bark and wood samples of a weakened Mondell pine in Kermanshah Province, western Iran. The new species has body length of 332 to 400 µm (females) and 365 to 395 µm (males). Lip region set off from body contour. The cuticle is weakly annulated, and there are four lines in the lateral field. The stylet is 8 to 10 μm long and has small basal swellings. The excretory pore is located ca one body diam. posterior to metacorpus valve or 51 to 62 μm from the head. The postuterine sac well developed (60-90 µm). Spicules are relatively short (15-16 μm in dorsal limb) with apex and rostrum rounded, well developed, and the end of the dorsal limb clearly curved ventrad like a hook. The male tail has usual three pairs of caudal papillae (2+2+2) and a well-developed mucro. The female tail is conical, terminating in a complicated step-like projection, usually with many tiny nodular protuberances. The new species belongs to the Group 2 sensu Shahina, category of Aphelenchoides species. Phylogenetic analysis based on small subunit (SSU) and partial large subunit (LSU) sequences of rRNA supported the morphological results.