Query Text
stringlengths
10
40.4k
Ranking 1
stringlengths
12
40.4k
Ranking 2
stringlengths
12
36.2k
Ranking 3
stringlengths
10
36.2k
Ranking 4
stringlengths
13
40.4k
Ranking 5
stringlengths
12
36.2k
Ranking 6
stringlengths
13
36.2k
Ranking 7
stringlengths
10
40.4k
Ranking 8
stringlengths
12
36.2k
Ranking 9
stringlengths
12
36.2k
Ranking 10
stringlengths
12
36.2k
Ranking 11
stringlengths
20
6.21k
Ranking 12
stringlengths
14
8.24k
Ranking 13
stringlengths
28
4.03k
score_0
float64
1
1.25
score_1
float64
0
0.25
score_2
float64
0
0.25
score_3
float64
0
0.25
score_4
float64
0
0.25
score_5
float64
0
0.25
score_6
float64
0
0.25
score_7
float64
0
0.24
score_8
float64
0
0.2
score_9
float64
0
0.03
score_10
float64
0
0
score_11
float64
0
0
score_12
float64
0
0
score_13
float64
0
0
Li-function activated ZNN with finite-time convergence applied to redundant-manipulator kinematic control via time-varying Jacobian matrix pseudoinversion. This paper presents and investigates the application of Zhang neural network (ZNN) activated by Li function to kinematic control of redundant robot manipulators via time-varying Jacobian matrix pseudoinversion. That is, by using Li activation function and by computing the time-varying pseudoinverse of the Jacobian matrix (of the robot manipulator), the resultant ZNN model is applied to redundant-manipulator kinematic control. Note that there are nine novelties and differences of ZNN from the conventional gradient neural network in the research methodology. More importantly, such a Li-function activated ZNN (LFAZNN) model has the property of finite-time convergence (showing its feasibility to redundant-manipulator kinematic control). Simulation results based on a four-link planar robot manipulator and a PA10 robot manipulator further demonstrate the effectiveness of the presented LFAZNN model, as well as show the LFAZNN application prospect.
Kinematic model to control the end-effector of a continuum robot for multi-axis processing This paper presents a novel kinematic approach for controlling the end-effector of a continuum robot for in-situ repair/inspection in restricted and hazardous environments. Forward and inverse kinematic (IK) models have been developed to control the last segment of the continuum robot for performing multi-axis processing tasks using the last six Degrees of Freedom (DoF). The forward kinematics (FK) is proposed using a combination of Euler angle representation and homogeneous matrices. Due to the redundancy of the system, different constraints are proposed to solve the IK for different cases; therefore, the IK model is solved for bending and direction angles between (-pi/2 to + pi/2) radians. In addition, a novel method to calculate the Jacobian matrix is proposed for this type of hyper-redundant kinematics. The error between the results calculated using the proposed Jacobian algorithm and using the partial derivative equations of the FK map (with respect to linear and angular velocity) is evaluated. The error between the two models is found to be insignificant, thus, the Jacobian is validated as a method of calculating the IK for six DoF.
Neural Network and Jacobian Method for Solving the Inverse Statics of a Cable-Driven Soft Arm With Nonconstant Curvature The solution of the inverse kinematics problem of soft manipulators is essential to generate paths in the task space. The inverse kinematics problem of constant curvature or piecewise constant curvature manipulators has already been solved by using different methods, which include closed-form analytical approaches and iterative methods based on the Jacobian method. On the other hand, the inverse kinematics problem of nonconstant curvature manipulators remains unsolved. This study represents one of the first attempts in this direction. It presents both a model-based method and a supervised learning method to solve the inverse statics of nonconstant curvature soft manipulators. In particular, a Jacobian-based method and a feedforward neural network are chosen and tested experimentally. A comparative analysis has been conducted in terms of accuracy and computational time.
Nonlinear recurrent neural networks for finite-time solution of general time-varying linear matrix equations. In order to solve general time-varying linear matrix equations (LMEs) more efficiently, this paper proposes two nonlinear recurrent neural networks based on two nonlinear activation functions. According to Lyapunov theory, such two nonlinear recurrent neural networks are proved to be convergent within finite-time. Besides, by solving differential equation, the upper bounds of the finite convergence time are determined analytically. Compared with existing recurrent neural networks, the proposed two nonlinear recurrent neural networks have a better convergence property (i.e., the upper bound is lower), and thus the accurate solutions of general time-varying LMEs can be obtained with less time. At last, various different situations have been considered by setting different coefficient matrices of general time-varying LMEs and a great variety of computer simulations (including the application to robot manipulators) have been conducted to validate the better finite-time convergence of the proposed two nonlinear recurrent neural networks.
Robust Zeroing Neural-Dynamics and Its Time-Varying Disturbances Suppression Model Applied to Mobile Robot Manipulators. This paper proposes a novel robust zeroing neural-dynamics (RZND) approach as well as its associated model for solving the inverse kinematics problem of mobile robot manipulators. Unlike existing works based on the assumption that neural network models are free of external disturbances, four common forms of time-varying disturbances suppressed by the proposed RZND model are investigated in this pa...
A New Inequality-Based Obstacle-Avoidance MVN Scheme and Its Application to Redundant Robot Manipulators This paper proposes a new inequality-based criterion/constraint with its algorithmic and computational details for obstacle avoidance of redundant robot manipulators. By incorporating such a dynamically updated inequality constraint and the joint physical constraints (such as joint-angle limits and joint-velocity limits), a novel minimum-velocity-norm (MVN) scheme is presented and investigated for robotic redundancy resolution. The resultant obstacle-avoidance MVN scheme resolved at the joint-velocity level is further reformulated as a general quadratic program (QP). Two QP solvers, i.e., a simplified primal–dual neural network based on linear variational inequalities (LVI) and an LVI-based numerical algorithm, are developed and applied for online solution of the QP problem as well as the inequality-based obstacle-avoidance MVN scheme. Simulative results that are based on PA10 robot manipulator and a six-link planar robot manipulator in the presence of window-shaped and point obstacles demonstrate the efficacy and superiority of the proposed obstacle-avoidance MVN scheme. Moreover, experimental results of the proposed MVN scheme implemented on the practical six-link planar robot manipulator substantiate the physical realizability and effectiveness of such a scheme for obstacle avoidance of redundant robot manipulator.
Event-Triggered Finite-Time Control for Networked Switched Linear Systems With Asynchronous Switching. This paper is concerned with the event-triggered finite-time control problem for networked switched linear systems by using an asynchronous switching scheme. Not only the problem of finite-time boundedness, but also the problem of input-output finite-time stability is considered in this paper. Compared with the existing event-triggered results of the switched systems, a new type of event-triggered...
Tabu Search - Part I
Joint Optimization of Radio and Computational Resources for Multicell Mobile-Edge Computing Migrating computational intensive tasks from mobile devices to more resourceful cloud servers is a promising technique to increase the computational capacity of mobile devices while saving their battery energy. In this paper, we consider a MIMO multicell system where multiple mobile users (MUs) ask for computation offloading to a common cloud server. We formulate the offloading problem as the joint optimization of the radio resources􀀀the transmit precoding matrices of the MUs􀀀and the computational resources􀀀the CPU cycles/second assigned by the cloud to each MU􀀀in order to minimize the overall users’ energy consumption, while meeting latency constraints. The resulting optimization problem is nonconvex (in the objective function and constraints). Nevertheless, in the single-user case, we are able to compute the global optimal solution in closed form. In the more challenging multiuser scenario, we propose an iterative algorithm, based on a novel successive convex approximation technique, converging to a local optimal solution of the original nonconvex problem. We then show that the proposed algorithmic framework naturally leads to a distributed and parallel implementation across the radio access points, requiring only a limited coordination/signaling with the cloud. Numerical results show that the proposed schemes outperform disjoint optimization algorithms.
Symbolic model checking for real-time systems We describe finite-state programs over real-numbered time in a guarded-command language with real-valued clocks or, equivalently, as finite automata with real-valued clocks. Model checking answers the question which states of a real-time program satisfy a branching-time specification (given in an extension of CTL with clock variables). We develop an algorithm that computes this set of states symbolically as a fixpoint of a functional on state predicates, without constructing the state space. For this purpose, we introduce a μ-calculus on computation trees over real-numbered time. Unfortunately, many standard program properties, such as response for all nonzeno execution sequences (during which time diverges), cannot be characterized by fixpoints: we show that the expressiveness of the timed μ-calculus is incomparable to the expressiveness of timed CTL. Fortunately, this result does not impair the symbolic verification of "implementable" real-time programs-those whose safety constraints are machine-closed with respect to diverging time and whose fairness constraints are restricted to finite upper bounds on clock values. All timed CTL properties of such programs are shown to be computable as finitely approximable fixpoints in a simple decidable theory.
The industrial indoor channel: large-scale and temporal fading at 900, 2400, and 5200 MHz In this paper, large-scale fading and temporal fading characteristics of the industrial radio channel at 900, 2400, and 5200 MHz are determined. In contrast to measurements performed in houses and in office buildings, few attempts have been made until now to model propagation in industrial environments. In this paper, the industrial environment is categorized into different topographies. Industrial topographies are defined separately for large-scale and temporal fading, and their definition is based upon the specific physical characteristics of the local surroundings affecting both types of fading. Large-scale fading is well expressed by a one-slope path-loss model and excellent agreement with a lognormal distribution is obtained. Temporal fading is found to be Ricean and Ricean K-factors have been determined. Ricean K-factors are found to follow a lognormal distribution.
Cost-Effective Authentic and Anonymous Data Sharing with Forward Security Data sharing has never been easier with the advances of cloud computing, and an accurate analysis on the shared data provides an array of benefits to both the society and individuals. Data sharing with a large number of participants must take into account several issues, including efficiency, data integrity and privacy of data owner. Ring signature is a promising candidate to construct an anonymous and authentic data sharing system. It allows a data owner to anonymously authenticate his data which can be put into the cloud for storage or analysis purpose. Yet the costly certificate verification in the traditional public key infrastructure (PKI) setting becomes a bottleneck for this solution to be scalable. Identity-based (ID-based) ring signature, which eliminates the process of certificate verification, can be used instead. In this paper, we further enhance the security of ID-based ring signature by providing forward security: If a secret key of any user has been compromised, all previous generated signatures that include this user still remain valid. This property is especially important to any large scale data sharing system, as it is impossible to ask all data owners to reauthenticate their data even if a secret key of one single user has been compromised. We provide a concrete and efficient instantiation of our scheme, prove its security and provide an implementation to show its practicality.
Adaptive fuzzy tracking control for switched uncertain strict-feedback nonlinear systems. •Adaptive tracking control for switched strict-feedback nonlinear systems is proposed.•The generalized fuzzy hyperbolic model is used to approximate nonlinear functions.•The designed controller has fewer design parameters comparing with existing methods.
Hardware Circuits Design and Performance Evaluation of a Soft Lower Limb Exoskeleton Soft lower limb exoskeletons (LLEs) are wearable devices that have good potential in walking rehabilitation and augmentation. While a few studies focused on the structure design and assistance force optimization of the soft LLEs, rarely work has been conducted on the hardware circuits design. The main purpose of this work is to present a new soft LLE for walking efficiency improvement and introduce its hardware circuits design. A soft LLE for hip flexion assistance and a hardware circuits system with scalability were proposed. To assess the efficacy of the soft LLE, the experimental tests that evaluate the sensor data acquisition, force tracking performance, lower limb muscle activity and metabolic cost were conducted. The time error in the peak assistance force was just 1%. The reduction in the normalized root-mean-square EMG of the rectus femoris was 7.1%. The net metabolic cost in exoskeleton on condition was reduced by 7.8% relative to walking with no exoskeleton. The results show that the designed hardware circuits can be applied to the soft LLE and the soft LLE is able to improve walking efficiency of wearers.
1.2
0.2
0.2
0.2
0.2
0.1
0.018182
0
0
0
0
0
0
0
A Fuzzy Based Roi Selection For Encryption And Watermarking In Medical Image Using Dwt And Svd Nowadays secure medical image watermarking had become a stringent task in telemedicine. This paper presents a novel medical image watermarking method by fuzzy based Region of Interest (ROI) selection and wavelet transformation approach to embed encrypted watermark. First, the source image will undergo fuzzification to determine the critical points through central and final intensity along the radial line for selecting region of interest (ROI). Second, watermark image is altered to time-frequency domain through wavelet decomposition where the sub-bands are swapped based on the magnitude value obtained through logistic mapping. In the each sub-band all the pixels get swapped, results in fully encrypted image which guarantees the watermark to a secure, reliable and an unbreakable form. In order to provide more robustness to watermark image, singular values are obtained for encrypted watermark image and key component is calculated for avoiding false positive error. Singular values of the source and watermark image are modified through key component. Experimental results reveal that the proposed algorithm attains high robustness and improved security to the watermarked image against various kinds of attacks.
Geometric attacks on image watermarking systems Synchronization errors can lead to significant performance loss in image watermarking methods, as the geometric attacks in the Stirmark benchmark software show. The authors describe the most common types of geometric attacks and survey proposed solutions.
Genetic Optimization Of Radial Basis Probabilistic Neural Networks This paper discusses using genetic algorithms (CA) to optimize the structure of radial basis probabilistic neural networks (RBPNN), including how to select hidden centers of the first hidden layer and to determine the controlling parameter of Gaussian kernel functions. In the process of constructing the genetic algorithm, a novel encoding method is proposed for optimizing the RBPNN structure. This encoding method can not only make the selected hidden centers sufficiently reflect the key distribution characteristic in the space of training samples set and reduce the hidden centers number as few as possible, but also simultaneously determine the optimum controlling parameters of Gaussian kernel functions matching the selected hidden centers. Additionally, we also constructively propose a new fitness function so as to make the designed RBPNN as simple as possible in the network structure in the case of not losing the network performance. Finally, we take the two benchmark problems of discriminating two-spiral problem and classifying the iris data, for example, to test and evaluate this designed GA. The experimental results illustrate that our designed CA can significantly reduce the required hidden centers number, compared with the recursive orthogonal least square algorithm (ROLSA) and the modified K-means algorithm (MKA). In particular, by means of statistical experiments it was proved that the optimized RBPNN by our designed GA, have still a better generalization performance with respect to the ones by the ROLSA and the MKA, in spite of the network scale having been greatly reduced. Additionally, our experimental results also demonstrate that our designed CA is also suitable for optimizing the radial basis function neural networks (RBFNN).
Current status and key issues in image steganography: A survey. Steganography and steganalysis are the prominent research fields in information hiding paradigm. Steganography is the science of invisible communication while steganalysis is the detection of steganography. Steganography means “covered writing” that hides the existence of the message itself. Digital steganography provides potential for private and secure communication that has become the necessity of most of the applications in today’s world. Various multimedia carriers such as audio, text, video, image can act as cover media to carry secret information. In this paper, we have focused only on image steganography. This article provides a review of fundamental concepts, evaluation measures and security aspects of steganography system, various spatial and transform domain embedding schemes. In addition, image quality metrics that can be used for evaluation of stego images and cover selection measures that provide additional security to embedding scheme are also highlighted. Current research trends and directions to improve on existing methods are suggested.
Hybrid local and global descriptor enhanced with colour information. Feature extraction is one of the most important steps in computer vision tasks such as object recognition, image retrieval and image classification. It describes an image by a set of descriptors where the best one gives a high quality description and a low computation. In this study, the authors propose a novel descriptor called histogram of local and global features using speeded up robust featur...
Secure visual cryptography for medical image using modified cuckoo search. Optimal secure visual cryptography for brain MRI medical image is proposed in this paper. Initially, the brain MRI images are selected and then discrete wavelet transform is applied to the brain MRI image for partitioning the image into blocks. Then Gaussian based cuckoo search algorithm is utilized to select the optimal position for every block. Next the proposed technique creates the dual shares from the secret image. Then the secret shares are embedded in the corresponding positions of the blocks. After embedding, the extraction operation is carried out. Here visual cryptographic design is used for the purpose of image authentication and verification. The extracted secret image has dual shares, based on that the receiver views the input image. The authentication and verification of medical image are assisted with the help of target database. All the secret images are registered previously in the target database. The performance of the proposed method is estimated by Peak Signal to Noise Ratio (PSNR), Mean square error (MSE) and normalized correlation. The implementation is done by MATLAB platform.
Multiscale Transform-Based Secured Joint Efficient Medical Image Compression-Encryption Using Symmetric Key Cryptography And Ebcot Encoding Technique Due to the huge advancement in technology, digitizing the multimedia content like text, images and videos has become easier. Everyday huge amounts of multimedia content are shared through the social networks using internet. Sometimes this multimedia content can be hacked by the hackers. This will lead to the misuse of the data. On the other hand, the medical content needs high security and privacy. Motivated by this, joint secured medical image compression-encryption mechanisms are proposed in this paper using multiscale transforms and symmetric key encryption techniques. The multiscale transforms involved in this paper are wavelet transform, bandelet transform and curvelet transform. The encryption techniques involved in this paper are international data encryption algorithm (IDEA), Rivest Cipher (RC5) and Blowfish. The encoding technique used in this paper is embedded block coding with truncation (EBCOT). Experimental results are done for the proposed works and evaluated by using various parameters like Peak Signal-to-Noise Ratio (PSNR), Mean Square Error (MSE), Image Quality Index (IQI) and Structural Similarity Index (SSIM), Average Difference (AD), Normalized Cross-correlation (NK), Structural Content (SC), Maximum difference (MD), Laplacian Mean Squared Error (LMSE) and Normalized Absolute Error (NAE). It is justified that the proposed approaches in this paper yield good results.
A New Efficient Medical Image Cipher Based On Hybrid Chaotic Map And Dna Code In this paper, we propose a novel medical image encryption algorithm based on a hybrid model of deoxyribonucleic acid (DNA) masking, a Secure Hash Algorithm SHA-2 and a new hybrid chaotic map. Our study uses DNA sequences and operations and the chaotic hybrid map to strengthen the cryptosystem. The significant advantages of this approach consist in improving the information entropy which is the most important feature of randomness, resisting against various typical attacks and getting good experimental results. The theoretical analysis and experimental results show that the algorithm improves the encoding efficiency, enhances the security of the ciphertext, has a large key space and a high key sensitivity, and is able to resist against the statistical and exhaustive attacks.
On Multi-Access Edge Computing: A Survey of the Emerging 5G Network Edge Cloud Architecture and Orchestration. Multi-access edge computing (MEC) is an emerging ecosystem, which aims at converging telecommunication and IT services, providing a cloud computing platform at the edge of the radio access network. MEC offers storage and computational resources at the edge, reducing latency for mobile end users and utilizing more efficiently the mobile backhaul and core networks. This paper introduces a survey on ...
An effective implementation of the Lin–Kernighan traveling salesman heuristic This paper describes an implementation of the Lin–Kernighan heuristic, one of the most successful methods for generating optimal or near-optimal solutions for the symmetric traveling salesman problem (TSP). Computational tests show that the implementation is highly effective. It has found optimal solutions for all solved problem instances we have been able to obtain, including a 13,509-city problem (the largest non-trivial problem instance solved to optimality today).
Exoskeletons for human power augmentation The first load-bearing and energetically autonomous exoskeleton, called the Berkeley Lower Extremity Exoskeleton (BLEEX) walks at the average speed of two miles per hour while carrying 75 pounds of load. The project, funded in 2000 by the Defense Advanced Research Project Agency (DARPA) tackled four fundamental technologies: the exoskeleton architectural design, a control algorithm, a body LAN to host the control algorithm, and an on-board power unit to power the actuators, sensors and the computers. This article gives an overview of the BLEEX project.
Assist-As-Needed Training Paradigms For Robotic Rehabilitation Of Spinal Cord Injuries This paper introduces a new "assist-as-needed" (AAN) training paradigm for rehabilitation of spinal cord injuries via robotic training devices. In the pilot study reported in this paper, nine female adult Swiss-Webster mice were divided into three groups, each experiencing a different robotic training control strategy: a fixed training trajectory (Fixed Group, A), an AAN training method without interlimb coordination (Band Group, B), and an AAN training method with bilateral hindlimb coordination (Window Group, C). Fourteen days after complete transection at the mid-thoracic level, the mice were robotically trained to step in the presence of an acutely administered serotonin agonist, quipazine, for a period of six weeks. The mice that received AAN training (Groups B and C) show higher levels of recovery than Group A mice, as measured by the number, consistency, and periodicity of steps realized during testing sessions. Group C displays a higher incidence of alternating stepping than Group B. These results indicate that this training approach may be more effective than fixed trajectory paradigms in promoting robust post-injury stepping behavior. Furthermore, the constraint of interlimb coordination appears to be an important contribution to successful training.
An ID-Based Linearly Homomorphic Signature Scheme and Its Application in Blockchain. Identity-based cryptosystems mean that public keys can be directly derived from user identifiers, such as telephone numbers, email addresses, and social insurance number, and so on. So they can simplify key management procedures of certificate-based public key infrastructures and can be used to realize authentication in blockchain. Linearly homomorphic signature schemes allow to perform linear computations on authenticated data. And the correctness of the computation can be publicly verified. Although a series of homomorphic signature schemes have been designed recently, there are few homomorphic signature schemes designed in identity-based cryptography. In this paper, we construct a new ID-based linear homomorphic signature scheme, which avoids the shortcomings of the use of public-key certificates. The scheme is proved secure against existential forgery on adaptively chosen message and ID attack under the random oracle model. The ID-based linearly homomorphic signature schemes can be applied in e-business and cloud computing. Finally, we show how to apply it to realize authentication in blockchain.
Robot tutor and pupils’ educational ability: Teaching the times tables Research shows promising results of educational robots in language and STEM tasks. In language, more research is available, occasionally in view of individual differences in pupils’ educational ability levels, and learning seems to improve with more expressive robot behaviors. In STEM, variations in robots’ behaviors have been examined with inconclusive results and never while systematically investigating how differences in educational abilities match with different robot behaviors. We applied an autonomously tutoring robot (without tablet, partly WOz) in a 2 × 2 experiment of social vs. neutral behavior in above-average vs. below-average schoolchildren (N = 86; age 8–10 years) while rehearsing the multiplication tables on a one-to-one basis. The standard school test showed that on average, pupils significantly improved their performance even after 3 occasions of 5-min exercises. Beyond-average pupils profited most from a robot tutor, whereas those below average in multiplication benefited more from a robot that showed neutral rather than more social behavior.
1.2
0.2
0.2
0.2
0.2
0.2
0.2
0.05
0
0
0
0
0
0
Blockchain-Enabled Distributed Security Framework for Next-Generation IoT: An Edge Cloud and Software-Defined Network-Integrated Approach The Internet of Things (IoT) plays a vital role in the real world by providing autonomous support for communications and operations, thus enabling and promoting novel services that are commonly used in day-to-day life. It is important to do research on security frameworks for next-generation IoT and develop state-of-the-art confidentiality protection schemes to deal with various attacks on IoT networks. In order to offer prominent features like continuous confidentiality, authentication, and robustness, the blockchain technology comes out as a sustainable solution. A blockchain-enabled distributed security framework using edge cloud and software-defined networking (SDN) is presented in this article. The security attack detection is achieved at the cloud layer, and security attacks are consequently reduced at the edge layer of the IoT network. The SDN-enabled gateway offers dynamic network traffic flow management, which contributes to the security attack recognition through determining doubtful network traffic flows and diminishes security attacks through hindering doubtful flows. The results obtained show that the proposed security framework can efficiently and effectively meet the data confidentiality challenges introduced by the integration of blockchain, edge cloud, and SDN paradigm.
Stochastic Optimal Operation of Microgrid Based on Chaotic Binary Particle Swarm Optimization Based on fuzzy mathematics theory, this paper proposes a fuzzy multi-objective optimization model with related constraints to minimize the total economic cost and network loss of microgrid. Uncontrollable microsources are considered as negative load, and stochastic net load scenarios are generated for taking the uncertainty of their output power and load into account. Cooperating with storage devices of the optimal capacity controllable microsources are treated as variables in the optimization process with the consideration of their start and stop strategy. Chaos optimization algorithm is introduced into binary particle swarm optimization (BPSO) to propose chaotic BPSO (CBPSO). Search capability of BPSO is improved via the chaotic search approach of chaos optimization algorithm. Tests of four benchmark functions show that the proposed CBPSO has better convergence performance than BPSO. Simulation results validate the correctness of the proposed model and the effectiveness of CBPSO.
A Technical Approach to the Energy Blockchain in Microgrids. The present paper considers some technical issues related to the “energy blockchain” paradigm applied to microgrids. In particular, what appears from the study is that the superposition of energy transactions in a microgrid creates a variation of the power losses in all the branches of the microgrid. Traditional power losses allocation in distribution systems takes into account only generators whi...
Energy-Aware Green Adversary Model for Cyberphysical Security in Industrial System Adversary models have been fundamental to the various cryptographic protocols and methods. However, their use in most of the branches of research in computer science is comparatively restricted, primarily in case of the research in cyberphysical security (e.g., vulnerability studies, position confidentiality). In this article, we propose an energy-aware green adversary model for its use in smart industrial environment through achieving confidentiality. Even though, mutually the hardware and the software parts of cyberphysical systems can be improved to decrease its energy consumption, this article focuses on aspects of conserving position and information confidentiality. On the basis of our findings (assumptions, adversary goals, and capabilities) from the literature, we give some testimonials to help practitioners and researchers working in cyberphysical security. The proposed model that runs on real-time anticipatory position-based query scheduling in order to minimize the communication and computation cost for each query, thus, facilitating energy consumption minimization. Moreover, we calculate the transferring/acceptance slots required for each query to avoid deteriorating slots. The experimental results confirm that the proposed approach can diminish energy consumption up to five times in comparison to existing approaches
Energy-Efficient and Trustworthy Data Collection Protocol Based on Mobile Fog Computing in Internet of Things The tremendous growth of interconnected things/devices in the whole world advances to the new paradigm, i.e., Internet of Things (IoT). The IoT use sensor-based embedded systems to interact with others, providing a wide range of applications and services to upper-level users. Undoubtedly, the data collected by the underlying IoTs are the basis of the upper-layer decision and the foundation for all the applications, which requires efficient energy protocols. Moreover, if the collected data are erroneous and untrustworthy, the data protection and application becomes an unrealistic goal, which further leads to unnecessary energy cost. However, the traditional methods cannot solve this problem efficiently and trustworthily. To achieve this goal, in this paper we design a novel energy-efficient and trustworthy protocol based on mobile fog computing. By establishing a trust model on fog elements to evaluate the sensor nodes, the mobile data collection path with the largest utility value is generated, which can avoid visiting unnecessary sensors and collecting untrustworthy data. Theoretical analysis and experimental results validate that our proposed architecture and method outperform traditional data collection methods in both energy and delay.
BLEU: a method for automatic evaluation of machine translation Human evaluations of machine translation are extensive but expensive. Human evaluations can take months to finish and involve human labor that can not be reused. We propose a method of automatic machine translation evaluation that is quick, inexpensive, and language-independent, that correlates highly with human evaluation, and that has little marginal cost per run. We present this method as an automated understudy to skilled human judges which substitutes for them when there is need for quick or frequent evaluations.
Massive MIMO for next generation wireless systems Multi-user MIMO offers big advantages over conventional point-to-point MIMO: it works with cheap single-antenna terminals, a rich scattering environment is not required, and resource allocation is simplified because every active terminal utilizes all of the time-frequency bins. However, multi-user MIMO, as originally envisioned, with roughly equal numbers of service antennas and terminals and frequency-division duplex operation, is not a scalable technology. Massive MIMO (also known as large-scale antenna systems, very large MIMO, hyper MIMO, full-dimension MIMO, and ARGOS) makes a clean break with current practice through the use of a large excess of service antennas over active terminals and time-division duplex operation. Extra antennas help by focusing energy into ever smaller regions of space to bring huge improvements in throughput and radiated energy efficiency. Other benefits of massive MIMO include extensive use of inexpensive low-power components, reduced latency, simplification of the MAC layer, and robustness against intentional jamming. The anticipated throughput depends on the propagation environment providing asymptotically orthogonal channels to the terminals, but so far experiments have not disclosed any limitations in this regard. While massive MIMO renders many traditional research problems irrelevant, it uncovers entirely new problems that urgently need attention: the challenge of making many low-cost low-precision components that work effectively together, acquisition and synchronization for newly joined terminals, the exploitation of extra degrees of freedom provided by the excess of service antennas, reducing internal power consumption to achieve total energy efficiency reductions, and finding new deployment scenarios. This article presents an overview of the massive MIMO concept and contemporary research on the topic.
Deep Residual Learning for Image Recognition Deeper neural networks are more difficult to train. We present a residual learning framework to ease the training of networks that are substantially deeper than those used previously. We explicitly reformulate the layers as learning residual functions with reference to the layer inputs, instead of learning unreferenced functions. We provide comprehensive empirical evidence showing that these residual networks are easier to optimize, and can gain accuracy from considerably increased depth. On the ImageNet dataset we evaluate residual nets with a depth of up to 152 layers - 8× deeper than VGG nets [40] but still having lower complexity. An ensemble of these residual nets achieves 3.57% error on the ImageNet test set. This result won the 1st place on the ILSVRC 2015 classification task. We also present analysis on CIFAR-10 with 100 and 1000 layers. The depth of representations is of central importance for many visual recognition tasks. Solely due to our extremely deep representations, we obtain a 28% relative improvement on the COCO object detection dataset. Deep residual nets are foundations of our submissions to ILSVRC & COCO 2015 competitions1, where we also won the 1st places on the tasks of ImageNet detection, ImageNet localization, COCO detection, and COCO segmentation.
Reaching Agreement in the Presence of Faults The problem addressed here concerns a set of isolated processors, some unknown subset of which may be faulty, that communicate only by means of two-party messages. Each nonfaulty processor has a private value of information that must be communicated to each other nonfaulty processor. Nonfaulty processors always communicate honestly, whereas faulty processors may lie. The problem is to devise an algorithm in which processors communicate their own values and relay values received from others that allows each nonfaulty processor to infer a value for each other processor. The value inferred for a nonfaulty processor must be that processor's private value, and the value inferred for a faulty one must be consistent with the corresponding value inferred by each other nonfaulty processor.It is shown that the problem is solvable for, and only for, n ≥ 3m + 1, where m is the number of faulty processors and n is the total number. It is also shown that if faulty processors can refuse to pass on information but cannot falsely relay information, the problem is solvable for arbitrary n ≥ m ≥ 0. This weaker assumption can be approximated in practice using cryptographic methods.
Reservoir computing approaches to recurrent neural network training Echo State Networks and Liquid State Machines introduced a new paradigm in artificial recurrent neural network (RNN) training, where an RNN (the reservoir) is generated randomly and only a readout is trained. The paradigm, becoming known as reservoir computing, greatly facilitated the practical application of RNNs and outperformed classical fully trained RNNs in many tasks. It has lately become a vivid research field with numerous extensions of the basic idea, including reservoir adaptation, thus broadening the initial paradigm to using different methods for training the reservoir and the readout. This review systematically surveys both current ways of generating/adapting the reservoirs and training different types of readouts. It offers a natural conceptual classification of the techniques, which transcends boundaries of the current “brand-names” of reservoir methods, and thus aims to help in unifying the field and providing the reader with a detailed “map” of it.
Implementing Vehicle Routing Algorithms
Finite-approximation-error-based discrete-time iterative adaptive dynamic programming. In this paper, a new iterative adaptive dynamic programming (ADP) algorithm is developed to solve optimal control problems for infinite horizon discrete-time nonlinear systems with finite approximation errors. First, a new generalized value iteration algorithm of ADP is developed to make the iterative performance index function converge to the solution of the Hamilton-Jacobi-Bellman equation. The ...
An evolutionary programming approach for securing medical images using watermarking scheme in invariant discrete wavelet transformation. •The proposed watermarking scheme utilized improved discrete wavelet transformation (IDWT) to retrieve the invariant wavelet domain.•The entropy mechanism is used to identify the suitable region for insertion of watermark. This will improve the imperceptibility and robustness of the watermarking procedure.•The scaling factors such as PSNR and NC are considered for evaluation of the proposed method and the Particle Swarm Optimization is employed to optimize the scaling factors.
Robot tutor and pupils’ educational ability: Teaching the times tables Research shows promising results of educational robots in language and STEM tasks. In language, more research is available, occasionally in view of individual differences in pupils’ educational ability levels, and learning seems to improve with more expressive robot behaviors. In STEM, variations in robots’ behaviors have been examined with inconclusive results and never while systematically investigating how differences in educational abilities match with different robot behaviors. We applied an autonomously tutoring robot (without tablet, partly WOz) in a 2 × 2 experiment of social vs. neutral behavior in above-average vs. below-average schoolchildren (N = 86; age 8–10 years) while rehearsing the multiplication tables on a one-to-one basis. The standard school test showed that on average, pupils significantly improved their performance even after 3 occasions of 5-min exercises. Beyond-average pupils profited most from a robot tutor, whereas those below average in multiplication benefited more from a robot that showed neutral rather than more social behavior.
1.2
0.2
0.2
0.2
0.1
0
0
0
0
0
0
0
0
0
High capacity reversible data hiding in encrypted image based on image encoding and POB Reversible data-hiding in encrypted image (RDH-EI) can be used for protection of original image by image owner; data hider can embed secret data in the encrypted image without necessity of knowing the original image content. In order to enhance the capacity of embedding in encrypted image, a novel RDH-EI framework based on JPEG-LS and permutation ordered binary (POB) number system is proposed. In the proposed scheme, the original image is first encoded by JPEG-LS, and then the encoded image is encrypted by bit planes. In the phase of data hiding, the encrypted image is first shuffled, and then non-encoded encrypted pixels are totally replaced by the secret data. Lastly, the above generated image is divided into some non-overlapping 2  ×  2 blocks, the block feature of 3-bits which is calculated from the pixels within the block and 5-bits secret data are embedded into the every pixel of the block, and then the pixels with 10-bits data in the block are transformed into the POB value. Experimental results show that the proposed scheme obtains the higher embedding rate compared with some of the state-of-the-art schemes, the weak robustness and performance of verification for tampered parts in the marked image are also achieved for the proposed scheme.
Probabilistic encryption A new probabilistic model of data encryption is introduced. For this model, under suitable complexity assumptions, it is proved that extracting any information about the cleartext from the cyphertext is hard on the average for an adversary with polynomially bounded computational resources. The proof holds for any message space with any probability distribution. The first implementation of this model is presented. The security of this implementation is proved under the interactability assumptin of deciding Quadratic Residuosity modulo composite numbers whose factorization is unknown.
Secure and privacy preserving keyword searching for cloud storage services Cloud storage services enable users to remotely access data in a cloud anytime and anywhere, using any device, in a pay-as-you-go manner. Moving data into a cloud offers great convenience to users since they do not have to care about the large capital investment in both the deployment and management of the hardware infrastructures. However, allowing a cloud service provider (CSP), whose purpose is mainly for making a profit, to take the custody of sensitive data, raises underlying security and privacy issues. To keep user data confidential against an untrusted CSP, a natural way is to apply cryptographic approaches, by disclosing the data decryption key only to authorized users. However, when a user wants to retrieve files containing certain keywords using a thin client, the adopted encryption system should not only support keyword searching over encrypted data, but also provide high performance. In this paper, we investigate the characteristics of cloud storage services and propose a secure and privacy preserving keyword searching (SPKS) scheme, which allows the CSP to participate in the decipherment, and to return only files containing certain keywords specified by the users, so as to reduce both the computational and communication overhead in decryption for users, on the condition of preserving user data privacy and user querying privacy. Performance analysis shows that the SPKS scheme is applicable to a cloud environment.
Integrating Encryption and Marking for Remote Sensing Image Based on Orthogonal Decomposition For the special characters, remote sensing image has higher requirements not only in the security but also in the management; it requires not only the active encryption during storage and transmission for preventing information leakage but also the marking technology to prevent illegal usage as well as copyright protection or even source tracing. Therefore, this paper proposes to integrate encryption and marking technology by the independence and fusion of orthogonal decomposition for the comprehensive security protection of remote sensing image. Under the proposed scheme, encryption and marking technology can achieve the operation independence and content mergence; moreover, there is no special requirement in selecting encryption and marking algorithms. It makes up the shortage of recent integration of encryption and watermarking based on spatial scrambling in applicability and security. According to the experimental results, integration of encryption and marking technology based on orthogonal decomposition satisfies the common constraints of encryption, and marking technology, furthermore, has little impact on remote sensing image data characters and later applications.
Separable reversible data hiding in encrypted images via adaptive embedding strategy with block selection. •An adaptive, separable reversible data hiding scheme in encrypted image is proposed.•Analogues stream-cipher and block permutation are used to encrypt original image.•Classification and selection for encrypted blocks are conducted during embedding.•An accurate prediction strategy was employed to achieve perfect image recovery.•Our scheme has better rate-distortion performance than some state-of-the-art schemes.
Image Feature Extraction in Encrypted Domain With Privacy-Preserving SIFT Privacy has received considerable attention but is still largely ignored in the multimedia community. Consider a cloud computing scenario where the server is resource-abundant, and is capable of finishing the designated tasks. It is envisioned that secure media applications with privacy preservation will be treated seriously. In view of the fact that scale-invariant feature transform (SIFT) has been widely adopted in various fields, this paper is the first to target the importance of privacy-preserving SIFT (PPSIFT) and to address the problem of secure SIFT feature extraction and representation in the encrypted domain. As all of the operations in SIFT must be moved to the encrypted domain, we propose a privacy-preserving realization of the SIFT method based on homomorphic encryption. We show through the security analysis based on the discrete logarithm problem and RSA that PPSIFT is secure against ciphertext only attack and known plaintext attack. Experimental results obtained from different case studies demonstrate that the proposed homomorphic encryption-based privacy-preserving SIFT performs comparably to the original SIFT and that our method is useful in SIFT-based privacy-preserving applications.
Distinctive Image Features from Scale-Invariant Keypoints This paper presents a method for extracting distinctive invariant features from images that can be used to perform reliable matching between different views of an object or scene. The features are invariant to image scale and rotation, and are shown to provide robust matching across a substantial range of affine distortion, change in 3D viewpoint, addition of noise, and change in illumination. The features are highly distinctive, in the sense that a single feature can be correctly matched with high probability against a large database of features from many images. This paper also describes an approach to using these features for object recognition. The recognition proceeds by matching individual features to a database of features from known objects using a fast nearest-neighbor algorithm, followed by a Hough transform to identify clusters belonging to a single object, and finally performing verification through least-squares solution for consistent pose parameters. This approach to recognition can robustly identify objects among clutter and occlusion while achieving near real-time performance.
An introduction to ROC analysis Receiver operating characteristics (ROC) graphs are useful for organizing classifiers and visualizing their performance. ROC graphs are commonly used in medical decision making, and in recent years have been used increasingly in machine learning and data mining research. Although ROC graphs are apparently simple, there are some common misconceptions and pitfalls when using them in practice. The purpose of this article is to serve as an introduction to ROC graphs and as a guide for using them in research.
Toward Massive Machine Type Communications in Ultra-Dense Cellular IoT Networks: Current Issues and Machine Learning-Assisted Solutions The ever-increasing number of resource-constrained machine-type communication (MTC) devices is leading to the critical challenge of fulfilling diverse communication requirements in dynamic and ultra-dense wireless environments. Among different application scenarios that the upcoming 5G and beyond cellular networks are expected to support, such as enhanced mobile broadband (eMBB), massive machine type communications (mMTCs), and ultra-reliable and low latency communications (URLLCs), the mMTC brings the unique technical challenge of supporting a huge number of MTC devices in cellular networks, which is the main focus of this paper. The related challenges include quality of service (QoS) provisioning, handling highly dynamic and sporadic MTC traffic, huge signalling overhead, and radio access network (RAN) congestion. In this regard, this paper aims to identify and analyze the involved technical issues, to review recent advances, to highlight potential solutions and to propose new research directions. First, starting with an overview of mMTC features and QoS provisioning issues, we present the key enablers for mMTC in cellular networks. Along with the highlights on the inefficiency of the legacy random access (RA) procedure in the mMTC scenario, we then present the key features and channel access mechanisms in the emerging cellular IoT standards, namely, LTE-M and narrowband IoT (NB-IoT). Subsequently, we present a framework for the performance analysis of transmission scheduling with the QoS support along with the issues involved in short data packet transmission. Next, we provide a detailed overview of the existing and emerging solutions toward addressing RAN congestion problem, and then identify potential advantages, challenges, and use cases for the applications of emerging machine learning (ML) techniques in ultra-dense cellular networks. Out of several ML techniques, we focus on the application of low-complexity <inline-formula xmlns:mml="http://www.w3.org/1998/Math/MathML" xmlns:xlink="http://www.w3.org/1999/xlink"> <tex-math notation="LaTeX">$Q$ </tex-math></inline-formula> -learning approach in the mMTC scenario along with the recent advances toward enhancing its learning performance and convergence. Finally, we discuss some open research challenges and promising future research directions.
Priced Oblivious Transfer: How to Sell Digital Goods We consider the question of protecting the privacy of customers buying digital goods. More specifically, our goal is to allow a buyer to purchase digital goods from a vendor without letting the vendor learn what, and to the extent possible also when and how much, it is buying. We propose solutions which allow the buyer, after making an initial deposit, to engage in an unlimited number of priced oblivious-transfer protocols, satisfying the following requirements: As long as the buyer's balance contains sufficient funds, it will successfully retrieve the selected item and its balance will be debited by the item's price. However, the buyer should be unable to retrieve an item whose cost exceeds its remaining balance. The vendor should learn nothing except what must inevitably be learned, namely, the amount of interaction and the initial deposit amount (which imply upper bounds on the quantity and total price of all information obtained by the buyer). In particular, the vendor should be unable to learn what the buyer's current balance is or when it actually runs out of its funds. The technical tools we develop, in the process of solving this problem, seem to be of independent interest. In particular, we present the first one-round (two-pass) protocol for oblivious transfer that does not rely on the random oracle model (a very similar protocol was independently proposed by Naor and Pinkas [21]). This protocol is a special case of a more general "conditional disclosure" methodology, which extends a previous approach from [11] and adapts it to the 2-party setting.
Minimum acceleration criterion with constraints implies bang-bang control as an underlying principle for optimal trajectories of arm reaching movements. Rapid arm-reaching movements serve as an excellent test bed for any theory about trajectory formation. How are these movements planned? A minimum acceleration criterion has been examined in the past, and the solution obtained, based on the Euler-Poisson equation, failed to predict that the hand would begin and end the movement at rest (i.e., with zero acceleration). Therefore, this criterion was rejected in favor of the minimum jerk, which was proved to be successful in describing many features of human movements. This letter follows an alternative approach and solves the minimum acceleration problem with constraints using Pontryagin's minimum principle. We use the minimum principle to obtain minimum acceleration trajectories and use the jerk as a control signal. In order to find a solution that does not include nonphysiological impulse functions, constraints on the maximum and minimum jerk values are assumed. The analytical solution provides a three-phase piecewise constant jerk signal (bang-bang control) where the magnitude of the jerk and the two switching times depend on the magnitude of the maximum and minimum available jerk values. This result fits the observed trajectories of reaching movements and takes into account both the extrinsic coordinates and the muscle limitations in a single framework. The minimum acceleration with constraints principle is discussed as a unifying approach for many observations about the neural control of movements.
Online Prediction of Driver Distraction Based on Brain Activity Patterns This paper presents a new computational framework for early detection of driver distractions (map viewing) using brain activity measured by electroencephalographic (EEG) signals. Compared with most studies in the literature, which are mainly focused on the classification of distracted and nondistracted periods, this study proposes a new framework to prospectively predict the start and end of a distraction period, defined by map viewing. The proposed prediction algorithm was tested on a data set of continuous EEG signals recorded from 24 subjects. During the EEG recordings, the subjects were asked to drive from an initial position to a destination using a city map in a simulated driving environment. The overall accuracy values for the prediction of the start and the end of map viewing were 81% and 70%, respectively. The experimental results demonstrated that the proposed algorithm can predict the start and end of map viewing with relatively high accuracy and can be generalized to individual subjects. The outcome of this study has a high potential to improve the design of future intelligent navigation systems. Prediction of the start of map viewing can be used to provide route information based on a driver's needs and consequently avoid map-viewing activities. Prediction of the end of map viewing can be used to provide warnings for potential long map-viewing durations. Further development of the proposed framework and its applications in driver-distraction predictions are also discussed.
A blind medical image watermarking: DWT-SVD based robust and secure approach for telemedicine applications. In this paper, a blind image watermarking scheme based on discrete wavelet transform (DWT) and singular value decomposition (SVD) is proposed. In this scheme, DWT is applied on ROI (region of interest) of the medical image to get different frequency subbands of its wavelet decomposition. On the low frequency subband LL of the ROI, block-SVD is applied to get different singular matrices. A pair of elements with similar values is identified from the left singular value matrix of these selected blocks. The values of these pairs are modified using certain threshold to embed a bit of watermark content. Appropriate threshold is chosen to achieve the imperceptibility and robustness of medical image and watermark contents respectively. For authentication and identification of original medical image, one watermark image (logo) and other text watermark have been used. The watermark image provides authentication whereas the text data represents electronic patient record (EPR) for identification. At receiving end, blind recovery of both watermark contents is performed by a similar comparison scheme used during the embedding process. The proposed algorithm is applied on various groups of medical images like X-ray, CT scan and mammography. This scheme offers better visibility of watermarked image and recovery of watermark content due to DWT-SVD combination. Moreover, use of Hamming error correcting code (ECC) on EPR text bits reduces the BER and thus provides better recovery of EPR. The performance of proposed algorithm with EPR data coding by Hamming code is compared with the BCH error correcting code and it is found that later one perform better. A result analysis shows that imperceptibility of watermarked image is better as PSNR is above 43 dB and WPSNR is above 52 dB for all set of images. In addition, robustness of the scheme is better than existing scheme for similar set of medical images in terms of normalized correlation coefficient (NCC) and bit-error-rate (BER). An analysis is also carried out to verify the performance of the proposed scheme for different size of watermark contents (image and EPR data). It is observed from analysis that the proposed scheme is also appropriate for watermarking of color image. Using proposed scheme, watermark contents are extracted successfully under various noise attacks like JPEG compression, filtering, Gaussian noise, Salt and pepper noise, cropping, filtering and rotation. Performance comparison of proposed scheme with existing schemes shows proposed scheme has better robustness against different types of attacks. Moreover, the proposed scheme is also robust under set of benchmark attacks known as checkmark attacks.
Energy harvesting algorithm considering max flow problem in wireless sensor networks. In Wireless Sensor Networks (WSNs), sensor nodes with poor energy always have bad effect on the data rate or max flow. These nodes are called bottleneck nodes. In this paper, in order to increase the max flow, we assume an energy harvesting WSNs environment to investigate the cooperation of multiple Mobile Chargers (MCs). MCs are mobile robots that use wireless charging technology to charge sensor nodes in WSNs. This means that in energy harvesting WSNs environments, sensor nodes can obtain energy replenishment by using MCs or collecting energy from nature by themselves. In our research, we use MCs to improve the energy of the sensor nodes by performing multiple rounds of unified scheduling, and finally achieve the purpose of increasing the max flow at sinks. Firstly, we model this problem as a Linear Programming (LP) to search the max flow in a round of charging scheduling and prove that the problem is NP-hard. In order to solve the problem, we propose a heuristic approach: deploying MCs in units of paths with the lowest energy node priority. To reduce the energy consumption of MCs and increase the charging efficiency, we also take the optimization of MCs’ moving distance into our consideration. Finally, we extend the method to multiple rounds of scheduling called BottleNeck. Simulation results show that Bottleneck performs well at increasing max flow.
1.2
0.2
0.2
0.2
0.2
0.1
0
0
0
0
0
0
0
0
ReRobApp: A modular and open-source software framework for robotic rehabilitation and human-robot interaction Physical therapy combined with robotic, electrical, and/or multimedia systems is a promising approach for the development of innovative, engaging, and more effective reha-bilitation tasks. To foster the development of new rehabilitation strategies and protocols, we propose a fully-customizable and open-source software framework named ReRobApp for robotic rehabilitation and other human-robot interaction applications. ReRobApp allows, e.g., an easy integration of different robotic devices and sensors in the same rehabilitation task, and the development of control strategies based on an block programming intuitive architecture. To showcase the versatility of our proposed framework, we performed case studies using serious games, functional electrical stimulation, electromyography, and two different lower limb robotic devices for ankle rehabilitation in a single multiplayer task with different subjects.
The Sybil Attack Large-scale peer-to-peer systems facesecurity threats from faulty or hostile remotecomputing elements. To resist these threats, manysuch systems employ redundancy. However, if asingle faulty entity can present multiple identities,it can control a substantial fraction of the system,thereby undermining this redundancy. Oneapproach to preventing these &quot;Sybil attacks&quot; is tohave a trusted agency certify identities. Thispaper shows that, without a logically centralizedauthority, Sybil...
BLEU: a method for automatic evaluation of machine translation Human evaluations of machine translation are extensive but expensive. Human evaluations can take months to finish and involve human labor that can not be reused. We propose a method of automatic machine translation evaluation that is quick, inexpensive, and language-independent, that correlates highly with human evaluation, and that has little marginal cost per run. We present this method as an automated understudy to skilled human judges which substitutes for them when there is need for quick or frequent evaluations.
Computational thinking Summary form only given. My vision for the 21st century, Computational Thinking, will be a fundamental skill used by everyone in the world. To reading, writing, and arithmetic, we should add computational thinking to every child's analytical ability. Computational thinking involves solving problems, designing systems, and understanding human behavior by drawing on the concepts fundamental to computer science. Thinking like a computer scientist means more than being able to program a computer. It requires the ability to abstract and thus to think at multiple levels of abstraction. In this talk I will give many examples of computational thinking, argue that it has already influenced other disciplines, and promote the idea that teaching computational thinking can not only inspire future generations to enter the field of computer science but benefit people in all fields.
Fuzzy logic in control systems: fuzzy logic controller. I.
Switching between stabilizing controllers This paper deals with the problem of switching between several linear time-invariant (LTI) controllers—all of them capable of stabilizing a speci4c LTI process—in such a way that the stability of the closed-loop system is guaranteed for any switching sequence. We show that it is possible to 4nd realizations for any given family of controller transfer matrices so that the closed-loop system remains stable, no matter how we switch among the controller. The motivation for this problem is the control of complex systems where con8icting requirements make a single LTI controller unsuitable. ? 2002 Published by Elsevier Science Ltd.
Tabu Search - Part I
Bidirectional recurrent neural networks In the first part of this paper, a regular recurrent neural network (RNN) is extended to a bidirectional recurrent neural network (BRNN). The BRNN can be trained without the limitation of using input information just up to a preset future frame. This is accomplished by training it simultaneously in positive and negative time direction. Structure and training procedure of the proposed network are explained. In regression and classification experiments on artificial data, the proposed structure gives better results than other approaches. For real data, classification experiments for phonemes from the TIMIT database show the same tendency. In the second part of this paper, it is shown how the proposed bidirectional structure can be easily modified to allow efficient estimation of the conditional posterior probability of complete symbol sequences without making any explicit assumption about the shape of the distribution. For this part, experiments on real data are reported
An intensive survey of fair non-repudiation protocols With the phenomenal growth of the Internet and open networks in general, security services, such as non-repudiation, become crucial to many applications. Non-repudiation services must ensure that when Alice sends some information to Bob over a network, neither Alice nor Bob can deny having participated in a part or the whole of this communication. Therefore a fair non-repudiation protocol has to generate non-repudiation of origin evidences intended to Bob, and non-repudiation of receipt evidences destined to Alice. In this paper, we clearly define the properties a fair non-repudiation protocol must respect, and give a survey of the most important non-repudiation protocols without and with trusted third party (TTP). For the later ones we discuss the evolution of the TTP's involvement and, between others, describe the most recent protocol using a transparent TTP. We also discuss some ad-hoc problems related to the management of non-repudiation evidences.
Dynamic movement and positioning of embodied agents in multiparty conversations For embodied agents to engage in realistic multiparty conversation, they must stand in appropriate places with respect to other agents and the environment. When these factors change, such as an agent joining the conversation, the agents must dynamically move to a new location and/or orientation to accommodate. This paper presents an algorithm for simulating movement of agents based on observed human behavior using techniques developed for pedestrian movement in crowd simulations. We extend a previous group conversation simulation to include an agent motion algorithm. We examine several test cases and show how the simulation generates results that mirror real-life conversation settings.
An improved genetic algorithm with conditional genetic operators and its application to set-covering problem The genetic algorithm (GA) is a popular, biologically inspired optimization method. However, in the GA there is no rule of thumb to design the GA operators and select GA parameters. Instead, trial-and-error has to be applied. In this paper we present an improved genetic algorithm in which crossover and mutation are performed conditionally instead of probability. Because there are no crossover rate and mutation rate to be selected, the proposed improved GA can be more easily applied to a problem than the conventional genetic algorithms. The proposed improved genetic algorithm is applied to solve the set-covering problem. Experimental studies show that the improved GA produces better results over the conventional one and other methods.
Lane-level traffic estimations using microscopic traffic variables This paper proposes a novel inference method to estimate lane-level traffic flow, time occupancy and vehicle inter-arrival time on road segments where local information could not be measured and assessed directly. The main contributions of the proposed method are 1) the ability to perform lane-level estimations of traffic flow, time occupancy and vehicle inter-arrival time and 2) the ability to adapt to different traffic regimes by assessing only microscopic traffic variables. We propose a modified Kriging estimation model which explicitly takes into account both spatial and temporal variability. Performance evaluations are conducted using real-world data under different traffic regimes and it is shown that the proposed method outperforms a Kalman filter-based approach.
Convolutional Neural Network-Based Classification of Driver's Emotion during Aggressive and Smooth Driving Using Multi-Modal Camera Sensors. Because aggressive driving often causes large-scale loss of life and property, techniques for advance detection of adverse driver emotional states have become important for the prevention of aggressive driving behaviors. Previous studies have primarily focused on systems for detecting aggressive driver emotion via smart-phone accelerometers and gyro-sensors, or they focused on methods of detecting physiological signals using electroencephalography (EEG) or electrocardiogram (ECG) sensors. Because EEG and ECG sensors cause discomfort to drivers and can be detached from the driver's body, it becomes difficult to focus on bio-signals to determine their emotional state. Gyro-sensors and accelerometers depend on the performance of GPS receivers and cannot be used in areas where GPS signals are blocked. Moreover, if driving on a mountain road with many quick turns, a driver's emotional state can easily be misrecognized as that of an aggressive driver. To resolve these problems, we propose a convolutional neural network (CNN)-based method of detecting emotion to identify aggressive driving using input images of the driver's face, obtained using near-infrared (NIR) light and thermal camera sensors. In this research, we conducted an experiment using our own database, which provides a high classification accuracy for detecting driver emotion leading to either aggressive or smooth (i.e., relaxed) driving. Our proposed method demonstrates better performance than existing methods.
Ethical Considerations Of Applying Robots In Kindergarten Settings: Towards An Approach From A Macroperspective In child-robot interaction (cHRI) research, many studies pursue the goal to develop interactive systems that can be applied in everyday settings. For early education, increasingly, the setting of a kindergarten is targeted. However, when cHRI and research are brought into a kindergarten, a range of ethical and related procedural aspects have to be considered and dealt with. While ethical models elaborated within other human-robot interaction settings, e.g., assisted living contexts, can provide some important indicators for relevant issues, we argue that it is important to start developing a systematic approach to identify and tackle those ethical issues which rise with cHRI in kindergarten settings on a more global level and address the impact of the technology from a macroperspective beyond the effects on the individual. Based on our experience in conducting studies with children in general and pedagogical considerations on the role of the institution of kindergarten in specific, in this paper, we enfold some relevant aspects that have barely been addressed in an explicit way in current cHRI research. Four areas are analyzed and key ethical issues are identified in each area: (1) the institutional setting of a kindergarten, (2) children as a vulnerable group, (3) the caregivers' role, and (4) pedagogical concepts. With our considerations, we aim at (i) broadening the methodology of the current studies within the area of cHRI, (ii) revalidate it based on our comprehensive empirical experience with research in kindergarten settings, both laboratory and real-world contexts, and (iii) provide a framework for the development of a more systematic approach to address the ethical issues in cHRI research within kindergarten settings.
1
0
0
0
0
0
0
0
0
0
0
0
0
0
Probabilistic Forecasting with Temporal Convolutional Neural Network. We present a probabilistic forecasting framework based on convolutional neural network (CNN) for multiple related time series forecasting. The framework can be applied to estimate probability density under both parametric and non-parametric settings. More specifically, stacked residual blocks based on dilated causal convolutional nets are constructed to capture the temporal dependencies of the series. Combined with representation learning, our approach is able to learn complex patterns such as seasonality, holiday effects within and across series, and to leverage those patterns for more accurate forecasts, especially when historical data is sparse or unavailable. Extensive empirical studies are performed on several real-world datasets, including datasets from JD.com, China’s largest online retailer. The results show that our framework compares favorably to the state-of-the-art in both point and probabilistic forecasting.
Estimation of prediction error by using K-fold cross-validation Estimation of prediction accuracy is important when our aim is prediction. The training error is an easy estimate of prediction error, but it has a downward bias. On the other hand, K-fold cross-validation has an upward bias. The upward bias may be negligible in leave-one-out cross-validation, but it sometimes cannot be neglected in 5-fold or 10-fold cross-validation, which are favored from a computational standpoint. Since the training error has a downward bias and K-fold cross-validation has an upward bias, there will be an appropriate estimate in a family that connects the two estimates. In this paper, we investigate two families that connect the training error and K-fold cross-validation.
Traffic Flow Forecasting for Urban Work Zones None of numerous existing traffic flow forecasting models focus on work zones. Work zone events create conditions that are different from both normal operating conditions and incident conditions. In this paper, four models were developed for forecasting traffic flow for planned work zone events. The four models are random forest, regression tree, multilayer feedforward neural network, and nonparametric regression. Both long-term and short-term traffic flow forecasting applications were investigated. Long-term forecast involves forecasting 24 h in advance using historical traffic data, and short-term forecasts involves forecasting 1 h and 45, 30, and 15 min in advance using real-time temporal and spatial traffic data. Models were evaluated using data from work zone events on two types of roadways, a freeway, i.e., I-270, and a signalized arterial, i.e., MO-141, in St. Louis, MO, USA. The results showed that the random forest model yielded the most accurate long-term and short-term work zone traffic flow forecasts. For freeway data, the most influential variables were the latest interval's look-back traffic flows at the upstream, downstream, and current locations. For arterial data, the most influential variables were the traffic flows from the three look-back intervals at the current location only.
Daily long-term traffic flow forecasting based on a deep neural network. •A new deep learning algorithm to predict daily long-term traffic flow data using contextual factors.•Deep neutral network to mine the relationship between traffic flow data and contextual factors.•Advanced batch training can effectively improve convergence of the training process.
A 3D CNN-LSTM-Based Image-to-Image Foreground Segmentation The video-based separation of foreground (FG) and background (BG) has been widely studied due to its vital role in many applications, including intelligent transportation and video surveillance. Most of the existing algorithms are based on traditional computer vision techniques that perform pixel-level processing assuming that FG and BG possess distinct visual characteristics. Recently, state-of-the-art solutions exploit deep learning models targeted originally for image classification. Major drawbacks of such a strategy are the lacking delineation of FG regions due to missing temporal information as they segment the FG based on a single frame object detection strategy. To grapple with this issue, we excogitate a 3D convolutional neural network (3D CNN) with long short-term memory (LSTM) pipelines that harness seminal ideas, viz., fully convolutional networking, 3D transpose convolution, and residual feature flows. Thence, an FG-BG segmenter is implemented in an encoder-decoder fashion and trained on representative FG-BG segments. The model devises a strategy called double encoding and slow decoding, which fuses the learned spatio-temporal cues with appropriate feature maps both in the down-sampling and up-sampling paths for achieving well generalized FG object representation. Finally, from the Sigmoid confidence map generated by the 3D CNN-LSTM model, the FG is identified automatically by using Nobuyuki Otsu’s method and an empirical global threshold. The analysis of experimental results via standard quantitative metrics on 16 benchmark datasets including both indoor and outdoor scenes validates that the proposed 3D CNN-LSTM achieves competitive performance in terms of figure of merit evaluated against prior and state-of-the-art methods. Besides, a failure analysis is conducted on 20 video sequences from the DAVIS 2016 dataset.
Improving Traffic Flow Prediction With Weather Information in Connected Cars: A Deep Learning Approach. Transportation systems might be heavily affected by factors such as accidents and weather. Specifically, inclement weather conditions may have a drastic impact on travel time and traffic flow. This study has two objectives: first, to investigate a correlation between weather parameters and traffic flow and, second, to improve traffic flow prediction by proposing a novel holistic architecture. It i...
Accurate Self-Localization in RFID Tag Information Grids Using FIR Filtering Grid navigation spaces nested with the radio-frequency identification (RFID) tags are promising for industrial and other needs, because each tag can deliver information about a local two-dimensional or three-dimensional surrounding. The approach, however, requires high accuracy in vehicle self-localization. Otherwise, errors may lead to collisions; possibly even fatal. We propose a new extended finite impulse response (EFIR) filtering algorithm and show that it meets this need. The EFIR filter requires an optimal averaging interval, but does not involve the noise statistics which are often not well known to the engineer. It is more accurate than the extended Kalman filter (EKF) under real operation conditions and its iterative algorithm has the Kalman form. Better performance of the proposed EFIR filter is demonstrated based on extensive simulations in a comparison to EKF, which is widely used in RFID tag grids. We also show that errors in noise covariances may provoke divergence in EKF, whereas the EFIR filter remains stable and is thus more robust.
Evolutionary computation: comments on the history and current state Evolutionary computation has started to receive significant attention during the last decade, although the origins can be traced back to the late 1950's. This article surveys the history as well as the current state of this rapidly growing field. We describe the purpose, the general structure, and the working principles of different approaches, including genetic algorithms (GA) (with links to genetic programming (GP) and classifier systems (CS)), evolution strategies (ES), and evolutionary programming (EP) by analysis and comparison of their most important constituents (i.e. representations, variation operators, reproduction, and selection mechanism). Finally, we give a brief overview on the manifold of application domains, although this necessarily must remain incomplete
Supporting social navigation on the World Wide Web This paper discusses a navigation behavior on Internet information services, in particular the World Wide Web, which is characterized by pointing out of information using various communication tools. We call this behavior social navigation as it is based on communication and interaction with other users, be that through email, or any other means of communication. Social navigation phenomena are quite common although most current tools (like Web browsers or email clients) offer very little support for it. We describe why social navigation is useful and how it can be better supported in future systems. We further describe two prototype systems that, although originally not designed explicitly as tools for social navigation, provide features that are typical for social navigation systems. One of these systems, the Juggler system, is a combination of a textual virtual environment and a Web client. The other system is a prototype of a Web- hotlist organizer, called Vortex. We use both systems to describe fundamental principles of social navigation systems.
Proofs of Storage from Homomorphic Identification Protocols Proofs of storage (PoS) are interactive protocols allowing a client to verify that a server faithfully stores a file. Previous work has shown that proofs of storage can be constructed from any homomorphic linear authenticator (HLA). The latter, roughly speaking, are signature/message authentication schemes where `tags' on multiple messages can be homomorphically combined to yield a `tag' on any linear combination of these messages. We provide a framework for building public-key HLAs from any identification protocol satisfying certain homomorphic properties. We then show how to turn any public-key HLA into a publicly-verifiable PoS with communication complexity independent of the file length and supporting an unbounded number of verifications. We illustrate the use of our transformations by applying them to a variant of an identification protocol by Shoup, thus obtaining the first unbounded-use PoS based on factoring (in the random oracle model).
Well-Solvable Special Cases of the Traveling Salesman Problem: A Survey. The traveling salesman problem (TSP) belongs to the most basic, most important, and most investigated problems in combinatorial optimization. Although it is an ${\cal NP}$-hard problem, many of its special cases can be solved efficiently in polynomial time. We survey these special cases with emphasis on the results that have been obtained during the decade 1985--1995. This survey complements an earlier survey from 1985 compiled by Gilmore, Lawler, and Shmoys [The Traveling Salesman Problem---A Guided Tour of Combinatorial Optimization, Wiley, Chichester, pp. 87--143].
Reinforcement Q-learning for optimal tracking control of linear discrete-time systems with unknown dynamics. In this paper, a novel approach based on the Q-learning algorithm is proposed to solve the infinite-horizon linear quadratic tracker (LQT) for unknown discrete-time systems in a causal manner. It is assumed that the reference trajectory is generated by a linear command generator system. An augmented system composed of the original system and the command generator is constructed and it is shown that the value function for the LQT is quadratic in terms of the state of the augmented system. Using the quadratic structure of the value function, a Bellman equation and an augmented algebraic Riccati equation (ARE) for solving the LQT are derived. In contrast to the standard solution of the LQT, which requires the solution of an ARE and a noncausal difference equation simultaneously, in the proposed method the optimal control input is obtained by only solving an augmented ARE. A Q-learning algorithm is developed to solve online the augmented ARE without any knowledge about the system dynamics or the command generator. Convergence to the optimal solution is shown. A simulation example is used to verify the effectiveness of the proposed control scheme.
An indoor localization solution using Bluetooth RSSI and multiple sensors on a smartphone. In this paper, we propose an indoor positioning system using a Bluetooth receiver, an accelerometer, a magnetic field sensor, and a barometer on a smartphone. The Bluetooth receiver is used to estimate distances from beacons. The accelerometer and magnetic field sensor are used to trace the movement of moving people in the given space. The horizontal location of the person is determined by received signal strength indications (RSSIs) and the traced movement. The barometer is used to measure the vertical position where a person is located. By combining RSSIs, the traced movement, and the vertical position, the proposed system estimates the indoor position of moving people. In experiments, the proposed approach showed excellent performance in localization with an overall error of 4.8%.
Attitudes Towards Social Robots In Education: Enthusiast, Practical, Troubled, Sceptic, And Mindfully Positive While social robots bring new opportunities for education, they also come with moral challenges. Therefore, there is a need for moral guidelines for the responsible implementation of these robots. When developing such guidelines, it is important to include different stakeholder perspectives. Existing (qualitative) studies regarding these perspectives however mainly focus on single stakeholders. In this exploratory study, we examine and compare the attitudes of multiple stakeholders on the use of social robots in primary education, using a novel questionnaire that covers various aspects of moral issues mentioned in earlier studies. Furthermore, we also group the stakeholders based on similarities in attitudes and examine which socio-demographic characteristics influence these attitude types. Based on the results, we identify five distinct attitude profiles and show that the probability of belonging to a specific profile is affected by such characteristics as stakeholder type, age, education and income. Our results also indicate that social robots have the potential to be implemented in education in a morally responsible way that takes into account the attitudes of various stakeholders, although there are multiple moral issues that need to be addressed first. Finally, we present seven (practical) implications for a responsible application of social robots in education following from our results. These implications provide valuable insights into how social robots should be implemented.
1.2
0.2
0.2
0.2
0.2
0.05
0
0
0
0
0
0
0
0
Radial shifted Legendre moments for image analysis and invariant image recognition. The rotation, scaling and translation invariant property of image moments has a high significance in image recognition. Legendre moments as a classical orthogonal moment have been widely used in image analysis and recognition. Since Legendre moments are defined in Cartesian coordinate, the rotation invariance is difficult to achieve. In this paper, we first derive two types of transformed Legendre polynomial: substituted and weighted radial shifted Legendre polynomials. Based on these two types of polynomials, two radial orthogonal moments, named substituted radial shifted Legendre moments and weighted radial shifted Legendre moments (SRSLMs and WRSLMs) are proposed. The proposed moments are orthogonal in polar coordinate domain and can be thought as generalized and orthogonalized complex moments. They have better image reconstruction performance, lower information redundancy and higher noise robustness than the existing radial orthogonal moments. At last, a mathematical framework for obtaining the rotation, scaling and translation invariants of these two types of radial shifted Legendre moments is provided. Theoretical and experimental results show the superiority of the proposed methods in terms of image reconstruction capability and invariant recognition accuracy under both noisy and noise-free conditions.
Combined invariants to similarity transformation and to blur using orthogonal Zernike moments. The derivation of moment invariants has been extensively investigated in the past decades. In this paper, we construct a set of invariants derived from Zernike moments which is simultaneously invariant to similarity transformation and to convolution with circularly symmetric point spread function (PSF). Two main contributions are provided: the theoretical framework for deriving the Zernike moments of a blurred image and the way to construct the combined geometric-blur invariants. The performance of the proposed descriptors is evaluated with various PSFs and similarity transformations. The comparison of the proposed method with the existing ones is also provided in terms of pattern recognition accuracy, template matching and robustness to noise. Experimental results show that the proposed descriptors perform on the overall better.
Fast computation of Jacobi-Fourier moments for invariant image recognition The Jacobi-Fourier moments (JFMs) provide a wide class of orthogonal rotation invariant moments (ORIMs) which are useful for many image processing, pattern recognition and computer vision applications. They, however, suffer from high time complexity and numerical instability at high orders of moment. In this paper, a fast method based on the recursive computation of radial kernel function of JFMs is proposed which not only reduces time complexity but also improves their numerical stability. Fast recursive method for the computation of Jacobi-Fourier moments is proposed.The proposed method not only reduces time complexity but also improves numerical stability of moments.Better image reconstruction is achieved with lower reconstruction error.Proposed method is useful for many image processing, pattern recognition and computer vision applications.
Image representation by harmonic transforms with parameters in SL(2, R). •New kinds of invariant harmonic transforms with parameters in SL(2,R) are proposed.•The capabilities of the proposed PLCT and the 2-D LCTS on image representation are analyzed.
Lossless medical image watermarking method based on significant difference of cellular automata transform coefficient. Conventional medical image watermarking techniques focus on improving invisibility and robustness of the watermarking mechanism to prevent medical disputes. This paper proposes a medical image watermarking algorithm based on the significant difference of cellular automata transform (CAT) for copyright protection. The medical image is firstly subsampled into four subimages, and two images are randomly chosen to obtain two low-frequency bandwidths using CAT. Coefficients within a low-frequency bandwidth are important information in an image. Hence, the difference between two low-frequency bandwidths is used as an important feature in the medical image. From these important features, watermarks and cover images can be used to generate an ownership share image (OSI) used for verifying the medical image. Besides appearing like cover images, the OSI will also be able to register with a third party. When a suspected medical image requires verification, the important features from the suspected medical image are first extracted. The master share image (MSI) can be generated from the important features from the suspected medical image. Lastly, the OSI and MSI can be combined to extract the watermark to verify the suspected medical image. The advantage of our proposed method is that the medical image does not require alteration to protect the copyright of the medical image. This means that while the image is protected, medical disputes will be unlikely and the appearance of the registered OSI will carry significant data to make management more convenient. Lastly, the proposed method has the features of having better security, invisibility, and robustness. Moreover, experimental results have demonstrated that our method results in good performance.
Robust watermarking scheme for color image based on quaternion-type moment invariants and visual cryptography. This paper introduces a novel robust watermarking scheme for copyright protection of color image based on quaternion-type moment invariants and visual cryptography. As a secure way to allow secret sharing of images, visual cryptography realizes encryption of classified information and the decryption is performed through human visual system. The proposed scheme represents the color image into a quaternion matrix, so that it can deal with the multichannel information in a holistic way. Then the quaternion moments are applied to extract the invariant features, which are crucial to generate the master share. Together with the scrambled watermark, they are used for constructing the ownership share based on visual cryptography. Afterwards, the ownership share is registered and responsible for authentication. A set of experiments has been conducted to illustrate the validity and feasibility of the proposed scheme as well as better robustness against different attacks.
Two-Dimensional Polar Harmonic Transforms for Invariant Image Representation This paper introduces a set of 2D transforms, based on a set of orthogonal projection bases, to generate a set of features which are invariant to rotation. We call these transforms Polar Harmonic Transforms (PHTs). Unlike the well-known Zernike and pseudo-Zernike moments, the kernel computation of PHTs is extremely simple and has no numerical stability issue whatsoever. This implies that PHTs encompass the orthogonality and invariance advantages of Zernike and pseudo-Zernike moments, but are free from their inherent limitations. This also means that PHTs are well suited for application where maximal discriminant information is needed. Furthermore, PHTs make available a large set of features for further feature selection in the process of seeking for the best discriminative or representative features for a particular application.
Tabu Search - Part I
Joint Optimization of Radio and Computational Resources for Multicell Mobile-Edge Computing Migrating computational intensive tasks from mobile devices to more resourceful cloud servers is a promising technique to increase the computational capacity of mobile devices while saving their battery energy. In this paper, we consider a MIMO multicell system where multiple mobile users (MUs) ask for computation offloading to a common cloud server. We formulate the offloading problem as the joint optimization of the radio resources􀀀the transmit precoding matrices of the MUs􀀀and the computational resources􀀀the CPU cycles/second assigned by the cloud to each MU􀀀in order to minimize the overall users’ energy consumption, while meeting latency constraints. The resulting optimization problem is nonconvex (in the objective function and constraints). Nevertheless, in the single-user case, we are able to compute the global optimal solution in closed form. In the more challenging multiuser scenario, we propose an iterative algorithm, based on a novel successive convex approximation technique, converging to a local optimal solution of the original nonconvex problem. We then show that the proposed algorithmic framework naturally leads to a distributed and parallel implementation across the radio access points, requiring only a limited coordination/signaling with the cloud. Numerical results show that the proposed schemes outperform disjoint optimization algorithms.
Symbolic model checking for real-time systems We describe finite-state programs over real-numbered time in a guarded-command language with real-valued clocks or, equivalently, as finite automata with real-valued clocks. Model checking answers the question which states of a real-time program satisfy a branching-time specification (given in an extension of CTL with clock variables). We develop an algorithm that computes this set of states symbolically as a fixpoint of a functional on state predicates, without constructing the state space. For this purpose, we introduce a μ-calculus on computation trees over real-numbered time. Unfortunately, many standard program properties, such as response for all nonzeno execution sequences (during which time diverges), cannot be characterized by fixpoints: we show that the expressiveness of the timed μ-calculus is incomparable to the expressiveness of timed CTL. Fortunately, this result does not impair the symbolic verification of "implementable" real-time programs-those whose safety constraints are machine-closed with respect to diverging time and whose fairness constraints are restricted to finite upper bounds on clock values. All timed CTL properties of such programs are shown to be computable as finitely approximable fixpoints in a simple decidable theory.
Teaching-Learning-Based Optimization: An optimization method for continuous non-linear large scale problems An efficient optimization method called 'Teaching-Learning-Based Optimization (TLBO)' is proposed in this paper for large scale non-linear optimization problems for finding the global solutions. The proposed method is based on the effect of the influence of a teacher on the output of learners in a class. The basic philosophy of the method is explained in detail. The effectiveness of the method is tested on many benchmark problems with different characteristics and the results are compared with other population based methods.
An Automatic Screening Approach for Obstructive Sleep Apnea Diagnosis Based on Single-Lead Electrocardiogram Traditional approaches for obstructive sleep apnea (OSA) diagnosis are apt to using multiple channels of physiological signals to detect apnea events by dividing the signals into equal-length segments, which may lead to incorrect apnea event detection and weaken the performance of OSA diagnosis. This paper proposes an automatic-segmentation-based screening approach with the single channel of Electrocardiogram (ECG) signal for OSA subject diagnosis, and the main work of the proposed approach lies in three aspects: (i) an automatic signal segmentation algorithm is adopted for signal segmentation instead of the equal-length segmentation rule; (ii) a local median filter is improved for reduction of the unexpected RR intervals before signal segmentation; (iii) the designed OSA severity index and additional admission information of OSA suspects are plugged into support vector machine (SVM) for OSA subject diagnosis. A real clinical example from PhysioNet database is provided to validate the proposed approach and an average accuracy of 97.41% for subject diagnosis is obtained which demonstrates the effectiveness for OSA diagnosis.
Automated Detection of Obstructive Sleep Apnea Events from a Single-Lead Electrocardiogram Using a Convolutional Neural Network. In this study, we propose a method for the automated detection of obstructive sleep apnea (OSA) from a single-lead electrocardiogram (ECG) using a convolutional neural network (CNN). A CNN model was designed with six optimized convolution layers including activation, pooling, and dropout layers. One-dimensional (1D) convolution, rectified linear units (ReLU), and max pooling were applied to the convolution, activation, and pooling layers, respectively. For training and evaluation of the CNN model, a single-lead ECG dataset was collected from 82 subjects with OSA and was divided into training (including data from 63 patients with 34,281 events) and testing (including data from 19 patients with 8571 events) datasets. Using this CNN model, a precision of 0.99%, a recall of 0.99%, and an F-score of 0.99% were attained with the training dataset; these values were all 0.96% when the CNN was applied to the testing dataset. These results show that the proposed CNN model can be used to detect OSA accurately on the basis of a single-lead ECG. Ultimately, this CNN model may be used as a screening tool for those suspected to suffer from OSA.
Hardware Circuits Design and Performance Evaluation of a Soft Lower Limb Exoskeleton Soft lower limb exoskeletons (LLEs) are wearable devices that have good potential in walking rehabilitation and augmentation. While a few studies focused on the structure design and assistance force optimization of the soft LLEs, rarely work has been conducted on the hardware circuits design. The main purpose of this work is to present a new soft LLE for walking efficiency improvement and introduce its hardware circuits design. A soft LLE for hip flexion assistance and a hardware circuits system with scalability were proposed. To assess the efficacy of the soft LLE, the experimental tests that evaluate the sensor data acquisition, force tracking performance, lower limb muscle activity and metabolic cost were conducted. The time error in the peak assistance force was just 1%. The reduction in the normalized root-mean-square EMG of the rectus femoris was 7.1%. The net metabolic cost in exoskeleton on condition was reduced by 7.8% relative to walking with no exoskeleton. The results show that the designed hardware circuits can be applied to the soft LLE and the soft LLE is able to improve walking efficiency of wearers.
1.2
0.2
0.2
0.2
0.2
0.1
0.022222
0
0
0
0
0
0
0
Redundancy resolution of a human arm for controlling a seven DOF wearable robotic system. The human arm including the shoulder, elbow, wrist joints and exclusion scapular motion has 7 Degrees of Freedom (DOF) while positioning of the wrist in space and orientating the palm is a task that requires 6 DOF. As such it includes one more DOF than is needed to complete the task. Given the redundant nature of the arm, multiple arm configurations can be used to complete a task, which is expressed mathematically by none unique solution for the inverse kinematics. Despite this mathematical difficulty, the motor control provides a unique solution for the arm redundancy as the arm is moved in space. Resolving this redundancy is becoming critical as the human interacts with a wearable robotic system(exoskeleton) which includes the same redundancy as the human arm. Therefore, the inverse kinematics solution resolving the redundancy of these two coupled systems must be identical in order to guarantee a seamless integration. The redundancy of the arm can be formulated kinematically by defining the swivel angle - the rotation angle of the plane including the upper and lower arm around a virtual axis connecting the shoulder and wrist joints which are fixed in space. Analyzing reaching tasks recorded with a motion capture lab indicates that the swivel angle is selected such that when the elbow joint is flexed, the palm points the head. Based on these experimental results, selecting the point around the center of the head as a stationary target allows to calculate the swivel angle and in that way to resolve the human arm redundancy. Experimental results indicated that by using the proposed redundancy resolution criteria the error between the predicted swivel angle and the actual swivel angle adopted by the motor control system is less then 5 Deg. This criterion or a synthesis of several additional criteria may improve the synergistic relationships between an operator and a wearable robotic system.
Parametrization and Range of Motion of the Ball-and-Socket Joint The ball-and-socket joint model is used to represent articulations with three rotational degrees of free- dom (DOF), such as the human shoulder and the hip. The goal of this paper is to discuss two related prob- lems: the parametrization and the definition of realistic joint boundaries for ball-and-socket joints. Doing this accurately is difficult, yet important for motion generators (such as inverse kinematics and dynamics engines) and for motion manipulators (such as motion retargeting), since the resulting motions should satisfy the anatomic constraints. The difficulty mainly comes from the complex nature of 3D orientations and of human articulations. The underlying question of parametrization must be addressed before realis- tic and meaningful boundaries can be defined over the set of 3D orientations. In this paper, we review and compare several known methods, and advocate the use of the swing-and-twist parametrization, that parti- tions an arbitrary orientation into two meaningful components. The related problem of induced twist is discussed. Finally, we review some joint boundaries representations based on this decomposition, and show an example.
Positional kinematics of humanoid arms We present the positional abilities of a humanoid manipulator based on an improved kinematical model of the human arm. This was synthesized from electro-optical measurements of healthy female and male subjects. The model possesses three joints: inner shoulder joint, outer shoulder joint and elbow joint. The first functions as the human sternoclavicular joint, the second functions as the human glenohumeral joint, and the last replicates the human humeroulnar rotation. There are three links included, the forearm and the upper arm link which are of a constant length, and the shoulder link which is expandable. Mathematical interrelations between the joint coordinates are also taken into consideration. We determined the reachability of a humanoid arm, treated its orienting redundancy in the shoulder complex and the positional redundancy in the shoulder-elbow complexes, and discussed optimum configurations in executing different tasks. The results are important for the design and control of humanoid robots, in medicine and sports.
Design of a Bio-Inspired Wearable Exoskeleton for Applications in Robotics In this paper we explain the methodology we adopted to design the kinematics structure of a multi-contact points haptic interface. We based our concept on the analysis of the human arm anatomy and kinematics with the intend to synthesize a system that will be able to interface with the human limb in a very natural way. We proposed a simplified kinematic model of the human arm using a notation coming from the robotics field. To find out the best kinematics architecture we employed real movement data, measured from a human subject, and integrated them with the kinematic model of the exoskeleton, this allow us to test the system before its construction and to formalize specific requirements. We also implemented and tested a first passive version of the shoulder joint.
A Minimal Set Of Coordinates For Describing Humanoid Shoulder Motion The kinematics of the anatomical shoulder are analysed and modelled as a parallel mechanism similar to a Stewart platform. A new method is proposed to describe the shoulder kinematics with minimal coordinates and solve the indeterminacy. The minimal coordinates are defined from bony landmarks and the scapulothoracic kinematic constraints. Independent from one another, they uniquely characterise the shoulder motion. A humanoid mechanism is then proposed with identical kinematic properties. It is then shown how minimal coordinates can be obtained for this mechanism and how the coordinates simplify both the motion-planning task and trajectory-tracking control. Lastly, the coordinates are also shown to have an application in the field of biomechanics where they can be used to model the scapulohumeral rhythm.
Elbow Musculoskeletal Model for Industrial Exoskeleton with Modulated Impedance Based on Operator's Arm Stiffness.
Minimum acceleration criterion with constraints implies bang-bang control as an underlying principle for optimal trajectories of arm reaching movements. Rapid arm-reaching movements serve as an excellent test bed for any theory about trajectory formation. How are these movements planned? A minimum acceleration criterion has been examined in the past, and the solution obtained, based on the Euler-Poisson equation, failed to predict that the hand would begin and end the movement at rest (i.e., with zero acceleration). Therefore, this criterion was rejected in favor of the minimum jerk, which was proved to be successful in describing many features of human movements. This letter follows an alternative approach and solves the minimum acceleration problem with constraints using Pontryagin's minimum principle. We use the minimum principle to obtain minimum acceleration trajectories and use the jerk as a control signal. In order to find a solution that does not include nonphysiological impulse functions, constraints on the maximum and minimum jerk values are assumed. The analytical solution provides a three-phase piecewise constant jerk signal (bang-bang control) where the magnitude of the jerk and the two switching times depend on the magnitude of the maximum and minimum available jerk values. This result fits the observed trajectories of reaching movements and takes into account both the extrinsic coordinates and the muscle limitations in a single framework. The minimum acceleration with constraints principle is discussed as a unifying approach for many observations about the neural control of movements.
Switching between stabilizing controllers This paper deals with the problem of switching between several linear time-invariant (LTI) controllers—all of them capable of stabilizing a speci4c LTI process—in such a way that the stability of the closed-loop system is guaranteed for any switching sequence. We show that it is possible to 4nd realizations for any given family of controller transfer matrices so that the closed-loop system remains stable, no matter how we switch among the controller. The motivation for this problem is the control of complex systems where con8icting requirements make a single LTI controller unsuitable. ? 2002 Published by Elsevier Science Ltd.
A Privacy-Preserving and Copy-Deterrence Content-Based Image Retrieval Scheme in Cloud Computing. With the increasing importance of images in people’s daily life, content-based image retrieval (CBIR) has been widely studied. Compared with text documents, images consume much more storage space. Hence, its maintenance is considered to be a typical example for cloud storage outsourcing. For privacy-preserving purposes, sensitive images, such as medical and personal images, need to be encrypted before outsourcing, which makes the CBIR technologies in plaintext domain to be unusable. In this paper, we propose a scheme that supports CBIR over encrypted images without leaking the sensitive information to the cloud server. First, feature vectors are extracted to represent the corresponding images. After that, the pre-filter tables are constructed by locality-sensitive hashing to increase search efficiency. Moreover, the feature vectors are protected by the secure kNN algorithm, and image pixels are encrypted by a standard stream cipher. In addition, considering the case that the authorized query users may illegally copy and distribute the retrieved images to someone unauthorized, we propose a watermark-based protocol to deter such illegal distributions. In our watermark-based protocol, a unique watermark is directly embedded into the encrypted images by the cloud server before images are sent to the query user. Hence, when image copy is found, the unlawful query user who distributed the image can be traced by the watermark extraction. The security analysis and the experiments show the security and efficiency of the proposed scheme.
Firefly algorithm, stochastic test functions and design optimisation Modern optimisation algorithms are often metaheuristic, and they are very promising in solving NP-hard optimisation problems. In this paper, we show how to use the recently developed firefly algorithm to solve non-linear design problems. For the standard pressure vessel design optimisation, the optimal solution found by FA is far better than the best solution obtained previously in the literature. In addition, we also propose a few new test functions with either singularity or stochastic components but with known global optimality and thus they can be used to validate new optimisation algorithms. Possible topics for further research are also discussed.
Multi-stage genetic programming: A new strategy to nonlinear system modeling This paper presents a new multi-stage genetic programming (MSGP) strategy for modeling nonlinear systems. The proposed strategy is based on incorporating the individual effect of predictor variables and the interactions among them to provide more accurate simulations. According to the MSGP strategy, an efficient formulation for a problem comprises different terms. In the first stage of the MSGP-based analysis, the output variable is formulated in terms of an influencing variable. Thereafter, the error between the actual and the predicted value is formulated in terms of a new variable. Finally, the interaction term is derived by formulating the difference between the actual values and the values predicted by the individually developed terms. The capabilities of MSGP are illustrated by applying it to the formulation of different complex engineering problems. The problems analyzed herein include the following: (i) simulation of pH neutralization process, (ii) prediction of surface roughness in end milling, and (iii) classification of soil liquefaction conditions. The validity of the proposed strategy is confirmed by applying the derived models to the parts of the experimental results that were not included in the analyses. Further, the external validation of the models is verified using several statistical criteria recommended by other researchers. The MSGP-based solutions are capable of effectively simulating the nonlinear behavior of the investigated systems. The results of MSGP are found to be more accurate than those of standard GP and artificial neural network-based models.
Finite-approximation-error-based discrete-time iterative adaptive dynamic programming. In this paper, a new iterative adaptive dynamic programming (ADP) algorithm is developed to solve optimal control problems for infinite horizon discrete-time nonlinear systems with finite approximation errors. First, a new generalized value iteration algorithm of ADP is developed to make the iterative performance index function converge to the solution of the Hamilton-Jacobi-Bellman equation. The ...
Scalable and Privacy-Preserving Data Sharing Based on Blockchain. With the development of network technology and cloud computing, data sharing is becoming increasingly popular, and many scholars have conducted in-depth research to promote its flourish. As the scale of data sharing expands, its privacy protection has become a hot issue in research. Moreover, in data sharing, the data is usually maintained in multiple parties, which brings new challenges to protect the privacy of these multi-party data. In this paper, we propose a trusted data sharing scheme using blockchain. We use blockchain to prevent the shared data from being tampered, and use the Paillier cryptosystem to realize the confidentiality of the shared data. In the proposed scheme, the shared data can be traded, and the transaction information is protected by using the (p, t)-threshold Paillier cryptosystem. We conduct experiments in cloud storage scenarios and the experimental results demonstrate the efficiency and effectiveness of the proposed scheme.
Robot tutor and pupils’ educational ability: Teaching the times tables Research shows promising results of educational robots in language and STEM tasks. In language, more research is available, occasionally in view of individual differences in pupils’ educational ability levels, and learning seems to improve with more expressive robot behaviors. In STEM, variations in robots’ behaviors have been examined with inconclusive results and never while systematically investigating how differences in educational abilities match with different robot behaviors. We applied an autonomously tutoring robot (without tablet, partly WOz) in a 2 × 2 experiment of social vs. neutral behavior in above-average vs. below-average schoolchildren (N = 86; age 8–10 years) while rehearsing the multiplication tables on a one-to-one basis. The standard school test showed that on average, pupils significantly improved their performance even after 3 occasions of 5-min exercises. Beyond-average pupils profited most from a robot tutor, whereas those below average in multiplication benefited more from a robot that showed neutral rather than more social behavior.
1.084489
0.08
0.08
0.08
0.08
0.08
0.0044
0
0
0
0
0
0
0
Recommendation Effects of a Social Robot for Advertisement-Use Context in a Shopping Mall. We developed a coupon-giving robot system for a shopping mall to explore possible applications using social robots in daily environments, particularly for advertising. The system provided information through conversations with people. The robot was semi-autonomous, which means that it was partly controlled by a human operator, to cope with the difficulty of speech recognition in real environments. We conducted two field trials to investigate two kinds of effectiveness related to recommendations: the presence of a robot and different conversation schemas. Although a robot can strongly attract people with its presence and interaction, it remains unknown whether it can increase the effects of advertisements in real environments. Our field trial results show that a small robot increased the number of people who printed coupons more than a normal-sized robot. The number of people who printed coupons also increased when the robot asked visitors to freely select from all coupon candidates or to listen to its recommendation.
Spatial augmented reality as a method for a mobile robot to communicate intended movement. •Communication strategies are to allow robots to convey upcoming movements to humans.•Arrows for conveying direction of movement are understood by humans.•Simple maps depicting a sequence of upcoming movements are useful to humans.•Robots projecting arrows and a map can effectively communicate upcoming movement.
From Proxemics Theory to Socially-Aware Navigation: A Survey. In the context of a growing interest in modelling human behavior to increase the robots’ social abilities, this article presents a survey related to socially-aware robot navigation. It presents a review from sociological concepts to social robotics and human-aware navigation. Social cues, signals and proxemics are discussed. Socially aware behavior in terms of navigation is tackled also. Finally, recent robotic experiments focusing on the way social conventions and robotics must be linked is presented.
Human Mobility Modeling for Robot-Assisted Evacuation in Complex Indoor Environments. A large number of injuries or deaths may occur when an emergency happens in a crowded public place. The congestion at exits may slow down the egress rate due to the effect of “faster-is-slower”. This inspires us to study how human behavior dynamically changes over time at an emergency in a complex indoor environment. In this paper, we refer the panic of evacuees to their perception of the threat a...
Human-Like Guide Robot that Proactively Explains Exhibits We developed an autonomous human-like guide robot for a science museum. Its identifies individuals, estimates the exhibits at which visitors are looking, and proactively approaches them to provide explanations with gaze autonomously, using our new approach called speak-and-retreat interaction. The robot also performs such relation-building behaviors as greeting visitors by their names and expressing a friendlier attitude to repeat visitors. We conducted a field study in a science museum at which our system basically operated autonomously and the visitors responded quite positively. First-time visitors on average interacted with the robot for about 9 min, and 94.74% expressed a desire to interact with it again in the future. Repeat visitors noticed its relation-building capability and perceived a closer relationship with it.
Computational thinking Summary form only given. My vision for the 21st century, Computational Thinking, will be a fundamental skill used by everyone in the world. To reading, writing, and arithmetic, we should add computational thinking to every child's analytical ability. Computational thinking involves solving problems, designing systems, and understanding human behavior by drawing on the concepts fundamental to computer science. Thinking like a computer scientist means more than being able to program a computer. It requires the ability to abstract and thus to think at multiple levels of abstraction. In this talk I will give many examples of computational thinking, argue that it has already influenced other disciplines, and promote the idea that teaching computational thinking can not only inspire future generations to enter the field of computer science but benefit people in all fields.
JPEG Error Analysis and Its Applications to Digital Image Forensics JPEG is one of the most extensively used image formats. Understanding the inherent characteristics of JPEG may play a useful role in digital image forensics. In this paper, we introduce JPEG error analysis to the study of image forensics. The main errors of JPEG include quantization, rounding, and truncation errors. Through theoretically analyzing the effects of these errors on single and double JPEG compression, we have developed three novel schemes for image forensics including identifying whether a bitmap image has previously been JPEG compressed, estimating the quantization steps of a JPEG image, and detecting the quantization table of a JPEG image. Extensive experimental results show that our new methods significantly outperform existing techniques especially for the images of small sizes. We also show that the new method can reliably detect JPEG image blocks which are as small as 8 × 8 pixels and compressed with quality factors as high as 98. This performance is important for analyzing and locating small tampered regions within a composite image.
Highly dynamic Destination-Sequenced Distance-Vector routing (DSDV) for mobile computers An ad-hoc network is the cooperative engagement of a collection of Mobile Hosts without the required intervention of any centralized Access Point. In this paper we present an innovative design for the operation of such ad-hoc networks. The basic idea of the design is to operate each Mobile Host as a specialized router, which periodically advertises its view of the interconnection topology with other Mobile Hosts within the network. This amounts to a new sort of routing protocol. We have investigated modifications to the basic Bellman-Ford routing mechanisms, as specified by RIP [5], to make it suitable for a dynamic and self-starting network mechanism as is required by users wishing to utilize ad hoc networks. Our modifications address some of the previous objections to the use of Bellman-Ford, related to the poor looping properties of such algorithms in the face of broken links and the resulting time dependent nature of the interconnection topology describing the links between the Mobile Hosts. Finally, we describe the ways in which the basic network-layer routing can be modified to provide MAC-layer support for ad-hoc networks.
The FERET Evaluation Methodology for Face-Recognition Algorithms Two of the most critical requirements in support of producing reliable face-recognition systems are a large database of facial images and a testing procedure to evaluate systems. The Face Recognition Technology (FERET) program has addressed both issues through the FERET database of facial images and the establishment of the FERET tests. To date, 14,126 images from 1,199 individuals are included in the FERET database, which is divided into development and sequestered portions of the database. In September 1996, the FERET program administered the third in a series of FERET face-recognition tests. The primary objectives of the third test were to 1) assess the state of the art, 2) identify future areas of research, and 3) measure algorithm performance.
Communication in reactive multiagent robotic systems Multiple cooperating robots are able to complete many tasks more quickly and reliably than one robot alone. Communication between the robots can multiply their capabilities and effectiveness, but to what extent? In this research, the importance of communication in robotic societies is investigated through experiments on both simulated and real robots. Performance was measured for three different types of communication for three different tasks. The levels of communication are progressively more complex and potentially more expensive to implement. For some tasks, communication can significantly improve performance, but for others inter-agent communication is apparently unnecessary. In cases where communication helps, the lowest level of communication is almost as effective as the more complex type. The bulk of these results are derived from thousands of simulations run with randomly generated initial conditions. The simulation results help determine appropriate parameters for the reactive control system which was ported for tests on Denning mobile robots.
Lower Extremity Exoskeletons and Active Orthoses: Challenges and State-of-the-Art In the nearly six decades since researchers began to explore methods of creating them, exoskeletons have progressed from the stuff of science fiction to nearly commercialized products. While there are still many challenges associated with exoskeleton development that have yet to be perfected, the advances in the field have been enormous. In this paper, we review the history and discuss the state-of-the-art of lower limb exoskeletons and active orthoses. We provide a design overview of hardware, actuation, sensory, and control systems for most of the devices that have been described in the literature, and end with a discussion of the major advances that have been made and hurdles yet to be overcome.
A Model Predictive Control Approach to Microgrid Operation Optimization. Microgrids are subsystems of the distribution grid, which comprises generation capacities, storage devices, and controllable loads, operating as a single controllable system either connected or isolated from the utility grid. In this paper, we present a study on applying a model predictive control approach to the problem of efficiently optimizing microgrid operations while satisfying a time-varying request and operation constraints. The overall problem is formulated using mixed-integer linear programming (MILP), which can be solved in an efficient way by using commercial solvers without resorting to complex heuristics or decompositions techniques. Then, the MILP formulation leads to significant improvements in solution quality and computational burden. A case study of a microgrid is employed to assess the performance of the online optimization-based control strategy and the simulation results are discussed. The method is applied to an experimental microgrid located in Athens, Greece. The experimental results show the feasibility and the effectiveness of the proposed approach.
Convolutional Neural Network-Based Classification of Driver's Emotion during Aggressive and Smooth Driving Using Multi-Modal Camera Sensors. Because aggressive driving often causes large-scale loss of life and property, techniques for advance detection of adverse driver emotional states have become important for the prevention of aggressive driving behaviors. Previous studies have primarily focused on systems for detecting aggressive driver emotion via smart-phone accelerometers and gyro-sensors, or they focused on methods of detecting physiological signals using electroencephalography (EEG) or electrocardiogram (ECG) sensors. Because EEG and ECG sensors cause discomfort to drivers and can be detached from the driver's body, it becomes difficult to focus on bio-signals to determine their emotional state. Gyro-sensors and accelerometers depend on the performance of GPS receivers and cannot be used in areas where GPS signals are blocked. Moreover, if driving on a mountain road with many quick turns, a driver's emotional state can easily be misrecognized as that of an aggressive driver. To resolve these problems, we propose a convolutional neural network (CNN)-based method of detecting emotion to identify aggressive driving using input images of the driver's face, obtained using near-infrared (NIR) light and thermal camera sensors. In this research, we conducted an experiment using our own database, which provides a high classification accuracy for detecting driver emotion leading to either aggressive or smooth (i.e., relaxed) driving. Our proposed method demonstrates better performance than existing methods.
Ethical Considerations Of Applying Robots In Kindergarten Settings: Towards An Approach From A Macroperspective In child-robot interaction (cHRI) research, many studies pursue the goal to develop interactive systems that can be applied in everyday settings. For early education, increasingly, the setting of a kindergarten is targeted. However, when cHRI and research are brought into a kindergarten, a range of ethical and related procedural aspects have to be considered and dealt with. While ethical models elaborated within other human-robot interaction settings, e.g., assisted living contexts, can provide some important indicators for relevant issues, we argue that it is important to start developing a systematic approach to identify and tackle those ethical issues which rise with cHRI in kindergarten settings on a more global level and address the impact of the technology from a macroperspective beyond the effects on the individual. Based on our experience in conducting studies with children in general and pedagogical considerations on the role of the institution of kindergarten in specific, in this paper, we enfold some relevant aspects that have barely been addressed in an explicit way in current cHRI research. Four areas are analyzed and key ethical issues are identified in each area: (1) the institutional setting of a kindergarten, (2) children as a vulnerable group, (3) the caregivers' role, and (4) pedagogical concepts. With our considerations, we aim at (i) broadening the methodology of the current studies within the area of cHRI, (ii) revalidate it based on our comprehensive empirical experience with research in kindergarten settings, both laboratory and real-world contexts, and (iii) provide a framework for the development of a more systematic approach to address the ethical issues in cHRI research within kindergarten settings.
1.24
0.24
0.24
0.24
0.12
0
0
0
0
0
0
0
0
0
Intermittent Estimator-Based Mixed Passive and <italic>H<sub>∞</sub></italic> Control for High-Speed Train With Actuator Stochastic Fault This article is concerned with the intermittent estimator-based mixed passive and <inline-formula xmlns:mml="http://www.w3.org/1998/Math/MathML" xmlns:xlink="http://www.w3.org/1999/xlink"> <tex-math notation="LaTeX">$\mathscr {H}_{\infty }$ </tex-math></inline-formula> control for the high-speed train (HST) with multiple noises, actuator stochastic fault, and sensor packet loss. First, an intermittent estimator is designed to track the undetectable status of HSTs in response to only partial information available due to sensor failures. Then, two different stability criteria are developed by adopting two different Lyapunov function strategies. Simultaneously, in order to reduce the control cost and accelerate the convergence time, two different algorithms are designed. It is worth emphasizing that different from the existing results of HST subject to actuator fault, this article adopts a more flexible fault representation mode, namely, semi-Markov switching mode, which is more in line with the practical background and has a higher valuable application. Especially, the Lyapunov function designed in this article can drive the system state to decrease monotonically in both the “working interval” and the “rest interval,” so as to avoid the phenomenon of state impulsive jump. Finally, through the test of HST experimental value of Japan’s Shinkansen, the simulation results show the effectiveness and rationality of the proposed control method and also make a comparative analysis with related works, to prove the advantages of the control technology proposed in this article.
The Sybil Attack Large-scale peer-to-peer systems facesecurity threats from faulty or hostile remotecomputing elements. To resist these threats, manysuch systems employ redundancy. However, if asingle faulty entity can present multiple identities,it can control a substantial fraction of the system,thereby undermining this redundancy. Oneapproach to preventing these &quot;Sybil attacks&quot; is tohave a trusted agency certify identities. Thispaper shows that, without a logically centralizedauthority, Sybil...
BLEU: a method for automatic evaluation of machine translation Human evaluations of machine translation are extensive but expensive. Human evaluations can take months to finish and involve human labor that can not be reused. We propose a method of automatic machine translation evaluation that is quick, inexpensive, and language-independent, that correlates highly with human evaluation, and that has little marginal cost per run. We present this method as an automated understudy to skilled human judges which substitutes for them when there is need for quick or frequent evaluations.
Computational thinking Summary form only given. My vision for the 21st century, Computational Thinking, will be a fundamental skill used by everyone in the world. To reading, writing, and arithmetic, we should add computational thinking to every child's analytical ability. Computational thinking involves solving problems, designing systems, and understanding human behavior by drawing on the concepts fundamental to computer science. Thinking like a computer scientist means more than being able to program a computer. It requires the ability to abstract and thus to think at multiple levels of abstraction. In this talk I will give many examples of computational thinking, argue that it has already influenced other disciplines, and promote the idea that teaching computational thinking can not only inspire future generations to enter the field of computer science but benefit people in all fields.
Fuzzy logic in control systems: fuzzy logic controller. I.
Switching between stabilizing controllers This paper deals with the problem of switching between several linear time-invariant (LTI) controllers—all of them capable of stabilizing a speci4c LTI process—in such a way that the stability of the closed-loop system is guaranteed for any switching sequence. We show that it is possible to 4nd realizations for any given family of controller transfer matrices so that the closed-loop system remains stable, no matter how we switch among the controller. The motivation for this problem is the control of complex systems where con8icting requirements make a single LTI controller unsuitable. ? 2002 Published by Elsevier Science Ltd.
Tabu Search - Part I
Bidirectional recurrent neural networks In the first part of this paper, a regular recurrent neural network (RNN) is extended to a bidirectional recurrent neural network (BRNN). The BRNN can be trained without the limitation of using input information just up to a preset future frame. This is accomplished by training it simultaneously in positive and negative time direction. Structure and training procedure of the proposed network are explained. In regression and classification experiments on artificial data, the proposed structure gives better results than other approaches. For real data, classification experiments for phonemes from the TIMIT database show the same tendency. In the second part of this paper, it is shown how the proposed bidirectional structure can be easily modified to allow efficient estimation of the conditional posterior probability of complete symbol sequences without making any explicit assumption about the shape of the distribution. For this part, experiments on real data are reported
An intensive survey of fair non-repudiation protocols With the phenomenal growth of the Internet and open networks in general, security services, such as non-repudiation, become crucial to many applications. Non-repudiation services must ensure that when Alice sends some information to Bob over a network, neither Alice nor Bob can deny having participated in a part or the whole of this communication. Therefore a fair non-repudiation protocol has to generate non-repudiation of origin evidences intended to Bob, and non-repudiation of receipt evidences destined to Alice. In this paper, we clearly define the properties a fair non-repudiation protocol must respect, and give a survey of the most important non-repudiation protocols without and with trusted third party (TTP). For the later ones we discuss the evolution of the TTP's involvement and, between others, describe the most recent protocol using a transparent TTP. We also discuss some ad-hoc problems related to the management of non-repudiation evidences.
Dynamic movement and positioning of embodied agents in multiparty conversations For embodied agents to engage in realistic multiparty conversation, they must stand in appropriate places with respect to other agents and the environment. When these factors change, such as an agent joining the conversation, the agents must dynamically move to a new location and/or orientation to accommodate. This paper presents an algorithm for simulating movement of agents based on observed human behavior using techniques developed for pedestrian movement in crowd simulations. We extend a previous group conversation simulation to include an agent motion algorithm. We examine several test cases and show how the simulation generates results that mirror real-life conversation settings.
An improved genetic algorithm with conditional genetic operators and its application to set-covering problem The genetic algorithm (GA) is a popular, biologically inspired optimization method. However, in the GA there is no rule of thumb to design the GA operators and select GA parameters. Instead, trial-and-error has to be applied. In this paper we present an improved genetic algorithm in which crossover and mutation are performed conditionally instead of probability. Because there are no crossover rate and mutation rate to be selected, the proposed improved GA can be more easily applied to a problem than the conventional genetic algorithms. The proposed improved genetic algorithm is applied to solve the set-covering problem. Experimental studies show that the improved GA produces better results over the conventional one and other methods.
Lane-level traffic estimations using microscopic traffic variables This paper proposes a novel inference method to estimate lane-level traffic flow, time occupancy and vehicle inter-arrival time on road segments where local information could not be measured and assessed directly. The main contributions of the proposed method are 1) the ability to perform lane-level estimations of traffic flow, time occupancy and vehicle inter-arrival time and 2) the ability to adapt to different traffic regimes by assessing only microscopic traffic variables. We propose a modified Kriging estimation model which explicitly takes into account both spatial and temporal variability. Performance evaluations are conducted using real-world data under different traffic regimes and it is shown that the proposed method outperforms a Kalman filter-based approach.
Convolutional Neural Network-Based Classification of Driver's Emotion during Aggressive and Smooth Driving Using Multi-Modal Camera Sensors. Because aggressive driving often causes large-scale loss of life and property, techniques for advance detection of adverse driver emotional states have become important for the prevention of aggressive driving behaviors. Previous studies have primarily focused on systems for detecting aggressive driver emotion via smart-phone accelerometers and gyro-sensors, or they focused on methods of detecting physiological signals using electroencephalography (EEG) or electrocardiogram (ECG) sensors. Because EEG and ECG sensors cause discomfort to drivers and can be detached from the driver's body, it becomes difficult to focus on bio-signals to determine their emotional state. Gyro-sensors and accelerometers depend on the performance of GPS receivers and cannot be used in areas where GPS signals are blocked. Moreover, if driving on a mountain road with many quick turns, a driver's emotional state can easily be misrecognized as that of an aggressive driver. To resolve these problems, we propose a convolutional neural network (CNN)-based method of detecting emotion to identify aggressive driving using input images of the driver's face, obtained using near-infrared (NIR) light and thermal camera sensors. In this research, we conducted an experiment using our own database, which provides a high classification accuracy for detecting driver emotion leading to either aggressive or smooth (i.e., relaxed) driving. Our proposed method demonstrates better performance than existing methods.
Robot tutor and pupils’ educational ability: Teaching the times tables Research shows promising results of educational robots in language and STEM tasks. In language, more research is available, occasionally in view of individual differences in pupils’ educational ability levels, and learning seems to improve with more expressive robot behaviors. In STEM, variations in robots’ behaviors have been examined with inconclusive results and never while systematically investigating how differences in educational abilities match with different robot behaviors. We applied an autonomously tutoring robot (without tablet, partly WOz) in a 2 × 2 experiment of social vs. neutral behavior in above-average vs. below-average schoolchildren (N = 86; age 8–10 years) while rehearsing the multiplication tables on a one-to-one basis. The standard school test showed that on average, pupils significantly improved their performance even after 3 occasions of 5-min exercises. Beyond-average pupils profited most from a robot tutor, whereas those below average in multiplication benefited more from a robot that showed neutral rather than more social behavior.
1
0
0
0
0
0
0
0
0
0
0
0
0
0
An Efficient Cad System For All Cell Identification From Microscopic Blood Images Computer-aided diagnosis (CAD) becomes a common tool for identifying diseases, especially various cancers, from medical images. Thus, digital image processing plays a significant role in this research area. This paper concerns with developing an efficient automatic system for the identification of acute lymphoblastic leukemia (ALL) cells. The proposed approach involves two steps. The first step focuses on segmenting the white blood cells (WBCs). In the second step, significant features such as shape, geometrical, statistical, and discrete cosine transform (DCT) are extracted from the segmented cells. Various classification techniques are applied to the extracted features to classify the segmented cells into normal and abnormal cells. The performance of the proposed approach has been evaluated via extensive experiments conducted on the well-known ALL-IDB dataset of microscopic images of blood. The experimental results demonstrate that the proposed approach realizes an accuracy rate 97.45% and outperforms other existing approaches.
Geometric attacks on image watermarking systems Synchronization errors can lead to significant performance loss in image watermarking methods, as the geometric attacks in the Stirmark benchmark software show. The authors describe the most common types of geometric attacks and survey proposed solutions.
Genetic Optimization Of Radial Basis Probabilistic Neural Networks This paper discusses using genetic algorithms (CA) to optimize the structure of radial basis probabilistic neural networks (RBPNN), including how to select hidden centers of the first hidden layer and to determine the controlling parameter of Gaussian kernel functions. In the process of constructing the genetic algorithm, a novel encoding method is proposed for optimizing the RBPNN structure. This encoding method can not only make the selected hidden centers sufficiently reflect the key distribution characteristic in the space of training samples set and reduce the hidden centers number as few as possible, but also simultaneously determine the optimum controlling parameters of Gaussian kernel functions matching the selected hidden centers. Additionally, we also constructively propose a new fitness function so as to make the designed RBPNN as simple as possible in the network structure in the case of not losing the network performance. Finally, we take the two benchmark problems of discriminating two-spiral problem and classifying the iris data, for example, to test and evaluate this designed GA. The experimental results illustrate that our designed CA can significantly reduce the required hidden centers number, compared with the recursive orthogonal least square algorithm (ROLSA) and the modified K-means algorithm (MKA). In particular, by means of statistical experiments it was proved that the optimized RBPNN by our designed GA, have still a better generalization performance with respect to the ones by the ROLSA and the MKA, in spite of the network scale having been greatly reduced. Additionally, our experimental results also demonstrate that our designed CA is also suitable for optimizing the radial basis function neural networks (RBFNN).
Current status and key issues in image steganography: A survey. Steganography and steganalysis are the prominent research fields in information hiding paradigm. Steganography is the science of invisible communication while steganalysis is the detection of steganography. Steganography means “covered writing” that hides the existence of the message itself. Digital steganography provides potential for private and secure communication that has become the necessity of most of the applications in today’s world. Various multimedia carriers such as audio, text, video, image can act as cover media to carry secret information. In this paper, we have focused only on image steganography. This article provides a review of fundamental concepts, evaluation measures and security aspects of steganography system, various spatial and transform domain embedding schemes. In addition, image quality metrics that can be used for evaluation of stego images and cover selection measures that provide additional security to embedding scheme are also highlighted. Current research trends and directions to improve on existing methods are suggested.
Hybrid local and global descriptor enhanced with colour information. Feature extraction is one of the most important steps in computer vision tasks such as object recognition, image retrieval and image classification. It describes an image by a set of descriptors where the best one gives a high quality description and a low computation. In this study, the authors propose a novel descriptor called histogram of local and global features using speeded up robust featur...
Secure visual cryptography for medical image using modified cuckoo search. Optimal secure visual cryptography for brain MRI medical image is proposed in this paper. Initially, the brain MRI images are selected and then discrete wavelet transform is applied to the brain MRI image for partitioning the image into blocks. Then Gaussian based cuckoo search algorithm is utilized to select the optimal position for every block. Next the proposed technique creates the dual shares from the secret image. Then the secret shares are embedded in the corresponding positions of the blocks. After embedding, the extraction operation is carried out. Here visual cryptographic design is used for the purpose of image authentication and verification. The extracted secret image has dual shares, based on that the receiver views the input image. The authentication and verification of medical image are assisted with the help of target database. All the secret images are registered previously in the target database. The performance of the proposed method is estimated by Peak Signal to Noise Ratio (PSNR), Mean square error (MSE) and normalized correlation. The implementation is done by MATLAB platform.
Digital watermarking techniques for image security: a review Multimedia technology usages is increasing day by day and to provide authorized data and protecting the secret information from unauthorized use is highly difficult and involves a complex process. By using the watermarking technique, only authorized user can use the data. Digital watermarking is a widely used technology for the protection of digital data. Digital watermarking deals with the embedding of secret data into actual information. Digital watermarking techniques are classified into three major categories, and they were based on domain, type of document (text, image, music or video) and human perception. Performance of the watermarked images is analysed using Peak signal to noise ratio, mean square error and bit error rate. Watermarking of images has been researched profoundly for its specialized and modern achievability in all media applications such as copyrights protection, medical reports (MRI scan and X-ray), annotation and privacy control. This paper reviews the watermarking technique and its merits and demerits.
A New Efficient Medical Image Cipher Based On Hybrid Chaotic Map And Dna Code In this paper, we propose a novel medical image encryption algorithm based on a hybrid model of deoxyribonucleic acid (DNA) masking, a Secure Hash Algorithm SHA-2 and a new hybrid chaotic map. Our study uses DNA sequences and operations and the chaotic hybrid map to strengthen the cryptosystem. The significant advantages of this approach consist in improving the information entropy which is the most important feature of randomness, resisting against various typical attacks and getting good experimental results. The theoretical analysis and experimental results show that the algorithm improves the encoding efficiency, enhances the security of the ciphertext, has a large key space and a high key sensitivity, and is able to resist against the statistical and exhaustive attacks.
On Multi-Access Edge Computing: A Survey of the Emerging 5G Network Edge Cloud Architecture and Orchestration. Multi-access edge computing (MEC) is an emerging ecosystem, which aims at converging telecommunication and IT services, providing a cloud computing platform at the edge of the radio access network. MEC offers storage and computational resources at the edge, reducing latency for mobile end users and utilizing more efficiently the mobile backhaul and core networks. This paper introduces a survey on ...
An effective implementation of the Lin–Kernighan traveling salesman heuristic This paper describes an implementation of the Lin–Kernighan heuristic, one of the most successful methods for generating optimal or near-optimal solutions for the symmetric traveling salesman problem (TSP). Computational tests show that the implementation is highly effective. It has found optimal solutions for all solved problem instances we have been able to obtain, including a 13,509-city problem (the largest non-trivial problem instance solved to optimality today).
Exoskeletons for human power augmentation The first load-bearing and energetically autonomous exoskeleton, called the Berkeley Lower Extremity Exoskeleton (BLEEX) walks at the average speed of two miles per hour while carrying 75 pounds of load. The project, funded in 2000 by the Defense Advanced Research Project Agency (DARPA) tackled four fundamental technologies: the exoskeleton architectural design, a control algorithm, a body LAN to host the control algorithm, and an on-board power unit to power the actuators, sensors and the computers. This article gives an overview of the BLEEX project.
Assist-As-Needed Training Paradigms For Robotic Rehabilitation Of Spinal Cord Injuries This paper introduces a new "assist-as-needed" (AAN) training paradigm for rehabilitation of spinal cord injuries via robotic training devices. In the pilot study reported in this paper, nine female adult Swiss-Webster mice were divided into three groups, each experiencing a different robotic training control strategy: a fixed training trajectory (Fixed Group, A), an AAN training method without interlimb coordination (Band Group, B), and an AAN training method with bilateral hindlimb coordination (Window Group, C). Fourteen days after complete transection at the mid-thoracic level, the mice were robotically trained to step in the presence of an acutely administered serotonin agonist, quipazine, for a period of six weeks. The mice that received AAN training (Groups B and C) show higher levels of recovery than Group A mice, as measured by the number, consistency, and periodicity of steps realized during testing sessions. Group C displays a higher incidence of alternating stepping than Group B. These results indicate that this training approach may be more effective than fixed trajectory paradigms in promoting robust post-injury stepping behavior. Furthermore, the constraint of interlimb coordination appears to be an important contribution to successful training.
An ID-Based Linearly Homomorphic Signature Scheme and Its Application in Blockchain. Identity-based cryptosystems mean that public keys can be directly derived from user identifiers, such as telephone numbers, email addresses, and social insurance number, and so on. So they can simplify key management procedures of certificate-based public key infrastructures and can be used to realize authentication in blockchain. Linearly homomorphic signature schemes allow to perform linear computations on authenticated data. And the correctness of the computation can be publicly verified. Although a series of homomorphic signature schemes have been designed recently, there are few homomorphic signature schemes designed in identity-based cryptography. In this paper, we construct a new ID-based linear homomorphic signature scheme, which avoids the shortcomings of the use of public-key certificates. The scheme is proved secure against existential forgery on adaptively chosen message and ID attack under the random oracle model. The ID-based linearly homomorphic signature schemes can be applied in e-business and cloud computing. Finally, we show how to apply it to realize authentication in blockchain.
Robot tutor and pupils’ educational ability: Teaching the times tables Research shows promising results of educational robots in language and STEM tasks. In language, more research is available, occasionally in view of individual differences in pupils’ educational ability levels, and learning seems to improve with more expressive robot behaviors. In STEM, variations in robots’ behaviors have been examined with inconclusive results and never while systematically investigating how differences in educational abilities match with different robot behaviors. We applied an autonomously tutoring robot (without tablet, partly WOz) in a 2 × 2 experiment of social vs. neutral behavior in above-average vs. below-average schoolchildren (N = 86; age 8–10 years) while rehearsing the multiplication tables on a one-to-one basis. The standard school test showed that on average, pupils significantly improved their performance even after 3 occasions of 5-min exercises. Beyond-average pupils profited most from a robot tutor, whereas those below average in multiplication benefited more from a robot that showed neutral rather than more social behavior.
1.2
0.2
0.2
0.2
0.2
0.2
0.2
0.05
0
0
0
0
0
0
Efficient LMI Conditions for Enhanced Stabilization of Discrete-Time Takagi-Sugeno Models via Delayed Nonquadratic Lyapunov Functions This paper is concerned with the reduction of conservatism on stabilization conditions of discrete-time Takagi–Sugeno fuzzy models via delayed nonquadratic Lyapunov functions. A fruitful approach recently appeared in the literature, which, despite its benefits, dramatically increases the number of linear matrix inequalities needed to synthesize a controller. The two sets of sufficient conditions hereby proposed tackle this numerical problem while reducing conservatism and increasing the stabilization domain of existing conditions in literature, as illustrated in several examples.
Controller design for TS models using delayed nonquadratic Lyapunov functions. In the last few years, nonquadratic Lyapunov functions have been more and more frequently used in the analysis and controller design for Takagi-Sugeno fuzzy models. In this paper, we developed relaxed conditions for controller design using nonquadratic Lyapunov functions and delayed controllers and give a general framework for the use of such Lyapunov functions. The two controller design methods developed in this framework outperform and generalize current state-of-the-art methods. The proposed methods are extended to robust and H∞ control and α -sample variation.
Stability analysis of T-S fuzzy control systems by using set theory This paper is concerned with the stability analysis for T-S fuzzy control systems. By exploiting the property of the structure of fuzzy inference engine, an equivalence relation on index set of the product of fuzzy rule weights is defined. Further, a new stability criterion is proposed by using the equivalence relation, and formulated into progressively less conservative sets of linear matrix inequalities. By using an extension of P´olya’s Theorem, the new criterion is proved to be with no conservatism for quadratic stability analysis of T-S fuzzy control systems with a product inference engine and any possible fuzzy membership functions. A numerical example is given to illustrate the effectiveness of the proposed method.
The Generalized TP Model Transformation for T-S Fuzzy Model Manipulation and Generalized Stability Verification. This paper integrates various ideas about the tensor product (TP) model transformation into one conceptual framework and formulates it in terms of the Takagi-Sugeno (T-S) fuzzy model manipulation and control design framework. Several new extensions of the TP model transformation are proposed, such as the quasi and “full,” compact and rank-reduced higher order singular-value-decomposition-based can...
Adaptive Fuzzy Observer-Based Fault Estimation for a Class of Nonlinear Stochastic Hybrid Systems This article studies the fault estimation problem for a class of continuous-time nonlinear Markovian jump systems with unmeasured states, unknown bounded sensor faults, and unknown nonlinearities simultaneously. In this article, a new adaptive fuzzy observer design scheme is developed, where the completely unknown nonlinear terms are approximated by adaptive fuzzy logic systems. By means of a nove...
Homogeneous polynomially nonquadratic stabilization of discrete-time TakagiSugeno systems via nonparallel distributed compensation law This paper considers stability of discrete-time nonlinear systems in TakagiSugeno (TS) form. This problem has been studied for more than 20 years with many sufficient conditions, and the asymptotically necessary and sufficient (ANS) conditions with respect to the common-quadratic Lyapunov, function, having being obtained. This paper considers general forms of homogeneous polynomially nonquadratic Lyapunov (HPNQL) function and homogeneous polynomially parameterized nonparallel distributed compensation (HPP-non-PDC) law. By generalization of the procedure based on Plyas theorem and techniques used for parameter-dependent linear matrix inequality (PD-LMI) which have been studied previously in different contexts, ANS stability conditions with respect to the general HPNQL function are obtained. © 2010 IEEE.
Grey Wolf Optimizer Algorithm-Based Tuning of Fuzzy Control Systems With Reduced Parametric Sensitivity. This paper proposes an innovative tuning approach for fuzzy control systems (CSs) with a reduced parametric sensitivity using the Grey Wolf Optimizer (GWO) algorithm. The CSs consist of servo system processes controlled by Takagi-Sugeno-Kang proportional-integral fuzzy controllers (TSK PI-FCs). The process models have second-order dynamics with an integral component, variable parameters, a saturat...
Adaptive Consensus Control for a Class of Nonlinear Multiagent Time-Delay Systems Using Neural Networks Because of the complicity of consensus control of nonlinear multiagent systems in state time-delay, most of previous works focused only on linear systems with input time-delay. An adaptive neural network (NN) consensus control method for a class of nonlinear multiagent systems with state time-delay is proposed in this paper. The approximation property of radial basis function neural networks (RBFNNs) is used to neutralize the uncertain nonlinear dynamics in agents. An appropriate Lyapunov-Krasovskii functional, which is obtained from the derivative of an appropriate Lyapunov function, is used to compensate the uncertainties of unknown time delays. It is proved that our proposed approach guarantees the convergence on the basis of Lyapunov stability theory. The simulation results of a nonlinear multiagent time-delay system and a multiple collaborative manipulators system show the effectiveness of the proposed consensus control algorithm.
Switching between stabilizing controllers This paper deals with the problem of switching between several linear time-invariant (LTI) controllers—all of them capable of stabilizing a speci4c LTI process—in such a way that the stability of the closed-loop system is guaranteed for any switching sequence. We show that it is possible to 4nd realizations for any given family of controller transfer matrices so that the closed-loop system remains stable, no matter how we switch among the controller. The motivation for this problem is the control of complex systems where con8icting requirements make a single LTI controller unsuitable. ? 2002 Published by Elsevier Science Ltd.
Using Social Psychology to Motivate Contributions to Online Communities Under-contribution is a problem for many online communities. Social psychology theories of social loafing and goal-setting can lead to mid-level design goals to address this problem. We tested design principles derived from these theories in four field experiments involving members of an online movie recommender community. In each of the experiments participated were given different explanations for the value of their contributions. As predicted by theory, individuals contributed when they were reminded of their uniqueness and when they were given specific and challenging goals. However, other predictions were disconfirmed. For example, in one experiment, participants given group goals contributed more than those given individual goals. The article ends with suggestions and challenges for mining design implications from social science theories.
Tabu search based multi-watermarks embedding algorithm with multiple description coding Digital watermarking is a useful solution for digital rights management systems, and it has been a popular research topic in the last decade. Most watermarking related literature focuses on how to resist deliberate attacks by applying benchmarks to watermarked media that assess the effectiveness of the watermarking algorithm. Only a few papers have concentrated on the error-resilient transmission of watermarked media. In this paper, we propose an innovative algorithm for vector quantization (VQ) based image watermarking, which is suitable for error-resilient transmission over noisy channels. By incorporating watermarking with multiple description coding (MDC), the scheme we propose to embed multiple watermarks can effectively overcome channel impairments while retaining the capability for copyright and ownership protection. In addition, we employ an optimization technique, called tabu search, to optimize both the watermarked image quality and the robustness of the extracted watermarks. We have obtained promising simulation results that demonstrate the utility and practicality of our algorithm. (C) 2011 Elsevier Inc. All rights reserved.
An efficient conditionally anonymous ring signature in the random oracle model A conditionally anonymous ring signature is an exception since the anonymity is conditional. Specifically, it allows an entity to confirm/refute the signature that he generated before. A group signature also shares the same property since a group manager can revoke a signer's anonymity using the trapdoor information. However, the special node (i.e., group manager) does not exist in the group in order to satisfy the ad hoc fashion. In this paper, we construct a new conditionally anonymous ring signature, in which the actual signer can be traced without the help of the group manager. The big advantage of the confirmation and disavowal protocols designed by us are non-interactive with constant costs while the known schemes suffer from the linear cost in terms of the ring size n or security parameter s.
The ApolloScape Dataset for Autonomous Driving Scene parsing aims to assign a class (semantic) label for each pixel in an image. It is a comprehensive analysis of an image. Given the rise of autonomous driving, pixel-accurate environmental perception is expected to be a key enabling technical piece. However, providing a large scale dataset for the design and evaluation of scene parsing algorithms, in particular for outdoor scenes, has been difficult. The per-pixel labelling process is prohibitively expensive, limiting the scale of existing ones. In this paper, we present a large-scale open dataset, ApolloScape, that consists of RGB videos and corresponding dense 3D point clouds. Comparing with existing datasets, our dataset has the following unique properties. The first is its scale, our initial release contains over 140K images - each with its per-pixel semantic mask, up to 1M is scheduled. The second is its complexity. Captured in various traffic conditions, the number of moving objects averages from tens to over one hundred (Figure 1). And the third is the 3D attribute, each image is tagged with high-accuracy pose information at cm accuracy and the static background point cloud has mm relative accuracy. We are able to label these many images by an interactive and efficient labelling pipeline that utilizes the high-quality 3D point cloud. Moreover, our dataset also contains different lane markings based on the lane colors and styles. We expect our new dataset can deeply benefit various autonomous driving related applications that include but not limited to 2D/3D scene understanding, localization, transfer learning, and driving simulation.
Social Robots for (Second) Language Learning in (Migrant) Primary School Children Especially these days, innovation and support from technology to relieve pressure in education is highly urgent. This study tested the potential advantage of a social robot over a tablet in (second) language learning on performance, engagement, and enjoyment. Shortages in primary education call for new technology solutions. Previous studies combined robots with tablets, to compensate for robot’s limitations, however, this study applied direct human–robot interaction. Primary school children (N = 63, aged 4–6) participated in a 3-wave field experiment with story-telling exercises, either with a semi-autonomous robot (without tablet, using WOz) or a tablet. Results showed increased learning gains over time when training with a social robot, compared to the tablet. Children who trained with a robot were more engaged in the story-telling task and enjoyed it more. Robot’s behavioral style (social or neutral) hardly differed overall, however, seems to vary for high versus low educational abilities. While social robots need sophistication before being implemented in schools, our study shows the potential of social robots as tutors in (second) language learning.
1.1124
0.1124
0.1124
0.1124
0.104
0.050507
0.0052
0.001044
0
0
0
0
0
0
Near-Optimal Scheduling Mechanisms for Deadline-Sensitive Jobs in Large Computing Clusters We consider a market-based resource allocation model for batch jobs in cloud computing clusters. In our model, we incorporate the importance of the due date of a job rather than the number of servers allocated to it at any given time. Each batch job is characterized by the work volume of total computing units (e.g., CPU hours) along with a bound on maximum degree of parallelism. Users specify, along with these job characteristics, their desired due date and a value for finishing the job by its deadline. Given this specification, the primary goal is to determine the scheduling of cloud computing instances under capacity constraints in order to maximize the social welfare (i.e., sum of values gained by allocated users). Our main result is a new (CC-k&cdot;ss-1)-approximation algorithm for this objective, where C denotes cloud capacity, k is the maximal bound on parallelized execution (in practical settings, k < C) and s is the slackness on the job completion time, that is, the minimal ratio between a specified deadline and the earliest finish time of a job. Our algorithm is based on utilizing dual fitting arguments over a strengthened linear program to the problem. Based on the new approximation algorithm, we construct truthful allocation and pricing mechanisms, in which reporting the true value and other properties of the job (deadline, work volume, and the parallelism bound) is a dominant strategy for all users. To that end, we extend known results for single-value settings to provide a general framework for transforming allocation algorithms into truthful mechanisms in domains of single-value and multi-properties. We then show that the basic mechanism can be extended under proper Bayesian assumptions to the objective of maximizing revenues, which is important for public clouds. We empirically evaluate the benefits of our approach through simulations on data-center job traces, and show that the revenues obtained under our mechanism are comparable with an ideal fixed-price mechanism, which sets an on-demand price using oracle knowledge of users’ valuations. Finally, we discuss how our model can be extended to accommodate uncertainties in job work volumes, which is a practical challenge in cloud settings.
Competitive on-line scheduling with level of service Motivated by an application in thinwire visualization, we study an abstract on-line scheduling problem where the size of each requested service can be scaled down by the scheduler. Thus, our problem embodies a notion of "Level of Service" that is increasingly important in multimedia applications. We give two schedulers FirstFit and EndFit based on two simple heuristics, and generalize them into a class of greedy schedulers. We show that both FirstFit and EndFit are 2-competitive, and any greedy scheduler is 3-competitive. These bounds are shown to be tight.
Coordinated Charging of Electric Vehicles for Congestion Prevention in the Distribution Grid Distributed energy resources (DERs), like electric vehicles (EVs), can offer valuable services to power systems, such as enabling renewable energy to the electricity producer and providing ancillary services to the system operator. However, these new DERs may challenge the distribution grid due to insufficient capacity in peak hours. This paper aims to coordinate the valuable services and operation constraints of three actors: the EV owner, the Fleet operator (FO) and the Distribution system operator (DSO), considering the individual EV owner's driving requirement, the charging cost of EV and thermal limits of cables and transformers in the proposed market framework. Firstly, a theoretical market framework is described. Within this framework, FOs who represent their customer's (EV owners) interests will centrally guarantee the EV owners' driving requirements and procure the energy for their vehicles with lower cost. The congestion problem will be solved by a coordination between DSO and FOs through a distribution grid capacity market scheme. Then, a mathematical formulation of the market scheme is presented. Further, some case studies are shown to illustrate the effectiveness of the proposed solutions.
Real-Time PEV Charging/Discharging Coordination in Smart Distribution Systems. This paper proposes a novel online coordination method for the charging of plug-in electric vehicles (PEVs) in smart distribution networks. The goal of the proposed method is to optimally charge the PEVs in order to maximize the PEV owners&#39; satisfaction and to minimize system operating costs without violating power system constraints. Unlike the solutions reported in the literature, the proposed c...
Decentralized Plug-in Electric Vehicle Charging Selection Algorithm in Power Systems This paper uses a charging selection concept for plug-in electric vehicles (PEVs) to maximize user convenience levels while meeting predefined circuit-level demand limits. The optimal PEV-charging selection problem requires an exhaustive search for all possible combinations of PEVs in a power system, which cannot be solved for the practical number of PEVs. Inspired by the efficiency of the convex relaxation optimization tool in finding close-to-optimal results in huge search spaces, this paper proposes the application of the convex relaxation optimization method to solve the PEV-charging selection problem. Compared with the results of the uncontrolled case, the simulated results indicate that the proposed PEV-charging selection algorithm only slightly reduces user convenience levels, but significantly mitigates the impact of the PEV-charging on the power system. We also develop a distributed optimization algorithm to solve the PEV-charging selection problem in a decentralized manner, i.e., the binary charging decisions (charged or not charged) are made locally by each vehicle. Using the proposed distributed optimization algorithm, each vehicle is only required to report its power demand rather than report several of its private user state information, mitigating the security problems inherent in such problem. The proposed decentralized algorithm only requires low-speed communication capability, making it suitable for real-time implementation.
Online Coordinated Charging Decision Algorithm for Electric Vehicles Without Future Information The large-scale integration of plug-in electric vehicles (PEVs) to the power grid spurs the need for efficient charging coordination mechanisms. It can be shown that the optimal charging schedule smooths out the energy consumption over time so as to minimize the total energy cost. In practice, however, it is hard to smooth out the energy consumption perfectly, because the future PEV charging demand is unknown at the moment when the charging rate of an existing PEV needs to be determined. In this paper, we propose an online coordinated charging decision (ORCHARD) algorithm, which minimizes the energy cost without knowing the future information. Through rigorous proof, we show that ORCHARD is strictly feasible in the sense that it guarantees to fulfill all charging demands before due time. Meanwhile, it achieves the best known competitive ratio of 2.39. By exploiting the problem structure, we propose a novel reduced-complexity algorithm to replace the standard convex optimization techniques used in ORCHARD. Through extensive simulations, we show that the average performance gap between ORCHARD and the offline optimal solution, which utilizes the complete future information, is as small as 6.5%. By setting a proper speeding factor, the average performance gap can be further reduced to 5%.
Distributed Optimization and Statistical Learning via the Alternating Direction Method of Multipliers Many problems of recent interest in statistics and machine learning can be posed in the framework of convex optimization. Due to the explosion in size and complexity of modern datasets, it is increasingly important to be able to solve problems with a very large number of features or training examples. As a result, both the decentralized collection or storage of these datasets as well as accompanying distributed solution methods are either necessary or at least highly desirable. In this review, we argue that the alternating direction method of multipliers is well suited to distributed convex optimization, and in particular to large-scale problems arising in statistics, machine learning, and related areas. The method was developed in the 1970s, with roots in the 1950s, and is equivalent or closely related to many other algorithms, such as dual decomposition, the method of multipliers, Douglas–Rachford splitting, Spingarn's method of partial inverses, Dykstra's alternating projections, Bregman iterative algorithms for ℓ1 problems, proximal methods, and others. After briefly surveying the theory and history of the algorithm, we discuss applications to a wide variety of statistical and machine learning problems of recent interest, including the lasso, sparse logistic regression, basis pursuit, covariance selection, support vector machines, and many others. We also discuss general distributed optimization, extensions to the nonconvex setting, and efficient implementation, including some details on distributed MPI and Hadoop MapReduce implementations.
Switching between stabilizing controllers This paper deals with the problem of switching between several linear time-invariant (LTI) controllers—all of them capable of stabilizing a speci4c LTI process—in such a way that the stability of the closed-loop system is guaranteed for any switching sequence. We show that it is possible to 4nd realizations for any given family of controller transfer matrices so that the closed-loop system remains stable, no matter how we switch among the controller. The motivation for this problem is the control of complex systems where con8icting requirements make a single LTI controller unsuitable. ? 2002 Published by Elsevier Science Ltd.
LMM: latency-aware micro-service mashup in mobile edge computing environment Internet of Things (IoT) applications introduce a set of stringent requirements (e.g., low latency, high bandwidth) to network and computing paradigm. 5G networks are faced with great challenges for supporting IoT services. The centralized cloud computing paradigm also becomes inefficient for those stringent requirements. Only extending spectrum resources cannot solve the problem effectively. Mobile edge computing offers an IT service environment at the Radio Access Network edge and presents great opportunities for the development of IoT applications. With the capability to reduce latency and offer an improved user experience, mobile edge computing becomes a key technology toward 5G. To achieve abundant sharing, complex IoT applications have been implemented as a set of lightweight micro-services that are distributed among containers over the mobile edge network. How to produce the optimal collocation of suitable micro-service for an application in mobile edge computing environment is an important issue that should be addressed. To address this issue, we propose a latency-aware micro-service mashup approach in this paper. Firstly, the problem is formulated into an integer nonlinear programming. Then, we prove the NP-hardness of the problem by reducing it into the delay constrained least cost problem. Finally, we propose an approximation latency-aware micro-service mashup approach to solve the problem. Experiment results show that the proposed approach achieves a substantial reduction in network resource consumption while still ensuring the latency constraint.
Symbolic model checking for real-time systems We describe finite-state programs over real-numbered time in a guarded-command language with real-valued clocks or, equivalently, as finite automata with real-valued clocks. Model checking answers the question which states of a real-time program satisfy a branching-time specification (given in an extension of CTL with clock variables). We develop an algorithm that computes this set of states symbolically as a fixpoint of a functional on state predicates, without constructing the state space. For this purpose, we introduce a μ-calculus on computation trees over real-numbered time. Unfortunately, many standard program properties, such as response for all nonzeno execution sequences (during which time diverges), cannot be characterized by fixpoints: we show that the expressiveness of the timed μ-calculus is incomparable to the expressiveness of timed CTL. Fortunately, this result does not impair the symbolic verification of "implementable" real-time programs-those whose safety constraints are machine-closed with respect to diverging time and whose fairness constraints are restricted to finite upper bounds on clock values. All timed CTL properties of such programs are shown to be computable as finitely approximable fixpoints in a simple decidable theory.
Dynamic transfer among alternative controllers and its relation to antiwindup controller design Advanced control strategies and modern consulting provide new challenges for the classical problem of bumpless transfer. It can, for example, be necessary to transfer between an only approximately known existing analog controller and a new digital or adaptive controller without accessing any states. Transfer ought to be bidirectional and not presuppose steady state, so that an immediate back-transfer is possible if the new controller should drive the plant unstable. We present a scheme that meets these requirements. By casting the problem of bidirectional transfer into an associated tracking control problem, systematic analysis and design procedures from control theory can be applied. The associated control problem also has a correspondence to the design of antiwindup controllers. The paper includes laboratory and industrial applications.
Online Prediction of Driver Distraction Based on Brain Activity Patterns This paper presents a new computational framework for early detection of driver distractions (map viewing) using brain activity measured by electroencephalographic (EEG) signals. Compared with most studies in the literature, which are mainly focused on the classification of distracted and nondistracted periods, this study proposes a new framework to prospectively predict the start and end of a distraction period, defined by map viewing. The proposed prediction algorithm was tested on a data set of continuous EEG signals recorded from 24 subjects. During the EEG recordings, the subjects were asked to drive from an initial position to a destination using a city map in a simulated driving environment. The overall accuracy values for the prediction of the start and the end of map viewing were 81% and 70%, respectively. The experimental results demonstrated that the proposed algorithm can predict the start and end of map viewing with relatively high accuracy and can be generalized to individual subjects. The outcome of this study has a high potential to improve the design of future intelligent navigation systems. Prediction of the start of map viewing can be used to provide route information based on a driver's needs and consequently avoid map-viewing activities. Prediction of the end of map viewing can be used to provide warnings for potential long map-viewing durations. Further development of the proposed framework and its applications in driver-distraction predictions are also discussed.
An efficient scheduling scheme for mobile charger in on-demand wireless rechargeable sensor networks. Existing studies on wireless sensor networks (WSNs) have revealed that the limited battery capacity of sensor nodes (SNs) hinders their perpetual operation. Recent findings in the domain of wireless energy transfer (WET) have attracted a lot of attention of academia and industry to cater the lack of energy in the WSNs. The main idea of WET is to restore the energy of SNs using one or more wireless mobile chargers (MCs), which leads to a new paradigm of wireless rechargeable sensor networks (WRSNs). The determination of an optimal order of charging the SNs (i.e., charging schedule) in an on-demand WRSN is a well-known NP-hard problem. Moreover, care must be taken while designing the charging schedule of an MC as requesting SNs introduce both spatial and temporal constraints. In this paper, we first present a Linear Programming (LP) formulation for the problem of scheduling an MC and then propose an efficient solution based on gravitational search algorithm (GSA). Our method is presented with a novel agent representation scheme and an efficient fitness function. We perform extensive simulations on the proposed scheme to demonstrate its effectiveness over two state-of-the-art algorithms, namely first come first serve (FCFS) and nearest job next with preemption (NJNP). The simulation results reveal that the proposed scheme outperforms both the existing algorithms in terms of charging latency. The virtue of our scheme is also proved by the well-known statistical test, analysis of variance (ANOVA), followed by post hoc analysis.
Social Robots for (Second) Language Learning in (Migrant) Primary School Children Especially these days, innovation and support from technology to relieve pressure in education is highly urgent. This study tested the potential advantage of a social robot over a tablet in (second) language learning on performance, engagement, and enjoyment. Shortages in primary education call for new technology solutions. Previous studies combined robots with tablets, to compensate for robot’s limitations, however, this study applied direct human–robot interaction. Primary school children (N = 63, aged 4–6) participated in a 3-wave field experiment with story-telling exercises, either with a semi-autonomous robot (without tablet, using WOz) or a tablet. Results showed increased learning gains over time when training with a social robot, compared to the tablet. Children who trained with a robot were more engaged in the story-telling task and enjoyed it more. Robot’s behavioral style (social or neutral) hardly differed overall, however, seems to vary for high versus low educational abilities. While social robots need sophistication before being implemented in schools, our study shows the potential of social robots as tutors in (second) language learning.
1.114534
0.12
0.12
0.12
0.06175
0.04
0.000475
0
0
0
0
0
0
0
Long-Term Traffic Speed Prediction Based on Multiscale Spatio-Temporal Feature Learning Network Speed plays a significant role in evaluating the evolution of traffic status, and predicting speed is one of the fundamental tasks for the intelligent transportation system. There exists a large number of works on speed forecast; however, the problem of long-term prediction for the next day is still not well addressed. In this paper, we propose a multiscale spatio-temporal feature learning network (MSTFLN) as the model to handle the challenging task of long-term traffic speed prediction for elevated highways. Raw traffic speed data collected from loop detectors every 5 min are transformed into spatial-temporal matrices; each matrix represents the one-day speed information, rows of the matrix indicate the numbers of loop detectors, and time intervals are denoted by columns. To predict the traffic speed of a certain day, nine speed matrices of three historical days with three different time scales are served as the input of MSTFLN. The proposed MSTFLN model consists of convolutional long short-term memories and convolutional neural networks. Experiments are evaluated using the data of three main elevated highways in Shanghai, China. The presented results demonstrate that our approach outperforms the state-of-the-art work and it can effectively predict the long-term speed information.
Forecasting holiday daily tourist flow based on seasonal support vector regression with adaptive genetic algorithm. •The model of support vector regression with adaptive genetic algorithm and the seasonal mechanism is proposed.•Parameters selection and seasonal adjustment should be carefully selected.•We focus on latest and representative holiday daily data in China.•Two experiments are used to prove the effect of the model.•The AGASSVR is superior to AGA-SVR and BPNN.
Regression conformal prediction with random forests Regression conformal prediction produces prediction intervals that are valid, i.e., the probability of excluding the correct target value is bounded by a predefined confidence level. The most important criterion when comparing conformal regressors is efficiency; the prediction intervals should be as tight (informative) as possible. In this study, the use of random forests as the underlying model for regression conformal prediction is investigated and compared to existing state-of-the-art techniques, which are based on neural networks and k-nearest neighbors. In addition to their robust predictive performance, random forests allow for determining the size of the prediction intervals by using out-of-bag estimates instead of requiring a separate calibration set. An extensive empirical investigation, using 33 publicly available data sets, was undertaken to compare the use of random forests to existing state-of-the-art conformal predictors. The results show that the suggested approach, on almost all confidence levels and using both standard and normalized nonconformity functions, produced significantly more efficient conformal predictors than the existing alternatives.
Learning to Predict Bus Arrival Time From Heterogeneous Measurements via Recurrent Neural Network Bus arrival time prediction intends to improve the level of the services provided by transportation agencies. Intuitively, many stochastic factors affect the predictability of the arrival time, e.g., weather and local events. Moreover, the arrival time prediction for a current station is closely correlated with that of multiple passed stations. Motivated by the observations above, this paper propo...
Hybrid Spatio-Temporal Graph Convolutional Network: Improving Traffic Prediction with Navigation Data Traffic forecasting has recently attracted increasing interest due to the popularity of online navigation services, ridesharing and smart city projects. Owing to the non-stationary nature of road traffic, forecasting accuracy is fundamentally limited by the lack of contextual information. To address this issue, we propose the Hybrid Spatio-Temporal Graph Convolutional Network (H-STGCN), which is able to "deduce" future travel time by exploiting the data of upcoming traffic volume. Specifically, we propose an algorithm to acquire the upcoming traffic volume from an online navigation engine. Taking advantage of the piecewise-linear flow-density relationship, a novel transformer structure converts the upcoming volume into its equivalent in travel time. We combine this signal with the commonly-utilized travel-time signal, and then apply graph convolution to capture the spatial dependency. Particularly, we construct a compound adjacency matrix which reflects the innate traffic proximity. We conduct extensive experiments on real-world datasets. The results show that H-STGCN remarkably outperforms state-of-the-art methods in various metrics, especially for the prediction of non-recurring congestion.
Learning Interpretable Deep State Space Model for Probabilistic Time Series Forecasting. Probabilistic time series forecasting involves estimating the distribution of future based on its history, which is essential for risk management in downstream decision-making. We propose a deep state space model for probabilistic time series forecasting whereby the non-linear emission model and transition model are parameterized by networks and the dependency is modeled by recurrent neural nets. We take the automatic relevance determination (ARD) view and devise a network to exploit the exogenous variables in addition to time series. In particular, our ARD network can incorporate the uncertainty of the exogenous variables and eventually helps identify useful exogenous variables and suppress those irrelevant for forecasting. The distribution of multi-step ahead forecasts are approximated by Monte Carlo simulation. We show in experiments that our model produces accurate and sharp probabilistic forecasts. The estimated uncertainty of our forecasting also realistically increases over time, in a spontaneous manner.
Transfer Knowledge between Cities The rapid urbanization has motivated extensive research on urban computing. It is critical for urban computing tasks to unlock the power of the diversity of data modalities generated by different sources in urban spaces, such as vehicles and humans. However, we are more likely to encounter the label scarcity problem and the data insufficiency problem when solving an urban computing task in a city where services and infrastructures are not ready or just built. In this paper, we propose a FLexible multimOdal tRAnsfer Learning (FLORAL) method to transfer knowledge from a city where there exist sufficient multimodal data and labels, to this kind of cities to fully alleviate the two problems. FLORAL learns semantically related dictionaries for multiple modalities from a source domain, and simultaneously transfers the dictionaries and labelled instances from the source into a target domain. We evaluate the proposed method with a case study of air quality prediction.
Space-time modeling of traffic flow. This paper discusses the application of space-time autoregressive integrated moving average (STARIMA) methodology for representing traffic flow patterns. Traffic flow data are in the form of spatial time series and are collected at specific locations at constant intervals of time. Important spatial characteristics of the space-time process are incorporated in the STARIMA model through the use of weighting matrices estimated on the basis of the distances among the various locations where data are collected. These matrices distinguish the space-time approach from the vector autoregressive moving average (VARMA) methodology and enable the model builders to control the number of the parameters that have to be estimated. The proposed models can be used for short-term forecasting of space-time stationary traffic-flow processes and for assessing the impact of traffic-flow changes on other parts of the network. The three-stage iterative space-time model building procedure is illustrated using 7.5min average traffic flow data for a set of 25 loop-detectors located at roads that direct to the centre of the city of Athens, Greece. Data for two months with different traffic-flow characteristics are modelled in order to determine the stability of the parameter estimation.
Model-based periodic event-triggered control for linear systems Periodic event-triggered control (PETC) is a control strategy that combines ideas from conventional periodic sampled-data control and event-triggered control. By communicating periodically sampled sensor and controller data only when needed to guarantee stability or performance properties, PETC is capable of reducing the number of transmissions significantly, while still retaining a satisfactory closed-loop behavior. In this paper, we will study observer-based controllers for linear systems and propose advanced event-triggering mechanisms (ETMs) that will reduce communication in both the sensor-to-controller channels and the controller-to-actuator channels. By exploiting model-based computations, the new classes of ETMs will outperform existing ETMs in the literature. To model and analyze the proposed classes of ETMs, we present two frameworks based on perturbed linear and piecewise linear systems, leading to conditions for global exponential stability and @?"2-gain performance of the resulting closed-loop systems in terms of linear matrix inequalities. The proposed analysis frameworks can be used to make tradeoffs between the network utilization on the one hand and the performance in terms of @?"2-gains on the other. In addition, we will show that the closed-loop performance realized by an observer-based controller, implemented in a conventional periodic time-triggered fashion, can be recovered arbitrarily closely by a PETC implementation. This provides a justification for emulation-based design. Next to centralized model-based ETMs, we will also provide a decentralized setup suitable for large-scale systems, where sensors and actuators are physically distributed over a wide area. The improvements realized by the proposed model-based ETMs will be demonstrated using numerical examples.
Affective social robots For human-robot interaction to proceed in a smooth, natural manner, robots must adhere to human social norms. One such human convention is the use of expressive moods and emotions as an integral part of social interaction. Such expressions are used to convey messages such as ''I'm happy to see you'' or ''I want to be comforted,'' and people's long-term relationships depend heavily on shared emotional experiences. Thus, we have developed an affective model for social robots. This generative model attempts to create natural, human-like affect and includes distinctions between immediate emotional responses, the overall mood of the robot, and long-term attitudes toward each visitor to the robot, with a focus on developing long-term human-robot relationships. This paper presents the general affect model as well as particular details of our implementation of the model on one robot, the Roboceptionist. In addition, we present findings from two studies that demonstrate the model's potential.
Rich Models for Steganalysis of Digital Images We describe a novel general strategy for building steganography detectors for digital images. The process starts with assembling a rich model of the noise component as a union of many diverse submodels formed by joint distributions of neighboring samples from quantized image noise residuals obtained using linear and nonlinear high-pass filters. In contrast to previous approaches, we make the model assembly a part of the training process driven by samples drawn from the corresponding cover- and stego-sources. Ensemble classifiers are used to assemble the model as well as the final steganalyzer due to their low computational complexity and ability to efficiently work with high-dimensional feature spaces and large training sets. We demonstrate the proposed framework on three steganographic algorithms designed to hide messages in images represented in the spatial domain: HUGO, edge-adaptive algorithm by Luo , and optimally coded ternary $\\pm {\\hbox{1}}$ embedding. For each algorithm, we apply a simple submodel-selection technique to increase the detection accuracy per model dimensionality and show how the detection saturates with increasing complexity of the rich model. By observing the differences between how different submodels engage in detection, an interesting interplay between the embedding and detection is revealed. Steganalysis built around rich image models combined with ensemble classifiers is a promising direction towards automatizing steganalysis for a wide spectrum of steganographic schemes.
Heterogeneous ensemble for feature drifts in data streams The nature of data streams requires classification algorithms to be real-time, efficient, and able to cope with high-dimensional data that are continuously arriving. It is a known fact that in high-dimensional datasets, not all features are critical for training a classifier. To improve the performance of data stream classification, we propose an algorithm called HEFT-Stream (H eterogeneous E nsemble with F eature drifT for Data Streams ) that incorporates feature selection into a heterogeneous ensemble to adapt to different types of concept drifts. As an example of the proposed framework, we first modify the FCBF [13] algorithm so that it dynamically update the relevant feature subsets for data streams. Next, a heterogeneous ensemble is constructed based on different online classifiers, including Online Naive Bayes and CVFDT [5]. Empirical results show that our ensemble classifier outperforms state-of-the-art ensemble classifiers (AWE [15] and OnlineBagging [21]) in terms of accuracy, speed, and scalability. The success of HEFT-Stream opens new research directions in understanding the relationship between feature selection techniques and ensemble learning to achieve better classification performance.
Orientation-aware RFID tracking with centimeter-level accuracy. RFID tracking attracts a lot of research efforts in recent years. Most of the existing approaches, however, adopt an orientation-oblivious model. When tracking a target whose orientation changes, those approaches suffer from serious accuracy degradation. In order to achieve target tracking with pervasive applicability in various scenarios, we in this paper propose OmniTrack, an orientation-aware RFID tracking approach. Our study discovers the linear relationship between the tag orientation and the phase change of the backscattered signals. Based on this finding, we propose an orientation-aware phase model to explicitly quantify the respective impact of the read-tag distance and the tag's orientation. OmniTrack addresses practical challenges in tracking the location and orientation of a mobile tag. Our experimental results demonstrate that OmniTrack achieves centimeter-level location accuracy and has significant advantages in tracking targets with varing orientations, compared to the state-of-the-art approaches.
Robot tutor and pupils’ educational ability: Teaching the times tables Research shows promising results of educational robots in language and STEM tasks. In language, more research is available, occasionally in view of individual differences in pupils’ educational ability levels, and learning seems to improve with more expressive robot behaviors. In STEM, variations in robots’ behaviors have been examined with inconclusive results and never while systematically investigating how differences in educational abilities match with different robot behaviors. We applied an autonomously tutoring robot (without tablet, partly WOz) in a 2 × 2 experiment of social vs. neutral behavior in above-average vs. below-average schoolchildren (N = 86; age 8–10 years) while rehearsing the multiplication tables on a one-to-one basis. The standard school test showed that on average, pupils significantly improved their performance even after 3 occasions of 5-min exercises. Beyond-average pupils profited most from a robot tutor, whereas those below average in multiplication benefited more from a robot that showed neutral rather than more social behavior.
1.22
0.22
0.22
0.22
0.22
0.22
0.073333
0.003542
0
0
0
0
0
0
Space-time super-resolution. We propose a method for constructing a video sequence of high space-time resolution by combining information from multiple low-resolution video sequences of the same dynamic scene. Super-resolution is performed simultaneously in time and in space. By "temporal super-resolution," we mean recovering rapid dynamic events that occur faster than regular frame-rate. Such dynamic events are not visible (or else are observed incorrectly) in any of the input sequences, even if these are played in "slow-motion." The spatial and temporal dimensions are very different in nature, yet are interrelated. This leads to interesting visual trade-offs in time and space and to new video applications. These include: 1) treatment of spatial artifacts (e.g., motion-blur) by increasing the temporal resolution and 2) combination of input sequences of different space-time resolutions (e.g., NTSC, PAL, and even high quality still images) to generate a high quality video sequence. We further analyze and compare characteristics of temporal super-resolution to those of spatial super-resolution. These include: How many video cameras are needed to obtain increased resolution? What is the upper bound on resolution improvement via super-resolution? What is the temporal analogue to the spatial "ringing" effect?
Transient attributes for high-level understanding and editing of outdoor scenes We live in a dynamic visual world where the appearance of scenes changes dramatically from hour to hour or season to season. In this work we study \"transient scene attributes\" -- high level properties which affect scene appearance, such as \"snow\", \"autumn\", \"dusk\", \"fog\". We define 40 transient attributes and use crowdsourcing to annotate thousands of images from 101 webcams. We use this \"transient attribute database\" to train regressors that can predict the presence of attributes in novel images. We demonstrate a photo organization method based on predicted attributes. Finally we propose a high-level image editing method which allows a user to adjust the attributes of a scene, e.g. change a scene to be \"snowy\" or \"sunset\". To support attribute manipulation we introduce a novel appearance transfer technique which is simple and fast yet competitive with the state-of-the-art. We show that we can convincingly modify many transient attributes in outdoor scenes.
Combining Markov Random Fields And Convolutional Neural Networks For Image Synthesis This paper studies a combination of generative Markov random field (MRF) models and discriminatively trained deep convolutional neural networks (dCNNs) for synthesizing 2D images. The generative MRF acts on higher-levels of a dCNN feature pyramid, controling the image layout at an abstract level. We apply the method to both photographic and non-photo-realistic (artwork) synthesis tasks. The MRF regularizer prevents over-excitation artifacts and reduces implausible feature mixtures common to previous dCNN inversion approaches, permitting synthezing photographic content with increased visual plausibility. Unlike standard MRF-based texture synthesis, the combined system can both match and adapt local features with considerable variability, yielding results far out of reach of classic generative MRF methods.
Deepfashion: Powering Robust Clothes Recognition And Retrieval With Rich Annotations Recent advances in clothes recognition have been driven by the construction of clothes datasets. Existing datasets are limited in the amount of annotations and are difficult to cope with the various challenges in real-world applications. In this work, we introduce DeepFashion(1), a large-scale clothes dataset with comprehensive annotations. It contains over 800,000 images, which are richly annotated with massive attributes, clothing landmarks, and correspondence of images taken under different scenarios including store, street snapshot, and consumer. Such rich annotations enable the development of powerful algorithms in clothes recognition and facilitating future researches. To demonstrate the advantages of DeepFashion, we propose a new deep model, namely FashionNet, which learns clothing features by jointly predicting clothing attributes and landmarks. The estimated landmarks are then employed to pool or gate the learned features. It is optimized in an iterative manner. Extensive experiments demonstrate the effectiveness of FashionNet and the usefulness of DeepFashion.
Flymap: Interacting With Maps Projected From A Drone Interactive maps have become ubiquitous in our daily lives, helping us reach destinations and discovering our surroundings. Yet, designing map interactions is not straightforward and depends on the device being used. As mobile devices evolve and become independent from users, such as with robots and drones, how will we interact with the maps they provide? We propose FlyMap as a novel user experience for drone-based interactive maps. We designed and developed three interaction techniques for FlyMap's usage scenarios. In a comprehensive indoor study (N = 16), we show the strengths and weaknesses of two techniques on users' cognition, task load, and satisfaction. FlyMap was then pilot tested with the third technique outdoors in real world conditions with four groups of participants (N = 13). We show that FlyMap's interactivity is exciting to users and opens the space for more direct interactions with drones.
Maskgan: Towards Diverse And Interactive Facial Image Manipulation Facial image manipulation has achieved great progress in recent years. However; previous methods either operate on a predefined set of face attributes or leave users little freedom to interactively manipulate images. To overcome these drawbacks, we propose a novel framework termed MaskGAN, enabling diverse and interactive face manipulation. Our key insight is that semantic masks serve as a suitable intermediate representation for flexible face manipulation with fidelity preservation. MaskGAN has two main components: 1) Dense Mapping Network (DMN) and 2) Editing Behavior Simulated Training (EBST). Specifically, DMN learns style mapping between a free-form user modified mask and a target image, enabling diverse generation results. EBST models the user editing behavior on the source mask, making the overall framework more robust to various manipulated inputs. Specifically, it introduces dual-editing consistency as the auxiliary supervision signal. To facilitate extensive studies, we construct a large-scale high-resolution face dataset with fine-grained mask annotations named CelebAMask-HQ. MaskGAN is comprehensively evaluated on two challenging tasks: attribute transfer and style copy, demonstrating superior performance over other state-of-the-art methods. The code, models, and dataset are available at https://github.com/switchablenorms/CeleAMask-HQ.
GLEAN: Generative Latent Bank for Large-Factor Image Super-Resolution We show that pre-trained Generative Adversarial Networks (GANs), e.g., StyleGAN, can be used as a latent bank to improve the restoration quality of large-factor image super-resolution (SR). While most existing SR approaches attempt to generate realistic textures through learning with adversarial loss, our method, Generative LatEnt bANk (GLEAN), goes beyond existing practices by directly leveraging rich and diverse priors encapsulated in a pre-trained GAN. But unlike prevalent GAN inversion methods that require expensive image-specific optimization at runtime, our approach only needs a single forward pass to generate the upscaled image. GLEAN can be easily incorporated in a simple encoder-bank-decoder architecture with multi-resolution skip connections. Switching the bank allows the method to deal with images from diverse categories, e.g., cat, building, human face, and car. Images upscaled by GLEAN show clear improvements in terms of fidelity and texture faithfulness in comparison to existing methods as shown in Fig. 1.
End-To-End Time-Lapse Video Synthesis From A Single Outdoor Image Time-lapse videos usually contain visually appealing content but are often difficult and costly to create. In this paper, we present an end-to-end solution to synthesize a time-lapse video from a single outdoor image using deep neural networks. Our key idea is to train a conditional generative adversarial network based on existing datasets of time-lapse videos and image sequences. We propose a multi-frame joint conditional generation framework to effectively learn the correlation between the illumination change of an outdoor scene and the time of the day. We further present a multi-domain training scheme for robust training of our generative models from two datasets with different distributions and missing timestamp labels. Compared to alternative time-lapse video synthesis algorithms, our method uses the timestamp as the control variable and does not require a reference video to guide the synthesis of the final output. We conduct ablation studies to validate our algorithm and compare with state-of-the-art techniques both qualitatively and quantitatively.
Sequence to Sequence Learning with Neural Networks. Deep Neural Networks (DNNs) are powerful models that have achieved excellent performance on difficult learning tasks. Although DNNs work well whenever large labeled training sets are available, they cannot be used to map sequences to sequences. In this paper, we present a general end-to-end approach to sequence learning that makes minimal assumptions on the sequence structure. Our method uses a multilayered Long Short-Term Memory (LSTM) to map the input sequence to a vector of a fixed dimensionality, and then another deep LSTM to decode the target sequence from the vector. Our main result is that on an English to French translation task from the WMT-14 dataset, the translations produced by the LSTM achieve a BLEU score of 34.8 on the entire test set, where the LSTM's BLEU score was penalized on out-of-vocabulary words. Additionally, the LSTM did not have difficulty on long sentences. For comparison, a phrase-based SMT system achieves a BLEU score of 33.3 on the same dataset. When we used the LSTM to rerank the 1000 hypotheses produced by the aforementioned SMT system, its BLEU score increases to 36.5, which is close to the previous state of the art. The LSTM also learned sensible phrase and sentence representations that are sensitive to word order and are relatively invariant to the active and the passive voice. Finally, we found that reversing the order of the words in all source sentences (but not target sentences) improved the LSTM's performance markedly, because doing so introduced many short term dependencies between the source and the target sentence which made the optimization problem easier.
A General Equilibrium Model for Industries with Price and Service Competition This paper develops a stochastic general equilibrium inventory model for an oligopoly, in which all inventory constraint parameters are endogenously determined. We propose several systems of demand processes whose distributions are functions of all retailers' prices and all retailers' service levels. We proceed with the investigation of the equilibrium behavior of infinite-horizon models for industries facing this type of generalized competition, under demand uncertainty.We systematically consider the following three competition scenarios. (1) Price competition only: Here, we assume that the firms' service levels are exogenously chosen, but characterize how the price and inventory strategy equilibrium vary with the chosen service levels. (2) Simultaneous price and service-level competition: Here, each of the firms simultaneously chooses a service level and a combined price and inventory strategy. (3) Two-stage competition: The firms make their competitive choices sequentially. In a first stage, all firms simultaneously choose a service level; in a second stage, the firms simultaneously choose a combined pricing and inventory strategy with full knowledge of the service levels selected by all competitors. We show that in all of the above settings a Nash equilibrium of infinite-horizon stationary strategies exists and that it is of a simple structure, provided a Nash equilibrium exists in a so-called reduced game.We pay particular attention to the question of whether a firm can choose its service level on the basis of its own (input) characteristics (i.e., its cost parameters and demand function) only. We also investigate under which of the demand models a firm, under simultaneous competition, responds to a change in the exogenously specified characteristics of the various competitors by either: (i) adjusting its service level and price in the same direction, thereby compensating for price increases (decreases) by offering improved (inferior) service, or (ii) adjusting them in opposite directions, thereby simultaneously offering better or worse prices and service.
Mobile cloud computing: A survey Despite increasing usage of mobile computing, exploiting its full potential is difficult due to its inherent problems such as resource scarcity, frequent disconnections, and mobility. Mobile cloud computing can address these problems by executing mobile applications on resource providers external to the mobile device. In this paper, we provide an extensive survey of mobile cloud computing research, while highlighting the specific concerns in mobile cloud computing. We present a taxonomy based on the key issues in this area, and discuss the different approaches taken to tackle these issues. We conclude the paper with a critical analysis of challenges that have not yet been fully met, and highlight directions for future work.
Eye-vergence visual servoing enhancing Lyapunov-stable trackability Visual servoing methods for hand---eye configuration are vulnerable for hand's dynamical oscillation, since nonlinear dynamical effects of whole manipulator stand against the stable tracking ability (trackability). Our proposal to solve this problem is that the controller for visual servoing of the hand and the one for eye-vergence should be separated independently based on decoupling each other, where the trackability is verified by Lyapunov analysis. Then the effectiveness of the decoupled hand and eye-vergence visual servoing method is evaluated through simulations incorporated with actual dynamics of 7-DoF robot with additional 3-DoF for eye-vergence mechanism by amplitude and phase frequency analysis.
An improved E-DRM scheme for mobile environments. With the rapid development of information science and network technology, Internet has become an important platform for the dissemination of digital content, which can be easily copied and distributed through the Internet. Although convenience is increased, it causes significant damage to authors of digital content. Digital rights management system (DRM system) is an access control system that is designed to protect digital content and ensure illegal users from maliciously spreading digital content. Enterprise Digital Rights Management system (E-DRM system) is a DRM system that prevents unauthorized users from stealing the enterprise's confidential data. User authentication is the most important method to ensure digital rights management. In order to verify the validity of user, the biometrics-based authentication protocol is widely used due to the biological characteristics of each user are unique. By using biometric identification, it can ensure the correctness of user identity. In addition, due to the popularity of mobile device and Internet, user can access digital content and network information at anytime and anywhere. Recently, Mishra et al. proposed an anonymous and secure biometric-based enterprise digital rights management system for mobile environment. Although biometrics-based authentication is used to prevent users from being forged, the anonymity of users and the preservation of digital content are not ensured in their proposed system. Therefore, in this paper, we will propose a more efficient and secure biometric-based enterprise digital rights management system with user anonymity for mobile environments.
Intention-detection strategies for upper limb exosuits: model-based myoelectric vs dynamic-based control The cognitive human-robot interaction between an exosuit and its wearer plays a key role in determining both the biomechanical effects of the device on movements and its perceived effectiveness. There is a lack of evidence, however, on the comparative performance of different control methods, implemented on the same device. Here, we compare two different control approaches on the same robotic suit: a model-based myoelectric control (myoprocessor), which estimates the joint torque from the activation of target muscles, and a dynamic-based control that provides support against gravity using an inverse dynamic model. Tested on a cohort of four healthy participants, assistance from the exosuit results in a marked reduction in the effort of muscles working against gravity with both control approaches (peak reduction of 68.6±18.8%, for the dynamic arm model and 62.4±25.1% for the myoprocessor), when compared to an unpowered condition. Neither of the two controllers had an affect on the performance of their users in a joint-angle tracking task (peak errors of 15.4° and 16.4° for the dynamic arm model and myoprocessor, respectively, compared to 13.1o in the unpowered condition). However, our results highlight the remarkable adaptability of the myoprocessor to seamlessly adapt to changing external dynamics.
1.2
0.2
0.2
0.2
0.2
0.2
0.2
0.1
0
0
0
0
0
0
Ear recognition system using adaptive approach Runge–Kutta (AARK) threshold segmentation with ANFIS classification In the field of biometrics, ear recognition is a niche idea of recognition for human authentication, which has several merits compared to the other biometric recognitions like face and finger print. The contour of an ear is distinctive for each person, which is the main reason for choosing this recognition technique. Only in very few studies, the ear recognition algorithms were presented. There still remains a large space for research in the field of ear biometrics. All recent papers have implemented ear recognition algorithms using 2-D ear images. The ear recognition algorithms should be efficient in order to provide accurate results, owing to issues like multiple poses and directional related. This paper proposes a novel method for segmentation based on adaptive approach Runge–Kutta (AARK) to recognize ear images. AARK threshold segmentation technique is used for finding the threshold value to determine the region to be segmented. The utilization of AARK’s numerical methods in computing the threshold value for ear recognition process improves the result accuracy. Firstly, preprocessing has been carried out for the dataset. The following steps are carried out sequentially: ring projection, information normalization, morphological operation, AARK segmentation, feature extraction of DWT and finally ANFIS classifier are used. Among the various steps mentioned, ring projection converts the two dimensions into single dimensions. The self-adaptive discrete wavelet transform is used to extract features from the segmented region. Then the ANFIS classifier is used to recognize the ear region from the image by taking the features form the test image and the training images. The proposed method obtained 72% improvement in PSNR and accuracy is improved to 63.3%. Moreover, the speed and space occupation of the self-adaptive DWT technique and the conventional DWT technique are measured by implementing the methods in FPGA Spartan 6 device. Comparing with the implementation of conventional DWT, the area is reduced to 361 from 7021 while implementing the proposed self-adaptive DWT method.
Joint discriminative dimensionality reduction and dictionary learning for face recognition In linear representation based face recognition (FR), it is expected that a discriminative dictionary can be learned from the training samples so that the query sample can be better represented for classification. On the other hand, dimensionality reduction is also an important issue for FR. It cannot only reduce significantly the storage space of face images, but also enhance the discrimination of face feature. Existing methods mostly perform dimensionality reduction and dictionary learning separately, which may not fully exploit the discriminative information in the training samples. In this paper, we propose to learn jointly the projection matrix for dimensionality reduction and the discriminative dictionary for face representation. The joint learning makes the learned projection and dictionary better fit with each other so that a more effective face classification can be obtained. The proposed algorithm is evaluated on benchmark face databases in comparison with existing linear representation based methods, and the results show that the joint learning improves the FR rate, particularly when the number of training samples per class is small.
Sparse Representation Based Fisher Discrimination Dictionary Learning for Image Classification The employed dictionary plays an important role in sparse representation or sparse coding based image reconstruction and classification, while learning dictionaries from the training data has led to state-of-the-art results in image classification tasks. However, many dictionary learning models exploit only the discriminative information in either the representation coefficients or the representation residual, which limits their performance. In this paper we present a novel dictionary learning method based on the Fisher discrimination criterion. A structured dictionary, whose atoms have correspondences to the subject class labels, is learned, with which not only the representation residual can be used to distinguish different classes, but also the representation coefficients have small within-class scatter and big between-class scatter. The classification scheme associated with the proposed Fisher discrimination dictionary learning (FDDL) model is consequently presented by exploiting the discriminative information in both the representation residual and the representation coefficients. The proposed FDDL model is extensively evaluated on various image datasets, and it shows superior performance to many state-of-the-art dictionary learning methods in a variety of classification tasks.
Feature and Rank Level Fusion for Privacy Preserved Multi-Biometric System AbstractPrivacy protection in biometric system is a newly emerging biometric technology that can provide the protection against various attacks by intruders. In this paper, the authors have presented a multi-level of random projection method based on face and ear biometric traits. Privacy preserved templates are used in the proposed system. The main idea behind the privacy preserve computation is the random projection algorithm. Multiple random projection matrixes are used to generate multiple templates for biometric authentication. Newly introduced random fusion method is used in the proposed system; therefore, proposed method can provide better template security, privacy and feature quality. Multiple randomly fused templates are used for recognition purpose and finally decision fusion is applied to generate the final classification result. The proposed method works in a similar way human cognition for face recognition works, furthermore it preserve privacy and multimodality of the system.
Ear recognition using local binary patterns: A comparative experimental study. •A comparative study of ear recognition using local binary patterns variants is done.•A new texture operator is proposed and used as an ear feature descriptor.•Detailed analysis on Identification and verification is conducted separately.•An approximated recognition rate of 99% is achieved by some texture descriptors.•The study has significant insights and can benefit researchers in future works.
Online Palmprint Identification Biometrics-based personal identification is regarded as an effective method for automatically recognizing, with a high confidence, a person's identity. This paper presents a new biometric approach to online personal identification using palmprint technology. In contrast to the existing methods, our online palmprint identification system employs low-resolution palmprint images to achieve effective personal identification. The system consists of two parts: a novel device for online palmprint image acquisition and an efficient algorithm for fast palmprint recognition. A robust image coordinate system is defined to facilitate image alignment for feature extraction. In addition, a 2D Gabor phase encoding scheme is proposed for palmprint feature extraction and representation. The experimental results demonstrate the feasibility of the proposed system.
Improved ear verification after surgery - An approach based on collaborative representation of locally competitive features. •Presents a comprehensive study for biometric verification performance of ears before and after surgery.•Extensive study on different type of ear-surgery is presented along with a new public ear database.•Presents a new feature extraction technique based on Topographic Locally Competitive Algorithm.•Demonstrates superior verification performance on both normal ear database and surgically modified ear database.•Discussion on computational complexity and state-of-art performance.
Rich Feature Hierarchies for Accurate Object Detection and Semantic Segmentation Object detection performance, as measured on the canonical PASCAL VOC dataset, has plateaued in the last few years. The best-performing methods are complex ensemble systems that typically combine multiple low-level image features with high-level context. In this paper, we propose a simple and scalable detection algorithm that improves mean average precision (mAP) by more than 30% relative to the previous best result on VOC 2012 -- achieving a mAP of 53.3%. Our approach combines two key insights: (1) one can apply high-capacity convolutional neural networks (CNNs) to bottom-up region proposals in order to localize and segment objects and (2) when labeled training data is scarce, supervised pre-training for an auxiliary task, followed by domain-specific fine-tuning, yields a significant performance boost. Since we combine region proposals with CNNs, we call our method R-CNN: Regions with CNN features. We also present experiments that provide insight into what the network learns, revealing a rich hierarchy of image features. Source code for the complete system is available at http://www.cs.berkeley.edu/~rbg/rcnn.
Adaptive Learning-Based Task Offloading for Vehicular Edge Computing Systems. The vehicular edge computing system integrates the computing resources of vehicles, and provides computing services for other vehicles and pedestrians with task offloading. However, the vehicular task offloading environment is dynamic and uncertain, with fast varying network topologies, wireless channel states, and computing workloads. These uncertainties bring extra challenges to task offloading. In this paper, we consider the task offloading among vehicles, and propose a solution that enables vehicles to learn the offloading delay performance of their neighboring vehicles while offloading computation tasks. We design an adaptive learning based task offloading (ALTO) algorithm based on the multi-armed bandit theory, in order to minimize the average offloading delay. ALTO works in a distributed manner without requiring frequent state exchange, and is augmented with input-awareness and occurrence-awareness to adapt to the dynamic environment. The proposed algorithm is proved to have a sublinear learning regret. Extensive simulations are carried out under both synthetic scenario and realistic highway scenario, and results illustrate that the proposed algorithm achieves low delay performance, and decreases the average delay up to <inline-formula xmlns:mml="http://www.w3.org/1998/Math/MathML" xmlns:xlink="http://www.w3.org/1999/xlink"><tex-math notation="LaTeX">$30\%$</tex-math></inline-formula> compared with the existing upper confidence bound based learning algorithm.
Visual cryptography for general access structures A visual cryptography scheme for a set P of n participants is a method of encoding a secret image SI into n shadow images called shares, where each participant in P receives one share. Certain qualified subsets of participants can “visually” recover the secret image, but other, forbidden, sets of participants have no information (in an information-theoretic sense) on SI . A “visual” recovery for a set X ⊆ P consists of xeroxing the shares given to the participants in X onto transparencies, and then stacking them. The participants in a qualified set X will be able to see the secret image without any knowledge of cryptography and without performing any cryptographic computation. In this paper we propose two techniques for constructing visual cryptography schemes for general access structures. We analyze the structure of visual cryptography schemes and we prove bounds on the size of the shares distributed to the participants in the scheme. We provide a novel technique for realizing k out of n threshold visual cryptography schemes. Our construction for k out of n visual cryptography schemes is better with respect to pixel expansion than the one proposed by M. Naor and A. Shamir (Visual cryptography, in “Advances in Cryptology—Eurocrypt '94” CA. De Santis, Ed.), Lecture Notes in Computer Science, Vol. 950, pp. 1–12, Springer-Verlag, Berlin, 1995) and for the case of 2 out of n is the best possible. Finally, we consider graph-based access structures, i.e., access structures in which any qualified set of participants contains at least an edge of a given graph whose vertices represent the participants of the scheme.
Multi-stage genetic programming: A new strategy to nonlinear system modeling This paper presents a new multi-stage genetic programming (MSGP) strategy for modeling nonlinear systems. The proposed strategy is based on incorporating the individual effect of predictor variables and the interactions among them to provide more accurate simulations. According to the MSGP strategy, an efficient formulation for a problem comprises different terms. In the first stage of the MSGP-based analysis, the output variable is formulated in terms of an influencing variable. Thereafter, the error between the actual and the predicted value is formulated in terms of a new variable. Finally, the interaction term is derived by formulating the difference between the actual values and the values predicted by the individually developed terms. The capabilities of MSGP are illustrated by applying it to the formulation of different complex engineering problems. The problems analyzed herein include the following: (i) simulation of pH neutralization process, (ii) prediction of surface roughness in end milling, and (iii) classification of soil liquefaction conditions. The validity of the proposed strategy is confirmed by applying the derived models to the parts of the experimental results that were not included in the analyses. Further, the external validation of the models is verified using several statistical criteria recommended by other researchers. The MSGP-based solutions are capable of effectively simulating the nonlinear behavior of the investigated systems. The results of MSGP are found to be more accurate than those of standard GP and artificial neural network-based models.
Collaborative Mobile Charging The limited battery capacity of sensor nodes has become one of the most critical impediments that stunt the deployment of wireless sensor networks (WSNs). Recent breakthroughs in wireless energy transfer and rechargeable lithium batteries provide a promising alternative to power WSNs: mobile vehicles/robots carrying high volume batteries serve as mobile chargers to periodically deliver energy to sensor nodes. In this paper, we consider how to schedule multiple mobile chargers to optimize energy usage effectiveness, such that every sensor will not run out of energy. We introduce a novel charging paradigm, collaborative mobile charging, where mobile chargers are allowed to intentionally transfer energy between themselves. To provide some intuitive insights into the problem structure, we first consider a scenario that satisfies three conditions, and propose a scheduling algorithm, PushWait, which is proven to be optimal and can cover a one-dimensional WSN of infinite length. Then, we remove the conditions one by one, investigating chargers' scheduling in a series of scenarios ranging from the most restricted one to a general 2D WSN. Through theoretical analysis and simulations, we demonstrate the advantages of the proposed algorithms in energy usage effectiveness and charging coverage.
A blind medical image watermarking: DWT-SVD based robust and secure approach for telemedicine applications. In this paper, a blind image watermarking scheme based on discrete wavelet transform (DWT) and singular value decomposition (SVD) is proposed. In this scheme, DWT is applied on ROI (region of interest) of the medical image to get different frequency subbands of its wavelet decomposition. On the low frequency subband LL of the ROI, block-SVD is applied to get different singular matrices. A pair of elements with similar values is identified from the left singular value matrix of these selected blocks. The values of these pairs are modified using certain threshold to embed a bit of watermark content. Appropriate threshold is chosen to achieve the imperceptibility and robustness of medical image and watermark contents respectively. For authentication and identification of original medical image, one watermark image (logo) and other text watermark have been used. The watermark image provides authentication whereas the text data represents electronic patient record (EPR) for identification. At receiving end, blind recovery of both watermark contents is performed by a similar comparison scheme used during the embedding process. The proposed algorithm is applied on various groups of medical images like X-ray, CT scan and mammography. This scheme offers better visibility of watermarked image and recovery of watermark content due to DWT-SVD combination. Moreover, use of Hamming error correcting code (ECC) on EPR text bits reduces the BER and thus provides better recovery of EPR. The performance of proposed algorithm with EPR data coding by Hamming code is compared with the BCH error correcting code and it is found that later one perform better. A result analysis shows that imperceptibility of watermarked image is better as PSNR is above 43 dB and WPSNR is above 52 dB for all set of images. In addition, robustness of the scheme is better than existing scheme for similar set of medical images in terms of normalized correlation coefficient (NCC) and bit-error-rate (BER). An analysis is also carried out to verify the performance of the proposed scheme for different size of watermark contents (image and EPR data). It is observed from analysis that the proposed scheme is also appropriate for watermarking of color image. Using proposed scheme, watermark contents are extracted successfully under various noise attacks like JPEG compression, filtering, Gaussian noise, Salt and pepper noise, cropping, filtering and rotation. Performance comparison of proposed scheme with existing schemes shows proposed scheme has better robustness against different types of attacks. Moreover, the proposed scheme is also robust under set of benchmark attacks known as checkmark attacks.
Robot tutor and pupils’ educational ability: Teaching the times tables Research shows promising results of educational robots in language and STEM tasks. In language, more research is available, occasionally in view of individual differences in pupils’ educational ability levels, and learning seems to improve with more expressive robot behaviors. In STEM, variations in robots’ behaviors have been examined with inconclusive results and never while systematically investigating how differences in educational abilities match with different robot behaviors. We applied an autonomously tutoring robot (without tablet, partly WOz) in a 2 × 2 experiment of social vs. neutral behavior in above-average vs. below-average schoolchildren (N = 86; age 8–10 years) while rehearsing the multiplication tables on a one-to-one basis. The standard school test showed that on average, pupils significantly improved their performance even after 3 occasions of 5-min exercises. Beyond-average pupils profited most from a robot tutor, whereas those below average in multiplication benefited more from a robot that showed neutral rather than more social behavior.
1.2
0.2
0.2
0.2
0.2
0.1
0.066667
0
0
0
0
0
0
0
Variable universe fuzzy control for vehicle semi-active suspension system with MR damper combining fuzzy neural network and particle swarm optimization. This study proposes a novel variable universe fuzzy control design for vehicle semi-active suspension system with magnetorheological (MR) damper through the combination of fuzzy neural network (FNN) and particle swarm optimization (PSO). By constructing a quarter-vehicle test rig equipped with MR damper and then collecting the measured data, a non-parametric model of MR damper based on adaptive neuro-fuzzy inference system is first presented. And then a Takagi–Sugeno (T–S) fuzzy controller is designed to achieve the effective control of the input current in MR damper by using the contraction-expansion factors. Furthermore, an appropriate FNN controller is proposed to obtain the contraction-expansion factors, in which particle swarm optimization and back propagation are introduced as the learning and training algorithm for the FNN controller. Lastly, a simulation investigation is provided to validate the proposed control scheme. The results of this study can provide the technical foundation for the development of vehicle semi-active suspension system.
Mobile cloud computing: A survey Despite increasing usage of mobile computing, exploiting its full potential is difficult due to its inherent problems such as resource scarcity, frequent disconnections, and mobility. Mobile cloud computing can address these problems by executing mobile applications on resource providers external to the mobile device. In this paper, we provide an extensive survey of mobile cloud computing research, while highlighting the specific concerns in mobile cloud computing. We present a taxonomy based on the key issues in this area, and discuss the different approaches taken to tackle these issues. We conclude the paper with a critical analysis of challenges that have not yet been fully met, and highlight directions for future work.
Harmony search algorithm for solving Sudoku Harmony search (HS) algorithm was applied to solving Sudoku puzzle. The HS is an evolutionary algorithm which mimics musicians' behaviors such as random play, memory-based play, and pitch-adjusted play when they perform improvisation. Sudoku puzzles in this study were formulated as an optimization problem with number-uniqueness penalties. HS could successfully solve the optimization problem after 285 function evaluations, taking 9 seconds. Also, sensitivity analysis of HS parameters was performed to obtain a better idea of algorithm parameter values.
DEC: dynamically evolving clustering and its application to structure identification of evolving fuzzy models. Identification of models from input-output data essentially requires estimation of appropriate cluster centers. In this paper, a new online evolving clustering approach for streaming data is proposed. Unlike other approaches that consider either the data density or distance from existing cluster centers, this approach uses cluster weight and distance before generating new clusters. To capture the dynamics of the data stream, the cluster weight is defined in both data and time space in such a way that it decays exponentially with time. It also applies concepts from computational geometry to determine the neighborhood information while forming clusters. A distinction is made between core and noncore clusters to effectively identify the real outliers. The approach efficiently estimates cluster centers upon which evolving Takagi-Sugeno models are developed. The experimental results with developed models show that the proposed approach attains results at par or better than existing approaches and significantly reduces the computational overhead, which makes it suitable for real-time applications.
An Easily Understandable Grey Wolf Optimizer and Its Application to Fuzzy Controller Tuning. This paper proposes an easily understandable Grey Wolf Optimizer (GWO) applied to the optimal tuning of the parameters of Takagi-Sugeno proportional-integral fuzzy controllers (T-S PI-FCs). GWO is employed for solving optimization problems focused on the minimization of discrete-time objective functions defined as the weighted sum of the absolute value of the control error and of the squared output sensitivity function, and the vector variable consists of the tuning parameters of the T-S PI-FCs. Since the sensitivity functions are introduced with respect to the parametric variations of the process, solving these optimization problems is important as it leads to fuzzy control systems with a reduced process parametric sensitivity obtained by a GWO-based fuzzy controller tuning approach. GWO algorithms applied with this regard are formulated in easily understandable terms for both vector and scalar operations, and discussions on stability, convergence, and parameter settings are offered. The controlled processes referred to in the course of this paper belong to a family of nonlinear servo systems, which are modeled by second order dynamics plus a saturation and dead zone static nonlinearity. Experimental results concerning the angular position control of a laboratory servo system are included for validating the proposed method.
Evolutionary Wall-Following Hexapod Robot Using Advanced Multiobjective Continuous Ant Colony Optimized Fuzzy Controller. This paper proposes an evolutionary wall-following hexapod robot, where a new multiobjective evolutionary fuzzy control approach is proposed to control both walking orientation and speed of a hexapod robot for a wall-following task. According to the measurements of four distance sensors, a fuzzy controller (FC) controls the walking speed of the robot by changing the common swing angles of its six ...
Fuzzy Logic in Dynamic Parameter Adaptation of Harmony Search Optimization for Benchmark Functions and Fuzzy Controllers. Nowadays the use of fuzzy logic has been increasing in popularity, and this is mainly due to the inference mechanism that allows simulating human reasoning in knowledge-based systems. The main contribution of this work is using the concepts of fuzzy logic in a method for dynamically adapting the main parameters of the harmony search algorithm during execution. Dynamical adaptation of parameters in metaheuristics has been shown to improve performance and accuracy in a wide range of applications. For this reason, we propose and approach for fuzzy adaptation of parameters in harmony search. Two case studies are considered for testing the proposed approach, the optimization of mathematical functions, which are unimodal, multimodal, hybrid, and composite functions and a control problem without noise and when noise is considered. A statistical comparison between the harmony search algorithm and the fuzzy harmony search algorithm is presented to verify the advantages of the proposed approach.
Finite-Time Input-to-State Stability and Applications to Finite-Time Control Design This paper extends the well-known concept, Sontag's input-to-state stability (ISS), to finite-time control problems. In other words, a new concept, finite-time input-to-state stability (FTISS), is proposed and then is applied to both the analysis of finite-time stability and the design of finite-time stabilizing feedback laws of control systems. With finite-time stability, nonsmoothness has to be considered, and serious technical challenges arise in the design of finite-time controllers and the stability analysis of the closed-loop system. It is found that FTISS plays an important role as the conventional ISS in the context of asymptotic stability analysis and smooth feedback stabilization. Moreover, a robust adaptive controller is proposed to handle nonlinear systems with parametric and dynamic uncertainties by virtue of FTISS and related arguments.
Adam: A Method for Stochastic Optimization. We introduce Adam, an algorithm for first-order gradient-based optimization of stochastic objective functions, based on adaptive estimates of lower-order moments. The method is straightforward to implement, is computationally efficient, has little memory requirements, is invariant to diagonal rescaling of the gradients, and is well suited for problems that are large in terms of data and/or parameters. The method is also appropriate for non-stationary objectives and problems with very noisy and/or sparse gradients. The hyper-parameters have intuitive interpretations and typically require little tuning. Some connections to related algorithms, on which Adam was inspired, are discussed. We also analyze the theoretical convergence properties of the algorithm and provide a regret bound on the convergence rate that is comparable to the best known results under the online convex optimization framework. Empirical results demonstrate that Adam works well in practice and compares favorably to other stochastic optimization methods. Finally, we discuss AdaMax, a variant of Adam based on the infinity norm.
Multiple Lyapunov functions and other analysis tools for switched and hybrid systems In this paper, we introduce some analysis tools for switched and hybrid systems. We first present work on stability analysis. We introduce multiple Lyapunov functions as a tool for analyzing Lyapunov stability and use iterated function systems (IFS) theory as a tool for Lagrange stability. We also discuss the case where the switched systems are indexed by an arbitrary compact set. Finally, we extend Bendixson's theorem to the case of Lipschitz continuous vector fields, allowing limit cycle analysis of a class of "continuous switched" systems.
Learning to Predict Driver Route and Destination Intent For many people, driving is a routine activity where people drive to the same destinations using the same routes on a regular basis. Many drivers, for example, will drive to and from work along a small set of routes, at about the same time every day of the working week. Similarly, although a person may shop on different days or at different times, they will often visit the same grocery store(s). In this paper, we present a novel approach to predicting driver intent that exploits the predictable nature of everyday driving. Our approach predicts a driver's intended route and destination through the use of a probabilistic model learned from observation of their driving habits. We show that by using a low-cost GPS sensor and a map database, it is possible to build a hidden Markov model (HMM) of the routes and destinations used by the driver. Furthermore, we show that this model can be used to make accurate predictions of the driver's destination and route through on-line observation of their GPS position during the trip. We present a thorough evaluation of our approach using a corpus of almost a month of real, everyday driving. Our results demonstrate the effectiveness of the approach, achieving approximately 98% accuracy in most cases. Such high performance suggests that the method can be harnessed for improved safety monitoring, route planning taking into account traffic density, and better trip duration prediction
Software-Defined Networking: A Comprehensive Survey The Internet has led to the creation of a digital society, where (almost) everything is connected and is accessible from anywhere. However, despite their widespread adoption, traditional IP networks are complex and very hard to manage. It is both difficult to configure the network according to predefined policies, and to reconfigure it to respond to faults, load, and changes. To make matters even more difficult, current networks are also vertically integrated: the control and data planes are bundled together. Software-defined networking (SDN) is an emerging paradigm that promises to change this state of affairs, by breaking vertical integration, separating the network's control logic from the underlying routers and switches, promoting (logical) centralization of network control, and introducing the ability to program the network. The separation of concerns, introduced between the definition of network policies, their implementation in switching hardware, and the forwarding of traffic, is key to the desired flexibility: by breaking the network control problem into tractable pieces, SDN makes it easier to create and introduce new abstractions in networking, simplifying network management and facilitating network evolution. In this paper, we present a comprehensive survey on SDN. We start by introducing the motivation for SDN, explain its main concepts and how it differs from traditional networking, its roots, and the standardization activities regarding this novel paradigm. Next, we present the key building blocks of an SDN infrastructure using a bottom-up, layered approach. We provide an in-depth analysis of the hardware infrastructure, southbound and northbound application programming interfaces (APIs), network virtualization layers, network operating systems (SDN controllers), network programming languages, and network applications. We also look at cross-layer problems such as debugging and troubleshooting. In an effort to anticipate the future evolution of this - ew paradigm, we discuss the main ongoing research efforts and challenges of SDN. In particular, we address the design of switches and control platforms - with a focus on aspects such as resiliency, scalability, performance, security, and dependability - as well as new opportunities for carrier transport networks and cloud providers. Last but not least, we analyze the position of SDN as a key enabler of a software-defined environment.
Deep Learning in Mobile and Wireless Networking: A Survey. The rapid uptake of mobile devices and the rising popularity of mobile applications and services pose unprecedented demands on mobile and wireless networking infrastructure. Upcoming 5G systems are evolving to support exploding mobile traffic volumes, real-time extraction of fine-grained analytics, and agile management of network resources, so as to maximize user experience. Fulfilling these tasks is challenging, as mobile environments are increasingly complex, heterogeneous, and evolving. One potential solution is to resort to advanced machine learning techniques, in order to help manage the rise in data volumes and algorithm-driven applications. The recent success of deep learning underpins new and powerful tools that tackle problems in this space. In this paper we bridge the gap between deep learning and mobile and wireless networking research, by presenting a comprehensive survey of the crossovers between the two areas. We first briefly introduce essential background and state-of-the-art in deep learning techniques with potential applications to networking. We then discuss several techniques and platforms that facilitate the efficient deployment of deep learning onto mobile systems. Subsequently, we provide an encyclopedic review of mobile and wireless networking research based on deep learning, which we categorize by different domains. Drawing from our experience, we discuss how to tailor deep learning to mobile environments. We complete this survey by pinpointing current challenges and open future directions for research.
Attitudes Towards Social Robots In Education: Enthusiast, Practical, Troubled, Sceptic, And Mindfully Positive While social robots bring new opportunities for education, they also come with moral challenges. Therefore, there is a need for moral guidelines for the responsible implementation of these robots. When developing such guidelines, it is important to include different stakeholder perspectives. Existing (qualitative) studies regarding these perspectives however mainly focus on single stakeholders. In this exploratory study, we examine and compare the attitudes of multiple stakeholders on the use of social robots in primary education, using a novel questionnaire that covers various aspects of moral issues mentioned in earlier studies. Furthermore, we also group the stakeholders based on similarities in attitudes and examine which socio-demographic characteristics influence these attitude types. Based on the results, we identify five distinct attitude profiles and show that the probability of belonging to a specific profile is affected by such characteristics as stakeholder type, age, education and income. Our results also indicate that social robots have the potential to be implemented in education in a morally responsible way that takes into account the attitudes of various stakeholders, although there are multiple moral issues that need to be addressed first. Finally, we present seven (practical) implications for a responsible application of social robots in education following from our results. These implications provide valuable insights into how social robots should be implemented.
1.24
0.24
0.24
0.24
0.24
0.24
0.12
0.013333
0
0
0
0
0
0
M-LionWhale: multi-objective optimisation model for secure routing in mobile ad-hoc network. Mobile ad-hoc network (MANET) is an emerging technology that comes under the category of wireless network. Even though the network assumes that all its mobile nodes are trusted, it is impossible in the real world as few nodes may be malicious. Therefore, it is essential to put forward a mechanism that can provide security by selecting an optimal route for data forwarding. In this study, a goal pro...
MOSOA: A new multi-objective seagull optimization algorithm •A novel Multi-objective Seagull Optimization Algorithm is proposed.•The algorithm is tested on 24 real challenging benchmark test function.•The results show the superior convergence behaviour of proposed algorithm.•The results on engineering design problems prove its efficiency and applicability.
Intelligent Ad-Hoc-On Demand Multipath Distance Vector for Wormhole Attack in Clustered WSN In Wireless Senor Networks, security is the most significant issue when sending such an essential message via wireless connection. This helps attackers to access the network and execute several potential attacks to intercept or modify real data/information. Because network sensors do not have routers, the same routing protocol must be split by all the nodes participating in the network to assist each other with packet transmission. In complex topology, its unguided existence often renders it open to all forms of protection attack, presenting a degree of security difficulties. Wormhole is a popular illustration of attacks, due to its difficulties in detecting and stopping, poses the greatest danger. A new routing technique is being presented in this paper which works towards ensuring the secured path for the data transmission. In the research work wormhole type of attack is being considered and the technique works towards the detection and prevention of the defined type of attack. The proposed methodology is validated based on certain performance related parameters for WSN as energy efficiency, delay from end to end, throughput, delivery ratio of packets. The generated outcomes are compared with some recent techniques over the same domain for the efficiency and the presented work has proven to be best among the described techniques for considered parameters. The methodology defined simulated using NS2 for various performance related parameters like energy efficiency, packet loss, throughput, etc.
Proactive fault-tolerant wireless mesh networks for mission-critical control systems Although wireless networks are becoming a fundamental infrastructure for various control applications, they are inherently exposed to network faults such as lossy links and node failures in environments such as mining, outdoor monitoring, and chemical process control. In this paper, we propose a proactive fault-tolerant mechanism to protect the wireless network against temporal faults without any explicit network state information for mission-critical control systems. Specifically, the proposed mechanism optimizes the multiple routing paths, link scheduling, and traffic generation rate such that it meets the control stability demands even if it experiences multiple link faults and node faults. The proactive network relies on a constrained optimization problem, where the objective function is the network robustness, and the main constraints are the set of the traffic demand, link, and routing layer requirements. To analyze the robustness, we propose a novel performance metric called stability margin ratio, based on the network performance and the stability boundary. Our numerical and experimental performance evaluation shows that the traffic generation rate and the delay of wireless networks are found as critical as the network reliability to guarantee the stability of control systems. Furthermore, the proposed proactive network provides more robust performance than practical state-of-the-art solutions while maintaining high energy efficiency.
Energy-efficient and balanced routing in low-power wireless sensor networks for data collection Cost-based routing protocols are the main approach used in practical wireless sensor network (WSN) and Internet of Things (IoT) deployments for data collection applications with energy constraints; however, those routing protocols lead to the concentration of most of the data traffic on some specific nodes which provide the best available routes, thus significantly increasing their energy consumption. Consequently, nodes providing the best routes are potentially the first ones to deplete their batteries and stop working. In this paper, we introduce a novel routing strategy for energy efficient and balanced data collection in WSNs/IoT, which can be applied to any cost-based routing solution to exploit suboptimal network routing alternatives based on the parent set concept. While still taking advantage of the stable routing topologies built in cost-based routing protocols, our approach adds a random component into the process of packet forwarding to achieve a better network lifetime in WSNs. We evaluate the implementation of our approach against other state-of-the-art WSN routing protocols through thorough real-world testbed experiments and simulations, and demonstrate that our approach achieves a significant reduction in the energy consumption of the routing layer in the busiest nodes ranging from 11% to 59%, while maintaining over 99% reliability. Furthermore, we conduct the field deployment of our approach in a heterogeneous WSN for environmental monitoring in a forest area, report the experimental results and illustrate the effectiveness of our approach in detail. Our EER based routing protocol CTP+EER is made available as open source to the community for evaluation and adoption.
Computational thinking Summary form only given. My vision for the 21st century, Computational Thinking, will be a fundamental skill used by everyone in the world. To reading, writing, and arithmetic, we should add computational thinking to every child's analytical ability. Computational thinking involves solving problems, designing systems, and understanding human behavior by drawing on the concepts fundamental to computer science. Thinking like a computer scientist means more than being able to program a computer. It requires the ability to abstract and thus to think at multiple levels of abstraction. In this talk I will give many examples of computational thinking, argue that it has already influenced other disciplines, and promote the idea that teaching computational thinking can not only inspire future generations to enter the field of computer science but benefit people in all fields.
JPEG Error Analysis and Its Applications to Digital Image Forensics JPEG is one of the most extensively used image formats. Understanding the inherent characteristics of JPEG may play a useful role in digital image forensics. In this paper, we introduce JPEG error analysis to the study of image forensics. The main errors of JPEG include quantization, rounding, and truncation errors. Through theoretically analyzing the effects of these errors on single and double JPEG compression, we have developed three novel schemes for image forensics including identifying whether a bitmap image has previously been JPEG compressed, estimating the quantization steps of a JPEG image, and detecting the quantization table of a JPEG image. Extensive experimental results show that our new methods significantly outperform existing techniques especially for the images of small sizes. We also show that the new method can reliably detect JPEG image blocks which are as small as 8 × 8 pixels and compressed with quality factors as high as 98. This performance is important for analyzing and locating small tampered regions within a composite image.
Highly dynamic Destination-Sequenced Distance-Vector routing (DSDV) for mobile computers An ad-hoc network is the cooperative engagement of a collection of Mobile Hosts without the required intervention of any centralized Access Point. In this paper we present an innovative design for the operation of such ad-hoc networks. The basic idea of the design is to operate each Mobile Host as a specialized router, which periodically advertises its view of the interconnection topology with other Mobile Hosts within the network. This amounts to a new sort of routing protocol. We have investigated modifications to the basic Bellman-Ford routing mechanisms, as specified by RIP [5], to make it suitable for a dynamic and self-starting network mechanism as is required by users wishing to utilize ad hoc networks. Our modifications address some of the previous objections to the use of Bellman-Ford, related to the poor looping properties of such algorithms in the face of broken links and the resulting time dependent nature of the interconnection topology describing the links between the Mobile Hosts. Finally, we describe the ways in which the basic network-layer routing can be modified to provide MAC-layer support for ad-hoc networks.
The FERET Evaluation Methodology for Face-Recognition Algorithms Two of the most critical requirements in support of producing reliable face-recognition systems are a large database of facial images and a testing procedure to evaluate systems. The Face Recognition Technology (FERET) program has addressed both issues through the FERET database of facial images and the establishment of the FERET tests. To date, 14,126 images from 1,199 individuals are included in the FERET database, which is divided into development and sequestered portions of the database. In September 1996, the FERET program administered the third in a series of FERET face-recognition tests. The primary objectives of the third test were to 1) assess the state of the art, 2) identify future areas of research, and 3) measure algorithm performance.
Neural fitted q iteration – first experiences with a data efficient neural reinforcement learning method This paper introduces NFQ, an algorithm for efficient and effective training of a Q-value function represented by a multi-layer perceptron. Based on the principle of storing and reusing transition experiences, a model-free, neural network based Reinforcement Learning algorithm is proposed. The method is evaluated on three benchmark problems. It is shown empirically, that reasonably few interactions with the plant are needed to generate control policies of high quality.
Labels and event processes in the Asbestos operating system Asbestos, a new operating system, provides novel labeling and isolation mechanisms that help contain the effects of exploitable software flaws. Applications can express a wide range of policies with Asbestos's kernel-enforced labels, including controls on interprocess communication and system-wide information flow. A new event process abstraction defines lightweight, isolated contexts within a single process, allowing one process to act on behalf of multiple users while preventing it from leaking any single user's data to others. A Web server demonstration application uses these primitives to isolate private user data. Since the untrusted workers that respond to client requests are constrained by labels, exploited workers cannot directly expose user data except as allowed by application policy. The server application requires 1.4 memory pages per user for up to 145,000 users and achieves connection rates similar to Apache, demonstrating that additional security can come at an acceptable cost.
Switching Stabilization for a Class of Slowly Switched Systems In this technical note, the problem of switching stabilization for slowly switched linear systems is investigated. In particular, the considered systems can be composed of all unstable subsystems. Based on the invariant subspace theory, the switching signal with mode-dependent average dwell time (MDADT) property is designed to exponentially stabilize the underlying system. Furthermore, sufficient condition of stabilization for switched systems with all stable subsystems under MDADT switching is also given. The correctness and effectiveness of the proposed approaches are illustrated by a numerical example.
Automated Detection of Obstructive Sleep Apnea Events from a Single-Lead Electrocardiogram Using a Convolutional Neural Network. In this study, we propose a method for the automated detection of obstructive sleep apnea (OSA) from a single-lead electrocardiogram (ECG) using a convolutional neural network (CNN). A CNN model was designed with six optimized convolution layers including activation, pooling, and dropout layers. One-dimensional (1D) convolution, rectified linear units (ReLU), and max pooling were applied to the convolution, activation, and pooling layers, respectively. For training and evaluation of the CNN model, a single-lead ECG dataset was collected from 82 subjects with OSA and was divided into training (including data from 63 patients with 34,281 events) and testing (including data from 19 patients with 8571 events) datasets. Using this CNN model, a precision of 0.99%, a recall of 0.99%, and an F-score of 0.99% were attained with the training dataset; these values were all 0.96% when the CNN was applied to the testing dataset. These results show that the proposed CNN model can be used to detect OSA accurately on the basis of a single-lead ECG. Ultimately, this CNN model may be used as a screening tool for those suspected to suffer from OSA.
Convert Harm Into Benefit: A Coordination-Learning Based Dynamic Spectrum Anti-Jamming Approach This paper mainly investigates the multi-user anti-jamming spectrum access problem. Using the idea of “converting harm into benefit,” the malicious jamming signals projected by the enemy are utilized by the users as the coordination signals to guide spectrum coordination. An “internal coordination-external confrontation” multi-user anti-jamming access game model is constructed, and the existence of Nash equilibrium (NE) as well as correlated equilibrium (CE) is demonstrated. A coordination-learning based anti-jamming spectrum access algorithm (CLASA) is designed to achieve the CE of the game. Simulation results show the convergence, and effectiveness of the proposed CLASA algorithm, and indicate that our approach can help users confront the malicious jammer, and coordinate internal spectrum access simultaneously without information exchange. Last but not least, the fairness of the proposed approach under different jamming attack patterns is analyzed, which illustrates that this approach provides fair anti-jamming spectrum access opportunities under complicated jamming pattern.
1.2
0.2
0.2
0.2
0.2
0
0
0
0
0
0
0
0
0
Efficient and Secure Service-Oriented Authentication Supporting Network Slicing for 5G-Enabled IoT. 5G network is considered as a key enabler in meeting continuously increasing demands for the future Internet of Things (IoT) services, including high data rate, numerous devices connection, and low service latency. To satisfy these demands, network slicing and fog computing have been envisioned as the promising solutions in service-oriented 5G architecture. However, security paradigms enabling aut...
Privacy Enabled Digital Rights Management Without Trusted Third Party Assumption Digital rights management systems are required to provide security and accountability without violating the privacy of the entities involved. However, achieving privacy along with accountability in the same framework is hard as these attributes are mutually contradictory. Thus, most of the current digital rights management systems rely on trusted third parties to provide privacy to the entities involved. However, a trusted third party can become malicious and break the privacy protection of the entities in the system. Hence, in this paper, we propose a novel privacy preserving content distribution mechanism for digital rights management without relying on the trusted third party assumption. We use simple primitives such as blind decryption and one way hash chain to avoid the trusted third party assumption. We prove that our scheme is not prone to the “oracle problem” of the blind decryption mechanism. The proposed mechanism supports access control without degrading user's privacy as well as allows revocation of even malicious users without violating their privacy.
An efficient conditionally anonymous ring signature in the random oracle model A conditionally anonymous ring signature is an exception since the anonymity is conditional. Specifically, it allows an entity to confirm/refute the signature that he generated before. A group signature also shares the same property since a group manager can revoke a signer's anonymity using the trapdoor information. However, the special node (i.e., group manager) does not exist in the group in order to satisfy the ad hoc fashion. In this paper, we construct a new conditionally anonymous ring signature, in which the actual signer can be traced without the help of the group manager. The big advantage of the confirmation and disavowal protocols designed by us are non-interactive with constant costs while the known schemes suffer from the linear cost in terms of the ring size n or security parameter s.
Threats to Networking Cloud and Edge Datacenters in the Internet of Things. Several application domains are collecting data using Internet of Things sensing devices and shipping it to remote cloud datacenters for analysis (fusion, storage, and processing). Data analytics activities raise a new set of technical challenges from the perspective of ensuring end-to-end security and privacy of data as it travels from an edge datacenter (EDC) to a cloud datacenter (CDC) (or vice...
SAFE: Secure Appliance Scheduling for Flexible and Efficient Energy Consumption for Smart Home IoT Smart homes (SHs) aim at forming an energy optimized environment that can efficiently regulate the use of various Internet of Things (IoT) devices in its network. Real-time electricity pricing models along with SHs provide users an opportunity to reduce their electricity expenditure by responding to the pricing that varies with different times of the day, resulting in reducing the expenditure at both customers’ and utility provider’s end. However, responding to such prices and effectively scheduling the appliances under such complex dynamics is a challenging optimization problem to be solved by the provider or by third party services. As communication in SH-IoT environment is extremely sensitive and private, reporting of such usage information to the provider to solve the optimization has a potential risk that the provider or third party services may track users’ energy consumption profile which compromises users’ privacy. To address these issues, we developed a homomorphic encryption-based alternating direction method of multipliers approach to solve the cost-aware appliance scheduling optimization in a distributed manner and schedule home appliances without leaking users’ privacy. Through extensive simulation study considering real-world datasets, we show that the proposed secure appliance scheduling for flexible and efficient energy consumption scheme, namely SAFE, effectively lowers electricity cost while preserving users’ privacy.
A Blockchain-Based Scheme For Privacy-Preserving And Secure Sharing Of Medical Data How to alleviate the contradiction between the patient's privacy and the research or com-mercial demands of health data has become the challenging problem of intelligent medical system with the exponential increase of medical data. In this paper, a blockchainbased privacy-preserving scheme is proposed, which realizes secure sharing of medical data between several entities involved patients, research institutions and semi-trusted cloud servers. And meanwhile, it achieves the data availability and consistency between patients and research institutions, where zero-knowledge proof is employed to verify whether the patient's medical data meets the specific requirements proposed by research institutions without revealing patients' privacy, and then the proxy re-encryption technology is adopted to ensure that research institutions can decrypt the intermediary ciphertext. In addition, this proposal can execute distributed consensus based on PBFT algorithm for transactions between patients and research institutions according to the prearranged terms. Theoretical analysis shows the proposed scheme can satisfy security and privacy requirements such as confidentiality, integrity and availability, as well as performance evaluation demonstrates it is feasible and efficient in contrast with other typical schemes. (C) 2020 Elsevier Ltd. All rights reserved.
Chaos-Based Content Distribution Framework for Digital Rights Management System Multimedia contents are digitally utilized these days. Thus, the development of an effective method to access the content is becoming the topmost priority of the entertainment industry to protect the digital content from unauthorized access. Digital rights management (DRM) systems are the technique that makes digital content accessible only to the legal rights holders. As the Internet of Things environment is used in the distribution and access of digital content, a secure and efficient content delivery mechanism is also required. Keeping the focus on these points, this article proposes a content distribution framework for DRM system using chaotic map. Formal security verification under the random oracle model, which uncovers the proposed protocol's capability to resist the critical attacks is given. Moreover, simulation study for security verification is performed using the broadly accepted “automated validation of Internet security protocols and applications,” which indicates that the protocol is safe. Moreover, the detailed comparative study with related protocols demonstrates that it provides better security and improves the computational and communication efficiency.
Constrained Kalman filtering for indoor localization of transport vehicles using floor-installed HF RFID transponders Localization of transport vehicles is an important issue for many intralogistics applications. The paper presents an inexpensive solution for indoor localization of vehicles. Global localization is realized by detection of RFID transponders, which are integrated in the floor. The paper presents a novel algorithm for fusing RFID readings with odometry using Constraint Kalman filtering. The paper presents experimental results with a Mecanum based omnidirectional vehicle on a NaviFloor® installation, which includes passive HF RFID transponders. The experiments show that the proposed Constraint Kalman filter provides a similar localization accuracy compared to a Particle filter but with much lower computational expense.
Constrained Multiobjective Optimization for IoT-Enabled Computation Offloading in Collaborative Edge and Cloud Computing Internet-of-Things (IoT) applications are becoming more resource-hungry and latency-sensitive, which are severely constrained by limited resources of current mobile hardware. Mobile cloud computing (MCC) can provide abundant computation resources, while mobile-edge computing (MEC) aims to reduce the transmission latency by offloading complex tasks from IoT devices to nearby edge servers. It is sti...
MOEA/D: A Multiobjective Evolutionary Algorithm Based on Decomposition Decomposition is a basic strategy in traditional multiobjective optimization. However, it has not yet been widely used in multiobjective evolutionary optimization. This paper proposes a multiobjective evolutionary algorithm based on decomposition (MOEA/D). It decomposes a multiobjective optimization problem into a number of scalar optimization subproblems and optimizes them simultaneously. Each subproblem is optimized by only using information from its several neighboring subproblems, which makes MOEA/D have lower computational complexity at each generation than MOGLS and nondominated sorting genetic algorithm II (NSGA-II). Experimental results have demonstrated that MOEA/D with simple decomposition methods outperforms or performs similarly to MOGLS and NSGA-II on multiobjective 0-1 knapsack problems and continuous multiobjective optimization problems. It has been shown that MOEA/D using objective normalization can deal with disparately-scaled objectives, and MOEA/D with an advanced decomposition method can generate a set of very evenly distributed solutions for 3-objective test instances. The ability of MOEA/D with small population, the scalability and sensitivity of MOEA/D have also been experimentally investigated in this paper.
A Model for Understanding How Virtual Reality Aids Complex Conceptual Learning Designers and evaluators of immersive virtual reality systems have many ideas concerning how virtual reality can facilitate learning. However, we have little information concerning which of virtual reality's features provide the most leverage for enhancing understanding or how to customize those affordances for different learning environments. In part, this reflects the truly complex nature of learning. Features of a learning environment do not act in isolation; other factors such as the concepts or skills to be learned, individual characteristics, the learning experience, and the interaction experience all play a role in shaping the learning process and its outcomes. Through Project Science Space, we have been trying to identify, use, and evaluate immersive virtual reality's affordances as a means to facilitate the mastery of complex, abstract concepts. In doing so, we are beginning to understand the interplay between virtual reality's features and other important factors in shaping the learning process and learning outcomes for this type of material. In this paper, we present a general model that describes how we think these factors work together and discuss some of the lessons we are learning about virtual reality's affordances in the context of this model for complex conceptual learning.
Solving the data sparsity problem in destination prediction Destination prediction is an essential task for many emerging location-based applications such as recommending sightseeing places and targeted advertising according to destinations. A common approach to destination prediction is to derive the probability of a location being the destination based on historical trajectories. However, almost all the existing techniques use various kinds of extra information such as road network, proprietary travel planner, statistics requested from government, and personal driving habits. Such extra information, in most circumstances, is unavailable or very costly to obtain. Thereby we approach the task of destination prediction by using only historical trajectory dataset. However, this approach encounters the \"data sparsity problem\", i.e., the available historical trajectories are far from enough to cover all possible query trajectories, which considerably limits the number of query trajectories that can obtain predicted destinations. We propose a novel method named Sub-Trajectory Synthesis (SubSyn) to address the data sparsity problem. SubSyn first decomposes historical trajectories into sub-trajectories comprising two adjacent locations, and then connects the sub-trajectories into \"synthesised\" trajectories. This process effectively expands the historical trajectory dataset to contain much more trajectories. Experiments based on real datasets show that SubSyn can predict destinations for up to ten times more query trajectories than a baseline prediction algorithm. Furthermore, the running time of the SubSyn-training algorithm is almost negligible for a large set of 1.9 million trajectories, and the SubSyn-prediction algorithm runs over two orders of magnitude faster than the baseline prediction algorithm constantly.
Design of robust fuzzy fault detection filter for polynomial fuzzy systems with new finite frequency specifications This paper investigates the problem of fault detection filter design for discrete-time polynomial fuzzy systems with faults and unknown disturbances. The frequency ranges of the faults and the disturbances are assumed to be known beforehand and to reside in low, middle or high frequency ranges. Thus, the proposed filter is designed in the finite frequency range to overcome the conservatism generated by those designed in the full frequency domain. Being of polynomial fuzzy structure, the proposed filter combines the H−/H∞ performances in order to ensure the best robustness to the disturbance and the best sensitivity to the fault. Design conditions are derived in Sum Of Squares formulations that can be easily solved via available software tools. Two illustrative examples are introduced to demonstrate the effectiveness of the proposed method and a comparative study with LMI method is also provided.
Intention-detection strategies for upper limb exosuits: model-based myoelectric vs dynamic-based control The cognitive human-robot interaction between an exosuit and its wearer plays a key role in determining both the biomechanical effects of the device on movements and its perceived effectiveness. There is a lack of evidence, however, on the comparative performance of different control methods, implemented on the same device. Here, we compare two different control approaches on the same robotic suit: a model-based myoelectric control (myoprocessor), which estimates the joint torque from the activation of target muscles, and a dynamic-based control that provides support against gravity using an inverse dynamic model. Tested on a cohort of four healthy participants, assistance from the exosuit results in a marked reduction in the effort of muscles working against gravity with both control approaches (peak reduction of 68.6±18.8%, for the dynamic arm model and 62.4±25.1% for the myoprocessor), when compared to an unpowered condition. Neither of the two controllers had an affect on the performance of their users in a joint-angle tracking task (peak errors of 15.4° and 16.4° for the dynamic arm model and myoprocessor, respectively, compared to 13.1o in the unpowered condition). However, our results highlight the remarkable adaptability of the myoprocessor to seamlessly adapt to changing external dynamics.
1.2
0.2
0.2
0.2
0.2
0.2
0.1
0
0
0
0
0
0
0
Understanding Back-Translation at Scale. An effective method to improve neural machine translation with monolingual data is to augment the parallel training corpus with back-translations of target language sentences. This work broadens the understanding of back-translation and investigates a number of methods to generate synthetic source sentences. We find that in all but resource poor settings back-translations obtained via sampling or noised beam outputs are most effective. Our analysis shows that sampling or noisy synthetic data gives a much stronger training signal than data generated by beam or greedy search. We also compare how synthetic data compares to genuine bitext and study various domain effects. Finally, we scale to hundreds of millions of monolingual sentences and achieve a new state of the art of 35 BLEU on the WMTu002714 English-German test set.
NLTK: the natural language toolkit The Natural Language Toolkit is a suite of program modules, data sets, tutorials and exercises, covering symbolic and statistical natural language processing. NLTK is written in Python and distributed under the GPL open source license. Over the past three years, NLTK has become popular in teaching and research. We describe the toolkit and report on its current state of development.
WP:clubhouse?: an exploration of Wikipedia's gender imbalance Wikipedia has rapidly become an invaluable destination for millions of information-seeking users. However, media reports suggest an important challenge: only a small fraction of Wikipedia's legion of volunteer editors are female. In the current work, we present a scientific exploration of the gender imbalance in the English Wikipedia's population of editors. We look at the nature of the imbalance itself, its effects on the quality of the encyclopedia, and several conflict-related factors that may be contributing to the gender gap. Our findings confirm the presence of a large gender gap among editors and a corresponding gender-oriented disparity in the content of Wikipedia's articles. Further, we find evidence hinting at a culture that may be resistant to female participation.
Word Emotion Induction for Multiple Languages as a Deep Multi-Task Learning Problem.
Amazon mechanical turk: Gold mine or coal mine?
Gender and Dialect Bias in YouTube's Automatic Captions.
A standalone RFID Indoor Positioning System Using Passive Tags Indoor positioning systems (IPSs) locate objects in closed structures such as office buildings, hospitals, stores, factories, and warehouses, where Global Positioning System devices generally do not work. Most available systems apply wireless concepts, optical tracking, and/or ultrasound. This paper presents a standalone IPS using radio frequency identification (RFID) technology. The concept is ba...
Model-based periodic event-triggered control for linear systems Periodic event-triggered control (PETC) is a control strategy that combines ideas from conventional periodic sampled-data control and event-triggered control. By communicating periodically sampled sensor and controller data only when needed to guarantee stability or performance properties, PETC is capable of reducing the number of transmissions significantly, while still retaining a satisfactory closed-loop behavior. In this paper, we will study observer-based controllers for linear systems and propose advanced event-triggering mechanisms (ETMs) that will reduce communication in both the sensor-to-controller channels and the controller-to-actuator channels. By exploiting model-based computations, the new classes of ETMs will outperform existing ETMs in the literature. To model and analyze the proposed classes of ETMs, we present two frameworks based on perturbed linear and piecewise linear systems, leading to conditions for global exponential stability and @?"2-gain performance of the resulting closed-loop systems in terms of linear matrix inequalities. The proposed analysis frameworks can be used to make tradeoffs between the network utilization on the one hand and the performance in terms of @?"2-gains on the other. In addition, we will show that the closed-loop performance realized by an observer-based controller, implemented in a conventional periodic time-triggered fashion, can be recovered arbitrarily closely by a PETC implementation. This provides a justification for emulation-based design. Next to centralized model-based ETMs, we will also provide a decentralized setup suitable for large-scale systems, where sensors and actuators are physically distributed over a wide area. The improvements realized by the proposed model-based ETMs will be demonstrated using numerical examples.
Hierarchical mesh segmentation based on fitting primitives In this paper, we describe a hierarchical face clustering algorithm for triangle meshes based on fitting primitives belonging to an arbitrary set. The method proposed is completely automatic, and generates a binary tree of clusters, each of which is fitted by one of the primitives employed. Initially, each triangle represents a single cluster; at every iteration, all the pairs of adjacent clusters are considered, and the one that can be better approximated by one of the primitives forms a new single cluster. The approximation error is evaluated using the same metric for all the primitives, so that it makes sense to choose which is the most suitable primitive to approximate the set of triangles in a cluster.Based on this approach, we have implemented a prototype that uses planes, spheres and cylinders, and have experimented that for meshes made of 100 K faces, the whole binary tree of clusters can be built in about 8 s on a standard PC.The framework described here has natural application in reverse engineering processes, but it has also been tested for surface denoising, feature recovery and character skinning.
2 Algorithms For Constructing A Delaunay Triangulation This paper provides a unified discussion of the Delaunay triangulation. Its geometric properties are reviewed and several applications are discussed. Two algorithms are presented for constructing the triangulation over a planar set ofN points. The first algorithm uses a divide-and-conquer approach. It runs inO(N logN) time, which is asymptotically optimal. The second algorithm is iterative and requiresO(N2) time in the worst case. However, its average case performance is comparable to that of the first algorithm.
Design, Implementation, and Experimental Results of a Quaternion-Based Kalman Filter for Human Body Motion Tracking Real-time tracking of human body motion is an important technology in synthetic environments, robotics, and other human-computer interaction applications. This paper presents an extended Kalman filter designed for real-time estimation of the orientation of human limb segments. The filter processes data from small inertial/magnetic sensor modules containing triaxial angular rate sensors, accelerometers, and magnetometers. The filter represents rotation using quaternions rather than Euler angles or axis/angle pairs. Preprocessing of the acceleration and magnetometer measurements using the Quest algorithm produces a computed quaternion input for the filter. This preprocessing reduces the dimension of the state vector and makes the measurement equations linear. Real-time implementation and testing results of the quaternion-based Kalman filter are presented. Experimental results validate the filter design, and show the feasibility of using inertial/magnetic sensor modules for real-time human body motion tracking
Reinforcement Q-learning for optimal tracking control of linear discrete-time systems with unknown dynamics. In this paper, a novel approach based on the Q-learning algorithm is proposed to solve the infinite-horizon linear quadratic tracker (LQT) for unknown discrete-time systems in a causal manner. It is assumed that the reference trajectory is generated by a linear command generator system. An augmented system composed of the original system and the command generator is constructed and it is shown that the value function for the LQT is quadratic in terms of the state of the augmented system. Using the quadratic structure of the value function, a Bellman equation and an augmented algebraic Riccati equation (ARE) for solving the LQT are derived. In contrast to the standard solution of the LQT, which requires the solution of an ARE and a noncausal difference equation simultaneously, in the proposed method the optimal control input is obtained by only solving an augmented ARE. A Q-learning algorithm is developed to solve online the augmented ARE without any knowledge about the system dynamics or the command generator. Convergence to the optimal solution is shown. A simulation example is used to verify the effectiveness of the proposed control scheme.
A multi-objective and PSO based energy efficient path design for mobile sink in wireless sensor networks. Data collection through mobile sink (MS) in wireless sensor networks (WSNs) is an effective solution to the hot-spot or sink-hole problem caused by multi-hop routing using the static sink. Rendezvous point (RP) based MS path design is a common and popular technique used in this regard. However, design of the optimal path is a well-known NP-hard problem. Therefore, an evolutionary approach like multi-objective particle swarm optimization (MOPSO) can prove to be a very promising and reasonable approach to solve the same. In this paper, we first present a Linear Programming formulation for the stated problem and then, propose an MOPSO-based algorithm to design an energy efficient trajectory for the MS. The algorithm is presented with an efficient particle encoding scheme and derivation of a proficient multi-objective fitness function. We use Pareto dominance in MOPSO for obtaining both local and global best guides for each particle. We carry out rigorous simulation experiments on the proposed algorithm and compare the results with two existing algorithms namely, tree cluster based data gathering algorithm (TCBDGA) and energy aware sink relocation (EASR). The results demonstrate that the proposed algorithm performs better than both of them in terms of various performance metrics. The results are also validated through the statistical test, analysis of variance (ANOVA) and its least significant difference (LSD) post hoc analysis.
Robot tutor and pupils’ educational ability: Teaching the times tables Research shows promising results of educational robots in language and STEM tasks. In language, more research is available, occasionally in view of individual differences in pupils’ educational ability levels, and learning seems to improve with more expressive robot behaviors. In STEM, variations in robots’ behaviors have been examined with inconclusive results and never while systematically investigating how differences in educational abilities match with different robot behaviors. We applied an autonomously tutoring robot (without tablet, partly WOz) in a 2 × 2 experiment of social vs. neutral behavior in above-average vs. below-average schoolchildren (N = 86; age 8–10 years) while rehearsing the multiplication tables on a one-to-one basis. The standard school test showed that on average, pupils significantly improved their performance even after 3 occasions of 5-min exercises. Beyond-average pupils profited most from a robot tutor, whereas those below average in multiplication benefited more from a robot that showed neutral rather than more social behavior.
1.24
0.24
0.24
0.24
0.12
0.01
0
0
0
0
0
0
0
0
Data Statements for Natural Language Processing: Toward Mitigating System Bias and Enabling Better Science. In this paper, we propose data statements as a design solution and professional practice for natural language processing technologists, in both research and development—through the adoption and widespread use of data statements, the field can begin to address critical scientific and ethical issues that result from the use of data from certain populations in the development of technology for other populations. We present a form that data statements can take and explore the implications of adopting them as part of regular practice. We argue that data statements will help alleviate issues related to exclusion and bias in language technology; lead to better precision in claims about how NLP research can generalize and thus better engineering results; protect companies from public embarrassment; and ultimately lead to language technology that meets its users in their own preferred linguistic style and furthermore does not misrepresent them to others.
Factual and Counterfactual Explanations for Black Box Decision Making. The rise of sophisticated machine learning models has brought accurate but obscure decision systems, which hide their logic, thus undermining transparency, trust, and the adoption of artificial intelligence (AI) in socially sensitive and safety-critical contexts. We introduce a local rule-based explanation method, providing faithful explanations of the decision made by a black box classifier on a ...
Optimal classification trees. State-of-the-art decision tree methods apply heuristics recursively to create each split in isolation, which may not capture well the underlying characteristics of the dataset. The optimal decision tree problem attempts to resolve this by creating the entire decision tree at once to achieve global optimality. In the last 25 years, algorithmic advances in integer optimization coupled with hardware improvements have resulted in an astonishing 800 billion factor speedup in mixed-integer optimization (MIO). Motivated by this speedup, we present optimal classification trees, a novel formulation of the decision tree problem using modern MIO techniques that yields the optimal decision tree for axes-aligned splits. We also show the richness of this MIO formulation by adapting it to give optimal classification trees with hyperplanes that generates optimal decision trees with multivariate splits. Synthetic tests demonstrate that these methods recover the true decision tree more closely than heuristics, refuting the notion that optimal methods overfit the training data. We comprehensively benchmark these methods on a sample of 53 datasets from the UCI machine learning repository. We establish that these MIO methods are practically solvable on real-world datasets with sizes in the 1000s, and give average absolute improvements in out-of-sample accuracy over CART of 1---2 and 3---5% for the univariate and multivariate cases, respectively. Furthermore, we identify that optimal classification trees are likely to outperform CART by 1.2---1.3% in situations where the CART accuracy is high and we have sufficient training data, while the multivariate version outperforms CART by 4---7% when the CART accuracy or dimension of the dataset is low.
Bias in data-driven artificial intelligence systems - An introductory survey. Artificial Intelligence (AI)-based systems are widely employed nowadays to make decisions that have far-reaching impact on individuals and society. Their decisions might affect everyone, everywhere, and anytime, entailing concerns about potential human rights issues. Therefore, it is necessary to move beyond traditional AI algorithms optimized for predictive performance and embed ethical and legal principles in their design, training, and deployment to ensure social good while still benefiting from the huge potential of the AI technology. The goal of this survey is to provide a broad multidisciplinary overview of the area of bias in AI systems, focusing on technical challenges and solutions as well as to suggest new research directions towards approaches well-grounded in a legal frame. In this survey, we focus on data-driven AI, as a large part of AI is powered nowadays by (big) data and powerful machine learning algorithms. If otherwise not specified, we use the general term bias to describe problems related to the gathering or processing of data that might result in prejudiced decisions on the bases of demographic features such as race, sex, and so forth. This article is categorized under: Commercial, Legal, and Ethical Issues > Fairness in Data Mining Commercial, Legal, and Ethical Issues > Ethical Considerations Commercial, Legal, and Ethical Issues > Legal Issues
Explanation in Artificial Intelligence: Insights from the Social Sciences. There has been a recent resurgence in the area of explainable artificial intelligence as researchers and practitioners seek to provide more transparency to their algorithms. Much of this research is focused on explicitly explaining decisions or actions to a human observer, and it should not be controversial to say that looking at how humans explain to each other can serve as a useful starting point for explanation in artificial intelligence. However, it is fair to say that most work in explainable artificial intelligence uses only the researchers' intuition of what constitutes a ‘good’ explanation. There exist vast and valuable bodies of research in philosophy, psychology, and cognitive science of how people define, generate, select, evaluate, and present explanations, which argues that people employ certain cognitive biases and social expectations to the explanation process. This paper argues that the field of explainable artificial intelligence can build on this existing research, and reviews relevant papers from philosophy, cognitive psychology/science, and social psychology, which study these topics. It draws out some important findings, and discusses ways that these can be infused with work on explainable artificial intelligence.
Computational thinking Summary form only given. My vision for the 21st century, Computational Thinking, will be a fundamental skill used by everyone in the world. To reading, writing, and arithmetic, we should add computational thinking to every child's analytical ability. Computational thinking involves solving problems, designing systems, and understanding human behavior by drawing on the concepts fundamental to computer science. Thinking like a computer scientist means more than being able to program a computer. It requires the ability to abstract and thus to think at multiple levels of abstraction. In this talk I will give many examples of computational thinking, argue that it has already influenced other disciplines, and promote the idea that teaching computational thinking can not only inspire future generations to enter the field of computer science but benefit people in all fields.
JPEG Error Analysis and Its Applications to Digital Image Forensics JPEG is one of the most extensively used image formats. Understanding the inherent characteristics of JPEG may play a useful role in digital image forensics. In this paper, we introduce JPEG error analysis to the study of image forensics. The main errors of JPEG include quantization, rounding, and truncation errors. Through theoretically analyzing the effects of these errors on single and double JPEG compression, we have developed three novel schemes for image forensics including identifying whether a bitmap image has previously been JPEG compressed, estimating the quantization steps of a JPEG image, and detecting the quantization table of a JPEG image. Extensive experimental results show that our new methods significantly outperform existing techniques especially for the images of small sizes. We also show that the new method can reliably detect JPEG image blocks which are as small as 8 × 8 pixels and compressed with quality factors as high as 98. This performance is important for analyzing and locating small tampered regions within a composite image.
Highly dynamic Destination-Sequenced Distance-Vector routing (DSDV) for mobile computers An ad-hoc network is the cooperative engagement of a collection of Mobile Hosts without the required intervention of any centralized Access Point. In this paper we present an innovative design for the operation of such ad-hoc networks. The basic idea of the design is to operate each Mobile Host as a specialized router, which periodically advertises its view of the interconnection topology with other Mobile Hosts within the network. This amounts to a new sort of routing protocol. We have investigated modifications to the basic Bellman-Ford routing mechanisms, as specified by RIP [5], to make it suitable for a dynamic and self-starting network mechanism as is required by users wishing to utilize ad hoc networks. Our modifications address some of the previous objections to the use of Bellman-Ford, related to the poor looping properties of such algorithms in the face of broken links and the resulting time dependent nature of the interconnection topology describing the links between the Mobile Hosts. Finally, we describe the ways in which the basic network-layer routing can be modified to provide MAC-layer support for ad-hoc networks.
Efficient Signature Generation by Smart Cards We present a new public-key signature scheme and a corresponding authentication scheme that are based on discrete logarithms in a subgroup of units in Zp where p is a sufficiently large prime, e.g., p = 2512. A key idea is to use for the base of the discrete logarithm an integer a in Zp such that the order of a is a sufficiently large prime q, e.g., q = 2140. In this way we improve the ElGamal signature scheme in the speed of the procedures for the generation and the verification of signatures and also in the bit length of signatures. We present an efficient algorithm that preprocesses the exponentiation of a random residue modulo p.
Stabilizing a linear system by switching control with dwell time The use of networks in control systems to connect controllers and sensors/actuators has become common practice in many applications. This new technology has also posed a theoretical control problem of how to use the limited data rate of the network effectively. We consider a system where its sensor and actuator are connected by a finite data rate channel. A design method to stabilize a continuous-time, linear plant using a switching controller is proposed. In particular, to prevent the actuator from fast switching, or chattering, which can not only increase the necessary data rate but also damage the system, we employ a dwell-time switching scheme. It is shown that a systematic partition of the state-space enables us to reduce the complexity of the design problem
Empirical Modelling of Genetic Algorithms This paper addresses the problem of reliably setting genetic algorithm parameters for consistent labelling problems. Genetic algorithm parameters are notoriously difficult to determine. This paper proposes a robust empirical framework, based on the analysis of factorial experiments. The use of a graeco-latin square permits an initial study of a wide range of parameter settings. This is followed by fully crossed factorial experiments with narrower ranges, which allow detailed analysis by logistic regression. The empirical models derived can be used to determine optimal algorithm parameters and to shed light on interactions between the parameters and their relative importance. Re-fined models are produced, which are shown to be robust under extrapolation to up to triple the problem size.
Biologically-inspired soft exosuit. In this paper, we present the design and evaluation of a novel soft cable-driven exosuit that can apply forces to the body to assist walking. Unlike traditional exoskeletons which contain rigid framing elements, the soft exosuit is worn like clothing, yet can generate moments at the ankle and hip with magnitudes of 18% and 30% of those naturally generated by the body during walking, respectively. Our design uses geared motors to pull on Bowden cables connected to the suit near the ankle. The suit has the advantages over a traditional exoskeleton in that the wearer's joints are unconstrained by external rigid structures, and the worn part of the suit is extremely light, which minimizes the suit's unintentional interference with the body's natural biomechanics. However, a soft suit presents challenges related to actuation force transfer and control, since the body is compliant and cannot support large pressures comfortably. We discuss the design of the suit and actuation system, including principles by which soft suits can transfer force to the body effectively and the biological inspiration for the design. For a soft exosuit, an important design parameter is the combined effective stiffness of the suit and its interface to the wearer. We characterize the exosuit's effective stiffness, and present preliminary results from it generating assistive torques to a subject during walking. We envision such an exosuit having broad applicability for assisting healthy individuals as well as those with muscle weakness.
Inter-class sparsity based discriminative least square regression Least square regression is a very popular supervised classification method. However, two main issues greatly limit its performance. The first one is that it only focuses on fitting the input features to the corresponding output labels while ignoring the correlations among samples. The second one is that the used label matrix, i.e., zero–one label matrix is inappropriate for classification. To solve these problems and improve the performance, this paper presents a novel method, i.e., inter-class sparsity based discriminative least square regression (ICS_DLSR), for multi-class classification. Different from other methods, the proposed method pursues that the transformed samples have a common sparsity structure in each class. For this goal, an inter-class sparsity constraint is introduced to the least square regression model such that the margins of samples from the same class can be greatly reduced while those of samples from different classes can be enlarged. In addition, an error term with row-sparsity constraint is introduced to relax the strict zero–one label matrix, which allows the method to be more flexible in learning the discriminative transformation matrix. These factors encourage the method to learn a more compact and discriminative transformation for regression and thus has the potential to perform better than other methods. Extensive experimental results show that the proposed method achieves the best performance in comparison with other methods for multi-class classification.
Robot tutor and pupils’ educational ability: Teaching the times tables Research shows promising results of educational robots in language and STEM tasks. In language, more research is available, occasionally in view of individual differences in pupils’ educational ability levels, and learning seems to improve with more expressive robot behaviors. In STEM, variations in robots’ behaviors have been examined with inconclusive results and never while systematically investigating how differences in educational abilities match with different robot behaviors. We applied an autonomously tutoring robot (without tablet, partly WOz) in a 2 × 2 experiment of social vs. neutral behavior in above-average vs. below-average schoolchildren (N = 86; age 8–10 years) while rehearsing the multiplication tables on a one-to-one basis. The standard school test showed that on average, pupils significantly improved their performance even after 3 occasions of 5-min exercises. Beyond-average pupils profited most from a robot tutor, whereas those below average in multiplication benefited more from a robot that showed neutral rather than more social behavior.
1.2
0.2
0.2
0.2
0.066667
0
0
0
0
0
0
0
0
0
Stability and $l_2$ -Gain Analysis of Discrete-Time Switched Systems with Mode-Dependent Average Dwell Time. In this paper, the problems of stability and l2-gain analysis are investigated for discrete-time switched systems with mode-dependent average dwell time (DT). By proposing a novel multiple discontinuous Lyapunov function (MDLF) approach, the conditions of stability and weighted l2-gain are established for nonlinear systems. Next, the corresponding results for linear switched systems are also propo...
The Sybil Attack Large-scale peer-to-peer systems facesecurity threats from faulty or hostile remotecomputing elements. To resist these threats, manysuch systems employ redundancy. However, if asingle faulty entity can present multiple identities,it can control a substantial fraction of the system,thereby undermining this redundancy. Oneapproach to preventing these &quot;Sybil attacks&quot; is tohave a trusted agency certify identities. Thispaper shows that, without a logically centralizedauthority, Sybil...
BLEU: a method for automatic evaluation of machine translation Human evaluations of machine translation are extensive but expensive. Human evaluations can take months to finish and involve human labor that can not be reused. We propose a method of automatic machine translation evaluation that is quick, inexpensive, and language-independent, that correlates highly with human evaluation, and that has little marginal cost per run. We present this method as an automated understudy to skilled human judges which substitutes for them when there is need for quick or frequent evaluations.
Computational thinking Summary form only given. My vision for the 21st century, Computational Thinking, will be a fundamental skill used by everyone in the world. To reading, writing, and arithmetic, we should add computational thinking to every child's analytical ability. Computational thinking involves solving problems, designing systems, and understanding human behavior by drawing on the concepts fundamental to computer science. Thinking like a computer scientist means more than being able to program a computer. It requires the ability to abstract and thus to think at multiple levels of abstraction. In this talk I will give many examples of computational thinking, argue that it has already influenced other disciplines, and promote the idea that teaching computational thinking can not only inspire future generations to enter the field of computer science but benefit people in all fields.
Fuzzy logic in control systems: fuzzy logic controller. I.
Switching between stabilizing controllers This paper deals with the problem of switching between several linear time-invariant (LTI) controllers—all of them capable of stabilizing a speci4c LTI process—in such a way that the stability of the closed-loop system is guaranteed for any switching sequence. We show that it is possible to 4nd realizations for any given family of controller transfer matrices so that the closed-loop system remains stable, no matter how we switch among the controller. The motivation for this problem is the control of complex systems where con8icting requirements make a single LTI controller unsuitable. ? 2002 Published by Elsevier Science Ltd.
Tabu Search - Part I
Bidirectional recurrent neural networks In the first part of this paper, a regular recurrent neural network (RNN) is extended to a bidirectional recurrent neural network (BRNN). The BRNN can be trained without the limitation of using input information just up to a preset future frame. This is accomplished by training it simultaneously in positive and negative time direction. Structure and training procedure of the proposed network are explained. In regression and classification experiments on artificial data, the proposed structure gives better results than other approaches. For real data, classification experiments for phonemes from the TIMIT database show the same tendency. In the second part of this paper, it is shown how the proposed bidirectional structure can be easily modified to allow efficient estimation of the conditional posterior probability of complete symbol sequences without making any explicit assumption about the shape of the distribution. For this part, experiments on real data are reported
An intensive survey of fair non-repudiation protocols With the phenomenal growth of the Internet and open networks in general, security services, such as non-repudiation, become crucial to many applications. Non-repudiation services must ensure that when Alice sends some information to Bob over a network, neither Alice nor Bob can deny having participated in a part or the whole of this communication. Therefore a fair non-repudiation protocol has to generate non-repudiation of origin evidences intended to Bob, and non-repudiation of receipt evidences destined to Alice. In this paper, we clearly define the properties a fair non-repudiation protocol must respect, and give a survey of the most important non-repudiation protocols without and with trusted third party (TTP). For the later ones we discuss the evolution of the TTP's involvement and, between others, describe the most recent protocol using a transparent TTP. We also discuss some ad-hoc problems related to the management of non-repudiation evidences.
Dynamic movement and positioning of embodied agents in multiparty conversations For embodied agents to engage in realistic multiparty conversation, they must stand in appropriate places with respect to other agents and the environment. When these factors change, such as an agent joining the conversation, the agents must dynamically move to a new location and/or orientation to accommodate. This paper presents an algorithm for simulating movement of agents based on observed human behavior using techniques developed for pedestrian movement in crowd simulations. We extend a previous group conversation simulation to include an agent motion algorithm. We examine several test cases and show how the simulation generates results that mirror real-life conversation settings.
An improved genetic algorithm with conditional genetic operators and its application to set-covering problem The genetic algorithm (GA) is a popular, biologically inspired optimization method. However, in the GA there is no rule of thumb to design the GA operators and select GA parameters. Instead, trial-and-error has to be applied. In this paper we present an improved genetic algorithm in which crossover and mutation are performed conditionally instead of probability. Because there are no crossover rate and mutation rate to be selected, the proposed improved GA can be more easily applied to a problem than the conventional genetic algorithms. The proposed improved genetic algorithm is applied to solve the set-covering problem. Experimental studies show that the improved GA produces better results over the conventional one and other methods.
Lane-level traffic estimations using microscopic traffic variables This paper proposes a novel inference method to estimate lane-level traffic flow, time occupancy and vehicle inter-arrival time on road segments where local information could not be measured and assessed directly. The main contributions of the proposed method are 1) the ability to perform lane-level estimations of traffic flow, time occupancy and vehicle inter-arrival time and 2) the ability to adapt to different traffic regimes by assessing only microscopic traffic variables. We propose a modified Kriging estimation model which explicitly takes into account both spatial and temporal variability. Performance evaluations are conducted using real-world data under different traffic regimes and it is shown that the proposed method outperforms a Kalman filter-based approach.
Convolutional Neural Network-Based Classification of Driver's Emotion during Aggressive and Smooth Driving Using Multi-Modal Camera Sensors. Because aggressive driving often causes large-scale loss of life and property, techniques for advance detection of adverse driver emotional states have become important for the prevention of aggressive driving behaviors. Previous studies have primarily focused on systems for detecting aggressive driver emotion via smart-phone accelerometers and gyro-sensors, or they focused on methods of detecting physiological signals using electroencephalography (EEG) or electrocardiogram (ECG) sensors. Because EEG and ECG sensors cause discomfort to drivers and can be detached from the driver's body, it becomes difficult to focus on bio-signals to determine their emotional state. Gyro-sensors and accelerometers depend on the performance of GPS receivers and cannot be used in areas where GPS signals are blocked. Moreover, if driving on a mountain road with many quick turns, a driver's emotional state can easily be misrecognized as that of an aggressive driver. To resolve these problems, we propose a convolutional neural network (CNN)-based method of detecting emotion to identify aggressive driving using input images of the driver's face, obtained using near-infrared (NIR) light and thermal camera sensors. In this research, we conducted an experiment using our own database, which provides a high classification accuracy for detecting driver emotion leading to either aggressive or smooth (i.e., relaxed) driving. Our proposed method demonstrates better performance than existing methods.
Robot tutor and pupils’ educational ability: Teaching the times tables Research shows promising results of educational robots in language and STEM tasks. In language, more research is available, occasionally in view of individual differences in pupils’ educational ability levels, and learning seems to improve with more expressive robot behaviors. In STEM, variations in robots’ behaviors have been examined with inconclusive results and never while systematically investigating how differences in educational abilities match with different robot behaviors. We applied an autonomously tutoring robot (without tablet, partly WOz) in a 2 × 2 experiment of social vs. neutral behavior in above-average vs. below-average schoolchildren (N = 86; age 8–10 years) while rehearsing the multiplication tables on a one-to-one basis. The standard school test showed that on average, pupils significantly improved their performance even after 3 occasions of 5-min exercises. Beyond-average pupils profited most from a robot tutor, whereas those below average in multiplication benefited more from a robot that showed neutral rather than more social behavior.
1
0
0
0
0
0
0
0
0
0
0
0
0
0
Multiple UAVs as Relays: Multi-Hop Single Link Versus Multiple Dual-Hop Links. Unmanned aerial vehicles (UAVs) have found many important applications in communications. They can serve as either aerial base stations or mobile relays to improve the quality of services. In this paper, we study the use of multiple UAVs in relaying. Considering two typical uses of multiple UAVs as relays that form either a single multi-hop link or multiple dual-hop links, we first optimize the pl...
Interior point methods 25 years later. Interior point methods for optimization have been around for more than 25 years now. Their presence has shaken up the field of optimization. Interior point methods for linear and (convex) quadratic programming display several features which make them particularly attractive for very large scale optimization. Among the most impressive of them are their low-degree polynomial worst-case complexity and an unrivalled ability to deliver optimal solutions in an almost constant number of iterations which depends very little, if at all, on the problem dimension. Interior point methods are competitive when dealing with small problems of dimensions below one million constraints and variables and are beyond competition when applied to large problems of dimensions going into millions of constraints and variables. In this survey we will discuss several issues related to interior point methods including the proof of the worst-case complexity result, the reasons for their amazingly fast practical convergence and the features responsible for their ability to solve very large problems. The ever-growing sizes of optimization problems impose new requirements on optimization methods and software. In the final part of this paper we will therefore address a redesign of interior point methods to allow them to work in a matrix-free regime and to make them well-suited to solving even larger problems. (C) 2011 Elsevier B.V. All rights reserved.
Genetic Algorithms in Wireless Networking: Techniques, Applications, and Issues. In recent times, wireless access technology is becoming increasingly commonplace due to the ease of operation and installation of untethered wireless media. The design of wireless networking is challenging due to the highly dynamic environmental condition that makes parameter optimization a complex task. Due to the dynamic, and often unknown, operating conditions, modern wireless networking standards increasingly rely on machine learning and artificial intelligence algorithms. Genetic algorithms (GAs) provide a well-established framework for implementing artificial intelligence tasks such as classification, learning, and optimization. GAs are well known for their remarkable generality and versatility and have been applied in a wide variety of settings in wireless networks. In this paper, we provide a comprehensive survey of the applications of GAs in wireless networks. We provide both an exposition of common GA models and configuration and provide a broad-ranging survey of GA techniques in wireless networks. We also point out open research issues and define potential future work. While various surveys on GAs exist in the literature, our paper is the first paper, to the best of our knowledge, which focuses on their application in wireless networks.
Fundamentals of the Downlink Green Coverage and Energy Efficiency in Heterogeneous Networks. This paper studies the proposed green (energy-efficient) coverage probability, link and network energy efficiencies in the downlink of a heterogeneous cellular network (HetNet) consisting of $K$ independent Poisson point processes of base stations (BSs). The important statistical properties of the universal (general) cell association functions are first studied, and the cell load statistics for power-law cell association functions, which can characterize the accurate void cell probability of a BS in every tier, are also derived. A simple and feasible green channel-aware cell association (GCA) scheme is proposed and the green coverage probability is also proposed for any particular cell association scheme, such as the maximum received power association (MRPA) and the nearest BS association (NBA) schemes. Then, the link and network energy efficiencies are proposed to characterize the mean spectrum efficiency per unit power consumption for a BS and the mean area spectrum efficiency for the HetNet, respectively. All the tight bounds on the green coverage probability, link, and network energy efficiencies for the GCA, MRPA, and NBA schemes are found. They are theoretically shown to pose the fundamental maximum limits on the link and network energy efficiencies achieved by any other cell association schemes, and such a fact is validated by numerical results as well.
Predictive Deployment of UAV Base Stations in Wireless Networks: Machine Learning Meets Contract Theory In this paper, a novel framework is proposed to enable a predictive deployment of unmanned aerial vehicles (UAVs) as temporary base stations (BSs) to complement ground cellular systems in face of downlink traffic overload. First, a novel learning approach, based on the weighted expectation maximization (WEM) algorithm, is proposed to estimate the user distribution and the downlink traffic demand. Next, to guarantee a truthful information exchange between the BS and UAVs, using the framework of contract theory, an offload contract is developed, and the sufficient and necessary conditions for having a feasible contract are analytically derived. Subsequently, an optimization problem is formulated to deploy an optimal UAV onto the hotspot area in a way that the utility of the overloaded BS is maximized. Simulation results show that the proposed WEM approach yields a prediction error of around 10%. Compared with the expectation maximization and k-mean approaches, the WEM method shows a significant advantage on the prediction accuracy, as the traffic load in the cellular system becomes spatially uneven. Furthermore, compared with two event-driven deployment schemes based on the closest-distance and maximal-energy metrics, the proposed predictive approach enables UAV operators to provide efficient communication service for hotspot users in terms of the downlink capacity, energy consumption and service delay. Simulation results also show that the proposed method significantly improves the revenues of both the BS and UAV networks, compared with two baseline schemes.
Millimeter Wave Cellular Wireless Networks: Potentials and Challenges. Millimeter-wave (mmW) frequencies between 30 and 300 GHz are a new frontier for cellular communication that offers the promise of orders of magnitude greater bandwidths combined with further gains via beamforming and spatial multiplexing from multielement antenna arrays. This paper surveys measurements and capacity studies to assess this technology with a focus on small cell deployments in urban e...
Footprints: history-rich tools for information foraging Inspired by Hill and Hollans original work [7], we have beendeveloping a theory of interaction history and building tools toapply this theory to navigation in a complex information space. Wehave built a series of tools - map, paths, annota- tions andsignposts - based on a physical-world navigation metaphor. Thesetools have been in use for over a year. Our user study involved acontrolled browse task and showed that users were able to get thesame amount of work done with significantly less effort.
Very Deep Convolutional Networks for Large-Scale Image Recognition. In this work we investigate the effect of the convolutional network depth on its accuracy in the large-scale image recognition setting. Our main contribution is a thorough evaluation of networks of increasing depth using an architecture with very small (3x3) convolution filters, which shows that a significant improvement on the prior-art configurations can be achieved by pushing the depth to 16-19 weight layers. These findings were the basis of our ImageNet Challenge 2014 submission, where our team secured the first and the second places in the localisation and classification tracks respectively. We also show that our representations generalise well to other datasets, where they achieve state-of-the-art results. We have made our two best-performing ConvNet models publicly available to facilitate further research on the use of deep visual representations in computer vision.
Chimp optimization algorithm. •A novel optimizer called Chimp Optimization Algorithm (ChOA) is proposed.•ChOA is inspired by individual intelligence and sexual motivation of chimps.•ChOA alleviates the problems of slow convergence rate and trapping in local optima.•The four main steps of Chimp hunting are implemented.
Space-time modeling of traffic flow. This paper discusses the application of space-time autoregressive integrated moving average (STARIMA) methodology for representing traffic flow patterns. Traffic flow data are in the form of spatial time series and are collected at specific locations at constant intervals of time. Important spatial characteristics of the space-time process are incorporated in the STARIMA model through the use of weighting matrices estimated on the basis of the distances among the various locations where data are collected. These matrices distinguish the space-time approach from the vector autoregressive moving average (VARMA) methodology and enable the model builders to control the number of the parameters that have to be estimated. The proposed models can be used for short-term forecasting of space-time stationary traffic-flow processes and for assessing the impact of traffic-flow changes on other parts of the network. The three-stage iterative space-time model building procedure is illustrated using 7.5min average traffic flow data for a set of 25 loop-detectors located at roads that direct to the centre of the city of Athens, Greece. Data for two months with different traffic-flow characteristics are modelled in order to determine the stability of the parameter estimation.
A simplified dual neural network for quadratic programming with its KWTA application. The design, analysis, and application of a new recurrent neural network for quadratic programming, called simplified dual neural network, are discussed. The analysis mainly concentrates on the convergence property and the computational complexity of the neural network. The simplified dual neural network is shown to be globally convergent to the exact optimal solution. The complexity of the neural network architecture is reduced with the number of neurons equal to the number of inequality constraints. Its application to k-winners-take-all (KWTA) operation is discussed to demonstrate how to solve problems with this neural network.
Understanding Taxi Service Strategies From Taxi GPS Traces Taxi service strategies, as the crowd intelligence of massive taxi drivers, are hidden in their historical time-stamped GPS traces. Mining GPS traces to understand the service strategies of skilled taxi drivers can benefit the drivers themselves, passengers, and city planners in a number of ways. This paper intends to uncover the efficient and inefficient taxi service strategies based on a large-scale GPS historical database of approximately 7600 taxis over one year in a city in China. First, we separate the GPS traces of individual taxi drivers and link them with the revenue generated. Second, we investigate the taxi service strategies from three perspectives, namely, passenger-searching strategies, passenger-delivery strategies, and service-region preference. Finally, we represent the taxi service strategies with a feature matrix and evaluate the correlation between service strategies and revenue, informing which strategies are efficient or inefficient. We predict the revenue of taxi drivers based on their strategies and achieve a prediction residual as less as 2.35 RMB/h,1 which demonstrates that the extracted taxi service strategies with our proposed approach well characterize the driving behavior and performance of taxi drivers.
An indoor localization solution using Bluetooth RSSI and multiple sensors on a smartphone. In this paper, we propose an indoor positioning system using a Bluetooth receiver, an accelerometer, a magnetic field sensor, and a barometer on a smartphone. The Bluetooth receiver is used to estimate distances from beacons. The accelerometer and magnetic field sensor are used to trace the movement of moving people in the given space. The horizontal location of the person is determined by received signal strength indications (RSSIs) and the traced movement. The barometer is used to measure the vertical position where a person is located. By combining RSSIs, the traced movement, and the vertical position, the proposed system estimates the indoor position of moving people. In experiments, the proposed approach showed excellent performance in localization with an overall error of 4.8%.
A Hierarchical Architecture Using Biased Min-Consensus for USV Path Planning This paper proposes a hierarchical architecture using the biased min-consensus (BMC) method, to solve the path planning problem of unmanned surface vessel (USV). We take the fixed-point monitoring mission as an example, where a series of intermediate monitoring points should be visited once by USV. The whole framework incorporates the low-level layer planning the standard path between any two intermediate points, and the high-level fashion determining their visiting sequence. First, the optimal standard path in terms of voyage time and risk measure is planned by the BMC protocol, given that the corresponding graph is constructed with node state and edge weight. The USV will avoid obstacles or keep a certain distance safely, and arrive at the target point quickly. It is proven theoretically that the state of the graph will converge to be stable after finite iterations, i.e., the optimal solution can be found by BMC with low calculation complexity. Second, by incorporating the constraint of intermediate points, their visiting sequence is optimized by BMC again with the reconstruction of a new virtual graph based on the former planned results. The extensive simulation results in various scenarios also validate the feasibility and effectiveness of our method for autonomous navigation.
1.2
0.2
0.2
0.2
0.2
0.066667
0
0
0
0
0
0
0
0
Deepfake Detection on Videos Based on Ratio Images Deepfake detection comes as a countermeasure to identify fake media content to reduce its harmful implications. Most detection approaches rely on identifying specific artifacts that can quickly become obsolete due to the fast advance in facial forgery methods. Some facial manipulation detection methods use temporal information to classify the video as real or fake. These methods mainly rely on 3D CNN architectures or two-stream networks using frame and video features. Our method not only considers temporal aspects, but it comes from a different perspective: extracting features that can account for inter-frame changes on a video. Inspired by the concept of ratio images, we extract features based on the ratio between adjacent frames for the face and its background. The experimental evaluation showed better results in intra- and cross-dataset tests on FaceForensics++ (FF++) and CelebDF datasets compared to the state-of-the-art deepfake detection approaches in the assessment with seen and unseen facial manipulation methods, as well as in seen and unseen video settings. In the intra-dataset experiment, the model resulted in an AUC of 100% for both CelebDF and FF++ datasets. In the cross dataset experiment, the model resulted in an AUC of 98% when trained with CelebDF and tested with FF++ and 86% when trained with FF++ and tested with CelebDF.
The Sybil Attack Large-scale peer-to-peer systems facesecurity threats from faulty or hostile remotecomputing elements. To resist these threats, manysuch systems employ redundancy. However, if asingle faulty entity can present multiple identities,it can control a substantial fraction of the system,thereby undermining this redundancy. Oneapproach to preventing these &quot;Sybil attacks&quot; is tohave a trusted agency certify identities. Thispaper shows that, without a logically centralizedauthority, Sybil...
BLEU: a method for automatic evaluation of machine translation Human evaluations of machine translation are extensive but expensive. Human evaluations can take months to finish and involve human labor that can not be reused. We propose a method of automatic machine translation evaluation that is quick, inexpensive, and language-independent, that correlates highly with human evaluation, and that has little marginal cost per run. We present this method as an automated understudy to skilled human judges which substitutes for them when there is need for quick or frequent evaluations.
Computational thinking Summary form only given. My vision for the 21st century, Computational Thinking, will be a fundamental skill used by everyone in the world. To reading, writing, and arithmetic, we should add computational thinking to every child's analytical ability. Computational thinking involves solving problems, designing systems, and understanding human behavior by drawing on the concepts fundamental to computer science. Thinking like a computer scientist means more than being able to program a computer. It requires the ability to abstract and thus to think at multiple levels of abstraction. In this talk I will give many examples of computational thinking, argue that it has already influenced other disciplines, and promote the idea that teaching computational thinking can not only inspire future generations to enter the field of computer science but benefit people in all fields.
Fuzzy logic in control systems: fuzzy logic controller. I.
Switching between stabilizing controllers This paper deals with the problem of switching between several linear time-invariant (LTI) controllers—all of them capable of stabilizing a speci4c LTI process—in such a way that the stability of the closed-loop system is guaranteed for any switching sequence. We show that it is possible to 4nd realizations for any given family of controller transfer matrices so that the closed-loop system remains stable, no matter how we switch among the controller. The motivation for this problem is the control of complex systems where con8icting requirements make a single LTI controller unsuitable. ? 2002 Published by Elsevier Science Ltd.
Tabu Search - Part I
Bidirectional recurrent neural networks In the first part of this paper, a regular recurrent neural network (RNN) is extended to a bidirectional recurrent neural network (BRNN). The BRNN can be trained without the limitation of using input information just up to a preset future frame. This is accomplished by training it simultaneously in positive and negative time direction. Structure and training procedure of the proposed network are explained. In regression and classification experiments on artificial data, the proposed structure gives better results than other approaches. For real data, classification experiments for phonemes from the TIMIT database show the same tendency. In the second part of this paper, it is shown how the proposed bidirectional structure can be easily modified to allow efficient estimation of the conditional posterior probability of complete symbol sequences without making any explicit assumption about the shape of the distribution. For this part, experiments on real data are reported
An intensive survey of fair non-repudiation protocols With the phenomenal growth of the Internet and open networks in general, security services, such as non-repudiation, become crucial to many applications. Non-repudiation services must ensure that when Alice sends some information to Bob over a network, neither Alice nor Bob can deny having participated in a part or the whole of this communication. Therefore a fair non-repudiation protocol has to generate non-repudiation of origin evidences intended to Bob, and non-repudiation of receipt evidences destined to Alice. In this paper, we clearly define the properties a fair non-repudiation protocol must respect, and give a survey of the most important non-repudiation protocols without and with trusted third party (TTP). For the later ones we discuss the evolution of the TTP's involvement and, between others, describe the most recent protocol using a transparent TTP. We also discuss some ad-hoc problems related to the management of non-repudiation evidences.
Dynamic movement and positioning of embodied agents in multiparty conversations For embodied agents to engage in realistic multiparty conversation, they must stand in appropriate places with respect to other agents and the environment. When these factors change, such as an agent joining the conversation, the agents must dynamically move to a new location and/or orientation to accommodate. This paper presents an algorithm for simulating movement of agents based on observed human behavior using techniques developed for pedestrian movement in crowd simulations. We extend a previous group conversation simulation to include an agent motion algorithm. We examine several test cases and show how the simulation generates results that mirror real-life conversation settings.
An improved genetic algorithm with conditional genetic operators and its application to set-covering problem The genetic algorithm (GA) is a popular, biologically inspired optimization method. However, in the GA there is no rule of thumb to design the GA operators and select GA parameters. Instead, trial-and-error has to be applied. In this paper we present an improved genetic algorithm in which crossover and mutation are performed conditionally instead of probability. Because there are no crossover rate and mutation rate to be selected, the proposed improved GA can be more easily applied to a problem than the conventional genetic algorithms. The proposed improved genetic algorithm is applied to solve the set-covering problem. Experimental studies show that the improved GA produces better results over the conventional one and other methods.
Lane-level traffic estimations using microscopic traffic variables This paper proposes a novel inference method to estimate lane-level traffic flow, time occupancy and vehicle inter-arrival time on road segments where local information could not be measured and assessed directly. The main contributions of the proposed method are 1) the ability to perform lane-level estimations of traffic flow, time occupancy and vehicle inter-arrival time and 2) the ability to adapt to different traffic regimes by assessing only microscopic traffic variables. We propose a modified Kriging estimation model which explicitly takes into account both spatial and temporal variability. Performance evaluations are conducted using real-world data under different traffic regimes and it is shown that the proposed method outperforms a Kalman filter-based approach.
Convolutional Neural Network-Based Classification of Driver's Emotion during Aggressive and Smooth Driving Using Multi-Modal Camera Sensors. Because aggressive driving often causes large-scale loss of life and property, techniques for advance detection of adverse driver emotional states have become important for the prevention of aggressive driving behaviors. Previous studies have primarily focused on systems for detecting aggressive driver emotion via smart-phone accelerometers and gyro-sensors, or they focused on methods of detecting physiological signals using electroencephalography (EEG) or electrocardiogram (ECG) sensors. Because EEG and ECG sensors cause discomfort to drivers and can be detached from the driver's body, it becomes difficult to focus on bio-signals to determine their emotional state. Gyro-sensors and accelerometers depend on the performance of GPS receivers and cannot be used in areas where GPS signals are blocked. Moreover, if driving on a mountain road with many quick turns, a driver's emotional state can easily be misrecognized as that of an aggressive driver. To resolve these problems, we propose a convolutional neural network (CNN)-based method of detecting emotion to identify aggressive driving using input images of the driver's face, obtained using near-infrared (NIR) light and thermal camera sensors. In this research, we conducted an experiment using our own database, which provides a high classification accuracy for detecting driver emotion leading to either aggressive or smooth (i.e., relaxed) driving. Our proposed method demonstrates better performance than existing methods.
Ethical Considerations Of Applying Robots In Kindergarten Settings: Towards An Approach From A Macroperspective In child-robot interaction (cHRI) research, many studies pursue the goal to develop interactive systems that can be applied in everyday settings. For early education, increasingly, the setting of a kindergarten is targeted. However, when cHRI and research are brought into a kindergarten, a range of ethical and related procedural aspects have to be considered and dealt with. While ethical models elaborated within other human-robot interaction settings, e.g., assisted living contexts, can provide some important indicators for relevant issues, we argue that it is important to start developing a systematic approach to identify and tackle those ethical issues which rise with cHRI in kindergarten settings on a more global level and address the impact of the technology from a macroperspective beyond the effects on the individual. Based on our experience in conducting studies with children in general and pedagogical considerations on the role of the institution of kindergarten in specific, in this paper, we enfold some relevant aspects that have barely been addressed in an explicit way in current cHRI research. Four areas are analyzed and key ethical issues are identified in each area: (1) the institutional setting of a kindergarten, (2) children as a vulnerable group, (3) the caregivers' role, and (4) pedagogical concepts. With our considerations, we aim at (i) broadening the methodology of the current studies within the area of cHRI, (ii) revalidate it based on our comprehensive empirical experience with research in kindergarten settings, both laboratory and real-world contexts, and (iii) provide a framework for the development of a more systematic approach to address the ethical issues in cHRI research within kindergarten settings.
1
0
0
0
0
0
0
0
0
0
0
0
0
0
Combined invariants to similarity transformation and to blur using orthogonal Zernike moments. The derivation of moment invariants has been extensively investigated in the past decades. In this paper, we construct a set of invariants derived from Zernike moments which is simultaneously invariant to similarity transformation and to convolution with circularly symmetric point spread function (PSF). Two main contributions are provided: the theoretical framework for deriving the Zernike moments of a blurred image and the way to construct the combined geometric-blur invariants. The performance of the proposed descriptors is evaluated with various PSFs and similarity transformations. The comparison of the proposed method with the existing ones is also provided in terms of pattern recognition accuracy, template matching and robustness to noise. Experimental results show that the proposed descriptors perform on the overall better.
Fast computation of Jacobi-Fourier moments for invariant image recognition The Jacobi-Fourier moments (JFMs) provide a wide class of orthogonal rotation invariant moments (ORIMs) which are useful for many image processing, pattern recognition and computer vision applications. They, however, suffer from high time complexity and numerical instability at high orders of moment. In this paper, a fast method based on the recursive computation of radial kernel function of JFMs is proposed which not only reduces time complexity but also improves their numerical stability. Fast recursive method for the computation of Jacobi-Fourier moments is proposed.The proposed method not only reduces time complexity but also improves numerical stability of moments.Better image reconstruction is achieved with lower reconstruction error.Proposed method is useful for many image processing, pattern recognition and computer vision applications.
Magnitude-Phase of Quaternion Wavelet Transform for Texture Representation Using Multilevel Copula A Copula-based joint statistical model is proposed for texture representation on the domain of quaternion wavelet transform (QWT). Both the magnitude and the three phases of QWT are used as the discriminative information for constructing a 2-level Copula model, which consists of inner and outer Copulas. The inner Copula is applied to the intra-scale dependence while the outer Copula to capture the magnitude-phase dependence. Texture retrieval experiments based on Bayesian framework are carried out to evaluate our model. Experimental results show that, based on the multilevel Copula model, the performance of texture representation is improved by taking into account both the magnitude-phase dependence and the intra-scale dependence. © 1994-2012 IEEE.
Image representation by harmonic transforms with parameters in SL(2, R). •New kinds of invariant harmonic transforms with parameters in SL(2,R) are proposed.•The capabilities of the proposed PLCT and the 2-D LCTS on image representation are analyzed.
Lossless medical image watermarking method based on significant difference of cellular automata transform coefficient. Conventional medical image watermarking techniques focus on improving invisibility and robustness of the watermarking mechanism to prevent medical disputes. This paper proposes a medical image watermarking algorithm based on the significant difference of cellular automata transform (CAT) for copyright protection. The medical image is firstly subsampled into four subimages, and two images are randomly chosen to obtain two low-frequency bandwidths using CAT. Coefficients within a low-frequency bandwidth are important information in an image. Hence, the difference between two low-frequency bandwidths is used as an important feature in the medical image. From these important features, watermarks and cover images can be used to generate an ownership share image (OSI) used for verifying the medical image. Besides appearing like cover images, the OSI will also be able to register with a third party. When a suspected medical image requires verification, the important features from the suspected medical image are first extracted. The master share image (MSI) can be generated from the important features from the suspected medical image. Lastly, the OSI and MSI can be combined to extract the watermark to verify the suspected medical image. The advantage of our proposed method is that the medical image does not require alteration to protect the copyright of the medical image. This means that while the image is protected, medical disputes will be unlikely and the appearance of the registered OSI will carry significant data to make management more convenient. Lastly, the proposed method has the features of having better security, invisibility, and robustness. Moreover, experimental results have demonstrated that our method results in good performance.
Robust watermarking scheme for color image based on quaternion-type moment invariants and visual cryptography. This paper introduces a novel robust watermarking scheme for copyright protection of color image based on quaternion-type moment invariants and visual cryptography. As a secure way to allow secret sharing of images, visual cryptography realizes encryption of classified information and the decryption is performed through human visual system. The proposed scheme represents the color image into a quaternion matrix, so that it can deal with the multichannel information in a holistic way. Then the quaternion moments are applied to extract the invariant features, which are crucial to generate the master share. Together with the scrambled watermark, they are used for constructing the ownership share based on visual cryptography. Afterwards, the ownership share is registered and responsible for authentication. A set of experiments has been conducted to illustrate the validity and feasibility of the proposed scheme as well as better robustness against different attacks.
Two-Dimensional Polar Harmonic Transforms for Invariant Image Representation This paper introduces a set of 2D transforms, based on a set of orthogonal projection bases, to generate a set of features which are invariant to rotation. We call these transforms Polar Harmonic Transforms (PHTs). Unlike the well-known Zernike and pseudo-Zernike moments, the kernel computation of PHTs is extremely simple and has no numerical stability issue whatsoever. This implies that PHTs encompass the orthogonality and invariance advantages of Zernike and pseudo-Zernike moments, but are free from their inherent limitations. This also means that PHTs are well suited for application where maximal discriminant information is needed. Furthermore, PHTs make available a large set of features for further feature selection in the process of seeking for the best discriminative or representative features for a particular application.
Tabu Search - Part I
Joint Optimization of Radio and Computational Resources for Multicell Mobile-Edge Computing Migrating computational intensive tasks from mobile devices to more resourceful cloud servers is a promising technique to increase the computational capacity of mobile devices while saving their battery energy. In this paper, we consider a MIMO multicell system where multiple mobile users (MUs) ask for computation offloading to a common cloud server. We formulate the offloading problem as the joint optimization of the radio resources􀀀the transmit precoding matrices of the MUs􀀀and the computational resources􀀀the CPU cycles/second assigned by the cloud to each MU􀀀in order to minimize the overall users’ energy consumption, while meeting latency constraints. The resulting optimization problem is nonconvex (in the objective function and constraints). Nevertheless, in the single-user case, we are able to compute the global optimal solution in closed form. In the more challenging multiuser scenario, we propose an iterative algorithm, based on a novel successive convex approximation technique, converging to a local optimal solution of the original nonconvex problem. We then show that the proposed algorithmic framework naturally leads to a distributed and parallel implementation across the radio access points, requiring only a limited coordination/signaling with the cloud. Numerical results show that the proposed schemes outperform disjoint optimization algorithms.
Symbolic model checking for real-time systems We describe finite-state programs over real-numbered time in a guarded-command language with real-valued clocks or, equivalently, as finite automata with real-valued clocks. Model checking answers the question which states of a real-time program satisfy a branching-time specification (given in an extension of CTL with clock variables). We develop an algorithm that computes this set of states symbolically as a fixpoint of a functional on state predicates, without constructing the state space. For this purpose, we introduce a μ-calculus on computation trees over real-numbered time. Unfortunately, many standard program properties, such as response for all nonzeno execution sequences (during which time diverges), cannot be characterized by fixpoints: we show that the expressiveness of the timed μ-calculus is incomparable to the expressiveness of timed CTL. Fortunately, this result does not impair the symbolic verification of "implementable" real-time programs-those whose safety constraints are machine-closed with respect to diverging time and whose fairness constraints are restricted to finite upper bounds on clock values. All timed CTL properties of such programs are shown to be computable as finitely approximable fixpoints in a simple decidable theory.
Teaching-Learning-Based Optimization: An optimization method for continuous non-linear large scale problems An efficient optimization method called 'Teaching-Learning-Based Optimization (TLBO)' is proposed in this paper for large scale non-linear optimization problems for finding the global solutions. The proposed method is based on the effect of the influence of a teacher on the output of learners in a class. The basic philosophy of the method is explained in detail. The effectiveness of the method is tested on many benchmark problems with different characteristics and the results are compared with other population based methods.
An Automatic Screening Approach for Obstructive Sleep Apnea Diagnosis Based on Single-Lead Electrocardiogram Traditional approaches for obstructive sleep apnea (OSA) diagnosis are apt to using multiple channels of physiological signals to detect apnea events by dividing the signals into equal-length segments, which may lead to incorrect apnea event detection and weaken the performance of OSA diagnosis. This paper proposes an automatic-segmentation-based screening approach with the single channel of Electrocardiogram (ECG) signal for OSA subject diagnosis, and the main work of the proposed approach lies in three aspects: (i) an automatic signal segmentation algorithm is adopted for signal segmentation instead of the equal-length segmentation rule; (ii) a local median filter is improved for reduction of the unexpected RR intervals before signal segmentation; (iii) the designed OSA severity index and additional admission information of OSA suspects are plugged into support vector machine (SVM) for OSA subject diagnosis. A real clinical example from PhysioNet database is provided to validate the proposed approach and an average accuracy of 97.41% for subject diagnosis is obtained which demonstrates the effectiveness for OSA diagnosis.
Automated Detection of Obstructive Sleep Apnea Events from a Single-Lead Electrocardiogram Using a Convolutional Neural Network. In this study, we propose a method for the automated detection of obstructive sleep apnea (OSA) from a single-lead electrocardiogram (ECG) using a convolutional neural network (CNN). A CNN model was designed with six optimized convolution layers including activation, pooling, and dropout layers. One-dimensional (1D) convolution, rectified linear units (ReLU), and max pooling were applied to the convolution, activation, and pooling layers, respectively. For training and evaluation of the CNN model, a single-lead ECG dataset was collected from 82 subjects with OSA and was divided into training (including data from 63 patients with 34,281 events) and testing (including data from 19 patients with 8571 events) datasets. Using this CNN model, a precision of 0.99%, a recall of 0.99%, and an F-score of 0.99% were attained with the training dataset; these values were all 0.96% when the CNN was applied to the testing dataset. These results show that the proposed CNN model can be used to detect OSA accurately on the basis of a single-lead ECG. Ultimately, this CNN model may be used as a screening tool for those suspected to suffer from OSA.
Hardware Circuits Design and Performance Evaluation of a Soft Lower Limb Exoskeleton Soft lower limb exoskeletons (LLEs) are wearable devices that have good potential in walking rehabilitation and augmentation. While a few studies focused on the structure design and assistance force optimization of the soft LLEs, rarely work has been conducted on the hardware circuits design. The main purpose of this work is to present a new soft LLE for walking efficiency improvement and introduce its hardware circuits design. A soft LLE for hip flexion assistance and a hardware circuits system with scalability were proposed. To assess the efficacy of the soft LLE, the experimental tests that evaluate the sensor data acquisition, force tracking performance, lower limb muscle activity and metabolic cost were conducted. The time error in the peak assistance force was just 1%. The reduction in the normalized root-mean-square EMG of the rectus femoris was 7.1%. The net metabolic cost in exoskeleton on condition was reduced by 7.8% relative to walking with no exoskeleton. The results show that the designed hardware circuits can be applied to the soft LLE and the soft LLE is able to improve walking efficiency of wearers.
1.2
0.2
0.2
0.2
0.2
0.1
0.022222
0
0
0
0
0
0
0
Energy efficient scheduling of open-pit coal mine trucks. •Open-pit coal mine truck activity is scheduled to minimise energy consumption.•A solution technique is developed to solve the model in reasonable time.•A case study is performed to validate the practical utility of the methodologies.•Opportunities for aiding short to long term decisions are presented.
Development of a scenario-based robust model for the optimal truck-shovel allocation in open-pit mining •A scenario-based robust stochastic optimization model for optimal truck–shovel allocation is developed.•The model selects the shortest path between origins and destinations to minimize the operational cycle and costs of trucks.•The model is applicable for trucks with different capacities and different paths between origins and destinations.•Compared with the current status of the studied mine, shovel output was increased and operational costs were decreased.
Simulation-based optimization approach for material dispatching in continuous mining systems •Simulation-based optimization approach is capable of optimizing dispatch decisions.•It is highly efficient for short-term production scheduling of opencast mines.•It combines deterministic optimization with stochastic simulation in a closed loop.•Number of feasible solutions increases as the sim-opt iteration loop progresses.
Parallel End-to-End Autonomous Mining: An IoT-Oriented Approach. This article proposes a new solution for end-to-end autonomous mining operations: Internet of Things (IoT)-based parallel mining, consisting of the concept definition, the solution given, and the concrete realization. The proposed parallel mining is inspired by the artificial societies (A) for modeling, computational experiments (C) for analysis, and parallel execution (P) for control (ACP) approa...
Personalized Driver/Vehicle Lane Change Models for ADAS Lane changes are stressful maneuvers for drivers, especially during high speed traffic flows. Advanced Driver Assistance Systems (ADAS) aim to assist drivers during lane change maneuvers. A system that is developed for an average driver or all drivers will have to be conservative for safety reasons in order to cover all driver/vehicle types. Such a conservative system may not be acceptable to aggressive drivers and could be perceived as too aggressive by the more passive drivers. An ADAS that takes into account the dynamics and characteristics of each individual vehicle/driver system during lane change maneuvers will be more effective and more acceptable to drivers without sacrificing safety. In this paper, we develop a methodology which learns the characteristics of an individual driver/vehicle response before and during lane changes and under different driving environments. These characteristics are captured by a set of models whose parameters are adjusted online to fit the individual vehicle/driver response during lane changes. We develop a two-layer model to describe the maneuver kinematics. The lower layer describes lane change as a kinematic model. The higher layer model establishes the kinematic model parameter values for the particular driver and represents their dependence on the configuration of the surrounding vehicles. The proposed modeling framework can be used as a kernel component of ADAS to provide more personalized recommendations to the driver, increasing the potential for more widespread acceptance and use of ADAS. We evaluated the proposed methodology using an actual vehicle and three different drivers. We demonstrated that the method is effective in modeling individual driver/vehicle responses during lane change by showing a consistency of matching between the model outputs and raw data.
Parallel learning: a perspective and a framework The development of machine learning in complex system is hindered by two problems nowadays. The first problem is the inefficiency of exploration in state and action space, which leads to the data-hungry of some state-of-art data-driven algorithm. The second problem is the lack of a general theory which can be used to analyze and implement a complex learning system. In this paper, we proposed a gen...
Probabilistic encryption A new probabilistic model of data encryption is introduced. For this model, under suitable complexity assumptions, it is proved that extracting any information about the cleartext from the cyphertext is hard on the average for an adversary with polynomially bounded computational resources. The proof holds for any message space with any probability distribution. The first implementation of this model is presented. The security of this implementation is proved under the interactability assumptin of deciding Quadratic Residuosity modulo composite numbers whose factorization is unknown.
ImageNet Classification with Deep Convolutional Neural Networks. We trained a large, deep convolutional neural network to classify the 1.2 million high-resolution images in the ImageNet LSVRC-2010 contest into the 1000 different classes. On the test data, we achieved top-1 and top-5 error rates of 37.5% and 17.0%, respectively, which is considerably better than the previous state-of-the-art. The neural network, which has 60 million parameters and 650,000 neurons, consists of five convolutional layers, some of which are followed by max-pooling layers, and three fully connected layers with a final 1000-way softmax. To make training faster, we used non-saturating neurons and a very efficient GPU implementation of the convolution operation. To reduce overfitting in the fully connected layers we employed a recently developed regularization method called \"dropout\" that proved to be very effective. We also entered a variant of this model in the ILSVRC-2012 competition and achieved a winning top-5 test error rate of 15.3%, compared to 26.2% achieved by the second-best entry.
The Whale Optimization Algorithm. The Whale Optimization Algorithm inspired by humpback whales is proposed.The WOA algorithm is benchmarked on 29 well-known test functions.The results on the unimodal functions show the superior exploitation of WOA.The exploration ability of WOA is confirmed by the results on multimodal functions.The results on structural design problems confirm the performance of WOA in practice. This paper proposes a novel nature-inspired meta-heuristic optimization algorithm, called Whale Optimization Algorithm (WOA), which mimics the social behavior of humpback whales. The algorithm is inspired by the bubble-net hunting strategy. WOA is tested with 29 mathematical optimization problems and 6 structural design problems. Optimization results prove that the WOA algorithm is very competitive compared to the state-of-art meta-heuristic algorithms as well as conventional methods. The source codes of the WOA algorithm are publicly available at http://www.alimirjalili.com/WOA.html
Scalable and efficient provable data possession. Storage outsourcing is a rising trend which prompts a number of interesting security issues, many of which have been extensively investigated in the past. However, Provable Data Possession (PDP) is a topic that has only recently appeared in the research literature. The main issue is how to frequently, efficiently and securely verify that a storage server is faithfully storing its client's (potentially very large) outsourced data. The storage server is assumed to be untrusted in terms of both security and reliability. (In other words, it might maliciously or accidentally erase hosted data; it might also relegate it to slow or off-line storage.) The problem is exacerbated by the client being a small computing device with limited resources. Prior work has addressed this problem using either public key cryptography or requiring the client to outsource its data in encrypted form. In this paper, we construct a highly efficient and provably secure PDP technique based entirely on symmetric key cryptography, while not requiring any bulk encryption. Also, in contrast with its predecessors, our PDP technique allows outsourcing of dynamic data, i.e, it efficiently supports operations, such as block modification, deletion and append.
Well-Solvable Special Cases of the Traveling Salesman Problem: A Survey. The traveling salesman problem (TSP) belongs to the most basic, most important, and most investigated problems in combinatorial optimization. Although it is an ${\cal NP}$-hard problem, many of its special cases can be solved efficiently in polynomial time. We survey these special cases with emphasis on the results that have been obtained during the decade 1985--1995. This survey complements an earlier survey from 1985 compiled by Gilmore, Lawler, and Shmoys [The Traveling Salesman Problem---A Guided Tour of Combinatorial Optimization, Wiley, Chichester, pp. 87--143].
Wireless Networks with RF Energy Harvesting: A Contemporary Survey Radio frequency (RF) energy transfer and harvesting techniques have recently become alternative methods to power the next generation wireless networks. As this emerging technology enables proactive energy replenishment of wireless devices, it is advantageous in supporting applications with quality of service (QoS) requirements. In this paper, we present a comprehensive literature review on the research progresses in wireless networks with RF energy harvesting capability, referred to as RF energy harvesting networks (RF-EHNs). First, we present an overview of the RF-EHNs including system architecture, RF energy harvesting techniques and existing applications. Then, we present the background in circuit design as well as the state-of-the-art circuitry implementations, and review the communication protocols specially designed for RF-EHNs. We also explore various key design issues in the development of RFEHNs according to the network types, i.e., single-hop networks, multi-antenna networks, relay networks, and cognitive radio networks. Finally, we envision some open research directions.
Multi-stream CNN: Learning representations based on human-related regions for action recognition. •Presenting a multi-stream CNN architecture to incorporate multiple complementary features trained in appearance and motion networks.•Demonstrating that using full-frame, human body, and motion-salient body part regions together is effective to improve recognition performance.•Proposing methods to detect the actor and motion-salient body part precisely.•Verifying that high-quality flow is critically important to learn accurate video representations for action recognition.
Social Robots for (Second) Language Learning in (Migrant) Primary School Children Especially these days, innovation and support from technology to relieve pressure in education is highly urgent. This study tested the potential advantage of a social robot over a tablet in (second) language learning on performance, engagement, and enjoyment. Shortages in primary education call for new technology solutions. Previous studies combined robots with tablets, to compensate for robot’s limitations, however, this study applied direct human–robot interaction. Primary school children (N = 63, aged 4–6) participated in a 3-wave field experiment with story-telling exercises, either with a semi-autonomous robot (without tablet, using WOz) or a tablet. Results showed increased learning gains over time when training with a social robot, compared to the tablet. Children who trained with a robot were more engaged in the story-telling task and enjoyed it more. Robot’s behavioral style (social or neutral) hardly differed overall, however, seems to vary for high versus low educational abilities. While social robots need sophistication before being implemented in schools, our study shows the potential of social robots as tutors in (second) language learning.
1.24
0.24
0.24
0.083333
0.004444
0.002222
0
0
0
0
0
0
0
0
Optimal Scheduling for Electric Vehicle Charging With Discrete Charging Levels in Distribution Grid. To accommodate the increasing electric vehicle (EV) penetration in distribution grid, coordinated EV charging has been extensively studied in the literature. However, most of the existing works optimistically consider the EV charging rate as a continuous variable and implicitly ignore the capacity limitation in distribution transformers, which both have great impact on the efficiency and stability...
Smoothed Least-laxity-first Algorithm for EV Charging. We formulate EV charging as a feasibility problem that meets all EVs' energy demands before departure under charging rate constraints and total power constraint. We propose an online algorithm, the smoothed least-laxity-first (sLLF) algorithm, that decides on the current charging rates based on only the information up to the current time. We characterize the performance of the sLLF algorithm analytically and numerically. Numerical experiments with real-world data show that it has significantly higher rate of generating feasible EV charging than several other common EV charging algorithms.
Robust Online Algorithms for Peak-Minimizing EV Charging Under Multistage Uncertainty. In this paper, we study how to utilize forecasts to design online electrical vehicle (EV) charging algorithms that can attain strong performance guarantees. We consider the scenario of an aggregator serving a large number of EVs together with its background load, using both its own renewable energy (for free) and the energy procured from the external grid. The goal of the aggregator is to minimize its peak procurement from the grid, subject to the constraint that each EV has to be fully charged before its deadline. Further, the aggregator can predict the future demand and the renewable energy supply with some levels of uncertainty. We show that such prediction can be very effective in reducing the competitive ratios of online control algorithms, and even allow online algorithms to achieve close-to-offline-optimal peak. Specifically, we first propose a 2-level increasing precision model (2-IPM), to model forecasts with different levels of accuracy. We then develop a powerful computational approach that can compute the optimal competitive ratio under 2-IPM over any online algorithm, and also online algorithms that can achieve the optimal competitive ratio. Simulation results show that, even with up to 20% day-ahead prediction errors, our online algorithms still achieve competitive ratios fairly close to 1, which are much better than the classic results in the literature with a competitive ratio of e. The second contribution of this paper is that we solve a dilemma for online algorithm design, e.g., an online algorithm with good competitive ratio may exhibit poor average-case performance. We propose a new Algorithm-Robustification procedure that can convert an online algorithm with good average-case performance to one with both the optimal competitive ratio and good average-case performance. We demonstrate via trace-based simulations the superior performance of the robustified version of a well-known heuristic algorithm based on model predictive control.
Definition and Evaluation of Model-Free Coordination of Electrical Vehicle Charging With Reinforcement Learning Initial DR studies mainly adopt model predictive control and thus require accurate models of the control problem (e.g., a customer behavior model), which are to a large extent uncertain for the EV scenario. Hence, model-free approaches, especially based on reinforcement learning (RL) are an attractive alternative. In this paper, we propose a new Markov decision process (MDP) formulation in the RL framework, to jointly coordinate a set of EV charging stations. State-of-the-art algorithms either focus on a single EV, or perform the control of an aggregate of EVs in multiple steps (e.g., aggregate load decisions in one step, then a step translating the aggregate decision to individual connected EVs). On the contrary, we propose an RL approach to jointly control the whole set of EVs at once. We contribute a new MDP formulation, with a scalable state representation that is independent of the number of EV charging stations. Further, we use a batch reinforcement learning algorithm, i.e., an instance of fitted Q-iteration, to learn the optimal charging policy. We analyze its performance using simulation experiments based on a real-world EV charging data. More specifically, we (i) explore the various settings in training the RL policy (e.g., duration of the period with training data), (ii) compare its performance to an oracle all-knowing benchmark (which provides an upper bound for performance, relying on information that is not available or at least imperfect in practice), (iii) analyze performance over time, over the course of a full year to evaluate possible performance fluctuations (e.g, across different seasons), and (iv) demonstrate the generalization capacity of a learned control policy to larger sets of charging stations.
ACN-Data - Analysis and Applications of an Open EV Charging Dataset. We are releasing ACN-Data, a dynamic dataset of workplace EV charging which currently includes over 30,000 sessions with more added daily. In this paper we describe the dataset, as well as some interesting user behavior it exhibits. To demonstrate the usefulness of the dataset, we present three examples, learning and predicting user behavior using Gaussian mixture models, optimally sizing on-site solar generation for adaptive electric vehicle charging, and using workplace charging to smooth the net demand Duck Curve.
Hierarchical Electric Vehicle Charging Aggregator Strategy Using Dantzig-Wolfe Decomposition. This article focuses on reducing a charging cost for electric vehicles (EVs). A charging strategy is proposed to minimize the charging cost of EVs within the charging station constraints.
The Sybil Attack Large-scale peer-to-peer systems facesecurity threats from faulty or hostile remotecomputing elements. To resist these threats, manysuch systems employ redundancy. However, if asingle faulty entity can present multiple identities,it can control a substantial fraction of the system,thereby undermining this redundancy. Oneapproach to preventing these &quot;Sybil attacks&quot; is tohave a trusted agency certify identities. Thispaper shows that, without a logically centralizedauthority, Sybil...
Long short-term memory. Learning to store information over extended time intervals by recurrent backpropagation takes a very long time, mostly because of insufficient, decaying error backflow. We briefly review Hochreiter's (1991) analysis of this problem, then address it by introducing a novel, efficient, gradient-based method called long short-term memory (LSTM). Truncating the gradient where this does not do harm, LSTM can learn to bridge minimal time lags in excess of 1000 discrete-time steps by enforcing constant error flow through constant error carousels within special units. Multiplicative gate units learn to open and close access to the constant error flow. LSTM is local in space and time; its computational complexity per time step and weight is O(1). Our experiments with artificial data involve local, distributed, real-valued, and noisy pattern representations. In comparisons with real-time recurrent learning, back propagation through time, recurrent cascade correlation, Elman nets, and neural sequence chunking, LSTM leads to many more successful runs, and learns much faster. LSTM also solves complex, artificial long-time-lag tasks that have never been solved by previous recurrent network algorithms.
Toward Integrating Vehicular Clouds with IoT for Smart City Services Vehicular ad hoc networks, cloud computing, and the Internet of Things are among the emerging technology enablers offering a wide array of new application possibilities in smart urban spaces. These applications consist of smart building automation systems, healthcare monitoring systems, and intelligent and connected transportation, among others. The integration of IoT-based vehicular technologies will enrich services that are eventually going to ignite the proliferation of exciting and even more advanced technological marvels. However, depending on different requirements and design models for networking and architecture, such integration needs the development of newer communication architectures and frameworks. This work proposes a novel framework for architectural and communication design to effectively integrate vehicular networking clouds with IoT, referred to as VCoT, to materialize new applications that provision various IoT services through vehicular clouds. In this article, we particularly put emphasis on smart city applications deployed, operated, and controlled through LoRaWAN-based vehicular networks. LoraWAN, being a new technology, provides efficient and long-range communication possibilities. The article also discusses possible research issues in such an integration including data aggregation, security, privacy, data quality, and network coverage. These issues must be addressed in order to realize the VCoT paradigm deployment, and to provide insights for investors and key stakeholders in VCoT service provisioning. The article presents deep insights for different real-world application scenarios (i.e., smart homes, intelligent traffic light, and smart city) using VCoT for general control and automation along with their associated challenges. It also presents initial insights, through preliminary results, regarding data and resource management in IoT-based resource constrained environments through vehicular clouds.
Distributed multirobot localization In this paper, we present a new approach to the problem of simultaneously localizing a group of mobile robots capable of sensing one another. Each of the robots collects sensor data regarding its own motion and shares this information with the rest of the team during the update cycles. A single estimator, in the form of a Kalman filter, processes the available positioning information from all the members of the team and produces a pose estimate for every one of them. The equations for this centralized estimator can be written in a decentralized form, therefore allowing this single Kalman filter to be decomposed into a number of smaller communicating filters. Each of these filters processes the sensor data collected by its host robot. Exchange of information between the individual filters is necessary only when two robots detect each other and measure their relative pose. The resulting decentralized estimation schema, which we call collective localization, constitutes a unique means for fusing measurements collected from a variety of sensors with minimal communication and processing requirements. The distributed localization algorithm is applied to a group of three robots and the improvement in localization accuracy is presented. Finally, a comparison to the equivalent decentralized information filter is provided.
A simplified dual neural network for quadratic programming with its KWTA application. The design, analysis, and application of a new recurrent neural network for quadratic programming, called simplified dual neural network, are discussed. The analysis mainly concentrates on the convergence property and the computational complexity of the neural network. The simplified dual neural network is shown to be globally convergent to the exact optimal solution. The complexity of the neural network architecture is reduced with the number of neurons equal to the number of inequality constraints. Its application to k-winners-take-all (KWTA) operation is discussed to demonstrate how to solve problems with this neural network.
NETWRAP: An NDN Based Real-TimeWireless Recharging Framework for Wireless Sensor Networks Using vehicles equipped with wireless energy transmission technology to recharge sensor nodes over the air is a game-changer for traditional wireless sensor networks. The recharging policy regarding when to recharge which sensor nodes critically impacts the network performance. So far only a few works have studied such recharging policy for the case of using a single vehicle. In this paper, we propose NETWRAP, an N DN based Real Time Wireless Rech arging Protocol for dynamic wireless recharging in sensor networks. The real-time recharging framework supports single or multiple mobile vehicles. Employing multiple mobile vehicles provides more scalability and robustness. To efficiently deliver sensor energy status information to vehicles in real-time, we leverage concepts and mechanisms from named data networking (NDN) and design energy monitoring and reporting protocols. We derive theoretical results on the energy neutral condition and the minimum number of mobile vehicles required for perpetual network operations. Then we study how to minimize the total traveling cost of vehicles while guaranteeing all the sensor nodes can be recharged before their batteries deplete. We formulate the recharge optimization problem into a Multiple Traveling Salesman Problem with Deadlines (m-TSP with Deadlines), which is NP-hard. To accommodate the dynamic nature of node energy conditions with low overhead, we present an algorithm that selects the node with the minimum weighted sum of traveling time and residual lifetime. Our scheme not only improves network scalability but also ensures the perpetual operation of networks. Extensive simulation results demonstrate the effectiveness and efficiency of the proposed design. The results also validate the correctness of the theoretical analysis and show significant improvements that cut the number of nonfunctional nodes by half compared to the static scheme while maintaining the network overhead at the same level.
Finite-Time Adaptive Fuzzy Tracking Control Design for Nonlinear Systems. This paper addresses the finite-time tracking problem of nonlinear pure-feedback systems. Unlike the literature on traditional finite-time stabilization, in this paper the nonlinear system functions, including the bounding functions, are all totally unknown. Fuzzy logic systems are used to model those unknown functions. To present a finite-time control strategy, a criterion of semiglobal practical...
Convert Harm Into Benefit: A Coordination-Learning Based Dynamic Spectrum Anti-Jamming Approach This paper mainly investigates the multi-user anti-jamming spectrum access problem. Using the idea of “converting harm into benefit,” the malicious jamming signals projected by the enemy are utilized by the users as the coordination signals to guide spectrum coordination. An “internal coordination-external confrontation” multi-user anti-jamming access game model is constructed, and the existence of Nash equilibrium (NE) as well as correlated equilibrium (CE) is demonstrated. A coordination-learning based anti-jamming spectrum access algorithm (CLASA) is designed to achieve the CE of the game. Simulation results show the convergence, and effectiveness of the proposed CLASA algorithm, and indicate that our approach can help users confront the malicious jammer, and coordinate internal spectrum access simultaneously without information exchange. Last but not least, the fairness of the proposed approach under different jamming attack patterns is analyzed, which illustrates that this approach provides fair anti-jamming spectrum access opportunities under complicated jamming pattern.
1.2
0.2
0.2
0.2
0.2
0.2
0
0
0
0
0
0
0
0
A multi-objective green UAV routing problem. •Introduce a new time-dependent UAV heterogeneous fleet routing problem.•Consider several objective functions and respect drones operational requirements.•Design a MILP model in order to find sets of non-dominated solutions.•Consider a model able to tackle multi-layer scenarios with package exchanging points.•Integrate UAV into the new concepts of mini/microgrid systems inside smart cities.
People detection and tracking from aerial thermal views Detection and tracking of people in visible-light images has been subject to extensive research in the past decades with applications ranging from surveillance to search-and-rescue. Following the growing availability of thermal cameras and the distinctive thermal signature of humans, research effort has been focusing on developing people detection and tracking methodologies applicable to this sensing modality. However, a plethora of challenges arise on the transition from visible-light to thermal images, especially with the recent trend of employing thermal cameras onboard aerial platforms (e.g. in search-and-rescue research) capturing oblique views of the scenery. This paper presents a new, publicly available dataset of annotated thermal image sequences, posing a multitude of challenges for people detection and tracking. Moreover, we propose a new particle filter based framework for tracking people in aerial thermal images. Finally, we evaluate the performance of this pipeline on our dataset, incorporating a selection of relevant, state-of-the-art methods and present a comprehensive discussion of the merits spawning from our study.
Pareto-Optimization for Scheduling of Crude Oil Operations in Refinery via Genetic Algorithm. With the interaction of discrete-event and continuous processes, it is challenging to schedule crude oil operations in a refinery. This paper studies the optimization problem of finding a detailed schedule to realize a given refining schedule. This is a multiobjective optimization problem with a combinatorial nature. Since the original problem cannot be directly solved by using heuristics and meta-heuristics, the problem is transformed into an assignment problem of charging tanks and distillers. Based on such a transformation, by analyzing the properties of the problem, this paper develops a chromosome that can describe a feasible schedule such that meta-heuristics can be applied. Then, it innovatively adopts an improved nondominated sorting genetic algorithm to solve the problem for the first time. An industrial case study is used to test the proposed solution method. The results show that the method makes a significant performance improvement and is applicable to real-life refinery scheduling problems.
The unmanned aerial vehicle routing and trajectory optimisation problem, a taxonomic review. •We introduce the UAV Routing and Trajectory Optimisation Problem.•We provide a taxonomy for UAV routing, TO and other variants.•We apply the proposed taxonomy to 70 scientific articles.•A lack of research about integrating UAV routing and TO is identified.
A vehicle routing problem arising in unmanned aerial monitoring. •We consider a routing problem arising in unmanned aerial monitoring.•The problem possesses several military and civil applications.•The problem is to construct vehicle routes and to select monitoring heights.•We model the problem as an integer linear program and solve it by tabu search.•Extensive tests confirm the efficiency of the heuristic.
Taking Drones to the Next Level: Cooperative Distributed Unmanned-Aerial-Vehicular Networks for Small and Mini Drones Unmanned aerial vehicles (UAVs) have been widely used in both military and civilian applications. However, the cooperation of small and mini drones in a network is capable of further improving the performance and the coverage area of UAVs. There are numerous new challenges to be solved before the widespread introduction of multi-UAV-based heterogeneous flying ad hoc networks (FANET), including the...
Completely derandomized self-adaptation in evolution strategies. This paper puts forward two useful methods for self-adaptation of the mutation distribution - the concepts of derandomization and cumulation. Principle shortcomings of the concept of mutative strategy parameter control and two levels of derandomization are reviewed. Basic demands on the self-adaptation of arbitrary (normal) mutation distributions are developed. Applying arbitrary, normal mutation distributions is equivalent to applying a general, linear problem encoding. The underlying objective of mutative strategy parameter control is roughly to favor previously selected mutation steps in the future. If this objective is pursued rigorously, a completely derandomized self-adaptation scheme results, which adapts arbitrary normal mutation distributions. This scheme, called covariance matrix adaptation (CMA), meets the previously stated demands. It can still be considerably improved by cumulation - utilizing an evolution path rather than single search steps. Simulations on various test functions reveal local and global search properties of the evolution strategy with and without covariance matrix adaptation. Their performances are comparable only on perfectly scaled functions. On badly scaled, non-separable functions usually a speed up factor of several orders of magnitude is observed. On moderately mis-scaled functions a speed up factor of three to ten can be expected.
A survey of socially interactive robots This paper reviews “socially interactive robots”: robots for which social human–robot interaction is important. We begin by discussing the context for socially interactive robots, emphasizing the relationship to other research fields and the different forms of “social robots”. We then present a taxonomy of design methods and system components used to build socially interactive robots. Finally, we describe the impact of these robots on humans and discuss open issues. An expanded version of this paper, which contains a survey and taxonomy of current applications, is available as a technical report [T. Fong, I. Nourbakhsh, K. Dautenhahn, A survey of socially interactive robots: concepts, design and applications, Technical Report No. CMU-RI-TR-02-29, Robotics Institute, Carnegie Mellon University, 2002].
IoT-U: Cellular Internet-of-Things Networks Over Unlicensed Spectrum. In this paper, we consider an uplink cellular Internet-of-Things (IoT) network, where a cellular user (CU) can serve as the mobile data aggregator for a cluster of IoT devices. To be specific, the IoT devices can either transmit the sensory data to the base station (BS) directly by cellular communications, or first aggregate the data to a CU through machine-to-machine (M2M) communications before t...
Multi-column Deep Neural Networks for Image Classification Traditional methods of computer vision and machine learning cannot match human performance on tasks such as the recognition of handwritten digits or traffic signs. Our biologically plausible deep artificial neural network architectures can. Small (often minimal) receptive fields of convolutional winner-take-all neurons yield large network depth, resulting in roughly as many sparsely connected neural layers as found in mammals between retina and visual cortex. Only winner neurons are trained. Several deep neural columns become experts on inputs preprocessed in different ways; their predictions are averaged. Graphics cards allow for fast training. On the very competitive MNIST handwriting benchmark, our method is the first to achieve near-human performance. On a traffic sign recognition benchmark it outperforms humans by a factor of two. We also improve the state-of-the-art on a plethora of common image classification benchmarks.
A novel full structure optimization algorithm for radial basis probabilistic neural networks. In this paper, a novel full structure optimization algorithm for radial basis probabilistic neural networks (RBPNN) is proposed. Firstly, a minimum volume covering hyperspheres (MVCH) algorithm is proposed to heuristically select the initial hidden layer centers of the RBPNN, and then the recursive orthogonal least square (ROLS) algorithm combined with the particle swarm optimization (PSO) algorithm is adopted to further optimize the initial structure of the RBPNN. Finally, the effectiveness and efficiency of our proposed algorithm are evaluated through a plant species identification task involving 50 plant species.
An Automatic Screening Approach for Obstructive Sleep Apnea Diagnosis Based on Single-Lead Electrocardiogram Traditional approaches for obstructive sleep apnea (OSA) diagnosis are apt to using multiple channels of physiological signals to detect apnea events by dividing the signals into equal-length segments, which may lead to incorrect apnea event detection and weaken the performance of OSA diagnosis. This paper proposes an automatic-segmentation-based screening approach with the single channel of Electrocardiogram (ECG) signal for OSA subject diagnosis, and the main work of the proposed approach lies in three aspects: (i) an automatic signal segmentation algorithm is adopted for signal segmentation instead of the equal-length segmentation rule; (ii) a local median filter is improved for reduction of the unexpected RR intervals before signal segmentation; (iii) the designed OSA severity index and additional admission information of OSA suspects are plugged into support vector machine (SVM) for OSA subject diagnosis. A real clinical example from PhysioNet database is provided to validate the proposed approach and an average accuracy of 97.41% for subject diagnosis is obtained which demonstrates the effectiveness for OSA diagnosis.
A robust medical image watermarking against salt and pepper noise for brain MRI images. The ever-growing numbers of medical digital images and the need to share them among specialists and hospitals for better and more accurate diagnosis require that patients' privacy be protected. During the transmission of medical images between hospitals or specialists through the network, the main priority is to protect a patient's documents against any act of tampering by unauthorised individuals. Because of this, there is a need for medical image authentication scheme to enable proper diagnosis on patient. In addition, medical images are also susceptible to salt and pepper impulse noise through the transmission in communication channels. This noise may also be intentionally used by the invaders to corrupt the embedded watermarks inside the medical images. A common drawback of existing watermarking methods is their weakness against salt and pepper noise. The research carried out in this work addresses the issue of designing a new watermarking method that can withstand high density of salt and pepper noise for brain MRI images. For this purpose, combination of a spatial domain watermarking method, channel coding and noise filtering schemes are used. The region of non-interest (RONI) of MRI images from five different databases are used as embedding area and electronic patient record (EPR) is considered as embedded data. The quality of watermarked image is evaluated using Peak Signal-to-Noise Ratio (PSNR) and Structural Similarity Index (SSIM), and the accuracy of the extracted watermark is assessed in terms of Bit Error Rate (BER).
A Hierarchical Architecture Using Biased Min-Consensus for USV Path Planning This paper proposes a hierarchical architecture using the biased min-consensus (BMC) method, to solve the path planning problem of unmanned surface vessel (USV). We take the fixed-point monitoring mission as an example, where a series of intermediate monitoring points should be visited once by USV. The whole framework incorporates the low-level layer planning the standard path between any two intermediate points, and the high-level fashion determining their visiting sequence. First, the optimal standard path in terms of voyage time and risk measure is planned by the BMC protocol, given that the corresponding graph is constructed with node state and edge weight. The USV will avoid obstacles or keep a certain distance safely, and arrive at the target point quickly. It is proven theoretically that the state of the graph will converge to be stable after finite iterations, i.e., the optimal solution can be found by BMC with low calculation complexity. Second, by incorporating the constraint of intermediate points, their visiting sequence is optimized by BMC again with the reconstruction of a new virtual graph based on the former planned results. The extensive simulation results in various scenarios also validate the feasibility and effectiveness of our method for autonomous navigation.
1.071111
0.066667
0.066667
0.066667
0.066667
0.022222
0
0
0
0
0
0
0
0
A new effective robust nonlinear controller based on PSO for interleaved DC-DC boost converters for fuel cell voltage regulation. Output voltage regulation of DC-DC converters has recently gained an increasing attention to face the many system nonidealities. The fast switching behavior is nonlinear time varying, the presence of model and measurement uncertainties, and large variations, are all inherited challenges. The aim of the present work is to design a robust nonlinear controller that ensures satisfactory and robust output voltage regulation for a proton-exchange membrane fuel cell (PEMFC) based on a DC-DC Interleaved Boost Converter (IBC). A state-space model of the DC-DC IBC is first derived using the state-space averaging technique, and a mathematical model is constructed for the PEFMC. In this regard, a robust nonlinear controller and a proportional integral controller are proposed. The controllers are tuned though particle swarm optimization algorithm to estimate their good parameters assuring the desired performance is met. The integral of absolute error criterion is used to improve the dynamic performance of the overall controlled system. Furthermore, the closed-loop stability is analyzed using the Lyapunov stability theorem, and the effectiveness of the closed-loop system is validated under various operating conditions of the PEMFC and load perturbations. Compared to other methods, the obtained results demonstrate a superior performance of the proposed control strategy in terms of its robustness to variations and uncertainties, smooth tracking of a varying set-point and faster transients.
Multi-stage genetic programming: A new strategy to nonlinear system modeling This paper presents a new multi-stage genetic programming (MSGP) strategy for modeling nonlinear systems. The proposed strategy is based on incorporating the individual effect of predictor variables and the interactions among them to provide more accurate simulations. According to the MSGP strategy, an efficient formulation for a problem comprises different terms. In the first stage of the MSGP-based analysis, the output variable is formulated in terms of an influencing variable. Thereafter, the error between the actual and the predicted value is formulated in terms of a new variable. Finally, the interaction term is derived by formulating the difference between the actual values and the values predicted by the individually developed terms. The capabilities of MSGP are illustrated by applying it to the formulation of different complex engineering problems. The problems analyzed herein include the following: (i) simulation of pH neutralization process, (ii) prediction of surface roughness in end milling, and (iii) classification of soil liquefaction conditions. The validity of the proposed strategy is confirmed by applying the derived models to the parts of the experimental results that were not included in the analyses. Further, the external validation of the models is verified using several statistical criteria recommended by other researchers. The MSGP-based solutions are capable of effectively simulating the nonlinear behavior of the investigated systems. The results of MSGP are found to be more accurate than those of standard GP and artificial neural network-based models.
An improved genetic algorithm with conditional genetic operators and its application to set-covering problem The genetic algorithm (GA) is a popular, biologically inspired optimization method. However, in the GA there is no rule of thumb to design the GA operators and select GA parameters. Instead, trial-and-error has to be applied. In this paper we present an improved genetic algorithm in which crossover and mutation are performed conditionally instead of probability. Because there are no crossover rate and mutation rate to be selected, the proposed improved GA can be more easily applied to a problem than the conventional genetic algorithms. The proposed improved genetic algorithm is applied to solve the set-covering problem. Experimental studies show that the improved GA produces better results over the conventional one and other methods.
Hybrid Whale Optimization Algorithm with simulated annealing for feature selection. •Four hybrid feature selection methods for classification task are proposed.•Our hybrid method combines Whale Optimization Algorithm with simulated annealing.•Eighteen UCI datasets were used in the experiments.•Our approaches result a higher accuracy by using less number of features.
Solving the dynamic weapon target assignment problem by an improved artificial bee colony algorithm with heuristic factor initialization. •Put forward an improved artificial bee colony algorithm based on ranking selection and elite guidance.•Put forward 4 rule-based heuristic factors: Wc, Rc, TRc and TRcL.•The heuristic factors are used in population initialization to improve the quality of the initial solutions in DWTA solving.•The heuristic factor initialization method is combined with the improved ABC algorithm to solve the DWTA problem.
A monarch butterfly optimization-based neural network simulator for prediction of siro-spun yarn tenacity Yarn tenacity directly affects the winding and knitting efficiency as well as warp and weft breakages during weaving process and therefore, is considered as the most important parameter to be controlled during yarn spinning process. Yarn tenacity is dependent on fiber properties and process parameters. Exploring the relationship between fiber properties, process parameters and yarn tenacity is very important to optimize the selection of raw materials and improve yarn quality. In this study, an efficient monarch butterfly optimization-based neural network simulator called MBONN was developed to predict the tenacity of siro-spun yarns from some process parameters and fiber properties. To this end, an experimental dataset was obtained with fiber fineness, yarn twist factor, yarn linear density and strand spacing as the input variables and yarn tenacity as the output parameter. In the proposed MBONN, a monarch butterfly optimization algorithm is applied as a global search method to evolve weights of a multilayer perception (MLP) neural network. The prediction accuracy of the MBONN was compared with that of a MLP neural network trained with back propagation algorithm, MLP neural network trained with genetic algorithms and linear regression model. The results indicated that the prediction accuracy of the proposed MBONN is statistically superior to that of other models. The effect of fiber fineness, yarn linear density, twist factor and strand spacing on yarn tenacity was investigated using the proposed MBONN. Additionally, the observed trends in variation of yarn tenacity with fiber and process parameters were discussed with reference to the yarn internal structure. It was established that higher migration parameters result in increasing the siro-spun yarn tenacity. It was found that the yarns with higher migration parameters benefit from a more coherent self-locking structure which severely restricts fiber slippage, thereby increasing the yarn tenacity.
Cooperation Search Algorithm: A Novel Metaheuristic Evolutionary Intelligence Algorithm For Numerical Optimization And Engineering Optimization Problems This paper develops a novel population-based evolutionary method called cooperation search algorithm (CSA) to address the complex global optimization problem. Inspired by the team cooperation behaviors in modern enterprise, the CSA method randomly generates a set of candidate solutions in the problem space, and then three operators are repeatedly executed until the stopping criterion is met: the team communication operator is used to improve the global exploration and determine the promising search area; the reflective learning operator is used to achieve a comprise between exploration and exploitation; the internal competition operator is used to choose solutions with better performances for the next cycle. Firstly, three kinds of mathematical optimization problems (including 24 famous test functions, 25 CEC2005 test problems and 30 CEC2014 test problems) are used to test the convergence speed and search accuracy of the CSA method. Then, several famous engineering optimization problems (like Gear train design, Welded beam design and Speed reducer design) are chosen to testify the engineering practicality of the CSA method. The results in different scenarios demonstrate that as compared with several existing evolutionary algorithms, the CSA method can effectively explore the decision space and produce competitive results in terms of various performance evaluation indicators. Thus, an effective tool is provided for solving the complex global optimization problems. (C) 2020 Elsevier B.V. All rights reserved.
On the security of public key protocols Recently the use of public key encryption to provide secure network communication has received considerable attention. Such public key systems are usually effective against passive eavesdroppers, who merely tap the lines and try to decipher the message. It has been pointed out, however, that an improperly designed protocol could be vulnerable to an active saboteur, one who may impersonate another user or alter the message being transmitted. Several models are formulated in which the security of protocols can be discussed precisely. Algorithms and characterizations that can be used to determine protocol security in these models are given.
QoE-Driven Edge Caching in Vehicle Networks Based on Deep Reinforcement Learning The Internet of vehicles (IoV) is a large information interaction network that collects information on vehicles, roads and pedestrians. One of the important uses of vehicle networks is to meet the entertainment needs of driving users through communication between vehicles and roadside units (RSUs). Due to the limited storage space of RSUs, determining the content cached in each RSU is a key challenge. With the development of 5G and video editing technology, short video systems have become increasingly popular. Current widely used cache update methods, such as partial file precaching and content popularity- and user interest-based determination, are inefficient for such systems. To solve this problem, this paper proposes a QoE-driven edge caching method for the IoV based on deep reinforcement learning. First, a class-based user interest model is established. Compared with the traditional file popularity- and user interest distribution-based cache update methods, the proposed method is more suitable for systems with a large number of small files. Second, a quality of experience (QoE)-driven RSU cache model is established based on the proposed class-based user interest model. Third, a deep reinforcement learning method is designed to address the QoE-driven RSU cache update issue effectively. The experimental results verify the effectiveness of the proposed algorithm.
Image information and visual quality Measurement of visual quality is of fundamental importance to numerous image and video processing applications. The goal of quality assessment (QA) research is to design algorithms that can automatically assess the quality of images or videos in a perceptually consistent manner. Image QA algorithms generally interpret image quality as fidelity or similarity with a "reference" or "perfect" image in some perceptual space. Such "full-reference" QA methods attempt to achieve consistency in quality prediction by modeling salient physiological and psychovisual features of the human visual system (HVS), or by signal fidelity measures. In this paper, we approach the image QA problem as an information fidelity problem. Specifically, we propose to quantify the loss of image information to the distortion process and explore the relationship between image information and visual quality. QA systems are invariably involved with judging the visual quality of "natural" images and videos that are meant for "human consumption." Researchers have developed sophisticated models to capture the statistics of such natural signals. Using these models, we previously presented an information fidelity criterion for image QA that related image quality with the amount of information shared between a reference and a distorted image. In this paper, we propose an image information measure that quantifies the information that is present in the reference image and how much of this reference information can be extracted from the distorted image. Combining these two quantities, we propose a visual information fidelity measure for image QA. We validate the performance of our algorithm with an extensive subjective study involving 779 images and show that our method outperforms recent state-of-the-art image QA algorithms by a sizeable margin in our simulations. The code and the data from the subjective study are available at the LIVE website.
Stabilization of switched continuous-time systems with all modes unstable via dwell time switching Stabilization of switched systems composed fully of unstable subsystems is one of the most challenging problems in the field of switched systems. In this brief paper, a sufficient condition ensuring the asymptotic stability of switched continuous-time systems with all modes unstable is proposed. The main idea is to exploit the stabilization property of switching behaviors to compensate the state divergence made by unstable modes. Then, by using a discretized Lyapunov function approach, a computable sufficient condition for switched linear systems is proposed in the framework of dwell time; it is shown that the time intervals between two successive switching instants are required to be confined by a pair of upper and lower bounds to guarantee the asymptotic stability. Based on derived results, an algorithm is proposed to compute the stability region of admissible dwell time. A numerical example is proposed to illustrate our approach.
Software-Defined Networking: A Comprehensive Survey The Internet has led to the creation of a digital society, where (almost) everything is connected and is accessible from anywhere. However, despite their widespread adoption, traditional IP networks are complex and very hard to manage. It is both difficult to configure the network according to predefined policies, and to reconfigure it to respond to faults, load, and changes. To make matters even more difficult, current networks are also vertically integrated: the control and data planes are bundled together. Software-defined networking (SDN) is an emerging paradigm that promises to change this state of affairs, by breaking vertical integration, separating the network's control logic from the underlying routers and switches, promoting (logical) centralization of network control, and introducing the ability to program the network. The separation of concerns, introduced between the definition of network policies, their implementation in switching hardware, and the forwarding of traffic, is key to the desired flexibility: by breaking the network control problem into tractable pieces, SDN makes it easier to create and introduce new abstractions in networking, simplifying network management and facilitating network evolution. In this paper, we present a comprehensive survey on SDN. We start by introducing the motivation for SDN, explain its main concepts and how it differs from traditional networking, its roots, and the standardization activities regarding this novel paradigm. Next, we present the key building blocks of an SDN infrastructure using a bottom-up, layered approach. We provide an in-depth analysis of the hardware infrastructure, southbound and northbound application programming interfaces (APIs), network virtualization layers, network operating systems (SDN controllers), network programming languages, and network applications. We also look at cross-layer problems such as debugging and troubleshooting. In an effort to anticipate the future evolution of this - ew paradigm, we discuss the main ongoing research efforts and challenges of SDN. In particular, we address the design of switches and control platforms - with a focus on aspects such as resiliency, scalability, performance, security, and dependability - as well as new opportunities for carrier transport networks and cloud providers. Last but not least, we analyze the position of SDN as a key enabler of a software-defined environment.
An ID-Based Linearly Homomorphic Signature Scheme and Its Application in Blockchain. Identity-based cryptosystems mean that public keys can be directly derived from user identifiers, such as telephone numbers, email addresses, and social insurance number, and so on. So they can simplify key management procedures of certificate-based public key infrastructures and can be used to realize authentication in blockchain. Linearly homomorphic signature schemes allow to perform linear computations on authenticated data. And the correctness of the computation can be publicly verified. Although a series of homomorphic signature schemes have been designed recently, there are few homomorphic signature schemes designed in identity-based cryptography. In this paper, we construct a new ID-based linear homomorphic signature scheme, which avoids the shortcomings of the use of public-key certificates. The scheme is proved secure against existential forgery on adaptively chosen message and ID attack under the random oracle model. The ID-based linearly homomorphic signature schemes can be applied in e-business and cloud computing. Finally, we show how to apply it to realize authentication in blockchain.
Robot tutor and pupils’ educational ability: Teaching the times tables Research shows promising results of educational robots in language and STEM tasks. In language, more research is available, occasionally in view of individual differences in pupils’ educational ability levels, and learning seems to improve with more expressive robot behaviors. In STEM, variations in robots’ behaviors have been examined with inconclusive results and never while systematically investigating how differences in educational abilities match with different robot behaviors. We applied an autonomously tutoring robot (without tablet, partly WOz) in a 2 × 2 experiment of social vs. neutral behavior in above-average vs. below-average schoolchildren (N = 86; age 8–10 years) while rehearsing the multiplication tables on a one-to-one basis. The standard school test showed that on average, pupils significantly improved their performance even after 3 occasions of 5-min exercises. Beyond-average pupils profited most from a robot tutor, whereas those below average in multiplication benefited more from a robot that showed neutral rather than more social behavior.
1.2
0.2
0.2
0.2
0.2
0.2
0.2
0
0
0
0
0
0
0
Machine-Learning-Enabled DDoS Attacks Detection in P4 Programmable Networks Distributed Denial of Service (DDoS) attacks represent a major concern in modern Software Defined Networking (SDN), as SDN controllers are sensitive points of failures in the whole SDN architecture. Recently, research on DDoS attacks detection in SDN has focused on investigation of how to leverage data plane programmability, enabled by P4 language, to detect attacks directly in network switches, with marginal involvement of SDN controllers. In order to effectively address cybersecurity management in SDN architectures, we investigate the potential of Artificial Intelligence and Machine Learning (ML) algorithms to perform automated DDoS Attacks Detection (DAD), specifically focusing on Transmission Control Protocol SYN flood attacks. We compare two different DAD architectures, called Standalone and Correlated DAD, where traffic features collection and attack detection are performed locally at network switches or in a single entity (e.g., in SDN controller), respectively. We combine the capability of ML and P4-enabled data planes to implement real-time DAD. Illustrative numerical results show that, for all tested ML algorithms, accuracy, precision, recall and F1-score are above 98% in most cases, and classification time is in the order of few hundreds of mu s in the worst case. Considering real-time DAD implementation, significant latency reduction is obtained when features are extracted at the data plane by using P4 language.
Adaptive Clustering with Feature Ranking for DDoS Attacks Detection Distributed Denial of Service (DDoS) attacks pose an increasing threat to the current internet. The detection of such attacks plays an important role in maintaining the security of networks. In this paper, we propose a novel adaptive clustering method combined with feature ranking for DDoS attacks detection. First, based on the analysis of network traffic, preliminary variables are selected. Second, the Modified Global K-means algorithm (MGKM) is used as the basic incremental clustering algorithm to identify the cluster structure of the target data. Third, the linear correlation coefficient is used for feature ranking. Lastly, the feature ranking result is used to inform and recalculate the clusters. This adaptive process can make worthwhile adjustments to the working feature vector according to different patterns of DDoS attacks, and can improve the quality of the clusters and the effectiveness of the clustering algorithm. The experimental results demonstrate that our method is effective and adaptive in detecting the separate phases of DDoS attacks.
The role of KL divergence in anomaly detection We study the role of Kullback-Leibler divergence in the framework of anomaly detection, where its abilities as a statistic underlying detection have never been investigated in depth. We give an in-principle analysis of network attack detection, showing explicitly attacks may be masked at minimal cost through 'camouflage'. We illustrate on both synthetic distributions and ones taken from real traffic.
An Anomaly Detection Model Based on One-Class SVM to Detect Network Intrusions. Intrusion detection occupies a decision position in solving the network security problems. Support Vector Machines (SVMs) are one of the widely used intrusion detection techniques. However, the commonly used two-class SVM algorithms are facing difficulties of constructing the training dataset. That is because in many real application scenarios, normal connection records are easy to be obtained, but attack records are not so. We propose an anomaly detection model based on One-class SVM to detect network intrusions. The one-class SVM adopts only normal network connection records as the training dataset. But after being trained, it is able to recognize normal from various attacks. This just meets the requirements of the anomaly detection. Experimental results on KDDCUP99 dataset show that compared to Probabilistic Neural Network (PNN) and C-SVM, our anomaly detection model based on One-class SVM achieves higher detection rates and yields average better performance in terms of precision, recall and F-value.
Investigating Real-Time Entropy Features of DDoS Attack Based on Categorized Partial-Flows With the advent of IoT devices and exponential growth of nodes on the internet, computer networks are facing new challenges, with one of the more important ones being DDoS attacks. In this paper, new features to detect initiation and termination of DDoS attacks are investigated. The method to extract these features is devised with respect to some openflowbased switch capabilities. These features provide us with a higher resolution to view and process packet count entropies, thus improving DDoS attack detection capabilities. Although some of the technical assumptions are based on SDN technology and openflow protocol, the methodology can be applied in other networking paradigms as well.
Augmented In-Band Telemetry to the User Equipment for Beyond 5G Converged Packet-Optical Networks Traffic monitoring through in-band telemetry is extended up to the User Equipment (UE), providing accurate e2e latency measurement. The UE becomes aware of its experienced service performance, enabling autonomous operations for faster automatic source-based Edge-Cloud steering.
Image quality assessment: from error visibility to structural similarity. Objective methods for assessing perceptual image quality traditionally attempted to quantify the visibility of errors (differences) between a distorted image and a reference image using a variety of known properties of the human visual system. Under the assumption that human visual perception is highly adapted for extracting structural information from a scene, we introduce an alternative complementary framework for quality assessment based on the degradation of structural information. As a specific example of this concept, we develop a Structural Similarity Index and demonstrate its promise through a set of intuitive examples, as well as comparison to both subjective ratings and state-of-the-art objective methods on a database of images compressed with JPEG and JPEG2000.
A survey of socially interactive robots This paper reviews “socially interactive robots”: robots for which social human–robot interaction is important. We begin by discussing the context for socially interactive robots, emphasizing the relationship to other research fields and the different forms of “social robots”. We then present a taxonomy of design methods and system components used to build socially interactive robots. Finally, we describe the impact of these robots on humans and discuss open issues. An expanded version of this paper, which contains a survey and taxonomy of current applications, is available as a technical report [T. Fong, I. Nourbakhsh, K. Dautenhahn, A survey of socially interactive robots: concepts, design and applications, Technical Report No. CMU-RI-TR-02-29, Robotics Institute, Carnegie Mellon University, 2002].
IoT-U: Cellular Internet-of-Things Networks Over Unlicensed Spectrum. In this paper, we consider an uplink cellular Internet-of-Things (IoT) network, where a cellular user (CU) can serve as the mobile data aggregator for a cluster of IoT devices. To be specific, the IoT devices can either transmit the sensory data to the base station (BS) directly by cellular communications, or first aggregate the data to a CU through machine-to-machine (M2M) communications before t...
Multi-column Deep Neural Networks for Image Classification Traditional methods of computer vision and machine learning cannot match human performance on tasks such as the recognition of handwritten digits or traffic signs. Our biologically plausible deep artificial neural network architectures can. Small (often minimal) receptive fields of convolutional winner-take-all neurons yield large network depth, resulting in roughly as many sparsely connected neural layers as found in mammals between retina and visual cortex. Only winner neurons are trained. Several deep neural columns become experts on inputs preprocessed in different ways; their predictions are averaged. Graphics cards allow for fast training. On the very competitive MNIST handwriting benchmark, our method is the first to achieve near-human performance. On a traffic sign recognition benchmark it outperforms humans by a factor of two. We also improve the state-of-the-art on a plethora of common image classification benchmarks.
A novel full structure optimization algorithm for radial basis probabilistic neural networks. In this paper, a novel full structure optimization algorithm for radial basis probabilistic neural networks (RBPNN) is proposed. Firstly, a minimum volume covering hyperspheres (MVCH) algorithm is proposed to heuristically select the initial hidden layer centers of the RBPNN, and then the recursive orthogonal least square (ROLS) algorithm combined with the particle swarm optimization (PSO) algorithm is adopted to further optimize the initial structure of the RBPNN. Finally, the effectiveness and efficiency of our proposed algorithm are evaluated through a plant species identification task involving 50 plant species.
An Automatic Screening Approach for Obstructive Sleep Apnea Diagnosis Based on Single-Lead Electrocardiogram Traditional approaches for obstructive sleep apnea (OSA) diagnosis are apt to using multiple channels of physiological signals to detect apnea events by dividing the signals into equal-length segments, which may lead to incorrect apnea event detection and weaken the performance of OSA diagnosis. This paper proposes an automatic-segmentation-based screening approach with the single channel of Electrocardiogram (ECG) signal for OSA subject diagnosis, and the main work of the proposed approach lies in three aspects: (i) an automatic signal segmentation algorithm is adopted for signal segmentation instead of the equal-length segmentation rule; (ii) a local median filter is improved for reduction of the unexpected RR intervals before signal segmentation; (iii) the designed OSA severity index and additional admission information of OSA suspects are plugged into support vector machine (SVM) for OSA subject diagnosis. A real clinical example from PhysioNet database is provided to validate the proposed approach and an average accuracy of 97.41% for subject diagnosis is obtained which demonstrates the effectiveness for OSA diagnosis.
Adaptive fuzzy tracking control for switched uncertain strict-feedback nonlinear systems. •Adaptive tracking control for switched strict-feedback nonlinear systems is proposed.•The generalized fuzzy hyperbolic model is used to approximate nonlinear functions.•The designed controller has fewer design parameters comparing with existing methods.
Sustainable and Efficient Data Collection from WSNs to Cloud. The development of cloud computing pours great vitality into traditional wireless sensor networks (WSNs). The integration of WSNs and cloud computing has received a lot of attention from both academia and industry. However, collecting data from WSNs to cloud is not sustainable. Due to the weak communication ability of WSNs, uploading big sensed data to the cloud within the limited time becomes a b...
1.1
0.1
0.1
0.1
0.1
0.1
0
0
0
0
0
0
0
0
HSACMA: a hierarchical scalable adaptive cloud monitoring architecture. Monitoring for cloud is the key technology to know the status and the availability of the resources and services present in the current infrastructure. However, cloud monitoring faces a lot of challenges due to inefficient monitoring capability and enormous resource consumption. We study the adaptive monitoring for cloud computing platform, and focus on the problem of balancing monitoring capability and resource consumption. We proposed HSACMA, a hierarchical scalable adaptive monitoring architecture, that (1) monitors the physical and virtual infrastructure at the infrastructure layer, the middleware running at the platform layer, and the application services at the application layer; (2) achieves the scalability of the monitoring based on microservices; and (3) adaptively adjusts the monitoring interval and data transmission strategy according to the running state of the cloud computing system. Moreover, we study a case of real production system deployed and running on the cloud computing platform called CloudStack, to verify the effectiveness of applying our architecture in practice. The results show that HSACMA can guarantee the accuracy and real-time performance of monitoring and reduces resource consumption.
Infrastructure as a Service and Cloud Technologies To choose the most appropriate cloud-computing model for your organization, you must analyze your IT infrastructure, usage, and needs. To help with this, this article describes cloud computing's current status.
Dynamic Management of Virtual Infrastructures Cloud infrastructures are becoming an appropriate solution to address the computational needs of scientific applications. However, the use of public or on-premises Infrastructure as a Service (IaaS) clouds requires users to have non-trivial system administration skills. Resource provisioning systems provide facilities to choose the most suitable Virtual Machine Images (VMI) and basic configuration of multiple instances and subnetworks. Other tasks such as the configuration of cluster services, computational frameworks or specific applications are not trivial on the cloud, and normally users have to manually select the VMI that best fits, including undesired additional services and software packages. This paper presents a set of components that ease the access and the usability of IaaS clouds by automating the VMI selection, deployment, configuration, software installation, monitoring and update of Virtual Appliances. It supports APIs from a large number of virtual platforms, making user applications cloud-agnostic. In addition it integrates a contextualization system to enable the installation and configuration of all the user required applications providing the user with a fully functional infrastructure. Therefore, golden VMIs and configuration recipes can be easily reused across different deployments. Moreover, the contextualization agent included in the framework supports horizontal (increase/decrease the number of resources) and vertical (increase/decrease resources within a running Virtual Machine) by properly reconfiguring the software installed, considering the configuration of the multiple resources running. This paves the way for automatic virtual infrastructure deployment, customization and elastic modification at runtime for IaaS clouds.
Virtual machine placement quality estimation in cloud infrastructures using integer linear programming This paper is devoted to the quality estimation of virtual machine (VM) placement in cloud infrastructures, i.e., to choose the best hosts for a given set of VMs. We focus on test generation and monitoring techniques for comparing the placement result of a given implementation with an optimal solution with respect to given criteria. We show how Integer Linear Programming problems can be formulated and utilized for deriving test suites and optimal solutions to provide verdicts concerning the quality of VM placement implementations; the quality is calculated as the distance from an optimal placement for a given criterion (or a set of criteria). The presented approach is generic and showcased on resource utilization, energy consumption, and resource over-commitment cost. Experiments performed with different VM placement algorithms (including the VM placement algorithms implemented in widely used platforms, such as OpenStack) exhibit the competence of such algorithms with respect to different criteria.
Integrating security and privacy in software development As a consequence to factors such as progress made by the attackers, release of new technologies and use of increasingly complex systems, and threats to applications security have been continuously evolving. Security of code and privacy of data must be implemented in both design and programming practice to face such scenarios. In such a context, this paper proposes a software development approach, Privacy Oriented Software Development (POSD), that complements traditional development processes by integrating the activities needed for addressing security and privacy management in software systems. The approach is based on 5 key elements (Privacy by Design, Privacy Design Strategies, Privacy Pattern, Vulnerabilities, Context). The approach can be applied in two directions forward and backward, for developing new software systems or re-engineering an existing one. This paper presents the POSD approach in the backward mode together with an application in the context of an industrial project. Results show that POSD is able to discover software vulnerabilities, identify the remediation patterns needed for addressing them in the source code, and design the target architecture to be used for guiding privacy-oriented system re-engineering.
Image quality assessment: from error visibility to structural similarity. Objective methods for assessing perceptual image quality traditionally attempted to quantify the visibility of errors (differences) between a distorted image and a reference image using a variety of known properties of the human visual system. Under the assumption that human visual perception is highly adapted for extracting structural information from a scene, we introduce an alternative complementary framework for quality assessment based on the degradation of structural information. As a specific example of this concept, we develop a Structural Similarity Index and demonstrate its promise through a set of intuitive examples, as well as comparison to both subjective ratings and state-of-the-art objective methods on a database of images compressed with JPEG and JPEG2000.
Vision meets robotics: The KITTI dataset We present a novel dataset captured from a VW station wagon for use in mobile robotics and autonomous driving research. In total, we recorded 6 hours of traffic scenarios at 10-100 Hz using a variety of sensor modalities such as high-resolution color and grayscale stereo cameras, a Velodyne 3D laser scanner and a high-precision GPS/IMU inertial navigation system. The scenarios are diverse, capturing real-world traffic situations, and range from freeways over rural areas to inner-city scenes with many static and dynamic objects. Our data is calibrated, synchronized and timestamped, and we provide the rectified and raw image sequences. Our dataset also contains object labels in the form of 3D tracklets, and we provide online benchmarks for stereo, optical flow, object detection and other tasks. This paper describes our recording platform, the data format and the utilities that we provide.
A tutorial on support vector regression In this tutorial we give an overview of the basic ideas underlying Support Vector (SV) machines for function estimation. Furthermore, we include a summary of currently used algorithms for training SV machines, covering both the quadratic (or convex) programming part and advanced methods for dealing with large datasets. Finally, we mention some modifications and extensions that have been applied to the standard SV algorithm, and discuss the aspect of regularization from a SV perspective.
GameFlow: a model for evaluating player enjoyment in games Although player enjoyment is central to computer games, there is currently no accepted model of player enjoyment in games. There are many heuristics in the literature, based on elements such as the game interface, mechanics, gameplay, and narrative. However, there is a need to integrate these heuristics into a validated model that can be used to design, evaluate, and understand enjoyment in games. We have drawn together the various heuristics into a concise model of enjoyment in games that is structured by flow. Flow, a widely accepted model of enjoyment, includes eight elements that, we found, encompass the various heuristics from the literature. Our new model, GameFlow, consists of eight elements -- concentration, challenge, skills, control, clear goals, feedback, immersion, and social interaction. Each element includes a set of criteria for achieving enjoyment in games. An initial investigation and validation of the GameFlow model was carried out by conducting expert reviews of two real-time strategy games, one high-rating and one low-rating, using the GameFlow criteria. The result was a deeper understanding of enjoyment in real-time strategy games and the identification of the strengths and weaknesses of the GameFlow model as an evaluation tool. The GameFlow criteria were able to successfully distinguish between the high-rated and low-rated games and identify why one succeeded and the other failed. We concluded that the GameFlow model can be used in its current form to review games; further work will provide tools for designing and evaluating enjoyment in games.
Adapting visual category models to new domains Domain adaptation is an important emerging topic in computer vision. In this paper, we present one of the first studies of domain shift in the context of object recognition. We introduce a method that adapts object models acquired in a particular visual domain to new imaging conditions by learning a transformation that minimizes the effect of domain-induced changes in the feature distribution. The transformation is learned in a supervised manner and can be applied to categories for which there are no labeled examples in the new domain. While we focus our evaluation on object recognition tasks, the transform-based adaptation technique we develop is general and could be applied to nonimage data. Another contribution is a new multi-domain object database, freely available for download. We experimentally demonstrate the ability of our method to improve recognition on categories with few or no target domain labels and moderate to large changes in the imaging conditions.
A Web-Based Tool For Control Engineering Teaching In this article a new tool for control engineering teaching is presented. The tool was implemented using Java applets and is freely accessible through Web. It allows the analysis and simulation of linear control systems and was created to complement the theoretical lectures in basic control engineering courses. The article is not only centered in the description of the tool but also in the methodology to use it and its evaluation in an electrical engineering degree. Two practical problems are included in the manuscript to illustrate the use of the main functions implemented. The developed web-based tool can be accessed through the link http://www.controlweb.cyc.ull.es. (C) 2006 Wiley Periodicals, Inc.
Adaptive Consensus Control for a Class of Nonlinear Multiagent Time-Delay Systems Using Neural Networks Because of the complicity of consensus control of nonlinear multiagent systems in state time-delay, most of previous works focused only on linear systems with input time-delay. An adaptive neural network (NN) consensus control method for a class of nonlinear multiagent systems with state time-delay is proposed in this paper. The approximation property of radial basis function neural networks (RBFNNs) is used to neutralize the uncertain nonlinear dynamics in agents. An appropriate Lyapunov-Krasovskii functional, which is obtained from the derivative of an appropriate Lyapunov function, is used to compensate the uncertainties of unknown time delays. It is proved that our proposed approach guarantees the convergence on the basis of Lyapunov stability theory. The simulation results of a nonlinear multiagent time-delay system and a multiple collaborative manipulators system show the effectiveness of the proposed consensus control algorithm.
5G Virtualized Multi-access Edge Computing Platform for IoT Applications. The next generation of fifth generation (5G) network, which is implemented using Virtualized Multi-access Edge Computing (vMEC), Network Function Virtualization (NFV) and Software Defined Networking (SDN) technologies, is a flexible and resilient network that supports various Internet of Things (IoT) devices. While NFV provides flexibility by allowing network functions to be dynamically deployed and inter-connected, vMEC provides intelligence at the edge of the mobile network reduces latency and increases the available capacity. With the diverse development of networking applications, the proposed vMEC use of Container-based Virtualization Technology (CVT) as gateway with IoT devices for flow control mechanism in scheduling and analysis methods will effectively increase the application Quality of Service (QoS). In this work, the proposed IoT gateway is analyzed. The combined effect of simultaneously deploying Virtual Network Functions (VNFs) and vMEC applications on a single network infrastructure, and critically in effecting exhibits low latency, high bandwidth and agility that will be able to connect large scale of devices. The proposed platform efficiently exploiting resources from edge computing and cloud computing, and takes IoT applications that adapt to network conditions to degrade an average 30% of end to end network latency.
Energy harvesting algorithm considering max flow problem in wireless sensor networks. In Wireless Sensor Networks (WSNs), sensor nodes with poor energy always have bad effect on the data rate or max flow. These nodes are called bottleneck nodes. In this paper, in order to increase the max flow, we assume an energy harvesting WSNs environment to investigate the cooperation of multiple Mobile Chargers (MCs). MCs are mobile robots that use wireless charging technology to charge sensor nodes in WSNs. This means that in energy harvesting WSNs environments, sensor nodes can obtain energy replenishment by using MCs or collecting energy from nature by themselves. In our research, we use MCs to improve the energy of the sensor nodes by performing multiple rounds of unified scheduling, and finally achieve the purpose of increasing the max flow at sinks. Firstly, we model this problem as a Linear Programming (LP) to search the max flow in a round of charging scheduling and prove that the problem is NP-hard. In order to solve the problem, we propose a heuristic approach: deploying MCs in units of paths with the lowest energy node priority. To reduce the energy consumption of MCs and increase the charging efficiency, we also take the optimization of MCs’ moving distance into our consideration. Finally, we extend the method to multiple rounds of scheduling called BottleNeck. Simulation results show that Bottleneck performs well at increasing max flow.
1.2
0.2
0.2
0.2
0.2
0
0
0
0
0
0
0
0
0
Data-Based Reinforcement Learning for Nonzero-Sum Games With Unknown Drift Dynamics. This paper is concerned about the nonlinear optimization problem of nonzero-sum (NZS) games with unknown drift dynamics. The data-based integral reinforcement learning (IRL) method is proposed to approximate the Nash equilibrium of NZS games iteratively. Furthermore, we prove that the data-based IRL method is equivalent to the model-based policy iteration algorithm, which guarantees the convergenc...
Inexact Kleinman-Newton Method for Riccati Equations In this paper we consider the numerical solution of the algebraic Riccati equation using Newton's method. We propose an inexact variant which allows one control the number of the inner iterates used in an iterative solver for each Newton step. Conditions are given under which the monotonicity and global convergence result of Kleinman also hold for the inexact Newton iterates. Numerical results illustrate the efficiency of this method.
Adaptive Dynamic Programming for Stochastic Systems With State and Control Dependent Noise. In this technical note, the adaptive optimal control problem is investigated for a class of continuous-time stochastic systems subject to multiplicative noise. A novel non-model-based optimal control design methodology is employed to iteratively update the control policy on-line by using directly the data of the system state and input. Both adaptive dynamic programming (ADP) and robust ADP algorithms are developed, along with rigorous stability and convergence analysis. The effectiveness of the obtained methods is illustrated by an example arising from biological sensorimotor control.
Stability Analysis of Optimal Adaptive Control using Value Iteration with Approximation Errors. Effects of the presence of approximation errors are analyzed on the stability of adaptive optimal control using value iteration, initiated from a stabilizing control policy. This analysis includes the system operated using any single/constant resulting control policy and also using an evolving/time-varying control policy. Sufficient conditions on the `per iteration&#39; approximation errors are obtain...
Policy Iteration Q-Learning for Data-Based Two-Player Zero-Sum Game of Linear Discrete-Time Systems. In this article, the data-based two-player zero-sum game problem is considered for linear discrete-time systems. This problem theoretically depends on solving the discrete-time game algebraic Riccati equation (DTGARE), while it requires complete system dynamics. To avoid solving the DTGARE, the $Q$ -function is introduced and a data-based policy iteration $Q$ -learning (PIQL) algorithm is develo...
Hamiltonian-Driven Adaptive Dynamic Programming for Continuous Nonlinear Dynamical Systems. This paper presents a Hamiltonian-driven framework of adaptive dynamic programming (ADP) for continuous time nonlinear systems, which consists of evaluation of an admissible control, comparison between two different admissible policies with respect to the corresponding the performance function, and the performance improvement of an admissible control. It is showed that the Hamiltonian can serve as...
Galerkin approximations of the generalized Hamilton-Jacobi-Bellman equation In this paper we study the convergence of the Galerkin approximation method applied to the generalized Hamilton-Jacobi-Bellman (GHJB) equation over a compact set containing the origin. The GHJB equation gives the cost of an arbitrary control law and can be used to improve the performance of this control. The GHJB equation can also be used to successively approximate the Hamilton-Jacobi-Bellman equation. We state sufficient conditions that guarantee that the Galerkin approximation converges to the solution of the GHJB equation and that the resulting approximate control is stabilizing on the same region as the initial control. The method is demonstrated on a simple nonlinear system and is compared to a result obtained by using exact feedback linearization in conjunction with the LQR design method. (C) 1997 Elsevier Science Ltd. All rights reserved.
Adaptive Consensus Control for a Class of Nonlinear Multiagent Time-Delay Systems Using Neural Networks Because of the complicity of consensus control of nonlinear multiagent systems in state time-delay, most of previous works focused only on linear systems with input time-delay. An adaptive neural network (NN) consensus control method for a class of nonlinear multiagent systems with state time-delay is proposed in this paper. The approximation property of radial basis function neural networks (RBFNNs) is used to neutralize the uncertain nonlinear dynamics in agents. An appropriate Lyapunov-Krasovskii functional, which is obtained from the derivative of an appropriate Lyapunov function, is used to compensate the uncertainties of unknown time delays. It is proved that our proposed approach guarantees the convergence on the basis of Lyapunov stability theory. The simulation results of a nonlinear multiagent time-delay system and a multiple collaborative manipulators system show the effectiveness of the proposed consensus control algorithm.
Bidirectional recurrent neural networks In the first part of this paper, a regular recurrent neural network (RNN) is extended to a bidirectional recurrent neural network (BRNN). The BRNN can be trained without the limitation of using input information just up to a preset future frame. This is accomplished by training it simultaneously in positive and negative time direction. Structure and training procedure of the proposed network are explained. In regression and classification experiments on artificial data, the proposed structure gives better results than other approaches. For real data, classification experiments for phonemes from the TIMIT database show the same tendency. In the second part of this paper, it is shown how the proposed bidirectional structure can be easily modified to allow efficient estimation of the conditional posterior probability of complete symbol sequences without making any explicit assumption about the shape of the distribution. For this part, experiments on real data are reported
Social Perception and Steering for Online Avatars This paper presents work on a new platform for producing realistic group conversation dynamics in shared virtual environments. An avatar, representing users, should perceive the surrounding social environment just as humans would, and use the perceptual information for driving low level reactive behaviors. Unconscious reactions serve as evidence of life, and can also signal social availability and spatial awareness to others. These behaviors get lost when avatar locomotion requires explicit user control. For automating such behaviors we propose a steering layer in the avatars that manages a set of prioritized behaviors executed at different frequencies, which can be activated or deactivated and combined together. This approach gives us enough flexibility to model the group dynamics of social interactions as a set of social norms that activate relevant steering behaviors. A basic set of behaviors is described for conversations, some of which generate a social force field that makes the formation of conversation groups fluidly adapt to external and internal noise, through avatar repositioning and reorientations. The resulting social group behavior appears relatively robust, but perhaps more importantly, it starts to bring a new sense of relevance and continuity to the virtual bodies that often get separated from the ongoing conversation in the chat window.
Surrogate-assisted evolutionary computation: Recent advances and future challenges Surrogate-assisted, or meta-model based evolutionary computation uses efficient computational models, often known as surrogates or meta-models, for approximating the fitness function in evolutionary algorithms. Research on surrogate-assisted evolutionary computation began over a decade ago and has received considerably increasing interest in recent years. Very interestingly, surrogate-assisted evolutionary computation has found successful applications not only in solving computationally expensive single- or multi-objective optimization problems, but also in addressing dynamic optimization problems, constrained optimization problems and multi-modal optimization problems. This paper provides a concise overview of the history and recent developments in surrogate-assisted evolutionary computation and suggests a few future trends in this research area.
A novel data hiding for color images based on pixel value difference and modulus function This paper proposes a novel data hiding method using pixel-value difference and modulus function for color image with the large embedding capacity(hiding 810757 bits in a 512 512 host image at least) and a high-visual-quality of the cover image. The proposed method has fully taken into account the correlation of the R, G and B plane of a color image. The amount of information embedded the R plane and the B plane determined by the difference of the corresponding pixel value between the G plane and the median of G pixel value in each pixel block. Furthermore, two sophisticated pixel value adjustment processes are provided to maintain the division consistency and to solve underflow and overflow problems. The most importance is that the secret data are completely extracted through the mathematical theoretical proof.
An indoor localization solution using Bluetooth RSSI and multiple sensors on a smartphone. In this paper, we propose an indoor positioning system using a Bluetooth receiver, an accelerometer, a magnetic field sensor, and a barometer on a smartphone. The Bluetooth receiver is used to estimate distances from beacons. The accelerometer and magnetic field sensor are used to trace the movement of moving people in the given space. The horizontal location of the person is determined by received signal strength indications (RSSIs) and the traced movement. The barometer is used to measure the vertical position where a person is located. By combining RSSIs, the traced movement, and the vertical position, the proposed system estimates the indoor position of moving people. In experiments, the proposed approach showed excellent performance in localization with an overall error of 4.8%.
Robot tutor and pupils’ educational ability: Teaching the times tables Research shows promising results of educational robots in language and STEM tasks. In language, more research is available, occasionally in view of individual differences in pupils’ educational ability levels, and learning seems to improve with more expressive robot behaviors. In STEM, variations in robots’ behaviors have been examined with inconclusive results and never while systematically investigating how differences in educational abilities match with different robot behaviors. We applied an autonomously tutoring robot (without tablet, partly WOz) in a 2 × 2 experiment of social vs. neutral behavior in above-average vs. below-average schoolchildren (N = 86; age 8–10 years) while rehearsing the multiplication tables on a one-to-one basis. The standard school test showed that on average, pupils significantly improved their performance even after 3 occasions of 5-min exercises. Beyond-average pupils profited most from a robot tutor, whereas those below average in multiplication benefited more from a robot that showed neutral rather than more social behavior.
1.12
0.12
0.12
0.12
0.12
0.04
0.009231
0.0005
0
0
0
0
0
0
Impact of Electric Vehicles on the Expansion Planning of Distribution Systems Considering Renewable Energy, Storage, and Charging Stations. Energy storage systems (ESS) have adopted a new role with the increasing penetration of electric vehicles (EV) and renewable energy sources (RES). EV introduce new charging demands that change the traditional demand profiles and RES are characterized by their high variability. This paper presents a new multistage distribution expansion planning model where investments in distribution network asset...
Competitive on-line scheduling with level of service Motivated by an application in thinwire visualization, we study an abstract on-line scheduling problem where the size of each requested service can be scaled down by the scheduler. Thus, our problem embodies a notion of "Level of Service" that is increasingly important in multimedia applications. We give two schedulers FirstFit and EndFit based on two simple heuristics, and generalize them into a class of greedy schedulers. We show that both FirstFit and EndFit are 2-competitive, and any greedy scheduler is 3-competitive. These bounds are shown to be tight.
Coordinated Charging of Electric Vehicles for Congestion Prevention in the Distribution Grid Distributed energy resources (DERs), like electric vehicles (EVs), can offer valuable services to power systems, such as enabling renewable energy to the electricity producer and providing ancillary services to the system operator. However, these new DERs may challenge the distribution grid due to insufficient capacity in peak hours. This paper aims to coordinate the valuable services and operation constraints of three actors: the EV owner, the Fleet operator (FO) and the Distribution system operator (DSO), considering the individual EV owner's driving requirement, the charging cost of EV and thermal limits of cables and transformers in the proposed market framework. Firstly, a theoretical market framework is described. Within this framework, FOs who represent their customer's (EV owners) interests will centrally guarantee the EV owners' driving requirements and procure the energy for their vehicles with lower cost. The congestion problem will be solved by a coordination between DSO and FOs through a distribution grid capacity market scheme. Then, a mathematical formulation of the market scheme is presented. Further, some case studies are shown to illustrate the effectiveness of the proposed solutions.
Real-Time PEV Charging/Discharging Coordination in Smart Distribution Systems. This paper proposes a novel online coordination method for the charging of plug-in electric vehicles (PEVs) in smart distribution networks. The goal of the proposed method is to optimally charge the PEVs in order to maximize the PEV owners&#39; satisfaction and to minimize system operating costs without violating power system constraints. Unlike the solutions reported in the literature, the proposed c...
Optimizing the Deployment of Electric Vehicle Charging Stations Using Pervasive Mobility Data. With the recent advances in battery technology and the resulting decrease in the charging times, public charging stations are becoming a viable option for Electric Vehicle (EV) drivers. Concurrently, emergence and the wide-spread use of location-tracking devices in mobile phones and wearable devices has paved the way to track individual-level human movements to an unprecedented spatial and temporal grain. Motivated by these developments, we propose a novel methodology to perform data-driven optimization of EV charging station locations. We formulate the problem as a discrete optimization problem on a geographical grid, with the objective of covering the entire demand region while minimizing a measure of drivers’ total excess driving distance to reach charging stations, the related energy overhead, and the number of charging stations. Since optimally solving the problem is computationally infeasible, we present computationally efficient solutions based on the genetic algorithm. We then apply the proposed methodology to optimize EV charging stations layout in the city of Boston, starting from Call Detail Records (CDR) of one million users over the span of 4 months. The results show that the genetic algorithm provides solutions that significantly reduce drivers’ excess driving distance to charging stations, energy overhead, and the number of charging stations required compared to both a locally-optimized feasible solution and the current charging station deployment in the Boston metro area. We further investigate the robustness of the proposed methodology and show that building upon well-known regularity of aggregate human mobility patterns, the layout computed for demands based on the single day movements preserves its advantage also in later days and months. When collectively considered, the results presented in this paper indicate the potential of data-driven approaches for optimally placing public charging facilities at urban scale.
Design and Planning of a Multiple-Charger Multiple-Port Charging System for PEV Charging Station. Investment of charging facilities is facing deficit problems in many countries at the initial development stage of plug-in electric vehicles (PEVs). In this paper, we study the charging facility planning problem faced by a PEV charging station investor who aims to serve PEV customers with random behaviors and demands (but follow a series of predicted distributions) with lower economic costs of bot...
Online Coordinated Charging Decision Algorithm for Electric Vehicles Without Future Information The large-scale integration of plug-in electric vehicles (PEVs) to the power grid spurs the need for efficient charging coordination mechanisms. It can be shown that the optimal charging schedule smooths out the energy consumption over time so as to minimize the total energy cost. In practice, however, it is hard to smooth out the energy consumption perfectly, because the future PEV charging demand is unknown at the moment when the charging rate of an existing PEV needs to be determined. In this paper, we propose an online coordinated charging decision (ORCHARD) algorithm, which minimizes the energy cost without knowing the future information. Through rigorous proof, we show that ORCHARD is strictly feasible in the sense that it guarantees to fulfill all charging demands before due time. Meanwhile, it achieves the best known competitive ratio of 2.39. By exploiting the problem structure, we propose a novel reduced-complexity algorithm to replace the standard convex optimization techniques used in ORCHARD. Through extensive simulations, we show that the average performance gap between ORCHARD and the offline optimal solution, which utilizes the complete future information, is as small as 6.5%. By setting a proper speeding factor, the average performance gap can be further reduced to 5%.
A standalone RFID Indoor Positioning System Using Passive Tags Indoor positioning systems (IPSs) locate objects in closed structures such as office buildings, hospitals, stores, factories, and warehouses, where Global Positioning System devices generally do not work. Most available systems apply wireless concepts, optical tracking, and/or ultrasound. This paper presents a standalone IPS using radio frequency identification (RFID) technology. The concept is ba...
TripRes: Traffic Flow Prediction Driven Resource Reservation for Multimedia IoV with Edge Computing AbstractThe Internet of Vehicles (IoV) connects vehicles, roadside units (RSUs) and other intelligent objects, enabling data sharing among them, thereby improving the efficiency of urban traffic and safety. Currently, collections of multimedia content, generated by multimedia surveillance equipment, vehicles, and so on, are transmitted to edge servers for implementation, because edge computing is a formidable paradigm for accommodating multimedia services with low-latency resource provisioning. However, the uneven or discrete distribution of the traffic flow covered by edge servers negatively affects the service performance (e.g., overload and underload) of edge servers in multimedia IoV systems. Therefore, how to accurately schedule and dynamically reserve proper numbers of resources for multimedia services in edge servers is still challenging. To address this challenge, a traffic flow prediction driven resource reservation method, called TripRes, is developed in this article. Specifically, the city map is divided into different regions, and the edge servers in a region are treated as a “big edge server” to simplify the complex distribution of edge servers. Then, future traffic flows are predicted using the deep spatiotemporal residual network (ST-ResNet), and future traffic flows are used to estimate the amount of multimedia services each region needs to offload to the edge servers. With the number of services to be offloaded in each region, their offloading destinations are determined through latency-sensitive transmission path selection. Finally, the performance of TripRes is evaluated using real-world big data with over 100M multimedia surveillance records from RSUs in Nanjing China.
A new optimization method: big bang-big crunch Nature is the principal source for proposing new optimization methods such as genetic algorithms (GA) and simulated annealing (SA) methods. All traditional evolutionary algorithms are heuristic population-based search procedures that incorporate random variation and selection. The main contribution of this study is that it proposes a novel optimization method that relies on one of the theories of the evolution of the universe; namely, the Big Bang and Big Crunch Theory. In the Big Bang phase, energy dissipation produces disorder and randomness is the main feature of this phase; whereas, in the Big Crunch phase, randomly distributed particles are drawn into an order. Inspired by this theory, an optimization algorithm is constructed, which will be called the Big Bang-Big Crunch (BB-BC) method that generates random points in the Big Bang phase and shrinks those points to a single representative point via a center of mass or minimal cost approach in the Big Crunch phase. It is shown that the performance of the new (BB-BC) method demonstrates superiority over an improved and enhanced genetic search algorithm also developed by the authors of this study, and outperforms the classical genetic algorithm (GA) for many benchmark test functions.
Using noise inconsistencies for blind image forensics A commonly used tool to conceal the traces of tampering is the addition of locally random noise to the altered image regions. The noise degradation is the main cause of failure of many active or passive image forgery detection methods. Typically, the amount of noise is uniform across the entire authentic image. Adding locally random noise may cause inconsistencies in the image's noise. Therefore, the detection of various noise levels in an image may signify tampering. In this paper, we propose a novel method capable of dividing an investigated image into various partitions with homogenous noise levels. In other words, we introduce a segmentation method detecting changes in noise level. We assume the additive white Gaussian noise. Several examples are shown to demonstrate the proposed method's output. An extensive quantitative measure of the efficiency of the noise estimation part as a function of different noise standard deviations, region sizes and various JPEG compression qualities is proposed as well.
Dynamic Management of Virtual Infrastructures Cloud infrastructures are becoming an appropriate solution to address the computational needs of scientific applications. However, the use of public or on-premises Infrastructure as a Service (IaaS) clouds requires users to have non-trivial system administration skills. Resource provisioning systems provide facilities to choose the most suitable Virtual Machine Images (VMI) and basic configuration of multiple instances and subnetworks. Other tasks such as the configuration of cluster services, computational frameworks or specific applications are not trivial on the cloud, and normally users have to manually select the VMI that best fits, including undesired additional services and software packages. This paper presents a set of components that ease the access and the usability of IaaS clouds by automating the VMI selection, deployment, configuration, software installation, monitoring and update of Virtual Appliances. It supports APIs from a large number of virtual platforms, making user applications cloud-agnostic. In addition it integrates a contextualization system to enable the installation and configuration of all the user required applications providing the user with a fully functional infrastructure. Therefore, golden VMIs and configuration recipes can be easily reused across different deployments. Moreover, the contextualization agent included in the framework supports horizontal (increase/decrease the number of resources) and vertical (increase/decrease resources within a running Virtual Machine) by properly reconfiguring the software installed, considering the configuration of the multiple resources running. This paves the way for automatic virtual infrastructure deployment, customization and elastic modification at runtime for IaaS clouds.
Adaptive fuzzy tracking control for switched uncertain strict-feedback nonlinear systems. •Adaptive tracking control for switched strict-feedback nonlinear systems is proposed.•The generalized fuzzy hyperbolic model is used to approximate nonlinear functions.•The designed controller has fewer design parameters comparing with existing methods.
Social Robots for (Second) Language Learning in (Migrant) Primary School Children Especially these days, innovation and support from technology to relieve pressure in education is highly urgent. This study tested the potential advantage of a social robot over a tablet in (second) language learning on performance, engagement, and enjoyment. Shortages in primary education call for new technology solutions. Previous studies combined robots with tablets, to compensate for robot’s limitations, however, this study applied direct human–robot interaction. Primary school children (N = 63, aged 4–6) participated in a 3-wave field experiment with story-telling exercises, either with a semi-autonomous robot (without tablet, using WOz) or a tablet. Results showed increased learning gains over time when training with a social robot, compared to the tablet. Children who trained with a robot were more engaged in the story-telling task and enjoyed it more. Robot’s behavioral style (social or neutral) hardly differed overall, however, seems to vary for high versus low educational abilities. While social robots need sophistication before being implemented in schools, our study shows the potential of social robots as tutors in (second) language learning.
1.11
0.12
0.12
0.12
0.1
0.1
0.04
0
0
0
0
0
0
0
Evolutionary Methods for Generating Synthetic MasterPrint Templates: Dictionary Attack in Fingerprint Recognition Recent research has demonstrated the possibility of generating "Masterprints" that can be used by an adversary to launch a dictionary attack against a fingerprint recognition system. Masterprints are fingerprint images that fortuitously match with a large number of other fingerprints thereby compromising the security of a fingerprint-based biometric system, especially those equipped with small-sized fingerprint sensors. This work presents new methods for creating a synthetic MasterPrint dictionary that sequentially maximizes the probability of matching a large number of target fingerprints. Three techniques, namely Covariance Matrix Adaptation Evolution Strategy (CMA-ES), Differential Evolution (DE) and Particle Swarm Optimization (PSO), are explored. Experiments carried out using a commercial fingerprint verification software, and public datasets, show that the proposed approaches performed quite well compared to the previously known MasterPrint generation methods.
MasterPrint: Exploring the Vulnerability of Partial Fingerprint-Based Authentication Systems. This paper investigates the security of partial fingerprint-based authentication systems, especially when multiple fingerprints of a user are enrolled. A number of consumer electronic devices, such as smartphones, are beginning to incorporate fingerprint sensors for user authentication. The sensors embedded in these devices are generally small and the resulting images are, therefore, limited in si...
Attacks on biometric systems: a case study in fingerprints In spite of numerous advantages of biometrics-based personal authentication systems over traditional security systems based on token or knowledge, they are vulnerable to attacks that can decrease their security considerably. In this paper, we analyze these attacks in the realm of a fingerprint biometric system. We propose an attack system that uses a hill climbing procedure to synthesize the target minutia templates and evaluate its feasibility with extensive experimental results conducted on a large fingerprint database. Several measures that can be utilized to decrease the probability of such attacks and their ramifications are also presented.
Generating Master Faces for Use in Performing Wolf Attacks on Face Recognition Systems Due to its convenience, biometric authentication, especial face authentication, has become increasingly mainstream and thus is now a prime target for attackers. Presentation attacks and face morphing are typical types of attack. Previous research has shown that finger- vein- and fingerprint-based authentication methods are susceptible to wolf attacks, in which a wolf sample matches many enrolled user templates. In this work, we demonstrated that wolf (generic) faces, which we call “master faces,” can also compromise face recognition systems and that the master face concept can be generalized in some cases. Motivated by recent similar work in the fingerprint domain, we generated high-quality master faces by using the state-of-the-art face generator StyleGAN in a process called latent variable evolution. Experiments demonstrated that even attackers with limited resources using only pre-trained models available on the Internet can initiate master face attacks. The results, in addition to demonstrating performance from the attacker's point of view, can also be used to clarify and improve the performance of face recognition systems and harden face authentication systems.
BLEU: a method for automatic evaluation of machine translation Human evaluations of machine translation are extensive but expensive. Human evaluations can take months to finish and involve human labor that can not be reused. We propose a method of automatic machine translation evaluation that is quick, inexpensive, and language-independent, that correlates highly with human evaluation, and that has little marginal cost per run. We present this method as an automated understudy to skilled human judges which substitutes for them when there is need for quick or frequent evaluations.
Wireless sensor networks: a survey This paper describes the concept of sensor networks which has been made viable by the convergence of micro-electro-mechanical systems technology, wireless communications and digital electronics. First, the sensing tasks and the potential sensor networks applications are explored, and a review of factors influencing the design of sensor networks is provided. Then, the communication architecture for sensor networks is outlined, and the algorithms and protocols developed for each layer in the literature are explored. Open research issues for the realization of sensor networks are also discussed.
Constrained Kalman filtering for indoor localization of transport vehicles using floor-installed HF RFID transponders Localization of transport vehicles is an important issue for many intralogistics applications. The paper presents an inexpensive solution for indoor localization of vehicles. Global localization is realized by detection of RFID transponders, which are integrated in the floor. The paper presents a novel algorithm for fusing RFID readings with odometry using Constraint Kalman filtering. The paper presents experimental results with a Mecanum based omnidirectional vehicle on a NaviFloor® installation, which includes passive HF RFID transponders. The experiments show that the proposed Constraint Kalman filter provides a similar localization accuracy compared to a Particle filter but with much lower computational expense.
On Multi-Access Edge Computing: A Survey of the Emerging 5G Network Edge Cloud Architecture and Orchestration. Multi-access edge computing (MEC) is an emerging ecosystem, which aims at converging telecommunication and IT services, providing a cloud computing platform at the edge of the radio access network. MEC offers storage and computational resources at the edge, reducing latency for mobile end users and utilizing more efficiently the mobile backhaul and core networks. This paper introduces a survey on ...
Precomputing Oblivious Transfer Alice and Bob are too untrusting of computer scientists to let their privacy depend on unproven assumptions such as the existence of one-way functions. Firm believers in Schrödinger and Heisenberg, they might accept a quantum OT device, but IBM’s prototype is not yet portable. Instead, as part of their prenuptial agreement, they decide to visit IBM and perform some OT’s in advance, so that any later divorces, coin-flipping or other important interactions can be done more conveniently, without needing expensive third parties. Unfortunately, OT can’t be done in advance in a direct way, because even though Bob might not know what bit Alice will later send (even if she first sends a random bit and later corrects it, for example), he would already know which bit or bits he will receive. We address the problem of precomputing oblivious transfer and show that OT can be precomputed at a cost of Θ(κ) prior transfers (a tight bound). In contrast, we show that variants of OT, such as one-out-of-two OT, can be precomputed using only one prior transfer. Finally, we show that all variants can be reduced to a single precomputed one-out-of-two oblivious transfer.
From structure-from-motion point clouds to fast location recognition Efficient view registration with respect to a given 3D re- construction has many applications like inside-out tracking in indoor and outdoor environments, and geo-locating im- ages from large photo collections. We present a fast loca- tion recognition technique based on structure from motion point clouds. Vocabulary tree-based indexing of features directly returns relevant fragments of 3D models instead of documents from the images database. Additionally, we pro- pose a compressed 3D scene representation which improves recognition rates while simultaneously reducing the compu- tation time and the memory consumption. The design of our method is based on algorithms that efficiently utilize mod- ern graphics processing units to deliver real-time perfor- mance for view registration. We demonstrate the approach by matching hand-held outdoor videos to known 3D urban models, and by registering images from online photo collec- tions to the corresponding landmarks.
Large System Analysis of Cooperative Multi-Cell Downlink Transmission via Regularized Channel Inversion with Imperfect CSIT In this paper, we analyze the ergodic sum-rate of a multi-cell downlink system with base station (BS) cooperation using regularized zero-forcing (RZF) precoding. Our model assumes that the channels between BSs and users have independent spatial correlations and imperfect channel state information at the transmitter (CSIT) is available. Our derivations are based on large dimensional random matrix theory (RMT) under the assumption that the numbers of antennas at the BS and users approach to infinity with some fixed ratios. In particular, a deterministic equivalent expression of the ergodic sum-rate is obtained and is instrumental in getting insight about the joint operations of BSs, which leads to an efficient method to find the asymptotic-optimal regularization parameter for the RZF. In another application, we use the deterministic channel rate to study the optimal feedback bit allocation among the BSs for maximizing the ergodic sum-rate, subject to a total number of feedback bits constraint. By inspecting the properties of the allocation, we further propose a scheme to greatly reduce the search space for optimization. Simulation results demonstrate that the ergodic sum-rates achievable by a subspace search provides comparable results to those by an exhaustive search under various typical settings.
A Model Predictive Control Approach to Microgrid Operation Optimization. Microgrids are subsystems of the distribution grid, which comprises generation capacities, storage devices, and controllable loads, operating as a single controllable system either connected or isolated from the utility grid. In this paper, we present a study on applying a model predictive control approach to the problem of efficiently optimizing microgrid operations while satisfying a time-varying request and operation constraints. The overall problem is formulated using mixed-integer linear programming (MILP), which can be solved in an efficient way by using commercial solvers without resorting to complex heuristics or decompositions techniques. Then, the MILP formulation leads to significant improvements in solution quality and computational burden. A case study of a microgrid is employed to assess the performance of the online optimization-based control strategy and the simulation results are discussed. The method is applied to an experimental microgrid located in Athens, Greece. The experimental results show the feasibility and the effectiveness of the proposed approach.
Quaternion polar harmonic Fourier moments for color images. •Quaternion polar harmonic Fourier moments (QPHFM) is proposed.•Complex Chebyshev-Fourier moments (CHFM) is extended to quaternion QCHFM.•Comparison experiments between QPHFM and QZM, QPZM, QOFMM, QCHFM and QRHFM are conducted.•QPHFM performs superbly in image reconstruction and invariant object recognition.•The importance of phase information of QPHFM in image reconstruction are discussed.
Ethical Considerations Of Applying Robots In Kindergarten Settings: Towards An Approach From A Macroperspective In child-robot interaction (cHRI) research, many studies pursue the goal to develop interactive systems that can be applied in everyday settings. For early education, increasingly, the setting of a kindergarten is targeted. However, when cHRI and research are brought into a kindergarten, a range of ethical and related procedural aspects have to be considered and dealt with. While ethical models elaborated within other human-robot interaction settings, e.g., assisted living contexts, can provide some important indicators for relevant issues, we argue that it is important to start developing a systematic approach to identify and tackle those ethical issues which rise with cHRI in kindergarten settings on a more global level and address the impact of the technology from a macroperspective beyond the effects on the individual. Based on our experience in conducting studies with children in general and pedagogical considerations on the role of the institution of kindergarten in specific, in this paper, we enfold some relevant aspects that have barely been addressed in an explicit way in current cHRI research. Four areas are analyzed and key ethical issues are identified in each area: (1) the institutional setting of a kindergarten, (2) children as a vulnerable group, (3) the caregivers' role, and (4) pedagogical concepts. With our considerations, we aim at (i) broadening the methodology of the current studies within the area of cHRI, (ii) revalidate it based on our comprehensive empirical experience with research in kindergarten settings, both laboratory and real-world contexts, and (iii) provide a framework for the development of a more systematic approach to address the ethical issues in cHRI research within kindergarten settings.
1.2
0.1
0.066667
0.066667
0
0
0
0
0
0
0
0
0
0
Blockchain Simulators: A Systematic Mapping Study Recently, distributed ledger technologies like blockchain have been proliferating and have attracted interest from the academic community, government, and industry. A wide range of blockchain solutions has been introduced, such as Bitcoin, Ethereum, and Hyperledger technologies in the literature. However, tools for evaluating these solutions and their applications are still lacking, limiting the exploration of their potentiality and associated challenges/limitations. That is, experimenting with real blockchain networks usually requires a solid budget; and thus, sophisticated blockchain simulators can facilitate designing and evaluating solutions before the actual implementation stage. The quality of such simulators depends on several factors such as usability, reliability, provided capabilities, and supported features. This paper aims to provide a systemic mapping review of blockchain simulators focusing on these quality factors. This paper also sheds light on the configuration parameters (inputs) and produced metrics (outputs) supported by each simulator. Furthermore, it investigates which metrics supported by each simulator are scientifically validated/evaluated. Moreover, code quality comparison is carried out to assess the source code of the covered simulators. The results reveal that no simulator fully covers the wide operational range of features and capabilities of existing blockchain technologies. However, several promising efforts exist in the domain of blockchain simulation with interesting and useful features. Finally, we discuss the subject of blockchain simulation and provide our insight into the matter.
The Sybil Attack Large-scale peer-to-peer systems facesecurity threats from faulty or hostile remotecomputing elements. To resist these threats, manysuch systems employ redundancy. However, if asingle faulty entity can present multiple identities,it can control a substantial fraction of the system,thereby undermining this redundancy. Oneapproach to preventing these &quot;Sybil attacks&quot; is tohave a trusted agency certify identities. Thispaper shows that, without a logically centralizedauthority, Sybil...
BLEU: a method for automatic evaluation of machine translation Human evaluations of machine translation are extensive but expensive. Human evaluations can take months to finish and involve human labor that can not be reused. We propose a method of automatic machine translation evaluation that is quick, inexpensive, and language-independent, that correlates highly with human evaluation, and that has little marginal cost per run. We present this method as an automated understudy to skilled human judges which substitutes for them when there is need for quick or frequent evaluations.
Computational thinking Summary form only given. My vision for the 21st century, Computational Thinking, will be a fundamental skill used by everyone in the world. To reading, writing, and arithmetic, we should add computational thinking to every child's analytical ability. Computational thinking involves solving problems, designing systems, and understanding human behavior by drawing on the concepts fundamental to computer science. Thinking like a computer scientist means more than being able to program a computer. It requires the ability to abstract and thus to think at multiple levels of abstraction. In this talk I will give many examples of computational thinking, argue that it has already influenced other disciplines, and promote the idea that teaching computational thinking can not only inspire future generations to enter the field of computer science but benefit people in all fields.
Fuzzy logic in control systems: fuzzy logic controller. I.
Switching between stabilizing controllers This paper deals with the problem of switching between several linear time-invariant (LTI) controllers—all of them capable of stabilizing a speci4c LTI process—in such a way that the stability of the closed-loop system is guaranteed for any switching sequence. We show that it is possible to 4nd realizations for any given family of controller transfer matrices so that the closed-loop system remains stable, no matter how we switch among the controller. The motivation for this problem is the control of complex systems where con8icting requirements make a single LTI controller unsuitable. ? 2002 Published by Elsevier Science Ltd.
Tabu Search - Part I
Bidirectional recurrent neural networks In the first part of this paper, a regular recurrent neural network (RNN) is extended to a bidirectional recurrent neural network (BRNN). The BRNN can be trained without the limitation of using input information just up to a preset future frame. This is accomplished by training it simultaneously in positive and negative time direction. Structure and training procedure of the proposed network are explained. In regression and classification experiments on artificial data, the proposed structure gives better results than other approaches. For real data, classification experiments for phonemes from the TIMIT database show the same tendency. In the second part of this paper, it is shown how the proposed bidirectional structure can be easily modified to allow efficient estimation of the conditional posterior probability of complete symbol sequences without making any explicit assumption about the shape of the distribution. For this part, experiments on real data are reported
An intensive survey of fair non-repudiation protocols With the phenomenal growth of the Internet and open networks in general, security services, such as non-repudiation, become crucial to many applications. Non-repudiation services must ensure that when Alice sends some information to Bob over a network, neither Alice nor Bob can deny having participated in a part or the whole of this communication. Therefore a fair non-repudiation protocol has to generate non-repudiation of origin evidences intended to Bob, and non-repudiation of receipt evidences destined to Alice. In this paper, we clearly define the properties a fair non-repudiation protocol must respect, and give a survey of the most important non-repudiation protocols without and with trusted third party (TTP). For the later ones we discuss the evolution of the TTP's involvement and, between others, describe the most recent protocol using a transparent TTP. We also discuss some ad-hoc problems related to the management of non-repudiation evidences.
Dynamic movement and positioning of embodied agents in multiparty conversations For embodied agents to engage in realistic multiparty conversation, they must stand in appropriate places with respect to other agents and the environment. When these factors change, such as an agent joining the conversation, the agents must dynamically move to a new location and/or orientation to accommodate. This paper presents an algorithm for simulating movement of agents based on observed human behavior using techniques developed for pedestrian movement in crowd simulations. We extend a previous group conversation simulation to include an agent motion algorithm. We examine several test cases and show how the simulation generates results that mirror real-life conversation settings.
An improved genetic algorithm with conditional genetic operators and its application to set-covering problem The genetic algorithm (GA) is a popular, biologically inspired optimization method. However, in the GA there is no rule of thumb to design the GA operators and select GA parameters. Instead, trial-and-error has to be applied. In this paper we present an improved genetic algorithm in which crossover and mutation are performed conditionally instead of probability. Because there are no crossover rate and mutation rate to be selected, the proposed improved GA can be more easily applied to a problem than the conventional genetic algorithms. The proposed improved genetic algorithm is applied to solve the set-covering problem. Experimental studies show that the improved GA produces better results over the conventional one and other methods.
Lane-level traffic estimations using microscopic traffic variables This paper proposes a novel inference method to estimate lane-level traffic flow, time occupancy and vehicle inter-arrival time on road segments where local information could not be measured and assessed directly. The main contributions of the proposed method are 1) the ability to perform lane-level estimations of traffic flow, time occupancy and vehicle inter-arrival time and 2) the ability to adapt to different traffic regimes by assessing only microscopic traffic variables. We propose a modified Kriging estimation model which explicitly takes into account both spatial and temporal variability. Performance evaluations are conducted using real-world data under different traffic regimes and it is shown that the proposed method outperforms a Kalman filter-based approach.
Convolutional Neural Network-Based Classification of Driver's Emotion during Aggressive and Smooth Driving Using Multi-Modal Camera Sensors. Because aggressive driving often causes large-scale loss of life and property, techniques for advance detection of adverse driver emotional states have become important for the prevention of aggressive driving behaviors. Previous studies have primarily focused on systems for detecting aggressive driver emotion via smart-phone accelerometers and gyro-sensors, or they focused on methods of detecting physiological signals using electroencephalography (EEG) or electrocardiogram (ECG) sensors. Because EEG and ECG sensors cause discomfort to drivers and can be detached from the driver's body, it becomes difficult to focus on bio-signals to determine their emotional state. Gyro-sensors and accelerometers depend on the performance of GPS receivers and cannot be used in areas where GPS signals are blocked. Moreover, if driving on a mountain road with many quick turns, a driver's emotional state can easily be misrecognized as that of an aggressive driver. To resolve these problems, we propose a convolutional neural network (CNN)-based method of detecting emotion to identify aggressive driving using input images of the driver's face, obtained using near-infrared (NIR) light and thermal camera sensors. In this research, we conducted an experiment using our own database, which provides a high classification accuracy for detecting driver emotion leading to either aggressive or smooth (i.e., relaxed) driving. Our proposed method demonstrates better performance than existing methods.
Ethical Considerations Of Applying Robots In Kindergarten Settings: Towards An Approach From A Macroperspective In child-robot interaction (cHRI) research, many studies pursue the goal to develop interactive systems that can be applied in everyday settings. For early education, increasingly, the setting of a kindergarten is targeted. However, when cHRI and research are brought into a kindergarten, a range of ethical and related procedural aspects have to be considered and dealt with. While ethical models elaborated within other human-robot interaction settings, e.g., assisted living contexts, can provide some important indicators for relevant issues, we argue that it is important to start developing a systematic approach to identify and tackle those ethical issues which rise with cHRI in kindergarten settings on a more global level and address the impact of the technology from a macroperspective beyond the effects on the individual. Based on our experience in conducting studies with children in general and pedagogical considerations on the role of the institution of kindergarten in specific, in this paper, we enfold some relevant aspects that have barely been addressed in an explicit way in current cHRI research. Four areas are analyzed and key ethical issues are identified in each area: (1) the institutional setting of a kindergarten, (2) children as a vulnerable group, (3) the caregivers' role, and (4) pedagogical concepts. With our considerations, we aim at (i) broadening the methodology of the current studies within the area of cHRI, (ii) revalidate it based on our comprehensive empirical experience with research in kindergarten settings, both laboratory and real-world contexts, and (iii) provide a framework for the development of a more systematic approach to address the ethical issues in cHRI research within kindergarten settings.
1
0
0
0
0
0
0
0
0
0
0
0
0
0
Weakly-Supervised Video Object Grounding by Exploring Spatio-Temporal Contexts Grounding objects in visual context from natural language queries is a crucial yet challenging vision-and-language task, which has gained increasing attention in recent years. Existing work has primarily investigated this task in the context of still images. Despite their effectiveness, these methods cannot be directly migrated into the video context, mainly due to 1) the complex spatio-temporal structure of videos and 2) the scarcity of fine-grained annotations of videos. To effectively ground objects in videos is profoundly more challenging and less explored. To fill the research gap, this paper presents a weakly-supervised framework for linking objects mentioned in a sentence with the corresponding regions in videos. It mainly considers two types of video characteristics: 1) objects are dynamically distributed across multiple frames and have diverse temporal durations, and 2) object regions in videos are spatially correlated with each other. Specifically, we propose a weakly-supervised video object grounding approach which mainly consists of three modules: 1) a temporal localization module to model the latent relation between queried objects and frames with a temporal attention network, 2) a spatial interaction module to capture feature correlation among object regions for learning context-aware region representation, and 3) a hierarchical video multiple instance learning algorithm to estimate the sentence-segment grounding score for discriminative training. Extensive experiments demonstrate that our method can achieve consistent improvement over the state-of-the-arts.
Deconfounded Video Moment Retrieval with Causal Intervention ABSTRACTWe tackle the task of video moment retrieval (VMR), which aims to localize a specific moment in a video according to a textual query. Existing methods primarily model the matching relationship between query and moment by complex cross-modal interactions. Despite their effectiveness, current models mostly exploit dataset biases while ignoring the video content, thus leading to poor generalizability. We argue that the issue is caused by the hidden confounder in VMR, i.e., temporal location of moments, that spuriously correlates the model input and prediction. How to design robust matching models against the temporal location biases is crucial but, as far as we know, has not been studied yet for VMR. To fill the research gap, we propose a causality-inspired VMR framework that builds structural causal model to capture the true effect of query and video content on the prediction. Specifically, we develop a Deconfounded Cross-modal Matching (DCM) method to remove the confounding effects of moment location. It first disentangles moment representation to infer the core feature of visual content, and then applies causal intervention on the disentangled multimodal input based on backdoor adjustment, which forces the model to fairly incorporate each possible location of the target into consideration. Extensive experiments clearly show that our approach can achieve significant improvement over the state-of-the-art methods in terms of both accuracy and generalization.
Annotating Objects and Relations in User-Generated Videos. Understanding the objects and relations between them is indispensable to fine-grained video content analysis, which is widely studied in recent research works in multimedia and computer vision. However, existing works are limited to evaluating with either small datasets or indirect metrics, such as the performance over images. The underlying reason is that the construction of a large-scale video dataset with dense annotation is tricky and costly. In this paper, we address several main issues in annotating objects and relations in user-generated videos, and propose an annotation pipeline that can be executed at a modest cost. As a result, we present a new dataset, named VidOR, consisting of 10k videos (84 hours) together with dense annotations that localize 80 categories of objects and 50 categories of predicates in each video. We have made the training and validation set public and extendable for more tasks to facilitate future research on video object and relation recognition.
Movie2Comics: Towards a Lively Video Content Presentation a type of artwork, comics is prevalent and popular around the world. However, despite the availability of assistive software and tools, the creation of comics is still a labor-intensive and time-consuming process. This paper proposes a scheme that is able to automatically turn a movie clip to comics. Two principles are followed in the scheme: 1) optimizing the information preservation of the movie; and 2) generating outputs following the rules and the styles of comics. The scheme mainly contains three components: script-face mapping, descriptive picture extraction, and cartoonization. The script-face mapping utilizes face tracking and recognition techniques to accomplish the mapping between characters' faces and their scripts. The descriptive picture extraction then generates a sequence of frames for presentation. Finally, the cartoonization is accomplished via three steps: panel scaling, stylization, and comics layout design. Experiments are conducted on a set of movie clips and the results have demonstrated the usefulness and the effectiveness of the scheme.
Semantic Understanding of Scenes through the ADE20K Dataset. Semantic understanding of visual scenes is one of the holy grails of computer vision. Despite efforts of the community in data collection, there are still few image datasets covering a wide range of scenes and object categories with pixel-wise annotations for scene understanding. In this work, we present a densely annotated dataset ADE20K, which spans diverse annotations of scenes, objects, parts of objects, and in some cases even parts of parts. Totally there are 25k images of the complex everyday scenes containing a variety of objects in their natural spatial context. On average there are 19.5 instances and 10.5 object classes per image. Based on ADE20K, we construct benchmarks for scene parsing and instance segmentation. We provide baseline performances on both of the benchmarks and re-implement state-of-the-art models for open source. We further evaluate the effect of synchronized batch normalization and find that a reasonably large batch size is crucial for the semantic segmentation performance. We show that the networks trained on ADE20K are able to segment a wide variety of scenes and objects.
Cycle-Consistent Diverse Image Synthesis from Natural Language Text-to-image translation has become an attractive yet challenging task in computer vision. Previous approaches tend to generate similar, or even monotonous, images for distinctive texts and overlook the characteristics of specific sentences. In this paper, we aim to generate images from the given texts by preserving diverse appearances and modes of the objects or instances contained. To achieve that, a novel learning model named SuperGAN is proposed, which consists of two major components: an image synthesis network and a captioning model in a Cycle-GAN framework. SuperGAN adopts the cycle-consistent adversarial training strategy to learn an image generator where the feature distribution of the generated images complies with the distribution of the generic images. Meanwhile, a cycle-consistency loss is applied to constrain that the caption of the generated images is closed to the original texts. Extensive experiments on the benchmark dataset Oxford-flowers-102 demonstrate the validity and effectiveness of our proposed method. In addition, a new evaluation metric is proposed to measure the diversity of synthetic results.
Ciagan: Conditional Identity Anonymization Generative Adversarial Networks The unprecedented increase in the usage of computer vision technology in society goes hand in hand with an increased concern in data privacy. In many real-world scenarios like people tracking or action recognition, it is important to be able to process the data while taking careful consideration in protecting people's identity. We propose and develop CIAGAN, a model for image and video anonymization based on conditional generative adversarial networks. Our model is able to remove the identifying characteristics of faces and bodies while producing high-quality images and videos that can be used for any computer vision task, such as detection or tracking. Unlike previous methods, we have MI control over the de-identification (anonymization) procedure, ensuring both anonymization as well as diversity. We compare our method to several baselines and achieve state-of-the-art results. To facilitate further research, we make available the code and the mod els at https://github.com/dvl-tum/ciagan.
DiverGAN: An Efficient and Effective Single-Stage Framework for Diverse Text-to-Image Generation In this paper, we concentrate on the text-to-image synthesis task that aims at automatically producing perceptually realistic pictures from text descriptions. Recently, several single-stage methods have been proposed to deal with the problems of a more complicated multi-stage modular architecture. However, they often suffer from the lack-of-diversity issue, yielding similar outputs given a single textual sequence. To this end, we present an efficient and effective single-stage framework (DiverGAN) to generate diverse, plausible and semantically consistent images according to a natural-language description. DiverGAN adopts two novel word-level attention modules, i.e., a channel-attention module (CAM) and a pixel-attention module (PAM), which model the importance of each word in the given sentence while allowing the network to assign larger weights to the significant channels and pixels semantically aligning with the salient words. After that, Conditional Adaptive Instance-Layer Normalization (CAdaILN) is introduced to enable the linguistic cues from the sentence embedding to flexibly manipulate the amount of change in shape and texture, further improving visual-semantic representation and helping stabilize the training. Also, a dual-residual structure is developed to preserve more original visual features while allowing for deeper networks, resulting in faster convergence speed and more vivid details. Furthermore, we propose to plug a fully-connected layer into the pipeline to address the lack-of-diversity problem, since we observe that a dense layer will remarkably enhance the generative capability of the network, balancing the trade-off between a low-dimensional random latent code contributing to variants and modulation modules that use high-dimensional and textual contexts to strength feature maps. Inserting a linear layer after the second residual block achieves the best variety and quality. Both qualitative and quantitative results on benchmark data sets demonstrate the superiority of our DiverGAN for realizing diversity, without harming quality and semantic consistency.
The Sybil Attack Large-scale peer-to-peer systems facesecurity threats from faulty or hostile remotecomputing elements. To resist these threats, manysuch systems employ redundancy. However, if asingle faulty entity can present multiple identities,it can control a substantial fraction of the system,thereby undermining this redundancy. Oneapproach to preventing these &quot;Sybil attacks&quot; is tohave a trusted agency certify identities. Thispaper shows that, without a logically centralizedauthority, Sybil...
Mean Shift, Mode Seeking, and Clustering Mean shift, a simple iterative procedure that shifts each data point to the average of data points in its neighborhood, is generalized and analyzed in this paper. This generalization makes some k-means like clustering algorithms its special cases. It is shown that mean shift is a mode-seeking process on a surface constructed with a 驴shadow驴 kernel. For Gaussian kernels, mean shift is a gradient mapping. Convergence is studied for mean shift iterations. Cluster analysis is treated as a deterministic problem of finding a fixed point of mean shift that characterizes the data. Applications in clustering and Hough transform are demonstrated. Mean shift is also considered as an evolutionary strategy that performs multistart global optimization.
Student Models that Invite the Learner In: The SMILI:() Open Learner Modelling Framework In recent years, the learner models of some adaptive learning environments have been opened to the learners they represent. However, as yet there is no standard way of describing and analysing these 'open learner models'. This is, in part, due to the variety of issues that can be important or relevant in any particular learner model. The lack of a framework to discuss open learner models poses several difficulties: there is no systematic way to analyse and describe the open learner models of any one system; there is no systematic way to compare the features of open learner models in different systems; and the designers of each new adaptive learning system must repeatedly tread the same path of studying the many diverse uses and approaches of open learner modelling so that they might determine how to make use of open learner modelling in their system. We believe this is a serious barrier to the effective use of open learner models. This paper presents such a framework, and gives examples of its use to describe and compare adaptive educational systems.
Image Feature Extraction in Encrypted Domain With Privacy-Preserving SIFT Privacy has received considerable attention but is still largely ignored in the multimedia community. Consider a cloud computing scenario where the server is resource-abundant, and is capable of finishing the designated tasks. It is envisioned that secure media applications with privacy preservation will be treated seriously. In view of the fact that scale-invariant feature transform (SIFT) has been widely adopted in various fields, this paper is the first to target the importance of privacy-preserving SIFT (PPSIFT) and to address the problem of secure SIFT feature extraction and representation in the encrypted domain. As all of the operations in SIFT must be moved to the encrypted domain, we propose a privacy-preserving realization of the SIFT method based on homomorphic encryption. We show through the security analysis based on the discrete logarithm problem and RSA that PPSIFT is secure against ciphertext only attack and known plaintext attack. Experimental results obtained from different case studies demonstrate that the proposed homomorphic encryption-based privacy-preserving SIFT performs comparably to the original SIFT and that our method is useful in SIFT-based privacy-preserving applications.
Fast Map Matching, An Algorithm Integrating Hidden Markov Model With Precomputation Wide deployment of global positioning system (GPS) sensors has generated a large amount of data with numerous applications in transportation research. Due to the observation error, a map matching (MM) process is commonly performed to infer a path on a road network from a noisy GPS trajectory. The increasing data volume calls for the design of efficient and scalable MM algorithms. This article presents fast map matching (FMM), an algorithm integrating hidden Markov model with precomputation, and provides an open-source implementation. An upper bounded origin-destination table is precomputed to store all pairs of shortest paths within a certain length in the road network. As a benefit, repeated routing queries known as the bottleneck of MM are replaced with hash table search. Additionally, several degenerate cases and a problem of reverse movement are identified and addressed in FMM. Experiments on a large collection of real-world taxi trip trajectories demonstrate that FMM has achieved a considerable single-processor MM speed of 25,000-45,000 points/second varying with the output mode. Investigation on the running time of different steps in FMM reveals that after precomputation is employed, the new bottleneck is located in candidate search, and more specifically, the projection of a GPS point to the polyline of a road edge. Reverse movement in the result is also effectively reduced by applying a penalty.
Intention-detection strategies for upper limb exosuits: model-based myoelectric vs dynamic-based control The cognitive human-robot interaction between an exosuit and its wearer plays a key role in determining both the biomechanical effects of the device on movements and its perceived effectiveness. There is a lack of evidence, however, on the comparative performance of different control methods, implemented on the same device. Here, we compare two different control approaches on the same robotic suit: a model-based myoelectric control (myoprocessor), which estimates the joint torque from the activation of target muscles, and a dynamic-based control that provides support against gravity using an inverse dynamic model. Tested on a cohort of four healthy participants, assistance from the exosuit results in a marked reduction in the effort of muscles working against gravity with both control approaches (peak reduction of 68.6±18.8%, for the dynamic arm model and 62.4±25.1% for the myoprocessor), when compared to an unpowered condition. Neither of the two controllers had an affect on the performance of their users in a joint-angle tracking task (peak errors of 15.4° and 16.4° for the dynamic arm model and myoprocessor, respectively, compared to 13.1o in the unpowered condition). However, our results highlight the remarkable adaptability of the myoprocessor to seamlessly adapt to changing external dynamics.
1.066667
0.066667
0.066667
0.066667
0.066667
0.066667
0.066667
0.066667
0
0
0
0
0
0
Preference-Driven Texture Modeling Through Interactive Generation and Search Data-driven texture modeling and rendering has pushed the limit of realism in haptics. However, the lack of haptic texture databases, difficulties of model interpolation and expansion, and the complexity of real textures prevent data-driven methods from capturing a large variety of textures and from customizing models to suit specific output hardware or user needs. This work proposes an interactive texture generation and search framework driven by user input. We design a GAN-based texture model generator, which can create a wide range of texture models using Auto-Regressive processes. Our interactive texture search method, which we call “ <italic xmlns:mml="http://www.w3.org/1998/Math/MathML" xmlns:xlink="http://www.w3.org/1999/xlink">preference-driven</i> ,” follows an evolutionary strategy given guidance from user's preferred feedback within a set of generated texture models. We implemented this framework on a 3D haptic device and conducted a two-phase user study to evaluate the efficiency and accuracy of our method for previously unmodeled textures. The results showed that by comparing the feel of real and generated virtual textures, users can follow an evolutionary process to efficiently find a virtual texture model that matched or exceeded the realism of a data-driven model. Furthermore, for 4 out of 5 real textures, <inline-formula xmlns:mml="http://www.w3.org/1998/Math/MathML" xmlns:xlink="http://www.w3.org/1999/xlink"><tex-math notation="LaTeX">$\geq$</tex-math></inline-formula> 80% of the preference-driven models from participants were rated comparable to the data-driven models.
The Sybil Attack Large-scale peer-to-peer systems facesecurity threats from faulty or hostile remotecomputing elements. To resist these threats, manysuch systems employ redundancy. However, if asingle faulty entity can present multiple identities,it can control a substantial fraction of the system,thereby undermining this redundancy. Oneapproach to preventing these &quot;Sybil attacks&quot; is tohave a trusted agency certify identities. Thispaper shows that, without a logically centralizedauthority, Sybil...
BLEU: a method for automatic evaluation of machine translation Human evaluations of machine translation are extensive but expensive. Human evaluations can take months to finish and involve human labor that can not be reused. We propose a method of automatic machine translation evaluation that is quick, inexpensive, and language-independent, that correlates highly with human evaluation, and that has little marginal cost per run. We present this method as an automated understudy to skilled human judges which substitutes for them when there is need for quick or frequent evaluations.
Computational thinking Summary form only given. My vision for the 21st century, Computational Thinking, will be a fundamental skill used by everyone in the world. To reading, writing, and arithmetic, we should add computational thinking to every child's analytical ability. Computational thinking involves solving problems, designing systems, and understanding human behavior by drawing on the concepts fundamental to computer science. Thinking like a computer scientist means more than being able to program a computer. It requires the ability to abstract and thus to think at multiple levels of abstraction. In this talk I will give many examples of computational thinking, argue that it has already influenced other disciplines, and promote the idea that teaching computational thinking can not only inspire future generations to enter the field of computer science but benefit people in all fields.
Fuzzy logic in control systems: fuzzy logic controller. I.
Switching between stabilizing controllers This paper deals with the problem of switching between several linear time-invariant (LTI) controllers—all of them capable of stabilizing a speci4c LTI process—in such a way that the stability of the closed-loop system is guaranteed for any switching sequence. We show that it is possible to 4nd realizations for any given family of controller transfer matrices so that the closed-loop system remains stable, no matter how we switch among the controller. The motivation for this problem is the control of complex systems where con8icting requirements make a single LTI controller unsuitable. ? 2002 Published by Elsevier Science Ltd.
Tabu Search - Part I
Bidirectional recurrent neural networks In the first part of this paper, a regular recurrent neural network (RNN) is extended to a bidirectional recurrent neural network (BRNN). The BRNN can be trained without the limitation of using input information just up to a preset future frame. This is accomplished by training it simultaneously in positive and negative time direction. Structure and training procedure of the proposed network are explained. In regression and classification experiments on artificial data, the proposed structure gives better results than other approaches. For real data, classification experiments for phonemes from the TIMIT database show the same tendency. In the second part of this paper, it is shown how the proposed bidirectional structure can be easily modified to allow efficient estimation of the conditional posterior probability of complete symbol sequences without making any explicit assumption about the shape of the distribution. For this part, experiments on real data are reported
An intensive survey of fair non-repudiation protocols With the phenomenal growth of the Internet and open networks in general, security services, such as non-repudiation, become crucial to many applications. Non-repudiation services must ensure that when Alice sends some information to Bob over a network, neither Alice nor Bob can deny having participated in a part or the whole of this communication. Therefore a fair non-repudiation protocol has to generate non-repudiation of origin evidences intended to Bob, and non-repudiation of receipt evidences destined to Alice. In this paper, we clearly define the properties a fair non-repudiation protocol must respect, and give a survey of the most important non-repudiation protocols without and with trusted third party (TTP). For the later ones we discuss the evolution of the TTP's involvement and, between others, describe the most recent protocol using a transparent TTP. We also discuss some ad-hoc problems related to the management of non-repudiation evidences.
Dynamic movement and positioning of embodied agents in multiparty conversations For embodied agents to engage in realistic multiparty conversation, they must stand in appropriate places with respect to other agents and the environment. When these factors change, such as an agent joining the conversation, the agents must dynamically move to a new location and/or orientation to accommodate. This paper presents an algorithm for simulating movement of agents based on observed human behavior using techniques developed for pedestrian movement in crowd simulations. We extend a previous group conversation simulation to include an agent motion algorithm. We examine several test cases and show how the simulation generates results that mirror real-life conversation settings.
An improved genetic algorithm with conditional genetic operators and its application to set-covering problem The genetic algorithm (GA) is a popular, biologically inspired optimization method. However, in the GA there is no rule of thumb to design the GA operators and select GA parameters. Instead, trial-and-error has to be applied. In this paper we present an improved genetic algorithm in which crossover and mutation are performed conditionally instead of probability. Because there are no crossover rate and mutation rate to be selected, the proposed improved GA can be more easily applied to a problem than the conventional genetic algorithms. The proposed improved genetic algorithm is applied to solve the set-covering problem. Experimental studies show that the improved GA produces better results over the conventional one and other methods.
Lane-level traffic estimations using microscopic traffic variables This paper proposes a novel inference method to estimate lane-level traffic flow, time occupancy and vehicle inter-arrival time on road segments where local information could not be measured and assessed directly. The main contributions of the proposed method are 1) the ability to perform lane-level estimations of traffic flow, time occupancy and vehicle inter-arrival time and 2) the ability to adapt to different traffic regimes by assessing only microscopic traffic variables. We propose a modified Kriging estimation model which explicitly takes into account both spatial and temporal variability. Performance evaluations are conducted using real-world data under different traffic regimes and it is shown that the proposed method outperforms a Kalman filter-based approach.
Convolutional Neural Network-Based Classification of Driver's Emotion during Aggressive and Smooth Driving Using Multi-Modal Camera Sensors. Because aggressive driving often causes large-scale loss of life and property, techniques for advance detection of adverse driver emotional states have become important for the prevention of aggressive driving behaviors. Previous studies have primarily focused on systems for detecting aggressive driver emotion via smart-phone accelerometers and gyro-sensors, or they focused on methods of detecting physiological signals using electroencephalography (EEG) or electrocardiogram (ECG) sensors. Because EEG and ECG sensors cause discomfort to drivers and can be detached from the driver's body, it becomes difficult to focus on bio-signals to determine their emotional state. Gyro-sensors and accelerometers depend on the performance of GPS receivers and cannot be used in areas where GPS signals are blocked. Moreover, if driving on a mountain road with many quick turns, a driver's emotional state can easily be misrecognized as that of an aggressive driver. To resolve these problems, we propose a convolutional neural network (CNN)-based method of detecting emotion to identify aggressive driving using input images of the driver's face, obtained using near-infrared (NIR) light and thermal camera sensors. In this research, we conducted an experiment using our own database, which provides a high classification accuracy for detecting driver emotion leading to either aggressive or smooth (i.e., relaxed) driving. Our proposed method demonstrates better performance than existing methods.
Robot tutor and pupils’ educational ability: Teaching the times tables Research shows promising results of educational robots in language and STEM tasks. In language, more research is available, occasionally in view of individual differences in pupils’ educational ability levels, and learning seems to improve with more expressive robot behaviors. In STEM, variations in robots’ behaviors have been examined with inconclusive results and never while systematically investigating how differences in educational abilities match with different robot behaviors. We applied an autonomously tutoring robot (without tablet, partly WOz) in a 2 × 2 experiment of social vs. neutral behavior in above-average vs. below-average schoolchildren (N = 86; age 8–10 years) while rehearsing the multiplication tables on a one-to-one basis. The standard school test showed that on average, pupils significantly improved their performance even after 3 occasions of 5-min exercises. Beyond-average pupils profited most from a robot tutor, whereas those below average in multiplication benefited more from a robot that showed neutral rather than more social behavior.
1
0
0
0
0
0
0
0
0
0
0
0
0
0
Tuft: Tree Based Heuristic Data Dissemination for Mobile Sink Wireless Sensor Networks Wireless sensor networks (WSNs) with a static sink suffer from concentrated data traffic in the vicinity of the sink, which increases the burden on the nodes surrounding the sink, and impels them to deplete their batteries faster than other nodes in the network. Mobile sinks solve this corollary by providing a more balanced traffic dispersion, by shifting the traffic concentration with the mobilit...
IoT Elements, Layered Architectures and Security Issues: A Comprehensive Survey. The use of the Internet is growing in this day and age, so another area has developed to use the Internet, called Internet of Things (IoT). It facilitates the machines and objects to communicate, compute and coordinate with each other. It is an enabler for the intelligence affixed to several essential features of the modern world, such as homes, hospitals, buildings, transports and cities. The security and privacy are some of the critical issues related to the wide application of IoT. Therefore, these issues prevent the wide adoption of the IoT. In this paper, we are presenting an overview about different layered architectures of IoT and attacks regarding security from the perspective of layers. In addition, a review of mechanisms that provide solutions to these issues is presented with their limitations. Furthermore, we have suggested a new secure layered architecture of IoT to overcome these issues.
A Multicharger Cooperative Energy Provision Algorithm Based On Density Clustering In The Industrial Internet Of Things Wireless sensor networks (WSNs) are an important core of the Industrial Internet of Things (IIoT). Wireless rechargeable sensor networks (WRSNs) are sensor networks that are charged by mobile chargers (MCs), and can achieve self-sufficiency. Therefore, the development of WRSNs has begun to attract widespread attention in recent years. Most of the existing energy replenishment algorithms for MCs use one or more MCs to serve the whole network in WRSNs. However, a single MC is not suitable for large-scale network environments, and multiple MCs make the network cost too high. Thus, this paper proposes a collaborative charging algorithm based on network density clustering (CCA-NDC) in WRSNs. This algorithm uses the mean-shift algorithm based on density to cluster, and then the mother wireless charger vehicle (MWCV) carries multiple sub wireless charger vehicles (SWCVs) to charge the nodes in each cluster by using a gradient descent optimization algorithm. The experimental results confirm that the proposed algorithm can effectively replenish the energy of the network and make the network more stable.
Dynamic Charging Scheme Problem With Actor–Critic Reinforcement Learning The energy problem is one of the most important challenges in the application of sensor networks. With the development of wireless charging technology and intelligent mobile charger (MC), the energy problem can be solved by the wireless charging strategy. In the practical application of wireless rechargeable sensor networks (WRSNs), the energy consumption rate of nodes is dynamically changed due to many uncertainties, such as the death and different transmission tasks of sensor nodes. However, existing works focus on on-demand schemes, which not fully consider real-time global charging scheduling. In this article, a novel dynamic charging scheme (DCS) in WRSN based on the actor-critic reinforcement learning (ACRL) algorithm is proposed. In the ACRL, we introduce gated recurrent units (GRUs) to capture the relationships of charging actions in time sequence. Using the actor network with one GRU layer, we can pick up an optimal or near-optimal sensor node from candidates as the next charging target more quickly and speed up the training of the model. Meanwhile, we take the tour length and the number of dead nodes as the reward signal. Actor and critic networks are updated by the error criterion function of R and V. Compared with current on-demand charging scheduling algorithms, extensive simulations show that the proposed ACRL algorithm surpasses heuristic algorithms, such as the Greedy, DP, nearest job next with preemption, and TSCA in the average lifetime and tour length, especially against the size and complexity increasing of WRSNs.
Adaptive Wireless Power Transfer in Mobile Ad Hoc Networks. We investigate the interesting impact of mobility on the problem of efficient wireless power transfer in ad hoc networks. We consider a set of mobile agents (consuming energy to perform certain sensing and communication tasks), and a single static charger (with finite energy) which can recharge the agents when they get in its range. In particular, we focus on the problem of efficiently computing the appropriate range of the charger with the goal of prolonging the network lifetime. We first demonstrate (under the realistic assumption of fixed energy supplies) the limitations of any fixed charging range and, therefore, the need for (and power of) a dynamic selection of the charging range, by adapting to the behavior of the mobile agents which is revealed in an online manner. We investigate the complexity of optimizing the selection of such an adaptive charging range, by showing that two simplified offline optimization problems (closely related to the online one) are NP-hard. To effectively address the involved performance trade-offs, we finally present a variety of adaptive heuristics, assuming different levels of agent information regarding their mobility and energy.
Reducing delay and prolonging the lifetime of wireless sensor network using efficient routing protocol based on mobile sink and virtual infrastructure Mobile sink has been used by many routing algorithms to improve the network lifetime in wireless sensor networks. Sink mobility brings new challenges to the sensor network. It suggests that sink position is continuously propagated through the network area to keep all sensor nodes updated with the direction of forwarding data to the sink. Unfortunately, frequent sink position updates can lead to both high-energy consumption and increased collisions in the network. In this paper, a new routing algorithm for wireless sensor networks is proposed, which uses a virtual multi-ring shaped infrastructure to advertise the mobile sink position to the network. In addition, the proposed algorithm supports several mobile sinks in the network. Simulation results indicate that the proposed algorithm reduces energy consumption and delay, and prolongs the sensor network lifetime when compared with the other similar algorithms.
GTCCS: A Game Theoretical Collaborative Charging Scheduling for On-Demand Charging Architecture. Battery energy is always limited in most wireless networked nodes, but wireless power transfer has the potential to address the problem for good. In this paper, we study the problem of collaborative charging decisions in Wireless Rechargeable Sensor Networks (WRSNs). In these networks, multiple Wireless Charging Vehicles (WCVs) need to decide which of the requesting sensors to charge and in what o...
Mobile Charging in Wireless-Powered Sensor Networks: Optimal Scheduling and Experimental Implementation. Wireless radio frequency (RF) energy transfer is a promising technology to provide a reliability-guaranteed power supply for wireless sensor networks. In this paper, we consider a special wireless-powered sensor network consisting of a mobile energy station that can travel through a preplanned path to charge wireless-powered sensors located in the considered area. We develop a hardware platform us...
Sequence to Sequence Learning with Neural Networks. Deep Neural Networks (DNNs) are powerful models that have achieved excellent performance on difficult learning tasks. Although DNNs work well whenever large labeled training sets are available, they cannot be used to map sequences to sequences. In this paper, we present a general end-to-end approach to sequence learning that makes minimal assumptions on the sequence structure. Our method uses a multilayered Long Short-Term Memory (LSTM) to map the input sequence to a vector of a fixed dimensionality, and then another deep LSTM to decode the target sequence from the vector. Our main result is that on an English to French translation task from the WMT-14 dataset, the translations produced by the LSTM achieve a BLEU score of 34.8 on the entire test set, where the LSTM's BLEU score was penalized on out-of-vocabulary words. Additionally, the LSTM did not have difficulty on long sentences. For comparison, a phrase-based SMT system achieves a BLEU score of 33.3 on the same dataset. When we used the LSTM to rerank the 1000 hypotheses produced by the aforementioned SMT system, its BLEU score increases to 36.5, which is close to the previous state of the art. The LSTM also learned sensible phrase and sentence representations that are sensitive to word order and are relatively invariant to the active and the passive voice. Finally, we found that reversing the order of the words in all source sentences (but not target sentences) improved the LSTM's performance markedly, because doing so introduced many short term dependencies between the source and the target sentence which made the optimization problem easier.
Fuzzy basis functions, universal approximation, and orthogonal least-squares learning. Fuzzy systems are represented as series expansions of fuzzy basis functions which are algebraic superpositions of fuzzy membership functions. Using the Stone-Weierstrass theorem, it is proved that linear combinations of the fuzzy basis functions are capable of uniformly approximating any real continuous function on a compact set to arbitrary accuracy. Based on the fuzzy basis function representations, an orthogonal least-squares (OLS) learning algorithm is developed for designing fuzzy systems based on given input-output pairs; then, the OLS algorithm is used to select significant fuzzy basis functions which are used to construct the final fuzzy system. The fuzzy basis function expansion is used to approximate a controller for the nonlinear ball and beam system, and the simulation results show that the control performance is improved by incorporating some common-sense fuzzy control rules.
A model of proximity control for information-presenting robots In this paper, we report a model that allows a robot to appropriately control its position as it presents information to a user. This capability is indispensable, since in the future, many robots will function in daily situations such as shopkeepers presenting products to customers or museum guides presenting information to visitors. Psychology research suggests that people adjust their positions to establish a joint view toward a target object. Similarly, when a robot presents an object, it should stand at an appropriate position that considers the positions of both the listener and the object to optimize the listener's field of view and establish a joint view. We observed human-human interaction situations, where people presented objects, and developed a model for an information-presenting robot to appropriately adjust its position. Our model consists of four constraints to establish O-space: 1) proximity to listener; 2) proximity to object; 3) listener's field of view; and 4) presenter's field of view.We also experimentally evaluate the effectiveness of our model.
Online Coordinated Charging Decision Algorithm for Electric Vehicles Without Future Information The large-scale integration of plug-in electric vehicles (PEVs) to the power grid spurs the need for efficient charging coordination mechanisms. It can be shown that the optimal charging schedule smooths out the energy consumption over time so as to minimize the total energy cost. In practice, however, it is hard to smooth out the energy consumption perfectly, because the future PEV charging demand is unknown at the moment when the charging rate of an existing PEV needs to be determined. In this paper, we propose an online coordinated charging decision (ORCHARD) algorithm, which minimizes the energy cost without knowing the future information. Through rigorous proof, we show that ORCHARD is strictly feasible in the sense that it guarantees to fulfill all charging demands before due time. Meanwhile, it achieves the best known competitive ratio of 2.39. By exploiting the problem structure, we propose a novel reduced-complexity algorithm to replace the standard convex optimization techniques used in ORCHARD. Through extensive simulations, we show that the average performance gap between ORCHARD and the offline optimal solution, which utilizes the complete future information, is as small as 6.5%. By setting a proper speeding factor, the average performance gap can be further reduced to 5%.
A Delay-Sensitive Multicast Protocol for Network Capacity Enhancement in Multirate MANETs. Due to significant advances in wireless modulation technologies, some MAC standards such as 802.11a, 802.11b, and 802.11g can operate with multiple data rates for QoS-constrained multimedia communication to utilize the limited resources of MANETs more efficiently. In this paper, by means of measuring the busy/idle ratio of the shared radio channel, a method for estimating one-hop delay is first su...
Robust PCA for Subspace Estimation in User-Centric Cell-Free Wireless Networks We consider a scalable user-centric cell-free massive MIMO network with distributed remote radio units (RUs), enabling macrodiversity and joint processing. Due to the limited uplink (UL) pilot dimension, multiuser interference in the UL pilot transmission phase makes channel estimation a non-trivial problem. We make use of two types of UL pilot signals, sounding reference signal (SRS) and demodulation reference signal (DMRS) pilots, for the estimation of the channel subspace and its instantaneous realization, respectively. The SRS pilots are transmitted over multiple time slots and resource blocks according to a Latin squares based hopping scheme, which aims at averaging out the interference of different SRS co-pilot users. We propose a robust principle component analysis approach for channel subspace estimation from the SRS signal samples, employed at the RUs for each associated user. The estimated subspace is further used at the RUs for DMRS pilot decontamination and instantaneous channel estimation. We provide numerical simulations to compare the system performance using our subspace and channel estimation scheme with the cases of ideal partial subspace/channel knowledge and pilot matching channel estimation. The results show that a system with a properly designed SRS pilot hopping scheme can closely approximate the performance of a genie-aided system.
1.11
0.12
0.12
0.12
0.12
0.1
0.06
0.006667
0
0
0
0
0
0
Exploring Human-Robot Trust Through the Investment Game - An Immersive Space Mission Scenario.
The Sybil Attack Large-scale peer-to-peer systems facesecurity threats from faulty or hostile remotecomputing elements. To resist these threats, manysuch systems employ redundancy. However, if asingle faulty entity can present multiple identities,it can control a substantial fraction of the system,thereby undermining this redundancy. Oneapproach to preventing these &quot;Sybil attacks&quot; is tohave a trusted agency certify identities. Thispaper shows that, without a logically centralizedauthority, Sybil...
BLEU: a method for automatic evaluation of machine translation Human evaluations of machine translation are extensive but expensive. Human evaluations can take months to finish and involve human labor that can not be reused. We propose a method of automatic machine translation evaluation that is quick, inexpensive, and language-independent, that correlates highly with human evaluation, and that has little marginal cost per run. We present this method as an automated understudy to skilled human judges which substitutes for them when there is need for quick or frequent evaluations.
Computational thinking Summary form only given. My vision for the 21st century, Computational Thinking, will be a fundamental skill used by everyone in the world. To reading, writing, and arithmetic, we should add computational thinking to every child's analytical ability. Computational thinking involves solving problems, designing systems, and understanding human behavior by drawing on the concepts fundamental to computer science. Thinking like a computer scientist means more than being able to program a computer. It requires the ability to abstract and thus to think at multiple levels of abstraction. In this talk I will give many examples of computational thinking, argue that it has already influenced other disciplines, and promote the idea that teaching computational thinking can not only inspire future generations to enter the field of computer science but benefit people in all fields.
Fuzzy logic in control systems: fuzzy logic controller. I.
Switching between stabilizing controllers This paper deals with the problem of switching between several linear time-invariant (LTI) controllers—all of them capable of stabilizing a speci4c LTI process—in such a way that the stability of the closed-loop system is guaranteed for any switching sequence. We show that it is possible to 4nd realizations for any given family of controller transfer matrices so that the closed-loop system remains stable, no matter how we switch among the controller. The motivation for this problem is the control of complex systems where con8icting requirements make a single LTI controller unsuitable. ? 2002 Published by Elsevier Science Ltd.
Tabu Search - Part I
Bidirectional recurrent neural networks In the first part of this paper, a regular recurrent neural network (RNN) is extended to a bidirectional recurrent neural network (BRNN). The BRNN can be trained without the limitation of using input information just up to a preset future frame. This is accomplished by training it simultaneously in positive and negative time direction. Structure and training procedure of the proposed network are explained. In regression and classification experiments on artificial data, the proposed structure gives better results than other approaches. For real data, classification experiments for phonemes from the TIMIT database show the same tendency. In the second part of this paper, it is shown how the proposed bidirectional structure can be easily modified to allow efficient estimation of the conditional posterior probability of complete symbol sequences without making any explicit assumption about the shape of the distribution. For this part, experiments on real data are reported
An intensive survey of fair non-repudiation protocols With the phenomenal growth of the Internet and open networks in general, security services, such as non-repudiation, become crucial to many applications. Non-repudiation services must ensure that when Alice sends some information to Bob over a network, neither Alice nor Bob can deny having participated in a part or the whole of this communication. Therefore a fair non-repudiation protocol has to generate non-repudiation of origin evidences intended to Bob, and non-repudiation of receipt evidences destined to Alice. In this paper, we clearly define the properties a fair non-repudiation protocol must respect, and give a survey of the most important non-repudiation protocols without and with trusted third party (TTP). For the later ones we discuss the evolution of the TTP's involvement and, between others, describe the most recent protocol using a transparent TTP. We also discuss some ad-hoc problems related to the management of non-repudiation evidences.
Dynamic movement and positioning of embodied agents in multiparty conversations For embodied agents to engage in realistic multiparty conversation, they must stand in appropriate places with respect to other agents and the environment. When these factors change, such as an agent joining the conversation, the agents must dynamically move to a new location and/or orientation to accommodate. This paper presents an algorithm for simulating movement of agents based on observed human behavior using techniques developed for pedestrian movement in crowd simulations. We extend a previous group conversation simulation to include an agent motion algorithm. We examine several test cases and show how the simulation generates results that mirror real-life conversation settings.
An improved genetic algorithm with conditional genetic operators and its application to set-covering problem The genetic algorithm (GA) is a popular, biologically inspired optimization method. However, in the GA there is no rule of thumb to design the GA operators and select GA parameters. Instead, trial-and-error has to be applied. In this paper we present an improved genetic algorithm in which crossover and mutation are performed conditionally instead of probability. Because there are no crossover rate and mutation rate to be selected, the proposed improved GA can be more easily applied to a problem than the conventional genetic algorithms. The proposed improved genetic algorithm is applied to solve the set-covering problem. Experimental studies show that the improved GA produces better results over the conventional one and other methods.
Lane-level traffic estimations using microscopic traffic variables This paper proposes a novel inference method to estimate lane-level traffic flow, time occupancy and vehicle inter-arrival time on road segments where local information could not be measured and assessed directly. The main contributions of the proposed method are 1) the ability to perform lane-level estimations of traffic flow, time occupancy and vehicle inter-arrival time and 2) the ability to adapt to different traffic regimes by assessing only microscopic traffic variables. We propose a modified Kriging estimation model which explicitly takes into account both spatial and temporal variability. Performance evaluations are conducted using real-world data under different traffic regimes and it is shown that the proposed method outperforms a Kalman filter-based approach.
Convolutional Neural Network-Based Classification of Driver's Emotion during Aggressive and Smooth Driving Using Multi-Modal Camera Sensors. Because aggressive driving often causes large-scale loss of life and property, techniques for advance detection of adverse driver emotional states have become important for the prevention of aggressive driving behaviors. Previous studies have primarily focused on systems for detecting aggressive driver emotion via smart-phone accelerometers and gyro-sensors, or they focused on methods of detecting physiological signals using electroencephalography (EEG) or electrocardiogram (ECG) sensors. Because EEG and ECG sensors cause discomfort to drivers and can be detached from the driver's body, it becomes difficult to focus on bio-signals to determine their emotional state. Gyro-sensors and accelerometers depend on the performance of GPS receivers and cannot be used in areas where GPS signals are blocked. Moreover, if driving on a mountain road with many quick turns, a driver's emotional state can easily be misrecognized as that of an aggressive driver. To resolve these problems, we propose a convolutional neural network (CNN)-based method of detecting emotion to identify aggressive driving using input images of the driver's face, obtained using near-infrared (NIR) light and thermal camera sensors. In this research, we conducted an experiment using our own database, which provides a high classification accuracy for detecting driver emotion leading to either aggressive or smooth (i.e., relaxed) driving. Our proposed method demonstrates better performance than existing methods.
Robot tutor and pupils’ educational ability: Teaching the times tables Research shows promising results of educational robots in language and STEM tasks. In language, more research is available, occasionally in view of individual differences in pupils’ educational ability levels, and learning seems to improve with more expressive robot behaviors. In STEM, variations in robots’ behaviors have been examined with inconclusive results and never while systematically investigating how differences in educational abilities match with different robot behaviors. We applied an autonomously tutoring robot (without tablet, partly WOz) in a 2 × 2 experiment of social vs. neutral behavior in above-average vs. below-average schoolchildren (N = 86; age 8–10 years) while rehearsing the multiplication tables on a one-to-one basis. The standard school test showed that on average, pupils significantly improved their performance even after 3 occasions of 5-min exercises. Beyond-average pupils profited most from a robot tutor, whereas those below average in multiplication benefited more from a robot that showed neutral rather than more social behavior.
1
0
0
0
0
0
0
0
0
0
0
0
0
0
Blockchain for Cybersecurity in Smart Grid: A Comprehensive Survey Blockchain is an immutable type of distributed ledger that is capable of storing data without relying on a third party. Blockchain technology has attracted significant interest in research areas, including its application in the smart grid for cybersecurity. Although significant efforts have been devoted to utilizing blockchain in the smart grid for cybersecurity, there is a lack of comprehensive survey on blockchain in the smart grid for cybersecurity in both application and technological perspectives. To fill this gap, we conducted a comprehensive survey on blockchain for smart gird cybersecurity. This conducted survey presents the latest insights of ideas, architectures, and techniques of implementation that are relevant to blockchain's application in the smart grid for cybersecurity. This article aims at providing helpful guidance and reference for future research efforts specific to blockchain for cybersecurity in the smart grid.
Local Load Redistribution Attacks in Power Systems With Incomplete Network Information Power grid is one of the most critical infrastructures in a nation and could suffer a variety of cyber attacks. Recent studies have shown that an attacker can inject pre-determined false data into smart meters such that it can pass the residue test of conventional state estimator. However, the calculation of the false data vector relies on the network (topology and parameter) information of the entire grid. In practice, it is impossible for an attacker to obtain all network information of a power grid. Unfortunately, this does not make power systems immune to false data injection attacks. In this paper, we propose a local load redistribution attacking model based on incomplete network information and show that an attacker only needs to obtain the network information of the local attacking region to inject false data into smart meters in the local region without being detected by the state estimator. Simulations on the modified IEEE 14-bus system demonstrate the correctness and effectiveness of the proposed model. The results of this paper reveal the mechanism of local false data injection attacks and highlight the importance and complexity of defending power systems against false data injection attacks.
Demand Response for Residential Appliances via Customer Reward Scheme. This paper proposes a reward based demand response algorithm for residential customers to shave network peaks. Customer survey information is used to calculate various criteria indices reflecting their priority and flexibility. Criteria indices and sensitivity based house ranking is used for appropriate load selection in the feeder for demand response. Customer Rewards (CR) are paid based on load shift and voltage improvement due to load adjustment. The proposed algorithm can be deployed in residential distribution networks using a two-level hierarchical control scheme. Realistic residential load model consisting of non-controllable and controllable appliances is considered in this study. The effectiveness of the proposed demand response scheme on the annual load growth of the feeder is also investigated. Simulation results show that reduced peak demand, improved network voltage performance, and customer satisfaction can be achieved.
Energy Crowdsourcing and Peer-to-Peer Energy Trading in Blockchain-Enabled Smart Grids. The power grid is rapidly transforming, and while recent grid innovations increased the utilization of advanced control methods, the next-generation grid demands technologies that enable the integration of distributed energy resources (DERs)—and consumers that both seamlessly buy and sell electricity. This paper develops an optimization model and blockchain-based architecture to manage the operation of crowdsourced energy systems (CESs), with peer-to-peer (P2P) energy trading transactions (ETTs). An operational model of CESs in distribution networks is presented considering various types of ETT and crowdsourcees. Then, a two-phase operation algorithm is presented: Phase I focuses on the day-ahead scheduling of generation and controllable DERs, whereas Phase II is developed for hour-ahead or real-time operation of distribution networks. The developed approach supports seamless P2P energy trading between individual prosumers and/or the utility. The presented operational model can also be used to operate islanded microgrids. The CES framework and the operation algorithm are then prototyped through an efficient blockchain implementation, namely, the IBM Hyperledger Fabric. This implementation allows the system operator to manage the network users to seamlessly trade energy. Case studies and prototype illustration are provided.
Deep Reinforcement Learning-based Capacity Scheduling for PV-Battery Storage System Investor-owned photovoltaic-battery storage systems (PV-BSS) can gain revenue by providing stacked services, including PV charging and frequency regulation, and by performing energy arbitrage. Capacity scheduling (CS) is a crucial component of PV-BSS energy management, aiming to ensure the secure and economic operation of the PV-BSS. This article proposes a Proximal Policy Optimization (PPO)-based...
Computational thinking Summary form only given. My vision for the 21st century, Computational Thinking, will be a fundamental skill used by everyone in the world. To reading, writing, and arithmetic, we should add computational thinking to every child's analytical ability. Computational thinking involves solving problems, designing systems, and understanding human behavior by drawing on the concepts fundamental to computer science. Thinking like a computer scientist means more than being able to program a computer. It requires the ability to abstract and thus to think at multiple levels of abstraction. In this talk I will give many examples of computational thinking, argue that it has already influenced other disciplines, and promote the idea that teaching computational thinking can not only inspire future generations to enter the field of computer science but benefit people in all fields.
JPEG Error Analysis and Its Applications to Digital Image Forensics JPEG is one of the most extensively used image formats. Understanding the inherent characteristics of JPEG may play a useful role in digital image forensics. In this paper, we introduce JPEG error analysis to the study of image forensics. The main errors of JPEG include quantization, rounding, and truncation errors. Through theoretically analyzing the effects of these errors on single and double JPEG compression, we have developed three novel schemes for image forensics including identifying whether a bitmap image has previously been JPEG compressed, estimating the quantization steps of a JPEG image, and detecting the quantization table of a JPEG image. Extensive experimental results show that our new methods significantly outperform existing techniques especially for the images of small sizes. We also show that the new method can reliably detect JPEG image blocks which are as small as 8 × 8 pixels and compressed with quality factors as high as 98. This performance is important for analyzing and locating small tampered regions within a composite image.
Highly dynamic Destination-Sequenced Distance-Vector routing (DSDV) for mobile computers An ad-hoc network is the cooperative engagement of a collection of Mobile Hosts without the required intervention of any centralized Access Point. In this paper we present an innovative design for the operation of such ad-hoc networks. The basic idea of the design is to operate each Mobile Host as a specialized router, which periodically advertises its view of the interconnection topology with other Mobile Hosts within the network. This amounts to a new sort of routing protocol. We have investigated modifications to the basic Bellman-Ford routing mechanisms, as specified by RIP [5], to make it suitable for a dynamic and self-starting network mechanism as is required by users wishing to utilize ad hoc networks. Our modifications address some of the previous objections to the use of Bellman-Ford, related to the poor looping properties of such algorithms in the face of broken links and the resulting time dependent nature of the interconnection topology describing the links between the Mobile Hosts. Finally, we describe the ways in which the basic network-layer routing can be modified to provide MAC-layer support for ad-hoc networks.
The FERET Evaluation Methodology for Face-Recognition Algorithms Two of the most critical requirements in support of producing reliable face-recognition systems are a large database of facial images and a testing procedure to evaluate systems. The Face Recognition Technology (FERET) program has addressed both issues through the FERET database of facial images and the establishment of the FERET tests. To date, 14,126 images from 1,199 individuals are included in the FERET database, which is divided into development and sequestered portions of the database. In September 1996, the FERET program administered the third in a series of FERET face-recognition tests. The primary objectives of the third test were to 1) assess the state of the art, 2) identify future areas of research, and 3) measure algorithm performance.
Neural fitted q iteration – first experiences with a data efficient neural reinforcement learning method This paper introduces NFQ, an algorithm for efficient and effective training of a Q-value function represented by a multi-layer perceptron. Based on the principle of storing and reusing transition experiences, a model-free, neural network based Reinforcement Learning algorithm is proposed. The method is evaluated on three benchmark problems. It is shown empirically, that reasonably few interactions with the plant are needed to generate control policies of high quality.
Labels and event processes in the Asbestos operating system Asbestos, a new operating system, provides novel labeling and isolation mechanisms that help contain the effects of exploitable software flaws. Applications can express a wide range of policies with Asbestos's kernel-enforced labels, including controls on interprocess communication and system-wide information flow. A new event process abstraction defines lightweight, isolated contexts within a single process, allowing one process to act on behalf of multiple users while preventing it from leaking any single user's data to others. A Web server demonstration application uses these primitives to isolate private user data. Since the untrusted workers that respond to client requests are constrained by labels, exploited workers cannot directly expose user data except as allowed by application policy. The server application requires 1.4 memory pages per user for up to 145,000 users and achieves connection rates similar to Apache, demonstrating that additional security can come at an acceptable cost.
Switching Stabilization for a Class of Slowly Switched Systems In this technical note, the problem of switching stabilization for slowly switched linear systems is investigated. In particular, the considered systems can be composed of all unstable subsystems. Based on the invariant subspace theory, the switching signal with mode-dependent average dwell time (MDADT) property is designed to exponentially stabilize the underlying system. Furthermore, sufficient condition of stabilization for switched systems with all stable subsystems under MDADT switching is also given. The correctness and effectiveness of the proposed approaches are illustrated by a numerical example.
An evolutionary programming approach for securing medical images using watermarking scheme in invariant discrete wavelet transformation. •The proposed watermarking scheme utilized improved discrete wavelet transformation (IDWT) to retrieve the invariant wavelet domain.•The entropy mechanism is used to identify the suitable region for insertion of watermark. This will improve the imperceptibility and robustness of the watermarking procedure.•The scaling factors such as PSNR and NC are considered for evaluation of the proposed method and the Particle Swarm Optimization is employed to optimize the scaling factors.
Social Robots for (Second) Language Learning in (Migrant) Primary School Children Especially these days, innovation and support from technology to relieve pressure in education is highly urgent. This study tested the potential advantage of a social robot over a tablet in (second) language learning on performance, engagement, and enjoyment. Shortages in primary education call for new technology solutions. Previous studies combined robots with tablets, to compensate for robot’s limitations, however, this study applied direct human–robot interaction. Primary school children (N = 63, aged 4–6) participated in a 3-wave field experiment with story-telling exercises, either with a semi-autonomous robot (without tablet, using WOz) or a tablet. Results showed increased learning gains over time when training with a social robot, compared to the tablet. Children who trained with a robot were more engaged in the story-telling task and enjoyed it more. Robot’s behavioral style (social or neutral) hardly differed overall, however, seems to vary for high versus low educational abilities. While social robots need sophistication before being implemented in schools, our study shows the potential of social robots as tutors in (second) language learning.
1.2
0.2
0.2
0.2
0.1
0
0
0
0
0
0
0
0
0
Observer-Based Fixed-Time Adaptive Fuzzy Bipartite Containment Control for Multiagent Systems With Unknown Hysteresis This article studies the fixed-time fuzzy adaptive bipartite containment quantized control problem for nonlinear multiagent systems subject to unknown external disturbances and unknown Bouc–Wen hysteresis. The output and input control signals of the systems are quantized by sector-bounded quantizer. A disturbance observer and a fuzzy state observer are simultaneously designed to estimate unknown e...
Distributed Containment Control for Multiple Unknown Second-Order Nonlinear Systems With Application to Networked Lagrangian Systems. In this paper, we consider the distributed containment control problem for multiagent systems with unknown nonlinear dynamics. More specifically, we focus on multiple second-order nonlinear systems and networked Lagrangian systems. We first study the distributed containment control problem for multiple second-order nonlinear systems with multiple dynamic leaders in the presence of unknown nonlinearities and external disturbances under a general directed graph that characterizes the interaction among the leaders and the followers. A distributed adaptive control algorithm with an adaptive gain design based on the approximation capability of neural networks is proposed. We present a necessary and sufficient condition on the directed graph such that the containment error can be reduced as small as desired. As a byproduct, the leaderless consensus problem is solved with asymptotical convergence. Because relative velocity measurements between neighbors are generally more difficult to obtain than relative position measurements, we then propose a distributed containment control algorithm without using neighbors' velocity information. A two-step Lyapunov-based method is used to study the convergence of the closed-loop system. Next, we apply the ideas to deal with the containment control problem for networked unknown Lagrangian systems under a general directed graph. All the proposed algorithms are distributed and can be implemented using only local measurements in the absence of communication. Finally, simulation examples are provided to show the effectiveness of the proposed control algorithms.
Fully distributed containment control of high-order multi-agent systems with nonlinear dynamics. In this paper, distributed containment control problems for high-order multi-agent systems with nonlinear dynamics are investigated under directed communication topology. The states of the leaders are only available to a subset of the followers and the inputs of the leaders are possibly nonzero and time varying. Distributed adaptive nonlinear protocol is proposed based only on the relative state information, under which the states of the followers converge to the dynamic convex hull spanned by those of the leaders. As the special case with only one dynamic leader, leader–follower consensus problem is also solved with the proposed protocol. The adaptive protocol here is independent of the eigenvalues of the Laplacian matrix, which means the protocol can be implemented by each agent in a fully distributed fashion. A simulation example is provided to illustrate the theoretical results.
Output Containment Control of Linear Heterogeneous Multi-Agent Systems Using Internal Model Principle. This paper studies the output containment control of linear heterogeneous multi-agent systems, where the system dynamics and even the state dimensions can generally be different. Since the states can have different dimensions, standard results from state containment control do not apply. Therefore, the control objective is to guarantee the convergence of the output of each follower to the dynamic ...
Finite-Time Consensus Tracking Neural Network FTC of Multi-Agent Systems The finite-time consensus fault-tolerant control (FTC) tracking problem is studied for the nonlinear multi-agent systems (MASs) in the nonstrict feedback form. The MASs are subject to unknown symmetric output dead zones, actuator bias and gain faults, and unknown control coefficients. According to the properties of the neural network (NN), the unstructured uncertainties problem is solved. The Nussbaum function is used to address the output dead zones and unknown control directions problems. By introducing an arbitrarily small positive number, the “singularity” problem caused by combining the finite-time control and backstepping design is solved. According to the backstepping design and Lyapunov stability theory, a finite-time adaptive NN FTC controller is obtained, which guarantees that the tracking error converges to a small neighborhood of zero in a finite time, and all signals in the closed-loop system are bounded. Finally, the effectiveness of the proposed method is illustrated via a physical example.
Observer-based Adaptive Fuzzy Containment Control for Multiple Uncertain Nonlinear Systems The adaptive fuzzy containment control problem is addressed for multiple uncertain nonlinear strict-feedback systems with immeasurable states and multiple leaders under directed communication graphs. By utilizing fuzzy logic systems to model the followers’ dynamics, a distributed fuzzy state observer is designed for the state estimation using only the relative position information. Then, an observer-based containment control scheme is constructed by the adaptive fuzzy control technique as well as the command filter. The filtering error loop is introduced to compensate the error arising from the command filter. The proposed adaptive fuzzy containment control scheme guarantees that all followers are driven into the dynamic convex hull spanned by the leaders with a bounded containment error, if there exists at least one of the leaders who has a directed path to the follower. Simulation results are given to illustrate the control performance of the proposed containment control method.
Distributed adaptive containment control of uncertain nonlinear multi-agent systems in strict-feedback form. This paper presents a distributed containment control approach for uncertain nonlinear strict-feedback systems with multiple dynamic leaders under a directed graph topology where the leaders are neighbors of only a subset of the followers. The strict-feedback followers with nonparametric uncertainties are considered and the local adaptive dynamic surface controller for each follower is designed using only neighbors’ information to guarantee that all followers converge to the dynamic convex hull spanned by the dynamic leaders where the derivatives of leader signals are not available to implement controllers, i.e., the position information of leaders is only required. The function approximation technique using neural networks is employed to estimate nonlinear uncertainty terms derived from the controller design procedure for the followers. It is shown that the containment control errors converge to an adjustable neighborhood of the origin.
Wireless sensor network survey A wireless sensor network (WSN) has important applications such as remote environmental monitoring and target tracking. This has been enabled by the availability, particularly in recent years, of sensors that are smaller, cheaper, and intelligent. These sensors are equipped with wireless interfaces with which they can communicate with one another to form a network. The design of a WSN depends significantly on the application, and it must consider factors such as the environment, the application's design objectives, cost, hardware, and system constraints. The goal of our survey is to present a comprehensive review of the recent literature since the publication of [I.F. Akyildiz, W. Su, Y. Sankarasubramaniam, E. Cayirci, A survey on sensor networks, IEEE Communications Magazine, 2002]. Following a top-down approach, we give an overview of several new applications and then review the literature on various aspects of WSNs. We classify the problems into three different categories: (1) internal platform and underlying operating system, (2) communication protocol stack, and (3) network services, provisioning, and deployment. We review the major development in these three categories and outline new challenges.
Mobile Edge Computing: A Survey. Mobile edge computing (MEC) is an emergent architecture where cloud computing services are extended to the edge of networks leveraging mobile base stations. As a promising edge technology, it can be applied to mobile, wireless, and wireline scenarios, using software and hardware platforms, located at the network edge in the vicinity of end-users. MEC provides seamless integration of multiple appli...
Computer intrusion detection through EWMA for autocorrelated and uncorrelated data Reliability and quality of service from information systems has been threatened by cyber intrusions. To protect information systems from intrusions and thus assure reliability and quality of service, it is highly desirable to develop techniques that detect intrusions. Many intrusions manifest in anomalous changes in intensity of events occurring in information systems. In this study, we apply, tes...
An evaluation of direct attacks using fake fingers generated from ISO templates This work reports a vulnerability evaluation of a highly competitive ISO matcher to direct attacks carried out with fake fingers generated from ISO templates. Experiments are carried out on a fingerprint database acquired in a real-life scenario and show that the evaluated system is highly vulnerable to the proposed attack scheme, granting access in over 75% of the attempts (for a high-security operating point). Thus, the study disproves the popular belief of minutiae templates non-reversibility and raises a key vulnerability issue in the use of non-encrypted standard templates. (This article is an extended version of Galbally et al., 2008, which was awarded with the IBM Best Student Paper Award in the track of Biometrics at ICPR 2008).
Collaborative Mobile Charging The limited battery capacity of sensor nodes has become one of the most critical impediments that stunt the deployment of wireless sensor networks (WSNs). Recent breakthroughs in wireless energy transfer and rechargeable lithium batteries provide a promising alternative to power WSNs: mobile vehicles/robots carrying high volume batteries serve as mobile chargers to periodically deliver energy to sensor nodes. In this paper, we consider how to schedule multiple mobile chargers to optimize energy usage effectiveness, such that every sensor will not run out of energy. We introduce a novel charging paradigm, collaborative mobile charging, where mobile chargers are allowed to intentionally transfer energy between themselves. To provide some intuitive insights into the problem structure, we first consider a scenario that satisfies three conditions, and propose a scheduling algorithm, PushWait, which is proven to be optimal and can cover a one-dimensional WSN of infinite length. Then, we remove the conditions one by one, investigating chargers' scheduling in a series of scenarios ranging from the most restricted one to a general 2D WSN. Through theoretical analysis and simulations, we demonstrate the advantages of the proposed algorithms in energy usage effectiveness and charging coverage.
Distributed Kalman consensus filter with event-triggered communication: Formulation and stability analysis. •The problem of distributed state estimation in sensor networks with event-triggered communication schedules on both sensor-to-estimator channel and estimator-to-estimator channel is studied.•An event-triggered KCF is designed by deriving the optimal Kalman gain matrix which minimizes the mean squared error.•A computational scalable form of the proposed filter is presented by some approximations.•An appropriate choice of the consensus gain matrix is provided to ensure the stochastic stability of the proposed filter.
Hardware Circuits Design and Performance Evaluation of a Soft Lower Limb Exoskeleton Soft lower limb exoskeletons (LLEs) are wearable devices that have good potential in walking rehabilitation and augmentation. While a few studies focused on the structure design and assistance force optimization of the soft LLEs, rarely work has been conducted on the hardware circuits design. The main purpose of this work is to present a new soft LLE for walking efficiency improvement and introduce its hardware circuits design. A soft LLE for hip flexion assistance and a hardware circuits system with scalability were proposed. To assess the efficacy of the soft LLE, the experimental tests that evaluate the sensor data acquisition, force tracking performance, lower limb muscle activity and metabolic cost were conducted. The time error in the peak assistance force was just 1%. The reduction in the normalized root-mean-square EMG of the rectus femoris was 7.1%. The net metabolic cost in exoskeleton on condition was reduced by 7.8% relative to walking with no exoskeleton. The results show that the designed hardware circuits can be applied to the soft LLE and the soft LLE is able to improve walking efficiency of wearers.
1.2
0.2
0.2
0.2
0.2
0.2
0.028571
0
0
0
0
0
0
0
Adaptive fixed-time bipartite tracking consensus control for unknown nonlinear multi-agent systems: An information classification mechanism. This paper is concerned with the problem of bipartite tracking consensus for high-order unknown nonlinear multi-agent systems with actuator faults. Unlike the traditional condition that the directed signed graph is structurally balanced, a directed signed graph containing a spanning tree is considered. Besides, the consensus errors are required to satisfy both the prescribed performance and fast convergence (fixed-time). By proposing an information classification mechanism, each agent selectively uses neighbor information such that agents in the system are divided into two styles, which transform the bipartite tracking consensus problem into a general tracking consensus problem. By using neural networks and adaptive technologies to approximate unknown functions, the adaptive fault-tolerant fixed-time consensus controllers are developed. All signals in the system are bounded within a fixed time. Moreover, the bipartite consensus errors satisfy the prescribed performance by selecting appropriately predefined performance functions. Stability analysis and simulation results further verify the effectiveness of the proposed method.
Distributed Containment Control for Multiple Unknown Second-Order Nonlinear Systems With Application to Networked Lagrangian Systems. In this paper, we consider the distributed containment control problem for multiagent systems with unknown nonlinear dynamics. More specifically, we focus on multiple second-order nonlinear systems and networked Lagrangian systems. We first study the distributed containment control problem for multiple second-order nonlinear systems with multiple dynamic leaders in the presence of unknown nonlinearities and external disturbances under a general directed graph that characterizes the interaction among the leaders and the followers. A distributed adaptive control algorithm with an adaptive gain design based on the approximation capability of neural networks is proposed. We present a necessary and sufficient condition on the directed graph such that the containment error can be reduced as small as desired. As a byproduct, the leaderless consensus problem is solved with asymptotical convergence. Because relative velocity measurements between neighbors are generally more difficult to obtain than relative position measurements, we then propose a distributed containment control algorithm without using neighbors' velocity information. A two-step Lyapunov-based method is used to study the convergence of the closed-loop system. Next, we apply the ideas to deal with the containment control problem for networked unknown Lagrangian systems under a general directed graph. All the proposed algorithms are distributed and can be implemented using only local measurements in the absence of communication. Finally, simulation examples are provided to show the effectiveness of the proposed control algorithms.
Fully distributed containment control of high-order multi-agent systems with nonlinear dynamics. In this paper, distributed containment control problems for high-order multi-agent systems with nonlinear dynamics are investigated under directed communication topology. The states of the leaders are only available to a subset of the followers and the inputs of the leaders are possibly nonzero and time varying. Distributed adaptive nonlinear protocol is proposed based only on the relative state information, under which the states of the followers converge to the dynamic convex hull spanned by those of the leaders. As the special case with only one dynamic leader, leader–follower consensus problem is also solved with the proposed protocol. The adaptive protocol here is independent of the eigenvalues of the Laplacian matrix, which means the protocol can be implemented by each agent in a fully distributed fashion. A simulation example is provided to illustrate the theoretical results.
Output Containment Control of Linear Heterogeneous Multi-Agent Systems Using Internal Model Principle. This paper studies the output containment control of linear heterogeneous multi-agent systems, where the system dynamics and even the state dimensions can generally be different. Since the states can have different dimensions, standard results from state containment control do not apply. Therefore, the control objective is to guarantee the convergence of the output of each follower to the dynamic ...
Finite-Time Consensus Tracking Neural Network FTC of Multi-Agent Systems The finite-time consensus fault-tolerant control (FTC) tracking problem is studied for the nonlinear multi-agent systems (MASs) in the nonstrict feedback form. The MASs are subject to unknown symmetric output dead zones, actuator bias and gain faults, and unknown control coefficients. According to the properties of the neural network (NN), the unstructured uncertainties problem is solved. The Nussbaum function is used to address the output dead zones and unknown control directions problems. By introducing an arbitrarily small positive number, the “singularity” problem caused by combining the finite-time control and backstepping design is solved. According to the backstepping design and Lyapunov stability theory, a finite-time adaptive NN FTC controller is obtained, which guarantees that the tracking error converges to a small neighborhood of zero in a finite time, and all signals in the closed-loop system are bounded. Finally, the effectiveness of the proposed method is illustrated via a physical example.
Observer-based Adaptive Fuzzy Containment Control for Multiple Uncertain Nonlinear Systems The adaptive fuzzy containment control problem is addressed for multiple uncertain nonlinear strict-feedback systems with immeasurable states and multiple leaders under directed communication graphs. By utilizing fuzzy logic systems to model the followers’ dynamics, a distributed fuzzy state observer is designed for the state estimation using only the relative position information. Then, an observer-based containment control scheme is constructed by the adaptive fuzzy control technique as well as the command filter. The filtering error loop is introduced to compensate the error arising from the command filter. The proposed adaptive fuzzy containment control scheme guarantees that all followers are driven into the dynamic convex hull spanned by the leaders with a bounded containment error, if there exists at least one of the leaders who has a directed path to the follower. Simulation results are given to illustrate the control performance of the proposed containment control method.
Harmony search algorithm for solving Sudoku Harmony search (HS) algorithm was applied to solving Sudoku puzzle. The HS is an evolutionary algorithm which mimics musicians' behaviors such as random play, memory-based play, and pitch-adjusted play when they perform improvisation. Sudoku puzzles in this study were formulated as an optimization problem with number-uniqueness penalties. HS could successfully solve the optimization problem after 285 function evaluations, taking 9 seconds. Also, sensitivity analysis of HS parameters was performed to obtain a better idea of algorithm parameter values.
Containment Control in Mobile Networks In this paper, the problem of driving a collection of mobile robots to a given target destination is studied. In particular, we are interested in achieving this transfer in an orderly manner so as to ensure that the agents remain in the convex polytope spanned by the leader-agents, while the remaining agents, only employ local interaction rules. To this aim we exploit the theory of partial difference equations and propose hybrid control schemes based on stop-go rules for the leader-agents. Non-Zenoness, liveness and convergence of the resulting system are also analyzed.
Probabilistic encryption A new probabilistic model of data encryption is introduced. For this model, under suitable complexity assumptions, it is proved that extracting any information about the cleartext from the cyphertext is hard on the average for an adversary with polynomially bounded computational resources. The proof holds for any message space with any probability distribution. The first implementation of this model is presented. The security of this implementation is proved under the interactability assumptin of deciding Quadratic Residuosity modulo composite numbers whose factorization is unknown.
Non-interactive verifiable computing: outsourcing computation to untrusted workers We introduce and formalize the notion of Verifiable Computation, which enables a computationally weak client to "outsource" the computation of a function F on various dynamically-chosen inputs x1, ...,xk to one or more workers. The workers return the result of the function evaluation, e.g., yi = F(xi), as well as a proof that the computation of F was carried out correctly on the given value xi. The primary constraint is that the verification of the proof should require substantially less computational effort than computing F(i) from scratch. We present a protocol that allows the worker to return a computationally-sound, non-interactive proof that can be verified in O(mċpoly(λ)) time, where m is the bit-length of the output of F, and λ is a security parameter. The protocol requires a one-time pre-processing stage by the client which takes O(|C|ċpoly(λ)) time, where C is the smallest known Boolean circuit computing F. Unlike previous work in this area, our scheme also provides (at no additional cost) input and output privacy for the client, meaning that the workers do not learn any information about the xi or yi values.
A web-based e-learning system for increasing study efficiency by stimulating learner's motivation Due to the opportunities provided by the Internet, more and more people are taking advantage of distance learning courses and during the last few years enormous research efforts have been dedicated to the development of distance learning systems. So far, many e-learning systems are proposed and used practically. However, in these systems the e-learning completion rate is about 30%. One of the reasons is the low study desire when the learner studies the learning materials. In this research, we propose an interactive Web-based e-learning system. The purpose of our system is to increase the e-learning completion rate by stimulating learner's motivation. The proposed system has three subsystems: the learning subsystem, learner support subsystem, and teacher support subsystem. The learning subsystem improves the learner's study desire. The learner support subsystem supports the learner during the study, and the teacher support subsystem supports the teacher to get the learner's study state. To evaluate the proposed system, we developed several experiments and surveys. By using new features such as: display of learner's study history, change of interface color, encourage function, ranking function, self-determination of the study materials, and grouping of learners, the proposed system can increase the learning efficiency.
Robust DT-CWT Watermarking for DIBR 3D Images The popularity of 3D content is on the rise since it provides an immersive experience to viewers. In this situation, depth-image-based rendering (DIBR) has taken on an important role in 3D technology due to its low bandwidth cost and ease of depth configuration. Noting that the viewer could record provided center view or synthesized views for illegal distribution, it is clear that copyright protection must be taken into account for the DIBR 3D content, including the possibility that one single view could be illegally distributed as 2D content. In this paper, we propose a robust watermarking scheme for DIBR 3D images by quantization on dual-tree complex wavelet transform (DT-CWT) coefficients with consideration of imperceptibility. To make the proposed scheme robust to DIBR process, two characteristics of DT-CWT are employed: approximate shift invariance and directional selectivity. We select certain coefficient sub-blocks and group the coefficient rows based on the properties of DIBR. On the extraction side, the threshold is carefully chosen with a low false positive rate. The simulation results show that the embedded watermark is stably extracted from the center view and the synthesized left and right views. In addition, even if the synthesized left and right views are distorted by general attacks, the watermark is successfully extracted. Furthermore, the proposed scheme is robust to pre-processing of the depth image and baseline adjusting, which are common processing on the DIBR system for better quality of 3D views.
Adaptive generation of challenging scenarios for testing and evaluation of autonomous vehicles. •A novel framework for generating test cases for autonomous vehicles is proposed.•Adaptive sampling significantly reduces the number of simulations required.•Adjacency clustering identifies performance boundaries of the system.•Approach successfully applied to complex unmanned underwater vehicle missions.
Learning Feature Recovery Transformer for Occluded Person Re-Identification One major issue that challenges person re-identification (Re-ID) is the ubiquitous occlusion over the captured persons. There are two main challenges for the occluded person Re-ID problem, i.e., the interference of noise during feature matching and the loss of pedestrian information brought by the occlusions. In this paper, we propose a new approach called Feature Recovery Transformer (FRT) to address the two challenges simultaneously, which mainly consists of visibility graph matching and feature recovery transformer. To reduce the interference of the noise during feature matching, we mainly focus on visible regions that appear in both images and develop a visibility graph to calculate the similarity. In terms of the second challenge, based on the developed graph similarity, for each query image, we propose a recovery transformer that exploits the feature sets of its k-nearest neighbors in the gallery to recover the complete features. Extensive experiments across different person Re-ID datasets, including occluded, partial and holistic datasets, demonstrate the effectiveness of FRT. Specifically, FRT significantly outperforms state-of-the-art results by at least 6.2% Rank- 1 accuracy and 7.2% mAP scores on the challenging Occluded-Duke dataset.
1.084444
0.066667
0.066667
0.066667
0.066667
0.066667
0.026667
0.013605
0
0
0
0
0
0
User-Centered Cooperative-Communication Strategy for 5G Internet of Vehicles The Internet of vehicles currently faces the challenges of spectrum interference in multimode, multiservice communications, and low efficiency in switched-vehicle selection for vehicular cooperative communication. To overcome these, a user-centered cooperative-communication strategy (UCCS) is proposed. A system model of the UCCS algorithm is established for three situations according to the vehicular cooperative-communication process. When there are service roadside units (SRSUs), cooperative RSUs (CRSUs), and cooperative vehicles (CVs) in the service vehicle (SV) communication range, a two-layer game method is proposed, which dynamically adjusts the transmission power, punishment factor, and other parameters of service and cooperative vehicle. When there are only CVs in the SV communication range, a multiparameter voting method is proposed, which selects the best cooperative-communication vehicle by calculating the normalized linear weighted value of CVs. When there are only CRSUs and CVs in the SV communication range, a station-vehicle voting method is proposed, which calculates the cooperative-communication value of CRSUs to select the optimal one, thus comparing it with the best CV to determine the optimal cooperative-communication object. The simulation results show that the average throughput of the UCCS algorithm is increased by 17% and 42% as compared to those based on the repeated game and network utility maximization algorithms, respectively, and its average waiting time decreased by 43% and 24% compared to those of the random selection and shortest path selection algorithms, respectively. The overall system revenue and communication performance of the proposed algorithm are significantly improved.
The Sybil Attack Large-scale peer-to-peer systems facesecurity threats from faulty or hostile remotecomputing elements. To resist these threats, manysuch systems employ redundancy. However, if asingle faulty entity can present multiple identities,it can control a substantial fraction of the system,thereby undermining this redundancy. Oneapproach to preventing these &quot;Sybil attacks&quot; is tohave a trusted agency certify identities. Thispaper shows that, without a logically centralizedauthority, Sybil...
BLEU: a method for automatic evaluation of machine translation Human evaluations of machine translation are extensive but expensive. Human evaluations can take months to finish and involve human labor that can not be reused. We propose a method of automatic machine translation evaluation that is quick, inexpensive, and language-independent, that correlates highly with human evaluation, and that has little marginal cost per run. We present this method as an automated understudy to skilled human judges which substitutes for them when there is need for quick or frequent evaluations.
Computational thinking Summary form only given. My vision for the 21st century, Computational Thinking, will be a fundamental skill used by everyone in the world. To reading, writing, and arithmetic, we should add computational thinking to every child's analytical ability. Computational thinking involves solving problems, designing systems, and understanding human behavior by drawing on the concepts fundamental to computer science. Thinking like a computer scientist means more than being able to program a computer. It requires the ability to abstract and thus to think at multiple levels of abstraction. In this talk I will give many examples of computational thinking, argue that it has already influenced other disciplines, and promote the idea that teaching computational thinking can not only inspire future generations to enter the field of computer science but benefit people in all fields.
Fuzzy logic in control systems: fuzzy logic controller. I.
Switching between stabilizing controllers This paper deals with the problem of switching between several linear time-invariant (LTI) controllers—all of them capable of stabilizing a speci4c LTI process—in such a way that the stability of the closed-loop system is guaranteed for any switching sequence. We show that it is possible to 4nd realizations for any given family of controller transfer matrices so that the closed-loop system remains stable, no matter how we switch among the controller. The motivation for this problem is the control of complex systems where con8icting requirements make a single LTI controller unsuitable. ? 2002 Published by Elsevier Science Ltd.
Tabu Search - Part I
Bidirectional recurrent neural networks In the first part of this paper, a regular recurrent neural network (RNN) is extended to a bidirectional recurrent neural network (BRNN). The BRNN can be trained without the limitation of using input information just up to a preset future frame. This is accomplished by training it simultaneously in positive and negative time direction. Structure and training procedure of the proposed network are explained. In regression and classification experiments on artificial data, the proposed structure gives better results than other approaches. For real data, classification experiments for phonemes from the TIMIT database show the same tendency. In the second part of this paper, it is shown how the proposed bidirectional structure can be easily modified to allow efficient estimation of the conditional posterior probability of complete symbol sequences without making any explicit assumption about the shape of the distribution. For this part, experiments on real data are reported
An intensive survey of fair non-repudiation protocols With the phenomenal growth of the Internet and open networks in general, security services, such as non-repudiation, become crucial to many applications. Non-repudiation services must ensure that when Alice sends some information to Bob over a network, neither Alice nor Bob can deny having participated in a part or the whole of this communication. Therefore a fair non-repudiation protocol has to generate non-repudiation of origin evidences intended to Bob, and non-repudiation of receipt evidences destined to Alice. In this paper, we clearly define the properties a fair non-repudiation protocol must respect, and give a survey of the most important non-repudiation protocols without and with trusted third party (TTP). For the later ones we discuss the evolution of the TTP's involvement and, between others, describe the most recent protocol using a transparent TTP. We also discuss some ad-hoc problems related to the management of non-repudiation evidences.
Dynamic movement and positioning of embodied agents in multiparty conversations For embodied agents to engage in realistic multiparty conversation, they must stand in appropriate places with respect to other agents and the environment. When these factors change, such as an agent joining the conversation, the agents must dynamically move to a new location and/or orientation to accommodate. This paper presents an algorithm for simulating movement of agents based on observed human behavior using techniques developed for pedestrian movement in crowd simulations. We extend a previous group conversation simulation to include an agent motion algorithm. We examine several test cases and show how the simulation generates results that mirror real-life conversation settings.
An improved genetic algorithm with conditional genetic operators and its application to set-covering problem The genetic algorithm (GA) is a popular, biologically inspired optimization method. However, in the GA there is no rule of thumb to design the GA operators and select GA parameters. Instead, trial-and-error has to be applied. In this paper we present an improved genetic algorithm in which crossover and mutation are performed conditionally instead of probability. Because there are no crossover rate and mutation rate to be selected, the proposed improved GA can be more easily applied to a problem than the conventional genetic algorithms. The proposed improved genetic algorithm is applied to solve the set-covering problem. Experimental studies show that the improved GA produces better results over the conventional one and other methods.
Lane-level traffic estimations using microscopic traffic variables This paper proposes a novel inference method to estimate lane-level traffic flow, time occupancy and vehicle inter-arrival time on road segments where local information could not be measured and assessed directly. The main contributions of the proposed method are 1) the ability to perform lane-level estimations of traffic flow, time occupancy and vehicle inter-arrival time and 2) the ability to adapt to different traffic regimes by assessing only microscopic traffic variables. We propose a modified Kriging estimation model which explicitly takes into account both spatial and temporal variability. Performance evaluations are conducted using real-world data under different traffic regimes and it is shown that the proposed method outperforms a Kalman filter-based approach.
Convolutional Neural Network-Based Classification of Driver's Emotion during Aggressive and Smooth Driving Using Multi-Modal Camera Sensors. Because aggressive driving often causes large-scale loss of life and property, techniques for advance detection of adverse driver emotional states have become important for the prevention of aggressive driving behaviors. Previous studies have primarily focused on systems for detecting aggressive driver emotion via smart-phone accelerometers and gyro-sensors, or they focused on methods of detecting physiological signals using electroencephalography (EEG) or electrocardiogram (ECG) sensors. Because EEG and ECG sensors cause discomfort to drivers and can be detached from the driver's body, it becomes difficult to focus on bio-signals to determine their emotional state. Gyro-sensors and accelerometers depend on the performance of GPS receivers and cannot be used in areas where GPS signals are blocked. Moreover, if driving on a mountain road with many quick turns, a driver's emotional state can easily be misrecognized as that of an aggressive driver. To resolve these problems, we propose a convolutional neural network (CNN)-based method of detecting emotion to identify aggressive driving using input images of the driver's face, obtained using near-infrared (NIR) light and thermal camera sensors. In this research, we conducted an experiment using our own database, which provides a high classification accuracy for detecting driver emotion leading to either aggressive or smooth (i.e., relaxed) driving. Our proposed method demonstrates better performance than existing methods.
Ethical Considerations Of Applying Robots In Kindergarten Settings: Towards An Approach From A Macroperspective In child-robot interaction (cHRI) research, many studies pursue the goal to develop interactive systems that can be applied in everyday settings. For early education, increasingly, the setting of a kindergarten is targeted. However, when cHRI and research are brought into a kindergarten, a range of ethical and related procedural aspects have to be considered and dealt with. While ethical models elaborated within other human-robot interaction settings, e.g., assisted living contexts, can provide some important indicators for relevant issues, we argue that it is important to start developing a systematic approach to identify and tackle those ethical issues which rise with cHRI in kindergarten settings on a more global level and address the impact of the technology from a macroperspective beyond the effects on the individual. Based on our experience in conducting studies with children in general and pedagogical considerations on the role of the institution of kindergarten in specific, in this paper, we enfold some relevant aspects that have barely been addressed in an explicit way in current cHRI research. Four areas are analyzed and key ethical issues are identified in each area: (1) the institutional setting of a kindergarten, (2) children as a vulnerable group, (3) the caregivers' role, and (4) pedagogical concepts. With our considerations, we aim at (i) broadening the methodology of the current studies within the area of cHRI, (ii) revalidate it based on our comprehensive empirical experience with research in kindergarten settings, both laboratory and real-world contexts, and (iii) provide a framework for the development of a more systematic approach to address the ethical issues in cHRI research within kindergarten settings.
1
0
0
0
0
0
0
0
0
0
0
0
0
0
Deep High-Resolution Representation Learning For Human Pose Estimation In this paper, we are interested in the human pose estimation problem with a focus on learning reliable high-resolution representations. Most existing methods recover high-resolution representations from low-resolution representations produced by a high-to-low resolution network. Instead, our proposed network maintains high-resolution representations through the whole process.We start from a high-resolution subnetwork as the first stage, gradually add high-to-low resolution subnetworks one by one to form more stages, and connect the mutli-resolution subnetworks in parallel. We conduct repeated multi-scale fusions such that each of the high-to-low resolution representations receives information from other parallel representations over and over, leading to rich high-resolution representations. As a result, the predicted keypoint heatmap is potentially more accurate and spatially more precise. We empirically demonstrate the effectiveness of our network through the superior pose estimation results over two benchmark datasets: the COCO keypoint detection dataset and the MPII Human Pose dataset. In addition, we show the superiority of our network in pose tracking on the PoseTrack dataset.
Deep Feature Learning via Structured Graph Laplacian Embedding for Person Re-Identification. •This paper is the first to formulates the structured distance relationships into the graph Laplacian form for deep feature learning.•Joint learning method is used in the framework to learn discriminative features.•The results show clear improvements on public benchmark datasets and some are the state-of-the-art.
Online Joint Multi-Metric Adaptation From Frequent Sharing-Subset Mining For Person Re-Identification Person Re-IDentification (P-RID), as an instance-level recognition problem, still remains challenging in computer vision community. Many P-RID works aim to learn faithful and discriminative features/metrics from offline training data and directly use them for the unseen online testing data. However, their performance is largely limited due to the severe data shifting issue between training and testing data. Therefore, we propose an online joint multi-metric adaptation model to adapt the offline learned P-RID models for the online data by learning a series of metrics for all the sharing-subsets. Each sharing-subset is obtained from the proposed novel frequent sharing-subset mining module and contains a group of testing samples which share strong visual similarity relationships to each other. Unlike existing online P-RID methods, our model simultaneously takes both the sample-specific discriminant and the set-based visual similarity among testing samples into consideration so that the adapted multiple metrics can refine the discriminant of all the given testing samples jointly via a multi-kernel late fusion framework. Our proposed model is generally suitable to any offline learned P-RID baselines for online boosting, the performance improvement by our model is not only verified by extensive experiments on several widely-used P-RID benchmarks (CUHK03, Market 1501, DukeMTMC-reID and MSMTI7) and state-of-the-art P-RID baselines but also guaranteed by the provided in-depth theoretical analyses.
Pose-Guided Visible Part Matching for Occluded Person ReID Occluded person re-identification is a challenging task as the appearance varies substantially with various obstacles, especially in the crowd scenario. To address this issue, we propose a Pose-guided Visible Part Matching (PVPM) method that jointly learns the discriminative features with pose-guided attention and self-mines the part visibility in an end-to-end framework. Specifically, the proposed PVPM includes two key components: 1) pose-guided attention (PGA) method for part feature pooling that exploits more discriminative local features; 2) pose-guided visibility predictor (PVP) that estimates whether a part suffers the occlusion or not. As there are no ground truth training annotations for the occluded part, we turn to utilize the characteristic of part correspondence in positive pairs and self-mining the correspondence scores via graph matching. The generated correspondence scores are then utilized as pseudo-labels for visibility predictor (PVP). Experimental results on three reported occluded benchmarks show that the proposed method achieves competitive performance to state-of-the-art methods. The source codes are available at https://github.com/hh23333/PVPM
Horizontal Pyramid Matching for Person Re-Identification Despite the remarkable progress in person re-identification (Re-ID), such approaches still suffer from the failure cases where the discriminative body parts are missing. To mitigate this type of failure, we propose a simple yet effective Horizontal Pyramid Matching (HPM) approach to fully exploit various partial information of a given person, so that correct person candidates can be identified even if some key parts are missing. With HPM, we make the following contributions to produce more robust feature representations for the Re-ID task: 1) we learn to classify using partial feature representations at different horizontal pyramid scales, which successfully enhance the discriminative capabilities of various person parts; 2) we exploit average and max pooling strategies to account for person-specific discriminative information in a global-local manner. To validate the effectiveness of our proposed HPM method, extensive experiments are conducted on three popular datasets including Market-1501, DukeMTMC-ReID and CUHK03. Respectively, we achieve mAP scores of 83.1%, 74.5% and 59.7% on these challenging benchmarks, which are the new state-of-the-arts.
Learning Context Graph For Person Search Person re-identification has achieved great progress with deep convolutional neural networks. However, most previous methods focus on learning individual appearance feature embedding, and it is hard for the models to handle difficult situations with different illumination, large pose variance and occlusion. In this work, we take a step further and consider employing context information for person search. For a probe-gallery pair, we first propose a contextual instance expansion module, which employs a relative attention module to search and filter useful context information in the scene. We also build a graph learning framework to effectively employ context pairs to update target similarity. These two modules are built on top of a joint detection and instance feature learning framework, which improves the discriminativeness of the learned features. The proposed framework achieves state-of-the-art performance on two widely used person search datasets.
Locally Aligned Feature Transforms across Views In this paper, we propose a new approach for matching images observed in different camera views with complex cross-view transforms and apply it to person re-identification. It jointly partitions the image spaces of two camera views into different configurations according to the similarity of cross-view transforms. The visual features of an image pair from different views are first locally aligned by being projected to a common feature space and then matched with softly assigned metrics which are locally optimized. The features optimal for recognizing identities are different from those for clustering cross-view transforms. They are jointly learned by utilizing sparsity-inducing norm and information theoretical regularization. This approach can be generalized to the settings where test images are from new camera views, not the same as those in the training set. Extensive experiments are conducted on public datasets and our own dataset. Comparisons with the state-of-the-art metric learning and person re-identification methods show the superior performance of our approach.
SPENCER: A Socially Aware Service Robot for Passenger Guidance and Help in Busy Airports. We present an ample description of a socially compliant mobile robotic platform, which is developed in the EU-funded project SPENCER. The purpose of this robot is to assist, inform and guide passengers in large and busy airports. One particular aim is to bring travellers of connecting flights conveniently and efficiently from their arrival gate to the passport control. The uniqueness of the project stems from the strong demand of service robots for this application with a large potential impact for the aviation industry on one side, and on the other side from the scientific advancements in social robotics, brought forward and achieved in SPENCER. The main contributions of SPENCER are novel methods to perceive, learn, and model human social behavior and to use this knowledge to plan appropriate actions in real-time for mobile platforms. In this paper, we describe how the project advances the fields of detection and tracking of individuals and groups, recognition of human social relations and activities, normative human behavior learning, socially-aware task and motion planning, learning socially annotated maps, and conducting empirical experiments to assess socio-psychological effects of normative robot behaviors.
New approach using ant colony optimization with ant set partition for fuzzy control design applied to the ball and beam system. In this paper we describe the design of a fuzzy logic controller for the ball and beam system using a modified Ant Colony Optimization (ACO) method for optimizing the type of membership functions, the parameters of the membership functions and the fuzzy rules. This is achieved by applying a systematic and hierarchical optimization approach modifying the conventional ACO algorithm using an ant set partition strategy. The simulation results show that the proposed algorithm achieves better results than the classical ACO algorithm for the design of the fuzzy controller.
EDUCO - A Collaborative Learning Environment Based on Social Navigation Web-based learning is primarily a lonesome activity, even when it involves working in groups. This is due to the fact that the majority of web-based learning relies on asynchronous forms of interacting with other people. In most of the cases, the chat discussion is the only form of synchronous interaction that adds to the feeling that there are other people present in the environment. EDUCO is a system that tries to bring in the sense of other users in a collaborative learning environment by making the other users and their the navigation visible to everyone else in the environment in real-time. The paper describes EDUCO and presents the first empirical evaluation as EDUCO was used in a university course.
A dual neural network for redundancy resolution of kinematically redundant manipulators subject to joint limits and joint velocity limits. In this paper, a recurrent neural network called the dual neural network is proposed for online redundancy resolution of kinematically redundant manipulators. Physical constraints such as joint limits and joint velocity limits, together with the drift-free criterion as a secondary task, are incorporated into the problem formulation of redundancy resolution. Compared to other recurrent neural networks, the dual neural network is piecewise linear and has much simpler architecture with only one layer of neurons. The dual neural network is shown to be globally (exponentially) convergent to optimal solutions. The dual neural network is simulated to control the PA10 robot manipulator with effectiveness demonstrated.
Software-Defined Networking: A Comprehensive Survey The Internet has led to the creation of a digital society, where (almost) everything is connected and is accessible from anywhere. However, despite their widespread adoption, traditional IP networks are complex and very hard to manage. It is both difficult to configure the network according to predefined policies, and to reconfigure it to respond to faults, load, and changes. To make matters even more difficult, current networks are also vertically integrated: the control and data planes are bundled together. Software-defined networking (SDN) is an emerging paradigm that promises to change this state of affairs, by breaking vertical integration, separating the network's control logic from the underlying routers and switches, promoting (logical) centralization of network control, and introducing the ability to program the network. The separation of concerns, introduced between the definition of network policies, their implementation in switching hardware, and the forwarding of traffic, is key to the desired flexibility: by breaking the network control problem into tractable pieces, SDN makes it easier to create and introduce new abstractions in networking, simplifying network management and facilitating network evolution. In this paper, we present a comprehensive survey on SDN. We start by introducing the motivation for SDN, explain its main concepts and how it differs from traditional networking, its roots, and the standardization activities regarding this novel paradigm. Next, we present the key building blocks of an SDN infrastructure using a bottom-up, layered approach. We provide an in-depth analysis of the hardware infrastructure, southbound and northbound application programming interfaces (APIs), network virtualization layers, network operating systems (SDN controllers), network programming languages, and network applications. We also look at cross-layer problems such as debugging and troubleshooting. In an effort to anticipate the future evolution of this - ew paradigm, we discuss the main ongoing research efforts and challenges of SDN. In particular, we address the design of switches and control platforms - with a focus on aspects such as resiliency, scalability, performance, security, and dependability - as well as new opportunities for carrier transport networks and cloud providers. Last but not least, we analyze the position of SDN as a key enabler of a software-defined environment.
Safe mutations for deep and recurrent neural networks through output gradients While neuroevolution (evolving neural networks) has been successful across a variety of domains from reinforcement learning, to artificial life, to evolutionary robotics, it is rarely applied to large, deep neural networks. A central reason is that while random mutation generally works in low dimensions, a random perturbation of thousands or millions of weights will likely break existing functionality. This paper proposes a solution: a family of safe mutation (SM) operators that facilitate exploration without dramatically altering network behavior or requiring additional interaction with the environment. The most effective SM variant scales the degree of mutation of each individual weight according to the sensitivity of the network's outputs to that weight, which requires computing the gradient of outputs with respect to the weights (instead of the gradient of error, as in conventional deep learning). This safe mutation through gradients (SM-G) operator dramatically increases the ability of a simple genetic algorithm-based neuroevolution method to find solutions in high-dimensional domains that require deep and/or recurrent neural networks, including domains that require processing raw pixels. By improving our ability to evolve deep neural networks, this new safer approach to mutation expands the scope of domains amenable to neuroevolution.
Robot tutor and pupils’ educational ability: Teaching the times tables Research shows promising results of educational robots in language and STEM tasks. In language, more research is available, occasionally in view of individual differences in pupils’ educational ability levels, and learning seems to improve with more expressive robot behaviors. In STEM, variations in robots’ behaviors have been examined with inconclusive results and never while systematically investigating how differences in educational abilities match with different robot behaviors. We applied an autonomously tutoring robot (without tablet, partly WOz) in a 2 × 2 experiment of social vs. neutral behavior in above-average vs. below-average schoolchildren (N = 86; age 8–10 years) while rehearsing the multiplication tables on a one-to-one basis. The standard school test showed that on average, pupils significantly improved their performance even after 3 occasions of 5-min exercises. Beyond-average pupils profited most from a robot tutor, whereas those below average in multiplication benefited more from a robot that showed neutral rather than more social behavior.
1.1055
0.1
0.1
0.1
0.056125
0.05
0.001686
0.000073
0
0
0
0
0
0
Restoration Of Stance Phase Knee Flexion During Walking After Spinal Cord Injury Using A Variable Impedance Orthosis A hybrid neuroprosthesis (HNP) combines lower extremity bracing with functional neuromuscular stimulation (FNS) to restore walking function and enhance the efficiency of ambulation. This report details the development of a novel HNP containing a variable impedance knee mechanism (VIKM) capable of supporting the knee against collapse while allowing controlled stance phase knee flexion. The design of a closed loop, finite state controller for coordination of VIKM activity with FNS-driven gait is presented. The controller is verified in testing during able bodied gait. The improved functionality provided by this system has the potential to delay the onset of fatigue and to expand FNS driven gait to allow walking over uneven terrains and down stairs.
Finite State Control of FES-Assisted Walking with Spring Brake Orthosis This paper presents finite state control (FSC) of paraplegic walking with wheel walker using functional electrical stimulation (FES) with spring brake orthosis (SBO). The work is a first effort towards restoring natural like swing phase in paraplegic gait through a new hybrid orthosis, referred to as spring brake orthosis (SBO). This mechanism simplifies the control task and results in smooth motion and more-natural like trajectory produced by the flexion reflex for gait in spinal cord injured subjects. The study is carried out with a model of humanoid with wheel walker using the Visual Nastran (Vn4D) dynamic simulation software. Stimulated muscle model of quadriceps is developed for knee extension. Fuzzy logic control (FLC) is developed in Matlab/Simulink to regulate the muscle stimulation pulse-width required to drive FES-assisted walking gait and the computed motion is visualised in graphic animation from Vn4D and finite state control is used to control the transaction between all walking states. Finite state control (FSC) is used to control the switching of brakes, FES and spring during walking cycle.
An FES-assisted training strategy combined with impedance control for a lower limb rehabilitation robot In order to investigate the feasibility of integrating functional electrical stimulation (FES) with robot-based rehabilitation training, this paper proposes an FES-assisted training strategy combined with impedance control for our self-made exoskeleton lower limb rehabilitation robot. This control strategy is carried out in a leg press task. Through impedance control, an active compliance of the robot is established, and the patient's voluntary effort to accomplish the task is inspired. During the training process, the patient's related muscles are applied with FES which provides an extra assistance to the patient. The intensity of the FES is properly chosen aiming to induce a desired active torque which is proportional to the voluntary effort of the patient. This kind of enhancement serves as a positive feedback which reminds the patient of the correct attempt to fulfill the desired motion. FES control is conducted by a combination of neural network-based feedforward controller and a PD feedback controller. The feasibility of this control strategy has been verified in Matlab.
A semi-active hybrid neuroprosthesis for restoring lower limb function in paraplegics. Through the application of functional electrical stimulation (FES) individuals with paraplegia can regain lost walking function. However, due to the rapid onset of muscle fatigue, the walking duration obtained with an FES-based neuroprosthesis is often relatively short. The rapid muscle fatigue can be compensated for by using a hybrid system that uses both FES and an active orthosis. In this paper, we demonstrate the initial testing of a semi-active hybrid walking neuroprosthesis. The semi-active hybrid orthosis (SEAHO) supports a user during the stance phase and standing while the electric motors attached to the hip section of the orthosis are used to generate hip flexion/extension. FES in SEAHO is mainly used to actuate knee flexion/extension and plantar flexion of the foot. SEAHO is controlled by a finite state machine that uses a recently developed nonlinear controller for position tracking control of the hip motors and cues from the hip angle to actuate FES and other components.
A Passive Gait-Based Weight-Support Lower Extremity Exoskeleton With Compliant Joints. This paper presents the design and analysis of a passive body weight (BW)-support lower extremity exoskeleton (LEE) with compliant joints to relieve compressive load in the knee. The biojoint-like mechanical knee decouples human gait into two phases, stance and swing, by a dual snap fit. The knee joint transfers the BW to the ground in the stance phases and is compliant to free the leg in the swin...
Preliminary assessment of the efficacy of supplementing knee extension capability in a lower limb exoskeleton with FES. The authors describe a cooperative controller that combines the knee joint actuation of an externally powered lower limb exoskeleton with the torque and power contribution from the electrically stimulated quadriceps muscle group. The efficacy of combining these efforts is experimentally validated with a series of weighted leg lift maneuvers. Measurements from these experiments indicate that the control approach effectively combines the respective efforts of the motor and muscle, such that good control performance is achieved, with substantial torque and energy contributions from both the biological and non-biological actuators.
One hundred data-driven haptic texture models and open-source methods for rendering on 3D objects This paper introduces the Penn Haptic Texture Toolkit (HaTT), a publicly available repository of haptic texture models for use by the research community. HaTT includes 100 haptic texture and friction models, the recorded data from which the models were made, images of the textures, and the code and methods necessary to render these textures using an impedance-type haptic interface such as a SensAble Phantom Omni. This paper reviews our previously developed methods for modeling haptic virtual textures, describes our technique for modeling Coulomb friction between a tooltip and a surface, discusses the adaptation of our rendering methods for display using an impedance-type haptic device, and provides an overview of the information included in the toolkit. Each texture and friction model was based on a ten-second recording of the force, speed, and high-frequency acceleration experienced by a handheld tool moved by an experimenter against the surface in a natural manner. We modeled each texture's recorded acceleration signal as a piecewise autoregressive (AR) process and stored the individual AR models in a Delaunay triangulation as a function of the force and speed used when recording the data. To increase the adaptability and utility of HaTT, we developed a method for resampling the texture models so they can be rendered at a sampling rate other than the 10 kHz used when recording data. Measurements of the user's instantaneous normal force and tangential speed are used to synthesize texture vibrations in real time. These vibrations are transformed into a texture force vector that is added to the friction and normal force vectors for display to the user.
Robust Indoor Positioning Provided by Real-Time RSSI Values in Unmodified WLAN Networks The positioning methods based on received signal strength (RSS) measurements, link the RSS values to the position of the mobile station(MS) to be located. Their accuracy depends on the suitability of the propagation models used for the actual propagation conditions. In indoor wireless networks, these propagation conditions are very difficult to predict due to the unwieldy and dynamic nature of the RSS. In this paper, we present a novel method which dynamically estimates the propagation models that best fit the propagation environments, by using only RSS measurements obtained in real time. This method is based on maximizing compatibility of the MS to access points (AP) distance estimates. Once the propagation models are estimated in real time, it is possible to accurately determine the distance between the MS and each AP. By means of these distance estimates, the location of the MS can be obtained by trilateration. The method proposed coupled with simulations and measurements in a real indoor environment, demonstrates its feasibility and suitability, since it outperforms conventional RSS-based indoor location methods without using any radio map information nor a calibration stage.
Adaptive Learning-Based Task Offloading for Vehicular Edge Computing Systems. The vehicular edge computing system integrates the computing resources of vehicles, and provides computing services for other vehicles and pedestrians with task offloading. However, the vehicular task offloading environment is dynamic and uncertain, with fast varying network topologies, wireless channel states, and computing workloads. These uncertainties bring extra challenges to task offloading. In this paper, we consider the task offloading among vehicles, and propose a solution that enables vehicles to learn the offloading delay performance of their neighboring vehicles while offloading computation tasks. We design an adaptive learning based task offloading (ALTO) algorithm based on the multi-armed bandit theory, in order to minimize the average offloading delay. ALTO works in a distributed manner without requiring frequent state exchange, and is augmented with input-awareness and occurrence-awareness to adapt to the dynamic environment. The proposed algorithm is proved to have a sublinear learning regret. Extensive simulations are carried out under both synthetic scenario and realistic highway scenario, and results illustrate that the proposed algorithm achieves low delay performance, and decreases the average delay up to <inline-formula xmlns:mml="http://www.w3.org/1998/Math/MathML" xmlns:xlink="http://www.w3.org/1999/xlink"><tex-math notation="LaTeX">$30\%$</tex-math></inline-formula> compared with the existing upper confidence bound based learning algorithm.
Visual cryptography for general access structures A visual cryptography scheme for a set P of n participants is a method of encoding a secret image SI into n shadow images called shares, where each participant in P receives one share. Certain qualified subsets of participants can “visually” recover the secret image, but other, forbidden, sets of participants have no information (in an information-theoretic sense) on SI . A “visual” recovery for a set X ⊆ P consists of xeroxing the shares given to the participants in X onto transparencies, and then stacking them. The participants in a qualified set X will be able to see the secret image without any knowledge of cryptography and without performing any cryptographic computation. In this paper we propose two techniques for constructing visual cryptography schemes for general access structures. We analyze the structure of visual cryptography schemes and we prove bounds on the size of the shares distributed to the participants in the scheme. We provide a novel technique for realizing k out of n threshold visual cryptography schemes. Our construction for k out of n visual cryptography schemes is better with respect to pixel expansion than the one proposed by M. Naor and A. Shamir (Visual cryptography, in “Advances in Cryptology—Eurocrypt '94” CA. De Santis, Ed.), Lecture Notes in Computer Science, Vol. 950, pp. 1–12, Springer-Verlag, Berlin, 1995) and for the case of 2 out of n is the best possible. Finally, we consider graph-based access structures, i.e., access structures in which any qualified set of participants contains at least an edge of a given graph whose vertices represent the participants of the scheme.
A Model for Understanding How Virtual Reality Aids Complex Conceptual Learning Designers and evaluators of immersive virtual reality systems have many ideas concerning how virtual reality can facilitate learning. However, we have little information concerning which of virtual reality's features provide the most leverage for enhancing understanding or how to customize those affordances for different learning environments. In part, this reflects the truly complex nature of learning. Features of a learning environment do not act in isolation; other factors such as the concepts or skills to be learned, individual characteristics, the learning experience, and the interaction experience all play a role in shaping the learning process and its outcomes. Through Project Science Space, we have been trying to identify, use, and evaluate immersive virtual reality's affordances as a means to facilitate the mastery of complex, abstract concepts. In doing so, we are beginning to understand the interplay between virtual reality's features and other important factors in shaping the learning process and learning outcomes for this type of material. In this paper, we present a general model that describes how we think these factors work together and discuss some of the lessons we are learning about virtual reality's affordances in the context of this model for complex conceptual learning.
NETWRAP: An NDN Based Real-TimeWireless Recharging Framework for Wireless Sensor Networks Using vehicles equipped with wireless energy transmission technology to recharge sensor nodes over the air is a game-changer for traditional wireless sensor networks. The recharging policy regarding when to recharge which sensor nodes critically impacts the network performance. So far only a few works have studied such recharging policy for the case of using a single vehicle. In this paper, we propose NETWRAP, an N DN based Real Time Wireless Rech arging Protocol for dynamic wireless recharging in sensor networks. The real-time recharging framework supports single or multiple mobile vehicles. Employing multiple mobile vehicles provides more scalability and robustness. To efficiently deliver sensor energy status information to vehicles in real-time, we leverage concepts and mechanisms from named data networking (NDN) and design energy monitoring and reporting protocols. We derive theoretical results on the energy neutral condition and the minimum number of mobile vehicles required for perpetual network operations. Then we study how to minimize the total traveling cost of vehicles while guaranteeing all the sensor nodes can be recharged before their batteries deplete. We formulate the recharge optimization problem into a Multiple Traveling Salesman Problem with Deadlines (m-TSP with Deadlines), which is NP-hard. To accommodate the dynamic nature of node energy conditions with low overhead, we present an algorithm that selects the node with the minimum weighted sum of traveling time and residual lifetime. Our scheme not only improves network scalability but also ensures the perpetual operation of networks. Extensive simulation results demonstrate the effectiveness and efficiency of the proposed design. The results also validate the correctness of the theoretical analysis and show significant improvements that cut the number of nonfunctional nodes by half compared to the static scheme while maintaining the network overhead at the same level.
An efficient scheduling scheme for mobile charger in on-demand wireless rechargeable sensor networks. Existing studies on wireless sensor networks (WSNs) have revealed that the limited battery capacity of sensor nodes (SNs) hinders their perpetual operation. Recent findings in the domain of wireless energy transfer (WET) have attracted a lot of attention of academia and industry to cater the lack of energy in the WSNs. The main idea of WET is to restore the energy of SNs using one or more wireless mobile chargers (MCs), which leads to a new paradigm of wireless rechargeable sensor networks (WRSNs). The determination of an optimal order of charging the SNs (i.e., charging schedule) in an on-demand WRSN is a well-known NP-hard problem. Moreover, care must be taken while designing the charging schedule of an MC as requesting SNs introduce both spatial and temporal constraints. In this paper, we first present a Linear Programming (LP) formulation for the problem of scheduling an MC and then propose an efficient solution based on gravitational search algorithm (GSA). Our method is presented with a novel agent representation scheme and an efficient fitness function. We perform extensive simulations on the proposed scheme to demonstrate its effectiveness over two state-of-the-art algorithms, namely first come first serve (FCFS) and nearest job next with preemption (NJNP). The simulation results reveal that the proposed scheme outperforms both the existing algorithms in terms of charging latency. The virtue of our scheme is also proved by the well-known statistical test, analysis of variance (ANOVA), followed by post hoc analysis.
Robot tutor and pupils’ educational ability: Teaching the times tables Research shows promising results of educational robots in language and STEM tasks. In language, more research is available, occasionally in view of individual differences in pupils’ educational ability levels, and learning seems to improve with more expressive robot behaviors. In STEM, variations in robots’ behaviors have been examined with inconclusive results and never while systematically investigating how differences in educational abilities match with different robot behaviors. We applied an autonomously tutoring robot (without tablet, partly WOz) in a 2 × 2 experiment of social vs. neutral behavior in above-average vs. below-average schoolchildren (N = 86; age 8–10 years) while rehearsing the multiplication tables on a one-to-one basis. The standard school test showed that on average, pupils significantly improved their performance even after 3 occasions of 5-min exercises. Beyond-average pupils profited most from a robot tutor, whereas those below average in multiplication benefited more from a robot that showed neutral rather than more social behavior.
1.205714
0.205714
0.205714
0.205714
0.205714
0.068571
0.000952
0
0
0
0
0
0
0
Deep Metric Learning for Person Re-identification Various hand-crafted features and metric learning methods prevail in the field of person re-identification. Compared to these methods, this paper proposes a more general way that can learn a similarity metric from image pixels directly. By using a \"siamese\" deep neural network, the proposed method can jointly learn the color feature, texture feature and metric in a unified framework. The network has a symmetry structure with two sub-networks which are connected by a cosine layer. Each sub network includes two convolutional layers and a full connected layer. To deal with the big variations of person images, binomial deviance is used to evaluate the cost between similarities and labels, which is proved to be robust to outliers. Experiments on VIPeR illustrate the superior performance of our method and a cross database experiment also shows its good generalization.
Group Reidentification with Multigrained Matching and Integration The task of reidentifying groups of people under different camera views is an important yet less-studied problem. Group reidentification (Re-ID) is a very challenging task since it is not only adversely affected by common issues in traditional single-object Re-ID problems, such as viewpoint and human pose variations, but also suffers from changes in group layout and group membership. In this paper, we propose a novel concept of group granularity by characterizing a group image by multigrained objects: individual people and subgroups of two and three people within a group. To achieve robust group Re-ID, we first introduce multigrained representations which can be extracted via the development of two separate schemes, that is, one with handcrafted descriptors and another with deep neural networks. The proposed representation seeks to characterize both appearance and spatial relations of multigrained objects, and is further equipped with importance weights which capture variations in intragroup dynamics. Optimal group-wise matching is facilitated by a multiorder matching process which, in turn, dynamically updates the importance weights in iterative fashion. We evaluated three multicamera group datasets containing complex scenarios and large dynamics, with experimental results demonstrating the effectiveness of our approach.
A Strong Baseline and Batch Normalization Neck for Deep Person Re-Identification This study proposes a simple but strong baseline for deep person re-identification (ReID). Deep person ReID has achieved great progress and high performance in recent years. However, many state-of-the-art methods design complex network structures and concatenate multi-branch features. In the literature, some effective training tricks briefly appear in several papers or source codes. The present study collects and evaluates these effective training tricks in person ReID. By combining these tricks, the model achieves 94.5% rank-1 and 85.9% mean average precision on Market1501 with only using the global features of ResNet50. The performance surpasses all existing global- and part-based baselines in person ReID. We propose a novel neck structure named as batch normalization neck (BNNeck). BNNeck adds a batch normalization layer after global pooling layer to separate metric and classification losses into two different feature spaces because we observe they are inconsistent in one embedding space. Extended experiments show that BNNeck can boost the baseline, and our baseline can improve the performance of existing state-of-the-art methods. Our codes and models are available at: <uri xmlns:mml="http://www.w3.org/1998/Math/MathML" xmlns:xlink="http://www.w3.org/1999/xlink">https://github.com/michuanhaohao/reid-strong-baseline</uri>
Omni-Scale Feature Learning For Person Re-Identification As an instance-level recognition problem, person reidentification (ReID) relies on discriminative features, which not only capture different spatial scales but also encapsulate an arbitrary combination of multiple scales. We callse features of both homogeneous and heterogeneous scales omni-scale features. In this paper, a novel deep ReID CNN is designed, termed Omni-Scale Network (OSNet), for omni-scale feature learning. This is achieved by designing a residual block composed of multiple convolutional feature streams, each detecting features at a certain scale. Importantly, a novel unified aggregation gate is introduced to dynamically fuse multiscale features with input-dependent channel-wise weights. To efficiently learn spatial-channel correlations and avoid overfitting, the building block uses both pointwise and depthwise convolutions. By stacking such blocks layer-by-layer, our OSNet is extremely lightweight and can be trained from scratch on existing ReID benchmarks. Despite its small model size, our OSNet achieves state-of-the-art performance on six person-ReID datasets. Code and models are available at: https://github.com/ KaiyangZhou/deep-person-reid.
Unsupervised domain adaptive re-identification: Theory and practice •We introduce the theoretical guarantees of unsupervised domain adaptive re-ID based on [2]. A DA-learnability result is shown under three assumptions that concerning the feature space. To the best of our knowledge, our paper is the first theoretical analysis work on domain adaptive re-ID tasks.•We theoretically turn the goal of satisfying the assumptions into tractable loss functions on the encoder network and data samples.•A self-training scheme is proposed to iteratively minimizing the loss functions. Our framework is applicable to all re-ID tasks and the effectiveness is verified on large-scale datasets for diverse re-ID tasks.
Unsupervised Person Re-Identification By Soft Multilabel Learning Although unsupervised person re-identification (RE-ID) has drawn increasing research attentions due to its potential to address the scalability problem of supervised RE-ID models, it is very challenging to learn discriminative information in the absence of pairwise labels across disjoint camera views. To overcome this problem, we propose a deep model for the soft multilabel learning for unsupervised RE-ID. The idea is to learn a soft multilabel (real-valued label likelihood vector) for each unlabeled person by comparing the unlabeled person with a set of known reference persons from an auxiliary domain. We propose the soft multilabel-guided hard negative mining to learn a discriminative embedding for the unlabeled target domain by exploring the similarity consistency of the visual features and the soft multilabels of unlabeled target pairs. Since most target pairs are cross-view pairs, we develop the cross-view consistent soft multilabel learning to achieve the learning goal that the soft multilabels are consistently good across different camera views. To enable effecient soft multilabel learning, we introduce the reference agent learning to represent each reference person by a reference agent in a joint embedding. We evaluate our unified deep model on Market-1501 and DukeMTMC-reID. Our model outperforms the state-of-the-art unsupervised RE-ID methods by clear margins.
Invariance Matters: Exemplar Memory for Domain Adaptive Person Re-Identification This paper considers the domain adaptive person re-identification (re-ID) problem: learning a re-ID model from a labeled source domain and an unlabeled target domain. Conventional methods are mainly to reduce feature distribution gap between the source and target domains. However, these studies largely neglect the intra-domain variations in the target domain, which contain critical factors influencing the testing performance on the target domain. In this work, we comprehensively investigate into the intra-domain variations of the target domain and propose to generalize the re-ID model w.r.t three types of the underlying invariance, i.e., exemplar-invariance, camera-invariance and neighborhood-invariance. To achieve this goal, an exemplar memory is introduced to store features of the target domain and accommodate the three invariance properties. The memory allows us to enforce the invariance constraints over global training batch without significantly increasing computation cost. Experiment demonstrates that the three invariance properties and the proposed memory are indispensable towards an effective domain adaptation system. Results on three re-ID domains show that our domain adaptation accuracy outperforms the state of the art by a large margin. Code is available at: https://github.com/zhunzhong07/ECN.
Glove: Global Vectors for Word Representation.
A survey on sensor networks The advancement in wireless communications and electronics has enabled the development of low-cost sensor networks. The sensor networks can be used for various application areas (e.g., health, military, home). For different application areas, there are different technical issues that researchers are currently resolving. The current state of the art of sensor networks is captured in this article, where solutions are discussed under their related protocol stack layer sections. This article also points out the open research issues and intends to spark new interests and developments in this field.
Affective social robots For human-robot interaction to proceed in a smooth, natural manner, robots must adhere to human social norms. One such human convention is the use of expressive moods and emotions as an integral part of social interaction. Such expressions are used to convey messages such as ''I'm happy to see you'' or ''I want to be comforted,'' and people's long-term relationships depend heavily on shared emotional experiences. Thus, we have developed an affective model for social robots. This generative model attempts to create natural, human-like affect and includes distinctions between immediate emotional responses, the overall mood of the robot, and long-term attitudes toward each visitor to the robot, with a focus on developing long-term human-robot relationships. This paper presents the general affect model as well as particular details of our implementation of the model on one robot, the Roboceptionist. In addition, we present findings from two studies that demonstrate the model's potential.
Robust Adaptive Control of Uncertain Nonlinear Systems in the Presence of Input Saturation and External Disturbance. In this technical note, we consider adaptive control of single input uncertain nonlinear systems in the presence of input saturation and unknown external disturbance. By using backstepping approaches, two new robust adaptive control algorithms are developed by introducing a well defined smooth function and using a Nussbaum function. The Nussbaum function is introduced to compensate for the nonlinear term arising from the input saturation. Unlike some existing control schemes for systems with input saturation, the developed controllers do not require assumptions on the uncertain parameters within a known compact set and a priori knowledge on the bound of the external disturbance. Besides showing global stability, transient performance is also established and can be adjusted by tuning certain design parameters.
Step length estimation using handheld inertial sensors. In this paper a novel step length model using a handheld Micro Electrical Mechanical System (MEMS) is presented. It combines the user's step frequency and height with a set of three parameters for estimating step length. The model has been developed and trained using 12 different subjects: six men and six women. For reliable estimation of the step frequency with a handheld device, the frequency content of the handheld sensor's signal is extracted by applying the Short Time Fourier Transform (STFT) independently from the step detection process. The relationship between step and hand frequencies is analyzed for different hand's motions and sensor carrying modes. For this purpose, the frequency content of synchronized signals collected with two sensors placed in the hand and on the foot of a pedestrian has been extracted. Performance of the proposed step length model is assessed with several field tests involving 10 test subjects different from the above 12. The percentages of error over the travelled distance using universal parameters and a set of parameters calibrated for each subject are compared. The fitted solutions show an error between 2.5 and 5% of the travelled distance, which is comparable with that achieved by models proposed in the literature for body fixed sensors only.
The ApolloScape Dataset for Autonomous Driving Scene parsing aims to assign a class (semantic) label for each pixel in an image. It is a comprehensive analysis of an image. Given the rise of autonomous driving, pixel-accurate environmental perception is expected to be a key enabling technical piece. However, providing a large scale dataset for the design and evaluation of scene parsing algorithms, in particular for outdoor scenes, has been difficult. The per-pixel labelling process is prohibitively expensive, limiting the scale of existing ones. In this paper, we present a large-scale open dataset, ApolloScape, that consists of RGB videos and corresponding dense 3D point clouds. Comparing with existing datasets, our dataset has the following unique properties. The first is its scale, our initial release contains over 140K images - each with its per-pixel semantic mask, up to 1M is scheduled. The second is its complexity. Captured in various traffic conditions, the number of moving objects averages from tens to over one hundred (Figure 1). And the third is the 3D attribute, each image is tagged with high-accuracy pose information at cm accuracy and the static background point cloud has mm relative accuracy. We are able to label these many images by an interactive and efficient labelling pipeline that utilizes the high-quality 3D point cloud. Moreover, our dataset also contains different lane markings based on the lane colors and styles. We expect our new dataset can deeply benefit various autonomous driving related applications that include but not limited to 2D/3D scene understanding, localization, transfer learning, and driving simulation.
Robot tutor and pupils’ educational ability: Teaching the times tables Research shows promising results of educational robots in language and STEM tasks. In language, more research is available, occasionally in view of individual differences in pupils’ educational ability levels, and learning seems to improve with more expressive robot behaviors. In STEM, variations in robots’ behaviors have been examined with inconclusive results and never while systematically investigating how differences in educational abilities match with different robot behaviors. We applied an autonomously tutoring robot (without tablet, partly WOz) in a 2 × 2 experiment of social vs. neutral behavior in above-average vs. below-average schoolchildren (N = 86; age 8–10 years) while rehearsing the multiplication tables on a one-to-one basis. The standard school test showed that on average, pupils significantly improved their performance even after 3 occasions of 5-min exercises. Beyond-average pupils profited most from a robot tutor, whereas those below average in multiplication benefited more from a robot that showed neutral rather than more social behavior.
1.032274
0.028571
0.028571
0.017298
0.015532
0.011532
0.004942
0.000432
0
0
0
0
0
0
Cooperative Control Framework for Human Driver and Active Rear Steering System to Advance Active Safety Driving safety is quite significant for human drivers, especially in the emergencies and extreme conditions. To advance the active safety of vehicles, a cooperative control framework is proposed for human driver and the active rear steering system (ARS). Human driver model is established considering different driving characteristics to simulate different drivers. Moreover, vehicle-road model is bu...
Analysing user physiological responses for affective video summarisation. Video summarisation techniques aim to abstract the most significant content from a video stream. This is typically achieved by processing low-level image, audio and text features which are still quite disparate from the high-level semantics that end users identify with (the ‘semantic gap’). Physiological responses are potentially rich indicators of memorable or emotionally engaging video content for a given user. Consequently, we investigate whether they may serve as a suitable basis for a video summarisation technique by analysing a range of user physiological response measures, specifically electro-dermal response (EDR), respiration amplitude (RA), respiration rate (RR), blood volume pulse (BVP) and heart rate (HR), in response to a range of video content in a variety of genres including horror, comedy, drama, sci-fi and action. We present an analysis framework for processing the user responses to specific sub-segments within a video stream based on percent rank value normalisation. The application of the analysis framework reveals that users respond significantly to the most entertaining video sub-segments in a range of content domains. Specifically, horror content seems to elicit significant EDR, RA, RR and BVP responses, and comedy content elicits comparatively lower levels of EDR, but does seem to elicit significant RA, RR, BVP and HR responses. Drama content seems to elicit less significant physiological responses in general, and both sci-fi and action content seem to elicit significant EDR responses. We discuss the implications this may have for future affective video summarisation approaches.
On the roles of eye gaze and head dynamics in predicting driver's intent to change lanes Driver behavioral cues may present a rich source of information and feedback for future intelligent advanced driver-assistance systems (ADASs). With the design of a simple and robust ADAS in mind, we are interested in determining the most important driver cues for distinguishing driver intent. Eye gaze may provide a more accurate proxy than head movement for determining driver attention, whereas the measurement of head motion is less cumbersome and more reliable in harsh driving conditions. We use a lane-change intent-prediction system (McCall et al., 2007) to determine the relative usefulness of each cue for determining intent. Various combinations of input data are presented to a discriminative classifier, which is trained to output a prediction of probable lane-change maneuver at a particular point in the future. Quantitative results from a naturalistic driving study are presented and show that head motion, when combined with lane position and vehicle dynamics, is a reliable cue for lane-change intent prediction. The addition of eye gaze does not improve performance as much as simpler head dynamics cues. The advantage of head data over eye data is shown to be statistically significant (p
Detection of Driver Fatigue Caused by Sleep Deprivation This paper aims to provide reliable indications of driver drowsiness based on the characteristics of driver-vehicle interaction. A test bed was built under a simulated driving environment, and a total of 12 subjects participated in two experiment sessions requiring different levels of sleep (partial sleep-deprivation versus no sleep-deprivation) before the experiment. The performance of the subjects was analyzed in a series of stimulus-response and routine driving tasks, which revealed the performance differences of drivers under different sleep-deprivation levels. The experiments further demonstrated that sleep deprivation had greater effect on rule-based than on skill-based cognitive functions: when drivers were sleep-deprived, their performance of responding to unexpected disturbances degraded, while they were robust enough to continue the routine driving tasks such as lane tracking, vehicle following, and lane changing. In addition, we presented both qualitative and quantitative guidelines for designing drowsy-driver detection systems in a probabilistic framework based on the paradigm of Bayesian networks. Temporal aspects of drowsiness and individual differences of subjects were addressed in the framework.
Online Prediction of Driver Distraction Based on Brain Activity Patterns This paper presents a new computational framework for early detection of driver distractions (map viewing) using brain activity measured by electroencephalographic (EEG) signals. Compared with most studies in the literature, which are mainly focused on the classification of distracted and nondistracted periods, this study proposes a new framework to prospectively predict the start and end of a distraction period, defined by map viewing. The proposed prediction algorithm was tested on a data set of continuous EEG signals recorded from 24 subjects. During the EEG recordings, the subjects were asked to drive from an initial position to a destination using a city map in a simulated driving environment. The overall accuracy values for the prediction of the start and the end of map viewing were 81% and 70%, respectively. The experimental results demonstrated that the proposed algorithm can predict the start and end of map viewing with relatively high accuracy and can be generalized to individual subjects. The outcome of this study has a high potential to improve the design of future intelligent navigation systems. Prediction of the start of map viewing can be used to provide route information based on a driver's needs and consequently avoid map-viewing activities. Prediction of the end of map viewing can be used to provide warnings for potential long map-viewing durations. Further development of the proposed framework and its applications in driver-distraction predictions are also discussed.
Keep Your Scanners Peeled: Gaze Behavior as a Measure of Automation Trust During Highly Automated Driving. Objective: The feasibility of measuring drivers' automation trust via gaze behavior during highly automated driving was assessed with eye tracking and validated with self-reported automation trust in a driving simulator study. Background: Earlier research from other domains indicates that drivers' automation trust might be inferred from gaze behavior, such as monitoring frequency. Method: The gaze behavior and self-reported automation trust of 35 participants attending to a visually demanding non-driving-related task (NDRT) during highly automated driving was evaluated. The relationship between dispositional, situational, and learned automation trust with gaze behavior was compared. Results: Overall, there was a consistent relationship between drivers' automation trust and gaze behavior. Participants reporting higher automation trust tended to monitor the automation less frequently. Further analyses revealed that higher automation trust was associated with lower monitoring frequency of the automation during NDRTs, and an increase in trust over the experimental session was connected with a decrease in monitoring frequency. Conclusion: We suggest that (a) the current results indicate a negative relationship between drivers' self-reported automation trust and monitoring frequency, (b) gaze behavior provides a more direct measure of automation trust than other behavioral measures, and (c) with further refinement, drivers' automation trust during highly automated driving might be inferred from gaze behavior. Application: Potential applications of this research include the estimation of drivers' automation trust and reliance during highly automated driving.
3D separable convolutional neural network for dynamic hand gesture recognition. •The Frame Difference method is used to pre-process the input in order to filter the background.•A 3D separable CNN is proposed for dynamic gesture recognition. The standard 3D convolution process is decomposed into two processes: 3D depth-wise and 3D point-wise.•By the application of skip connection and layer-wise learning rate, the undesirable gradient dispersion due to the separation operation is solved and the performance of the network is improved.•A dynamic hand gesture library is built through HoloLens.
Cooperative channel assignment for VANETs based on multiagent reinforcement learning. Dynamic channel assignment (DCA) plays a key role in extending vehicular ad-hoc network capacity and mitigating congestion. However, channel assignment under vehicular direct communication scenarios faces mutual influence of large-scale nodes, the lack of centralized coordination, unknown global state information, and other challenges. To solve this problem, a multiagent reinforcement learning (RL) based cooperative DCA (RL-CDCA) mechanism is proposed. Specifically, each vehicular node can successfully learn the proper strategies of channel selection and backoff adaptation from the real-time channel state information (CSI) using two cooperative RL models. In addition, neural networks are constructed as nonlinear Q-function approximators, which facilitates the mapping of the continuously sensed input to the mixed policy output. Nodes are driven to locally share and incorporate their individual rewards such that they can optimize their policies in a distributed collaborative manner. Simulation results show that the proposed multiagent RL-CDCA can better reduce the one-hop packet delay by no less than 73.73%, improve the packet delivery ratio by no less than 12.66% on average in a highly dense situation, and improve the fairness of the global network resource allocation.
Reinforcement learning based data fusion method for multi-sensors In order to improve detection system robustness and reliability, multi-sensors fusion is used in modern air combat. In this paper, a data fusion method based on reinforcement learning is developed for multi-sensors. Initially, the cubic B-spline interpolation is used to solve time alignment problems of multisource data. Then, the reinforcement learning based data fusion (RLBDF) method is proposed to obtain the fusion results. With the case that the priori knowledge of target is obtained, the fusion accuracy reinforcement is realized by the error between fused value and actual value. Furthermore, the Fisher information is instead used as the reward if the priori knowledge is unable to be obtained. Simulations results verify that the developed method is feasible and effective for the multi-sensors data fusion in air combat.
Mobile Edge Computing Enabled 5G Health Monitoring for Internet of Medical Things: A Decentralized Game Theoretic Approach The prompt evolution of Internet of Medical Things (IoMT) promotes pervasive in-home health monitoring networks. However, excessive requirements of patients result in insufficient spectrum resources and communication overload. Mobile Edge Computing (MEC) enabled 5G health monitoring is conceived as a favorable paradigm to tackle such an obstacle. In this paper, we construct a cost-efficient in-home health monitoring system for IoMT by dividing it into two sub-networks, i.e., intra-Wireless Body Area Networks (WBANs) and beyond-WBANs. Highlighting the characteristics of IoMT, the cost of patients depends on medical criticality, Age of Information (AoI) and energy consumption. For intra-WBANs, a cooperative game is formulated to allocate the wireless channel resources. While for beyond-WBANs, considering the individual rationality and potential selfishness, a decentralized non-cooperative game is proposed to minimize the system-wide cost in IoMT. We prove that the proposed algorithm can reach a Nash equilibrium. In addition, the upper bound of the algorithm time complexity and the number of patients benefiting from MEC is theoretically derived. Performance evaluations demonstrate the effectiveness of our proposed algorithm with respect to the system-wide cost and the number of patients benefiting from MEC.
Artificial fish swarm algorithm: a survey of the state-of-the-art, hybridization, combinatorial and indicative applications FSA (artificial fish-swarm algorithm) is one of the best methods of optimization among the swarm intelligence algorithms. This algorithm is inspired by the collective movement of the fish and their various social behaviors. Based on a series of instinctive behaviors, the fish always try to maintain their colonies and accordingly demonstrate intelligent behaviors. Searching for food, immigration and dealing with dangers all happen in a social form and interactions between all fish in a group will result in an intelligent social behavior.This algorithm has many advantages including high convergence speed, flexibility, fault tolerance and high accuracy. This paper is a review of AFSA algorithm and describes the evolution of this algorithm along with all improvements, its combination with various methods as well as its applications. There are many optimization methods which have a affinity with this method and the result of this combination will improve the performance of this method. Its disadvantages include high time complexity, lack of balance between global and local search, in addition to lack of benefiting from the experiences of group members for the next movements.
Short-Term Traffic Flow Forecasting: An Experimental Comparison of Time-Series Analysis and Supervised Learning The literature on short-term traffic flow forecasting has undergone great development recently. Many works, describing a wide variety of different approaches, which very often share similar features and ideas, have been published. However, publications presenting new prediction algorithms usually employ different settings, data sets, and performance measurements, making it difficult to infer a clear picture of the advantages and limitations of each model. The aim of this paper is twofold. First, we review existing approaches to short-term traffic flow forecasting methods under the common view of probabilistic graphical models, presenting an extensive experimental comparison, which proposes a common baseline for their performance analysis and provides the infrastructure to operate on a publicly available data set. Second, we present two new support vector regression models, which are specifically devised to benefit from typical traffic flow seasonality and are shown to represent an interesting compromise between prediction accuracy and computational efficiency. The SARIMA model coupled with a Kalman filter is the most accurate model; however, the proposed seasonal support vector regressor turns out to be highly competitive when performing forecasts during the most congested periods.
TSCA: A Temporal-Spatial Real-Time Charging Scheduling Algorithm for On-Demand Architecture in Wireless Rechargeable Sensor Networks. The collaborative charging issue in Wireless Rechargeable Sensor Networks (WRSNs) is a popular research problem. With the help of wireless power transfer technology, electrical energy can be transferred from wireless charging vehicles (WCVs) to sensors, providing a new paradigm to prolong network lifetime. Existing techniques on collaborative charging usually take the periodical and deterministic approach, but neglect influences of non-deterministic factors such as topological changes and node failures, making them unsuitable for large-scale WRSNs. In this paper, we develop a temporal-spatial charging scheduling algorithm, namely TSCA, for the on-demand charging architecture. We aim to minimize the number of dead nodes while maximizing energy efficiency to prolong network lifetime. First, after gathering charging requests, a WCV will compute a feasible movement solution. A basic path planning algorithm is then introduced to adjust the charging order for better efficiency. Furthermore, optimizations are made in a global level. Then, a node deletion algorithm is developed to remove low efficient charging nodes. Lastly, a node insertion algorithm is executed to avoid the death of abandoned nodes. Extensive simulations show that, compared with state-of-the-art charging scheduling algorithms, our scheme can achieve promising performance in charging throughput, charging efficiency, and other performance metrics.
A novel adaptive dynamic programming based on tracking error for nonlinear discrete-time systems In this paper, to eliminate the tracking error by using adaptive dynamic programming (ADP) algorithms, a novel formulation of the value function is presented for the optimal tracking problem (TP) of nonlinear discrete-time systems. Unlike existing ADP methods, this formulation introduces the control input into the tracking error, and ignores the quadratic form of the control input directly, which makes the boundedness and convergence of the value function independent of the discount factor. Based on the proposed value function, the optimal control policy can be deduced without considering the reference control input. Value iteration (VI) and policy iteration (PI) methods are applied to prove the optimality of the obtained control policy, and derived the monotonicity property and convergence of the iterative value function. Simulation examples realized with neural networks and the actor–critic structure are provided to verify the effectiveness of the proposed ADP algorithm.
1.2
0.2
0.2
0.2
0.2
0.2
0.2
0.2
0.1
0
0
0
0
0
Amazon, Google and Microsoft Solutions for IoT: Architectures and a Performance Comparison Internet of Things (IoT) aims to connect the real world made up of devices, sensors and actuators to the virtual world of Internet in order to interconnect devices with each other generating information from the gathered data. Devices, in general, have limited computational power and limited storage capacity. Cloud Computing (CC) has virtually unlimited capacity in terms of storage and computing power, and is based on sharing resources. Therefore, the integration between IoT and CC seems to be one of the most promising solutions. In fact, many of the biggest companies that offer Cloud Services are focusing on the IoT world to offer services also in this direction to their users. In this paper we compare the three main Cloud Platforms (Amazon Web Services, Google Cloud Platform and Microsoft Azure) regarding to the services made available for the IoT. After describing the typical architecture of an IoT application, we map the Cloud-IoT Platforms services with this architecture analyzing the key points for each platform. At the same time, in order to conduct a comparative analysis of performance, we focus on a service made available by all platforms (MQTT middleware) building the reference scenarios and the metrics to be taken into account. Finally, we provide an overview of platform costs based on different loads. The aim is not to declare a winner, but to provide a useful tool to developers to make an informed choice of a platform depending on the use case.
Infrastructure as a Service and Cloud Technologies To choose the most appropriate cloud-computing model for your organization, you must analyze your IT infrastructure, usage, and needs. To help with this, this article describes cloud computing's current status.
Dynamic Management of Virtual Infrastructures Cloud infrastructures are becoming an appropriate solution to address the computational needs of scientific applications. However, the use of public or on-premises Infrastructure as a Service (IaaS) clouds requires users to have non-trivial system administration skills. Resource provisioning systems provide facilities to choose the most suitable Virtual Machine Images (VMI) and basic configuration of multiple instances and subnetworks. Other tasks such as the configuration of cluster services, computational frameworks or specific applications are not trivial on the cloud, and normally users have to manually select the VMI that best fits, including undesired additional services and software packages. This paper presents a set of components that ease the access and the usability of IaaS clouds by automating the VMI selection, deployment, configuration, software installation, monitoring and update of Virtual Appliances. It supports APIs from a large number of virtual platforms, making user applications cloud-agnostic. In addition it integrates a contextualization system to enable the installation and configuration of all the user required applications providing the user with a fully functional infrastructure. Therefore, golden VMIs and configuration recipes can be easily reused across different deployments. Moreover, the contextualization agent included in the framework supports horizontal (increase/decrease the number of resources) and vertical (increase/decrease resources within a running Virtual Machine) by properly reconfiguring the software installed, considering the configuration of the multiple resources running. This paves the way for automatic virtual infrastructure deployment, customization and elastic modification at runtime for IaaS clouds.
Virtual machine placement quality estimation in cloud infrastructures using integer linear programming This paper is devoted to the quality estimation of virtual machine (VM) placement in cloud infrastructures, i.e., to choose the best hosts for a given set of VMs. We focus on test generation and monitoring techniques for comparing the placement result of a given implementation with an optimal solution with respect to given criteria. We show how Integer Linear Programming problems can be formulated and utilized for deriving test suites and optimal solutions to provide verdicts concerning the quality of VM placement implementations; the quality is calculated as the distance from an optimal placement for a given criterion (or a set of criteria). The presented approach is generic and showcased on resource utilization, energy consumption, and resource over-commitment cost. Experiments performed with different VM placement algorithms (including the VM placement algorithms implemented in widely used platforms, such as OpenStack) exhibit the competence of such algorithms with respect to different criteria.
Integrating security and privacy in software development As a consequence to factors such as progress made by the attackers, release of new technologies and use of increasingly complex systems, and threats to applications security have been continuously evolving. Security of code and privacy of data must be implemented in both design and programming practice to face such scenarios. In such a context, this paper proposes a software development approach, Privacy Oriented Software Development (POSD), that complements traditional development processes by integrating the activities needed for addressing security and privacy management in software systems. The approach is based on 5 key elements (Privacy by Design, Privacy Design Strategies, Privacy Pattern, Vulnerabilities, Context). The approach can be applied in two directions forward and backward, for developing new software systems or re-engineering an existing one. This paper presents the POSD approach in the backward mode together with an application in the context of an industrial project. Results show that POSD is able to discover software vulnerabilities, identify the remediation patterns needed for addressing them in the source code, and design the target architecture to be used for guiding privacy-oriented system re-engineering.
Image quality assessment: from error visibility to structural similarity. Objective methods for assessing perceptual image quality traditionally attempted to quantify the visibility of errors (differences) between a distorted image and a reference image using a variety of known properties of the human visual system. Under the assumption that human visual perception is highly adapted for extracting structural information from a scene, we introduce an alternative complementary framework for quality assessment based on the degradation of structural information. As a specific example of this concept, we develop a Structural Similarity Index and demonstrate its promise through a set of intuitive examples, as well as comparison to both subjective ratings and state-of-the-art objective methods on a database of images compressed with JPEG and JPEG2000.
Vision meets robotics: The KITTI dataset We present a novel dataset captured from a VW station wagon for use in mobile robotics and autonomous driving research. In total, we recorded 6 hours of traffic scenarios at 10-100 Hz using a variety of sensor modalities such as high-resolution color and grayscale stereo cameras, a Velodyne 3D laser scanner and a high-precision GPS/IMU inertial navigation system. The scenarios are diverse, capturing real-world traffic situations, and range from freeways over rural areas to inner-city scenes with many static and dynamic objects. Our data is calibrated, synchronized and timestamped, and we provide the rectified and raw image sequences. Our dataset also contains object labels in the form of 3D tracklets, and we provide online benchmarks for stereo, optical flow, object detection and other tasks. This paper describes our recording platform, the data format and the utilities that we provide.
A tutorial on support vector regression In this tutorial we give an overview of the basic ideas underlying Support Vector (SV) machines for function estimation. Furthermore, we include a summary of currently used algorithms for training SV machines, covering both the quadratic (or convex) programming part and advanced methods for dealing with large datasets. Finally, we mention some modifications and extensions that have been applied to the standard SV algorithm, and discuss the aspect of regularization from a SV perspective.
GameFlow: a model for evaluating player enjoyment in games Although player enjoyment is central to computer games, there is currently no accepted model of player enjoyment in games. There are many heuristics in the literature, based on elements such as the game interface, mechanics, gameplay, and narrative. However, there is a need to integrate these heuristics into a validated model that can be used to design, evaluate, and understand enjoyment in games. We have drawn together the various heuristics into a concise model of enjoyment in games that is structured by flow. Flow, a widely accepted model of enjoyment, includes eight elements that, we found, encompass the various heuristics from the literature. Our new model, GameFlow, consists of eight elements -- concentration, challenge, skills, control, clear goals, feedback, immersion, and social interaction. Each element includes a set of criteria for achieving enjoyment in games. An initial investigation and validation of the GameFlow model was carried out by conducting expert reviews of two real-time strategy games, one high-rating and one low-rating, using the GameFlow criteria. The result was a deeper understanding of enjoyment in real-time strategy games and the identification of the strengths and weaknesses of the GameFlow model as an evaluation tool. The GameFlow criteria were able to successfully distinguish between the high-rated and low-rated games and identify why one succeeded and the other failed. We concluded that the GameFlow model can be used in its current form to review games; further work will provide tools for designing and evaluating enjoyment in games.
Adapting visual category models to new domains Domain adaptation is an important emerging topic in computer vision. In this paper, we present one of the first studies of domain shift in the context of object recognition. We introduce a method that adapts object models acquired in a particular visual domain to new imaging conditions by learning a transformation that minimizes the effect of domain-induced changes in the feature distribution. The transformation is learned in a supervised manner and can be applied to categories for which there are no labeled examples in the new domain. While we focus our evaluation on object recognition tasks, the transform-based adaptation technique we develop is general and could be applied to nonimage data. Another contribution is a new multi-domain object database, freely available for download. We experimentally demonstrate the ability of our method to improve recognition on categories with few or no target domain labels and moderate to large changes in the imaging conditions.
A Web-Based Tool For Control Engineering Teaching In this article a new tool for control engineering teaching is presented. The tool was implemented using Java applets and is freely accessible through Web. It allows the analysis and simulation of linear control systems and was created to complement the theoretical lectures in basic control engineering courses. The article is not only centered in the description of the tool but also in the methodology to use it and its evaluation in an electrical engineering degree. Two practical problems are included in the manuscript to illustrate the use of the main functions implemented. The developed web-based tool can be accessed through the link http://www.controlweb.cyc.ull.es. (C) 2006 Wiley Periodicals, Inc.
Adaptive Consensus Control for a Class of Nonlinear Multiagent Time-Delay Systems Using Neural Networks Because of the complicity of consensus control of nonlinear multiagent systems in state time-delay, most of previous works focused only on linear systems with input time-delay. An adaptive neural network (NN) consensus control method for a class of nonlinear multiagent systems with state time-delay is proposed in this paper. The approximation property of radial basis function neural networks (RBFNNs) is used to neutralize the uncertain nonlinear dynamics in agents. An appropriate Lyapunov-Krasovskii functional, which is obtained from the derivative of an appropriate Lyapunov function, is used to compensate the uncertainties of unknown time delays. It is proved that our proposed approach guarantees the convergence on the basis of Lyapunov stability theory. The simulation results of a nonlinear multiagent time-delay system and a multiple collaborative manipulators system show the effectiveness of the proposed consensus control algorithm.
5G Virtualized Multi-access Edge Computing Platform for IoT Applications. The next generation of fifth generation (5G) network, which is implemented using Virtualized Multi-access Edge Computing (vMEC), Network Function Virtualization (NFV) and Software Defined Networking (SDN) technologies, is a flexible and resilient network that supports various Internet of Things (IoT) devices. While NFV provides flexibility by allowing network functions to be dynamically deployed and inter-connected, vMEC provides intelligence at the edge of the mobile network reduces latency and increases the available capacity. With the diverse development of networking applications, the proposed vMEC use of Container-based Virtualization Technology (CVT) as gateway with IoT devices for flow control mechanism in scheduling and analysis methods will effectively increase the application Quality of Service (QoS). In this work, the proposed IoT gateway is analyzed. The combined effect of simultaneously deploying Virtual Network Functions (VNFs) and vMEC applications on a single network infrastructure, and critically in effecting exhibits low latency, high bandwidth and agility that will be able to connect large scale of devices. The proposed platform efficiently exploiting resources from edge computing and cloud computing, and takes IoT applications that adapt to network conditions to degrade an average 30% of end to end network latency.
Energy harvesting algorithm considering max flow problem in wireless sensor networks. In Wireless Sensor Networks (WSNs), sensor nodes with poor energy always have bad effect on the data rate or max flow. These nodes are called bottleneck nodes. In this paper, in order to increase the max flow, we assume an energy harvesting WSNs environment to investigate the cooperation of multiple Mobile Chargers (MCs). MCs are mobile robots that use wireless charging technology to charge sensor nodes in WSNs. This means that in energy harvesting WSNs environments, sensor nodes can obtain energy replenishment by using MCs or collecting energy from nature by themselves. In our research, we use MCs to improve the energy of the sensor nodes by performing multiple rounds of unified scheduling, and finally achieve the purpose of increasing the max flow at sinks. Firstly, we model this problem as a Linear Programming (LP) to search the max flow in a round of charging scheduling and prove that the problem is NP-hard. In order to solve the problem, we propose a heuristic approach: deploying MCs in units of paths with the lowest energy node priority. To reduce the energy consumption of MCs and increase the charging efficiency, we also take the optimization of MCs’ moving distance into our consideration. Finally, we extend the method to multiple rounds of scheduling called BottleNeck. Simulation results show that Bottleneck performs well at increasing max flow.
1.2
0.2
0.2
0.2
0.2
0
0
0
0
0
0
0
0
0
Study of joint routing and wireless charging strategies in sensor networks In recent years, wireless charging (a.k.a. wireless energy transferring) [3] has been recognized as a promising alternative to address the energy constraint challenge in wireless sensor networks. Comparing to the conventional energy conservation or harvesting approaches, wireless charging can replenish energy in a more controllable manner and does not require accurate location of or physical alignment to sensor nodes. In spite of these advantages, there has been little research on how much potential performance improvement may be achieved by applying the wireless charging approach to sensor networks and how to fully leverage its potential. In this paper, as one of the first efforts to study these issues, we (1) formulate the problem of maximizing the sensor network lifetime via codetermining routing and charging (ML-JRC), (2) prove the NP-hardness nature of the problem and derive an upper bound of the maximum sensor network lifetime that is achievable with ML-JRC, and (3) present a set of heuristics to determine the wireless charging strategies under various routing schemes, and demonstrate their effectiveness via in-depth simulation.
Impacts of traveling paths on energy provisioning for industrial wireless rechargeable sensor networks Traditional Industrial Wireless Sensor Networks (IWSNs) are constrained by limited battery energy. Recent breakthroughs in wireless power transfer have inspired the emergence of Industrial Wireless Rechargeable Sensor Networks (IWRSNs). IWRSNs usually contain one or more mobile chargers which can traverse the network to replenish energy supply for sensor nodes. The essential problem in mobile energy provisioning is to find the optimum path along which the mobile chargers travel to improve charging performance, prolong the battery lifespan of nodes and reduce the charging latency as much as possible. In this paper, we introduce and analyze the impacts of four traveling paths, namely, SCAN, HILBERT, S-CURVES(ad) and Z-curve on energy provisioning for IWRSNs. This evaluation aims to embody effective and essential properties that a superior traveling path should possess. Our simulations show that S-CURVES(ad) outperforms the other traveling paths in the lifetime of nodes and traveling efficiency. And at the same time, it has relatively small charging latency.
Mobility in wireless sensor networks - Survey and proposal. Targeting an increasing number of potential application domains, wireless sensor networks (WSN) have been the subject of intense research, in an attempt to optimize their performance while guaranteeing reliability in highly demanding scenarios. However, hardware constraints have limited their application, and real deployments have demonstrated that WSNs have difficulties in coping with complex communication tasks – such as mobility – in addition to application-related tasks. Mobility support in WSNs is crucial for a very high percentage of application scenarios and, most notably, for the Internet of Things. It is, thus, important to know the existing solutions for mobility in WSNs, identifying their main characteristics and limitations. With this in mind, we firstly present a survey of models for mobility support in WSNs. We then present the Network of Proxies (NoP) assisted mobility proposal, which relieves resource-constrained WSN nodes from the heavy procedures inherent to mobility management. The presented proposal was implemented and evaluated in a real platform, demonstrating not only its advantages over conventional solutions, but also its very good performance in the simultaneous handling of several mobile nodes, leading to high handoff success rate and low handoff time.
A survey on cross-layer solutions for wireless sensor networks Ever since wireless sensor networks (WSNs) have emerged, different optimizations have been proposed to overcome their constraints. Furthermore, the proposal of new applications for WSNs have also created new challenges to be addressed. Cross-layer approaches have proven to be the most efficient optimization techniques for these problems, since they are able to take the behavior of the protocols at each layer into consideration. Thus, this survey proposes to identify the key problems of WSNs and gather available cross-layer solutions for them that have been proposed so far, in order to provide insights on the identification of open issues and provide guidelines for future proposals.
OPPC: An Optimal Path Planning Charging Scheme Based on Schedulability Evaluation for WRSNs. The lack of schedulability evaluation of previous charging schemes in wireless rechargeable sensor networks (WRSNs) degrades the charging efficiency, leading to node exhaustion. We propose an Optimal Path Planning Charging scheme, namely OPPC, for the on-demand charging architecture. OPPC evaluates the schedulability of a charging mission, which makes charging scheduling predictable. It provides an optimal charging path which maximizes charging efficiency. When confronted with a non-schedulable charging mission, a node discarding algorithm is developed to enable the schedulability. Experimental simulations demonstrate that OPPC can achieve better performance in successful charging rate as well as charging efficiency.
Machine learning algorithms for wireless sensor networks: A survey. •The survey of machine learning algorithms for WSNs from the period 2014 to March 2018.•Machine learning (ML) for WSNs with their advantages, features and limitations.•A statistical survey of ML-based algorithms for WSNs.•Reasons to choose a ML techniques to solve issues in WSNs.•The survey proposes a discussion on open issues.
Task Scheduling for Energy-Harvesting-Based IoT: A Survey and Critical Analysis The Internet of Things (IoT) has important applications in our daily lives, including health and fitness tracking, environmental monitoring, and transportation. However, sensor nodes in IoT suffer from the limited lifetime of batteries resulting from their finite energy availability. A promising solution is to harvest energy from environmental sources, such as solar, kinetic, thermal, and radio-fr...
Distributed Sensor Nodes Charged by Mobile Charger with Directional Antenna and by Energy Trading for Balancing. Provision of energy to wireless sensor networks is crucial for their sustainable operation. Sensor nodes are typically equipped with batteries as their operating energy sources. However, when the sensor nodes are sited in almost inaccessible locations, replacing their batteries incurs high maintenance cost. Under such conditions, wireless charging of sensor nodes by a mobile charger with an antenna can be an efficient solution. When charging distributed sensor nodes, a directional antenna, rather than an omnidirectional antenna, is more energy-efficient because of smaller proportion of off-target radiation. In addition, for densely distributed sensor nodes, it can be more effective for some undercharged sensor nodes to harvest energy from neighboring overcharged sensor nodes than from the remote mobile charger, because this reduces the pathloss of charging signal due to smaller distances. In this paper, we propose a hybrid charging scheme that combines charging by a mobile charger with a directional antenna, and energy trading, e.g., transferring and harvesting, between neighboring sensor nodes. The proposed scheme is compared with other charging scheme. Simulations demonstrate that the hybrid charging scheme with a directional antenna achieves a significant reduction in the total charging time required for all sensor nodes to reach a target energy level.
Data Collection in Wireless Sensor Networks with Mobile Elements: A Survey Wireless sensor networks (WSNs) have emerged as an effective solution for a wide range of applications. Most of the traditional WSN architectures consist of static nodes which are densely deployed over a sensing area. Recently, several WSN architectures based on mobile elements (MEs) have been proposed. Most of them exploit mobility to address the problem of data collection in WSNs. In this article we first define WSNs with MEs and provide a comprehensive taxonomy of their architectures, based on the role of the MEs. Then we present an overview of the data collection process in such a scenario, and identify the corresponding issues and challenges. On the basis of these issues, we provide an extensive survey of the related literature. Finally, we compare the underlying approaches and solutions, with hints to open problems and future research directions.
Sequence to Sequence Learning with Neural Networks. Deep Neural Networks (DNNs) are powerful models that have achieved excellent performance on difficult learning tasks. Although DNNs work well whenever large labeled training sets are available, they cannot be used to map sequences to sequences. In this paper, we present a general end-to-end approach to sequence learning that makes minimal assumptions on the sequence structure. Our method uses a multilayered Long Short-Term Memory (LSTM) to map the input sequence to a vector of a fixed dimensionality, and then another deep LSTM to decode the target sequence from the vector. Our main result is that on an English to French translation task from the WMT-14 dataset, the translations produced by the LSTM achieve a BLEU score of 34.8 on the entire test set, where the LSTM's BLEU score was penalized on out-of-vocabulary words. Additionally, the LSTM did not have difficulty on long sentences. For comparison, a phrase-based SMT system achieves a BLEU score of 33.3 on the same dataset. When we used the LSTM to rerank the 1000 hypotheses produced by the aforementioned SMT system, its BLEU score increases to 36.5, which is close to the previous state of the art. The LSTM also learned sensible phrase and sentence representations that are sensitive to word order and are relatively invariant to the active and the passive voice. Finally, we found that reversing the order of the words in all source sentences (but not target sentences) improved the LSTM's performance markedly, because doing so introduced many short term dependencies between the source and the target sentence which made the optimization problem easier.
Survey of NLOS identification and error mitigation problems in UWB-based positioning algorithms for dense environments In this survey, the currently available ultra-wideband-based non-line-of-sight (NLOS) identification and error mitigation methods are presented. They are classified into several categories and their comparison is presented in two tables: one each for NLOS identification and error mitigation. NLOS identification methods are classified based on range estimates, channel statistics, and the actual maps of the building and environment. NLOS error mitigation methods are categorized based on direct path and statistics-based detection.
Robust ear identification using sparse representation of local texture descriptors Automated personal identification using localized ear images has wide range of civilian and law-enforcement applications. This paper investigates a new approach for more accurate ear recognition and verification problem using the sparse representation of local gray-level orientations. We exploit the computational simplicity of localized Radon transform for the robust ear shape representation and also investigate the effectiveness of local curvature encoding using Hessian based feature representation. The ear representation problem is modeled as the sparse coding solution based on multi-orientation Radon transform dictionary whose solution is computed using the convex optimization approach. We also study the nonnegative formulation such problem, to address the limitations from the regularized optimization problem, in the sparse representation of localized ear features. The log-Gabor filter based approach and the localized Radon transform based feature representation has been used as baseline algorithm to ascertain the effectiveness of the proposed approach. We present experimental results from publically available UND and IITD ear databases which achieve significant improvement in the performance, both for the recognition and authentication problem, and confirm the usefulness of proposed approach for more accurate ear identification.
Distributed Adaptive Fuzzy Containment Control of Stochastic Pure-Feedback Nonlinear Multiagent Systems With Local Quantized Controller and Tracking Constraint This paper studies the distributed adaptive fuzzy containment tracking control for a class of high-order stochastic pure-feedback nonlinear multiagent systems with multiple dynamic leaders and performance constraint requirement. The control inputs are quantized by hysteresis quantizers. Mean value theorems are used to transfer the nonaffine systems into affine forms and a nonlinear decomposition is employed to solve the quantized input control problem. With a novel structure barrier Lyapunov function, the distributed control strategy is developed. It is strictly proved that the outputs of the followers converge to the convex hull spanned by the multiple dynamic leaders, the containment tracking errors satisfy the performance constraint requirement and the resulting leader-following multiagent system is stable in probability based on Lyapunov stability theory. At last, simulation is provided to show the validity and the advantages of the proposed techniques.
Active Suspension Control of Quarter-Car System With Experimental Validation A reliable, efficient, and simple control is presented and validated for a quarter-car active suspension system equipped with an electro-hydraulic actuator. Unlike the existing techniques, this control does not use any function approximation, e.g., neural networks (NNs) or fuzzy-logic systems (FLSs), while the unmolded dynamics, including the hydraulic actuator behavior, can be accommodated effectively. Hence, the heavy computational costs and tedious parameter tuning phase can be remedied. Moreover, both the transient and steady-state suspension performance can be retained by incorporating prescribed performance functions (PPFs) into the control implementation. This guaranteed performance is particularly useful for guaranteeing the safe operation of suspension systems. Apart from theoretical studies, some practical considerations of control implementation and several parameter tuning guidelines are suggested. Experimental results based on a practical quarter-car active suspension test-rig demonstrate that this control can obtain a superior performance and has better computational efficiency over several other control methods.
1.045401
0.044728
0.04
0.04
0.04
0.04
0.04
0.02
0.008206
0
0
0
0
0
Unmanned Aerial Vehicle Location Routing Problem With Charging Stations for Belt Conveyor Inspection System in the Mining Industry Technological advances have opened up the possibility of using unmanned aerial vehicles (UAVs) in diverse environments. The mining industry has been looking for solutions to handle periodic inspections of the belt conveyors that transport iron ore. The state of the art indicates the use of UAVs for this task as an attractive, low-cost and safe alternative, allowing for a significant increase in security. A new concise mixed-integer linear programming (MILP) model is developed to address UAV routing and charging station planning for belt conveyor inspection. We conduct computational tests covering a real conveyor belt system in Brazil to validate the model in practical applications. The loading terminal possesses approximately 120 km of belt conveyors, leading to 230 inspection points. Instances of different sizes were generated by randomly sampling a subset of these points and using two different drone specifications. The results show that the new optimization modeling satisfies the problem requirements and is a significant contribution to the automation of inspection in the mining industry.
Survey of NLOS identification and error mitigation problems in UWB-based positioning algorithms for dense environments In this survey, the currently available ultra-wideband-based non-line-of-sight (NLOS) identification and error mitigation methods are presented. They are classified into several categories and their comparison is presented in two tables: one each for NLOS identification and error mitigation. NLOS identification methods are classified based on range estimates, channel statistics, and the actual maps of the building and environment. NLOS error mitigation methods are categorized based on direct path and statistics-based detection.
Reliable Classification of Vehicle Types Based on Cascade Classifier Ensembles Vehicle-type recognition based on images is a challenging task. This paper comparatively studied two feature extraction methods for image description, i.e., the Gabor wavelet transform and the Pyramid Histogram of Oriented Gradients (PHOG). The Gabor transform has been widely adopted to extract image features for various vision tasks. PHOG has the superiority in its description of more discriminating information. A highly reliable classification scheme was proposed by cascade classifier ensembles with reject option to accommodate the situations where no decision should be made if there exists adequate ambiguity. The first ensemble is heterogeneous, consisting of several classifiers, including $k$-nearest neighbors (kNNs), multiple-layer perceptrons (MLPs), support vector machines (SVMs), and random forest. The classification reliability is further enhanced by a second classifier ensemble, which is composed of a set of base MLPs coordinated by an ensemble metalearning method called rotation forest (RF). For both of the ensembles, rejection option is accomplished by relating the consensus degree from majority voting to a confidence measure and by abstaining to classify ambiguous samples if the consensus degree is lower than a threshold. The final class label is assigned by dual majority voting from the two ensembles. Experimental results using more than 600 images from a variety of 21 makes of cars and vans demonstrated the effectiveness of the proposed approach. The cascade ensembles produce consistently reliable results. With a moderate ensemble size of 25 in the second ensemble, the two-stage classification scheme offers 98.65% accuracy with a rejection rate of 2.5%, exhibiting promising potential for real-world applications.
People detection and tracking from aerial thermal views Detection and tracking of people in visible-light images has been subject to extensive research in the past decades with applications ranging from surveillance to search-and-rescue. Following the growing availability of thermal cameras and the distinctive thermal signature of humans, research effort has been focusing on developing people detection and tracking methodologies applicable to this sensing modality. However, a plethora of challenges arise on the transition from visible-light to thermal images, especially with the recent trend of employing thermal cameras onboard aerial platforms (e.g. in search-and-rescue research) capturing oblique views of the scenery. This paper presents a new, publicly available dataset of annotated thermal image sequences, posing a multitude of challenges for people detection and tracking. Moreover, we propose a new particle filter based framework for tracking people in aerial thermal images. Finally, we evaluate the performance of this pipeline on our dataset, incorporating a selection of relevant, state-of-the-art methods and present a comprehensive discussion of the merits spawning from our study.
Pareto-Optimization for Scheduling of Crude Oil Operations in Refinery via Genetic Algorithm. With the interaction of discrete-event and continuous processes, it is challenging to schedule crude oil operations in a refinery. This paper studies the optimization problem of finding a detailed schedule to realize a given refining schedule. This is a multiobjective optimization problem with a combinatorial nature. Since the original problem cannot be directly solved by using heuristics and meta-heuristics, the problem is transformed into an assignment problem of charging tanks and distillers. Based on such a transformation, by analyzing the properties of the problem, this paper develops a chromosome that can describe a feasible schedule such that meta-heuristics can be applied. Then, it innovatively adopts an improved nondominated sorting genetic algorithm to solve the problem for the first time. An industrial case study is used to test the proposed solution method. The results show that the method makes a significant performance improvement and is applicable to real-life refinery scheduling problems.
Adaptive Consensus-Based Distributed Target Tracking With Dynamic Cluster in Sensor Networks. This paper is concerned with the target tracking problem over a filtering network with dynamic cluster and data fusion. A novel distributed consensus-based adaptive Kalman estimation is developed to track a linear moving target. Both optimal filtering gain and average disagreement of the estimates are considered in the filter design. In order to estimate the states of the target more precisely, an optimal Kalman gain is obtained by minimizing the mean-squared estimation error. An adaptive consensus factor is employed to adjust the optimal gain as well as to acquire a better filtering performance. In the filter's information exchange, dynamic cluster selection and two-stage hierarchical fusion structure are employed to get more accurate estimation. At the first stage, every sensor collects information from its neighbors and runs the Kalman estimation algorithm to obtain a local estimate of system states. At the second stage, each local sensor sends its estimate to the cluster head to get a fused estimation. Finally, an illustrative example is presented to validate the effectiveness of the proposed scheme.
3D Point Clouds Data Super Resolution Aided LiDAR Odometry for Vehicular Positioning in Urban Canyons LiDAR odometry algorithms are extensively studied for vehicular positioning. However, achieving high-precision positioning using low-cost 16-channel LiDAR in urban canyons remains a challenging problem due to the limited point cloud density from low-cost LiDAR and excessive dynamic surrounding objects. To fill this gap, this paper proposes enriching sparse 3D point clouds to denser clouds via a no...
A survey on ear biometrics Recognizing people by their ear has recently received significant attention in the literature. Several reasons account for this trend: first, ear recognition does not suffer from some problems associated with other non-contact biometrics, such as face recognition; second, it is the most promising candidate for combination with the face in the context of multi-pose face recognition; and third, the ear can be used for human recognition in surveillance videos where the face may be occluded completely or in part. Further, the ear appears to degrade little with age. Even though current ear detection and recognition systems have reached a certain level of maturity, their success is limited to controlled indoor conditions. In addition to variation in illumination, other open research problems include hair occlusion, earprint forensics, ear symmetry, ear classification, and ear individuality. This article provides a detailed survey of research conducted in ear detection and recognition. It provides an up-to-date review of the existing literature revealing the current state-of-art for not only those who are working in this area but also for those who might exploit this new approach. Furthermore, it offers insights into some unsolved ear recognition problems as well as ear databases available for researchers.
Delay-Aware Microservice Coordination in Mobile Edge Computing: A Reinforcement Learning Approach As an emerging service architecture, microservice enables decomposition of a monolithic web service into a set of independent lightweight services which can be executed independently. With mobile edge computing, microservices can be further deployed in edge clouds dynamically, launched quickly, and migrated across edge clouds easily, providing better services for users in proximity. However, the user mobility can result in frequent switch of nearby edge clouds, which increases the service delay when users move away from their serving edge clouds. To address this issue, this article investigates microservice coordination among edge clouds to enable seamless and real-time responses to service requests from mobile users. The objective of this work is to devise the optimal microservice coordination scheme which can reduce the overall service delay with low costs. To this end, we first propose a dynamic programming-based offline microservice coordination algorithm, that can achieve the globally optimal performance. However, the offline algorithm heavily relies on the availability of the prior information such as computation request arrivals, time-varying channel conditions and edge cloud's computation capabilities required, which is hard to be obtained. Therefore, we reformulate the microservice coordination problem using Markov decision process framework and then propose a reinforcement learning-based online microservice coordination algorithm to learn the optimal strategy. Theoretical analysis proves that the offline algorithm can find the optimal solution while the online algorithm can achieve near-optimal performance. Furthermore, based on two real-world datasets, i.e., the Telecom's base station dataset and Taxi Track dataset from Shanghai, experiments are conducted. The experimental results demonstrate that the proposed online algorithm outperforms existing algorithms in terms of service delay and migration costs, and the achieved performance is close to the optimal performance obtained by the offline algorithm.
Swarm Intelligence: From Natural to Artificial Systems
Energy Provisioning in Wireless Rechargeable Sensor Networks Wireless rechargeable sensor networks (WRSNs) have emerged as an alternative to solving the challenges of size and operation time posed by traditional battery-powered systems. In this paper, we study a WRSN built from the industrial wireless identification and sensing platform (WISP) and commercial off-the-shelf RFID readers. The paper-thin WISP tags serve as sensors and can harvest energy from RF signals transmitted by the readers. This kind of WRSNs is highly desirable for indoor sensing and activity recognition and is gaining attention in the research community. One fundamental question in WRSN design is how to deploy readers in a network to ensure that the WISP tags can harvest sufficient energy for continuous operation. We refer to this issue as the energy provisioning problem. Based on a practical wireless recharge model supported by experimental data, we investigate two forms of the problem: point provisioning and path provisioning. Point provisioning uses the least number of readers to ensure that a static tag placed in any position of the network will receive a sufficient recharge rate for sustained operation. Path provisioning exploits the potential mobility of tags (e.g., those carried by human users) to further reduce the number of readers necessary: mobile tags can harvest excess energy in power-rich regions and store it for later use in power-deficient regions. Our analysis shows that our deployment methods, by exploiting the physical characteristics of wireless recharging, can greatly reduce the number of readers compared with those assuming traditional coverage models.
A Fully Distributed Algorithm for Dynamic Channel Adaptation in Canonical Communication Networks In this letter, a low-complexity, fully distributed algorithm is designed for dynamic channel adaptation in a canonical communication network, where each player can independently update its action without any information exchange. Both the static and dynamic channel environments are studied. The proposed algorithm converges to a set of correlated equilibria with probability one. Moreover, the optimality property of the problem is analyzed. Simulation results demonstrate that the proposed algorithm achieves a near-optimal performance for interference mitigation.
Safe mutations for deep and recurrent neural networks through output gradients While neuroevolution (evolving neural networks) has been successful across a variety of domains from reinforcement learning, to artificial life, to evolutionary robotics, it is rarely applied to large, deep neural networks. A central reason is that while random mutation generally works in low dimensions, a random perturbation of thousands or millions of weights will likely break existing functionality. This paper proposes a solution: a family of safe mutation (SM) operators that facilitate exploration without dramatically altering network behavior or requiring additional interaction with the environment. The most effective SM variant scales the degree of mutation of each individual weight according to the sensitivity of the network's outputs to that weight, which requires computing the gradient of outputs with respect to the weights (instead of the gradient of error, as in conventional deep learning). This safe mutation through gradients (SM-G) operator dramatically increases the ability of a simple genetic algorithm-based neuroevolution method to find solutions in high-dimensional domains that require deep and/or recurrent neural networks, including domains that require processing raw pixels. By improving our ability to evolve deep neural networks, this new safer approach to mutation expands the scope of domains amenable to neuroevolution.
Sustainable and Efficient Data Collection from WSNs to Cloud. The development of cloud computing pours great vitality into traditional wireless sensor networks (WSNs). The integration of WSNs and cloud computing has received a lot of attention from both academia and industry. However, collecting data from WSNs to cloud is not sustainable. Due to the weak communication ability of WSNs, uploading big sensed data to the cloud within the limited time becomes a b...
1.05
0.05
0.05
0.05
0.05
0.05
0.05
0
0
0
0
0
0
0
Bi-level programming based real-time path planning for unmanned aerial vehicles This paper presents a novel real-time path planning approach for unmanned aerial vehicles (UAVs) based on bi-level programming (BLP), in which the planning problem is described as a leader-follower decision making model. The proposed approach can fulfill an integrated path planning requirement, including several realistic abilities of convergence to target, obstacle avoidance, path length optimization, flight path smoothing and adaptability to the changes of the UAV's kinematic and sensory properties. In the BLP model, opposition obstacles and UAV's interrelated performances are described to construct path searching constraints, and variable planning time intervals are introduced to generate navigable flight paths only when necessary. A discretization solution algorithm embedded with five heuristic optimization strategies is particularly designed to speed up the planning. Moreover, convergence and computational cost, as well as potential extensions, are discussed to expose the efficiency and applicability of the approach. Numerous simulations in stochastic and representative scenarios demonstrate not only the effectiveness of the approach in generating optimized flight paths for UAVs, but also its clear advantages through comparisons with four typical methods, while synthetically considering the crucial abilities.
Chimp optimization algorithm. •A novel optimizer called Chimp Optimization Algorithm (ChOA) is proposed.•ChOA is inspired by individual intelligence and sexual motivation of chimps.•ChOA alleviates the problems of slow convergence rate and trapping in local optima.•The four main steps of Chimp hunting are implemented.
Development of a UAV-LiDAR System with Application to Forest Inventory We present the development of a low-cost Unmanned Aerial Vehicle-Light Detecting and Ranging (UAV-LiDAR) system and an accompanying workflow to produce 3D point clouds. UAV systems provide an unrivalled combination of high temporal and spatial resolution datasets. The TerraLuma UAV-LiDAR system has been developed to take advantage of these properties and in doing so overcome some of the current limitations of the use of this technology within the forestry industry. A modified processing workflow including a novel trajectory determination algorithm fusing observations from a GPS receiver, an Inertial Measurement Unit (IMU) and a High Definition (HD) video camera is presented. The advantages of this workflow are demonstrated using a rigorous assessment of the spatial accuracy of the final point clouds. It is shown that due to the inclusion of video the horizontal accuracy of the final point cloud improves from 0.61 m to 0.34 m (RMS error assessed against ground control). The effect of the very high density point clouds (up to 62 points per m(2)) produced by the UAV-LiDAR system on the measurement of tree location, height and crown width are also assessed by performing repeat surveys over individual isolated trees. The standard deviation of tree height is shown to reduce from 0.26 m, when using data with a density of 8 points per m(2), to 0.15 m when the higher density data was used. Improvements in the uncertainty of the measurement of tree location, 0.80 m to 0.53 m, and crown width, 0.69 m to 0.61 m are also shown.
Solution of structural and mathematical optimization problems using a new hybrid swarm intelligence optimization algorithm. In this investigation a new optimization algorithm named as interactive search algorithm (ISA) is presented. This method is developed through modifying and hybridizing the affirmative features of recently developed integrated particle swarm optimization (iPSO) algorithm with the pairwise knowledge sharing mechanism of the teaching and learning based optimization (TLBO) method. Proposed ISA provides two different navigation schemes as Tracking and Interacting. Each agent based on its tendency factor can pick one of these two schemes for searching the domain. Additionally, ISA utilizes an improved fly-back technique to handle problem constraints. The proposed method is tested on a set of mathematical and structural optimization benchmark problems with discrete and continuous variables. Numerical results indicate that the new algorithm is competitive with other well-stablished metaheuristic algorithms.
Three-dimension path planning for UCAV using hybrid meta-heuristic ACO-DE algorithm Three-dimension path planning of uninhabited combat air vehicle (UCAV) is a complicated optimal problem, which mainly focuses on optimizing the flight route considering the different types of constrains under complicated combating environments. A new hybrid meta-heuristic ant colony optimization (ACO) and differential evolution (DE) algorithm is proposed to solve the UCAV three-dimension path planning problem. DE is applied to optimize the pheromone trail of the improved ACO model during the process of ant pheromone updating. Then, the UCAV can find the safe path by connecting the chosen nodes of the three-dimensional mesh while avoiding the threats area and costing minimum fuel. This new approach can accelerate the global convergence speed while preserving the strong robustness of the basic ACO. The realization procedure for this hybrid meta-heuristic approach is also presented in detail. In order to make the optimized UCAV path more feasible, the к-trajectory is adopted for smoothing the path. Finally, series experimental comparison results demonstrate that this proposed hybrid meta-heuristic method is more effective and feasible in UCAV three-dimension path planning than the basic ACO model.
Efficient Deployment of Multiple Unmanned Aerial Vehicles for Optimal Wireless Coverage. In this letter, the efficient deployment of multiple unmanned aerial vehicles (UAVs) acting as wireless base stations that provide coverage for ground users is analyzed. First, the downlink coverage probability for UAVs as a function of the altitude and the antenna gain is derived. Next, using circle packing theory, the 3-D locations of the UAVs is determined in a way that the total coverage area ...
Accurate Self-Localization in RFID Tag Information Grids Using FIR Filtering Grid navigation spaces nested with the radio-frequency identification (RFID) tags are promising for industrial and other needs, because each tag can deliver information about a local two-dimensional or three-dimensional surrounding. The approach, however, requires high accuracy in vehicle self-localization. Otherwise, errors may lead to collisions; possibly even fatal. We propose a new extended finite impulse response (EFIR) filtering algorithm and show that it meets this need. The EFIR filter requires an optimal averaging interval, but does not involve the noise statistics which are often not well known to the engineer. It is more accurate than the extended Kalman filter (EKF) under real operation conditions and its iterative algorithm has the Kalman form. Better performance of the proposed EFIR filter is demonstrated based on extensive simulations in a comparison to EKF, which is widely used in RFID tag grids. We also show that errors in noise covariances may provoke divergence in EKF, whereas the EFIR filter remains stable and is thus more robust.
Distributed Representations, Simple Recurrent Networks, And Grammatical Structure In this paper three problems for a connectionist account of language are considered:1. What is the nature of linguistic representations?2. How can complex structural relationships such as constituent structure be represented?3. How can the apparently open-ended nature of language be accommodated by a fixed-resource system?Using a prediction task, a simple recurrent network (SRN) is trained on multiclausal sentences which contain multiply-embedded relative clauses. Principal component analysis of the hidden unit activation patterns reveals that the network solves the task by developing complex distributed representations which encode the relevant grammatical relations and hierarchical constituent structure. Differences between the SRN state representations and the more traditional pushdown store are discussed in the final section.
Latent dirichlet allocation We describe latent Dirichlet allocation (LDA), a generative probabilistic model for collections of discrete data such as text corpora. LDA is a three-level hierarchical Bayesian model, in which each item of a collection is modeled as a finite mixture over an underlying set of topics. Each topic is, in turn, modeled as an infinite mixture over an underlying set of topic probabilities. In the context of text modeling, the topic probabilities provide an explicit representation of a document. We present efficient approximate inference techniques based on variational methods and an EM algorithm for empirical Bayes parameter estimation. We report results in document modeling, text classification, and collaborative filtering, comparing to a mixture of unigrams model and the probabilistic LSI model.
Reservoir computing approaches to recurrent neural network training Echo State Networks and Liquid State Machines introduced a new paradigm in artificial recurrent neural network (RNN) training, where an RNN (the reservoir) is generated randomly and only a readout is trained. The paradigm, becoming known as reservoir computing, greatly facilitated the practical application of RNNs and outperformed classical fully trained RNNs in many tasks. It has lately become a vivid research field with numerous extensions of the basic idea, including reservoir adaptation, thus broadening the initial paradigm to using different methods for training the reservoir and the readout. This review systematically surveys both current ways of generating/adapting the reservoirs and training different types of readouts. It offers a natural conceptual classification of the techniques, which transcends boundaries of the current “brand-names” of reservoir methods, and thus aims to help in unifying the field and providing the reader with a detailed “map” of it.
Empirical Modelling of Genetic Algorithms This paper addresses the problem of reliably setting genetic algorithm parameters for consistent labelling problems. Genetic algorithm parameters are notoriously difficult to determine. This paper proposes a robust empirical framework, based on the analysis of factorial experiments. The use of a graeco-latin square permits an initial study of a wide range of parameter settings. This is followed by fully crossed factorial experiments with narrower ranges, which allow detailed analysis by logistic regression. The empirical models derived can be used to determine optimal algorithm parameters and to shed light on interactions between the parameters and their relative importance. Re-fined models are produced, which are shown to be robust under extrapolation to up to triple the problem size.
Magnetic, Acceleration Fields and Gyroscope Quaternion (MAGYQ)-based attitude estimation with smartphone sensors for indoor pedestrian navigation. The dependence of proposed pedestrian navigation solutions on a dedicated infrastructure is a limiting factor to the deployment of location based services. Consequently self-contained Pedestrian Dead-Reckoning (PDR) approaches are gaining interest for autonomous navigation. Even if the quality of low cost inertial sensors and magnetometers has strongly improved, processing noisy sensor signals combined with high hand dynamics remains a challenge. Estimating accurate attitude angles for achieving long term positioning accuracy is targeted in this work. A new Magnetic, Acceleration fields and GYroscope Quaternion (MAGYQ)-based attitude angles estimation filter is proposed and demonstrated with handheld sensors. It benefits from a gyroscope signal modelling in the quaternion set and two new opportunistic updates: magnetic angular rate update (MARU) and acceleration gradient update (AGU). MAGYQ filter performances are assessed indoors, outdoors, with dynamic and static motion conditions. The heading error, using only the inertial solution, is found to be less than 10 degrees after 1.5 km walking. The performance is also evaluated in the positioning domain with trajectories computed following a PDR strategy.
Quaternion polar harmonic Fourier moments for color images. •Quaternion polar harmonic Fourier moments (QPHFM) is proposed.•Complex Chebyshev-Fourier moments (CHFM) is extended to quaternion QCHFM.•Comparison experiments between QPHFM and QZM, QPZM, QOFMM, QCHFM and QRHFM are conducted.•QPHFM performs superbly in image reconstruction and invariant object recognition.•The importance of phase information of QPHFM in image reconstruction are discussed.
Robot tutor and pupils’ educational ability: Teaching the times tables Research shows promising results of educational robots in language and STEM tasks. In language, more research is available, occasionally in view of individual differences in pupils’ educational ability levels, and learning seems to improve with more expressive robot behaviors. In STEM, variations in robots’ behaviors have been examined with inconclusive results and never while systematically investigating how differences in educational abilities match with different robot behaviors. We applied an autonomously tutoring robot (without tablet, partly WOz) in a 2 × 2 experiment of social vs. neutral behavior in above-average vs. below-average schoolchildren (N = 86; age 8–10 years) while rehearsing the multiplication tables on a one-to-one basis. The standard school test showed that on average, pupils significantly improved their performance even after 3 occasions of 5-min exercises. Beyond-average pupils profited most from a robot tutor, whereas those below average in multiplication benefited more from a robot that showed neutral rather than more social behavior.
1.11
0.1
0.1
0.1
0.033333
0.001667
0
0
0
0
0
0
0
0
Network Slicing and Softwarization: A Survey on Principles, Enabling Technologies, and Solutions. Network slicing has been identified as the backbone of the rapidly evolving 5G technology. However, as its consolidation and standardization progress, there are no literatures that comprehensively discuss its key principles, enablers, and research challenges. This paper elaborates network slicing from an end-to-end perspective detailing its historical heritage, principal concepts, enabling technol...
Resource Slicing in Virtual Wireless Networks: A Survey. New architectural and design approaches for radio access networks have appeared with the introduction of network virtualization in the wireless domain. One of these approaches splits the wireless network infrastructure into isolated virtual slices under their own management, requirements, and characteristics. Despite the advances in wireless virtualization, there are still many open issues regarding the resource allocation and isolation of wireless slices. Because of the dynamics and shared nature of the wireless medium, guaranteeing that the traffic on one slice will not affect the traffic on the others has proven to be difficult. In this paper, we focus on the detailed definition of the problem, discussing its challenges. We also provide a review of existing works that deal with the problem, analyzing how new trends such as software defined networking and network function virtualization can assist in the slicing. We will finally describe some research challenges on this topic.
Survey on Network Slicing for Internet of Things Realization in 5G Networks Internet of Things (IoT) is an emerging technology that makes people’s lives smart by conquering a plethora of diverse application and service areas. In near future, the fifth-generation (5G) wireless networks provide the connectivity for this IoT ecosystem. It has been carefully designed to facilitate the exponential growth in the IoT field. Network slicing is one of the key technologies in the 5...
On Service Resilience in Cloud-Native 5G Mobile Systems. To cope with the tremendous growth in mobile data traffic on one hand, and the modest average revenue per user on the other hand, mobile operators have been exploring network virtualization and cloud computing technologies to build cost-efficient and elastic mobile networks and to have them offered as a cloud service. In such cloud-based mobile networks, ensuring service resilience is an important challenge to tackle. Indeed, high availability and service reliability are important requirements of carrier grade, but not necessarily intrinsic features of cloud computing. Building a system that requires the five nines reliability on a platform that may not always grant it is, therefore, a hurdle. Effectively, in carrier cloud, service resilience can be heavily impacted by a failure of any network function (NF) running on a virtual machine (VM). In this paper, we introduce a framework, along with efficient and proactive restoration mechanisms, to ensure service resilience in carrier cloud. As restoration of a NF failure impacts a potential number of users, adequate network overload control mechanisms are also proposed. A mathematical model is developed to evaluate the performance of the proposed mechanisms. The obtained results are encouraging and demonstrate that the proposed mechanisms efficiently achieve their design goals.
A Tutorial on Ultrareliable and Low-Latency Communications in 6G: Integrating Domain Knowledge Into Deep Learning As one of the key communication scenarios in the fifth-generation and also the sixth-generation (6G) mobile communication networks, ultrareliable and low-latency communications (URLLCs) will be central for the development of various emerging mission-critical applications. State-of-the-art mobile communication systems do not fulfill the end-to-end delay and overall reliability requirements of URLLCs. In particular, a holistic framework that takes into account latency, reliability, availability, scalability, and decision-making under uncertainty is lacking. Driven by recent breakthroughs in deep neural networks, deep learning algorithms have been considered as promising ways of developing enabling technologies for URLLCs in future 6G networks. This tutorial illustrates how domain knowledge (models, analytical tools, and optimization frameworks) of communications and networking can be integrated into different kinds of deep learning algorithms for URLLCs. We first provide some background of URLLCs and review promising network architectures and deep learning frameworks for 6G. To better illustrate how to improve learning algorithms with domain knowledge, we revisit model-based analytical tools and cross-layer optimization frameworks for URLLCs. Following this, we examine the potential of applying supervised/unsupervised deep learning and deep reinforcement learning in URLLCs and summarize related open problems. Finally, we provide simulation and experimental results to validate the effectiveness of different learning algorithms and discuss future directions.
A Survey on Security Aspects for 3GPP 5G Networks With the continuous development of mobile communication technologies, Third Generation Partnership Project (3GPP) has proposed related standards with the fifth generation mobile communication technology (5G), which marks the official start of the evolution from the current Long Term Evolution (LTE) system to the next generation mobile communication system (5GS). This paper makes a large number of contributions to the security aspects of 3GPP 5G networks. Firstly, we present an overview of the network architecture and security functionality of the 3GPP 5G networks. Subsequently, we focus on the new features and techniques including the support of massive Internet of Things (IoT) devices, Device to Device (D2D) communication, Vehicle to Everything (V2X) communication, and network slice, which incur the huge challenges for the security aspects in 3GPP 5G networks. Finally, we discuss in detail the security features, security requirements or security vulnerabilities, existing security solutions and some open research issues about the new features and techniques in 3GPP 5G network.
Revisiting traffic anomaly detection using software defined networking Despite their exponential growth, home and small office/home office networks continue to be poorly managed. Consequently, security of hosts in most home networks is easily compromised and these hosts are in turn used for largescale malicious activities without the home users' knowledge. We argue that the advent of Software Defined Networking (SDN) provides a unique opportunity to effectively detect and contain network security problems in home and home office networks. We show how four prominent traffic anomaly detection algorithms can be implemented in an SDN context using Openflow compliant switches and NOX as a controller. Our experiments indicate that these algorithms are significantly more accurate in identifying malicious activities in the home networks as compared to the ISP. Furthermore, the efficiency analysis of our SDN implementations on a programmable home network router indicates that the anomaly detectors can operate at line rates without introducing any performance penalties for the home network traffic.
Wireless sensor network survey A wireless sensor network (WSN) has important applications such as remote environmental monitoring and target tracking. This has been enabled by the availability, particularly in recent years, of sensors that are smaller, cheaper, and intelligent. These sensors are equipped with wireless interfaces with which they can communicate with one another to form a network. The design of a WSN depends significantly on the application, and it must consider factors such as the environment, the application's design objectives, cost, hardware, and system constraints. The goal of our survey is to present a comprehensive review of the recent literature since the publication of [I.F. Akyildiz, W. Su, Y. Sankarasubramaniam, E. Cayirci, A survey on sensor networks, IEEE Communications Magazine, 2002]. Following a top-down approach, we give an overview of several new applications and then review the literature on various aspects of WSNs. We classify the problems into three different categories: (1) internal platform and underlying operating system, (2) communication protocol stack, and (3) network services, provisioning, and deployment. We review the major development in these three categories and outline new challenges.
Energy-Aware Task Offloading and Resource Allocation for Time-Sensitive Services in Mobile Edge Computing Systems Mobile Edge Computing (MEC) is a promising architecture to reduce the energy consumption of mobile devices and provide satisfactory quality-of-service to time-sensitive services. How to jointly optimize task offloading and resource allocation to minimize the energy consumption subject to the latency requirement remains an open problem, which motivates this paper. When the latency constraint is tak...
Symbolic model checking for real-time systems We describe finite-state programs over real-numbered time in a guarded-command language with real-valued clocks or, equivalently, as finite automata with real-valued clocks. Model checking answers the question which states of a real-time program satisfy a branching-time specification (given in an extension of CTL with clock variables). We develop an algorithm that computes this set of states symbolically as a fixpoint of a functional on state predicates, without constructing the state space. For this purpose, we introduce a μ-calculus on computation trees over real-numbered time. Unfortunately, many standard program properties, such as response for all nonzeno execution sequences (during which time diverges), cannot be characterized by fixpoints: we show that the expressiveness of the timed μ-calculus is incomparable to the expressiveness of timed CTL. Fortunately, this result does not impair the symbolic verification of "implementable" real-time programs-those whose safety constraints are machine-closed with respect to diverging time and whose fairness constraints are restricted to finite upper bounds on clock values. All timed CTL properties of such programs are shown to be computable as finitely approximable fixpoints in a simple decidable theory.
Optimal distributed interference avoidance: potential game and learning. This article studies the problem of distributed interference avoidance (IA) through channel selection for distributed wireless networks, where mutual interference only occurs among nearby users. First, an interference graph is used to characterise the limited range of interference, and then the distributed IA problem is formulated as a graph colouring problem. Because solving the graph colouring problem is non-deterministic polynomial hard even in a centralised manner, the task of obtaining the optimal channel selection profile distributively is challenging. We formulate this problem as a channel selection game, which is proved to be an exact potential game with the weighted aggregate interference serving as the potential function. On the basis of this, a distributed learning algorithm is proposed to achieve the optimal channel selection profile that constitutes an optimal Nash equilibrium point of the channel selection game. The proposed learning algorithm is fully distributed because it needs information about neither the network topology nor the actions and the experienced interference of others. Simulation results show that the proposed potential game theoretic IA algorithm outperforms the existing algorithm because it minimises the aggregate weighted interference and achieves higher network rate. Copyright (c) 2012 John Wiley & Sons, Ltd.
A Framework of Joint Mobile Energy Replenishment and Data Gathering in Wireless Rechargeable Sensor Networks Recent years have witnessed the rapid development and proliferation of techniques on improving energy efficiency for wireless sensor networks. Although these techniques can relieve the energy constraint on wireless sensors to some extent, the lifetime of wireless sensor networks is still limited by sensor batteries. Recent studies have shown that energy rechargeable sensors have the potential to provide perpetual network operations by capturing renewable energy from external environments. However, the low output of energy capturing devices can only provide intermittent recharging opportunities to support low-rate data services due to spatial-temporal, geographical or environmental factors. To provide steady and high recharging rates and achieve energy efficient data gathering from sensors, in this paper, we propose to utilize mobility for joint energy replenishment and data gathering. In particular, a multi-functional mobile entity, called SenCarin this paper, is employed, which serves not only as a mobile data collector that roams over the field to gather data via short-range communication but also as an energy transporter that charges static sensors on its migration tour via wireless energy transmissions. Taking advantages of SenCar's controlled mobility, we focus on the joint optimization of effective energy charging and high-performance data collections. We first study this problem in general networks with random topologies. We give a two-step approach for the joint design. In the first step, the locations of a subset of sensors are periodically selected as anchor points, where the SenCar will sequentially visit to charge the sensors at these locations and gather data from nearby sensors in a multi-hop fashion. To achieve a desirable balance between energy replenishment amount and data gathering latency, we provide a selection algorithm to search for a maximum number of anchor points where sensors hold the least battery energy, and meanwhile by visiting them, - he tour length of the SenCar is no more than a threshold. In the second step, we consider data gathering performance when the SenCar migrates among these anchor points. We formulate the problem into a network utility maximization problem and propose a distributed algorithm to adjust data rates at which sensors send buffered data to the SenCar, link scheduling and flow routing so as to adapt to the up-to-date energy replenishing status of sensors. Besides general networks, we also study a special scenario where sensors are regularly deployed. For this case we can provide a simplified solution of lower complexity by exploiting the symmetry of the topology. Finally, we validate the effectiveness of our approaches by extensive numerical results, which show that our solutions can achieve perpetual network operations and provide high network utility.
Multiple switching-time-dependent discretized Lyapunov functions/functionals methods for stability analysis of switched time-delay stochastic systems. This paper presents novel approaches for stability analysis of switched linear time-delay stochastic systems under dwell time constraint. Instead of using comparison principle, piecewise switching-time-dependent discretized Lyapunov functions/functionals are introduced to analyze the stability of switched stochastic systems with constant or time-varying delays. These Lyapunov functions/functionals are decreasing during the dwell time and non-increasing at switching instants, which lead to two mode-dependent dwell-time-based delay-independent stability criteria for the switched systems without restricting the stability of the subsystems. Comparison and numerical examples are provided to show the efficiency of the proposed results.
Learning Feature Recovery Transformer for Occluded Person Re-Identification One major issue that challenges person re-identification (Re-ID) is the ubiquitous occlusion over the captured persons. There are two main challenges for the occluded person Re-ID problem, i.e., the interference of noise during feature matching and the loss of pedestrian information brought by the occlusions. In this paper, we propose a new approach called Feature Recovery Transformer (FRT) to address the two challenges simultaneously, which mainly consists of visibility graph matching and feature recovery transformer. To reduce the interference of the noise during feature matching, we mainly focus on visible regions that appear in both images and develop a visibility graph to calculate the similarity. In terms of the second challenge, based on the developed graph similarity, for each query image, we propose a recovery transformer that exploits the feature sets of its k-nearest neighbors in the gallery to recover the complete features. Extensive experiments across different person Re-ID datasets, including occluded, partial and holistic datasets, demonstrate the effectiveness of FRT. Specifically, FRT significantly outperforms state-of-the-art results by at least 6.2% Rank- 1 accuracy and 7.2% mAP scores on the challenging Occluded-Duke dataset.
1.11
0.11
0.11
0.1
0.1
0.06
0.0025
0
0
0
0
0
0
0
Deep Reinforcement Learning for User Association and Resource Allocation in Heterogeneous Cellular Networks. Heterogeneous cellular networks can offload the mobile traffic and reduce the deployment costs, which have been considered to be a promising technique in the next-generation wireless network. Due to the non-convex and combinatorial characteristics, it is challenging to obtain an optimal strategy for the joint user association and resource allocation issue. In this paper, a reinforcement learning (...
Trajectory Design and Power Control for Multi-UAV Assisted Wireless Networks: A Machine Learning Approach. A novel framework is proposed for the trajectory design of multiple unmanned aerial vehicles (UAVs) based on the prediction of usersu0027 mobility information. The problem of joint trajectory design and power control is formulated for maximizing the instantaneous sum transmit rate while satisfying the rate requirement of users. In an effort to solve this pertinent problem, a three-step approach is proposed which is based on machine learning techniques to obtain both the position information of users and the trajectory design of UAVs. Firstly, a multi-agent Q-learning based placement algorithm is proposed for determining the optimal positions of the UAVs based on the initial location of the users. Secondly, in an effort to determine the mobility information of users based on a real dataset, their position data is collected from Twitter to describe the anonymous user-trajectories in the physical world. In the meantime, an echo state network (ESN) based prediction algorithm is proposed for predicting the future positions of users based on the real dataset. Thirdly, a multi-agent Q-learning based algorithm is conceived for predicting the position of UAVs in each time slot based on the movement of users. The algorithm is proved to be able to converge to an optimal state. In this algorithm, multiple UAVs act as agents to find optimal actions by interacting with their environment and learn from their mistakes. Numerical results are provided to demonstrate that as the size of the reservoir increases, the proposed ESN approach improves the prediction accuracy. Finally, we demonstrate that throughput gains of about $17%$ are achieved.
On-Board Deep Q-Network for UAV-Assisted Online Power Transfer and Data Collection Unmanned Aerial Vehicles (UAVs) with Microwave Power Transfer (MPT) capability provide a practical means to deploy a large number of wireless powered sensing devices into areas with no access to persistent power supplies. The UAV can charge the sensing devices remotely and harvest their data. A key challenge is online MPT and data collection in the presence of on-board control of a UAV (e.g., patrolling velocity) for preventing battery drainage and data queue overflow of the devices, while up-to-date knowledge on battery level and data queue of the devices is not available at the UAV. In this paper, an on-board deep Q-network is developed to minimize the overall data packet loss of the sensing devices, by optimally deciding the device to be charged and interrogated for data collection, and the instantaneous patrolling velocity of the UAV. Specifically, we formulate a Markov Decision Process (MDP) with the states of battery level and data queue length of devices, channel conditions, and waypoints given the trajectory of the UAV; and solve it optimally with Q-learning. Furthermore, we propose the on-board deep Q-network that enlarges the state space of the MDP, and a deep reinforcement learning based scheduling algorithm that asymptotically derives the optimal solution online, even when the UAV has only outdated knowledge on the MDP states. Numerical results demonstrate that our deep reinforcement learning algorithm reduces the packet loss by at least 69.2%, as compared to existing non-learning greedy algorithms.
Unmanned Aerial Vehicle-Aided Communications: Joint Transmit Power and Trajectory Optimization. This letter investigates the transmit power and trajectory optimization problem for unmanned aerial vehicle (UAV)-aided networks. Different from majority of the existing studies with fixed communication infrastructure, a dynamic scenario is considered where a flying UAV provides wireless services for multiple ground nodes simultaneously. To fully exploit the controllable channel variations provide...
Interference Management for Cellular-Connected UAVs: A Deep Reinforcement Learning Approach In this paper, an interference-aware path planning scheme for a network of cellular-connected unmanned aerial vehicles (UAVs) is proposed. In particular, each UAV aims at achieving a tradeoff between maximizing energy efficiency and minimizing both wireless latency and the interference caused on the ground network along its path. The problem is cast as a dynamic game among UAVs. To solve this game, a deep reinforcement learning algorithm, based on echo state network (ESN) cells, is proposed. The introduced deep ESN architecture is trained to allow each UAV to map each observation of the network state to an action, with the goal of minimizing a sequence of time-dependent utility functions. Each UAV uses the ESN to learn its optimal path, transmission power, and cell association vector at different locations along its path. The proposed algorithm is shown to reach a subgame perfect Nash equilibrium upon convergence. Moreover, an upper bound and a lower bound for the altitude of the UAVs are derived thus reducing the computational complexity of the proposed algorithm. The simulation results show that the proposed scheme achieves better wireless latency per UAV and rate per ground user (UE) while requiring a number of steps that are comparable to a heuristic baseline that considers moving via the shortest distance toward the corresponding destinations. The results also show that the optimal altitude of the UAVs varies based on the ground network density and the UE data rate requirements and plays a vital role in minimizing the interference level on the ground UEs as well as the wireless transmission delay of the UAV.
Energy-Efficient Data Collection and Device Positioning in UAV-Assisted IoT. The Internet of Things (IoT) will significantly change both industrial manufacturing and our daily lives. Data collection and 3-D positioning of IoT devices are two indispensable services of such networks. However, in conventional networks, only terrestrial base stations (BSs) are used to provide these two services. On the one hand, this leads to high energy consumption for devices transmitting at...
Fuzzy logic in control systems: fuzzy logic controller. I.
An introduction to ROC analysis Receiver operating characteristics (ROC) graphs are useful for organizing classifiers and visualizing their performance. ROC graphs are commonly used in medical decision making, and in recent years have been used increasingly in machine learning and data mining research. Although ROC graphs are apparently simple, there are some common misconceptions and pitfalls when using them in practice. The purpose of this article is to serve as an introduction to ROC graphs and as a guide for using them in research.
Toward Massive Machine Type Communications in Ultra-Dense Cellular IoT Networks: Current Issues and Machine Learning-Assisted Solutions The ever-increasing number of resource-constrained machine-type communication (MTC) devices is leading to the critical challenge of fulfilling diverse communication requirements in dynamic and ultra-dense wireless environments. Among different application scenarios that the upcoming 5G and beyond cellular networks are expected to support, such as enhanced mobile broadband (eMBB), massive machine type communications (mMTCs), and ultra-reliable and low latency communications (URLLCs), the mMTC brings the unique technical challenge of supporting a huge number of MTC devices in cellular networks, which is the main focus of this paper. The related challenges include quality of service (QoS) provisioning, handling highly dynamic and sporadic MTC traffic, huge signalling overhead, and radio access network (RAN) congestion. In this regard, this paper aims to identify and analyze the involved technical issues, to review recent advances, to highlight potential solutions and to propose new research directions. First, starting with an overview of mMTC features and QoS provisioning issues, we present the key enablers for mMTC in cellular networks. Along with the highlights on the inefficiency of the legacy random access (RA) procedure in the mMTC scenario, we then present the key features and channel access mechanisms in the emerging cellular IoT standards, namely, LTE-M and narrowband IoT (NB-IoT). Subsequently, we present a framework for the performance analysis of transmission scheduling with the QoS support along with the issues involved in short data packet transmission. Next, we provide a detailed overview of the existing and emerging solutions toward addressing RAN congestion problem, and then identify potential advantages, challenges, and use cases for the applications of emerging machine learning (ML) techniques in ultra-dense cellular networks. Out of several ML techniques, we focus on the application of low-complexity <inline-formula xmlns:mml="http://www.w3.org/1998/Math/MathML" xmlns:xlink="http://www.w3.org/1999/xlink"> <tex-math notation="LaTeX">$Q$ </tex-math></inline-formula> -learning approach in the mMTC scenario along with the recent advances toward enhancing its learning performance and convergence. Finally, we discuss some open research challenges and promising future research directions.
Efficient k-out-of-n oblivious transfer schemes with adaptive and non-adaptive queries In this paper we propose efficient two-round k-out-of-n oblivious transfer schemes, in which R sends O(k) messages to S, and S sends O(n) messages back to R. The computation cost of R and S is reasonable. The choices of R are unconditionally secure. For the basic scheme, the secrecy of unchosen messages is guaranteed if the Decisional Diffie-Hellman problem is hard. When k=1, our basic scheme is as efficient as the most efficient 1-out-of-n oblivious transfer scheme. Our schemes have the nice property of universal parameters, that is each pair of R and S need neither hold any secret key nor perform any prior setup (initialization). The system parameters can be used by all senders and receivers without any trapdoor specification. Our k-out-of-n oblivious transfer schemes are the most efficient ones in terms of the communication cost, in both rounds and the number of messages. Moreover, one of our schemes can be extended in a straightforward way to an adaptivek-out-of-n oblivious transfer scheme, which allows the receiver R to choose the messages one by one adaptively. In our adaptive-query scheme, S sends O(n) messages to R in one round in the commitment phase. For each query of R, only O(1) messages are exchanged and O(1) operations are performed. In fact, the number k of queries need not be pre-fixed or known beforehand. This makes our scheme highly flexible.
Minimum acceleration criterion with constraints implies bang-bang control as an underlying principle for optimal trajectories of arm reaching movements. Rapid arm-reaching movements serve as an excellent test bed for any theory about trajectory formation. How are these movements planned? A minimum acceleration criterion has been examined in the past, and the solution obtained, based on the Euler-Poisson equation, failed to predict that the hand would begin and end the movement at rest (i.e., with zero acceleration). Therefore, this criterion was rejected in favor of the minimum jerk, which was proved to be successful in describing many features of human movements. This letter follows an alternative approach and solves the minimum acceleration problem with constraints using Pontryagin's minimum principle. We use the minimum principle to obtain minimum acceleration trajectories and use the jerk as a control signal. In order to find a solution that does not include nonphysiological impulse functions, constraints on the maximum and minimum jerk values are assumed. The analytical solution provides a three-phase piecewise constant jerk signal (bang-bang control) where the magnitude of the jerk and the two switching times depend on the magnitude of the maximum and minimum available jerk values. This result fits the observed trajectories of reaching movements and takes into account both the extrinsic coordinates and the muscle limitations in a single framework. The minimum acceleration with constraints principle is discussed as a unifying approach for many observations about the neural control of movements.
An Automatic Screening Approach for Obstructive Sleep Apnea Diagnosis Based on Single-Lead Electrocardiogram Traditional approaches for obstructive sleep apnea (OSA) diagnosis are apt to using multiple channels of physiological signals to detect apnea events by dividing the signals into equal-length segments, which may lead to incorrect apnea event detection and weaken the performance of OSA diagnosis. This paper proposes an automatic-segmentation-based screening approach with the single channel of Electrocardiogram (ECG) signal for OSA subject diagnosis, and the main work of the proposed approach lies in three aspects: (i) an automatic signal segmentation algorithm is adopted for signal segmentation instead of the equal-length segmentation rule; (ii) a local median filter is improved for reduction of the unexpected RR intervals before signal segmentation; (iii) the designed OSA severity index and additional admission information of OSA suspects are plugged into support vector machine (SVM) for OSA subject diagnosis. A real clinical example from PhysioNet database is provided to validate the proposed approach and an average accuracy of 97.41% for subject diagnosis is obtained which demonstrates the effectiveness for OSA diagnosis.
Modeling taxi driver anticipatory behavior. As part of a wider behavioral agent-based model that simulates taxi drivers' dynamic passenger-finding behavior under uncertainty, we present a model of strategic behavior of taxi drivers in anticipation of substantial time varying demand at locations such as airports and major train stations. The model assumes that, considering a particular decision horizon, a taxi driver decides to transfer to such a destination based on a reward function. The dynamic uncertainty of demand is captured by a time dependent pick-up probability, which is a cumulative distribution function of waiting time. The model allows for information learning by which taxi drivers update their beliefs from past experiences. A simulation on a real road network, applied to test the model, indicates that the formulated model dynamically improves passenger-finding strategies at the airport. Taxi drivers learn when to transfer to the airport in anticipation of the time-varying demand at the airport to minimize their waiting time.
Ethical Considerations Of Applying Robots In Kindergarten Settings: Towards An Approach From A Macroperspective In child-robot interaction (cHRI) research, many studies pursue the goal to develop interactive systems that can be applied in everyday settings. For early education, increasingly, the setting of a kindergarten is targeted. However, when cHRI and research are brought into a kindergarten, a range of ethical and related procedural aspects have to be considered and dealt with. While ethical models elaborated within other human-robot interaction settings, e.g., assisted living contexts, can provide some important indicators for relevant issues, we argue that it is important to start developing a systematic approach to identify and tackle those ethical issues which rise with cHRI in kindergarten settings on a more global level and address the impact of the technology from a macroperspective beyond the effects on the individual. Based on our experience in conducting studies with children in general and pedagogical considerations on the role of the institution of kindergarten in specific, in this paper, we enfold some relevant aspects that have barely been addressed in an explicit way in current cHRI research. Four areas are analyzed and key ethical issues are identified in each area: (1) the institutional setting of a kindergarten, (2) children as a vulnerable group, (3) the caregivers' role, and (4) pedagogical concepts. With our considerations, we aim at (i) broadening the methodology of the current studies within the area of cHRI, (ii) revalidate it based on our comprehensive empirical experience with research in kindergarten settings, both laboratory and real-world contexts, and (iii) provide a framework for the development of a more systematic approach to address the ethical issues in cHRI research within kindergarten settings.
1.24
0.24
0.24
0.12
0.08
0.048
0
0
0
0
0
0
0
0
A Joint Energy Replenishment and Data Collection Algorithm in Wireless Rechargeable Sensor Networks. Energy constraint is a critical issue in the development of wireless sensor networks (WSNs) because sensor nodes are generally powered by batteries. Recently, wireless rechargeable sensor networks (WRSNs), which introduce wireless mobile chargers (MCs) to replenish energy for nodes, have been proposed to resolve the root cause of energy limitations in WSNs. However, existing wireless charging algo...
IoT Elements, Layered Architectures and Security Issues: A Comprehensive Survey. The use of the Internet is growing in this day and age, so another area has developed to use the Internet, called Internet of Things (IoT). It facilitates the machines and objects to communicate, compute and coordinate with each other. It is an enabler for the intelligence affixed to several essential features of the modern world, such as homes, hospitals, buildings, transports and cities. The security and privacy are some of the critical issues related to the wide application of IoT. Therefore, these issues prevent the wide adoption of the IoT. In this paper, we are presenting an overview about different layered architectures of IoT and attacks regarding security from the perspective of layers. In addition, a review of mechanisms that provide solutions to these issues is presented with their limitations. Furthermore, we have suggested a new secure layered architecture of IoT to overcome these issues.
A Multicharger Cooperative Energy Provision Algorithm Based On Density Clustering In The Industrial Internet Of Things Wireless sensor networks (WSNs) are an important core of the Industrial Internet of Things (IIoT). Wireless rechargeable sensor networks (WRSNs) are sensor networks that are charged by mobile chargers (MCs), and can achieve self-sufficiency. Therefore, the development of WRSNs has begun to attract widespread attention in recent years. Most of the existing energy replenishment algorithms for MCs use one or more MCs to serve the whole network in WRSNs. However, a single MC is not suitable for large-scale network environments, and multiple MCs make the network cost too high. Thus, this paper proposes a collaborative charging algorithm based on network density clustering (CCA-NDC) in WRSNs. This algorithm uses the mean-shift algorithm based on density to cluster, and then the mother wireless charger vehicle (MWCV) carries multiple sub wireless charger vehicles (SWCVs) to charge the nodes in each cluster by using a gradient descent optimization algorithm. The experimental results confirm that the proposed algorithm can effectively replenish the energy of the network and make the network more stable.
Dynamic Charging Scheme Problem With Actor–Critic Reinforcement Learning The energy problem is one of the most important challenges in the application of sensor networks. With the development of wireless charging technology and intelligent mobile charger (MC), the energy problem can be solved by the wireless charging strategy. In the practical application of wireless rechargeable sensor networks (WRSNs), the energy consumption rate of nodes is dynamically changed due to many uncertainties, such as the death and different transmission tasks of sensor nodes. However, existing works focus on on-demand schemes, which not fully consider real-time global charging scheduling. In this article, a novel dynamic charging scheme (DCS) in WRSN based on the actor-critic reinforcement learning (ACRL) algorithm is proposed. In the ACRL, we introduce gated recurrent units (GRUs) to capture the relationships of charging actions in time sequence. Using the actor network with one GRU layer, we can pick up an optimal or near-optimal sensor node from candidates as the next charging target more quickly and speed up the training of the model. Meanwhile, we take the tour length and the number of dead nodes as the reward signal. Actor and critic networks are updated by the error criterion function of R and V. Compared with current on-demand charging scheduling algorithms, extensive simulations show that the proposed ACRL algorithm surpasses heuristic algorithms, such as the Greedy, DP, nearest job next with preemption, and TSCA in the average lifetime and tour length, especially against the size and complexity increasing of WRSNs.
Adaptive Wireless Power Transfer in Mobile Ad Hoc Networks. We investigate the interesting impact of mobility on the problem of efficient wireless power transfer in ad hoc networks. We consider a set of mobile agents (consuming energy to perform certain sensing and communication tasks), and a single static charger (with finite energy) which can recharge the agents when they get in its range. In particular, we focus on the problem of efficiently computing the appropriate range of the charger with the goal of prolonging the network lifetime. We first demonstrate (under the realistic assumption of fixed energy supplies) the limitations of any fixed charging range and, therefore, the need for (and power of) a dynamic selection of the charging range, by adapting to the behavior of the mobile agents which is revealed in an online manner. We investigate the complexity of optimizing the selection of such an adaptive charging range, by showing that two simplified offline optimization problems (closely related to the online one) are NP-hard. To effectively address the involved performance trade-offs, we finally present a variety of adaptive heuristics, assuming different levels of agent information regarding their mobility and energy.
Improving charging capacity for wireless sensor networks by deploying one mobile vehicle with multiple removable chargers. Wireless energy transfer is a promising technology to prolong the lifetime of wireless sensor networks (WSNs), by employing charging vehicles to replenish energy to lifetime-critical sensors. Existing studies on sensor charging assumed that one or multiple charging vehicles being deployed. Such an assumption may have its limitation for a real sensor network. On one hand, it usually is insufficient to employ just one vehicle to charge many sensors in a large-scale sensor network due to the limited charging capacity of the vehicle or energy expirations of some sensors prior to the arrival of the charging vehicle. On the other hand, although the employment of multiple vehicles can significantly improve the charging capability, it is too costly in terms of the initial investment and maintenance costs on these vehicles. In this paper, we propose a novel charging model that a charging vehicle can carry multiple low-cost removable chargers and each charger is powered by a portable high-volume battery. When there are energy-critical sensors to be charged, the vehicle can carry the chargers to charge multiple sensors simultaneously, by placing one portable charger in the vicinity of one sensor. Under this novel charging model, we study the scheduling problem of the charging vehicle so that both the dead duration of sensors and the total travel distance of the mobile vehicle per tour are minimized. Since this problem is NP-hard, we instead propose a (3+ϵ)-approximation algorithm if the residual lifetime of each sensor can be ignored; otherwise, we devise a novel heuristic algorithm, where ϵ is a given constant with 0 < ϵ ≤ 1. Finally, we evaluate the performance of the proposed algorithms through experimental simulations. Experimental results show that the performance of the proposed algorithms are very promising.
Peer-to-peer energy sharing in mobile networks: Applications, challenges, and open problems. Energy is a scarce resource in mobile wireless networks that consist of devices mainly powered by their batteries. Providing energy ubiquitously to these devices for making them functional for a long time is a challenging task. With the advent of energy sharing techniques, either by wired or wireless mediums, it is possible to extend the lifetime of such networks by utilizing the energy from other energy sources (e.g., chargers, other devices) within the network. In this paper, we explore the utilization of peer-to-peer energy sharing in various applications of mobile networks that consist of agents from low-power devices such as sensors to high-power ones such as electric vehicles. We provide an overview of the current research directions and developments of new protocols and algorithms that exploit the energy sharing techniques to solve the scarce energy problem in mobile networks. For each mobile networking domain, we highlight the specific challenges and describe the approaches followed to address them under various research problems. We also discuss open problems yet to be solved in each specific application.
A Periodic and Distributed Energy Supplement Method Based on Maximum Recharging Benefit in Sensor Networks The issue of using vehicles to wirelessly recharge nodes for energy supplement in wireless sensor networks has become a research hotspot in recent works. Unfortunately, most of the researches did not consider the rationality of the recharging request threshold (RRT) and also overlooked the difference of node's power consumption, which may lead to the premature death of nodes as well as low efficiency of wireless charging vehicles (WCVs). In order to solve the above problems, a periodic and distributed energy supplement method based on maximum recharging benefit (PDESM) is proposed in this article. First, to avoid frequent recharging requests from nodes, we put forward an annuluses-based cost-balanced data uploading strategy under deterministic deployment. Then, one WCV in each annulus periodically selects and recharges nodes located in this region which sends the energy supplement requests. In addition, the predicted values of power consumption of nodes are calculated out according to the real-time energy consumption rate, and thus the most appropriate RRT is obtained. Finally, a moving path optimization scheme based on the minimum spanning tree is constructed for distributed recharging. Simulation results show that PDESM performs well on enhancing the proportion of the alive nodes as well as the wireless recharging efficiency compared with node failure avoidance online charging and first come first served. Moreover, it also has an advantage in balancing the energy consumption of WCVs.
Efficient on-demand multi-node charging techniques for wireless sensor networks. This paper deals with wireless charging in sensor networks and explores efficient policies to perform simultaneous multi-node power transfer through a mobile charger (MC).The proposed solution, called On-demand Multi-node Charging (OMC), features an original threshold-based tour launching (TTL) strategy, using request grouping, and a path planning algorithm based on minimizing the number of stopping points in the charging tour. Contrary to existing solutions, which focus on shortening the charging delays, OMC groups incoming charging requests and optimizes the charging tour and the mobile charger energy consumption. Although slightly increasing the waiting time before nodes are charged, this allows taking advantage of multiple simultaneous charges and also reduces node failures. At the tour planning level, a new modeling approach is used. It leverages simultaneous energy transfer to multiple nodes by maximizing the number of sensors that are charged at each stop. Given its NP-hardness, tour planning is approximated through a clique partitioning problem, which is solved using a lightweight heuristic approach. The proposed schemes are evaluated in offline and on-demand scenarios and compared against relevant state-of-the-art protocols. The results in the offline scenario show that the path planning strategy reduces the number of stops and the energy consumed by the mobile charger, compared to existing offline solutions. This is with further reduction in time complexity, due to the simple heuristics that are used. The results in the on-demand scenario confirm the effectiveness of the path planning strategy. More importantly, they show the impact of path planning, TTL and multi-node charging on the efficiency of handling the requests, in a way that reduces node failures and the mobile charger energy expenditure.
Deep Residual Learning for Image Recognition Deeper neural networks are more difficult to train. We present a residual learning framework to ease the training of networks that are substantially deeper than those used previously. We explicitly reformulate the layers as learning residual functions with reference to the layer inputs, instead of learning unreferenced functions. We provide comprehensive empirical evidence showing that these residual networks are easier to optimize, and can gain accuracy from considerably increased depth. On the ImageNet dataset we evaluate residual nets with a depth of up to 152 layers - 8× deeper than VGG nets [40] but still having lower complexity. An ensemble of these residual nets achieves 3.57% error on the ImageNet test set. This result won the 1st place on the ILSVRC 2015 classification task. We also present analysis on CIFAR-10 with 100 and 1000 layers. The depth of representations is of central importance for many visual recognition tasks. Solely due to our extremely deep representations, we obtain a 28% relative improvement on the COCO object detection dataset. Deep residual nets are foundations of our submissions to ILSVRC & COCO 2015 competitions1, where we also won the 1st places on the tasks of ImageNet detection, ImageNet localization, COCO detection, and COCO segmentation.
Stability of Time-Delay Feedback Switched Linear Systems We address stability of state feedback switched linear systems in which delays are present in both the feedback state and the switching signal of the switched controller. For switched systems with average dwell-time switching signals, we provide a condition, in terms of upper bounds on the delays and in terms of a lower bound on the average dwell-time, to guarantee asymptotic stability of the closed loop. The condition also implies that, in general, feedback switched linear systems are robust with respect to both small state delays and small switching delays. Our approach combines existing multiple Lyapunov function techniques with the merging switching signal technique, which gives relationships between the average dwell times of two mismatched switching signals and their mismatched times. A methodology for numerical solution based on linear matrix inequality is also included.
Cooperative social robots to accompany groups of people This study proposes a new model for guiding people in urban settings using multiple robots that work cooperatively. More specifically, this investigation describes the circumstances in which people might stray from the formation when following different robots' instructions. To this end, we introduce a 'prediction and anticipation model' that predicts the position of the group using a particle filter, while determining the optimal robot behavior to help people stay in the group in areas where they may become distracted. As a result, this article presents a novel approach to locally optimizing the work performed by robots and people using the minimum robots' work criterion and determining human-friendly types of movements. The guidance missions were carried out in urban areas that included multiple conflict areas and obstacles. This study also provides an analysis of robots' behavioral reactions to people by simulating different situations in the locations that were used for the investigation. The method was tested through simulations that took into account the difficulties and technological constraints derived from real-life situations. Despite these problematic issues, we were able to demonstrate the robots' effect on people in real-life situations in terms of pushing and dragging forces.
Training and Testing Object Detectors with Virtual Images. In the area of computer vision, deep learning has produced a variety of state-of-the-art models that rely on massive labeled data. However, collecting and annotating images from the real world is too demanding in terms of labor and money investments, and is usually inflexible to build datasets with specific characteristics, such as small area of objects and high occlusion level. Under the framewor...
Methods for Flexible Management of Blockchain-Based Cryptocurrencies in Electricity Markets and Smart Grids The growing trend in the use of blockchain-based cryptocurrencies in modern communities provides several advantages, but also imposes several challenges to energy markets and power systems, in general. This paper aims at providing recommendations for efficient use of digital cryptocurrencies in today's and future smart power systems, in order to face the challenging aspects of this new technology. In this paper, existing issues and challenges of smart grids in the presence of blockchain-based cryptocurrencies are presented and some innovative approaches for efficient integration and management of blockchain-based cryptocurrencies in smart grids are proposed. Also some recommendations are given for improving the smart grids performance in the presence of digital cryptocurrencies and some future research directions are highlighted.
1.073556
0.08
0.08
0.08
0.08
0.066667
0.066667
0.048889
0.017167
0
0
0
0
0
Urban Traffic Flow Prediction Using a Spatio-Temporal Random Effects Model Traffic prediction is critical for the success of intelligent transportation systems (ITS). However, most spatio-temporal models suffer from high mathematical complexity and low tune-up flexibility. This article presents a novel spatio-temporal random effects (STRE) model that has a reduced computational complexity due to mathematical dimension reduction, with additional tune-up flexibility provided by a basis function capable of taking traffic patterns into account. Bellevue, WA, was selected as the model test site due to its widespread deployment of loop detectors. Data collected during the 2 weeks of July 2007 from 105 detectors in the downtown area were used in the modeling process and traffic volumes predicted for 14 detectors for the entire month of July 2008. The results show that the STRE model not only effectively predicts traffic volume but also outperforms three well-established volume prediction models, the enhanced versions of autoregressive moving average (ARMA) and spatiotemporal ARMA, and artificial neural network. Even without further model tuning, all the experimental links produced mean absolute percentage errors between 8% and 16% except for three atypical locations. Based on lessons learned, recommendations are provided for future applications and tune-up of the proposed STRE model.
Applied research of data sensing and service to ubiquitous intelligent transportation system High-efficiency transportation systems in urban environments are not only solutions for the growing public travel demands, but are also the premise for enlarging transportation capacity and narrowing the gap between urban and rural areas. Such transportation systems should have characteristics such as mobility, convenience and being accident-free. Ubiquitous-intelligent transportation systems (U-ITS) are next generation of intelligent transportation system (ITS). The key issue of U-ITS is providing better and more efficient services by providing vehicle to vehicle (V2V) or vehicle to infrastructure (V2I) interconnection. The emergence of cyber physical systems (CPS), which focus on information awareness technologies, provides technical assurance for the rapid development of U-ITS. This paper introduces the ongoing Beijing U-ITS project, which utilizes mobile sensors. Realization of universal interconnection between real-time information systems and large-scale detectors allows the system to maximize equipment efficiency and improve transportation efficiency through information services.
Catch Me If You Can: Detecting Pickpocket Suspects from Large-Scale Transit Records Massive data collected by automated fare collection (AFC) systems provide opportunities for studying both personal traveling behaviors and collective mobility patterns in the urban area. Existing studies on the AFC data have primarily focused on identifying passengers' movement patterns. In this paper, however, we creatively leveraged such data for identifying thieves in the public transit systems. Indeed, stopping pickpockets in the public transit systems has been critical for improving passenger satisfaction and public safety. However, it is challenging to tell thieves from regular passengers in practice. To this end, we developed a suspect detection and surveillance system, which can identify pick-pocket suspects based on their daily transit records. Specifically, we first extracted a number of features from each passenger's daily activities in the transit systems. Then, we took a two-step approach that exploits the strengths of unsupervised outlier detection and supervised classification models to identify thieves, who exhibit abnormal traveling behaviors. Experimental results demonstrated the effective- ness of our method. We also developed a prototype system with a user-friendly interface for the security personnel.
Computing Urban Traffic Congestions by Incorporating Sparse GPS Probe Data and Social Media Data Estimating urban traffic conditions of an arterial network with GPS probe data is a practically important while substantially challenging problem, and has attracted increasing research interests recently. Although GPS probe data is becoming a ubiquitous data source for various traffic related applications currently, they are usually insufficient for fully estimating traffic conditions of a large arterial network due to the low sampling frequency. To explore other data sources for more effectively computing urban traffic conditions, we propose to collect various traffic events such as traffic accident and jam from social media as complementary information. In addition, to further explore other factors that might affect traffic conditions, we also extract rich auxiliary information including social events, road features, Point of Interest (POI), and weather. With the enriched traffic data and auxiliary information collected from different sources, we first study the traffic co-congestion pattern mining problem with the aim of discovering which road segments geographically close to each other are likely to co-occur traffic congestion. A search tree based approach is proposed to efficiently discover the co-congestion patterns. These patterns are then used to help estimate traffic congestions and detect anomalies in a transportation network. To fuse the multisourced data, we finally propose a coupled matrix and tensor factorization model named TCE_R to more accurately complete the sparse traffic congestion matrix by collaboratively factorizing it with other matrices and tensors formed by other data. We evaluate the proposed model on the arterial network of downtown Chicago with 1,257 road segments whose total length is nearly 700 miles. The results demonstrate the superior performance of TCE_R by comprehensive comparison with existing approaches.
Dynamic Spatio-temporal Graph-based CNNs for Traffic Prediction. Accurate traffic forecast is a challenging problem due to the large-scale problem size, as well as the complex and dynamic nature of spatio-temporal dependency of traffic flow. Most existing graph-based CNNs attempt to capture the static relations while largely neglecting the dynamics underlying sequential data. In this paper, we present dynamic spatio-temporal graph-based CNNs (DST-GCNNs) by learning expressive features to represent spatio-temporal structures and predict future traffic from historical traffic flow. In particular, DST-GCNN is a two stream network. In the flow prediction stream, we present a novel graph-based spatio-temporal convolutional layer to extract features from a graph representation of traffic flow. Then several such layers are stacked together to predict future traffic over time. Meanwhile, the proximity relations between nodes in the graph are often time variant as the traffic condition changes over time. To capture the graph dynamics, we use the graph prediction stream to predict the dynamic graph structures, and the predicted structures are fed into the flow prediction stream. Experiments on real traffic datasets demonstrate that the proposed model achieves competitive performances compared with the other state-of-the-art methods.
Deep Multi-View Spatial-Temporal Network for Taxi Demand Prediction. Taxi demand prediction is an important building block to enabling intelligent transportation systems in a smart city. An accurate prediction model can help the city pre-allocate resources to meet travel demand and to reduce empty taxis on streets which waste energy and worsen the traffic congestion. With the increasing popularity of taxi requesting services such as Uber and Didi Chuxing (in China), we are able to collect large-scale taxi demand data continuously. How to utilize such big data to improve the demand prediction is an interesting and critical real-world problem. Traditional demand prediction methods mostly rely on time series forecasting techniques, which fail to model the complex non-linear spatial and temporal relations. Recent advances in deep learning have shown superior performance on traditionally challenging tasks such as image classification by learning the complex features and correlations from large-scale data. This breakthrough has inspired researchers to explore deep learning techniques on traffic prediction problems. However, existing methods on traffic prediction have only considered spatial relation (e.g., using CNN) or temporal relation (e.g., using LSTM) independently. We propose a Deep Multi-View Spatial-Temporal Network (DMVST-Net) framework to model both spatial and temporal relations. Specifically, our proposed model consists of three views: temporal view (modeling correlations between future demand values with near time points via LSTM), spatial view (modeling local spatial correlation via local CNN), and semantic view (modeling correlations among regions sharing similar temporal patterns). Experiments on large-scale real taxi demand data demonstrate effectiveness of our approach over state-of-the-art methods.
Heterogeneous ensemble for feature drifts in data streams The nature of data streams requires classification algorithms to be real-time, efficient, and able to cope with high-dimensional data that are continuously arriving. It is a known fact that in high-dimensional datasets, not all features are critical for training a classifier. To improve the performance of data stream classification, we propose an algorithm called HEFT-Stream (H eterogeneous E nsemble with F eature drifT for Data Streams ) that incorporates feature selection into a heterogeneous ensemble to adapt to different types of concept drifts. As an example of the proposed framework, we first modify the FCBF [13] algorithm so that it dynamically update the relevant feature subsets for data streams. Next, a heterogeneous ensemble is constructed based on different online classifiers, including Online Naive Bayes and CVFDT [5]. Empirical results show that our ensemble classifier outperforms state-of-the-art ensemble classifiers (AWE [15] and OnlineBagging [21]) in terms of accuracy, speed, and scalability. The success of HEFT-Stream opens new research directions in understanding the relationship between feature selection techniques and ensemble learning to achieve better classification performance.
On the Graph Fourier Transform for Directed Graphs. The analysis of signals defined over a graph is relevant in many applications, such as social and economic networks, big data or biological networks, and so on. A key tool for analyzing these signals is the so-called graph Fourier transform (GFT). Alternative definitions of GFT have been suggested in the literature, based on the eigen-decomposition of either the graph Laplacian or adjacency matrix...
A survey of socially interactive robots This paper reviews “socially interactive robots”: robots for which social human–robot interaction is important. We begin by discussing the context for socially interactive robots, emphasizing the relationship to other research fields and the different forms of “social robots”. We then present a taxonomy of design methods and system components used to build socially interactive robots. Finally, we describe the impact of these robots on humans and discuss open issues. An expanded version of this paper, which contains a survey and taxonomy of current applications, is available as a technical report [T. Fong, I. Nourbakhsh, K. Dautenhahn, A survey of socially interactive robots: concepts, design and applications, Technical Report No. CMU-RI-TR-02-29, Robotics Institute, Carnegie Mellon University, 2002].
Affective social robots For human-robot interaction to proceed in a smooth, natural manner, robots must adhere to human social norms. One such human convention is the use of expressive moods and emotions as an integral part of social interaction. Such expressions are used to convey messages such as ''I'm happy to see you'' or ''I want to be comforted,'' and people's long-term relationships depend heavily on shared emotional experiences. Thus, we have developed an affective model for social robots. This generative model attempts to create natural, human-like affect and includes distinctions between immediate emotional responses, the overall mood of the robot, and long-term attitudes toward each visitor to the robot, with a focus on developing long-term human-robot relationships. This paper presents the general affect model as well as particular details of our implementation of the model on one robot, the Roboceptionist. In addition, we present findings from two studies that demonstrate the model's potential.
Millimeter Wave Cellular Wireless Networks: Potentials and Challenges. Millimeter-wave (mmW) frequencies between 30 and 300 GHz are a new frontier for cellular communication that offers the promise of orders of magnitude greater bandwidths combined with further gains via beamforming and spatial multiplexing from multielement antenna arrays. This paper surveys measurements and capacity studies to assess this technology with a focus on small cell deployments in urban e...
Micro aerial vehicle networks: an experimental analysis of challenges and opportunities The need for aerial networks is growing with the recent advance of micro aerial vehicles, which enable a wide range of civilian applications. Our experimental analysis shows that wireless connectivity among MAVs is challenged by the mobility and heterogeneity of the nodes, lightweight antenna design, body blockage, constrained embedded resources, and limited battery power. However, the movement and location of MAVs are known and may be controlled to establish wireless links with the best transmission opportunities in time and space. This special ecosystem undoubtedly requires a rethinking of wireless communications and calls for novel networking approaches. Supported by empirical results, we identify important research questions, and introduce potential solutions and directions for investigation.
Gender Bias in Coreference Resolution. We present an empirical study of gender bias in coreference resolution systems. We first introduce a novel, Winograd schema-style set of minimal pair sentences that differ only by pronoun gender. With these Winogender schemas, we evaluate and confirm systematic gender bias in three publicly-available coreference resolution systems, and correlate this bias with real-world and textual gender statistics.
Robot tutor and pupils’ educational ability: Teaching the times tables Research shows promising results of educational robots in language and STEM tasks. In language, more research is available, occasionally in view of individual differences in pupils’ educational ability levels, and learning seems to improve with more expressive robot behaviors. In STEM, variations in robots’ behaviors have been examined with inconclusive results and never while systematically investigating how differences in educational abilities match with different robot behaviors. We applied an autonomously tutoring robot (without tablet, partly WOz) in a 2 × 2 experiment of social vs. neutral behavior in above-average vs. below-average schoolchildren (N = 86; age 8–10 years) while rehearsing the multiplication tables on a one-to-one basis. The standard school test showed that on average, pupils significantly improved their performance even after 3 occasions of 5-min exercises. Beyond-average pupils profited most from a robot tutor, whereas those below average in multiplication benefited more from a robot that showed neutral rather than more social behavior.
1.078519
0.068333
0.068333
0.068333
0.039167
0.022778
0.006667
0.000556
0
0
0
0
0
0
A Dual Fuzzy-Enhanced Neurodynamic Scheme for Model-Less Kinematic Control of Redundant and Hyperredundant Robots Tracking control of redundant and hyperredundant manipulators is a fundamental and critical problem in practical applications. In order to effectively decrease the end-effector position errors, a novel dual fuzzy-enhanced neurodynamic (DFEN) scheme is put forward for solving the position error accumulation problem followed by achieving accurate tracking control results. The proposed scheme is established based on a zeroing neurodynamic approach in conjunction with two fuzzy adjustment units that are capable of tuning the control parameters by monitoring the tracking error. Moreover, the DFEN scheme can effectively solve the tracking problem without requiring knowing <italic xmlns:mml="http://www.w3.org/1998/Math/MathML" xmlns:xlink="http://www.w3.org/1999/xlink">a priori</i> knowledge of the kinematic model of the robot. The convergence and the stability of the proposed approach are demonstrated by theoretical analysis. The effectiveness, accuracy, and robustness of the proposed DFEN scheme are verified on the simulative redundant manipulator, continuum robot, and hybrid robot (integrating the redundant manipulator and the continuum robot). A practical experiment is provided to validate the proposed scheme as well.
The Sybil Attack Large-scale peer-to-peer systems facesecurity threats from faulty or hostile remotecomputing elements. To resist these threats, manysuch systems employ redundancy. However, if asingle faulty entity can present multiple identities,it can control a substantial fraction of the system,thereby undermining this redundancy. Oneapproach to preventing these &quot;Sybil attacks&quot; is tohave a trusted agency certify identities. Thispaper shows that, without a logically centralizedauthority, Sybil...
BLEU: a method for automatic evaluation of machine translation Human evaluations of machine translation are extensive but expensive. Human evaluations can take months to finish and involve human labor that can not be reused. We propose a method of automatic machine translation evaluation that is quick, inexpensive, and language-independent, that correlates highly with human evaluation, and that has little marginal cost per run. We present this method as an automated understudy to skilled human judges which substitutes for them when there is need for quick or frequent evaluations.
Computational thinking Summary form only given. My vision for the 21st century, Computational Thinking, will be a fundamental skill used by everyone in the world. To reading, writing, and arithmetic, we should add computational thinking to every child's analytical ability. Computational thinking involves solving problems, designing systems, and understanding human behavior by drawing on the concepts fundamental to computer science. Thinking like a computer scientist means more than being able to program a computer. It requires the ability to abstract and thus to think at multiple levels of abstraction. In this talk I will give many examples of computational thinking, argue that it has already influenced other disciplines, and promote the idea that teaching computational thinking can not only inspire future generations to enter the field of computer science but benefit people in all fields.
Fuzzy logic in control systems: fuzzy logic controller. I.
Switching between stabilizing controllers This paper deals with the problem of switching between several linear time-invariant (LTI) controllers—all of them capable of stabilizing a speci4c LTI process—in such a way that the stability of the closed-loop system is guaranteed for any switching sequence. We show that it is possible to 4nd realizations for any given family of controller transfer matrices so that the closed-loop system remains stable, no matter how we switch among the controller. The motivation for this problem is the control of complex systems where con8icting requirements make a single LTI controller unsuitable. ? 2002 Published by Elsevier Science Ltd.
Tabu Search - Part I
Bidirectional recurrent neural networks In the first part of this paper, a regular recurrent neural network (RNN) is extended to a bidirectional recurrent neural network (BRNN). The BRNN can be trained without the limitation of using input information just up to a preset future frame. This is accomplished by training it simultaneously in positive and negative time direction. Structure and training procedure of the proposed network are explained. In regression and classification experiments on artificial data, the proposed structure gives better results than other approaches. For real data, classification experiments for phonemes from the TIMIT database show the same tendency. In the second part of this paper, it is shown how the proposed bidirectional structure can be easily modified to allow efficient estimation of the conditional posterior probability of complete symbol sequences without making any explicit assumption about the shape of the distribution. For this part, experiments on real data are reported
An intensive survey of fair non-repudiation protocols With the phenomenal growth of the Internet and open networks in general, security services, such as non-repudiation, become crucial to many applications. Non-repudiation services must ensure that when Alice sends some information to Bob over a network, neither Alice nor Bob can deny having participated in a part or the whole of this communication. Therefore a fair non-repudiation protocol has to generate non-repudiation of origin evidences intended to Bob, and non-repudiation of receipt evidences destined to Alice. In this paper, we clearly define the properties a fair non-repudiation protocol must respect, and give a survey of the most important non-repudiation protocols without and with trusted third party (TTP). For the later ones we discuss the evolution of the TTP's involvement and, between others, describe the most recent protocol using a transparent TTP. We also discuss some ad-hoc problems related to the management of non-repudiation evidences.
Dynamic movement and positioning of embodied agents in multiparty conversations For embodied agents to engage in realistic multiparty conversation, they must stand in appropriate places with respect to other agents and the environment. When these factors change, such as an agent joining the conversation, the agents must dynamically move to a new location and/or orientation to accommodate. This paper presents an algorithm for simulating movement of agents based on observed human behavior using techniques developed for pedestrian movement in crowd simulations. We extend a previous group conversation simulation to include an agent motion algorithm. We examine several test cases and show how the simulation generates results that mirror real-life conversation settings.
An improved genetic algorithm with conditional genetic operators and its application to set-covering problem The genetic algorithm (GA) is a popular, biologically inspired optimization method. However, in the GA there is no rule of thumb to design the GA operators and select GA parameters. Instead, trial-and-error has to be applied. In this paper we present an improved genetic algorithm in which crossover and mutation are performed conditionally instead of probability. Because there are no crossover rate and mutation rate to be selected, the proposed improved GA can be more easily applied to a problem than the conventional genetic algorithms. The proposed improved genetic algorithm is applied to solve the set-covering problem. Experimental studies show that the improved GA produces better results over the conventional one and other methods.
Lane-level traffic estimations using microscopic traffic variables This paper proposes a novel inference method to estimate lane-level traffic flow, time occupancy and vehicle inter-arrival time on road segments where local information could not be measured and assessed directly. The main contributions of the proposed method are 1) the ability to perform lane-level estimations of traffic flow, time occupancy and vehicle inter-arrival time and 2) the ability to adapt to different traffic regimes by assessing only microscopic traffic variables. We propose a modified Kriging estimation model which explicitly takes into account both spatial and temporal variability. Performance evaluations are conducted using real-world data under different traffic regimes and it is shown that the proposed method outperforms a Kalman filter-based approach.
Convolutional Neural Network-Based Classification of Driver's Emotion during Aggressive and Smooth Driving Using Multi-Modal Camera Sensors. Because aggressive driving often causes large-scale loss of life and property, techniques for advance detection of adverse driver emotional states have become important for the prevention of aggressive driving behaviors. Previous studies have primarily focused on systems for detecting aggressive driver emotion via smart-phone accelerometers and gyro-sensors, or they focused on methods of detecting physiological signals using electroencephalography (EEG) or electrocardiogram (ECG) sensors. Because EEG and ECG sensors cause discomfort to drivers and can be detached from the driver's body, it becomes difficult to focus on bio-signals to determine their emotional state. Gyro-sensors and accelerometers depend on the performance of GPS receivers and cannot be used in areas where GPS signals are blocked. Moreover, if driving on a mountain road with many quick turns, a driver's emotional state can easily be misrecognized as that of an aggressive driver. To resolve these problems, we propose a convolutional neural network (CNN)-based method of detecting emotion to identify aggressive driving using input images of the driver's face, obtained using near-infrared (NIR) light and thermal camera sensors. In this research, we conducted an experiment using our own database, which provides a high classification accuracy for detecting driver emotion leading to either aggressive or smooth (i.e., relaxed) driving. Our proposed method demonstrates better performance than existing methods.
Ethical Considerations Of Applying Robots In Kindergarten Settings: Towards An Approach From A Macroperspective In child-robot interaction (cHRI) research, many studies pursue the goal to develop interactive systems that can be applied in everyday settings. For early education, increasingly, the setting of a kindergarten is targeted. However, when cHRI and research are brought into a kindergarten, a range of ethical and related procedural aspects have to be considered and dealt with. While ethical models elaborated within other human-robot interaction settings, e.g., assisted living contexts, can provide some important indicators for relevant issues, we argue that it is important to start developing a systematic approach to identify and tackle those ethical issues which rise with cHRI in kindergarten settings on a more global level and address the impact of the technology from a macroperspective beyond the effects on the individual. Based on our experience in conducting studies with children in general and pedagogical considerations on the role of the institution of kindergarten in specific, in this paper, we enfold some relevant aspects that have barely been addressed in an explicit way in current cHRI research. Four areas are analyzed and key ethical issues are identified in each area: (1) the institutional setting of a kindergarten, (2) children as a vulnerable group, (3) the caregivers' role, and (4) pedagogical concepts. With our considerations, we aim at (i) broadening the methodology of the current studies within the area of cHRI, (ii) revalidate it based on our comprehensive empirical experience with research in kindergarten settings, both laboratory and real-world contexts, and (iii) provide a framework for the development of a more systematic approach to address the ethical issues in cHRI research within kindergarten settings.
1
0
0
0
0
0
0
0
0
0
0
0
0
0
Latent Space Expanded Variational Autoencoder For Sentence Generation Sentence generation is a key task in many natural language processing systems. Models based on a variational autoencoder (VAE) can generate plausible sentences from a continuous latent space. However, the VAE forces the latent distribution of each input sentence to match the same prior, which results in a large overlap among the latent subspaces of different sentences and a limited informative latent space. Therefore, the sentences generated by sampling from a subspace may have little correlation with the corresponding input, and the latent space cannot capture rich useful information from the input sentences, which leads to the failure of the model to generate diverse sentences from the latent space. Additionally, the Kullback-Leibler (KL) divergence collapse problem makes the VAE notoriously difficult to train. In this paper, a latent space expanded VAE (LSE-VAE) model is presented for sentence generation. The model maps each sentence to a continuous latent subspace under the constraint of its own prior distribution, and constrains nearby sentences to map to nearby subspaces. Sentences are dispersed to a large continuous latent space according to sentence similarity, where the latent subspaces of different sentences may be relatively far away from each other and arranged in an orderly manner. The experimental results show that the LSE-VAE improves the reconstruction ability of the VAE, generates plausible and more diverse sentences, and learns a larger informative latent space than the VAE with the properties of continuity and smoothness. The LSE-VAE does not suffer from the KL collapse problem, and it is robust to hyperparameters and much easier to train.
Contact personalization using a score understanding method This paper presents a method to interpret the output of a classification (or regression) model. The interpretation is based on two concepts: the variable importance and the value importance of the variable. Unlike most of the state of art interpretation methods, our approach allows the interpretation of the model output for every instance. Understanding the score given by a model for one instance can for example lead to an immediate decision in a customer relational management (CRM) system. Moreover the proposed method does not depend on a particular model and is therefore usable for any model or software used to produce the scores.
Image saliency: From intrinsic to extrinsic context We propose a novel framework for automatic saliency estimation in natural images. We consider saliency to be an anomaly with respect to a given context that can be global or local. In the case of global context, we estimate saliency in the whole image relative to a large dictionary of images. Unlike in some prior methods, this dictionary is not annotated, i.e., saliency is assumed unknown. In the case of local context, we partition the image into patches and estimate saliency in each patch relative to a large dictionary of un-annotated patches from the rest of the image. We propose a unified framework that applies to both cases in three steps. First, given an input (image or patch) we extract k nearest neighbors from the dictionary. Then, we geometrically warp each neighbor to match the input. Finally, we derive the saliency map from the mean absolute error between the input and all its warped neighbors. This algorithm is not only easy to implement but also outperforms state-of-the-art methods.
Diversity in Machine Learning. Machine learning methods have achieved good performance and been widely applied in various real-world applications. They can learn the model adaptively and be better fit for special requirements of different tasks. Generally, a good machine learning system is composed of plentiful training data, a good model training process, and an accurate inference. Many factors can affect the performance of the machine learning process, among which the diversity of the machine learning process is an important one. The diversity can help each procedure to guarantee a totally good machine learning: diversity of the training data ensures that the training data can provide more discriminative information for the model, diversity of the learned model (diversity in parameters of each model or diversity among different base models) makes each parameter/model capture unique or complement information and the diversity in inference can provide multiple choices each of which corresponds to a specific plausible local optimal result. Even though diversity plays an important role in the machine learning process, there is no systematical analysis of the diversification in the machine learning system. In this paper, we systematically summarize the methods to make data diversification, model diversification, and inference diversification in the machine learning process. In addition, the typical applications where the diversity technology improved the machine learning performance have been surveyed including the remote sensing imaging tasks, machine translation, camera relocalization, image segmentation, object detection, topic modeling, and others. Finally, we discuss some challenges of the diversity technology in machine learning and point out some directions in future work. Our analysis provides a deeper understanding of the diversity technology in machine learning tasks and hence can help design and learn more effective models for real-world applications.
Long short-term memory. Learning to store information over extended time intervals by recurrent backpropagation takes a very long time, mostly because of insufficient, decaying error backflow. We briefly review Hochreiter's (1991) analysis of this problem, then address it by introducing a novel, efficient, gradient-based method called long short-term memory (LSTM). Truncating the gradient where this does not do harm, LSTM can learn to bridge minimal time lags in excess of 1000 discrete-time steps by enforcing constant error flow through constant error carousels within special units. Multiplicative gate units learn to open and close access to the constant error flow. LSTM is local in space and time; its computational complexity per time step and weight is O(1). Our experiments with artificial data involve local, distributed, real-valued, and noisy pattern representations. In comparisons with real-time recurrent learning, back propagation through time, recurrent cascade correlation, Elman nets, and neural sequence chunking, LSTM leads to many more successful runs, and learns much faster. LSTM also solves complex, artificial long-time-lag tasks that have never been solved by previous recurrent network algorithms.
Fuzzy logic in control systems: fuzzy logic controller. I.
Robust Indoor Positioning Provided by Real-Time RSSI Values in Unmodified WLAN Networks The positioning methods based on received signal strength (RSS) measurements, link the RSS values to the position of the mobile station(MS) to be located. Their accuracy depends on the suitability of the propagation models used for the actual propagation conditions. In indoor wireless networks, these propagation conditions are very difficult to predict due to the unwieldy and dynamic nature of the RSS. In this paper, we present a novel method which dynamically estimates the propagation models that best fit the propagation environments, by using only RSS measurements obtained in real time. This method is based on maximizing compatibility of the MS to access points (AP) distance estimates. Once the propagation models are estimated in real time, it is possible to accurately determine the distance between the MS and each AP. By means of these distance estimates, the location of the MS can be obtained by trilateration. The method proposed coupled with simulations and measurements in a real indoor environment, demonstrates its feasibility and suitability, since it outperforms conventional RSS-based indoor location methods without using any radio map information nor a calibration stage.
Optimization Of Radio And Computational Resources For Energy Efficiency In Latency-Constrained Application Offloading Providing femto access points (FAPs) with computational capabilities will allow (either total or partial) offloading of highly demanding applications from smartphones to the so-called femto-cloud. Such offloading promises to be beneficial in terms of battery savings at the mobile terminal (MT) and/or in latency reduction in the execution of applications. However, for this promise to become a reality, the energy and/or the time required for the communication process must be compensated by the energy and/or the time savings that result from the remote computation at the FAPs. For this problem, we provide in this paper a framework for the joint optimization of the radio and computational resource usage exploiting the tradeoff between energy consumption and latency. Multiple antennas are assumed to be available at the MT and the serving FAP. As a result of the optimization, the optimal communication strategy (e.g., transmission power, rate, and precoder) is obtained, as well as the optimal distribution of the computational load between the handset and the serving FAP. This paper also establishes the conditions under which total or no offloading is optimal, determines which is the minimum affordable latency in the execution of the application, and analyzes, as a particular case, the minimization of the total consumed energy without latency constraints.
Integrating structured biological data by Kernel Maximum Mean Discrepancy Motivation: Many problems in data integration in bioinformatics can be posed as one common question: Are two sets of observations generated by the same distribution? We propose a kernel-based statistical test for this problem, based on the fact that two distributions are different if and only if there exists at least one function having different expectation on the two distributions. Consequently we use the maximum discrepancy between function means as the basis of a test statistic. The Maximum Mean Discrepancy (MMD) can take advantage of the kernel trick, which allows us to apply it not only to vectors, but strings, sequences, graphs, and other common structured data types arising in molecular biology. Results: We study the practical feasibility of an MMD-based test on three central data integration tasks: Testing cross-platform comparability of microarray data, cancer diagnosis, and data-content based schema matching for two different protein function classification schemas. In all of these experiments, including high-dimensional ones, MMD is very accurate in finding samples that were generated from the same distribution, and outperforms its best competitors. Conclusions: We have defined a novel statistical test of whether two samples are from the same distribution, compatible with both multivariate and structured data, that is fast, easy to implement, and works well, as confirmed by our experiments. Availability: Contact: kb@dbs.ifi.lmu.de
Noninterference for a Practical DIFC-Based Operating System The Flume system is an implementation of decentralized information flow control (DIFC) at the operating system level. Prior work has shown Flume can be implemented as a practical extension to the Linux operating system, allowing real Web applications to achieve useful security guarantees. However, the question remains if the Flume system is actually secure. This paper compares Flume with other recent DIFC systems like Asbestos, arguing that the latter is inherently susceptible to certain wide-bandwidth covert channels, and proving their absence in Flume by means of a noninterference proof in the communicating sequential processes formalism.
Large System Analysis of Cooperative Multi-Cell Downlink Transmission via Regularized Channel Inversion with Imperfect CSIT In this paper, we analyze the ergodic sum-rate of a multi-cell downlink system with base station (BS) cooperation using regularized zero-forcing (RZF) precoding. Our model assumes that the channels between BSs and users have independent spatial correlations and imperfect channel state information at the transmitter (CSIT) is available. Our derivations are based on large dimensional random matrix theory (RMT) under the assumption that the numbers of antennas at the BS and users approach to infinity with some fixed ratios. In particular, a deterministic equivalent expression of the ergodic sum-rate is obtained and is instrumental in getting insight about the joint operations of BSs, which leads to an efficient method to find the asymptotic-optimal regularization parameter for the RZF. In another application, we use the deterministic channel rate to study the optimal feedback bit allocation among the BSs for maximizing the ergodic sum-rate, subject to a total number of feedback bits constraint. By inspecting the properties of the allocation, we further propose a scheme to greatly reduce the search space for optimization. Simulation results demonstrate that the ergodic sum-rates achievable by a subspace search provides comparable results to those by an exhaustive search under various typical settings.
Global Adaptive Dynamic Programming for Continuous-Time Nonlinear Systems This paper presents a novel method of global adaptive dynamic programming (ADP) for the adaptive optimal control of nonlinear polynomial systems. The strategy consists of relaxing the problem of solving the Hamilton-Jacobi-Bellman (HJB) equation to an optimization problem, which is solved via a new policy iteration method. The proposed method distinguishes from previously known nonlinear ADP methods in that the neural network approximation is avoided, giving rise to signicant computational improvement. Instead of semiglobally or locally stabilizing, the resultant control policy is globally stabilizing for a general class of nonlinear polynomial systems. Furthermore, in the absence of the a priori knowledge of the system dynamics, an online learning method is devised to implement the proposed policy iteration technique by generalizing the current ADP theory. Finally, three numerical examples are provided to validate the effectiveness of the proposed method.
Quaternion polar harmonic Fourier moments for color images. •Quaternion polar harmonic Fourier moments (QPHFM) is proposed.•Complex Chebyshev-Fourier moments (CHFM) is extended to quaternion QCHFM.•Comparison experiments between QPHFM and QZM, QPZM, QOFMM, QCHFM and QRHFM are conducted.•QPHFM performs superbly in image reconstruction and invariant object recognition.•The importance of phase information of QPHFM in image reconstruction are discussed.
Robot tutor and pupils’ educational ability: Teaching the times tables Research shows promising results of educational robots in language and STEM tasks. In language, more research is available, occasionally in view of individual differences in pupils’ educational ability levels, and learning seems to improve with more expressive robot behaviors. In STEM, variations in robots’ behaviors have been examined with inconclusive results and never while systematically investigating how differences in educational abilities match with different robot behaviors. We applied an autonomously tutoring robot (without tablet, partly WOz) in a 2 × 2 experiment of social vs. neutral behavior in above-average vs. below-average schoolchildren (N = 86; age 8–10 years) while rehearsing the multiplication tables on a one-to-one basis. The standard school test showed that on average, pupils significantly improved their performance even after 3 occasions of 5-min exercises. Beyond-average pupils profited most from a robot tutor, whereas those below average in multiplication benefited more from a robot that showed neutral rather than more social behavior.
1.2
0.2
0.2
0.1
0.002597
0
0
0
0
0
0
0
0
0
Performance Guaranteed Consensus Tracking Control of Nonlinear Multiagent Systems: A Finite-Time Function-Based Approach In this article, we study the performance guaranteed consensus tracking problem for a class of high-order nonlinear multiagent systems subject to mismatched uncertainties and external disturbances. We first construct a finite-time function, with which a performance function is introduced that links the convergence time of the relative consensus errors with the neighbor agents. We then introduce two new lemmas that play a virtual role in addressing the consensus stability of closed-loop multiagent system, where a fully distributed adaptive control without using global information of the topology is developed. Different from most existing works for multiagent systems with prescribed performance that can only achieve uniformly ultimately bounded consensus, the proposed control scheme is able to ensure that the consensus errors converge to the pregiven compact sets within preassigned finite time rather than infinite time and the outputs of all the agents track the leader’s trajectory asymptotically. Simulation verification also confirms the effectiveness of the proposed approach.
Fuzzy Adaptive Tracking Control of Wheeled Mobile Robots With State-Dependent Kinematic and Dynamic Disturbances Unlike most works based on pure nonholonomic constraint, this paper proposes a fuzzy adaptive tracking control method for wheeled mobile robots, where unknown slippage occurs and violates the nonholononomic constraint in the form of state-dependent kinematic and dynamic disturbances. These disturbances degrade tracking performance significantly and, therefore, should be compensated. To this end, the kinematics with state-dependent disturbances are rigorously derived based on the general form of slippage in the mobile robots, and fuzzy adaptive observers together with parameter adaptation laws are designed to estimate the state-dependent disturbances in both kinematics and dynamics. Because of the modular structure of the proposed method, it can be easily combined with the previous controllers based on the model with the pure nonholonomic constraint, such that the combination of the fuzzy adaptive observers with the previously proposed backstepping-like feedback linearization controller can guarantee the trajectory tracking errors to be globally ultimately bounded, even when the nonholonomic constraint is violated, and their ultimate bounds can be adjusted appropriately for various types of trajectories in the presence of large initial tracking errors and disturbances. Both the stability analysis and simulation results are provided to validate the proposed controller.
Leader-following consensus in second-order multi-agent systems with input time delay: An event-triggered sampling approach. This paper analytically investigates an event-triggered leader-following consensus in second-order multi-agent systems with time delay in the control input. Each agent׳s update of control input is driven by properly defined event, which depends on the measurement error, the states of its neighboring agents at their individual time instants, and an exponential decay function. Necessary and sufficient conditions are presented to ensure a leader-following consensus. Moreover, the control is updated only when the event-triggered condition is satisfied, which significantly decreases the number of communication among nodes, avoided effectively the continuous communication of the information channel among agents and excluded the Zeno-behavior of triggering time sequences. A numerical simulation example is given to illustrate the theoretical results.
Adaptive neural control for a class of stochastic nonlinear systems by backstepping approach. This paper addresses adaptive neural control for a class of stochastic nonlinear systems which are not in strict-feedback form. Based on the structural characteristics of radial basis function (RBF) neural networks (NNs), a backstepping design approach is extended from stochastic strict-feedback systems to a class of more general stochastic nonlinear systems. In the control design procedure, RBF NNs are used to approximate unknown nonlinear functions and the backstepping technique is utilized to construct the desired controller. The proposed adaptive neural controller guarantees that all the closed-loop signals are bounded and the tracking error converges to a sufficiently small neighborhood of the origin. Two simulation examples are used to illustrate the effectiveness of the proposed approach.
Prescribed Performance Adaptive Fuzzy Containment Control for Nonlinear Multiagent Systems Using Disturbance Observer This article focuses on the containment control problem for nonlinear multiagent systems (MASs) with unknown disturbance and prescribed performance in the presence of dead-zone output. The fuzzy-logic systems (FLSs) are used to approximate the unknown nonlinear function, and a nonlinear disturbance observer is used to estimate unknown external disturbances. Meanwhile, a new distributed containment control scheme is developed by utilizing the adaptive compensation technique without assumption of the boundary value of unknown disturbance. Furthermore, a Nussbaum function is utilized to cope with the unknown control coefficient, which is caused by the nonlinearity in the output mechanism. Moreover, a second-order tracking differentiator (TD) is introduced to avoid the repeated differentiation of the virtual controller. The outputs of the followers converge to the convex hull spanned by the multiple dynamic leaders. It is shown that all the signals are semiglobally uniformly ultimately bounded (SGUUB), and the local neighborhood containment errors can converge into the prescribed boundary. Finally, the effectiveness of the approach proposed in this article is illustrated by simulation results.
Distributed Observer-Based Cooperative Control Approach for Uncertain Nonlinear MASs Under Event-Triggered Communication The distributed tracking problem for uncertain nonlinear multiagent systems (MASs) under event-triggered communication is an important issue. However, existing results provide solutions that can only ensure stability with bounded tracking errors, as asymptotic tracking is difficult to be achieved mainly due to the errors caused by event-triggering mechanisms and system uncertainties. In this artic...
Model-Based Adaptive Event-Triggered Control of Strict-Feedback Nonlinear Systems This paper is concerned with the adaptive event-triggered control problem of nonlinear continuous-time systems in strict-feedback form. By using the event-sampled neural network (NN) to approximate the unknown nonlinear function, an adaptive model and an associated event-triggered controller are designed by exploiting the backstepping method. In the proposed method, the feedback signals and the NN...
Massive MIMO for next generation wireless systems Multi-user MIMO offers big advantages over conventional point-to-point MIMO: it works with cheap single-antenna terminals, a rich scattering environment is not required, and resource allocation is simplified because every active terminal utilizes all of the time-frequency bins. However, multi-user MIMO, as originally envisioned, with roughly equal numbers of service antennas and terminals and frequency-division duplex operation, is not a scalable technology. Massive MIMO (also known as large-scale antenna systems, very large MIMO, hyper MIMO, full-dimension MIMO, and ARGOS) makes a clean break with current practice through the use of a large excess of service antennas over active terminals and time-division duplex operation. Extra antennas help by focusing energy into ever smaller regions of space to bring huge improvements in throughput and radiated energy efficiency. Other benefits of massive MIMO include extensive use of inexpensive low-power components, reduced latency, simplification of the MAC layer, and robustness against intentional jamming. The anticipated throughput depends on the propagation environment providing asymptotically orthogonal channels to the terminals, but so far experiments have not disclosed any limitations in this regard. While massive MIMO renders many traditional research problems irrelevant, it uncovers entirely new problems that urgently need attention: the challenge of making many low-cost low-precision components that work effectively together, acquisition and synchronization for newly joined terminals, the exploitation of extra degrees of freedom provided by the excess of service antennas, reducing internal power consumption to achieve total energy efficiency reductions, and finding new deployment scenarios. This article presents an overview of the massive MIMO concept and contemporary research on the topic.
Adaptive Federated Learning in Resource Constrained Edge Computing Systems Emerging technologies and applications including Internet of Things, social networking, and crowd-sourcing generate large amounts of data at the network edge. Machine learning models are often built from the collected data, to enable the detection, classification, and prediction of future events. Due to bandwidth, storage, and privacy concerns, it is often impractical to send all the data to a cen...
A new optimization method: big bang-big crunch Nature is the principal source for proposing new optimization methods such as genetic algorithms (GA) and simulated annealing (SA) methods. All traditional evolutionary algorithms are heuristic population-based search procedures that incorporate random variation and selection. The main contribution of this study is that it proposes a novel optimization method that relies on one of the theories of the evolution of the universe; namely, the Big Bang and Big Crunch Theory. In the Big Bang phase, energy dissipation produces disorder and randomness is the main feature of this phase; whereas, in the Big Crunch phase, randomly distributed particles are drawn into an order. Inspired by this theory, an optimization algorithm is constructed, which will be called the Big Bang-Big Crunch (BB-BC) method that generates random points in the Big Bang phase and shrinks those points to a single representative point via a center of mass or minimal cost approach in the Big Crunch phase. It is shown that the performance of the new (BB-BC) method demonstrates superiority over an improved and enhanced genetic search algorithm also developed by the authors of this study, and outperforms the classical genetic algorithm (GA) for many benchmark test functions.
Secure and privacy preserving keyword searching for cloud storage services Cloud storage services enable users to remotely access data in a cloud anytime and anywhere, using any device, in a pay-as-you-go manner. Moving data into a cloud offers great convenience to users since they do not have to care about the large capital investment in both the deployment and management of the hardware infrastructures. However, allowing a cloud service provider (CSP), whose purpose is mainly for making a profit, to take the custody of sensitive data, raises underlying security and privacy issues. To keep user data confidential against an untrusted CSP, a natural way is to apply cryptographic approaches, by disclosing the data decryption key only to authorized users. However, when a user wants to retrieve files containing certain keywords using a thin client, the adopted encryption system should not only support keyword searching over encrypted data, but also provide high performance. In this paper, we investigate the characteristics of cloud storage services and propose a secure and privacy preserving keyword searching (SPKS) scheme, which allows the CSP to participate in the decipherment, and to return only files containing certain keywords specified by the users, so as to reduce both the computational and communication overhead in decryption for users, on the condition of preserving user data privacy and user querying privacy. Performance analysis shows that the SPKS scheme is applicable to a cloud environment.
A review on interval type-2 fuzzy logic applications in intelligent control. A review of the applications of interval type-2 fuzzy logic in intelligent control has been considered in this paper. The fundamental focus of the paper is based on the basic reasons for using type-2 fuzzy controllers for different areas of application. Recently, bio-inspired methods have emerged as powerful optimization algorithms for solving complex problems. In the case of designing type-2 fuzzy controllers for particular applications, the use of bio-inspired optimization methods have helped in the complex task of finding the appropriate parameter values and structure of the fuzzy systems. In this review, we consider the application of genetic algorithms, particle swarm optimization and ant colony optimization as three different paradigms that help in the design of optimal type-2 fuzzy controllers. We also mention alternative approaches to designing type-2 fuzzy controllers without optimization techniques.
Design of robust fuzzy fault detection filter for polynomial fuzzy systems with new finite frequency specifications This paper investigates the problem of fault detection filter design for discrete-time polynomial fuzzy systems with faults and unknown disturbances. The frequency ranges of the faults and the disturbances are assumed to be known beforehand and to reside in low, middle or high frequency ranges. Thus, the proposed filter is designed in the finite frequency range to overcome the conservatism generated by those designed in the full frequency domain. Being of polynomial fuzzy structure, the proposed filter combines the H−/H∞ performances in order to ensure the best robustness to the disturbance and the best sensitivity to the fault. Design conditions are derived in Sum Of Squares formulations that can be easily solved via available software tools. Two illustrative examples are introduced to demonstrate the effectiveness of the proposed method and a comparative study with LMI method is also provided.
Hardware Circuits Design and Performance Evaluation of a Soft Lower Limb Exoskeleton Soft lower limb exoskeletons (LLEs) are wearable devices that have good potential in walking rehabilitation and augmentation. While a few studies focused on the structure design and assistance force optimization of the soft LLEs, rarely work has been conducted on the hardware circuits design. The main purpose of this work is to present a new soft LLE for walking efficiency improvement and introduce its hardware circuits design. A soft LLE for hip flexion assistance and a hardware circuits system with scalability were proposed. To assess the efficacy of the soft LLE, the experimental tests that evaluate the sensor data acquisition, force tracking performance, lower limb muscle activity and metabolic cost were conducted. The time error in the peak assistance force was just 1%. The reduction in the normalized root-mean-square EMG of the rectus femoris was 7.1%. The net metabolic cost in exoskeleton on condition was reduced by 7.8% relative to walking with no exoskeleton. The results show that the designed hardware circuits can be applied to the soft LLE and the soft LLE is able to improve walking efficiency of wearers.
1.2
0.2
0.2
0.2
0.2
0.2
0.05
0
0
0
0
0
0
0
A critical discussion into the core of swarm intelligence algorithms The literature is now filled with swarm intelligence algorithms developed by taking inspiration from a number of insects and other animals and phenomena, such as ants, termites, bees, fishes and cockroaches, to name just a few. Many, if not most, of these bioinspirations carry with them some common issues and features which happen at the individual level, promoting very similar collective emergent phenomena. Thus, despite using different biological metaphors as inspiration, most algorithms present a similar structure and it is possible to identify common macro-processes among them. In this context, this paper identifies a set of common features among some well-known swarm-based algorithms and how each of these approaches implement them. By doing this, we provide the community with the core features of swarm-intelligence algorithms. This diagnostic is crucial and timely to the field, because once we are able to list and explain these commonalities, we are also able to better analyze and design swarm intelligence algorithms.
On the History of the Minimum Spanning Tree Problem It is standard practice among authors discussing the minimum spanning tree problem to refer to the work of Kruskal(1956) and Prim (1957) as the sources of the problem and its first efficient solutions, despite the citation by both of Boruvka (1926) as a predecessor. In fact, there are several apparently independent sources and algorithmic solutions of the problem. They have appeared in Czechoslovakia, France, and Poland, going back to the beginning of this century. We shall explore and compare these works and their motivations, and relate them to the most recent advances on the minimum spanning tree problem.
Smart home energy management system using IEEE 802.15.4 and zigbee Wireless personal area network and wireless sensor networks are rapidly gaining popularity, and the IEEE 802.15 Wireless Personal Area Working Group has defined no less than different standards so as to cater to the requirements of different applications. The ubiquitous home network has gained widespread attentions due to its seamless integration into everyday life. This innovative system transparently unifies various home appliances, smart sensors and energy technologies. The smart energy market requires two types of ZigBee networks for device control and energy management. Today, organizations use IEEE 802.15.4 and ZigBee to effectively deliver solutions for a variety of areas including consumer electronic device control, energy management and efficiency, home and commercial building automation as well as industrial plant management. We present the design of a multi-sensing, heating and airconditioning system and actuation application - the home users: a sensor network-based smart light control system for smart home and energy control production. This paper designs smart home device descriptions and standard practices for demand response and load management "Smart Energy" applications needed in a smart energy based residential or light commercial environment. The control application domains included in this initial version are sensing device control, pricing and demand response and load control applications. This paper introduces smart home interfaces and device definitions to allow interoperability among ZigBee devices produced by various manufacturers of electrical equipment, meters, and smart energy enabling products. We introduced the proposed home energy control systems design that provides intelligent services for users and we demonstrate its implementation using a real testbad.
Bee life-based multi constraints multicast routing optimization for vehicular ad hoc networks. A vehicular ad hoc network (VANET) is a subclass of mobile ad hoc networks, considered as one of the most important approach of intelligent transportation systems (ITS). It allows inter-vehicle communication in which their movement is restricted by a VANET mobility model and supported by some roadside base stations as fixed infrastructures. Multicasting provides different traffic information to a limited number of vehicle drivers by a parallel transmission. However, it represents a very important challenge in the application of vehicular ad hoc networks especially, in the case of the network scalability. In the applications of this sensitive field, it is very essential to transmit correct data anywhere and at any time. Consequently, the VANET routing protocols should be adapted appropriately and meet effectively the quality of service (QoS) requirements in an optimized multicast routing. In this paper, we propose a novel bee colony optimization algorithm called bees life algorithm (BLA) applied to solve the quality of service multicast routing problem (QoS-MRP) for vehicular ad hoc networks as NP-Complete problem with multiple constraints. It is considered as swarm-based algorithm which imitates closely the life of the colony. It follows the two important behaviors in the nature of bees which are the reproduction and the food foraging. BLA is applied to solve QoS-MRP with four objectives which are cost, delay, jitter, and bandwidth. It is also submitted to three constraints which are maximum allowed delay, maximum allowed jitter and minimum requested bandwidth. In order to evaluate the performance and the effectiveness of this realized proposal using C++ and integrated at the routing protocol level, a simulation study has been performed using the network simulator (NS2) based on a mobility model of VANET. The comparisons of the experimental results show that the proposed algorithm outperformed in an efficient way genetic algorithm (GA), bees algorithm (BA) and marriage in honey bees optimization (MBO) algorithm as state-of-the-art conventional metaheuristics applied to QoS-MRP problem with the same simulation parameters.
On the Spatiotemporal Traffic Variation in Vehicle Mobility Modeling Several studies have shown the importance of realistic micromobility and macromobility modeling in vehicular ad hoc networks (VANETs). At the macroscopic level, most researchers focus on a detailed and accurate description of road topology. However, a key factor often overlooked is a spatiotemporal configuration of vehicular traffic. This factor greatly influences network topology and topology variations. Indeed, vehicle distribution has high spatial and temporal diversity that depends on the time of the day and place attraction. This diversity impacts the quality of radio links and, thus, network topology. In this paper, we propose a new mobility model for vehicular networks in urban and suburban environments. To reproduce realistic network topology and topological changes, the model uses real static and dynamic data on the environment. The data concern particularly the topographic and socioeconomic characteristics of infrastructures and the spatiotemporal population distribution. We validate our model by comparing the simulation results with real data derived from individual displacement survey. We also present statistics on network topology, which show the interest of taking into account the spatiotemporal mobility variation.
A bio-inspired clustering in mobile adhoc networks for internet of things based on honey bee and genetic algorithm In mobile adhoc networks for internet of things, the size of routing table can be reduced with the help of clustering structure. The dynamic nature of MANETs and its complexity make it a type of network with high topology changes. To reduce the topology maintenance overhead, the cluster based structure may be used. Hence, it is highly desirable to design an algorithm that adopts quickly to topology dynamics and form balanced and stable clusters. In this article, the formulation of clustering problem is carried out initially. Later, an algorithm on the basis of honey bee algorithm, genetic algorithm and tabu search (GBTC) for internet of things is proposed. In this algorithm, the individual (bee) represents a possbile clustering structure and its fitness is evaluated on the basis of its stability and load balancing. A method is presented by merging the properties of honey bee and genetic algorithms to help the population to cope with the topology dynamics and produce top quality solutions that are closely related to each other. The simulation results conducted for validation show that the proposed work forms balance and stable clusters. The simulation results are compared with algorithms that do not consider the dynamic optimization requirements. The GTBC outperform existing algorithms in terms of network lifetime and clustering overhead etc.
An enhanced QoS CBT multicast routing protocol based on Genetic Algorithm in a hybrid HAP-Satellite system A QoS multicast routing scheme based on Genetic Algorithms (GA) heuristic is presented in this paper. Our proposal, called Constrained Cost–Bandwidth–Delay Genetic Algorithm (CCBD-GA), is applied to a multilayer hybrid platform that includes High Altitude Platforms (HAPs) and a Satellite platform. This GA scheme has been compared with another GA well-known in the literature called Multi-Objective Genetic Algorithm (MOGA) in order to show the proposed algorithm goodness. In order to test the efficiency of GA schemes on a multicast routing protocol, these GA schemes are inserted into an enhanced version of the Core-Based Tree (CBT) protocol with QoS support. CBT and GA schemes are tested in a multilayer hybrid HAP and Satellite architecture and interesting results have been discovered. The joint bandwidth–delay metrics can be very useful in hybrid platforms such as that considered, because it is possible to take advantage of the single characteristics of the Satellite and HAP segments. The HAP segment offers low propagation delay permitting QoS constraints based on maximum end-to-end delay to be met. The Satellite segment, instead, offers high bandwidth capacity with higher propagation delay. The joint bandwidth–delay metric permits the balancing of the traffic load respecting both QoS constraints. Simulation results have been evaluated in terms of HAP and Satellite utilization, bandwidth, end-to-end delay, fitness function and cost of the GA schemes.
On the ratio of optimal integral and fractional covers It is shown that the ratio of optimal integral and fractional covers of a hypergraph does not exceed 1 + log d , where d is the maximum degree. This theorem may replace probabilistic methods in certain circumstances. Several applications are shown.
Task Offloading in Vehicular Edge Computing Networks: A Load-Balancing Solution Recently, the rapid advance of vehicular networks has led to the emergence of diverse delay-sensitive vehicular applications such as automatic driving, auto navigation. Note that existing resource-constrained vehicles cannot adequately meet these demands on low / ultra-low latency. By offloading parts of the vehicles’ compute-intensive tasks to the edge servers in proximity, mobile edge computing is envisioned as a promising paradigm, giving rise to the vehicular edge computing networks (VECNs). However, most existing works on task offloading in VECNs did not take the load balancing of the computation resources at the edge servers into account. To address these issues and given the high dynamics of vehicular networks, we introduce fiber-wireless (FiWi) technology to enhance VECNs, due to its advantages on centralized network management and supporting multiple communication techniques. Aiming to minimize the processing delay of the vehicles’ computation tasks, we propose a software-defined networking (SDN) based load-balancing task offloading scheme in FiWi enhanced VECNs, where SDN is introduced to provide supports for the centralized network and vehicle information management. Extensive analysis and numerical results corroborate that our proposed load-balancing scheme can achieve superior performance on processing delay reduction by utilizing the edge servers’ computation resources more efficiently.
A hierarchical type-2 fuzzy logic control architecture for autonomous mobile robots Autonomous mobile robots navigating in changing and dynamic unstructured environments like the outdoor environments need to cope with large amounts of uncertainties that are inherent of natural environments. The traditional type-1 fuzzy logic controller (FLC) using precise type-1 fuzzy sets cannot fully handle such uncertainties. A type-2 FLC using type-2 fuzzy sets can handle such uncertainties to produce a better performance. In this paper, we present a novel reactive control architecture for autonomous mobile robots that is based on type-2 FLC to implement the basic navigation behaviors and the coordination between these behaviors to produce a type-2 hierarchical FLC. In our experiments, we implemented this type-2 architecture in different types of mobile robots navigating in indoor and outdoor unstructured and challenging environments. The type-2-based control system dealt with the uncertainties facing mobile robots in unstructured environments and resulted in a very good performance that outperformed the type-1-based control system while achieving a significant rule reduction compared to the type-1 system.
Multi-stage genetic programming: A new strategy to nonlinear system modeling This paper presents a new multi-stage genetic programming (MSGP) strategy for modeling nonlinear systems. The proposed strategy is based on incorporating the individual effect of predictor variables and the interactions among them to provide more accurate simulations. According to the MSGP strategy, an efficient formulation for a problem comprises different terms. In the first stage of the MSGP-based analysis, the output variable is formulated in terms of an influencing variable. Thereafter, the error between the actual and the predicted value is formulated in terms of a new variable. Finally, the interaction term is derived by formulating the difference between the actual values and the values predicted by the individually developed terms. The capabilities of MSGP are illustrated by applying it to the formulation of different complex engineering problems. The problems analyzed herein include the following: (i) simulation of pH neutralization process, (ii) prediction of surface roughness in end milling, and (iii) classification of soil liquefaction conditions. The validity of the proposed strategy is confirmed by applying the derived models to the parts of the experimental results that were not included in the analyses. Further, the external validation of the models is verified using several statistical criteria recommended by other researchers. The MSGP-based solutions are capable of effectively simulating the nonlinear behavior of the investigated systems. The results of MSGP are found to be more accurate than those of standard GP and artificial neural network-based models.
Placing Virtual Machines to Optimize Cloud Gaming Experience Optimizing cloud gaming experience is no easy task due to the complex tradeoff between gamer quality of experience (QoE) and provider net profit. We tackle the challenge and study an optimization problem to maximize the cloud gaming provider's total profit while achieving just-good-enough QoE. We conduct measurement studies to derive the QoE and performance models. We formulate and optimally solve the problem. The optimization problem has exponential running time, and we develop an efficient heuristic algorithm. We also present an alternative formulation and algorithms for closed cloud gaming services with dedicated infrastructures, where the profit is not a concern and overall gaming QoE needs to be maximized. We present a prototype system and testbed using off-the-shelf virtualization software, to demonstrate the practicality and efficiency of our algorithms. Our experience on realizing the testbed sheds some lights on how cloud gaming providers may build up their own profitable services. Last, we conduct extensive trace-driven simulations to evaluate our proposed algorithms. The simulation results show that the proposed heuristic algorithms: (i) produce close-to-optimal solutions, (ii) scale to large cloud gaming services with 20,000 servers and 40,000 gamers, and (iii) outperform the state-of-the-art placement heuristic, e.g., by up to 3.5 times in terms of net profits.
Adversarial Example Generation with Syntactically Controlled Paraphrase Networks. We propose syntactically controlled paraphrase networks (SCPNs) and use them to generate adversarial examples. Given a sentence and a target syntactic form (e.g., a constituency parse), SCPNs are trained to produce a paraphrase of the sentence with the desired syntax. We show it is possible to create training data for this task by first doing backtranslation at a very large scale, and then using a parser to label the syntactic transformations that naturally occur during this process. Such data allows us to train a neural encoder-decoder model with extra inputs to specify the target syntax. A combination of automated and human evaluations show that SCPNs generate paraphrases that follow their target specifications without decreasing paraphrase quality when compared to baseline (uncontrolled) paraphrase systems. Furthermore, they are more capable of generating syntactically adversarial examples that both (1) fool pretrained models and (2) improve the robustness of these models to syntactic variation when used to augment their training data.
Hardware Circuits Design and Performance Evaluation of a Soft Lower Limb Exoskeleton Soft lower limb exoskeletons (LLEs) are wearable devices that have good potential in walking rehabilitation and augmentation. While a few studies focused on the structure design and assistance force optimization of the soft LLEs, rarely work has been conducted on the hardware circuits design. The main purpose of this work is to present a new soft LLE for walking efficiency improvement and introduce its hardware circuits design. A soft LLE for hip flexion assistance and a hardware circuits system with scalability were proposed. To assess the efficacy of the soft LLE, the experimental tests that evaluate the sensor data acquisition, force tracking performance, lower limb muscle activity and metabolic cost were conducted. The time error in the peak assistance force was just 1%. The reduction in the normalized root-mean-square EMG of the rectus femoris was 7.1%. The net metabolic cost in exoskeleton on condition was reduced by 7.8% relative to walking with no exoskeleton. The results show that the designed hardware circuits can be applied to the soft LLE and the soft LLE is able to improve walking efficiency of wearers.
1.2
0.2
0.2
0.2
0.2
0.2
0.1
0
0
0
0
0
0
0
Analysis of cooperative localisation performance under varying sensor qualities and communication rates. Cooperative Localisation (CL) is a robust technique used to improve localisation accuracy in multi-robot systems. However, there is a lack of research on how CL performs under different conditions. It is unclear when CL is worthwhile, and how CL performance is affected if the system changes. This information is particularly important for systems with robots that have limited power and processing, which cannot afford to constantly perform CL. This paper investigates CL under varying sensor qualities (position accuracy, yaw accuracy, sample rate), communication rates, and number of robots for both homogeneous and heterogeneous multi-robot systems. Trends are found in MATLAB simulations using the UTIAS dataset, and then validated on Kobuki robots using an OptiTrack-based system. We find that yaw accuracy has a substantial effect on performance, a communication rate that is too fast can be detrimental, and heterogeneous systems are greater candidates for cooperative localisation than homogeneous systems.
Self-triggered coordination of robotic networks for optimal deployment This paper studies a deployment problem for a group of robots where individual agents operate with outdated information about each other's locations. Our objective is to understand to what extent outdated information is still useful and at which point it becomes essential to obtain new, up-to-date information. We propose a self-triggered coordination algorithm based on spatial partitioning techniques with uncertain information. We analyze its correctness in synchronous and asynchronous scenarios, and establish the same convergence guarantees that a synchronous algorithm with perfect information at all times would achieve. The technical approach combines computational geometry, set-valued stability analysis, and event-based systems.
Robust Control for Mobility and Wireless Communication in Cyber-Physical Systems With Application to Robot Teams. In this paper, a system architecture to provide end-to-end network connectivity for autonomous teams of robots is discussed. The core of the proposed system is a cyber-physical controller whose goal is to ensure network connectivity as robots move to accomplish their assigned tasks. Due to channel quality uncertainties inherent to wireless propagation, we adopt a stochastic model where achievable ...
Consensus Control Under Communication Delay in a Three-Robot System: Design and Experiments A consensus-control protocol is designed and implemented here on a three-robot system arranged on a horizontal platform, in which a camera system is used to track robot positions, and a personal computer broadcasts commands to the robots based on this protocol via a Bluetooth connection, where such commands are affected by time delays. The design involves some salient features of this protocol based on a graph-based approach, an input–output linearization scheme, and addressing uncertainties in the control problem. By implementing this design on experiments, we show that consensus of the robots can be successfully achieved, and their speed of reaching consensus can be systematically improved. Experimental results also strongly agree with those obtained from nonlinear simulations.
Cooperative Localization Under Limited Connectivity In this article, we report two decentralized multiagent cooperative localization algorithms in which, to reduce the communication cost, interagent state estimate correlations are not maintained but accounted for implicitly. In our first algorithm, to guarantee filter consistency, we account for unknown interagent correlations via an upper bound on the joint covariance matrix of the agents. In the second method, we use an optimization framework to estimate the unknown interagent cross-covariance matrix. In our algorithms, each agent localizes itself in a global coordinate frame using a local filter driven by local dead reckoning and occasional absolute measurement updates, and opportunistically corrects its pose estimate whenever it can obtain relative measurements with respect to other mobile agents. To process any relative measurement, only the agent that has taken the measurement and the agent the measurement is taken from need to communicate with each other. Consequently, our algorithms are decentralized algorithms that do not impose restrictive network-wide connectivity condition. Moreover, we make no assumptions about the type of agents or relative measurements. We demonstrate our algorithms in simulation and a robotic experiment.
Communication in reactive multiagent robotic systems Multiple cooperating robots are able to complete many tasks more quickly and reliably than one robot alone. Communication between the robots can multiply their capabilities and effectiveness, but to what extent? In this research, the importance of communication in robotic societies is investigated through experiments on both simulated and real robots. Performance was measured for three different types of communication for three different tasks. The levels of communication are progressively more complex and potentially more expensive to implement. For some tasks, communication can significantly improve performance, but for others inter-agent communication is apparently unnecessary. In cases where communication helps, the lowest level of communication is almost as effective as the more complex type. The bulk of these results are derived from thousands of simulations run with randomly generated initial conditions. The simulation results help determine appropriate parameters for the reactive control system which was ported for tests on Denning mobile robots.
Footprints: history-rich tools for information foraging Inspired by Hill and Hollans original work [7], we have beendeveloping a theory of interaction history and building tools toapply this theory to navigation in a complex information space. Wehave built a series of tools - map, paths, annota- tions andsignposts - based on a physical-world navigation metaphor. Thesetools have been in use for over a year. Our user study involved acontrolled browse task and showed that users were able to get thesame amount of work done with significantly less effort.
Very Deep Convolutional Networks for Large-Scale Image Recognition. In this work we investigate the effect of the convolutional network depth on its accuracy in the large-scale image recognition setting. Our main contribution is a thorough evaluation of networks of increasing depth using an architecture with very small (3x3) convolution filters, which shows that a significant improvement on the prior-art configurations can be achieved by pushing the depth to 16-19 weight layers. These findings were the basis of our ImageNet Challenge 2014 submission, where our team secured the first and the second places in the localisation and classification tracks respectively. We also show that our representations generalise well to other datasets, where they achieve state-of-the-art results. We have made our two best-performing ConvNet models publicly available to facilitate further research on the use of deep visual representations in computer vision.
Chimp optimization algorithm. •A novel optimizer called Chimp Optimization Algorithm (ChOA) is proposed.•ChOA is inspired by individual intelligence and sexual motivation of chimps.•ChOA alleviates the problems of slow convergence rate and trapping in local optima.•The four main steps of Chimp hunting are implemented.
The contourlet transform: an efficient directional multiresolution image representation. The limitations of commonly used separable extensions of one-dimensional transforms, such as the Fourier and wavelet transforms, in capturing the geometry of image edges are well known. In this paper, we pursue a "true" two-dimensional transform that can capture the intrinsic geometrical structure that is key in visual information. The main challenge in exploring geometry in images comes from the discrete nature of the data. Thus, unlike other approaches, such as curvelets, that first develop a transform in the continuous domain and then discretize for sampled data, our approach starts with a discrete-domain construction and then studies its convergence to an expansion in the continuous domain. Specifically, we construct a discrete-domain multiresolution and multidirection expansion using nonseparable filter banks, in much the same way that wavelets were derived from filter banks. This construction results in a flexible multiresolution, local, and directional image expansion using contour segments, and, thus, it is named the contourlet transform. The discrete contourlet transform has a fast iterated filter bank algorithm that requires an order N operations for N-pixel images. Furthermore, we establish a precise link between the developed filter bank and the associated continuous-domain contourlet expansion via a directional multiresolution analysis framework. We show that with parabolic scaling and sufficient directional vanishing moments, contourlets achieve the optimal approximation rate for piecewise smooth functions with discontinuities along twice continuously differentiable curves. Finally, we show some numerical experiments demonstrating the potential of contourlets in several image processing applications. Index Terms-Contourlets, contours, filter banks, geometric image processing, multidirection, multiresolution, sparse representation, wavelets.
Cognitive Cars: A New Frontier for ADAS Research This paper provides a survey of recent works on cognitive cars with a focus on driver-oriented intelligent vehicle motion control. The main objective here is to clarify the goals and guidelines for future development in the area of advanced driver-assistance systems (ADASs). Two major research directions are investigated and discussed in detail: 1) stimuli–decisions–actions, which focuses on the driver side, and 2) perception enhancement–action-suggestion–function-delegation, which emphasizes the ADAS side. This paper addresses the important achievements and major difficulties of each direction and discusses how to combine the two directions into a single integrated system to obtain safety and comfort while driving. Other related topics, including driver training and infrastructure design, are also studied.
Finite-approximation-error-based discrete-time iterative adaptive dynamic programming. In this paper, a new iterative adaptive dynamic programming (ADP) algorithm is developed to solve optimal control problems for infinite horizon discrete-time nonlinear systems with finite approximation errors. First, a new generalized value iteration algorithm of ADP is developed to make the iterative performance index function converge to the solution of the Hamilton-Jacobi-Bellman equation. The ...
Robust Sparse Linear Discriminant Analysis Linear discriminant analysis (LDA) is a very popular supervised feature extraction method and has been extended to different variants. However, classical LDA has the following problems: 1) The obtained discriminant projection does not have good interpretability for features. 2) LDA is sensitive to noise. 3) LDA is sensitive to the selection of number of projection directions. In this paper, a novel feature extraction method called robust sparse linear discriminant analysis (RSLDA) is proposed to solve the above problems. Specifically, RSLDA adaptively selects the most discriminative features for discriminant analysis by introducing the l2;1 norm. An orthogonal matrix and a sparse matrix are also simultaneously introduced to guarantee that the extracted features can hold the main energy of the original data and enhance the robustness to noise, and thus RSLDA has the potential to perform better than other discriminant methods. Extensive experiments on six databases demonstrate that the proposed method achieves the competitive performance compared with other state-of-the-art feature extraction methods. Moreover, the proposed method is robust to the noisy data. IEEE
Ethical Considerations Of Applying Robots In Kindergarten Settings: Towards An Approach From A Macroperspective In child-robot interaction (cHRI) research, many studies pursue the goal to develop interactive systems that can be applied in everyday settings. For early education, increasingly, the setting of a kindergarten is targeted. However, when cHRI and research are brought into a kindergarten, a range of ethical and related procedural aspects have to be considered and dealt with. While ethical models elaborated within other human-robot interaction settings, e.g., assisted living contexts, can provide some important indicators for relevant issues, we argue that it is important to start developing a systematic approach to identify and tackle those ethical issues which rise with cHRI in kindergarten settings on a more global level and address the impact of the technology from a macroperspective beyond the effects on the individual. Based on our experience in conducting studies with children in general and pedagogical considerations on the role of the institution of kindergarten in specific, in this paper, we enfold some relevant aspects that have barely been addressed in an explicit way in current cHRI research. Four areas are analyzed and key ethical issues are identified in each area: (1) the institutional setting of a kindergarten, (2) children as a vulnerable group, (3) the caregivers' role, and (4) pedagogical concepts. With our considerations, we aim at (i) broadening the methodology of the current studies within the area of cHRI, (ii) revalidate it based on our comprehensive empirical experience with research in kindergarten settings, both laboratory and real-world contexts, and (iii) provide a framework for the development of a more systematic approach to address the ethical issues in cHRI research within kindergarten settings.
1.24
0.24
0.24
0.24
0.24
0.04
0
0
0
0
0
0
0
0
Smooth filtering identification based on convolutional neural networks The increasing prevalence of digital technology brings great convenience to human life, while also shows us the problems and challenges. Relying on easy-to-use image editing tools, some malicious manipulations, such as image forgery, have already threatened the authenticity of information, especially the electronic evidence in the crimes. As a result, digital forensics attracts more and more attention of researchers. Since some general post-operations, like widely used smooth filtering, can affect the reliability of forensic methods in various ways, it is also significant to detect them. Furthermore, the determination of detailed filtering parameters assists to recover the tampering history of an image. To deal with this problem, we propose a new approach based on convolutional neural networks (CNNs). Through adding a transform layer, obtained distinguishable frequency-domain features are put into a conventional CNN model, to identify the template parameters of various types of spatial smooth filtering operations, such as average, Gaussian and median filtering. Experimental results on a composite database show that putting the images directly into the conventional CNN model without transformation can not work well, and our method achieves better performance than some other applicable related methods, especially in the scenarios of small size and JPEG compression.
The redundant discrete wavelet transform and additive noise The behavior under additive noise of the redundant discrete wavelet transform (RDWT), which is a frame expansion that is essentially an undecimated discrete wavelet transform, is studied. Known prior results in the form of inequalities bound distortion energy in the original signal domain from additive noise in frame-expansion coefficients. In this letter, a precise relationship between RDWT-domai...
Deep learning for source camera identification on mobile devices. •The design of an efficient CNN architecture for the SCI problem on mobile devices.•The evaluation of different CNN configurations.•The usage of a unique dataset (MICHE-I) of images taken from several mobile devices.•A 98.1% of accuracy on model detection.•A 91.1% of accuracy on sensor detection.
Speech emotion recognition using deep 1D & 2D CNN LSTM networks. We aimed at learning deep emotion features to recognize speech emotion. Two convolutional neural network and long short-term memory (CNN LSTM) networks, one 1D CNN LSTM network and one 2D CNN LSTM network, were constructed to learn local and global emotion-related features from speech and log-mel spectrogram respectively. The two networks have the similar architecture, both consisting of four local feature learning blocks (LFLBs) and one long short-term memory (LSTM) layer. LFLB, which mainly contains one convolutional layer and one max-pooling layer, is built for learning local correlations along with extracting hierarchical correlations. LSTM layer is adopted to learn long-term dependencies from the learned local features. The designed networks, combinations of the convolutional neural network (CNN) and LSTM, can take advantage of the strengths of both networks and overcome the shortcomings of them, and are evaluated on two benchmark databases. The experimental results show that the designed networks achieve excellent performance on the task of recognizing speech emotion, especially the 2D CNN LSTM network outperforms the traditional approaches, Deep Belief Network (DBN) and CNN on the selected databases. The 2D CNN LSTM network achieves recognition accuracies of 95.33% and 95.89% on Berlin EmoDB of speaker-dependent and speaker-independent experiments respectively, which compare favourably to the accuracy of 91.6% and 92.9% obtained by traditional approaches; and also yields recognition accuracies of 89.16% and 52.14% on IEMOCAP database of speaker-dependent and speaker-independent experiments, which are much higher than the accuracy of 73.78% and 40.02% obtained by DBN and CNN.
Real-time detecting one specific tampering operation in multiple operator chains Currently, many forensic techniques have been developed to determine the processing history of given multimedia contents. However, because of the interaction among tampering operations, there are still fundamental limits on the determination of tampering order and type. Up to now, a few works consider the cases where multiple operation types are involved in. In these cases, we not only need to consider the interplay of operation order, but also should quantify the detectability of one specific operation. In this paper, we propose an efficient information theoretical framework to solve this problem. Specially, we analyze the operation detection problem from the perspective of set partitioning and detection theory. Then, under certain detectors, we present the information framework to contrast the detected hypotheses and true hypotheses. Some constraint criterions are designed to improve the detection performance of an operation. In addition, Maximum-Likelihood Estimation (MLE) is used to obtain the best detector. Finally, a multiple chain set is examined in this paper, where three efficient detection methods have been proposed and the effectiveness of our framework has been demonstrated by simulations.
Distinctive Image Features from Scale-Invariant Keypoints This paper presents a method for extracting distinctive invariant features from images that can be used to perform reliable matching between different views of an object or scene. The features are invariant to image scale and rotation, and are shown to provide robust matching across a substantial range of affine distortion, change in 3D viewpoint, addition of noise, and change in illumination. The features are highly distinctive, in the sense that a single feature can be correctly matched with high probability against a large database of features from many images. This paper also describes an approach to using these features for object recognition. The recognition proceeds by matching individual features to a database of features from known objects using a fast nearest-neighbor algorithm, followed by a Hough transform to identify clusters belonging to a single object, and finally performing verification through least-squares solution for consistent pose parameters. This approach to recognition can robustly identify objects among clutter and occlusion while achieving near real-time performance.
Rich Feature Hierarchies for Accurate Object Detection and Semantic Segmentation Object detection performance, as measured on the canonical PASCAL VOC dataset, has plateaued in the last few years. The best-performing methods are complex ensemble systems that typically combine multiple low-level image features with high-level context. In this paper, we propose a simple and scalable detection algorithm that improves mean average precision (mAP) by more than 30% relative to the previous best result on VOC 2012 -- achieving a mAP of 53.3%. Our approach combines two key insights: (1) one can apply high-capacity convolutional neural networks (CNNs) to bottom-up region proposals in order to localize and segment objects and (2) when labeled training data is scarce, supervised pre-training for an auxiliary task, followed by domain-specific fine-tuning, yields a significant performance boost. Since we combine region proposals with CNNs, we call our method R-CNN: Regions with CNN features. We also present experiments that provide insight into what the network learns, revealing a rich hierarchy of image features. Source code for the complete system is available at http://www.cs.berkeley.edu/~rbg/rcnn.
A new approach for dynamic fuzzy logic parameter tuning in Ant Colony Optimization and its application in fuzzy control of a mobile robot Central idea is to avoid or slow down full convergence through the dynamic variation of parameters.Performance of different ACO variants was observed to choose one as the basis to the proposed approach.Convergence fuzzy controller with the objective of maintaining diversity to avoid premature convergence was created. Ant Colony Optimization is a population-based meta-heuristic that exploits a form of past performance memory that is inspired by the foraging behavior of real ants. The behavior of the Ant Colony Optimization algorithm is highly dependent on the values defined for its parameters. Adaptation and parameter control are recurring themes in the field of bio-inspired optimization algorithms. The present paper explores a new fuzzy approach for diversity control in Ant Colony Optimization. The main idea is to avoid or slow down full convergence through the dynamic variation of a particular parameter. The performance of different variants of the Ant Colony Optimization algorithm is analyzed to choose one as the basis to the proposed approach. A convergence fuzzy logic controller with the objective of maintaining diversity at some level to avoid premature convergence is created. Encouraging results on several traveling salesman problem instances and its application to the design of fuzzy controllers, in particular the optimization of membership functions for a unicycle mobile robot trajectory control are presented with the proposed method.
Online Palmprint Identification Biometrics-based personal identification is regarded as an effective method for automatically recognizing, with a high confidence, a person's identity. This paper presents a new biometric approach to online personal identification using palmprint technology. In contrast to the existing methods, our online palmprint identification system employs low-resolution palmprint images to achieve effective personal identification. The system consists of two parts: a novel device for online palmprint image acquisition and an efficient algorithm for fast palmprint recognition. A robust image coordinate system is defined to facilitate image alignment for feature extraction. In addition, a 2D Gabor phase encoding scheme is proposed for palmprint feature extraction and representation. The experimental results demonstrate the feasibility of the proposed system.
Theory of Mind for a Humanoid Robot If we are to build human-like robots that can interact naturally with people, our robots must know not only about the properties of objects but also the properties of animate agents in the world. One of the fundamental social skills for humans is the attribution of beliefs, goals, and desires to other people. This set of skills has often been called a “theory of mind.” This paper presents the theories of Leslie (1994) and Baron-Cohen (1995) on the development of theory of mind in human children and discusses the potential application of both of these theories to building robots with similar capabilities. Initial implementation details and basic skills (such as finding faces and eyes and distinguishing animate from inanimate stimuli) are introduced. I further speculate on the usefulness of a robotic implementation in evaluating and comparing these two models.
Gravity-Balancing Leg Orthosis and Its Performance Evaluation In this paper, we propose a device to assist persons with hemiparesis to walk by reducing or eliminating the effects of gravity. The design of the device includes the following features: 1) it is passive, i.e., it does not include motors or actuators, but is only composed of links and springs; 2) it is safe and has a simple patient-machine interface to accommodate variability in geometry and inertia of the subjects. A number of methods have been proposed in the literature to gravity-balance a machine. Here, we use a hybrid method to achieve gravity balancing of a human leg over its range of motion. In the hybrid method, a mechanism is used to first locate the center of mass of the human limb and the orthosis. Springs are then added so that the system is gravity-balanced in every configuration. For a quantitative evaluation of the performance of the device, electromyographic (EMG) data of the key muscles, involved in the motion of the leg, were collected and analyzed. Further experiments involving leg-raising and walking tasks were performed, where data from encoders and force-torque sensors were used to compute joint torques. These experiments were performed on five healthy subjects and a stroke patient. The results showed that the EMG activity from the rectus femoris and hamstring muscles with the device was reduced by 75%, during static hip and knee flexion, respectively. For leg-raising tasks, the average torque for static positioning was reduced by 66.8% at the hip joint and 47.3% at the knee joint; however, if we include the transient portion of the leg-raising task, the average torque at the hip was reduced by 61.3%, and at the knee was increased by 2.7% at the knee joints. In the walking experiment, there was a positive impact on the range of movement at the hip and knee joints, especially for the stroke patient: the range of movement increased by 45% at the hip joint and by 85% at the knee joint. We believe that this orthosis can be potentially used to desig- - n rehabilitation protocols for patients with stroke
Biologically-inspired soft exosuit. In this paper, we present the design and evaluation of a novel soft cable-driven exosuit that can apply forces to the body to assist walking. Unlike traditional exoskeletons which contain rigid framing elements, the soft exosuit is worn like clothing, yet can generate moments at the ankle and hip with magnitudes of 18% and 30% of those naturally generated by the body during walking, respectively. Our design uses geared motors to pull on Bowden cables connected to the suit near the ankle. The suit has the advantages over a traditional exoskeleton in that the wearer's joints are unconstrained by external rigid structures, and the worn part of the suit is extremely light, which minimizes the suit's unintentional interference with the body's natural biomechanics. However, a soft suit presents challenges related to actuation force transfer and control, since the body is compliant and cannot support large pressures comfortably. We discuss the design of the suit and actuation system, including principles by which soft suits can transfer force to the body effectively and the biological inspiration for the design. For a soft exosuit, an important design parameter is the combined effective stiffness of the suit and its interface to the wearer. We characterize the exosuit's effective stiffness, and present preliminary results from it generating assistive torques to a subject during walking. We envision such an exosuit having broad applicability for assisting healthy individuals as well as those with muscle weakness.
Inter-class sparsity based discriminative least square regression Least square regression is a very popular supervised classification method. However, two main issues greatly limit its performance. The first one is that it only focuses on fitting the input features to the corresponding output labels while ignoring the correlations among samples. The second one is that the used label matrix, i.e., zero–one label matrix is inappropriate for classification. To solve these problems and improve the performance, this paper presents a novel method, i.e., inter-class sparsity based discriminative least square regression (ICS_DLSR), for multi-class classification. Different from other methods, the proposed method pursues that the transformed samples have a common sparsity structure in each class. For this goal, an inter-class sparsity constraint is introduced to the least square regression model such that the margins of samples from the same class can be greatly reduced while those of samples from different classes can be enlarged. In addition, an error term with row-sparsity constraint is introduced to relax the strict zero–one label matrix, which allows the method to be more flexible in learning the discriminative transformation matrix. These factors encourage the method to learn a more compact and discriminative transformation for regression and thus has the potential to perform better than other methods. Extensive experimental results show that the proposed method achieves the best performance in comparison with other methods for multi-class classification.
Social Robots for (Second) Language Learning in (Migrant) Primary School Children Especially these days, innovation and support from technology to relieve pressure in education is highly urgent. This study tested the potential advantage of a social robot over a tablet in (second) language learning on performance, engagement, and enjoyment. Shortages in primary education call for new technology solutions. Previous studies combined robots with tablets, to compensate for robot’s limitations, however, this study applied direct human–robot interaction. Primary school children (N = 63, aged 4–6) participated in a 3-wave field experiment with story-telling exercises, either with a semi-autonomous robot (without tablet, using WOz) or a tablet. Results showed increased learning gains over time when training with a social robot, compared to the tablet. Children who trained with a robot were more engaged in the story-telling task and enjoyed it more. Robot’s behavioral style (social or neutral) hardly differed overall, however, seems to vary for high versus low educational abilities. While social robots need sophistication before being implemented in schools, our study shows the potential of social robots as tutors in (second) language learning.
1.2
0.2
0.2
0.2
0.066667
0
0
0
0
0
0
0
0
0
Universal Adversarial Triggers for Attacking and Analyzing NLP
NLTK: the natural language toolkit The Natural Language Toolkit is a suite of program modules, data sets, tutorials and exercises, covering symbolic and statistical natural language processing. NLTK is written in Python and distributed under the GPL open source license. Over the past three years, NLTK has become popular in teaching and research. We describe the toolkit and report on its current state of development.
WP:clubhouse?: an exploration of Wikipedia's gender imbalance Wikipedia has rapidly become an invaluable destination for millions of information-seeking users. However, media reports suggest an important challenge: only a small fraction of Wikipedia's legion of volunteer editors are female. In the current work, we present a scientific exploration of the gender imbalance in the English Wikipedia's population of editors. We look at the nature of the imbalance itself, its effects on the quality of the encyclopedia, and several conflict-related factors that may be contributing to the gender gap. Our findings confirm the presence of a large gender gap among editors and a corresponding gender-oriented disparity in the content of Wikipedia's articles. Further, we find evidence hinting at a culture that may be resistant to female participation.
Understanding Back-Translation at Scale. An effective method to improve neural machine translation with monolingual data is to augment the parallel training corpus with back-translations of target language sentences. This work broadens the understanding of back-translation and investigates a number of methods to generate synthetic source sentences. We find that in all but resource poor settings back-translations obtained via sampling or noised beam outputs are most effective. Our analysis shows that sampling or noisy synthetic data gives a much stronger training signal than data generated by beam or greedy search. We also compare how synthetic data compares to genuine bitext and study various domain effects. Finally, we scale to hundreds of millions of monolingual sentences and achieve a new state of the art of 35 BLEU on the WMTu002714 English-German test set.
Language (Technology) is Power: A Critical Survey of "Bias" in NLP We survey 146 papers analyzing "bias" in NLP systems, finding that their motivations are often vague, inconsistent, and lacking in normative reasoning, despite the fact that analyzing "bias" is an inherently normative process. We further find that these papers' proposed quantitative techniques for measuring or mitigating "bias" are poorly matched to their motivations and do not engage with the relevant literature outside of NLP. Based on these findings, we describe the beginnings of a path forward by proposing three recommendations that should guide work analyzing "bias" in NLP systems. These recommendations rest on a greater recognition of the relationships between language and social hierarchies, encouraging researchers and practitioners to articulate their conceptualizations of "bias"---i.e., what kinds of system behaviors are harmful, in what ways, to whom, and why, as well as the normative reasoning underlying these statements---and to center work around the lived experiences of members of communities affected by NLP systems, while interrogating and reimagining the power relations between technologists and such communities.
Gender Bias in Contextualized Word Embeddings. In this paper, we quantify, analyze and mitigate gender bias exhibited in ELMou0027s contextualized word vectors. First, we conduct several intrinsic analyses and find that (1) training data for ELMo contains significantly more male than female entities, (2) the trained ELMo embeddings systematically encode gender information and (3) ELMo unequally encodes gender information about male and female entities. Then, we show that a state-of-the-art coreference system that depends on ELMo inherits its bias and demonstrates significant bias on the WinoBias probing corpus. Finally, we explore two methods to mitigate such gender bias and show that the bias demonstrated on WinoBias can be eliminated.
A Multi-Task Approach for Disentangling Syntax and Semantics in Sentence Representations. We propose a generative model for a sentence that uses two latent variables, with one intended to represent the syntax of the sentence and the other to represent its semantics. We show we can achieve better disentanglement between semantic and syntactic representations by training with multiple losses, including losses that exploit aligned paraphrastic sentences and word-order information. We also investigate the effect of moving from bag-of-words to recurrent neural network modules. We evaluate our models as well as several popular pretrained embeddings on standard semantic similarity tasks and novel syntactic similarity tasks. Empirically, we find that the model with the best performing syntactic and semantic representations also gives rise to the most disentangled representations.
BLEU: a method for automatic evaluation of machine translation Human evaluations of machine translation are extensive but expensive. Human evaluations can take months to finish and involve human labor that can not be reused. We propose a method of automatic machine translation evaluation that is quick, inexpensive, and language-independent, that correlates highly with human evaluation, and that has little marginal cost per run. We present this method as an automated understudy to skilled human judges which substitutes for them when there is need for quick or frequent evaluations.
Model-based periodic event-triggered control for linear systems Periodic event-triggered control (PETC) is a control strategy that combines ideas from conventional periodic sampled-data control and event-triggered control. By communicating periodically sampled sensor and controller data only when needed to guarantee stability or performance properties, PETC is capable of reducing the number of transmissions significantly, while still retaining a satisfactory closed-loop behavior. In this paper, we will study observer-based controllers for linear systems and propose advanced event-triggering mechanisms (ETMs) that will reduce communication in both the sensor-to-controller channels and the controller-to-actuator channels. By exploiting model-based computations, the new classes of ETMs will outperform existing ETMs in the literature. To model and analyze the proposed classes of ETMs, we present two frameworks based on perturbed linear and piecewise linear systems, leading to conditions for global exponential stability and @?"2-gain performance of the resulting closed-loop systems in terms of linear matrix inequalities. The proposed analysis frameworks can be used to make tradeoffs between the network utilization on the one hand and the performance in terms of @?"2-gains on the other. In addition, we will show that the closed-loop performance realized by an observer-based controller, implemented in a conventional periodic time-triggered fashion, can be recovered arbitrarily closely by a PETC implementation. This provides a justification for emulation-based design. Next to centralized model-based ETMs, we will also provide a decentralized setup suitable for large-scale systems, where sensors and actuators are physically distributed over a wide area. The improvements realized by the proposed model-based ETMs will be demonstrated using numerical examples.
A Probabilistic Approach to Collaborative Multi-Robot Localization This paper presents a statistical algorithm for collaborative mobile robot localization. Our approach uses a sample-based version of Markov localization, capable of localizing mobile robots in an any-time fashion. When teams of robots localize themselves in the same environment, probabilistic methods are employed to synchronize each robot's belief whenever one robot detects another. As a result, the robots localize themselves faster, maintain higher accuracy, and high-cost sensors are amortized across multiple robot platforms. The technique has been implemented and tested using two mobile robots equipped with cameras and laser range-finders for detecting other robots. The results, obtained with the real robots and in series of simulation runs, illustrate drastic improvements in localization speed and accuracy when compared to conventional single-robot localization. A further experiment demonstrates that under certain conditions, successful localization is only possible if teams of heterogeneous robots collaborate during localization.
Millimeter Wave Cellular Wireless Networks: Potentials and Challenges. Millimeter-wave (mmW) frequencies between 30 and 300 GHz are a new frontier for cellular communication that offers the promise of orders of magnitude greater bandwidths combined with further gains via beamforming and spatial multiplexing from multielement antenna arrays. This paper surveys measurements and capacity studies to assess this technology with a focus on small cell deployments in urban e...
Let me tell you! investigating the effects of robot communication strategies in advice-giving situations based on robot appearance, interaction modality and distance Recent proposals for how robots should talk to people when they give advice suggest that the same strategies humans employ with other humans are effective for robots as well. However, the evidence is exclusively based on people's observation of robot giving advice to other humans. Hence, it is not clear whether the results still apply when people actually participate in real interactions with robots. We address this shortcoming in a novel systematic mixed-methods study where we employ both survey-based subjective and brain-based objective measures (using functional near infrared spectroscopy). The results show that previous results from observation conditions do not transfer automatically to interaction conditions, and that robot appearance and interaction distance are important modulators of human perceptions of robot behavior in advice-giving contexts.
Safe mutations for deep and recurrent neural networks through output gradients While neuroevolution (evolving neural networks) has been successful across a variety of domains from reinforcement learning, to artificial life, to evolutionary robotics, it is rarely applied to large, deep neural networks. A central reason is that while random mutation generally works in low dimensions, a random perturbation of thousands or millions of weights will likely break existing functionality. This paper proposes a solution: a family of safe mutation (SM) operators that facilitate exploration without dramatically altering network behavior or requiring additional interaction with the environment. The most effective SM variant scales the degree of mutation of each individual weight according to the sensitivity of the network's outputs to that weight, which requires computing the gradient of outputs with respect to the weights (instead of the gradient of error, as in conventional deep learning). This safe mutation through gradients (SM-G) operator dramatically increases the ability of a simple genetic algorithm-based neuroevolution method to find solutions in high-dimensional domains that require deep and/or recurrent neural networks, including domains that require processing raw pixels. By improving our ability to evolve deep neural networks, this new safer approach to mutation expands the scope of domains amenable to neuroevolution.
Robot tutor and pupils’ educational ability: Teaching the times tables Research shows promising results of educational robots in language and STEM tasks. In language, more research is available, occasionally in view of individual differences in pupils’ educational ability levels, and learning seems to improve with more expressive robot behaviors. In STEM, variations in robots’ behaviors have been examined with inconclusive results and never while systematically investigating how differences in educational abilities match with different robot behaviors. We applied an autonomously tutoring robot (without tablet, partly WOz) in a 2 × 2 experiment of social vs. neutral behavior in above-average vs. below-average schoolchildren (N = 86; age 8–10 years) while rehearsing the multiplication tables on a one-to-one basis. The standard school test showed that on average, pupils significantly improved their performance even after 3 occasions of 5-min exercises. Beyond-average pupils profited most from a robot tutor, whereas those below average in multiplication benefited more from a robot that showed neutral rather than more social behavior.
1.073556
0.08
0.08
0.08
0.08
0.074
0.066667
0.006618
0
0
0
0
0
0
Bayesian AirComp with Sign-Alignment Precoding for Wireless Federated Learning In this paper, we consider the problem of wireless federated learning based on sign stochastic gradient descent (signSGD) algorithm via a multiple access channel. When sending locally computed gradient's sign information, each mobile device requires to apply precoding to circumvent wireless fading effects. In practice, however, acquiring perfect knowledge of channel state information (CSI) at all mobile devices is infeasible. In this paper, we present a simple yet effective precoding method with limited channel knowledge, called sign-alignment precoding. The idea of sign-alignment precoding is to protect sign-flipping errors from wireless fadings. Under the Gaussian prior assumption on the local gradients, we also derive the mean squared error (MSE)-optimal aggregation function called Bayesian over-the-air computation (BayAirComp). Our key finding is that one-bit precoding with BayAirComp aggregation can provide a better learning performance than the existing precoding method even using perfect CSI with AirComp aggregation.
The Sybil Attack Large-scale peer-to-peer systems facesecurity threats from faulty or hostile remotecomputing elements. To resist these threats, manysuch systems employ redundancy. However, if asingle faulty entity can present multiple identities,it can control a substantial fraction of the system,thereby undermining this redundancy. Oneapproach to preventing these &quot;Sybil attacks&quot; is tohave a trusted agency certify identities. Thispaper shows that, without a logically centralizedauthority, Sybil...
BLEU: a method for automatic evaluation of machine translation Human evaluations of machine translation are extensive but expensive. Human evaluations can take months to finish and involve human labor that can not be reused. We propose a method of automatic machine translation evaluation that is quick, inexpensive, and language-independent, that correlates highly with human evaluation, and that has little marginal cost per run. We present this method as an automated understudy to skilled human judges which substitutes for them when there is need for quick or frequent evaluations.
Computational thinking Summary form only given. My vision for the 21st century, Computational Thinking, will be a fundamental skill used by everyone in the world. To reading, writing, and arithmetic, we should add computational thinking to every child's analytical ability. Computational thinking involves solving problems, designing systems, and understanding human behavior by drawing on the concepts fundamental to computer science. Thinking like a computer scientist means more than being able to program a computer. It requires the ability to abstract and thus to think at multiple levels of abstraction. In this talk I will give many examples of computational thinking, argue that it has already influenced other disciplines, and promote the idea that teaching computational thinking can not only inspire future generations to enter the field of computer science but benefit people in all fields.
Fuzzy logic in control systems: fuzzy logic controller. I.
Switching between stabilizing controllers This paper deals with the problem of switching between several linear time-invariant (LTI) controllers—all of them capable of stabilizing a speci4c LTI process—in such a way that the stability of the closed-loop system is guaranteed for any switching sequence. We show that it is possible to 4nd realizations for any given family of controller transfer matrices so that the closed-loop system remains stable, no matter how we switch among the controller. The motivation for this problem is the control of complex systems where con8icting requirements make a single LTI controller unsuitable. ? 2002 Published by Elsevier Science Ltd.
Tabu Search - Part I
Bidirectional recurrent neural networks In the first part of this paper, a regular recurrent neural network (RNN) is extended to a bidirectional recurrent neural network (BRNN). The BRNN can be trained without the limitation of using input information just up to a preset future frame. This is accomplished by training it simultaneously in positive and negative time direction. Structure and training procedure of the proposed network are explained. In regression and classification experiments on artificial data, the proposed structure gives better results than other approaches. For real data, classification experiments for phonemes from the TIMIT database show the same tendency. In the second part of this paper, it is shown how the proposed bidirectional structure can be easily modified to allow efficient estimation of the conditional posterior probability of complete symbol sequences without making any explicit assumption about the shape of the distribution. For this part, experiments on real data are reported
An intensive survey of fair non-repudiation protocols With the phenomenal growth of the Internet and open networks in general, security services, such as non-repudiation, become crucial to many applications. Non-repudiation services must ensure that when Alice sends some information to Bob over a network, neither Alice nor Bob can deny having participated in a part or the whole of this communication. Therefore a fair non-repudiation protocol has to generate non-repudiation of origin evidences intended to Bob, and non-repudiation of receipt evidences destined to Alice. In this paper, we clearly define the properties a fair non-repudiation protocol must respect, and give a survey of the most important non-repudiation protocols without and with trusted third party (TTP). For the later ones we discuss the evolution of the TTP's involvement and, between others, describe the most recent protocol using a transparent TTP. We also discuss some ad-hoc problems related to the management of non-repudiation evidences.
Dynamic movement and positioning of embodied agents in multiparty conversations For embodied agents to engage in realistic multiparty conversation, they must stand in appropriate places with respect to other agents and the environment. When these factors change, such as an agent joining the conversation, the agents must dynamically move to a new location and/or orientation to accommodate. This paper presents an algorithm for simulating movement of agents based on observed human behavior using techniques developed for pedestrian movement in crowd simulations. We extend a previous group conversation simulation to include an agent motion algorithm. We examine several test cases and show how the simulation generates results that mirror real-life conversation settings.
An improved genetic algorithm with conditional genetic operators and its application to set-covering problem The genetic algorithm (GA) is a popular, biologically inspired optimization method. However, in the GA there is no rule of thumb to design the GA operators and select GA parameters. Instead, trial-and-error has to be applied. In this paper we present an improved genetic algorithm in which crossover and mutation are performed conditionally instead of probability. Because there are no crossover rate and mutation rate to be selected, the proposed improved GA can be more easily applied to a problem than the conventional genetic algorithms. The proposed improved genetic algorithm is applied to solve the set-covering problem. Experimental studies show that the improved GA produces better results over the conventional one and other methods.
Lane-level traffic estimations using microscopic traffic variables This paper proposes a novel inference method to estimate lane-level traffic flow, time occupancy and vehicle inter-arrival time on road segments where local information could not be measured and assessed directly. The main contributions of the proposed method are 1) the ability to perform lane-level estimations of traffic flow, time occupancy and vehicle inter-arrival time and 2) the ability to adapt to different traffic regimes by assessing only microscopic traffic variables. We propose a modified Kriging estimation model which explicitly takes into account both spatial and temporal variability. Performance evaluations are conducted using real-world data under different traffic regimes and it is shown that the proposed method outperforms a Kalman filter-based approach.
Convolutional Neural Network-Based Classification of Driver's Emotion during Aggressive and Smooth Driving Using Multi-Modal Camera Sensors. Because aggressive driving often causes large-scale loss of life and property, techniques for advance detection of adverse driver emotional states have become important for the prevention of aggressive driving behaviors. Previous studies have primarily focused on systems for detecting aggressive driver emotion via smart-phone accelerometers and gyro-sensors, or they focused on methods of detecting physiological signals using electroencephalography (EEG) or electrocardiogram (ECG) sensors. Because EEG and ECG sensors cause discomfort to drivers and can be detached from the driver's body, it becomes difficult to focus on bio-signals to determine their emotional state. Gyro-sensors and accelerometers depend on the performance of GPS receivers and cannot be used in areas where GPS signals are blocked. Moreover, if driving on a mountain road with many quick turns, a driver's emotional state can easily be misrecognized as that of an aggressive driver. To resolve these problems, we propose a convolutional neural network (CNN)-based method of detecting emotion to identify aggressive driving using input images of the driver's face, obtained using near-infrared (NIR) light and thermal camera sensors. In this research, we conducted an experiment using our own database, which provides a high classification accuracy for detecting driver emotion leading to either aggressive or smooth (i.e., relaxed) driving. Our proposed method demonstrates better performance than existing methods.
Robot tutor and pupils’ educational ability: Teaching the times tables Research shows promising results of educational robots in language and STEM tasks. In language, more research is available, occasionally in view of individual differences in pupils’ educational ability levels, and learning seems to improve with more expressive robot behaviors. In STEM, variations in robots’ behaviors have been examined with inconclusive results and never while systematically investigating how differences in educational abilities match with different robot behaviors. We applied an autonomously tutoring robot (without tablet, partly WOz) in a 2 × 2 experiment of social vs. neutral behavior in above-average vs. below-average schoolchildren (N = 86; age 8–10 years) while rehearsing the multiplication tables on a one-to-one basis. The standard school test showed that on average, pupils significantly improved their performance even after 3 occasions of 5-min exercises. Beyond-average pupils profited most from a robot tutor, whereas those below average in multiplication benefited more from a robot that showed neutral rather than more social behavior.
1
0
0
0
0
0
0
0
0
0
0
0
0
0
Detecting Compressed Deepfake Videos in Social Networks Using Frame-Temporality Two-Stream Convolutional Network The development of technologies that can generate Deepfake videos is expanding rapidly. These videos are easily synthesized without leaving obvious traces of manipulation. Though forensically detection in high-definition video datasets has achieved remarkable results, the forensics of compressed videos is worth further exploring. In fact, compressed videos are common in social networks, such as vi...
FaceForensics++: Learning to Detect Manipulated Facial Images The rapid progress in synthetic image generation and manipulation has now come to a point where it raises significant concerns for the implications towards society. At best, this leads to a loss of trust in digital content, but could potentially cause further harm by spreading false information or fake news. This paper examines the realism of state-of-the-art image manipulations, and how difficult it is to detect them, either automatically or by humans. To standardize the evaluation of detection methods, we propose an automated benchmark for facial manipulation detection. In particular, the benchmark is based on Deep-Fakes, Face2Face, FaceSwap and NeuralTextures as prominent representatives for facial manipulations at random compression level and size. The benchmark is publicly available and contains a hidden test set as well as a database of over 1.8 million manipulated images. This dataset is over an order of magnitude larger than comparable, publicly available, forgery datasets. Based on this data, we performed a thorough analysis of data-driven forgery detectors. We show that the use of additional domain-specific knowledge improves forgery detection to unprecedented accuracy, even in the presence of strong compression, and clearly outperforms human observers.
Identification of deep network generated images using disparities in color components •Disparities between real and GAN-generated images are analyzed in different color spaces.•Generated images are more distinguishable in residual domain of chrominance components.•Mismatched training-testing data and one-class classification are evaluated in experiments.
Deep High-Resolution Representation Learning for Visual Recognition High-resolution representations are essential for position-sensitive vision problems, such as human pose estimation, semantic segmentation, and object detection. Existing state-of-the-art frameworks first encode the input image as a low-resolution representation through a subnetwork that is formed by connecting high-to-low resolution convolutions in series (e.g., ResNet, VGGNet), ...
Spatiotemporal Trident Networks: Detection and Localization of Object Removal Tampering in Video Passive Forensics With the development of video and image processing technology, the field of video tampering forensics is facing enormous challenges. Specifically, as the fundamental basis of judicial forensics, passive forensics for object removal video forgery is particularly essential. To extract tampering traces in video more sufficiently, the author proposed a spatiotemporal trident network based on the spati...
Spatiotemporal Inconsistency Learning for DeepFake Video Detection ABSTRACTThe rapid development of facial manipulation techniques has aroused public concerns in recent years. Following the success of deep learning, existing methods always formulate DeepFake video detection as a binary classification problem and develop frame-based and video-based solutions. However, little attention has been paid to capturing the spatial-temporal inconsistency in forged videos. To address this issue, we term this task as a Spatial-Temporal Inconsistency Learning (STIL) process and instantiate it into a novel STIL block, which consists of a Spatial Inconsistency Module (SIM), a Temporal Inconsistency Module (TIM), and an Information Supplement Module (ISM). Specifically, we present a novel temporal modeling paradigm in TIM by exploiting the temporal difference over adjacent frames along with both horizontal and vertical directions. And the ISM simultaneously utilizes the spatial information from SIM and temporal information from TIM to establish a more comprehensive spatial-temporal representation. Moreover, our STIL block is flexible and could be plugged into existing 2D CNNs. Extensive experiments and visualizations are presented to demonstrate the effectiveness of our method against the state-of-the-art competitors.
Hybrid LSTM and Encoder-Decoder Architecture for Detection of Image Forgeries. With advanced image journaling tools, one can easily alter the semantic meaning of an image by exploiting certain manipulation techniques such as copy clone, object splicing, and removal, which mislead the viewers. In contrast, the identification of these manipulations becomes a very challenging task as manipulated regions are not visually apparent. This paper proposes a high-confidence manipulati...
Rich Feature Hierarchies for Accurate Object Detection and Semantic Segmentation Object detection performance, as measured on the canonical PASCAL VOC dataset, has plateaued in the last few years. The best-performing methods are complex ensemble systems that typically combine multiple low-level image features with high-level context. In this paper, we propose a simple and scalable detection algorithm that improves mean average precision (mAP) by more than 30% relative to the previous best result on VOC 2012 -- achieving a mAP of 53.3%. Our approach combines two key insights: (1) one can apply high-capacity convolutional neural networks (CNNs) to bottom-up region proposals in order to localize and segment objects and (2) when labeled training data is scarce, supervised pre-training for an auxiliary task, followed by domain-specific fine-tuning, yields a significant performance boost. Since we combine region proposals with CNNs, we call our method R-CNN: Regions with CNN features. We also present experiments that provide insight into what the network learns, revealing a rich hierarchy of image features. Source code for the complete system is available at http://www.cs.berkeley.edu/~rbg/rcnn.
Task Offloading in Vehicular Edge Computing Networks: A Load-Balancing Solution Recently, the rapid advance of vehicular networks has led to the emergence of diverse delay-sensitive vehicular applications such as automatic driving, auto navigation. Note that existing resource-constrained vehicles cannot adequately meet these demands on low / ultra-low latency. By offloading parts of the vehicles’ compute-intensive tasks to the edge servers in proximity, mobile edge computing is envisioned as a promising paradigm, giving rise to the vehicular edge computing networks (VECNs). However, most existing works on task offloading in VECNs did not take the load balancing of the computation resources at the edge servers into account. To address these issues and given the high dynamics of vehicular networks, we introduce fiber-wireless (FiWi) technology to enhance VECNs, due to its advantages on centralized network management and supporting multiple communication techniques. Aiming to minimize the processing delay of the vehicles’ computation tasks, we propose a software-defined networking (SDN) based load-balancing task offloading scheme in FiWi enhanced VECNs, where SDN is introduced to provide supports for the centralized network and vehicle information management. Extensive analysis and numerical results corroborate that our proposed load-balancing scheme can achieve superior performance on processing delay reduction by utilizing the edge servers’ computation resources more efficiently.
Visual cryptography for general access structures A visual cryptography scheme for a set P of n participants is a method of encoding a secret image SI into n shadow images called shares, where each participant in P receives one share. Certain qualified subsets of participants can “visually” recover the secret image, but other, forbidden, sets of participants have no information (in an information-theoretic sense) on SI . A “visual” recovery for a set X ⊆ P consists of xeroxing the shares given to the participants in X onto transparencies, and then stacking them. The participants in a qualified set X will be able to see the secret image without any knowledge of cryptography and without performing any cryptographic computation. In this paper we propose two techniques for constructing visual cryptography schemes for general access structures. We analyze the structure of visual cryptography schemes and we prove bounds on the size of the shares distributed to the participants in the scheme. We provide a novel technique for realizing k out of n threshold visual cryptography schemes. Our construction for k out of n visual cryptography schemes is better with respect to pixel expansion than the one proposed by M. Naor and A. Shamir (Visual cryptography, in “Advances in Cryptology—Eurocrypt '94” CA. De Santis, Ed.), Lecture Notes in Computer Science, Vol. 950, pp. 1–12, Springer-Verlag, Berlin, 1995) and for the case of 2 out of n is the best possible. Finally, we consider graph-based access structures, i.e., access structures in which any qualified set of participants contains at least an edge of a given graph whose vertices represent the participants of the scheme.
IntrospectiveViews: an interface for scrutinizing semantic user models User models are a key component for user-adaptive systems They represent information about users such as interests, expertise, goals, traits, etc This information is used to achieve various adaptation effects, e.g., recommending relevant documents or products To ensure acceptance by users, these models need to be scrutable, i.e., users must be able to view and alter them to understand and if necessary correct the assumptions the system makes about the user However, in most existing systems, this goal is not met In this paper, we introduce IntrospectiveViews, an interface that enables the user to view and edit her user model Furthermore, we present the results of a formative evaluation that show the importance users give in general to different aspects of scrutable user models and also substantiate our claim that IntrospectiveViews is an appropriate realization of an interface to such models.
A review on interval type-2 fuzzy logic applications in intelligent control. A review of the applications of interval type-2 fuzzy logic in intelligent control has been considered in this paper. The fundamental focus of the paper is based on the basic reasons for using type-2 fuzzy controllers for different areas of application. Recently, bio-inspired methods have emerged as powerful optimization algorithms for solving complex problems. In the case of designing type-2 fuzzy controllers for particular applications, the use of bio-inspired optimization methods have helped in the complex task of finding the appropriate parameter values and structure of the fuzzy systems. In this review, we consider the application of genetic algorithms, particle swarm optimization and ant colony optimization as three different paradigms that help in the design of optimal type-2 fuzzy controllers. We also mention alternative approaches to designing type-2 fuzzy controllers without optimization techniques.
Design of robust fuzzy fault detection filter for polynomial fuzzy systems with new finite frequency specifications This paper investigates the problem of fault detection filter design for discrete-time polynomial fuzzy systems with faults and unknown disturbances. The frequency ranges of the faults and the disturbances are assumed to be known beforehand and to reside in low, middle or high frequency ranges. Thus, the proposed filter is designed in the finite frequency range to overcome the conservatism generated by those designed in the full frequency domain. Being of polynomial fuzzy structure, the proposed filter combines the H−/H∞ performances in order to ensure the best robustness to the disturbance and the best sensitivity to the fault. Design conditions are derived in Sum Of Squares formulations that can be easily solved via available software tools. Two illustrative examples are introduced to demonstrate the effectiveness of the proposed method and a comparative study with LMI method is also provided.
A Hierarchical Architecture Using Biased Min-Consensus for USV Path Planning This paper proposes a hierarchical architecture using the biased min-consensus (BMC) method, to solve the path planning problem of unmanned surface vessel (USV). We take the fixed-point monitoring mission as an example, where a series of intermediate monitoring points should be visited once by USV. The whole framework incorporates the low-level layer planning the standard path between any two intermediate points, and the high-level fashion determining their visiting sequence. First, the optimal standard path in terms of voyage time and risk measure is planned by the BMC protocol, given that the corresponding graph is constructed with node state and edge weight. The USV will avoid obstacles or keep a certain distance safely, and arrive at the target point quickly. It is proven theoretically that the state of the graph will converge to be stable after finite iterations, i.e., the optimal solution can be found by BMC with low calculation complexity. Second, by incorporating the constraint of intermediate points, their visiting sequence is optimized by BMC again with the reconstruction of a new virtual graph based on the former planned results. The extensive simulation results in various scenarios also validate the feasibility and effectiveness of our method for autonomous navigation.
1.04
0.04
0.04
0.04
0.04
0.04
0.008
0
0
0
0
0
0
0
An Improved IOHMM-Based Stochastic Driver Lane-Changing Model The prediction and estimation of the lane-changing state of the host car and surrounding cars are important parts of an advanced driving assistant system, which mainly depend on the understanding of the driver lane-changing behavior. To learn driver lane-changing maneuver well, this article provides a novel stochastic driver lane-changing model based on an improved input–output hidden Markov model...
Analysing user physiological responses for affective video summarisation. Video summarisation techniques aim to abstract the most significant content from a video stream. This is typically achieved by processing low-level image, audio and text features which are still quite disparate from the high-level semantics that end users identify with (the ‘semantic gap’). Physiological responses are potentially rich indicators of memorable or emotionally engaging video content for a given user. Consequently, we investigate whether they may serve as a suitable basis for a video summarisation technique by analysing a range of user physiological response measures, specifically electro-dermal response (EDR), respiration amplitude (RA), respiration rate (RR), blood volume pulse (BVP) and heart rate (HR), in response to a range of video content in a variety of genres including horror, comedy, drama, sci-fi and action. We present an analysis framework for processing the user responses to specific sub-segments within a video stream based on percent rank value normalisation. The application of the analysis framework reveals that users respond significantly to the most entertaining video sub-segments in a range of content domains. Specifically, horror content seems to elicit significant EDR, RA, RR and BVP responses, and comedy content elicits comparatively lower levels of EDR, but does seem to elicit significant RA, RR, BVP and HR responses. Drama content seems to elicit less significant physiological responses in general, and both sci-fi and action content seem to elicit significant EDR responses. We discuss the implications this may have for future affective video summarisation approaches.
On the roles of eye gaze and head dynamics in predicting driver's intent to change lanes Driver behavioral cues may present a rich source of information and feedback for future intelligent advanced driver-assistance systems (ADASs). With the design of a simple and robust ADAS in mind, we are interested in determining the most important driver cues for distinguishing driver intent. Eye gaze may provide a more accurate proxy than head movement for determining driver attention, whereas the measurement of head motion is less cumbersome and more reliable in harsh driving conditions. We use a lane-change intent-prediction system (McCall et al., 2007) to determine the relative usefulness of each cue for determining intent. Various combinations of input data are presented to a discriminative classifier, which is trained to output a prediction of probable lane-change maneuver at a particular point in the future. Quantitative results from a naturalistic driving study are presented and show that head motion, when combined with lane position and vehicle dynamics, is a reliable cue for lane-change intent prediction. The addition of eye gaze does not improve performance as much as simpler head dynamics cues. The advantage of head data over eye data is shown to be statistically significant (p
Detection of Driver Fatigue Caused by Sleep Deprivation This paper aims to provide reliable indications of driver drowsiness based on the characteristics of driver-vehicle interaction. A test bed was built under a simulated driving environment, and a total of 12 subjects participated in two experiment sessions requiring different levels of sleep (partial sleep-deprivation versus no sleep-deprivation) before the experiment. The performance of the subjects was analyzed in a series of stimulus-response and routine driving tasks, which revealed the performance differences of drivers under different sleep-deprivation levels. The experiments further demonstrated that sleep deprivation had greater effect on rule-based than on skill-based cognitive functions: when drivers were sleep-deprived, their performance of responding to unexpected disturbances degraded, while they were robust enough to continue the routine driving tasks such as lane tracking, vehicle following, and lane changing. In addition, we presented both qualitative and quantitative guidelines for designing drowsy-driver detection systems in a probabilistic framework based on the paradigm of Bayesian networks. Temporal aspects of drowsiness and individual differences of subjects were addressed in the framework.
Online Prediction of Driver Distraction Based on Brain Activity Patterns This paper presents a new computational framework for early detection of driver distractions (map viewing) using brain activity measured by electroencephalographic (EEG) signals. Compared with most studies in the literature, which are mainly focused on the classification of distracted and nondistracted periods, this study proposes a new framework to prospectively predict the start and end of a distraction period, defined by map viewing. The proposed prediction algorithm was tested on a data set of continuous EEG signals recorded from 24 subjects. During the EEG recordings, the subjects were asked to drive from an initial position to a destination using a city map in a simulated driving environment. The overall accuracy values for the prediction of the start and the end of map viewing were 81% and 70%, respectively. The experimental results demonstrated that the proposed algorithm can predict the start and end of map viewing with relatively high accuracy and can be generalized to individual subjects. The outcome of this study has a high potential to improve the design of future intelligent navigation systems. Prediction of the start of map viewing can be used to provide route information based on a driver's needs and consequently avoid map-viewing activities. Prediction of the end of map viewing can be used to provide warnings for potential long map-viewing durations. Further development of the proposed framework and its applications in driver-distraction predictions are also discussed.
Keep Your Scanners Peeled: Gaze Behavior as a Measure of Automation Trust During Highly Automated Driving. Objective: The feasibility of measuring drivers' automation trust via gaze behavior during highly automated driving was assessed with eye tracking and validated with self-reported automation trust in a driving simulator study. Background: Earlier research from other domains indicates that drivers' automation trust might be inferred from gaze behavior, such as monitoring frequency. Method: The gaze behavior and self-reported automation trust of 35 participants attending to a visually demanding non-driving-related task (NDRT) during highly automated driving was evaluated. The relationship between dispositional, situational, and learned automation trust with gaze behavior was compared. Results: Overall, there was a consistent relationship between drivers' automation trust and gaze behavior. Participants reporting higher automation trust tended to monitor the automation less frequently. Further analyses revealed that higher automation trust was associated with lower monitoring frequency of the automation during NDRTs, and an increase in trust over the experimental session was connected with a decrease in monitoring frequency. Conclusion: We suggest that (a) the current results indicate a negative relationship between drivers' self-reported automation trust and monitoring frequency, (b) gaze behavior provides a more direct measure of automation trust than other behavioral measures, and (c) with further refinement, drivers' automation trust during highly automated driving might be inferred from gaze behavior. Application: Potential applications of this research include the estimation of drivers' automation trust and reliance during highly automated driving.
3D separable convolutional neural network for dynamic hand gesture recognition. •The Frame Difference method is used to pre-process the input in order to filter the background.•A 3D separable CNN is proposed for dynamic gesture recognition. The standard 3D convolution process is decomposed into two processes: 3D depth-wise and 3D point-wise.•By the application of skip connection and layer-wise learning rate, the undesirable gradient dispersion due to the separation operation is solved and the performance of the network is improved.•A dynamic hand gesture library is built through HoloLens.
Cooperative channel assignment for VANETs based on multiagent reinforcement learning. Dynamic channel assignment (DCA) plays a key role in extending vehicular ad-hoc network capacity and mitigating congestion. However, channel assignment under vehicular direct communication scenarios faces mutual influence of large-scale nodes, the lack of centralized coordination, unknown global state information, and other challenges. To solve this problem, a multiagent reinforcement learning (RL) based cooperative DCA (RL-CDCA) mechanism is proposed. Specifically, each vehicular node can successfully learn the proper strategies of channel selection and backoff adaptation from the real-time channel state information (CSI) using two cooperative RL models. In addition, neural networks are constructed as nonlinear Q-function approximators, which facilitates the mapping of the continuously sensed input to the mixed policy output. Nodes are driven to locally share and incorporate their individual rewards such that they can optimize their policies in a distributed collaborative manner. Simulation results show that the proposed multiagent RL-CDCA can better reduce the one-hop packet delay by no less than 73.73%, improve the packet delivery ratio by no less than 12.66% on average in a highly dense situation, and improve the fairness of the global network resource allocation.
Reinforcement learning based data fusion method for multi-sensors In order to improve detection system robustness and reliability, multi-sensors fusion is used in modern air combat. In this paper, a data fusion method based on reinforcement learning is developed for multi-sensors. Initially, the cubic B-spline interpolation is used to solve time alignment problems of multisource data. Then, the reinforcement learning based data fusion (RLBDF) method is proposed to obtain the fusion results. With the case that the priori knowledge of target is obtained, the fusion accuracy reinforcement is realized by the error between fused value and actual value. Furthermore, the Fisher information is instead used as the reward if the priori knowledge is unable to be obtained. Simulations results verify that the developed method is feasible and effective for the multi-sensors data fusion in air combat.
Mobile Edge Computing Enabled 5G Health Monitoring for Internet of Medical Things: A Decentralized Game Theoretic Approach The prompt evolution of Internet of Medical Things (IoMT) promotes pervasive in-home health monitoring networks. However, excessive requirements of patients result in insufficient spectrum resources and communication overload. Mobile Edge Computing (MEC) enabled 5G health monitoring is conceived as a favorable paradigm to tackle such an obstacle. In this paper, we construct a cost-efficient in-home health monitoring system for IoMT by dividing it into two sub-networks, i.e., intra-Wireless Body Area Networks (WBANs) and beyond-WBANs. Highlighting the characteristics of IoMT, the cost of patients depends on medical criticality, Age of Information (AoI) and energy consumption. For intra-WBANs, a cooperative game is formulated to allocate the wireless channel resources. While for beyond-WBANs, considering the individual rationality and potential selfishness, a decentralized non-cooperative game is proposed to minimize the system-wide cost in IoMT. We prove that the proposed algorithm can reach a Nash equilibrium. In addition, the upper bound of the algorithm time complexity and the number of patients benefiting from MEC is theoretically derived. Performance evaluations demonstrate the effectiveness of our proposed algorithm with respect to the system-wide cost and the number of patients benefiting from MEC.
Artificial fish swarm algorithm: a survey of the state-of-the-art, hybridization, combinatorial and indicative applications FSA (artificial fish-swarm algorithm) is one of the best methods of optimization among the swarm intelligence algorithms. This algorithm is inspired by the collective movement of the fish and their various social behaviors. Based on a series of instinctive behaviors, the fish always try to maintain their colonies and accordingly demonstrate intelligent behaviors. Searching for food, immigration and dealing with dangers all happen in a social form and interactions between all fish in a group will result in an intelligent social behavior.This algorithm has many advantages including high convergence speed, flexibility, fault tolerance and high accuracy. This paper is a review of AFSA algorithm and describes the evolution of this algorithm along with all improvements, its combination with various methods as well as its applications. There are many optimization methods which have a affinity with this method and the result of this combination will improve the performance of this method. Its disadvantages include high time complexity, lack of balance between global and local search, in addition to lack of benefiting from the experiences of group members for the next movements.
Short-Term Traffic Flow Forecasting: An Experimental Comparison of Time-Series Analysis and Supervised Learning The literature on short-term traffic flow forecasting has undergone great development recently. Many works, describing a wide variety of different approaches, which very often share similar features and ideas, have been published. However, publications presenting new prediction algorithms usually employ different settings, data sets, and performance measurements, making it difficult to infer a clear picture of the advantages and limitations of each model. The aim of this paper is twofold. First, we review existing approaches to short-term traffic flow forecasting methods under the common view of probabilistic graphical models, presenting an extensive experimental comparison, which proposes a common baseline for their performance analysis and provides the infrastructure to operate on a publicly available data set. Second, we present two new support vector regression models, which are specifically devised to benefit from typical traffic flow seasonality and are shown to represent an interesting compromise between prediction accuracy and computational efficiency. The SARIMA model coupled with a Kalman filter is the most accurate model; however, the proposed seasonal support vector regressor turns out to be highly competitive when performing forecasts during the most congested periods.
TSCA: A Temporal-Spatial Real-Time Charging Scheduling Algorithm for On-Demand Architecture in Wireless Rechargeable Sensor Networks. The collaborative charging issue in Wireless Rechargeable Sensor Networks (WRSNs) is a popular research problem. With the help of wireless power transfer technology, electrical energy can be transferred from wireless charging vehicles (WCVs) to sensors, providing a new paradigm to prolong network lifetime. Existing techniques on collaborative charging usually take the periodical and deterministic approach, but neglect influences of non-deterministic factors such as topological changes and node failures, making them unsuitable for large-scale WRSNs. In this paper, we develop a temporal-spatial charging scheduling algorithm, namely TSCA, for the on-demand charging architecture. We aim to minimize the number of dead nodes while maximizing energy efficiency to prolong network lifetime. First, after gathering charging requests, a WCV will compute a feasible movement solution. A basic path planning algorithm is then introduced to adjust the charging order for better efficiency. Furthermore, optimizations are made in a global level. Then, a node deletion algorithm is developed to remove low efficient charging nodes. Lastly, a node insertion algorithm is executed to avoid the death of abandoned nodes. Extensive simulations show that, compared with state-of-the-art charging scheduling algorithms, our scheme can achieve promising performance in charging throughput, charging efficiency, and other performance metrics.
A novel adaptive dynamic programming based on tracking error for nonlinear discrete-time systems In this paper, to eliminate the tracking error by using adaptive dynamic programming (ADP) algorithms, a novel formulation of the value function is presented for the optimal tracking problem (TP) of nonlinear discrete-time systems. Unlike existing ADP methods, this formulation introduces the control input into the tracking error, and ignores the quadratic form of the control input directly, which makes the boundedness and convergence of the value function independent of the discount factor. Based on the proposed value function, the optimal control policy can be deduced without considering the reference control input. Value iteration (VI) and policy iteration (PI) methods are applied to prove the optimality of the obtained control policy, and derived the monotonicity property and convergence of the iterative value function. Simulation examples realized with neural networks and the actor–critic structure are provided to verify the effectiveness of the proposed ADP algorithm.
1.2
0.2
0.2
0.2
0.2
0.2
0.2
0.2
0.1
0
0
0
0
0
MOSOA: A new multi-objective seagull optimization algorithm •A novel Multi-objective Seagull Optimization Algorithm is proposed.•The algorithm is tested on 24 real challenging benchmark test function.•The results show the superior convergence behaviour of proposed algorithm.•The results on engineering design problems prove its efficiency and applicability.
M-LionWhale: multi-objective optimisation model for secure routing in mobile ad-hoc network. Mobile ad-hoc network (MANET) is an emerging technology that comes under the category of wireless network. Even though the network assumes that all its mobile nodes are trusted, it is impossible in the real world as few nodes may be malicious. Therefore, it is essential to put forward a mechanism that can provide security by selecting an optimal route for data forwarding. In this study, a goal pro...
Intelligent Ad-Hoc-On Demand Multipath Distance Vector for Wormhole Attack in Clustered WSN In Wireless Senor Networks, security is the most significant issue when sending such an essential message via wireless connection. This helps attackers to access the network and execute several potential attacks to intercept or modify real data/information. Because network sensors do not have routers, the same routing protocol must be split by all the nodes participating in the network to assist each other with packet transmission. In complex topology, its unguided existence often renders it open to all forms of protection attack, presenting a degree of security difficulties. Wormhole is a popular illustration of attacks, due to its difficulties in detecting and stopping, poses the greatest danger. A new routing technique is being presented in this paper which works towards ensuring the secured path for the data transmission. In the research work wormhole type of attack is being considered and the technique works towards the detection and prevention of the defined type of attack. The proposed methodology is validated based on certain performance related parameters for WSN as energy efficiency, delay from end to end, throughput, delivery ratio of packets. The generated outcomes are compared with some recent techniques over the same domain for the efficiency and the presented work has proven to be best among the described techniques for considered parameters. The methodology defined simulated using NS2 for various performance related parameters like energy efficiency, packet loss, throughput, etc.
Proactive fault-tolerant wireless mesh networks for mission-critical control systems Although wireless networks are becoming a fundamental infrastructure for various control applications, they are inherently exposed to network faults such as lossy links and node failures in environments such as mining, outdoor monitoring, and chemical process control. In this paper, we propose a proactive fault-tolerant mechanism to protect the wireless network against temporal faults without any explicit network state information for mission-critical control systems. Specifically, the proposed mechanism optimizes the multiple routing paths, link scheduling, and traffic generation rate such that it meets the control stability demands even if it experiences multiple link faults and node faults. The proactive network relies on a constrained optimization problem, where the objective function is the network robustness, and the main constraints are the set of the traffic demand, link, and routing layer requirements. To analyze the robustness, we propose a novel performance metric called stability margin ratio, based on the network performance and the stability boundary. Our numerical and experimental performance evaluation shows that the traffic generation rate and the delay of wireless networks are found as critical as the network reliability to guarantee the stability of control systems. Furthermore, the proposed proactive network provides more robust performance than practical state-of-the-art solutions while maintaining high energy efficiency.
Energy-efficient and balanced routing in low-power wireless sensor networks for data collection Cost-based routing protocols are the main approach used in practical wireless sensor network (WSN) and Internet of Things (IoT) deployments for data collection applications with energy constraints; however, those routing protocols lead to the concentration of most of the data traffic on some specific nodes which provide the best available routes, thus significantly increasing their energy consumption. Consequently, nodes providing the best routes are potentially the first ones to deplete their batteries and stop working. In this paper, we introduce a novel routing strategy for energy efficient and balanced data collection in WSNs/IoT, which can be applied to any cost-based routing solution to exploit suboptimal network routing alternatives based on the parent set concept. While still taking advantage of the stable routing topologies built in cost-based routing protocols, our approach adds a random component into the process of packet forwarding to achieve a better network lifetime in WSNs. We evaluate the implementation of our approach against other state-of-the-art WSN routing protocols through thorough real-world testbed experiments and simulations, and demonstrate that our approach achieves a significant reduction in the energy consumption of the routing layer in the busiest nodes ranging from 11% to 59%, while maintaining over 99% reliability. Furthermore, we conduct the field deployment of our approach in a heterogeneous WSN for environmental monitoring in a forest area, report the experimental results and illustrate the effectiveness of our approach in detail. Our EER based routing protocol CTP+EER is made available as open source to the community for evaluation and adoption.
Computational thinking Summary form only given. My vision for the 21st century, Computational Thinking, will be a fundamental skill used by everyone in the world. To reading, writing, and arithmetic, we should add computational thinking to every child's analytical ability. Computational thinking involves solving problems, designing systems, and understanding human behavior by drawing on the concepts fundamental to computer science. Thinking like a computer scientist means more than being able to program a computer. It requires the ability to abstract and thus to think at multiple levels of abstraction. In this talk I will give many examples of computational thinking, argue that it has already influenced other disciplines, and promote the idea that teaching computational thinking can not only inspire future generations to enter the field of computer science but benefit people in all fields.
JPEG Error Analysis and Its Applications to Digital Image Forensics JPEG is one of the most extensively used image formats. Understanding the inherent characteristics of JPEG may play a useful role in digital image forensics. In this paper, we introduce JPEG error analysis to the study of image forensics. The main errors of JPEG include quantization, rounding, and truncation errors. Through theoretically analyzing the effects of these errors on single and double JPEG compression, we have developed three novel schemes for image forensics including identifying whether a bitmap image has previously been JPEG compressed, estimating the quantization steps of a JPEG image, and detecting the quantization table of a JPEG image. Extensive experimental results show that our new methods significantly outperform existing techniques especially for the images of small sizes. We also show that the new method can reliably detect JPEG image blocks which are as small as 8 × 8 pixels and compressed with quality factors as high as 98. This performance is important for analyzing and locating small tampered regions within a composite image.
Highly dynamic Destination-Sequenced Distance-Vector routing (DSDV) for mobile computers An ad-hoc network is the cooperative engagement of a collection of Mobile Hosts without the required intervention of any centralized Access Point. In this paper we present an innovative design for the operation of such ad-hoc networks. The basic idea of the design is to operate each Mobile Host as a specialized router, which periodically advertises its view of the interconnection topology with other Mobile Hosts within the network. This amounts to a new sort of routing protocol. We have investigated modifications to the basic Bellman-Ford routing mechanisms, as specified by RIP [5], to make it suitable for a dynamic and self-starting network mechanism as is required by users wishing to utilize ad hoc networks. Our modifications address some of the previous objections to the use of Bellman-Ford, related to the poor looping properties of such algorithms in the face of broken links and the resulting time dependent nature of the interconnection topology describing the links between the Mobile Hosts. Finally, we describe the ways in which the basic network-layer routing can be modified to provide MAC-layer support for ad-hoc networks.
The FERET Evaluation Methodology for Face-Recognition Algorithms Two of the most critical requirements in support of producing reliable face-recognition systems are a large database of facial images and a testing procedure to evaluate systems. The Face Recognition Technology (FERET) program has addressed both issues through the FERET database of facial images and the establishment of the FERET tests. To date, 14,126 images from 1,199 individuals are included in the FERET database, which is divided into development and sequestered portions of the database. In September 1996, the FERET program administered the third in a series of FERET face-recognition tests. The primary objectives of the third test were to 1) assess the state of the art, 2) identify future areas of research, and 3) measure algorithm performance.
Neural fitted q iteration – first experiences with a data efficient neural reinforcement learning method This paper introduces NFQ, an algorithm for efficient and effective training of a Q-value function represented by a multi-layer perceptron. Based on the principle of storing and reusing transition experiences, a model-free, neural network based Reinforcement Learning algorithm is proposed. The method is evaluated on three benchmark problems. It is shown empirically, that reasonably few interactions with the plant are needed to generate control policies of high quality.
Labels and event processes in the Asbestos operating system Asbestos, a new operating system, provides novel labeling and isolation mechanisms that help contain the effects of exploitable software flaws. Applications can express a wide range of policies with Asbestos's kernel-enforced labels, including controls on interprocess communication and system-wide information flow. A new event process abstraction defines lightweight, isolated contexts within a single process, allowing one process to act on behalf of multiple users while preventing it from leaking any single user's data to others. A Web server demonstration application uses these primitives to isolate private user data. Since the untrusted workers that respond to client requests are constrained by labels, exploited workers cannot directly expose user data except as allowed by application policy. The server application requires 1.4 memory pages per user for up to 145,000 users and achieves connection rates similar to Apache, demonstrating that additional security can come at an acceptable cost.
Switching Stabilization for a Class of Slowly Switched Systems In this technical note, the problem of switching stabilization for slowly switched linear systems is investigated. In particular, the considered systems can be composed of all unstable subsystems. Based on the invariant subspace theory, the switching signal with mode-dependent average dwell time (MDADT) property is designed to exponentially stabilize the underlying system. Furthermore, sufficient condition of stabilization for switched systems with all stable subsystems under MDADT switching is also given. The correctness and effectiveness of the proposed approaches are illustrated by a numerical example.
Automated Detection of Obstructive Sleep Apnea Events from a Single-Lead Electrocardiogram Using a Convolutional Neural Network. In this study, we propose a method for the automated detection of obstructive sleep apnea (OSA) from a single-lead electrocardiogram (ECG) using a convolutional neural network (CNN). A CNN model was designed with six optimized convolution layers including activation, pooling, and dropout layers. One-dimensional (1D) convolution, rectified linear units (ReLU), and max pooling were applied to the convolution, activation, and pooling layers, respectively. For training and evaluation of the CNN model, a single-lead ECG dataset was collected from 82 subjects with OSA and was divided into training (including data from 63 patients with 34,281 events) and testing (including data from 19 patients with 8571 events) datasets. Using this CNN model, a precision of 0.99%, a recall of 0.99%, and an F-score of 0.99% were attained with the training dataset; these values were all 0.96% when the CNN was applied to the testing dataset. These results show that the proposed CNN model can be used to detect OSA accurately on the basis of a single-lead ECG. Ultimately, this CNN model may be used as a screening tool for those suspected to suffer from OSA.
Convert Harm Into Benefit: A Coordination-Learning Based Dynamic Spectrum Anti-Jamming Approach This paper mainly investigates the multi-user anti-jamming spectrum access problem. Using the idea of “converting harm into benefit,” the malicious jamming signals projected by the enemy are utilized by the users as the coordination signals to guide spectrum coordination. An “internal coordination-external confrontation” multi-user anti-jamming access game model is constructed, and the existence of Nash equilibrium (NE) as well as correlated equilibrium (CE) is demonstrated. A coordination-learning based anti-jamming spectrum access algorithm (CLASA) is designed to achieve the CE of the game. Simulation results show the convergence, and effectiveness of the proposed CLASA algorithm, and indicate that our approach can help users confront the malicious jammer, and coordinate internal spectrum access simultaneously without information exchange. Last but not least, the fairness of the proposed approach under different jamming attack patterns is analyzed, which illustrates that this approach provides fair anti-jamming spectrum access opportunities under complicated jamming pattern.
1.2
0.2
0.2
0.2
0.2
0
0
0
0
0
0
0
0
0
Reducing delay and prolonging the lifetime of wireless sensor network using efficient routing protocol based on mobile sink and virtual infrastructure Mobile sink has been used by many routing algorithms to improve the network lifetime in wireless sensor networks. Sink mobility brings new challenges to the sensor network. It suggests that sink position is continuously propagated through the network area to keep all sensor nodes updated with the direction of forwarding data to the sink. Unfortunately, frequent sink position updates can lead to both high-energy consumption and increased collisions in the network. In this paper, a new routing algorithm for wireless sensor networks is proposed, which uses a virtual multi-ring shaped infrastructure to advertise the mobile sink position to the network. In addition, the proposed algorithm supports several mobile sinks in the network. Simulation results indicate that the proposed algorithm reduces energy consumption and delay, and prolongs the sensor network lifetime when compared with the other similar algorithms.
Reliable and Fast Hand-Offs in Low-Power Wireless Networks Hand-off (or hand-over), the process where mobile nodes select the best access point available to transfer data, has been well studied in wireless networks. The performance of a hand-off process depends on the specific characteristics of the wireless links. In the case of low-power wireless networks, hand-off decisions must be carefully taken by considering the unique properties of inexpensive low-power radios. This paper addresses the design, implementation and evaluation of smart-HOP, a hand-off mechanism tailored for low-power wireless networks. This work has three main contributions. First, it formulates the hard hand-off process for low-power networks (such as typical wireless sensor networks - WSNs) with a probabilistic model, to investigate the impact of the most relevant channel parameters through an analytical approach. Second, it confirms the probabilistic model through simulation and further elaborates on the impact of several hand-off parameters. Third, it fine-tunes the most relevant hand-off parameters via an extended set of experiments, in a realistic experimental scenario. The evaluation shows that smart-HOP performs well in the transitional region while achieving more than 98 percent relative delivery ratio and hand-off delays in the order of a few tens of a milliseconds.
An efficient medium access control protocol for WSN-UAV. Recent advances in Unmanned Aerial Vehicle (UAV) technologies have enhanced Wireless Sensor Networks (WSNs) by offering a UAV as a mobile data gathering node. These systems are called WSN-UAV that are well-suited for remote monitoring and emergency applications. Since previous Medium Access Control (MAC) protocols proposed in WSNs are not appropriate in the presence of a UAV, few researches have proposed new MAC protocols to meet WSN-UAV requirements. MAC protocols of WSN-UAV should be extremely efficient and fair due to the time-limited presence of the UAV in the neighborhood of each sensor. However, issues such as high throughput in dense networks, fairness among sensors, and efficiency have not been resolved yet in a satisfactory manner. Moreover, previous works lack analytical evaluation of their protocols. In this paper, we present a novel MAC protocol in WSN-UAV, called Advanced Prioritized MAC (AP-MAC), that can provide high throughput, fairness, and efficiency, especially in dense networks. We also analytically evaluate AP-MAC using a 3-dimensional Markov chain and validate its correctness using simulation. Simulation results under various scenarios confirm that AP-MAC can approximately improve throughput and fairness up to 20% and 25%, respectively, leading to higher efficiency compared with previous work in WSN-UAV systems such as Prioritized Frame Selection (PFS).
Wireless Powered Sensor Networks for Internet of Things: Maximum Throughput and Optimal Power Allocation. This paper investigates a wireless powered sensor network, where multiple sensor nodes are deployed to monitor a certain external environment. A multiantenna power station (PS) provides the power to these sensor nodes during wireless energy transfer phase, and consequently the sensor nodes employ the harvested energy to transmit their own monitoring information to a fusion center during wireless i...
A Novel Cluster Head Selection Technique for Edge-Computing based IoMT Systems •An innovative cluster head selection technique is developed for medical systems.•An energy efficient communication protocol for IoMT applications.•Development of clustering model for remote healthcare domains.
An Energy-Efficient Evolutionary Clustering Technique for Disaster Management in IoT Networks. Wireless Sensor Networks (WSNs) are key elements of Internet of Things (IoT) networks which provide sensing and wireless connectivity. Disaster management in smart cities is classified as a safety-critical application. Thus, it is important to ensure system availability by increasing the lifetime of WSNs. Clustering is one of the routing techniques that benefits energy efficiency in WSNs. This paper provides an evolutionary clustering and routing method which is capable of managing the energy consumption of nodes while considering the characteristics of a disaster area. The proposed method consists of two phases. First, we present a model with improved hybrid Particle Swarm Optimization (PSO) and Harmony Search Algorithm (HSA) for cluster head (CH) selection. Second, we design a PSO-based multi-hop routing system with enhanced tree encoding and a modified data packet format. The simulation results for disaster scenarios prove the efficiency of the proposed method in comparison with the state-of-the-art approaches in terms of the overall residual energy, number of live nodes, network coverage, and the packet delivery ratio.
Wireless sensor network survey A wireless sensor network (WSN) has important applications such as remote environmental monitoring and target tracking. This has been enabled by the availability, particularly in recent years, of sensors that are smaller, cheaper, and intelligent. These sensors are equipped with wireless interfaces with which they can communicate with one another to form a network. The design of a WSN depends significantly on the application, and it must consider factors such as the environment, the application's design objectives, cost, hardware, and system constraints. The goal of our survey is to present a comprehensive review of the recent literature since the publication of [I.F. Akyildiz, W. Su, Y. Sankarasubramaniam, E. Cayirci, A survey on sensor networks, IEEE Communications Magazine, 2002]. Following a top-down approach, we give an overview of several new applications and then review the literature on various aspects of WSNs. We classify the problems into three different categories: (1) internal platform and underlying operating system, (2) communication protocol stack, and (3) network services, provisioning, and deployment. We review the major development in these three categories and outline new challenges.
Long short-term memory. Learning to store information over extended time intervals by recurrent backpropagation takes a very long time, mostly because of insufficient, decaying error backflow. We briefly review Hochreiter's (1991) analysis of this problem, then address it by introducing a novel, efficient, gradient-based method called long short-term memory (LSTM). Truncating the gradient where this does not do harm, LSTM can learn to bridge minimal time lags in excess of 1000 discrete-time steps by enforcing constant error flow through constant error carousels within special units. Multiplicative gate units learn to open and close access to the constant error flow. LSTM is local in space and time; its computational complexity per time step and weight is O(1). Our experiments with artificial data involve local, distributed, real-valued, and noisy pattern representations. In comparisons with real-time recurrent learning, back propagation through time, recurrent cascade correlation, Elman nets, and neural sequence chunking, LSTM leads to many more successful runs, and learns much faster. LSTM also solves complex, artificial long-time-lag tasks that have never been solved by previous recurrent network algorithms.
Joint Task Offloading and Resource Allocation for Multi-Server Mobile-Edge Computing Networks Mobile-edge computing (MEC) is an emerging paradigm that provides a capillary distribution of cloud computing capabilities to the edge of the wireless access network, enabling rich services and applications in close proximity to the end users. In this paper, an MEC enabled multi-cell wireless network is considered where each base station (BS) is equipped with a MEC server that assists mobile users...
Space-time modeling of traffic flow. This paper discusses the application of space-time autoregressive integrated moving average (STARIMA) methodology for representing traffic flow patterns. Traffic flow data are in the form of spatial time series and are collected at specific locations at constant intervals of time. Important spatial characteristics of the space-time process are incorporated in the STARIMA model through the use of weighting matrices estimated on the basis of the distances among the various locations where data are collected. These matrices distinguish the space-time approach from the vector autoregressive moving average (VARMA) methodology and enable the model builders to control the number of the parameters that have to be estimated. The proposed models can be used for short-term forecasting of space-time stationary traffic-flow processes and for assessing the impact of traffic-flow changes on other parts of the network. The three-stage iterative space-time model building procedure is illustrated using 7.5min average traffic flow data for a set of 25 loop-detectors located at roads that direct to the centre of the city of Athens, Greece. Data for two months with different traffic-flow characteristics are modelled in order to determine the stability of the parameter estimation.
Dynamic transfer among alternative controllers and its relation to antiwindup controller design Advanced control strategies and modern consulting provide new challenges for the classical problem of bumpless transfer. It can, for example, be necessary to transfer between an only approximately known existing analog controller and a new digital or adaptive controller without accessing any states. Transfer ought to be bidirectional and not presuppose steady state, so that an immediate back-transfer is possible if the new controller should drive the plant unstable. We present a scheme that meets these requirements. By casting the problem of bidirectional transfer into an associated tracking control problem, systematic analysis and design procedures from control theory can be applied. The associated control problem also has a correspondence to the design of antiwindup controllers. The paper includes laboratory and industrial applications.
NETWRAP: An NDN Based Real-TimeWireless Recharging Framework for Wireless Sensor Networks Using vehicles equipped with wireless energy transmission technology to recharge sensor nodes over the air is a game-changer for traditional wireless sensor networks. The recharging policy regarding when to recharge which sensor nodes critically impacts the network performance. So far only a few works have studied such recharging policy for the case of using a single vehicle. In this paper, we propose NETWRAP, an N DN based Real Time Wireless Rech arging Protocol for dynamic wireless recharging in sensor networks. The real-time recharging framework supports single or multiple mobile vehicles. Employing multiple mobile vehicles provides more scalability and robustness. To efficiently deliver sensor energy status information to vehicles in real-time, we leverage concepts and mechanisms from named data networking (NDN) and design energy monitoring and reporting protocols. We derive theoretical results on the energy neutral condition and the minimum number of mobile vehicles required for perpetual network operations. Then we study how to minimize the total traveling cost of vehicles while guaranteeing all the sensor nodes can be recharged before their batteries deplete. We formulate the recharge optimization problem into a Multiple Traveling Salesman Problem with Deadlines (m-TSP with Deadlines), which is NP-hard. To accommodate the dynamic nature of node energy conditions with low overhead, we present an algorithm that selects the node with the minimum weighted sum of traveling time and residual lifetime. Our scheme not only improves network scalability but also ensures the perpetual operation of networks. Extensive simulation results demonstrate the effectiveness and efficiency of the proposed design. The results also validate the correctness of the theoretical analysis and show significant improvements that cut the number of nonfunctional nodes by half compared to the static scheme while maintaining the network overhead at the same level.
Adaptive Fuzzy Control With Prescribed Performance for Block-Triangular-Structured Nonlinear Systems. In this paper, an adaptive fuzzy control method with prescribed performance is proposed for multi-input and multioutput block-triangular-structured nonlinear systems with immeasurable states. Fuzzy logic systems are adopted to identify the unknown nonlinear system functions. Adaptive fuzzy state observers are designed to solve the problem of unmeasured states, and a new observer-based output-feedb...
Energy harvesting algorithm considering max flow problem in wireless sensor networks. In Wireless Sensor Networks (WSNs), sensor nodes with poor energy always have bad effect on the data rate or max flow. These nodes are called bottleneck nodes. In this paper, in order to increase the max flow, we assume an energy harvesting WSNs environment to investigate the cooperation of multiple Mobile Chargers (MCs). MCs are mobile robots that use wireless charging technology to charge sensor nodes in WSNs. This means that in energy harvesting WSNs environments, sensor nodes can obtain energy replenishment by using MCs or collecting energy from nature by themselves. In our research, we use MCs to improve the energy of the sensor nodes by performing multiple rounds of unified scheduling, and finally achieve the purpose of increasing the max flow at sinks. Firstly, we model this problem as a Linear Programming (LP) to search the max flow in a round of charging scheduling and prove that the problem is NP-hard. In order to solve the problem, we propose a heuristic approach: deploying MCs in units of paths with the lowest energy node priority. To reduce the energy consumption of MCs and increase the charging efficiency, we also take the optimization of MCs’ moving distance into our consideration. Finally, we extend the method to multiple rounds of scheduling called BottleNeck. Simulation results show that Bottleneck performs well at increasing max flow.
1.2
0.2
0.2
0.2
0.2
0.2
0.011111
0
0
0
0
0
0
0
Upper limb exosuit cable routing optimization Exosuits are emerging as promising in assisting with activities of daily living. In the design phase of an exosuit, it is fundamental to maximize its portability. The goal of this work was to identify the best cable routing configuration for an upper limb cable-driven exosuit to assist elbow flexion. Simulations were run in OpenSim. Different cable configurations were evaluated. The goal was to minimize the overall tension of the cables to reduce the device’s power consumption and torque requirements. The optimal configuration was evaluated in simulation for different percentages of assistance to study its effects in terms of muscle activation and joint reaction forces. We then tested three different configurations on a test bench to both evaluate the motor current and their effect on the pronation/supination of the elbow. Simulation results suggested that a double cable configuration might help to lower the motor torque and power consumption. This conclusion was supported by the experimental results, in which the motor current was reduced by 12.5% with respect to the single cable configuration. Simulation results also showed that the optimal configuration lowered muscle activation without greatly affecting joint reactions at the elbow, even though it might cause unwanted pronation/supination, as experimental results confirmed. However, since a double configuration results in greater complexity and reduced efficiency, single-cable solutions still represent a good option.
The Sybil Attack Large-scale peer-to-peer systems facesecurity threats from faulty or hostile remotecomputing elements. To resist these threats, manysuch systems employ redundancy. However, if asingle faulty entity can present multiple identities,it can control a substantial fraction of the system,thereby undermining this redundancy. Oneapproach to preventing these &quot;Sybil attacks&quot; is tohave a trusted agency certify identities. Thispaper shows that, without a logically centralizedauthority, Sybil...
BLEU: a method for automatic evaluation of machine translation Human evaluations of machine translation are extensive but expensive. Human evaluations can take months to finish and involve human labor that can not be reused. We propose a method of automatic machine translation evaluation that is quick, inexpensive, and language-independent, that correlates highly with human evaluation, and that has little marginal cost per run. We present this method as an automated understudy to skilled human judges which substitutes for them when there is need for quick or frequent evaluations.
Computational thinking Summary form only given. My vision for the 21st century, Computational Thinking, will be a fundamental skill used by everyone in the world. To reading, writing, and arithmetic, we should add computational thinking to every child's analytical ability. Computational thinking involves solving problems, designing systems, and understanding human behavior by drawing on the concepts fundamental to computer science. Thinking like a computer scientist means more than being able to program a computer. It requires the ability to abstract and thus to think at multiple levels of abstraction. In this talk I will give many examples of computational thinking, argue that it has already influenced other disciplines, and promote the idea that teaching computational thinking can not only inspire future generations to enter the field of computer science but benefit people in all fields.
Fuzzy logic in control systems: fuzzy logic controller. I.
Switching between stabilizing controllers This paper deals with the problem of switching between several linear time-invariant (LTI) controllers—all of them capable of stabilizing a speci4c LTI process—in such a way that the stability of the closed-loop system is guaranteed for any switching sequence. We show that it is possible to 4nd realizations for any given family of controller transfer matrices so that the closed-loop system remains stable, no matter how we switch among the controller. The motivation for this problem is the control of complex systems where con8icting requirements make a single LTI controller unsuitable. ? 2002 Published by Elsevier Science Ltd.
Tabu Search - Part I
Bidirectional recurrent neural networks In the first part of this paper, a regular recurrent neural network (RNN) is extended to a bidirectional recurrent neural network (BRNN). The BRNN can be trained without the limitation of using input information just up to a preset future frame. This is accomplished by training it simultaneously in positive and negative time direction. Structure and training procedure of the proposed network are explained. In regression and classification experiments on artificial data, the proposed structure gives better results than other approaches. For real data, classification experiments for phonemes from the TIMIT database show the same tendency. In the second part of this paper, it is shown how the proposed bidirectional structure can be easily modified to allow efficient estimation of the conditional posterior probability of complete symbol sequences without making any explicit assumption about the shape of the distribution. For this part, experiments on real data are reported
An intensive survey of fair non-repudiation protocols With the phenomenal growth of the Internet and open networks in general, security services, such as non-repudiation, become crucial to many applications. Non-repudiation services must ensure that when Alice sends some information to Bob over a network, neither Alice nor Bob can deny having participated in a part or the whole of this communication. Therefore a fair non-repudiation protocol has to generate non-repudiation of origin evidences intended to Bob, and non-repudiation of receipt evidences destined to Alice. In this paper, we clearly define the properties a fair non-repudiation protocol must respect, and give a survey of the most important non-repudiation protocols without and with trusted third party (TTP). For the later ones we discuss the evolution of the TTP's involvement and, between others, describe the most recent protocol using a transparent TTP. We also discuss some ad-hoc problems related to the management of non-repudiation evidences.
Dynamic movement and positioning of embodied agents in multiparty conversations For embodied agents to engage in realistic multiparty conversation, they must stand in appropriate places with respect to other agents and the environment. When these factors change, such as an agent joining the conversation, the agents must dynamically move to a new location and/or orientation to accommodate. This paper presents an algorithm for simulating movement of agents based on observed human behavior using techniques developed for pedestrian movement in crowd simulations. We extend a previous group conversation simulation to include an agent motion algorithm. We examine several test cases and show how the simulation generates results that mirror real-life conversation settings.
An improved genetic algorithm with conditional genetic operators and its application to set-covering problem The genetic algorithm (GA) is a popular, biologically inspired optimization method. However, in the GA there is no rule of thumb to design the GA operators and select GA parameters. Instead, trial-and-error has to be applied. In this paper we present an improved genetic algorithm in which crossover and mutation are performed conditionally instead of probability. Because there are no crossover rate and mutation rate to be selected, the proposed improved GA can be more easily applied to a problem than the conventional genetic algorithms. The proposed improved genetic algorithm is applied to solve the set-covering problem. Experimental studies show that the improved GA produces better results over the conventional one and other methods.
Lane-level traffic estimations using microscopic traffic variables This paper proposes a novel inference method to estimate lane-level traffic flow, time occupancy and vehicle inter-arrival time on road segments where local information could not be measured and assessed directly. The main contributions of the proposed method are 1) the ability to perform lane-level estimations of traffic flow, time occupancy and vehicle inter-arrival time and 2) the ability to adapt to different traffic regimes by assessing only microscopic traffic variables. We propose a modified Kriging estimation model which explicitly takes into account both spatial and temporal variability. Performance evaluations are conducted using real-world data under different traffic regimes and it is shown that the proposed method outperforms a Kalman filter-based approach.
Convolutional Neural Network-Based Classification of Driver's Emotion during Aggressive and Smooth Driving Using Multi-Modal Camera Sensors. Because aggressive driving often causes large-scale loss of life and property, techniques for advance detection of adverse driver emotional states have become important for the prevention of aggressive driving behaviors. Previous studies have primarily focused on systems for detecting aggressive driver emotion via smart-phone accelerometers and gyro-sensors, or they focused on methods of detecting physiological signals using electroencephalography (EEG) or electrocardiogram (ECG) sensors. Because EEG and ECG sensors cause discomfort to drivers and can be detached from the driver's body, it becomes difficult to focus on bio-signals to determine their emotional state. Gyro-sensors and accelerometers depend on the performance of GPS receivers and cannot be used in areas where GPS signals are blocked. Moreover, if driving on a mountain road with many quick turns, a driver's emotional state can easily be misrecognized as that of an aggressive driver. To resolve these problems, we propose a convolutional neural network (CNN)-based method of detecting emotion to identify aggressive driving using input images of the driver's face, obtained using near-infrared (NIR) light and thermal camera sensors. In this research, we conducted an experiment using our own database, which provides a high classification accuracy for detecting driver emotion leading to either aggressive or smooth (i.e., relaxed) driving. Our proposed method demonstrates better performance than existing methods.
Ethical Considerations Of Applying Robots In Kindergarten Settings: Towards An Approach From A Macroperspective In child-robot interaction (cHRI) research, many studies pursue the goal to develop interactive systems that can be applied in everyday settings. For early education, increasingly, the setting of a kindergarten is targeted. However, when cHRI and research are brought into a kindergarten, a range of ethical and related procedural aspects have to be considered and dealt with. While ethical models elaborated within other human-robot interaction settings, e.g., assisted living contexts, can provide some important indicators for relevant issues, we argue that it is important to start developing a systematic approach to identify and tackle those ethical issues which rise with cHRI in kindergarten settings on a more global level and address the impact of the technology from a macroperspective beyond the effects on the individual. Based on our experience in conducting studies with children in general and pedagogical considerations on the role of the institution of kindergarten in specific, in this paper, we enfold some relevant aspects that have barely been addressed in an explicit way in current cHRI research. Four areas are analyzed and key ethical issues are identified in each area: (1) the institutional setting of a kindergarten, (2) children as a vulnerable group, (3) the caregivers' role, and (4) pedagogical concepts. With our considerations, we aim at (i) broadening the methodology of the current studies within the area of cHRI, (ii) revalidate it based on our comprehensive empirical experience with research in kindergarten settings, both laboratory and real-world contexts, and (iii) provide a framework for the development of a more systematic approach to address the ethical issues in cHRI research within kindergarten settings.
1
0
0
0
0
0
0
0
0
0
0
0
0
0
Optimal tracking control of nonlinear partially-unknown constrained-input systems using integral reinforcement learning In this paper, a new formulation for the optimal tracking control problem (OTCP) of continuous-time nonlinear systems is presented. This formulation extends the integral reinforcement learning (IRL) technique, a method for solving optimal regulation problems, to learn the solution to the OTCP. Unlike existing solutions to the OTCP, the proposed method does not need to have or to identify knowledge of the system drift dynamics, and it also takes into account the input constraints a priori. An augmented system composed of the error system dynamics and the command generator dynamics is used to introduce a new nonquadratic discounted performance function for the OTCP. This encodes the input constrains into the optimization problem. A tracking Hamilton-Jacobi-Bellman (HJB) equation associated with this nonquadratic performance function is derived which gives the optimal control solution. An online IRL algorithm is presented to learn the solution to the tracking HJB equation without knowing the system drift dynamics. Convergence to a near-optimal control solution and stability of the whole system are shown under a persistence of excitation condition. Simulation examples are provided to show the effectiveness of the proposed method.
Self-Learning Optimal Regulation for Discrete-Time Nonlinear Systems Under Event-Driven Formulation The self-learning optimal regulation for discrete-time nonlinear systems under event-driven formulation is investigated. An event-based adaptive critic algorithm is developed with convergence discussion of the iterative process. The input-to-state stability (ISS) analysis for the present nonlinear plant is established. Then, a suitable triggering condition is proved to ensure the ISS of the controlled system. An iterative dual heuristic dynamic programming (DHP) strategy is adopted to implement the event-driven framework. Simulation examples are carried out to demonstrate the applicability of the constructed method. Compared with the traditional DHP algorithm, the even-based algorithm is able to substantially reduce the updating times of the control input, while still maintaining an impressive performance.
Asymptotically Stable Adaptive-Optimal Control Algorithm With Saturating Actuators and Relaxed Persistence of Excitation. This paper proposes a control algorithm based on adaptive dynamic programming to solve the infinite-horizon optimal control problem for known deterministic nonlinear systems with saturating actuators and nonquadratic cost functionals. The algorithm is based on an actor/critic framework, where a critic neural network (NN) is used to learn the optimal cost, and an actor NN is used to learn the optim...
Optimal control using adaptive resonance theory and Q-learning. Motivated by recent advancement in neurocognitive in brain modeling research, a multiple model-based Q-learning structure is proposed for optimal tracking control problem of time-varying discrete-time systems. This is achieved by utilizing a multiple-model scheme combined with adaptive resonance theory (ART). The ART algorithm generates sub-models based on the match-based clustering method. A responsibility signal governs the likelihood of contribution of each sub-model to the Q-function. The Q-function is learned using the batch least-square algorithm. Simulation results are added to show the performance and the effectiveness of the overall proposed control method.
Linear Quadratic Tracking Control of Partially-Unknown Continuous-Time Systems Using Reinforcement Learning. In this technical note, an online learning algorithm is developed to solve the linear quadratic tracking (LQT) problem for partially-unknown continuous-time systems. It is shown that the value function is quadratic in terms of the state of the system and the command generator. Based on this quadratic form, an LQT Bellman equation and an LQT algebraic Riccati equation (ARE) are derived to solve the LQT problem. The integral reinforcement learning technique is used to find the solution to the LQT ARE online and without requiring the knowledge of the system drift dynamics or the command generator dynamics. The convergence of the proposed online algorithm to the optimal control solution is verified. To show the efficiency of the proposed approach, a simulation example is provided.
Integral Barrier Lyapunov function-based adaptive control for switched nonlinear systems This paper presents an adaptive control method for a class of uncertain strict-feedback switched nonlinear systems. First, we consider the constraint characteristics in the switched nonlinear systems to ensure that all states in switched systems do not violate the constraint ranges. Second, we design the controller based on the backstepping technique, while integral Barrier Lyapunov functions (iBLFs) are adopted to solve the full state constraint problems in each step in order to realize the direct constraints on state variables. Furthermore, we introduce the Lyapunov stability theory to demonstrate that the adaptive controller achieves the desired control goals. Finally, we perform a numerical simulation, which further verifies the significance and feasibility of the presented control scheme.
Adaptive dynamic programming for finite-horizon optimal control of discrete-time nonlinear systems with ε-error bound. In this paper, we study the finite-horizon optimal control problem for discrete-time nonlinear systems using the adaptive dynamic programming (ADP) approach. The idea is to use an iterative ADP algorithm to obtain the optimal control law which makes the performance index function close to the greatest lower bound of all performance indices within an ε-error bound. The optimal number of control steps can also be obtained by the proposed ADP algorithms. A convergence analysis of the proposed ADP algorithms in terms of performance index function and control policy is made. In order to facilitate the implementation of the iterative ADP algorithms, neural networks are used for approximating the performance index function, computing the optimal control policy, and modeling the nonlinear system. Finally, two simulation examples are employed to illustrate the applicability of the proposed method.
Adaptive Fuzzy Decentralized Control for a Class of Strong Interconnected Nonlinear Systems With Unmodeled Dynamics. The state-feedback decentralized stabilization problem is considered for interconnected nonlinear systems in the presence of unmodeled dynamics. The functional relationship in affine form between the strong interconnected functions and error signals is established, which makes backstepping-based fuzzy control successfully generalized to strong interconnected nonlinear systems. By combining adaptiv...
A robust adaptive nonlinear control design An adaptive control design procedure for a class of nonlinear systems with both parametric uncertainty and unknown nonlinearities is presented. The unknown nonlinearities lie within some 'bounding functions', which are assumed to be partially known. The key assumption is that the uncertain terms satisfy a 'triangularity condition'. As illustrated by examples, the proposed design procedure expands the class of nonlinear systems for which global adaptive stabilization methods can be applied. The overall adaptive scheme is shown to guarantee global uniform ultimate boundedness.
Untangling Blockchain: A Data Processing View of Blockchain Systems. Blockchain technologies are gaining massive momentum in the last few years. Blockchains are distributed ledgers that enable parties who do not fully trust each other to maintain a set of global states. The parties agree on the existence, values, and histories of the states. As the technology landscape is expanding rapidly, it is both important and challenging to have a firm grasp of what the core ...
Attacks on biometric systems: a case study in fingerprints In spite of numerous advantages of biometrics-based personal authentication systems over traditional security systems based on token or knowledge, they are vulnerable to attacks that can decrease their security considerably. In this paper, we analyze these attacks in the realm of a fingerprint biometric system. We propose an attack system that uses a hill climbing procedure to synthesize the target minutia templates and evaluate its feasibility with extensive experimental results conducted on a large fingerprint database. Several measures that can be utilized to decrease the probability of such attacks and their ramifications are also presented.
Addictive links: engaging students through adaptive navigation support and open social student modeling Empirical studies of adaptive annotation in the educational context have demonstrated that it can help students to acquire knowledge faster, improve learning outcomes, reduce navigational overhead, and encourage non-sequential navigation. Over the last 8 years we have explored a lesser known effect of adaptive annotation -- its ability to significantly increase student engagement in working with non-mandatory educational content. In the presence of adaptive link annotation, students tend to access significantly more learning content; they stay with it longer, return to it more often and explore a wider variety of learning resources. This talk will present an overview of our exploration of the addictive links effect in many course-long studies, which we ran in several domains (C, SQL and Java programming), for several types of learning content (quizzes, problems, interactive examples). The first part of the talk will review our exploration of a more traditional knowledge-based personalization approach and the second part will focus on more recent studies of social navigation and open social student modeling.
Gender Bias in Coreference Resolution: Evaluation and Debiasing Methods. We introduce a new benchmark, WinoBias, for coreference resolution focused on gender bias. Our corpus contains Winograd-schema style sentences with entities corresponding to people referred by their occupation (e.g. the nurse, the doctor, the carpenter). We demonstrate that a rule-based, a feature-rich, and a neural coreference system all link gendered pronouns to pro-stereotypical entities with higher accuracy than anti-stereotypical entities, by an average difference of 21.1 in F1 score. Finally, we demonstrate a data-augmentation approach that, in combination with existing word-embedding debiasing techniques, removes the bias demonstrated by these systems in WinoBias without significantly affecting their performance on existing coreference benchmark datasets. Our dataset and code are available at this http URL
Social Robots for (Second) Language Learning in (Migrant) Primary School Children Especially these days, innovation and support from technology to relieve pressure in education is highly urgent. This study tested the potential advantage of a social robot over a tablet in (second) language learning on performance, engagement, and enjoyment. Shortages in primary education call for new technology solutions. Previous studies combined robots with tablets, to compensate for robot’s limitations, however, this study applied direct human–robot interaction. Primary school children (N = 63, aged 4–6) participated in a 3-wave field experiment with story-telling exercises, either with a semi-autonomous robot (without tablet, using WOz) or a tablet. Results showed increased learning gains over time when training with a social robot, compared to the tablet. Children who trained with a robot were more engaged in the story-telling task and enjoyed it more. Robot’s behavioral style (social or neutral) hardly differed overall, however, seems to vary for high versus low educational abilities. While social robots need sophistication before being implemented in schools, our study shows the potential of social robots as tutors in (second) language learning.
1.043253
0.042722
0.036306
0.033333
0.028278
0.016746
0.009966
0.002347
0.000013
0
0
0
0
0
Efficient Sampling-Based Maximum Entropy Inverse Reinforcement Learning with Application to Autonomous Driving In the past decades, we have witnessed significant progress in the domain of autonomous driving. Advanced techniques based on optimization and reinforcement learning become increasingly powerful when solving the forward problem: given designed reward/cost functions, how we should optimize them and obtain driving policies that interact with the environment safely and efficiently. Such progress has ...
Attention Based Vehicle Trajectory Prediction Self-driving vehicles need to continuously analyse the driving scene, understand the behavior of other road users and predict their future trajectories in order to plan a safe motion and reduce their reaction time. Motivated by this idea, this paper addresses the problem of vehicle trajectory prediction over an extended horizon. On highways, human drivers continuously adapt their speed and paths according to the behavior of their neighboring vehicles. Therefore, vehicles' trajectories are very correlated and considering vehicle interactions makes motion prediction possible even before the start of a clear maneuver pattern. To this end, we introduce and analyze trajectory prediction methods based on how they model the vehicles interactions. Inspired by human reasoning, we use an attention mechanism that explicitly highlights the importance of neighboring vehicles with respect to their future states. We go beyond pairwise vehicle interactions and model higher order interactions. Moreover, the existence of different goals and driving behaviors induces multiple potential futures. We exploit a combination of global and partial attention paid to surrounding vehicles to generate different possible trajectory. Experiments on highway datasets show that the proposed model outperforms the state-of-the-art performances.
Driving Behavior Modeling Using Naturalistic Human Driving Data With Inverse Reinforcement Learning Driving behavior modeling is of great importance for designing safe, smart, and personalized autonomous driving systems. In this paper, an internal reward function-based driving model that emulates the human’s decision-making mechanism is utilized. To infer the reward function parameters from naturalistic human driving data, we propose a structural assumption about human driving behavior that focuses on discrete latent driving intentions. It converts the continuous behavior modeling problem to a discrete setting and thus makes maximum entropy inverse reinforcement learning (IRL) tractable to learn reward functions. Specifically, a polynomial trajectory sampler is adopted to generate candidate trajectories considering high-level intentions and approximate the partition function in the maximum entropy IRL framework. An environment model considering interactive behaviors among the ego and surrounding vehicles is built to better estimate the generated trajectories. The proposed method is applied to learn personalized reward functions for individual human drivers from the NGSIM highway driving dataset. The qualitative results demonstrate that the learned reward functions are able to explicitly express the preferences of different drivers and interpret their decisions. The quantitative results reveal that the learned reward functions are robust, which is manifested by only a marginal decline in proximity to the human driving trajectories when applying the reward function in the testing conditions. For the testing performance, the personalized modeling method outperforms the general modeling approach, significantly reducing the modeling errors in human likeness (a custom metric to gauge accuracy), and these two methods deliver better results compared to other baseline methods. Moreover, it is found that predicting the response actions of surrounding vehicles and incorporating their potential decelerations caused by the ego vehicle are critical in estimating the generated trajectories, and the accuracy of personalized planning using the learned reward functions relies on the accuracy of the forecasting model.
Ontology-based methods for enhancing autonomous vehicle path planning We report the results of a first implementation demonstrating the use of an ontology to support reasoning about obstacles to improve the capabilities and performance of on-board route planning for autonomous vehicles. This is part of an overall effort to evaluate the performance of ontologies in different components of an autonomous vehicle within the 4D/RCS system architecture developed at NIST. Our initial focus has been on simple roadway driving scenarios where the controlled vehicle encounters potential obstacles in its path. As reported elsewhere [C. Schlenoff, S. Balakirsky, M. Uschold, R. Provine, S. Smith, Using ontologies to aid navigation planning in autonomous vehicles, Knowledge Engineering Review 18 (3) (2004) 243–255], our approach is to develop an ontology of objects in the environment, in conjunction with rules for estimating the damage that would be incurred by collisions with different objects in different situations. Automated reasoning is used to estimate collision damage; this information is fed to the route planner to help it decide whether to plan to avoid the object. We describe the results of the first implementation that integrates the ontology, the reasoner and the planner. We describe our insights and lessons learned and discuss resulting changes to our approach.
Integrated Simulation and Formal Verification of a Simple Autonomous Vehicle. This paper presents a proof-of-concept application of an approach to system development based on the integration of formal verification and co-simulation. A simple autonomous vehicle has the task of reaching an assigned straight path and then follow it, and it can be controlled by varying its turning speed. The correctness of the proposed control law has been formalized and verified by interactive theorem proving with the Prototype Verification System. Concurrently, the system has been co-simulated using the Prototype Verification System and the MathWorks Simulink tool: The vehicle kinematics have been simulated in Simulink, whereas the controller has been modeled in the logic language of the Prototype Verification System and simulated with the interpreter for the same language available in the theorem proving environment. With this approach, co-simulation and formal verification corroborate each other, thus strengthening developers' confidence in their analysis.
A Retargetable Fault Injection Framework for Safety Validation of Autonomous Vehicles Autonomous vehicles use Electronic Control Units running complex software to improve passenger comfort and safety. To test safety of in-vehicle electronics, the ISO 26262 standard on functional safety recommends using fault injection during component and system-level design. A Fault Injection Framework (FIF) induces hard-to-trigger hardware and software faults at runtime, enabling analysis of fault propagation effects. The growing number and complexity of diverse interacting components in vehicles demands a versatile FIF at the vehicle level. In this paper, we present a novel retargetable FIF based on debugger interfaces available on many target systems. We validated our FIF in three Hardware-In-the-Loop setups for autonomous driving based on the NXP BlueBox prototyping platform. To trigger a fault injection process, we developed an interactive user interface based on Robot Operating System, which also visualized vehicle system health. Our retargetable debugger-based fault injection mechanism confirmed safety properties and identified safety shortcomings of various automotive systems.
Test Scenario Generation and Optimization Technology for Intelligent Driving Systems In this paper, we propose a new scenario generation algorithm called Combinatorial Testing Based on Complexity (CTBC) based on both combinatorial testing (CT) method and Test Matrix (TM) technique for intelligent driving systems. To guide the generation procedure in the algorithm and evaluate the validity of the generated scenarios, we further propose a concept of complexity of test scenario. CTBC...
Decision-making in driver-automation shared control: A review and perspectives Shared control schemes allow a human driver to work with an automated driving agent in driver-vehicle systems while retaining the driver's abilities to control. The human driver, as an essential agent in the driver-vehicle shared control systems, should be precisely modeled regarding their cognitive processes, control strategies, and decision-making processes. The interactive strategy design between drivers and automated driving agents brings an excellent challenge for human-centric driver assistance systems due to the inherent characteristics of humans. Many open-ended questions arise, such as what proper role of human drivers should act in a shared control scheme? How to make an intelligent decision capable of balancing the benefits of agents in shared control systems? Due to the advent of these attentions and questions, it is desirable to present a survey on the decision making between human drivers and highly automated vehicles, to understand their architectures, human driver modeling, and interaction strategies under the driver-vehicle shared schemes. Finally, we give a further discussion on the key future challenges and opportunities. They are likely to shape new potential research directions.
Image Super-Resolution Using Deep Convolutional Networks We propose a deep learning method for single image super-resolution (SR). Our method directly learns an end-to-end mapping between the low/high-resolution images. The mapping is represented as a deep convolutional neural network (CNN) that takes the low-resolution image as the input and outputs the high-resolution one. We further show that traditional sparsecoding- based SR methods can also be viewed as a deep convolutional network. But unlike traditional methods that handle each component separately, our method jointly optimizes all layers. Our deep CNN has a lightweight structure, yet demonstrates state-of-the-art restoration quality, and achieves fast speed for practical on-line usage. We explore different network structures and parameter settings to achieve trade-offs between performance and speed. Moreover, we extend our network to cope with three color channels simultaneously, and show better overall reconstruction quality.
Some complexity questions related to distributive computing(Preliminary Report) Let M &equil; {0, 1, 2, ..., m—1} , N &equil; {0, 1, 2,..., n—1} , and f:M × N → {0, 1} a Boolean-valued function. We will be interested in the following problem and its related questions. Let i &egr; M, j &egr; N be integers known only to two persons P1 and P2, respectively. For P1 and P2 to determine cooperatively the value f(i, j), they send information to each other alternately, one bit at a time, according to some algorithm. The quantity of interest, which measures the information exchange necessary for computing f, is the minimum number of bits exchanged in any algorithm. For example, if f(i, j) &equil; (i + j) mod 2. then 1 bit of information (conveying whether i is odd) sent from P1 to P2 will enable P2 to determine f(i, j), and this is clearly the best possible. The above problem is a variation of a model of Abelson [1] concerning information transfer in distributive computions.
A constructive hybrid structure optimization methodology for radial basis probabilistic neural networks. In this paper, a novel heuristic structure optimization methodology for radial basis probabilistic neural networks (RBPNNs) is proposed. First, a minimum volume covering hyperspheres (MVCH) algorithm is proposed to select the initial hidden-layer centers of the RBPNN, and then the recursive orthogonal least square algorithm (ROLSA) combined with the particle swarm optimization (PSO) algorithm is adopted to further optimize the initial structure of the RBPNN. The proposed algorithms are evaluated through eight benchmark classification problems and two real-world application problems, a plant species identification task involving 50 plant species and a palmprint recognition task. Experimental results show that our proposed algorithm is feasible and efficient for the structure optimization of the RBPNN. The RBPNN achieves higher recognition rates and better classification efficiency than multilayer perceptron networks (MLPNs) and radial basis function neural networks (RBFNNs) in both tasks. Moreover, the experimental results illustrated that the generalization performance of the optimized RBPNN in the plant species identification task was markedly better than that of the optimized RBFNN.
Orthogonal moments based on exponent functions: Exponent-Fourier moments. In this paper, we propose a new set of orthogonal moments based on Exponent functions, named Exponent-Fourier moments (EFMs), which are suitable for image analysis and rotation invariant pattern recognition. Compared with Zernike polynomials of the same degree, the new radial functions have more zeros, and these zeros are evenly distributed, this property make EFMs have strong ability in describing image. Unlike Zernike moments, the kernel of computation of EFMs is extremely simple. Theoretical and experimental results show that Exponent-Fourier moments perform very well in terms of image reconstruction capability and invariant recognition accuracy in noise-free, noisy and smooth distortion conditions. The Exponent-Fourier moments can be thought of as generalized orthogonal complex moments.
When Vehicles See Pedestrians With Phones: A Multicue Framework for Recognizing Phone-Based Activities of Pedestrians The intelligent vehicle community has devoted considerable efforts to model driver behavior, and in particular, to detect and overcome driver distraction in an effort to reduce accidents caused by driver negligence. However, as the domain increasingly shifts toward autonomous and semiautonomous solutions, the driver is no longer integral to the decision-making process, indicating a need to refocus...
Pricing-Based Channel Selection for D2D Content Sharing in Dynamic Environment In order to make device-to-device (D2D) content sharing give full play to its advantage of improving local area services, one of the important issues is to decide the channels that D2D pairs occupy. Most existing works study this issue in static environment, and ignore the guidance for D2D pairs to select the channel adaptively. In this paper, we investigate this issue in dynamic environment where...
1.078
0.076667
0.076667
0.066667
0.066667
0.066667
0.066667
0.033333
0.000009
0
0
0
0
0
Accurate detection of sleep apnea with long short-term memory network based on RR interval signals Sleep apnea is a common condition that is characterized by sleep-disordered breathing. Worldwide the number of apnea cases has increased and there has been a growing number of patients suffering from apnea complications. Unfortunately, many cases remain undetected, because expensive and inconvenient examination methods are formidable barriers with regard to the diagnostics. Furthermore, treatment monitoring depends on the same methods which also underpin the initial diagnosis; hence issues related to the examination methods cause difficulties with managing sleep apnea as well. Computer-Aided Diagnosis (CAD) systems could be a tool to increase the efficiency and efficacy of diagnosis. To investigate this hypothesis, we designed a deep learning model that classifies beat-to-beat interval traces, medically known as RR intervals, into apnea versus non-apnea. The RR intervals were extracted from Electrocardiogram (ECG) signals contained in the Apnea-ECG benchmark Database. Before feeding the RR intervals to the classification algorithm, the signal was band-pass filtered with an Ornstein–Uhlenbeck third-order Gaussian process. 10-fold cross-validation indicated that the Long Short-Term Memory (LSTM) network has 99.80% accuracy, 99.85% sensitivity, and 99.73% specificity. With hold-out validation, the same network achieved 81.30% accuracy, 59.90% sensitivity, and 91.75% specificity. During the design, we learned that the band-pass filter improved classification accuracy by over 20%. The increased performance resulted from the fact that neural activation functions can process a DC free signal more efficiently. The result is likely transferable to the design of other RR interval based CAD systems, where the filter can help to improve classification performance.
The Sybil Attack Large-scale peer-to-peer systems facesecurity threats from faulty or hostile remotecomputing elements. To resist these threats, manysuch systems employ redundancy. However, if asingle faulty entity can present multiple identities,it can control a substantial fraction of the system,thereby undermining this redundancy. Oneapproach to preventing these &quot;Sybil attacks&quot; is tohave a trusted agency certify identities. Thispaper shows that, without a logically centralizedauthority, Sybil...
BLEU: a method for automatic evaluation of machine translation Human evaluations of machine translation are extensive but expensive. Human evaluations can take months to finish and involve human labor that can not be reused. We propose a method of automatic machine translation evaluation that is quick, inexpensive, and language-independent, that correlates highly with human evaluation, and that has little marginal cost per run. We present this method as an automated understudy to skilled human judges which substitutes for them when there is need for quick or frequent evaluations.
Computational thinking Summary form only given. My vision for the 21st century, Computational Thinking, will be a fundamental skill used by everyone in the world. To reading, writing, and arithmetic, we should add computational thinking to every child's analytical ability. Computational thinking involves solving problems, designing systems, and understanding human behavior by drawing on the concepts fundamental to computer science. Thinking like a computer scientist means more than being able to program a computer. It requires the ability to abstract and thus to think at multiple levels of abstraction. In this talk I will give many examples of computational thinking, argue that it has already influenced other disciplines, and promote the idea that teaching computational thinking can not only inspire future generations to enter the field of computer science but benefit people in all fields.
Fuzzy logic in control systems: fuzzy logic controller. I.
Switching between stabilizing controllers This paper deals with the problem of switching between several linear time-invariant (LTI) controllers—all of them capable of stabilizing a speci4c LTI process—in such a way that the stability of the closed-loop system is guaranteed for any switching sequence. We show that it is possible to 4nd realizations for any given family of controller transfer matrices so that the closed-loop system remains stable, no matter how we switch among the controller. The motivation for this problem is the control of complex systems where con8icting requirements make a single LTI controller unsuitable. ? 2002 Published by Elsevier Science Ltd.
Tabu Search - Part I
Bidirectional recurrent neural networks In the first part of this paper, a regular recurrent neural network (RNN) is extended to a bidirectional recurrent neural network (BRNN). The BRNN can be trained without the limitation of using input information just up to a preset future frame. This is accomplished by training it simultaneously in positive and negative time direction. Structure and training procedure of the proposed network are explained. In regression and classification experiments on artificial data, the proposed structure gives better results than other approaches. For real data, classification experiments for phonemes from the TIMIT database show the same tendency. In the second part of this paper, it is shown how the proposed bidirectional structure can be easily modified to allow efficient estimation of the conditional posterior probability of complete symbol sequences without making any explicit assumption about the shape of the distribution. For this part, experiments on real data are reported
An intensive survey of fair non-repudiation protocols With the phenomenal growth of the Internet and open networks in general, security services, such as non-repudiation, become crucial to many applications. Non-repudiation services must ensure that when Alice sends some information to Bob over a network, neither Alice nor Bob can deny having participated in a part or the whole of this communication. Therefore a fair non-repudiation protocol has to generate non-repudiation of origin evidences intended to Bob, and non-repudiation of receipt evidences destined to Alice. In this paper, we clearly define the properties a fair non-repudiation protocol must respect, and give a survey of the most important non-repudiation protocols without and with trusted third party (TTP). For the later ones we discuss the evolution of the TTP's involvement and, between others, describe the most recent protocol using a transparent TTP. We also discuss some ad-hoc problems related to the management of non-repudiation evidences.
Dynamic movement and positioning of embodied agents in multiparty conversations For embodied agents to engage in realistic multiparty conversation, they must stand in appropriate places with respect to other agents and the environment. When these factors change, such as an agent joining the conversation, the agents must dynamically move to a new location and/or orientation to accommodate. This paper presents an algorithm for simulating movement of agents based on observed human behavior using techniques developed for pedestrian movement in crowd simulations. We extend a previous group conversation simulation to include an agent motion algorithm. We examine several test cases and show how the simulation generates results that mirror real-life conversation settings.
An improved genetic algorithm with conditional genetic operators and its application to set-covering problem The genetic algorithm (GA) is a popular, biologically inspired optimization method. However, in the GA there is no rule of thumb to design the GA operators and select GA parameters. Instead, trial-and-error has to be applied. In this paper we present an improved genetic algorithm in which crossover and mutation are performed conditionally instead of probability. Because there are no crossover rate and mutation rate to be selected, the proposed improved GA can be more easily applied to a problem than the conventional genetic algorithms. The proposed improved genetic algorithm is applied to solve the set-covering problem. Experimental studies show that the improved GA produces better results over the conventional one and other methods.
Lane-level traffic estimations using microscopic traffic variables This paper proposes a novel inference method to estimate lane-level traffic flow, time occupancy and vehicle inter-arrival time on road segments where local information could not be measured and assessed directly. The main contributions of the proposed method are 1) the ability to perform lane-level estimations of traffic flow, time occupancy and vehicle inter-arrival time and 2) the ability to adapt to different traffic regimes by assessing only microscopic traffic variables. We propose a modified Kriging estimation model which explicitly takes into account both spatial and temporal variability. Performance evaluations are conducted using real-world data under different traffic regimes and it is shown that the proposed method outperforms a Kalman filter-based approach.
Convolutional Neural Network-Based Classification of Driver's Emotion during Aggressive and Smooth Driving Using Multi-Modal Camera Sensors. Because aggressive driving often causes large-scale loss of life and property, techniques for advance detection of adverse driver emotional states have become important for the prevention of aggressive driving behaviors. Previous studies have primarily focused on systems for detecting aggressive driver emotion via smart-phone accelerometers and gyro-sensors, or they focused on methods of detecting physiological signals using electroencephalography (EEG) or electrocardiogram (ECG) sensors. Because EEG and ECG sensors cause discomfort to drivers and can be detached from the driver's body, it becomes difficult to focus on bio-signals to determine their emotional state. Gyro-sensors and accelerometers depend on the performance of GPS receivers and cannot be used in areas where GPS signals are blocked. Moreover, if driving on a mountain road with many quick turns, a driver's emotional state can easily be misrecognized as that of an aggressive driver. To resolve these problems, we propose a convolutional neural network (CNN)-based method of detecting emotion to identify aggressive driving using input images of the driver's face, obtained using near-infrared (NIR) light and thermal camera sensors. In this research, we conducted an experiment using our own database, which provides a high classification accuracy for detecting driver emotion leading to either aggressive or smooth (i.e., relaxed) driving. Our proposed method demonstrates better performance than existing methods.
Robot tutor and pupils’ educational ability: Teaching the times tables Research shows promising results of educational robots in language and STEM tasks. In language, more research is available, occasionally in view of individual differences in pupils’ educational ability levels, and learning seems to improve with more expressive robot behaviors. In STEM, variations in robots’ behaviors have been examined with inconclusive results and never while systematically investigating how differences in educational abilities match with different robot behaviors. We applied an autonomously tutoring robot (without tablet, partly WOz) in a 2 × 2 experiment of social vs. neutral behavior in above-average vs. below-average schoolchildren (N = 86; age 8–10 years) while rehearsing the multiplication tables on a one-to-one basis. The standard school test showed that on average, pupils significantly improved their performance even after 3 occasions of 5-min exercises. Beyond-average pupils profited most from a robot tutor, whereas those below average in multiplication benefited more from a robot that showed neutral rather than more social behavior.
1
0
0
0
0
0
0
0
0
0
0
0
0
0
Fuzzy Logic in Dynamic Parameter Adaptation of Harmony Search Optimization for Benchmark Functions and Fuzzy Controllers. Nowadays the use of fuzzy logic has been increasing in popularity, and this is mainly due to the inference mechanism that allows simulating human reasoning in knowledge-based systems. The main contribution of this work is using the concepts of fuzzy logic in a method for dynamically adapting the main parameters of the harmony search algorithm during execution. Dynamical adaptation of parameters in metaheuristics has been shown to improve performance and accuracy in a wide range of applications. For this reason, we propose and approach for fuzzy adaptation of parameters in harmony search. Two case studies are considered for testing the proposed approach, the optimization of mathematical functions, which are unimodal, multimodal, hybrid, and composite functions and a control problem without noise and when noise is considered. A statistical comparison between the harmony search algorithm and the fuzzy harmony search algorithm is presented to verify the advantages of the proposed approach.
Mobile cloud computing: A survey Despite increasing usage of mobile computing, exploiting its full potential is difficult due to its inherent problems such as resource scarcity, frequent disconnections, and mobility. Mobile cloud computing can address these problems by executing mobile applications on resource providers external to the mobile device. In this paper, we provide an extensive survey of mobile cloud computing research, while highlighting the specific concerns in mobile cloud computing. We present a taxonomy based on the key issues in this area, and discuss the different approaches taken to tackle these issues. We conclude the paper with a critical analysis of challenges that have not yet been fully met, and highlight directions for future work.
Harmony search algorithm for solving Sudoku Harmony search (HS) algorithm was applied to solving Sudoku puzzle. The HS is an evolutionary algorithm which mimics musicians' behaviors such as random play, memory-based play, and pitch-adjusted play when they perform improvisation. Sudoku puzzles in this study were formulated as an optimization problem with number-uniqueness penalties. HS could successfully solve the optimization problem after 285 function evaluations, taking 9 seconds. Also, sensitivity analysis of HS parameters was performed to obtain a better idea of algorithm parameter values.
DEC: dynamically evolving clustering and its application to structure identification of evolving fuzzy models. Identification of models from input-output data essentially requires estimation of appropriate cluster centers. In this paper, a new online evolving clustering approach for streaming data is proposed. Unlike other approaches that consider either the data density or distance from existing cluster centers, this approach uses cluster weight and distance before generating new clusters. To capture the dynamics of the data stream, the cluster weight is defined in both data and time space in such a way that it decays exponentially with time. It also applies concepts from computational geometry to determine the neighborhood information while forming clusters. A distinction is made between core and noncore clusters to effectively identify the real outliers. The approach efficiently estimates cluster centers upon which evolving Takagi-Sugeno models are developed. The experimental results with developed models show that the proposed approach attains results at par or better than existing approaches and significantly reduces the computational overhead, which makes it suitable for real-time applications.
An Easily Understandable Grey Wolf Optimizer and Its Application to Fuzzy Controller Tuning. This paper proposes an easily understandable Grey Wolf Optimizer (GWO) applied to the optimal tuning of the parameters of Takagi-Sugeno proportional-integral fuzzy controllers (T-S PI-FCs). GWO is employed for solving optimization problems focused on the minimization of discrete-time objective functions defined as the weighted sum of the absolute value of the control error and of the squared output sensitivity function, and the vector variable consists of the tuning parameters of the T-S PI-FCs. Since the sensitivity functions are introduced with respect to the parametric variations of the process, solving these optimization problems is important as it leads to fuzzy control systems with a reduced process parametric sensitivity obtained by a GWO-based fuzzy controller tuning approach. GWO algorithms applied with this regard are formulated in easily understandable terms for both vector and scalar operations, and discussions on stability, convergence, and parameter settings are offered. The controlled processes referred to in the course of this paper belong to a family of nonlinear servo systems, which are modeled by second order dynamics plus a saturation and dead zone static nonlinearity. Experimental results concerning the angular position control of a laboratory servo system are included for validating the proposed method.
Stability Analysis and Estimation of Domain of Attraction for Positive Polynomial Fuzzy Systems With Input Saturation AbstractIn this paper, the stability and positivity of positive polynomial fuzzy model based (PPFMB) control system are investigated, in which the positive polynomial fuzzy model and positive polynomial fuzzy controller are allowed to have different premise membership functions from each other. These mismatched premise membership functions can increase the flexibility of controller design; however, it will lead to the conservative results when the stability is analyzed based on the Lyapunov stability theory. To relax the positivity/stability conditions, the improved Taylor-series-membership-functions-dependent (ITSMFD) method is introduced by introducing the sample points information of Taylor-series approximate membership functions, local error information and boundary information of substate space of premise variables into the stability/positivity conditions. Meanwhile, the ITSMFD method is extended to the PPFMB control system with input saturation to relax the estimation of domain of attraction. Finally, simulation examples are presented to verify the feasibility of this method.
A fuzzy hierarchical operator in the grey wolf optimizer algorithm. •The main goal is to study the performance of the Grey Wolf Optimizer algorithm when a new hierarchical operator is introduced.•This new operator is basically a hierarchical transformation that is inspired in the hierarchical social pyramid of the grey wolf.•This proposed operator is applied to the simulation of the hunting process in the algorithm and has 5 variants are presented.•Notably the variants having the greatest impact in the GWO performance are based on the use fuzzy logic.
Theory and Experiment on Formation-Containment Control of Multiple Multirotor Unmanned Aerial Vehicle Systems. Formation-containment control problems for multiple multirotor unmanned aerial vehicle (UAV) systems with directed topologies are studied, where the states of leaders form desired formation and the states of followers converge to the convex hull spanned by those of the leaders. First, formation-containment protocols are constructed based on the neighboring information of UAVs. Then, sufficient con...
Sequence to Sequence Learning with Neural Networks. Deep Neural Networks (DNNs) are powerful models that have achieved excellent performance on difficult learning tasks. Although DNNs work well whenever large labeled training sets are available, they cannot be used to map sequences to sequences. In this paper, we present a general end-to-end approach to sequence learning that makes minimal assumptions on the sequence structure. Our method uses a multilayered Long Short-Term Memory (LSTM) to map the input sequence to a vector of a fixed dimensionality, and then another deep LSTM to decode the target sequence from the vector. Our main result is that on an English to French translation task from the WMT-14 dataset, the translations produced by the LSTM achieve a BLEU score of 34.8 on the entire test set, where the LSTM's BLEU score was penalized on out-of-vocabulary words. Additionally, the LSTM did not have difficulty on long sentences. For comparison, a phrase-based SMT system achieves a BLEU score of 33.3 on the same dataset. When we used the LSTM to rerank the 1000 hypotheses produced by the aforementioned SMT system, its BLEU score increases to 36.5, which is close to the previous state of the art. The LSTM also learned sensible phrase and sentence representations that are sensitive to word order and are relatively invariant to the active and the passive voice. Finally, we found that reversing the order of the words in all source sentences (but not target sentences) improved the LSTM's performance markedly, because doing so introduced many short term dependencies between the source and the target sentence which made the optimization problem easier.
Lambertian Reflectance and Linear Subspaces We prove that the set of all Lambertian reflectance functions (the mapping from surface normals to intensities) obtained with arbitrary distant light sources lies close to a 9D linear subspace. This implies that, in general, the set of images of a convex Lambertian object obtained under a wide variety of lighting conditions can be approximated accurately by a low-dimensional linear subspace, explaining prior empirical results. We also provide a simple analytic characterization of this linear space. We obtain these results by representing lighting using spherical harmonics and describing the effects of Lambertian materials as the analog of a convolution. These results allow us to construct algorithms for object recognition based on linear methods as well as algorithms that use convex optimization to enforce nonnegative lighting functions. We also show a simple way to enforce nonnegative lighting when the images of an object lie near a 4D linear space. We apply these algorithms to perform face recognition by finding the 3D model that best matches a 2D query image.
Evolutionary dynamic optimization: A survey of the state of the art. Optimization in dynamic environments is a challenging but important task since many real-world optimization problems are changing over time. Evolutionary computation and swarm intelligence are good tools to address optimization problems in dynamic environments due to their inspiration from natural self-organized systems and biological evolution, which have always been subject to changing environments. Evolutionary optimization in dynamic environments, or evolutionary dynamic optimization (EDO), has attracted a lot of research effort during the last 20 years, and has become one of the most active research areas in the field of evolutionary computation. In this paper we carry out an in-depth survey of the state-of-the-art of academic research in the field of EDO and other meta-heuristics in four areas: benchmark problems/generators, performance measures, algorithmic approaches, and theoretical studies. The purpose is to for the first time (i) provide detailed explanations of how current approaches work; (ii) review the strengths and weaknesses of each approach; (iii) discuss the current assumptions and coverage of existing EDO research; and (iv) identify current gaps, challenges and opportunities in EDO.
FreeTagpaper: a pen-and-paper-based collaboration system using visual tags printed on paper We propose a novel collaboration system using pen and paper. FreeTagpaper uses a camera set above a work table to recognize the visual tags printed on the paper to acquire the paper's ID and posture. From the paper ID, FreeTagpaper can retrieve the paper's size and format from a document management database. It then uses the format information to transmit images of handwritten notes to the other side and simultaneously projects the other side's handwritten notes onto this paper. Based on this mutual image sharing scheme, users can collaborate in real time using paper of any size, format, and posture. This report describes preliminary work on FreeTagpaper including the current implementation and essential experimental results.
The ApolloScape Dataset for Autonomous Driving Scene parsing aims to assign a class (semantic) label for each pixel in an image. It is a comprehensive analysis of an image. Given the rise of autonomous driving, pixel-accurate environmental perception is expected to be a key enabling technical piece. However, providing a large scale dataset for the design and evaluation of scene parsing algorithms, in particular for outdoor scenes, has been difficult. The per-pixel labelling process is prohibitively expensive, limiting the scale of existing ones. In this paper, we present a large-scale open dataset, ApolloScape, that consists of RGB videos and corresponding dense 3D point clouds. Comparing with existing datasets, our dataset has the following unique properties. The first is its scale, our initial release contains over 140K images - each with its per-pixel semantic mask, up to 1M is scheduled. The second is its complexity. Captured in various traffic conditions, the number of moving objects averages from tens to over one hundred (Figure 1). And the third is the 3D attribute, each image is tagged with high-accuracy pose information at cm accuracy and the static background point cloud has mm relative accuracy. We are able to label these many images by an interactive and efficient labelling pipeline that utilizes the high-quality 3D point cloud. Moreover, our dataset also contains different lane markings based on the lane colors and styles. We expect our new dataset can deeply benefit various autonomous driving related applications that include but not limited to 2D/3D scene understanding, localization, transfer learning, and driving simulation.
Attitudes Towards Social Robots In Education: Enthusiast, Practical, Troubled, Sceptic, And Mindfully Positive While social robots bring new opportunities for education, they also come with moral challenges. Therefore, there is a need for moral guidelines for the responsible implementation of these robots. When developing such guidelines, it is important to include different stakeholder perspectives. Existing (qualitative) studies regarding these perspectives however mainly focus on single stakeholders. In this exploratory study, we examine and compare the attitudes of multiple stakeholders on the use of social robots in primary education, using a novel questionnaire that covers various aspects of moral issues mentioned in earlier studies. Furthermore, we also group the stakeholders based on similarities in attitudes and examine which socio-demographic characteristics influence these attitude types. Based on the results, we identify five distinct attitude profiles and show that the probability of belonging to a specific profile is affected by such characteristics as stakeholder type, age, education and income. Our results also indicate that social robots have the potential to be implemented in education in a morally responsible way that takes into account the attitudes of various stakeholders, although there are multiple moral issues that need to be addressed first. Finally, we present seven (practical) implications for a responsible application of social robots in education following from our results. These implications provide valuable insights into how social robots should be implemented.
1.11
0.12
0.12
0.12
0.12
0.12
0.06
0.006667
0
0
0
0
0
0
Multiadvisor Reinforcement Learning for Multiagent Multiobjective Smart Home Energy Control Effective automated smart home energy control is essential for smart grid approaches to demand response (DR). This is a multiobjective adaptive control problem because it balances an appliance’s primary objective with demand response objectives. One challenge comes from the heterogeneous nature of objectives, requiring tradeoffs between comfort, cost, and other objectives. Another challenge comes from the heterogeneous dynamics, which result from different environments and the different appliances used. Another challenge is nonstationary nature of dynamics and rewards due to seasonal changes and time-varying user preferences. Finally, we consider computational challenges, required by the real-time aspect of the control problem, particularly notable due to “the curse of dimensionality.” We propose a multiagent multiadvisor reinforcement learning framework to address these challenges. We design a smart-home simulation to demonstrate the performance (in terms of weighted reward) of our approach relative to competitive single-objective reinforcement learning algorithms. Furthermore, we theoretically and empirically demonstrate the linear computational scalability of the algorithm. Finally, we identify the need for key performance measures of the proposed system by considering the effect of selected preferences on agents. Overall, the proposed algorithm is reasonably competitive with conventional approaches while simultaneously enabling behavior changes with change in preferences without requiring more data.
The Sybil Attack Large-scale peer-to-peer systems facesecurity threats from faulty or hostile remotecomputing elements. To resist these threats, manysuch systems employ redundancy. However, if asingle faulty entity can present multiple identities,it can control a substantial fraction of the system,thereby undermining this redundancy. Oneapproach to preventing these &quot;Sybil attacks&quot; is tohave a trusted agency certify identities. Thispaper shows that, without a logically centralizedauthority, Sybil...
BLEU: a method for automatic evaluation of machine translation Human evaluations of machine translation are extensive but expensive. Human evaluations can take months to finish and involve human labor that can not be reused. We propose a method of automatic machine translation evaluation that is quick, inexpensive, and language-independent, that correlates highly with human evaluation, and that has little marginal cost per run. We present this method as an automated understudy to skilled human judges which substitutes for them when there is need for quick or frequent evaluations.
Computational thinking Summary form only given. My vision for the 21st century, Computational Thinking, will be a fundamental skill used by everyone in the world. To reading, writing, and arithmetic, we should add computational thinking to every child's analytical ability. Computational thinking involves solving problems, designing systems, and understanding human behavior by drawing on the concepts fundamental to computer science. Thinking like a computer scientist means more than being able to program a computer. It requires the ability to abstract and thus to think at multiple levels of abstraction. In this talk I will give many examples of computational thinking, argue that it has already influenced other disciplines, and promote the idea that teaching computational thinking can not only inspire future generations to enter the field of computer science but benefit people in all fields.
Fuzzy logic in control systems: fuzzy logic controller. I.
Switching between stabilizing controllers This paper deals with the problem of switching between several linear time-invariant (LTI) controllers—all of them capable of stabilizing a speci4c LTI process—in such a way that the stability of the closed-loop system is guaranteed for any switching sequence. We show that it is possible to 4nd realizations for any given family of controller transfer matrices so that the closed-loop system remains stable, no matter how we switch among the controller. The motivation for this problem is the control of complex systems where con8icting requirements make a single LTI controller unsuitable. ? 2002 Published by Elsevier Science Ltd.
Tabu Search - Part I
Bidirectional recurrent neural networks In the first part of this paper, a regular recurrent neural network (RNN) is extended to a bidirectional recurrent neural network (BRNN). The BRNN can be trained without the limitation of using input information just up to a preset future frame. This is accomplished by training it simultaneously in positive and negative time direction. Structure and training procedure of the proposed network are explained. In regression and classification experiments on artificial data, the proposed structure gives better results than other approaches. For real data, classification experiments for phonemes from the TIMIT database show the same tendency. In the second part of this paper, it is shown how the proposed bidirectional structure can be easily modified to allow efficient estimation of the conditional posterior probability of complete symbol sequences without making any explicit assumption about the shape of the distribution. For this part, experiments on real data are reported
An intensive survey of fair non-repudiation protocols With the phenomenal growth of the Internet and open networks in general, security services, such as non-repudiation, become crucial to many applications. Non-repudiation services must ensure that when Alice sends some information to Bob over a network, neither Alice nor Bob can deny having participated in a part or the whole of this communication. Therefore a fair non-repudiation protocol has to generate non-repudiation of origin evidences intended to Bob, and non-repudiation of receipt evidences destined to Alice. In this paper, we clearly define the properties a fair non-repudiation protocol must respect, and give a survey of the most important non-repudiation protocols without and with trusted third party (TTP). For the later ones we discuss the evolution of the TTP's involvement and, between others, describe the most recent protocol using a transparent TTP. We also discuss some ad-hoc problems related to the management of non-repudiation evidences.
Dynamic movement and positioning of embodied agents in multiparty conversations For embodied agents to engage in realistic multiparty conversation, they must stand in appropriate places with respect to other agents and the environment. When these factors change, such as an agent joining the conversation, the agents must dynamically move to a new location and/or orientation to accommodate. This paper presents an algorithm for simulating movement of agents based on observed human behavior using techniques developed for pedestrian movement in crowd simulations. We extend a previous group conversation simulation to include an agent motion algorithm. We examine several test cases and show how the simulation generates results that mirror real-life conversation settings.
An improved genetic algorithm with conditional genetic operators and its application to set-covering problem The genetic algorithm (GA) is a popular, biologically inspired optimization method. However, in the GA there is no rule of thumb to design the GA operators and select GA parameters. Instead, trial-and-error has to be applied. In this paper we present an improved genetic algorithm in which crossover and mutation are performed conditionally instead of probability. Because there are no crossover rate and mutation rate to be selected, the proposed improved GA can be more easily applied to a problem than the conventional genetic algorithms. The proposed improved genetic algorithm is applied to solve the set-covering problem. Experimental studies show that the improved GA produces better results over the conventional one and other methods.
Lane-level traffic estimations using microscopic traffic variables This paper proposes a novel inference method to estimate lane-level traffic flow, time occupancy and vehicle inter-arrival time on road segments where local information could not be measured and assessed directly. The main contributions of the proposed method are 1) the ability to perform lane-level estimations of traffic flow, time occupancy and vehicle inter-arrival time and 2) the ability to adapt to different traffic regimes by assessing only microscopic traffic variables. We propose a modified Kriging estimation model which explicitly takes into account both spatial and temporal variability. Performance evaluations are conducted using real-world data under different traffic regimes and it is shown that the proposed method outperforms a Kalman filter-based approach.
Convolutional Neural Network-Based Classification of Driver's Emotion during Aggressive and Smooth Driving Using Multi-Modal Camera Sensors. Because aggressive driving often causes large-scale loss of life and property, techniques for advance detection of adverse driver emotional states have become important for the prevention of aggressive driving behaviors. Previous studies have primarily focused on systems for detecting aggressive driver emotion via smart-phone accelerometers and gyro-sensors, or they focused on methods of detecting physiological signals using electroencephalography (EEG) or electrocardiogram (ECG) sensors. Because EEG and ECG sensors cause discomfort to drivers and can be detached from the driver's body, it becomes difficult to focus on bio-signals to determine their emotional state. Gyro-sensors and accelerometers depend on the performance of GPS receivers and cannot be used in areas where GPS signals are blocked. Moreover, if driving on a mountain road with many quick turns, a driver's emotional state can easily be misrecognized as that of an aggressive driver. To resolve these problems, we propose a convolutional neural network (CNN)-based method of detecting emotion to identify aggressive driving using input images of the driver's face, obtained using near-infrared (NIR) light and thermal camera sensors. In this research, we conducted an experiment using our own database, which provides a high classification accuracy for detecting driver emotion leading to either aggressive or smooth (i.e., relaxed) driving. Our proposed method demonstrates better performance than existing methods.
Ethical Considerations Of Applying Robots In Kindergarten Settings: Towards An Approach From A Macroperspective In child-robot interaction (cHRI) research, many studies pursue the goal to develop interactive systems that can be applied in everyday settings. For early education, increasingly, the setting of a kindergarten is targeted. However, when cHRI and research are brought into a kindergarten, a range of ethical and related procedural aspects have to be considered and dealt with. While ethical models elaborated within other human-robot interaction settings, e.g., assisted living contexts, can provide some important indicators for relevant issues, we argue that it is important to start developing a systematic approach to identify and tackle those ethical issues which rise with cHRI in kindergarten settings on a more global level and address the impact of the technology from a macroperspective beyond the effects on the individual. Based on our experience in conducting studies with children in general and pedagogical considerations on the role of the institution of kindergarten in specific, in this paper, we enfold some relevant aspects that have barely been addressed in an explicit way in current cHRI research. Four areas are analyzed and key ethical issues are identified in each area: (1) the institutional setting of a kindergarten, (2) children as a vulnerable group, (3) the caregivers' role, and (4) pedagogical concepts. With our considerations, we aim at (i) broadening the methodology of the current studies within the area of cHRI, (ii) revalidate it based on our comprehensive empirical experience with research in kindergarten settings, both laboratory and real-world contexts, and (iii) provide a framework for the development of a more systematic approach to address the ethical issues in cHRI research within kindergarten settings.
1
0
0
0
0
0
0
0
0
0
0
0
0
0
Modeling Aspects of Theory of Mind with Markov Random Fields We propose Markov random fields (MRFs) as a probabilistic mathematical model for incorporating the in- ternal states of other agents, both human and robotic, into ro- bot decision making. By using estimates of Theory of Mind (ToM), the mental states of other agents can be incorporated into decision making through statistical inference, allowing robots to balance their own goals and internal objectives with those of other collaborating agents. The MRF model is well-suited to domains in which the joint probability over latent (action) and observed (perceived) variables can be factored into pairwise interactions between these variables. Specifically, these interactions occur through functions that evaluate "local evidence" between an observed and latent variable and "compatibility" between a pair of latent vari- ables. We will describe how experimental findings from the ToM literature can be explained using MRF models, and then show how this framework can be applied to a social robotics task. We will also describe how to use belief prop- agation on a multi-robot MRF as a novel approach to multi- robot coordination, with parallels to human collaboration
Embodiment and Human-Robot Interaction: A Task-Based Perspective In this work, we further test the hypothesis that physical embodiment has a measurable effect on performance and impression of social interactions. Support for this hypothesis would suggest fundamental differences between virtual agents and robots from a social standpoint and would have significan t implications for human-robot interaction. We have refined our task-based metrics to give a measure- ment, not only of the participant's immediate impressions o f a coach for a task, but also of the participant's performance i n a given task. We measure task performance and participants' impression of a robot's social abilities in a structured tas k based on the Towers of Hanoi puzzle. Our experiment compares aspects of embodiment by evaluating: (1) the difference between a physical robot and a simulated one; and (2) the effect of physical presence through a co-located robot versus a remote, tele-present robot. With a participant pool (n=21) of roboticists and non- roboticists, we were able to show that participants felt that an embodied robot was more appealing and perceptive of the world than non-embodied robots. A larger pool of participants (n=32) also demonstrated that the embodied robot was seen as most helpful, watchful, and enjoyable when compared to a remote tele-present robot and a simulated robot.
The benefit of being physically present: A survey of experimental works comparing copresent robots, telepresent robots and virtual agents. The effects of physical embodiment and physical presence were explored through a survey of 33 experimental works comparing how people interacted with physical robots and virtual agents. A qualitative assessment of the direction of quantitative effects demonstrated that robots were more persuasive and perceived more positively when physically present in a user׳s environment than when digitally-displayed on a screen either as a video feed of the same robot or as a virtual character analog; robots also led to better user performance when they were collocated as opposed to shown via video on a screen. However, participants did not respond differently to physical robots and virtual agents when both were displayed digitally on a screen – suggesting that physical presence, rather than physical embodiment, characterizes people׳s responses to social robots. Implications for understanding psychological response to physical and virtual agents and for methodological design are discussed.
Help Me Please: Robot Politeness Strategies for Soliciting Help From Humans. Robots that can leverage help from people could accomplish much more than robots that cannot. We present the results of two experiments that examine how robots can more effectively request help from people. Study 1 is a video prototype experiment (N=354), investigating the effectiveness of four linguistic politeness strategies as well as the effects of social status (equal, low), size of request (large, small), and robot familiarity (high, low) on people's willingness to help a robot. The results of this study largely support Politeness Theory and the Computers as Social Actors paradigm. Study 2 is a physical human-robot interaction experiment (N=48), examining the impact of source orientation (autonomous, single operator, multiple operators) on people's behavioral willingness to help the robot. People were nearly 50% faster to help the robot if they perceived it to be autonomous rather than being teleoperated. Implications for research design, theory, and methods are discussed.
Deal or No Deal? End-to-End Learning of Negotiation Dialogues. Much of human dialogue occurs in semi-cooperative settings, where agents with different goals attempt to agree on common decisions. Negotiations require complex communication and reasoning skills, but success is easy to measure, making this an interesting task for AI. We gather a large dataset of human-human negotiations on a multi-issue bargaining task, where agents who cannot observe each otheru0027s reward functions must reach an agreement (or a deal) via natural language dialogue. For the first time, we show it is possible to train end-to-end models for negotiation, which must learn both linguistic and reasoning skills with no annotated dialogue states. We also introduce dialogue rollouts, in which the model plans ahead by simulating possible complete continuations of the conversation, and find that this technique dramatically improves performance. Our code and dataset are publicly available (this https URL).
No fair!!: an interaction with a cheating robot Using a humanoid robot and a simple children's game, we examine the degree to which variations in behavior result in attributions of mental state and intentionality. Participants play the well-known children's game ¿rock-paper-scissors¿ against a robot that either plays fairly, or that cheats in one of two ways. In the ¿verbal cheat¿ condition, the robot announces the wrong outcome on several rounds which it loses, declaring itself the winner. In the ¿action cheat¿ condition, the robot changes its gesture after seeing its opponent's play. We find that participants display a greater level of social engagement and make greater attributions of mental state when playing against the robot in the conditions in which it cheats.
Accurate Self-Localization in RFID Tag Information Grids Using FIR Filtering Grid navigation spaces nested with the radio-frequency identification (RFID) tags are promising for industrial and other needs, because each tag can deliver information about a local two-dimensional or three-dimensional surrounding. The approach, however, requires high accuracy in vehicle self-localization. Otherwise, errors may lead to collisions; possibly even fatal. We propose a new extended finite impulse response (EFIR) filtering algorithm and show that it meets this need. The EFIR filter requires an optimal averaging interval, but does not involve the noise statistics which are often not well known to the engineer. It is more accurate than the extended Kalman filter (EKF) under real operation conditions and its iterative algorithm has the Kalman form. Better performance of the proposed EFIR filter is demonstrated based on extensive simulations in a comparison to EKF, which is widely used in RFID tag grids. We also show that errors in noise covariances may provoke divergence in EKF, whereas the EFIR filter remains stable and is thus more robust.
Evolutionary computation: comments on the history and current state Evolutionary computation has started to receive significant attention during the last decade, although the origins can be traced back to the late 1950's. This article surveys the history as well as the current state of this rapidly growing field. We describe the purpose, the general structure, and the working principles of different approaches, including genetic algorithms (GA) (with links to genetic programming (GP) and classifier systems (CS)), evolution strategies (ES), and evolutionary programming (EP) by analysis and comparison of their most important constituents (i.e. representations, variation operators, reproduction, and selection mechanism). Finally, we give a brief overview on the manifold of application domains, although this necessarily must remain incomplete
Supporting social navigation on the World Wide Web This paper discusses a navigation behavior on Internet information services, in particular the World Wide Web, which is characterized by pointing out of information using various communication tools. We call this behavior social navigation as it is based on communication and interaction with other users, be that through email, or any other means of communication. Social navigation phenomena are quite common although most current tools (like Web browsers or email clients) offer very little support for it. We describe why social navigation is useful and how it can be better supported in future systems. We further describe two prototype systems that, although originally not designed explicitly as tools for social navigation, provide features that are typical for social navigation systems. One of these systems, the Juggler system, is a combination of a textual virtual environment and a Web client. The other system is a prototype of a Web- hotlist organizer, called Vortex. We use both systems to describe fundamental principles of social navigation systems.
Proofs of Storage from Homomorphic Identification Protocols Proofs of storage (PoS) are interactive protocols allowing a client to verify that a server faithfully stores a file. Previous work has shown that proofs of storage can be constructed from any homomorphic linear authenticator (HLA). The latter, roughly speaking, are signature/message authentication schemes where `tags' on multiple messages can be homomorphically combined to yield a `tag' on any linear combination of these messages. We provide a framework for building public-key HLAs from any identification protocol satisfying certain homomorphic properties. We then show how to turn any public-key HLA into a publicly-verifiable PoS with communication complexity independent of the file length and supporting an unbounded number of verifications. We illustrate the use of our transformations by applying them to a variant of an identification protocol by Shoup, thus obtaining the first unbounded-use PoS based on factoring (in the random oracle model).
Design, Implementation, and Experimental Results of a Quaternion-Based Kalman Filter for Human Body Motion Tracking Real-time tracking of human body motion is an important technology in synthetic environments, robotics, and other human-computer interaction applications. This paper presents an extended Kalman filter designed for real-time estimation of the orientation of human limb segments. The filter processes data from small inertial/magnetic sensor modules containing triaxial angular rate sensors, accelerometers, and magnetometers. The filter represents rotation using quaternions rather than Euler angles or axis/angle pairs. Preprocessing of the acceleration and magnetometer measurements using the Quest algorithm produces a computed quaternion input for the filter. This preprocessing reduces the dimension of the state vector and makes the measurement equations linear. Real-time implementation and testing results of the quaternion-based Kalman filter are presented. Experimental results validate the filter design, and show the feasibility of using inertial/magnetic sensor modules for real-time human body motion tracking
Switching Stabilization for a Class of Slowly Switched Systems In this technical note, the problem of switching stabilization for slowly switched linear systems is investigated. In particular, the considered systems can be composed of all unstable subsystems. Based on the invariant subspace theory, the switching signal with mode-dependent average dwell time (MDADT) property is designed to exponentially stabilize the underlying system. Furthermore, sufficient condition of stabilization for switched systems with all stable subsystems under MDADT switching is also given. The correctness and effectiveness of the proposed approaches are illustrated by a numerical example.
An efficient scheduling scheme for mobile charger in on-demand wireless rechargeable sensor networks. Existing studies on wireless sensor networks (WSNs) have revealed that the limited battery capacity of sensor nodes (SNs) hinders their perpetual operation. Recent findings in the domain of wireless energy transfer (WET) have attracted a lot of attention of academia and industry to cater the lack of energy in the WSNs. The main idea of WET is to restore the energy of SNs using one or more wireless mobile chargers (MCs), which leads to a new paradigm of wireless rechargeable sensor networks (WRSNs). The determination of an optimal order of charging the SNs (i.e., charging schedule) in an on-demand WRSN is a well-known NP-hard problem. Moreover, care must be taken while designing the charging schedule of an MC as requesting SNs introduce both spatial and temporal constraints. In this paper, we first present a Linear Programming (LP) formulation for the problem of scheduling an MC and then propose an efficient solution based on gravitational search algorithm (GSA). Our method is presented with a novel agent representation scheme and an efficient fitness function. We perform extensive simulations on the proposed scheme to demonstrate its effectiveness over two state-of-the-art algorithms, namely first come first serve (FCFS) and nearest job next with preemption (NJNP). The simulation results reveal that the proposed scheme outperforms both the existing algorithms in terms of charging latency. The virtue of our scheme is also proved by the well-known statistical test, analysis of variance (ANOVA), followed by post hoc analysis.
Hardware Circuits Design and Performance Evaluation of a Soft Lower Limb Exoskeleton Soft lower limb exoskeletons (LLEs) are wearable devices that have good potential in walking rehabilitation and augmentation. While a few studies focused on the structure design and assistance force optimization of the soft LLEs, rarely work has been conducted on the hardware circuits design. The main purpose of this work is to present a new soft LLE for walking efficiency improvement and introduce its hardware circuits design. A soft LLE for hip flexion assistance and a hardware circuits system with scalability were proposed. To assess the efficacy of the soft LLE, the experimental tests that evaluate the sensor data acquisition, force tracking performance, lower limb muscle activity and metabolic cost were conducted. The time error in the peak assistance force was just 1%. The reduction in the normalized root-mean-square EMG of the rectus femoris was 7.1%. The net metabolic cost in exoskeleton on condition was reduced by 7.8% relative to walking with no exoskeleton. The results show that the designed hardware circuits can be applied to the soft LLE and the soft LLE is able to improve walking efficiency of wearers.
1.2
0.2
0.2
0.2
0.1
0.04
0
0
0
0
0
0
0
0
A sub-constant error-probability low-degree test, and a sub-constant error-probability PCP characterization of NP
Minimum interference routing of bandwidth guaranteed tunnels with MPLS traffic engineering applications This paper presents new algorithms for dynamic routing of bandwidth guaranteed tunnels, where tunnel routing requests arrive one by one and there is no a priori knowledge regarding future requests. This problem is motivated by the service provider needs for fast deployment of bandwidth guaranteed services. Offline routing algorithms cannot be used since they require a priori knowledge of all tunnel requests that are to be rooted. Instead, on-line algorithms that handle requests arriving one by one and that satisfy as many potential future demands as possible are needed. The newly developed algorithms are on-line algorithms and are based on the idea that a newly routed tunnel must follow a route that does not “interfere too much” with a route that may he critical to satisfy a future demand. We show that this problem is NP-hard. We then develop path selection heuristics which are based on the idea of deferred loading of certain “critical” links. These critical links are identified by the algorithm as links that, if heavily loaded, would make it impossible to satisfy future demands between certain ingress-egress pairs. Like min-hop routing, the presented algorithm uses link-state information and some auxiliary capacity information for path selection. Unlike previous algorithms, the proposed algorithm exploits any available knowledge of the network ingress-egress points of potential future demands, even though the demands themselves are unknown. If all nodes are ingress-egress nodes, the algorithm can still be used, particularly to reduce the rejection rate of requests between a specified subset of important ingress-egress pairs. The algorithm performs well in comparison to previously proposed algorithms on several metrics like the number of rejected demands and successful rerouting of demands upon link failure
The set cover with pairs problem We consider a generalization of the set cover problem, in which elements are covered by pairs of objects, and we are required to find a minimum cost subset of objects that induces a collection of pairs covering all elements. Formally, let U be a ground set of elements and let ${\cal S}$ be a set of objects, where each object i has a non-negative cost wi. For every $\{ i, j \} \subseteq {\cal S}$, let ${\cal C}(i,j)$ be the collection of elements in U covered by the pair { i, j }. The set cover with pairs problem asks to find a subset $A \subseteq {\cal S}$ such that $\bigcup_{ \{ i, j \} \subseteq A } {\cal C}(i,j) = U$ and such that ∑i∈Awi is minimized. In addition to studying this general problem, we are also concerned with developing polynomial time approximation algorithms for interesting special cases. The problems we consider in this framework arise in the context of domination in metric spaces and separation of point sets.
Control Plane Latency With SDN Network Hypervisors: The Cost of Virtualization. Software defined networking (SDN) network hypervisors provide the functionalities needed for virtualizing software-defined networks. Hypervisors sit logically between the multiple virtual SDN networks (vSDNs), which reside on the underlying physical SDN network infrastructure, and the corresponding tenant (vSDN) controllers. Different SDN network hypervisor architectures have mainly been explored through proof-of-concept implementations. We fundamentally advance SDN network hypervisor research by conducting a model-based analysis of SDN hypervisor architectures. Specifically, we introduce mixed integer programming formulations for four different SDN network hypervisor architectures. Our model formulations can also optimize the placement of multi-controller switches in virtualized OpenFlow-enabled SDN networks. We employ our models to quantitatively examine the optimal placement of the hypervisor instances. We compare the control plane latencies of the different SDN hypervisor architectures and quantify the cost of virtualization, i.e., the latency overhead due to virtualizing SDN networks via hypervisors. For generalization, we quantify how the hypervisor architectures behave for different network topologies. Our model formulations and the insights drawn from our evaluations inform network operators about the trade-offs of the different hypervisor architectures and help choosing an architecture according to operator demands.
Flow Setup Latency in SDN Networks. In software-defined networking, the typical switch-controller cycle, from generating a network event notification at the controller until the flow rules are installed at the switches, is not an instantaneous activity. Our measurement results show that this has serious implications on the performance of flow setup procedure, specifically for larger networks: we observe that, even with software swit...
Near-Optimal Disjoint-Path Facility Location Through Set Cover by Pairs AbstractContent Distribution and End-to-End Monitoring with Set Cover by PairsDigital content is housed at data centers located on nodes of a data network. Consumers of this content are also located on network nodes. Content flows from a data center to a consumer on a path defined by a routing protocol, such as open shortest path first. A pair of data centers is said to feasibly serve content to a consumer if there are disjoint paths from each data center to the consumer. In “Near-Optimal Disjoint-Path Facility Location Through Set Cover by Pairs,” Johnson, Breslau, Diakonikolas, Duffield, Gu, Hajiaghayi, Karloff, Resende, and Sen study this problem when the goal is to minimize the number of required data centers to serve a set of consumers. They also study another facility location problem that arises in network traffic monitoring. Both problems are modeled as a set cover-by-pairs problem. The authors provide complexity results, a new lower-bounding integer programming formulation, and several heuristics. The lower bounds are easily computed with a commercial MIP solver and validate the claim of near-optimality of their heuristics.In this paper, we consider two special cases of the “cover-by-pairs” optimization problem that arises when we need to place facilities so that each customer is served by two facilities that reach it by disjoint shortest paths. These problems arise in a network traffic-monitoring scheme proposed by Breslau et al. and have potential applications to content distribution. The “set-disjoint” variant applies to networks that use the open shortest path first routing protocol, and the “path-disjoint” variant applies when multiprotocol label switching routing is enabled, making better solutions possible at the cost of greater operational expense. Although we can prove that no polynomial-time algorithm can guarantee good solutions for either version, we are able to provide heuristics that do very well in practice on instances with real-world network structure. Fast implementations of the heuristics, made possible by exploiting mathematical observations about the relationship between the network instances and the corresponding instances of the cover-by-pairs problem, allow us to perform an extensive experimental evaluation of the heuristics and what the solutions they produce tell us about the effectiveness of the proposed monitoring scheme. For the set-disjoint variant, we validate our claim of near-optimality via a new lower-bounding integer programming formulation. Although computing this lower bound requires solving the NP-hard hitting set problem and can underestimate the optimal value by a linear factor in the worst case, it can be computed quickly by CPLEX, and it equals the optimal solution value for all the instances in our extensive test bed.
Vehicle Routing Problems for Drone Delivery. Unmanned aerial vehicles, or drones, have the potential to significantly reduce the cost and time of making last-mile deliveries and responding to emergencies. Despite this potential, little work has gone into developing vehicle routing problems (VRPs) specifically for drone delivery scenarios. Existing VRPs are insufficient for planning drone deliveries: either multiple trips to the depot are not permitted, leading to solutions with excess drones, or the effect of battery and payload weight on energy consumption is not considered, leading to costly or infeasible routes. We propose two multitrip VRPs for drone delivery that address both issues. One minimizes costs subject to a delivery time limit, while the other minimizes the overall delivery time subject to a budget constraint. We mathematically derive and experimentally validate an energy consumption model for multirotor drones, demonstrating that energy consumption varies approximately linearly with payload and battery weight. We use this approximation to derive mixed integer linear programs for our VRPs. We propose a cost function that considers our energy consumption model and drone reuse, and apply it in a simulated annealing (SA) heuristic for finding suboptimal solutions to practical scenarios. To assist drone delivery practitioners with balancing cost and delivery time, the SA heuristic is used to show that the minimum cost has an inverse exponential relationship with the delivery time limit, and the minimum overall delivery time has an inverse exponential relationship with the budget. Numerical results confirm the importance of reusing drones and optimizing battery size in drone delivery VRPs.
Unsupervised Representation Learning with Deep Convolutional Generative Adversarial Networks In recent years, supervised learning with convolutional networks (CNNs) has seen huge adoption in computer vision applications. Comparatively, unsupervised learning with CNNs has received less attention. In this work we hope to help bridge the gap between the success of CNNs for supervised learning and unsupervised learning. We introduce a class of CNNs called deep convolutional generative adversarial networks (DCGANs), that have certain architectural constraints, and demonstrate that they are a strong candidate for unsupervised learning. Training on various image datasets, we show convincing evidence that our deep convolutional adversarial pair learns a hierarchy of representations from object parts to scenes in both the generator and discriminator. Additionally, we use the learned features for novel tasks - demonstrating their applicability as general image representations.
Wireless sensor networks: a survey This paper describes the concept of sensor networks which has been made viable by the convergence of micro-electro-mechanical systems technology, wireless communications and digital electronics. First, the sensing tasks and the potential sensor networks applications are explored, and a review of factors influencing the design of sensor networks is provided. Then, the communication architecture for sensor networks is outlined, and the algorithms and protocols developed for each layer in the literature are explored. Open research issues for the realization of sensor networks are also discussed.
Traveling Salesman Problems with Profits Traveling salesman problems with profits (TSPs with profits) are a generalization of the traveling salesman problem (TSP), where it is not necessary to visit all vertices. A profit is associated with each vertex. The overall goal is the simultaneous optimization of the collected profit and the travel costs. These two optimization criteria appear either in the objective function or as a constraint. In this paper, a classification of TSPs with profits is proposed, and the existing literature is surveyed. Different classes of applications, modeling approaches, and exact or heuristic solution techniques are identified and compared. Conclusions emphasize the interest of this class of problems, with respect to applications as well as theoretical results.
Linear quadratic bumpless transfer A method for bumpless transfer using ideas from LQ theory is presented and shown to reduce to the Hanus conditioning scheme under certain conditions.
A recent survey of reversible watermarking techniques. The art of secretly hiding and communicating information has gained immense importance in the last two decades due to the advances in generation, storage, and communication technology of digital content. Watermarking is one of the promising solutions for tamper detection and protection of digital content. However, watermarking can cause damage to the sensitive information present in the cover work. Therefore, at the receiving end, the exact recovery of cover work may not be possible. Additionally, there exist certain applications that may not tolerate even small distortions in cover work prior to the downstream processing. In such applications, reversible watermarking instead of conventional watermarking is employed. Reversible watermarking of digital content allows full extraction of the watermark along with the complete restoration of the cover work. For the last few years, reversible watermarking techniques are gaining popularity because of its increasing applications in some important and sensitive areas, i.e., military communication, healthcare, and law-enforcement. Due to the rapid evolution of reversible watermarking techniques, a latest review of recent research in this field is highly desirable. In this survey, the performances of different reversible watermarking schemes are discussed on the basis of various characteristics of watermarking. However, the major focus of this survey is on prediction-error expansion based reversible watermarking techniques, whereby the secret information is hidden in the prediction domain through error expansion. Comparison of the different reversible watermarking techniques is provided in tabular form, and an analysis is carried out. Additionally, experimental comparison of some of the recent reversible watermarking techniques, both in terms of watermarking properties and computational time, is provided on a dataset of 300 images. Future directions are also provided for this potentially important field of watermarking.
Adaptive fuzzy tracking control for switched uncertain strict-feedback nonlinear systems. •Adaptive tracking control for switched strict-feedback nonlinear systems is proposed.•The generalized fuzzy hyperbolic model is used to approximate nonlinear functions.•The designed controller has fewer design parameters comparing with existing methods.
Convert Harm Into Benefit: A Coordination-Learning Based Dynamic Spectrum Anti-Jamming Approach This paper mainly investigates the multi-user anti-jamming spectrum access problem. Using the idea of “converting harm into benefit,” the malicious jamming signals projected by the enemy are utilized by the users as the coordination signals to guide spectrum coordination. An “internal coordination-external confrontation” multi-user anti-jamming access game model is constructed, and the existence of Nash equilibrium (NE) as well as correlated equilibrium (CE) is demonstrated. A coordination-learning based anti-jamming spectrum access algorithm (CLASA) is designed to achieve the CE of the game. Simulation results show the convergence, and effectiveness of the proposed CLASA algorithm, and indicate that our approach can help users confront the malicious jammer, and coordinate internal spectrum access simultaneously without information exchange. Last but not least, the fairness of the proposed approach under different jamming attack patterns is analyzed, which illustrates that this approach provides fair anti-jamming spectrum access opportunities under complicated jamming pattern.
1.103595
0.1
0.1
0.1
0.1
0.1
0.000667
0.000001
0
0
0
0
0
0
Learning-Based Control: A Tutorial And Some Recent Results This monograph presents a new framework for learning-based control synthesis of continuous-time dynamical systems with unknown dynamics. The new design paradigm proposed here is fundamentally different from traditional control theory. In the classical paradigm, controllers are often designed for a given class of dynamical control systems; it is a model-based design. Under the learning-based control framework, controllers are learned online from real-time input-output data collected along the trajectories of the control system in question. An entanglement of techniques from reinforcement learning and model-based control theory is advocated to find a sequence of suboptimal controllers that converge to the optimal solution as learning steps increase. On the one hand, this learning-based design approach attempts to overcome the well-known "curse of dimensionality" and the "curse of modeling" associated with Bellman's Dynamic Programming. On the other hand, rigorous stability and robustness analysis can be derived for the closed-loop system with real-time learning-based controllers. The effectiveness of the proposed learning-based control framework is demonstrated via its applications to theoretical optimal control problems tied to various important classes of continuous-time dynamical systems and practical problems arising from biological motor control, connected and autonomous vehicles.
Inexact Kleinman-Newton Method for Riccati Equations In this paper we consider the numerical solution of the algebraic Riccati equation using Newton's method. We propose an inexact variant which allows one control the number of the inner iterates used in an iterative solver for each Newton step. Conditions are given under which the monotonicity and global convergence result of Kleinman also hold for the inexact Newton iterates. Numerical results illustrate the efficiency of this method.
Adaptive Dynamic Programming for Stochastic Systems With State and Control Dependent Noise. In this technical note, the adaptive optimal control problem is investigated for a class of continuous-time stochastic systems subject to multiplicative noise. A novel non-model-based optimal control design methodology is employed to iteratively update the control policy on-line by using directly the data of the system state and input. Both adaptive dynamic programming (ADP) and robust ADP algorithms are developed, along with rigorous stability and convergence analysis. The effectiveness of the obtained methods is illustrated by an example arising from biological sensorimotor control.
Stability Analysis of Optimal Adaptive Control using Value Iteration with Approximation Errors. Effects of the presence of approximation errors are analyzed on the stability of adaptive optimal control using value iteration, initiated from a stabilizing control policy. This analysis includes the system operated using any single/constant resulting control policy and also using an evolving/time-varying control policy. Sufficient conditions on the `per iteration&#39; approximation errors are obtain...
Policy Iteration Q-Learning for Data-Based Two-Player Zero-Sum Game of Linear Discrete-Time Systems. In this article, the data-based two-player zero-sum game problem is considered for linear discrete-time systems. This problem theoretically depends on solving the discrete-time game algebraic Riccati equation (DTGARE), while it requires complete system dynamics. To avoid solving the DTGARE, the $Q$ -function is introduced and a data-based policy iteration $Q$ -learning (PIQL) algorithm is develo...
Hamiltonian-Driven Adaptive Dynamic Programming for Continuous Nonlinear Dynamical Systems. This paper presents a Hamiltonian-driven framework of adaptive dynamic programming (ADP) for continuous time nonlinear systems, which consists of evaluation of an admissible control, comparison between two different admissible policies with respect to the corresponding the performance function, and the performance improvement of an admissible control. It is showed that the Hamiltonian can serve as...
Complementary Stability and Loop Shaping for Improved Human–Robot Interaction Robots intended for high-force interaction with humans face particular challenges to achieve performance and stability. They require low and tunable endpoint impedance as well as high force capacity, and demand actuators with low intrinsic impedance, the ability to exhibit high impedance (relative to the human subject), and a high ratio of force to weight. Force-feedback control can be used to improve actuator performance, but causes well-known interaction stability problems. This paper presents a novel method to design actuator controllers for physically interactive machines. A loop-shaping design method is developed from a study of fundamental differences between interaction control and the more common servo problem. This approach addresses the interaction problem by redefining stability and performance, using a computational approach to search parameter spaces and displaying variations in performance as control parameters are adjusted. A measure of complementary stability is introduced, and the coupled stability problem is transformed to a robust stability problem using limited knowledge of the environment dynamics (in this case, the human). Design examples show that this new measure improves performance beyond the current best-practice stability constraint (passivity). The controller was implemented on an interactive robot, verifying stability and performance. Testing showed that the new controller out-performed a state-of-the-art controller on the same system
Completely derandomized self-adaptation in evolution strategies. This paper puts forward two useful methods for self-adaptation of the mutation distribution - the concepts of derandomization and cumulation. Principle shortcomings of the concept of mutative strategy parameter control and two levels of derandomization are reviewed. Basic demands on the self-adaptation of arbitrary (normal) mutation distributions are developed. Applying arbitrary, normal mutation distributions is equivalent to applying a general, linear problem encoding. The underlying objective of mutative strategy parameter control is roughly to favor previously selected mutation steps in the future. If this objective is pursued rigorously, a completely derandomized self-adaptation scheme results, which adapts arbitrary normal mutation distributions. This scheme, called covariance matrix adaptation (CMA), meets the previously stated demands. It can still be considerably improved by cumulation - utilizing an evolution path rather than single search steps. Simulations on various test functions reveal local and global search properties of the evolution strategy with and without covariance matrix adaptation. Their performances are comparable only on perfectly scaled functions. On badly scaled, non-separable functions usually a speed up factor of several orders of magnitude is observed. On moderately mis-scaled functions a speed up factor of three to ten can be expected.
Constraint-handling in nature-inspired numerical optimization: Past, present and future. In their original versions, nature-inspired search algorithms such as evolutionary algorithms and those based on swarm intelligence, lack a mechanism to deal with the constraints of a numerical optimization problem. Nowadays, however, there exists a considerable amount of research devoted to design techniques for handling constraints within a nature-inspired algorithm. This paper presents an analysis of the most relevant types of constraint-handling techniques that have been adopted with nature-inspired algorithms. From them, the most popular approaches are analyzed in more detail. For each of them, some representative instantiations are further discussed. In the last part of the paper, some of the future trends in the area, which have been only scarcely explored, are briefly discussed and then the conclusions of this paper are presented.
Simultaneous localization and mapping: part I he simultaneous localization and mapping (SLAM) problem asks if it is possible for a mobile robot to be placed at an unknown location in an unknown envi- ronment and for the robot to incrementally build a consistent map of this environment while simultaneously determining its location within this map. A solution to the SLAM problem has been seen as a "holy grail" for the mobile robotics com- munity as it would provide the means to make a robot truly autonomous. The "solution" of the SLAM problem has been one of the notable successes of the robotics community over the past decade. SLAM has been formulated and solved as a theoretical problem in a number of different forms. SLAM has also been implemented in a number of different domains from indoor robots to outdoor, underwater, and airborne systems. At a theoretical and conceptual level, SLAM can now be consid- ered a solved problem. However, substantial issues remain in practically realizing more general SLAM solutions and notably in building and using perceptually rich maps as part of a SLAM algorithm. This two-part tutorial and survey of SLAM aims to provide a broad introduction to this rapidly growing field. Part I (this article) begins by providing a brief history of early develop- ments in SLAM. The formulation section introduces the struc- ture the SLAM problem in now standard Bayesian form, and explains the evolution of the SLAM process. The solution sec- tion describes the two key computational solutions to the SLAM problem through the use of the extended Kalman filter (EKF-SLAM) and through the use of Rao-Blackwellized par- ticle filters (FastSLAM). Other recent solutions to the SLAM problem are discussed in Part II of this tutorial. The application section describes a number of important real-world implemen- tations of SLAM and also highlights implementations where the sensor data and software are freely down-loadable for other researchers to study. Part II of this tutorial describes major issues in computation, convergence, and data association in SLAM. These are subjects that have been the main focus of the SLAM research community over the past five years.
Stability analysis of switched systems using variational principles: An introduction Many natural and artificial systems and processes encompass several modes of operation with a different dynamical behavior in each mode. Switched systems provide a suitable mathematical model for such processes, and their stability analysis is important for both theoretical and practical reasons. We review a specific approach for stability analysis based on using variational principles to characterize the ''most unstable'' solution of the switched system. We also discuss a link between the variational approach and the stability analysis of switched systems using Lie-algebraic considerations. Both approaches require the use of sophisticated tools from many different fields of applied mathematics. The purpose of this paper is to provide an accessible and self-contained review of these topics, emphasizing the intuitive and geometric underlying ideas.
Robust and Imperceptible Dual Watermarking for Telemedicine Applications In this paper, the effects of different error correction codes on the robustness and imperceptibility of discrete wavelet transform and singular value decomposition based dual watermarking scheme is investigated. Text and image watermarks are embedded into cover radiological image for their potential application in secure and compact medical data transmission. Four different error correcting codes such as Hamming, the Bose, Ray-Chaudhuri, Hocquenghem (BCH), the Reed---Solomon and hybrid error correcting (BCH and repetition code) codes are considered for encoding of text watermark in order to achieve additional robustness for sensitive text data such as patient identification code. Performance of the proposed algorithm is evaluated against number of signal processing attacks by varying the strength of watermarking and covers image modalities. The experimental results demonstrate that this algorithm provides better robustness without affecting the quality of watermarked image.This algorithm combines the advantages and removes the disadvantages of the two transform techniques. Out of the three error correcting codes tested, it has been found that Reed---Solomon shows the best performance. Further, a hybrid model of two of the error correcting codes (BCH and repetition code) is concatenated and implemented. It is found that the hybrid code achieves better results in terms of robustness. This paper provides a detailed analysis of the obtained experimental results.
Automated Detection of Obstructive Sleep Apnea Events from a Single-Lead Electrocardiogram Using a Convolutional Neural Network. In this study, we propose a method for the automated detection of obstructive sleep apnea (OSA) from a single-lead electrocardiogram (ECG) using a convolutional neural network (CNN). A CNN model was designed with six optimized convolution layers including activation, pooling, and dropout layers. One-dimensional (1D) convolution, rectified linear units (ReLU), and max pooling were applied to the convolution, activation, and pooling layers, respectively. For training and evaluation of the CNN model, a single-lead ECG dataset was collected from 82 subjects with OSA and was divided into training (including data from 63 patients with 34,281 events) and testing (including data from 19 patients with 8571 events) datasets. Using this CNN model, a precision of 0.99%, a recall of 0.99%, and an F-score of 0.99% were attained with the training dataset; these values were all 0.96% when the CNN was applied to the testing dataset. These results show that the proposed CNN model can be used to detect OSA accurately on the basis of a single-lead ECG. Ultimately, this CNN model may be used as a screening tool for those suspected to suffer from OSA.
Robot tutor and pupils’ educational ability: Teaching the times tables Research shows promising results of educational robots in language and STEM tasks. In language, more research is available, occasionally in view of individual differences in pupils’ educational ability levels, and learning seems to improve with more expressive robot behaviors. In STEM, variations in robots’ behaviors have been examined with inconclusive results and never while systematically investigating how differences in educational abilities match with different robot behaviors. We applied an autonomously tutoring robot (without tablet, partly WOz) in a 2 × 2 experiment of social vs. neutral behavior in above-average vs. below-average schoolchildren (N = 86; age 8–10 years) while rehearsing the multiplication tables on a one-to-one basis. The standard school test showed that on average, pupils significantly improved their performance even after 3 occasions of 5-min exercises. Beyond-average pupils profited most from a robot tutor, whereas those below average in multiplication benefited more from a robot that showed neutral rather than more social behavior.
1.24
0.24
0.24
0.24
0.24
0.08
0.013333
0
0
0
0
0
0
0
Urban Traffic Prediction from Spatio-Temporal Data Using Deep Meta Learning Predicting urban traffic is of great importance to intelligent transportation systems and public safety, yet is very challenging because of two aspects: 1) complex spatio-temporal correlations of urban traffic, including spatial correlations between locations along with temporal correlations among timestamps; 2) diversity of such spatio-temporal correlations, which vary from location to location and depend on the surrounding geographical information, e.g., points of interests and road networks. To tackle these challenges, we proposed a deep-meta-learning based model, entitled ST-MetaNet, to collectively predict traffic in all location at once. ST-MetaNet employs a sequence-to-sequence architecture, consisting of an encoder to learn historical information and a decoder to make predictions step by step. In specific, the encoder and decoder have the same network structure, consisting of a recurrent neural network to encode the traffic, a meta graph attention network to capture diverse spatial correlations, and a meta recurrent neural network to consider diverse temporal correlations. Extensive experiments were conducted based on two real-world datasets to illustrate the effectiveness of ST-MetaNet beyond several state-of-the-art methods.
Network-wide Crowd Flow Prediction of Sydney Trains via Customized Online Non-negative Matrix Factorization. Crowd Flow Prediction (CFP) is one major challenge in the intelligent transportation systems of the Sydney Trains Network. However, most advanced CFP methods only focus on entrance and exit flows at the major stations or a few subway lines, neglecting Crowd Flow Distribution (CFD) forecasting problem across the entire city network. CFD prediction plays an irreplaceable role in metro management as a tool that can help authorities plan route schedules and avoid congestion. In this paper, we propose three online non-negative matrix factorization (ONMF) models. ONMF-AO incorporates an Average Optimization strategy that adapts to stable passenger flows. ONMF-MR captures the Most Recent trends to achieve better performance when sudden changes in crowd flow occur. The Hybrid model, ONMF-H, integrates both ONMF-AO and ONMF-MR to exploit the strengths of each model in different scenarios and enhance the models' applicability to real-world situations. Given a series of CFD snapshots, both models learn the latent attributes of the train stations and, therefore, are able to capture transition patterns from one timestamp to the next by combining historic guidance. Intensive experiments on a large-scale, real-world dataset containing transactional data demonstrate the superiority of our ONMF models.
Origin-Destination Matrix Prediction via Graph Convolution: a New Perspective of Passenger Demand Modeling Ride-hailing applications are becoming more and more popular for providing drivers and passengers with convenient ride services, especially in metropolises like Beijing or New York. To obtain the passengers' mobility patterns, the online platforms of ride services need to predict the number of passenger demands from one region to another in advance. We formulate this problem as an Origin-Destination Matrix Prediction (ODMP) problem. Though this problem is essential to large-scale providers of ride services for helping them make decisions and some providers have already put it forward in public, existing studies have not solved this problem well. One of the main reasons is that the ODMP problem is more challenging than the common demand prediction. Besides the number of demands in a region, it also requires the model to predict the destinations of them. In addition, data sparsity is a severe issue. To solve the problem effectively, we propose a unified model, Grid-Embedding based Multi-task Learning (GEML) which consists of two components focusing on spatial and temporal information respectively. The Grid-Embedding part is designed to model the spatial mobility patterns of passengers and neighboring relationships of different areas, the pre-weighted aggregator of which aims to sense the sparsity and range of data. The Multi-task Learning framework focuses on modeling temporal attributes and capturing several objectives of the ODMP problem. The evaluation of our model is conducted on real operational datasets from UCAR and Didi. The experimental results demonstrate the superiority of our GEML against the state-of-the-art approaches.
Forecasting holiday daily tourist flow based on seasonal support vector regression with adaptive genetic algorithm. •The model of support vector regression with adaptive genetic algorithm and the seasonal mechanism is proposed.•Parameters selection and seasonal adjustment should be carefully selected.•We focus on latest and representative holiday daily data in China.•Two experiments are used to prove the effect of the model.•The AGASSVR is superior to AGA-SVR and BPNN.
Streaming Adaptation of Deep Forecasting Models using Adaptive Recurrent Units. We present ARU, an Adaptive Recurrent Unit for streaming adaptation of deep globally trained time-series forecasting models. The ARU combines the advantages of learning complex data transformations across multiple time series from deep global models, with per-series localization offered by closed-form linear models. Unlike existing methods of adaptation that are either memory-intensive or non-responsive after training, ARUs require only fixed sized state and adapt to streaming data via an easy RNN-like update operation. The core principle driving ARU is simple --- maintain sufficient statistics of conditional Gaussian distributions and use them to compute local parameters in closed form. Our contribution is in embedding such local linear models in globally trained deep models while allowing end-to-end training on the one hand, and easy RNN-like updates on the other. Across several datasets we show that ARU is more effective than recently proposed local adaptation methods that tax the global network to compute local parameters.
Vehicle Speed Prediction Using a Markov Chain With Speed Constraints Modern vehicles are designed to improve fuel consumption while satisfying emissions regulations. As a result, powertrains are becoming increasingly complex and changing rapidly. Optimal control based on the future vehicle speed is one way to address these changes. In this approach, accurate prediction of velocity is closely related to the performance of optimal control. However, there exists uncer...
A Survey of the Usages of Deep Learning for Natural Language Processing Over the last several years, the field of natural language processing has been propelled forward by an explosion in the use of deep learning models. This article provides a brief introduction to the field and a quick overview of deep learning architectures and methods. It then sifts through the plethora of recent studies and summarizes a large assortment of relevant contributions. Analyzed research areas include several core linguistic processing issues in addition to many applications of computational linguistics. A discussion of the current state of the art is then provided along with recommendations for future research in the field.
Network‐wide traffic speed forecasting: 3D convolutional neural network with ensemble empirical mode decomposition AbstractAbstractAccurate traffic speed forecasting is one of the most critical tasks in proactive traffic management and the deployment of advanced traveler information systems. This paper proposes a hybrid forecasting approach named DeepEnsemble by integrating the three‐dimensional convolutional neural network (3D CNN) with ensemble empirical mode decomposition (EEMD). There are four steps in this hybrid approach. First, EEMD is adopted to decompose the complex traffic speed time series data with noise into several intrinsic mode functions (IMFs) and a residue. Second, a three‐dimensional tensor is established and fed into 3D CNN for prediction. Third, the output of 3D CNN prediction is obtained by a linear combination of the results of all components. Finally, the 3D CNN prediction output, external features, and historical features are fused to predict the network‐wide traffic speed simultaneously. The proposed DeepEnsemble approach is tested on the three‐month traffic speed series data of a real‐world large‐scale urban expressway network with 308 traffic flow detectors in Beijing, China. The experimental results indicate that DeepEnsemble outperforms the state‐of‐the‐art network‐wide traffic speed forecasting models. 3D CNN learns temporal, spatial, and depth information better than 2D CNN. Moreover, forecasting accuracy can be improved by employing EEMD. DeepEnsemble is a promising model with scalability and portability for network‐wide traffic speed prediction and can be further extended to conduct traffic status monitoring and congestion mitigation strategies.
Predicting Taxi–Passenger Demand Using Streaming Data Informed driving is increasingly becoming a key feature for increasing the sustainability of taxi companies. The sensors that are installed in each vehicle are providing new opportunities for automatically discovering knowledge, which, in return, delivers information for real-time decision making. Intelligent transportation systems for taxi dispatching and for finding time-saving routes are already exploring these sensing data. This paper introduces a novel methodology for predicting the spatial distribution of taxi–passengers for a short-term time horizon using streaming data. First, the information was aggregated into a histogram time series. Then, three time-series forecasting techniques were combined to originate a prediction. Experimental tests were conducted using the online data that are transmitted by 441 vehicles of a fleet running in the city of Porto, Portugal. The results demonstrated that the proposed framework can provide effective insight into the spatiotemporal distribution of taxi–passenger demand for a 30-min horizon.
Multi-Armed Bandit-Based Client Scheduling for Federated Learning By exploiting the computing power and local data of distributed clients, federated learning (FL) features ubiquitous properties such as reduction of communication overhead and preserving data privacy. In each communication round of FL, the clients update local models based on their own data and upload their local updates via wireless channels. However, latency caused by hundreds to thousands of communication rounds remains a bottleneck in FL. To minimize the training latency, this work provides a multi-armed bandit-based framework for online client scheduling (CS) in FL without knowing wireless channel state information and statistical characteristics of clients. Firstly, we propose a CS algorithm based on the upper confidence bound policy (CS-UCB) for ideal scenarios where local datasets of clients are independent and identically distributed (i.i.d.) and balanced. An upper bound of the expected performance regret of the proposed CS-UCB algorithm is provided, which indicates that the regret grows logarithmically over communication rounds. Then, to address non-ideal scenarios with non-i.i.d. and unbalanced properties of local datasets and varying availability of clients, we further propose a CS algorithm based on the UCB policy and virtual queue technique (CS-UCB-Q). An upper bound is also derived, which shows that the expected performance regret of the proposed CS-UCB-Q algorithm can have a sub-linear growth over communication rounds under certain conditions. Besides, the convergence performance of FL training is also analyzed. Finally, simulation results validate the efficiency of the proposed algorithms.
Output-feedback stochastic nonlinear stabilization The authors present the first result on global output-feedback stabilization (in probability) for stochastic nonlinear continuous-time systems. The class of systems that they consider is a stochastic counterpart of the broadest class of deterministic systems for which globally stabilizing controllers are currently available. Their controllers are “inverse optimal” and possess an infinite gain margin. A reader of the paper needs no prior familiarity with techniques of stochastic control
Generalized dilations and numerically solving discrete-time homogeneous optimization problems We introduce generalized dilations, a broader class of operators than that of dilations, and consider homogeneity with respect to this new class of dilations. For discrete-time systems that are asymptotically controllable and homogeneous (with degree zero) we propose a method to numerically approximate any homogeneous value function (solution to an infinite horizon optimization problem) to arbitrary accuracy. We also show that the method can be used to generate an offline computed stabilizing feedback law.
Squeezed Convolutional Variational AutoEncoder for unsupervised anomaly detection in edge device industrial Internet of Things In this paper, we propose Squeezed Convolutional Variational AutoEncoder (SCVAE) for anomaly detection in time series data for Edge Computing in Industrial Internet of Things (IIoT). The proposed model is applied to labeled time series data from UCI datasets for exact performance evaluation, and applied to real world data for indirect model performance comparison. In addition, by comparing the models before and after applying Fire Modules from SqueezeNet, we show that model size and inference times are reduced while similar levels of performance is maintained.
Robot tutor and pupils’ educational ability: Teaching the times tables Research shows promising results of educational robots in language and STEM tasks. In language, more research is available, occasionally in view of individual differences in pupils’ educational ability levels, and learning seems to improve with more expressive robot behaviors. In STEM, variations in robots’ behaviors have been examined with inconclusive results and never while systematically investigating how differences in educational abilities match with different robot behaviors. We applied an autonomously tutoring robot (without tablet, partly WOz) in a 2 × 2 experiment of social vs. neutral behavior in above-average vs. below-average schoolchildren (N = 86; age 8–10 years) while rehearsing the multiplication tables on a one-to-one basis. The standard school test showed that on average, pupils significantly improved their performance even after 3 occasions of 5-min exercises. Beyond-average pupils profited most from a robot tutor, whereas those below average in multiplication benefited more from a robot that showed neutral rather than more social behavior.
1.033918
0.037857
0.037857
0.035714
0.035714
0.035714
0.028571
0.028571
0.004827
0
0
0
0
0
Decentralized data access control over consortium blockchains Blockchain is an emerging data management technology that enables people in a collaborative network to establish trusted connections with the other participants. Recently consortium blockchains have raised interest in a broader blockchain technology discussion. Instead of a fully public, autonomous network, consortium blockchain supports a network where participants can be limited to a subset of users and data access strictly controlled. Access control policies should be defined by the respective data owner and applied throughout the network without requiring a centralized data administrator. As a result, decentralized data access control (DDAC) emerges as a fundamental challenge for such systems. However, we show from a trust model for consortium collaborative networks that current consortium blockchain systems provide limited support for DDAC. Further, the distributed, replicated nature of blockchain makes it even more challenging to control data access, especially read access, compared with traditional DBMSes. We investigate possible strategies to protect data from being read by unauthorized users in consortium blockchain systems using combinations of ledger partitioning and encryption strategies. A general framework is proposed to help inexperienced users determine appropriate strategies under different application scenarios. The framework was implemented on top of Hyperledger Fabric to evaluate feasibility. Experimental results along with a real-world case study contrasted the performance of different strategies under various conditions and the practicality of the proposed framework.
GMDH-based networks for intelligent intrusion detection. Network intrusion detection has been an area of rapid advancement in recent times. Similar advances in the field of intelligent computing have led to the introduction of several classification techniques for accurately identifying and differentiating network traffic into normal and anomalous. Group Method for Data Handling (GMDH) is one such supervised inductive learning approach for the synthesis of neural network models. Through this paper, we propose a GMDH-based technique for classifying network traffic into normal and anomalous. Two variants of the technique, namely, Monolithic and Ensemble-based, were tested on the KDD-99 dataset. The dataset was preprocessed and all features were ranked based on three feature ranking techniques, namely, Information Gain, Gain Ratio, and GMDH by itself. The results obtained proved that the proposed intrusion detection scheme yields high attack detection rates, nearly 98%, when compared with other intelligent classification techniques for network intrusion detection.
Mining network data for intrusion detection through combining SVMs with ant colony networks. In this paper, we introduce a new machine-learning-based data classification algorithm that is applied to network intrusion detection. The basic task is to classify network activities (in the network log as connection records) as normal or abnormal while minimizing misclassification. Although different classification models have been developed for network intrusion detection, each of them has its strengths and weaknesses, including the most commonly applied Support Vector Machine (SVM) method and the Clustering based on Self-Organized Ant Colony Network (CSOACN). Our new approach combines the SVM method with CSOACNs to take the advantages of both while avoiding their weaknesses. Our algorithm is implemented and evaluated using a standard benchmark KDD99 data set. Experiments show that CSVAC (Combining Support Vectors with Ant Colony) outperforms SVM alone or CSOACN alone in terms of both classification rate and run-time efficiency.
A two-level hybrid approach for intrusion detection. To exploit the strengths of misuse detection and anomaly detection, an intensive focus on intrusion detection combines the two. From a novel perspective, in this paper, we proposed a hybrid approach toward achieving a high detection rate with a low false positive rate. The approach is a two-level hybrid solution consisting of two anomaly detection components and a misuse detection component. In stage 1, an anomaly detection method with low computing complexity is developed and employed to build the detection component. The k-nearest neighbors algorithm becomes crucial in building the two detection components for stage 2. In this hybrid approach, all of the detection components are well-coordinated. The detection component of stage 1 becomes involved in the course of building the two detection components of stage 2 that reduce the false positives and false negatives generated by the detection component of stage 1. Experimental results on the KDD'99 dataset and the Kyoto University Benchmark dataset confirm that the proposed hybrid approach can effectively detect network anomalies with a low false positive rate. HighlightsA novel two-level hybrid intrusion detection approach is proposed.A novel anomaly detection method based on change of cluster centres is proposed.Detection components in the two stages of the hybrid approach work well together.Experimental results show that our approach performs well in false positive rate.
A Blockchain Based Truthful Incentive Mechanism for Distributed P2P Applications. In distributed peer-to-peer (P2P) applications, peers self-organize and cooperate to effectively complete certain tasks such as forwarding files, delivering messages, or uploading data. Nevertheless, users are selfish in nature and they may refuse to cooperate due to their concerns on energy and bandwidth consumption. Thus each user should receive a satisfying reward to compensate its resource consumption for cooperation. However, suitable incentive mechanisms that can meet the diverse requirements of users in dynamic and distributed P2P environments are still missing. On the other hand, we observe that Blockchain is a decentralized secure digital ledger of economic transactions that can be programmed to record not just financial transactions and Blockchain-based cryptocurrencies get more and more market capitalization. Therefore in this paper, we propose a Blockchain based truthful incentive mechanism for distributed P2P applications that applies a cryptocurrency such as Bitcoin to incentivize users for cooperation. In this mechanism, users who help with a successful delivery get rewarded. As users and miners in the Blockchain P2P system may exhibit selfish actions or collude with each other, we propose a secure validation method and a pricing strategy, and integrate them into our incentive mechanism. Through a game theoretical analysis and evaluation study, we demonstrate the effectiveness and security strength of our proposed incentive mechanism.
A weighted voting ensemble of efficient regularized extreme learning machine The exact evaluation of Extreme Learning Machine (ELM) compactness is difficult due to the randomness in hidden layer nodes number, weight and bias values. To overcome this randomness, and other problems such as resultant overfitting and large variance, a selective weighted voting ensemble model based on regularized ELM is investigated. It can strongly enhance the overall performance including accuracy, variance and time consumption. Efficient Prediction Sum of Squares (PRESS) criteria that utilizing Singular Value Decomposition (SVD) is proposed to address the slow execution. Furthermore, an ensemble pruning approach based on the eigenvalues for the input weight matrix is developed. In this work, the ensemble base classifiers weights are calculated based on the same PRESS error metric used for the solutions of the output weight vector (β) in RELM, thus, it can reduce computational cost and space requirement. Different state-of-the-art learning approaches and various well-known facial expressions faces and object recognition benchmark datasets were examined in this work.
Computational thinking Summary form only given. My vision for the 21st century, Computational Thinking, will be a fundamental skill used by everyone in the world. To reading, writing, and arithmetic, we should add computational thinking to every child's analytical ability. Computational thinking involves solving problems, designing systems, and understanding human behavior by drawing on the concepts fundamental to computer science. Thinking like a computer scientist means more than being able to program a computer. It requires the ability to abstract and thus to think at multiple levels of abstraction. In this talk I will give many examples of computational thinking, argue that it has already influenced other disciplines, and promote the idea that teaching computational thinking can not only inspire future generations to enter the field of computer science but benefit people in all fields.
Adam: A Method for Stochastic Optimization. We introduce Adam, an algorithm for first-order gradient-based optimization of stochastic objective functions, based on adaptive estimates of lower-order moments. The method is straightforward to implement, is computationally efficient, has little memory requirements, is invariant to diagonal rescaling of the gradients, and is well suited for problems that are large in terms of data and/or parameters. The method is also appropriate for non-stationary objectives and problems with very noisy and/or sparse gradients. The hyper-parameters have intuitive interpretations and typically require little tuning. Some connections to related algorithms, on which Adam was inspired, are discussed. We also analyze the theoretical convergence properties of the algorithm and provide a regret bound on the convergence rate that is comparable to the best known results under the online convex optimization framework. Empirical results demonstrate that Adam works well in practice and compares favorably to other stochastic optimization methods. Finally, we discuss AdaMax, a variant of Adam based on the infinity norm.
Blockchain Meets IoT: An Architecture for Scalable Access Management in IoT. The Internet of Things (IoT) is stepping out of its infancy into full maturity and establishing itself as a part of the future Internet. One of the technical challenges of having billions of devices deployed worldwide is the ability to manage them. Although access management technologies exist in IoT, they are based on centralized models which introduce a new variety of technical limitations to ma...
Multivariate Short-Term Traffic Flow Forecasting Using Time-Series Analysis Existing time-series models that are used for short-term traffic condition forecasting are mostly univariate in nature. Generally, the extension of existing univariate time-series models to a multivariate regime involves huge computational complexities. A different class of time-series models called structural time-series model (STM) (in its multivariate form) has been introduced in this paper to develop a parsimonious and computationally simple multivariate short-term traffic condition forecasting algorithm. The different components of a time-series data set such as trend, seasonal, cyclical, and calendar variations can separately be modeled in STM methodology. A case study at the Dublin, Ireland, city center with serious traffic congestion is performed to illustrate the forecasting strategy. The results indicate that the proposed forecasting algorithm is an effective approach in predicting real-time traffic flow at multiple junctions within an urban transport network.
State resetting for bumpless switching in supervisory control In this paper the realization and implementation of a multi-controller scheme made of a finite set of linear single-input-single-output controllers, possibly having different state dimensions, is studied. The supervisory control framework is considered, namely a minimal parameter dependent realization of the set of controllers such that all controllers share the same state space is used. A specific state resetting strategy based on the behavioral approach to system theory is developed in order to master the transient upon controller switching.
Adaptive dynamic programming and optimal control of nonlinear nonaffine systems. In this paper, a novel optimal control design scheme is proposed for continuous-time nonaffine nonlinear dynamic systems with unknown dynamics by adaptive dynamic programming (ADP). The proposed methodology iteratively updates the control policy online by using the state and input information without identifying the system dynamics. An ADP algorithm is developed, and can be applied to a general class of nonlinear control design problems. The convergence analysis for the designed control scheme is presented, along with rigorous stability analysis for the closed-loop system. The effectiveness of this new algorithm is illustrated by two simulation examples.
Finite-Time Adaptive Fuzzy Tracking Control Design for Nonlinear Systems. This paper addresses the finite-time tracking problem of nonlinear pure-feedback systems. Unlike the literature on traditional finite-time stabilization, in this paper the nonlinear system functions, including the bounding functions, are all totally unknown. Fuzzy logic systems are used to model those unknown functions. To present a finite-time control strategy, a criterion of semiglobal practical...
Energy harvesting algorithm considering max flow problem in wireless sensor networks. In Wireless Sensor Networks (WSNs), sensor nodes with poor energy always have bad effect on the data rate or max flow. These nodes are called bottleneck nodes. In this paper, in order to increase the max flow, we assume an energy harvesting WSNs environment to investigate the cooperation of multiple Mobile Chargers (MCs). MCs are mobile robots that use wireless charging technology to charge sensor nodes in WSNs. This means that in energy harvesting WSNs environments, sensor nodes can obtain energy replenishment by using MCs or collecting energy from nature by themselves. In our research, we use MCs to improve the energy of the sensor nodes by performing multiple rounds of unified scheduling, and finally achieve the purpose of increasing the max flow at sinks. Firstly, we model this problem as a Linear Programming (LP) to search the max flow in a round of charging scheduling and prove that the problem is NP-hard. In order to solve the problem, we propose a heuristic approach: deploying MCs in units of paths with the lowest energy node priority. To reduce the energy consumption of MCs and increase the charging efficiency, we also take the optimization of MCs’ moving distance into our consideration. Finally, we extend the method to multiple rounds of scheduling called BottleNeck. Simulation results show that Bottleneck performs well at increasing max flow.
1.2
0.2
0.2
0.2
0.2
0.2
0
0
0
0
0
0
0
0
Faster R-CNN: Towards Real-Time Object Detection with Region Proposal Networks State-of-the-art object detection networks depend on region proposal algorithms to hypothesize object locations. Advances like SPPnet [1] and Fast R-CNN [2] have reduced the running time of these detection networks, exposing region proposal computation as a bottleneck. In this work, we introduce a Region Proposal Network (RPN) that shares full-image convolutional features with the detection network, thus enabling nearly cost-free region proposals. An RPN is a fully convolutional network that simultaneously predicts object bounds and objectness scores at each position. The RPN is trained end-to-end to generate high-quality region proposals, which are used by Fast R-CNN for detection. We further merge RPN and Fast R-CNN into a single network by sharing their convolutional features—using the recently popular terminology of neural networks with ’attention’ mechanisms, the RPN component tells the unified network where to look. For the very deep VGG-16 model [3] , our detection system has a frame rate of 5 fps (including all steps) on a GPU, while achieving state-of-the-art object detection accuracy on PASCAL VOC 2007, 2012, and MS COCO datasets with only 300 proposals per image. In ILSVRC and COCO 2015 competitions, Faster R-CNN and RPN are the foundations of the 1st-place winning entries in several tracks. Code has been made publicly available.
Auto-Alert: A Spatial and Temporal Architecture for Driving Assistance in Road Traffic Environments Over the last decade, the Advanced Driver Assistance System (ADAS) concept has evolved prominently. ADAS involves several advanced approaches such as automotive electronics, vehicular communication, RADAR, LIDAR, computer vision, and its associated aspects such as machine learning and deep learning. Of these, computer vision and machine learning-based solutions have mainly been effective that have allowed real-time vehicle control, driver-aided systems, etc. However, most of the existing works deal with ADAS deployment and autonomous driving functionality in countries with well-disciplined lane traffic. These solutions and frameworks do not work in countries and cities with less-disciplined/ chaotic traffic. Hence, critical ADAS functionalities and even L2/ L3 autonomy levels in driving remain a major open challenge. In this regard, this work proposes a novel framework called Auto-Alert. Auto-Alert performs a two-stage spatial and temporal analysis based on external traffic environment and tri-axial sensor system for safe driving assistance. This work investigates time-series analysis with deep learning models for driving events prediction and assistance. Further, as a basic premise, various essential design considerations towards the ADAS are discussed. Significantly, the Convolutional Neural Network (CNN) and Long-Short-Term-Memory (LSTM) models are applied in the proposed Auto-Alert. It is shown that the LSTM outperforms the CNN with 99% for the considered window length. Importantly, this also involves developing and demonstrating an efficient traffic monitoring and density estimation system. Further, this work provides the benchmark results for Indian Driving Dataset (IDD), specifically for the object detection task. The findings of this proposed work demonstrate the significance of using CNN and LSTM networks to assist the driver in the holistic traffic environment.
A Virtual-Real Interaction Approach to Object Instance Segmentation in Traffic Scenes Object instance segmentation in traffic scenes is an important research topic. For training instance segmentation models, synthetic data can potentially complement real data, alleviating manual effort on annotating real images. However, the data distribution discrepancy between synthetic data and real data hampers the wide applications of synthetic data. In light of that, we propose a virtual-real interaction method for object instance segmentation. This method works over synthetic images with accurate annotations and real images without any labels. The virtual-real interaction guides the model to learn useful information from synthetic data while keeping consistent with real data. We first analyze the data distribution discrepancy from a probabilistic perspective, and divide it into image-level and instance-level discrepancies. Then, we design two components to align these discrepancies, i.e., global-level alignment and local-level alignment. Furthermore, a consistency alignment component is proposed to encourage the consistency between the global-level and the local-level alignment components. We evaluate the proposed approach on the real Cityscapes dataset by adapting from virtual SYNTHIA, Virtual KITTI, and VIPER datasets. The experimental results demonstrate that it achieves significantly better performance than state-of-the-art methods.
Soft-Weighted-Average Ensemble Vehicle Detection Method Based on Single-Stage and Two-Stage Deep Learning Models The deep learning object detection algorithms have become one of the powerful tools for road vehicle detection in autonomous driving. However, the limitation of the number of high-quality labeled training samples makes the single-object detection algorithms unable to achieve satisfactory accuracy in road vehicle detection. In this paper, by comparing the pros and cons of various object detection algorithms, two different algorithms with a different emphasis are selected for a weighted ensemble. Besides, a new ensemble method named the Soft-Weighted-Average method is proposed. The proposed method is attenuated by the confidence, and it “punishes” the detection result of the corresponding relationship by the confidence attenuation, instead of by deleting the output of a certain model. The proposed method can further reduce the vehicle misdetection of the target detection algorithm, obtaining a better detection result. Lastly, the ensemble method can achieve an average accuracy of 94.75% for simple targets, which makes it the third-ranked method in the KITTI evaluation system.
Deep Multi-Modal Object Detection and Semantic Segmentation for Autonomous Driving: Datasets, Methods, and Challenges AbstractRecent advancements in perception for autonomous driving are driven by deep learning. In order to achieve robust and accurate scene understanding, autonomous vehicles are usually equipped with different sensors (e.g. cameras, LiDARs, Radars), and multiple sensing modalities can be fused to exploit their complementary properties. In this context, many methods have been proposed for deep multi-modal perception problems. However, there is no general guideline for network architecture design, and questions of “what to fuse”, “when to fuse”, and “how to fuse” remain open. This review paper attempts to systematically summarize methodologies and discuss challenges for deep multi-modal object detection and semantic segmentation in autonomous driving. To this end, we first provide an overview of on-board sensors on test vehicles, open datasets, and background information for object detection and semantic segmentation in autonomous driving research. We then summarize the fusion methodologies and discuss challenges and open questions. In the appendix, we provide tables that summarize topics and methods. We also provide an interactive online platform to navigate each reference: https://boschresearch.github.io/multimodalperception/.
Parallel control for optimal tracking via adaptive dynamic programming This paper studies the problem of optimal parallel tracking control for continuous-time general nonlinear systems. Unlike existing optimal state feedback control, the control input of the optimal parallel control is introduced into the feedback system. However, due to the introduction of control input into the feedback system, the optimal state feedback control methods can not be applied directly. To address this problem, an augmented system and an augmented performance index function are proposed firstly. Thus, the general nonlinear system is transformed into an affine nonlinear system. The difference between the optimal parallel control and the optimal state feedback control is analyzed theoretically. It is proven that the optimal parallel control with the augmented performance index function can be seen as the suboptimal state feedback control with the traditional performance index function. Moreover, an adaptive dynamic programming (ADP) technique is utilized to implement the optimal parallel tracking control using a critic neural network (NN) to approximate the value function online. The stability analysis of the closed-loop system is performed using the Lyapunov theory, and the tracking error and NN weights errors are uniformly ultimately bounded (UUB). Also, the optimal parallel controller guarantees the continuity of the control input under the circumstance that there are finite jump discontinuities in the reference signals. Finally, the effectiveness of the developed optimal parallel control method is verified in two cases.
Night-to-Day: Online Image-to-Image Translation for Object Detection Within Autonomous Driving by Night Object detectors are central to autonomous driving and are widely used in driver assistance systems. Object detectors are trained on a finite amount of data within a specific domain, hampering detection performance when applying object detectors to samples from other domains during inference, an effect known as domain gap. Domain gap is a concern for data-driven applications, evoking repetitive retraining of networks when the applications unfold into other domains. With object detectors that have been trained on day images only, a domain gap can be observed in object detection by night. Training object detectors on night images is critical because of the enormous effort required to generate an adequate amount of diversely labeled data, and existing data sets often tend to overfit specific domain characteristics. For the first time, this work proposes adapting domains by online image-to-image translation to expand an object detector's domain of operation. The domain gap is decreased without additional labeling effort and without having to retrain the object detector while unfolding into the target domain. The approach follows the concept of domain adaptation, shifting the target domain samples into the domain knownto the object detector (source domain). Firstly, the UNIT network is trained for domain adaptation and subsequently cast into an online domain adaptation module, which narrows down the domain gap. Domain adaptation capabilities are evaluated qualitatively by displaying translated samples and visualizing the domain shift through the 2D tSNE algorithm. We quantitatively benchmark the domain adaptation's influence on a state-of-the-art object detector, and on a retrained object detector, for mean average precision, mean recall, and the resulting F1-score. Our approach achieves an F1 score improvement of 5.27 % within object detection by night when applying online domain adaptation. The evaluation is executed on the BDD100K benchmark data set.
Reduction of Uncertainties for Safety Assessment of Automated Driving Under Parallel Simulations Many achievements concerning developments in the field of automated driving have been made. However, automated driving still faces the challenge of safety validation. Conventional methods are not suitable any more for this highly complex automation system. Therefore, the method named Virtual Assessment of Automation in Field Operation (VAAFO) is motivated. In this approach, automated driving system has no access to actuators but rather runs parallel to the human driver. Consequently, this approach is divided into two modules: online trajectory comparison and offline safety assessment. This paper focuses on the second module, in which uncertainties in world model are reduced and then the safety of Automated Vehicle (AV) is evaluated. Retrospective post-processing combined with Joint Integrated Probabilistic Data Association (JIPDA) tracker is put forward to reduce existence uncertainties. State uncertainties are reduced by an Unscented Rauch-Tung-Striebel smoother (URTSS). Furthermore, inverse TTC and remaining lateral distance are utilized to assess the safety of AV in the corrected world model. The results demonstrate that retrospective post-processing combined with JIPDA can reduce existence uncertainties greatly. URTSS is very useful for reducing state uncertainties. The studied case illustrates that the safety of AV can be assessed by parallel running and critical scenarios are found accordingly.
Visual Genome: Connecting Language and Vision Using Crowdsourced Dense Image Annotations. Despite progress in perceptual tasks such as image classification, computers still perform poorly on cognitive tasks such as image description and question answering. Cognition is core to tasks that involve not just recognizing, but reasoning about our visual world. However, models used to tackle the rich content in images for cognitive tasks are still being trained using the same datasets designed for perceptual tasks. To achieve success at cognitive tasks, models need to understand the interactions and relationships between objects in an image. When asked \"What vehicle is the person riding?\", computers will need to identify the objects in an image as well as the relationships riding(man, carriage) and pulling(horse, carriage) to answer correctly that \"the person is riding a horse-drawn carriage.\" In this paper, we present the Visual Genome dataset to enable the modeling of such relationships. We collect dense annotations of objects, attributes, and relationships within each image to learn these models. Specifically, our dataset contains over 108K images where each image has an average of $$35$$35 objects, $$26$$26 attributes, and $$21$$21 pairwise relationships between objects. We canonicalize the objects, attributes, relationships, and noun phrases in region descriptions and questions answer pairs to WordNet synsets. Together, these annotations represent the densest and largest dataset of image descriptions, objects, attributes, relationships, and question answer pairs.
Omni-Scale Feature Learning For Person Re-Identification As an instance-level recognition problem, person reidentification (ReID) relies on discriminative features, which not only capture different spatial scales but also encapsulate an arbitrary combination of multiple scales. We callse features of both homogeneous and heterogeneous scales omni-scale features. In this paper, a novel deep ReID CNN is designed, termed Omni-Scale Network (OSNet), for omni-scale feature learning. This is achieved by designing a residual block composed of multiple convolutional feature streams, each detecting features at a certain scale. Importantly, a novel unified aggregation gate is introduced to dynamically fuse multiscale features with input-dependent channel-wise weights. To efficiently learn spatial-channel correlations and avoid overfitting, the building block uses both pointwise and depthwise convolutions. By stacking such blocks layer-by-layer, our OSNet is extremely lightweight and can be trained from scratch on existing ReID benchmarks. Despite its small model size, our OSNet achieves state-of-the-art performance on six person-ReID datasets. Code and models are available at: https://github.com/ KaiyangZhou/deep-person-reid.
NIPS 2016 Tutorial: Generative Adversarial Networks. This report summarizes the tutorial presented by the author at NIPS 2016 on generative adversarial networks (GANs). The tutorial describes: (1) Why generative modeling is a topic worth studying, (2) how generative models work, and how GANs compare to other generative models, (3) the details of how GANs work, (4) research frontiers in GANs, and (5) state-of-the-art image models that combine GANs with other methods. Finally, the tutorial contains three exercises for readers to complete, and the solutions to these exercises.
Picbreeder: evolving pictures collaboratively online Picbreeder is an online service that allows users to collaboratively evolve images. Like in other Interactive Evolutionary Computation (IEC) programs, users evolve images on Picbreeder by selecting ones that appeal to them to produce a new generation. However, Picbreeder also offers an online community in which to share these images, and most importantly, the ability to continue evolving others' images. Through this process of branching from other images, and through continually increasing image complexity made possible by the NeuroEvolution of Augmenting Topologies (NEAT) algorithm, evolved images proliferate unlike in any other current IEC systems. Participation requires no explicit talent from the users, thereby opening Picbreeder to the entire Internet community. This paper details how Picbreeder encourages innovation, featuring images that were collaboratively evolved.
Low-Cost Collaborative Mobile Charging for Large-Scale Wireless Sensor Networks. In wireless rechargeable sensor networks (WRSNs), prior studies mainly focus on the optimization of power transfer efficiency. In this work, we consider the cost for building and operating WRSNs. In the network, sensor nodes can be charged by mobile chargers, that have limited energy which is used for charging and moving. We introduce a novel concept called shuttling and introduce an optimal charging algorithm, which is proven to achieve the minimum number of chargers in theory. We also point out the limitations of the optimal algorithm, which motivates the development of solutions named Push-Shuttle-Back (PSB). We formally prove that PSB achieves the minimum number of chargers and the optimal shuttling distance in a 1D scenario with negligible energy loss. When the loss in wireless charging is non-negligible, we propose to exploit detachable battery pack (DBP) and propose a DBP-PSB algorithm to avoid energy loss. We further extend the solution to 2D scenarios and introduce a new circle-based shortcutting scheme that improves charging efficiency and reduces the number of chargers needed to serve the sensor network. We carry out extensive simulations to demonstrate the performance of the proposed algorithms, and the results show the proposed algorithms achieve a low overall cost.
Learning Interpretable Deep State Space Model for Probabilistic Time Series Forecasting. Probabilistic time series forecasting involves estimating the distribution of future based on its history, which is essential for risk management in downstream decision-making. We propose a deep state space model for probabilistic time series forecasting whereby the non-linear emission model and transition model are parameterized by networks and the dependency is modeled by recurrent neural nets. We take the automatic relevance determination (ARD) view and devise a network to exploit the exogenous variables in addition to time series. In particular, our ARD network can incorporate the uncertainty of the exogenous variables and eventually helps identify useful exogenous variables and suppress those irrelevant for forecasting. The distribution of multi-step ahead forecasts are approximated by Monte Carlo simulation. We show in experiments that our model produces accurate and sharp probabilistic forecasts. The estimated uncertainty of our forecasting also realistically increases over time, in a spontaneous manner.
1.010512
0.012609
0.011934
0.011125
0.008696
0.008696
0.007417
0.004348
0.002531
0.000828
0.000097
0
0
0
People detection and tracking from aerial thermal views Detection and tracking of people in visible-light images has been subject to extensive research in the past decades with applications ranging from surveillance to search-and-rescue. Following the growing availability of thermal cameras and the distinctive thermal signature of humans, research effort has been focusing on developing people detection and tracking methodologies applicable to this sensing modality. However, a plethora of challenges arise on the transition from visible-light to thermal images, especially with the recent trend of employing thermal cameras onboard aerial platforms (e.g. in search-and-rescue research) capturing oblique views of the scenery. This paper presents a new, publicly available dataset of annotated thermal image sequences, posing a multitude of challenges for people detection and tracking. Moreover, we propose a new particle filter based framework for tracking people in aerial thermal images. Finally, we evaluate the performance of this pipeline on our dataset, incorporating a selection of relevant, state-of-the-art methods and present a comprehensive discussion of the merits spawning from our study.
Pareto-Optimization for Scheduling of Crude Oil Operations in Refinery via Genetic Algorithm. With the interaction of discrete-event and continuous processes, it is challenging to schedule crude oil operations in a refinery. This paper studies the optimization problem of finding a detailed schedule to realize a given refining schedule. This is a multiobjective optimization problem with a combinatorial nature. Since the original problem cannot be directly solved by using heuristics and meta-heuristics, the problem is transformed into an assignment problem of charging tanks and distillers. Based on such a transformation, by analyzing the properties of the problem, this paper develops a chromosome that can describe a feasible schedule such that meta-heuristics can be applied. Then, it innovatively adopts an improved nondominated sorting genetic algorithm to solve the problem for the first time. An industrial case study is used to test the proposed solution method. The results show that the method makes a significant performance improvement and is applicable to real-life refinery scheduling problems.
A multi-objective model for the green capacitated location-routing problem considering environmental impact. •The Capacitated Location-Routing Problem considering environmental impact is proposed.•A new model for computing greenhouse gas emissions in vehicle routing is proposed.•The Green CLRP is formulated as a bi-objective mixed integer linear programming.•Using more vehicles can lead to large fuel economy in the long term and hence less emission.•More vehicles in shorter routes and prioritizing high demand clients lead to less emission.
Persistent UAV delivery logistics: MILP formulation and efficient heuristic. •UAV delivery logistics with multiple recharge/reload stations was considered.•UAVs visit station, refill consumables and return to service persistently.•Amount of loaded product effects on the flight time of UAVs during delivery.•Validity of the proposed model was demonstrated via island area delivery example.•Performance of mathematical formulation and heuristic were tested and compared.
Multiperiod Asset Allocation Considering Dynamic Loss Aversion Behavior of Investors In order to study the effect of loss aversion behavior on multiperiod investment decisions, we first introduce some psychological characteristics of dynamic loss aversion and then construct a multiperiod portfolio model by considering a conditional value-at-risk (CVaR) constraint. We then design a variable neighborhood search-based hybrid genetic algorithm to solve the model. We finally study the optimal asset allocation and investment performance of the proposed multiperiod model. Some important metrics, such as the initial loss aversion coefficient and reference point, are used to test the robustness of the model. The result shows that investors with loss aversion tend to centralize most of their wealth and have a better performance than rational investors. The effects of CVaR on investment performance are given. When a market is falling, investors with a higher degree of risk aversion can avoid a large loss and can obtain higher gains.
Cooperative Aerial-Ground Vehicle Route Planning With Fuel Constraints for Coverage Applications. Low-cost unmanned aerial vehicles (UAVs) need multiple refuels to accomplish large area coverage. We propose the use of a mobile ground vehicle (GV), constrained to travel on a given road network, as a refueling station for the UAV. Determining optimal routes for a UAV and GV, and selecting rendezvous locations for refueling to minimize coverage time is NP-hard. We develop a two-stage strategy for...
Hamming Embedding and Weak Geometric Consistency for Large Scale Image Search This paper improves recent methods for large scale image search. State-of-the-art methods build on the bag-of-features image representation. We, first, analyze bag-of-features in the framework of approximate nearest neighbor search. This shows the sub-optimality of such a representation for matching descriptors and leads us to derive a more precise representation based on 1) Hamming embedding (HE) and 2) weak geometric consistency constraints (WGC). HE provides binary signatures that refine the matching based on visual words. WGC filters matching descriptors that are not consistent in terms of angle and scale. HE and WGC are integrated within the inverted file and are efficiently exploited for all images, even in the case of very large datasets. Experiments performed on a dataset of one million of images show a significant improvement due to the binary signature and the weak geometric consistency constraints, as well as their efficiency. Estimation of the full geometric transformation, i.e., a re-ranking step on a short list of images, is complementary to our weak geometric consistency constraints and allows to further improve the accuracy.
Microsoft Coco: Common Objects In Context We present a new dataset with the goal of advancing the state-of-the-art in object recognition by placing the question of object recognition in the context of the broader question of scene understanding. This is achieved by gathering images of complex everyday scenes containing common objects in their natural context. Objects are labeled using per-instance segmentations to aid in precise object localization. Our dataset contains photos of 91 objects types that would be easily recognizable by a 4 year old. With a total of 2.5 million labeled instances in 328k images, the creation of our dataset drew upon extensive crowd worker involvement via novel user interfaces for category detection, instance spotting and instance segmentation. We present a detailed statistical analysis of the dataset in comparison to PASCAL, ImageNet, and SUN. Finally, we provide baseline performance analysis for bounding box and segmentation detection results using a Deformable Parts Model.
The Whale Optimization Algorithm. The Whale Optimization Algorithm inspired by humpback whales is proposed.The WOA algorithm is benchmarked on 29 well-known test functions.The results on the unimodal functions show the superior exploitation of WOA.The exploration ability of WOA is confirmed by the results on multimodal functions.The results on structural design problems confirm the performance of WOA in practice. This paper proposes a novel nature-inspired meta-heuristic optimization algorithm, called Whale Optimization Algorithm (WOA), which mimics the social behavior of humpback whales. The algorithm is inspired by the bubble-net hunting strategy. WOA is tested with 29 mathematical optimization problems and 6 structural design problems. Optimization results prove that the WOA algorithm is very competitive compared to the state-of-art meta-heuristic algorithms as well as conventional methods. The source codes of the WOA algorithm are publicly available at http://www.alimirjalili.com/WOA.html
Collaborative privacy management The landscape of the World Wide Web with all its versatile services heavily relies on the disclosure of private user information. Unfortunately, the growing amount of personal data collected by service providers poses a significant privacy threat for Internet users. Targeting growing privacy concerns of users, privacy-enhancing technologies emerged. One goal of these technologies is the provision of tools that facilitate a more informative decision about personal data disclosures. A famous PET representative is the PRIME project that aims for a holistic privacy-enhancing identity management system. However, approaches like the PRIME privacy architecture require service providers to change their server infrastructure and add specific privacy-enhancing components. In the near future, service providers are not expected to alter internal processes. Addressing the dependency on service providers, this paper introduces a user-centric privacy architecture that enables the provider-independent protection of personal data. A central component of the proposed privacy infrastructure is an online privacy community, which facilitates the open exchange of privacy-related information about service providers. We characterize the benefits and the potentials of our proposed solution and evaluate a prototypical implementation.
Cognitive Cars: A New Frontier for ADAS Research This paper provides a survey of recent works on cognitive cars with a focus on driver-oriented intelligent vehicle motion control. The main objective here is to clarify the goals and guidelines for future development in the area of advanced driver-assistance systems (ADASs). Two major research directions are investigated and discussed in detail: 1) stimuli–decisions–actions, which focuses on the driver side, and 2) perception enhancement–action-suggestion–function-delegation, which emphasizes the ADAS side. This paper addresses the important achievements and major difficulties of each direction and discusses how to combine the two directions into a single integrated system to obtain safety and comfort while driving. Other related topics, including driver training and infrastructure design, are also studied.
Completely Pinpointing the Missing RFID Tags in a Time-Efficient Way Radio Frequency Identification (RFID) technology has been widely used in inventory management in many scenarios, e.g., warehouses, retail stores, hospitals, etc. This paper investigates a challenging problem of complete identification of missing tags in large-scale RFID systems. Although this problem has attracted extensive attention from academy and industry, the existing work can hardly satisfy the stringent real-time requirements. In this paper, a Slot Filter-based Missing Tag Identification (SFMTI) protocol is proposed to reconcile some expected collision slots into singleton slots and filter out the expected empty slots as well as the unreconcilable collision slots, thereby achieving the improved time-efficiency. The theoretical analysis is conducted to minimize the execution time of the proposed SFMTI. We then propose a cost-effective method to extend SFMTI to the multi-reader scenarios. The extensive simulation experiments and performance results demonstrate that the proposed SFMTI protocol outperforms the most promising Iterative ID-free Protocol (IIP) by reducing nearly 45% of the required execution time, and is just within a factor of 1.18 from the lower bound of the minimum execution time.
A blind medical image watermarking: DWT-SVD based robust and secure approach for telemedicine applications. In this paper, a blind image watermarking scheme based on discrete wavelet transform (DWT) and singular value decomposition (SVD) is proposed. In this scheme, DWT is applied on ROI (region of interest) of the medical image to get different frequency subbands of its wavelet decomposition. On the low frequency subband LL of the ROI, block-SVD is applied to get different singular matrices. A pair of elements with similar values is identified from the left singular value matrix of these selected blocks. The values of these pairs are modified using certain threshold to embed a bit of watermark content. Appropriate threshold is chosen to achieve the imperceptibility and robustness of medical image and watermark contents respectively. For authentication and identification of original medical image, one watermark image (logo) and other text watermark have been used. The watermark image provides authentication whereas the text data represents electronic patient record (EPR) for identification. At receiving end, blind recovery of both watermark contents is performed by a similar comparison scheme used during the embedding process. The proposed algorithm is applied on various groups of medical images like X-ray, CT scan and mammography. This scheme offers better visibility of watermarked image and recovery of watermark content due to DWT-SVD combination. Moreover, use of Hamming error correcting code (ECC) on EPR text bits reduces the BER and thus provides better recovery of EPR. The performance of proposed algorithm with EPR data coding by Hamming code is compared with the BCH error correcting code and it is found that later one perform better. A result analysis shows that imperceptibility of watermarked image is better as PSNR is above 43 dB and WPSNR is above 52 dB for all set of images. In addition, robustness of the scheme is better than existing scheme for similar set of medical images in terms of normalized correlation coefficient (NCC) and bit-error-rate (BER). An analysis is also carried out to verify the performance of the proposed scheme for different size of watermark contents (image and EPR data). It is observed from analysis that the proposed scheme is also appropriate for watermarking of color image. Using proposed scheme, watermark contents are extracted successfully under various noise attacks like JPEG compression, filtering, Gaussian noise, Salt and pepper noise, cropping, filtering and rotation. Performance comparison of proposed scheme with existing schemes shows proposed scheme has better robustness against different types of attacks. Moreover, the proposed scheme is also robust under set of benchmark attacks known as checkmark attacks.
Hardware Circuits Design and Performance Evaluation of a Soft Lower Limb Exoskeleton Soft lower limb exoskeletons (LLEs) are wearable devices that have good potential in walking rehabilitation and augmentation. While a few studies focused on the structure design and assistance force optimization of the soft LLEs, rarely work has been conducted on the hardware circuits design. The main purpose of this work is to present a new soft LLE for walking efficiency improvement and introduce its hardware circuits design. A soft LLE for hip flexion assistance and a hardware circuits system with scalability were proposed. To assess the efficacy of the soft LLE, the experimental tests that evaluate the sensor data acquisition, force tracking performance, lower limb muscle activity and metabolic cost were conducted. The time error in the peak assistance force was just 1%. The reduction in the normalized root-mean-square EMG of the rectus femoris was 7.1%. The net metabolic cost in exoskeleton on condition was reduced by 7.8% relative to walking with no exoskeleton. The results show that the designed hardware circuits can be applied to the soft LLE and the soft LLE is able to improve walking efficiency of wearers.
1.2
0.2
0.2
0.2
0.2
0.066667
0
0
0
0
0
0
0
0
Equality of Opportunity in Supervised Learning. We propose a criterion for discrimination against a specified sensitive attribute in supervised learning, where the goal is to predict some target based on available features. Assuming data about the predictor, target, and membership in the protected group are available, we show how to optimally adjust any learned predictor so as to remove discrimination according to our definition. Our framework also improves incentives by shifting the cost of poor classification from disadvantaged groups to the decision maker, who can respond by improving the classification accuracy. We enourage readers to consult the more complete manuscript on the arXiv.
Inherent Trade-Offs in the Fair Determination of Risk Scores. Recent discussion in the public sphere about algorithmic classification has involved tension between competing notions of what it means for a probabilistic classification to be fair to different groups. We formalize three fairness conditions that lie at the heart of these debates, and we prove that except in highly constrained special cases, there is no method that can satisfy these three conditions simultaneously. Moreover, even satisfying all three conditions approximately requires that the data lie in an approximate version of one of the constrained special cases identified by our theorem. These results suggest some of the ways in which key notions of fairness are incompatible with each other, and hence provide a framework for thinking about the trade-offs between them.
Racial Disparity in Natural Language Processing: A Case Study of Social Media African-American English. We highlight an important frontier in algorithmic fairness: disparity in the quality of natural language processing algorithms when applied to language from authors of different social groups. For example, current systems sometimes analyze the language of females and minorities more poorly than they do of whites and males. We conduct an empirical analysis of racial disparity in language identification for tweets written in African-American English, and discuss implications of disparity in NLP.
Gender and Dialect Bias in YouTube's Automatic Captions.
The Social Impact Of Natural Language Processing Medical sciences have long since established an ethics code for experiments, to minimize the risk of harm to subjects. Natural language processing (NLP) used to involve mostly anonymous corpora, with the goal of enriching linguistic analysis, and was therefore unlikely to raise ethical concerns. As NLP becomes increasingly wide-spread and uses more data from social media, however, the situation has changed: the outcome of NLP experiments and applications can now have a direct effect on individual users' lives. Until now, the discourse on this topic in the field has not followed the technological development, while public discourse was often focused on exaggerated dangers. This position paper tries to take back the initiative and start a discussion. We identify a number of social implications of NLP and discuss their ethical significance, as well as ways to address them.
Plan-And-Write: Towards Better Automatic Storytelling. Automatic storytelling is challenging since it requires generating long, coherent natural language to describes a sensible sequence of events. Despite considerable efforts on automatic story generation in the past, prior work either is restricted in plot planning, or can only generate stories in a narrow domain. In this paper, we explore open-domain story generation that writes stories given a title (topic) as input. We propose a plan-and-write hierarchical generation framework that first plans a storyline, and then generates a story based on the storyline. We compare two planning strategies. The dynamic schema interweaves story planning and its surface realization in text, while the static schema plans out the entire storyline before generating stories. Experiments show that with explicit storyline planning, the generated stories are more diverse, coherent, and on topic than those generated without creating a full plan, according to both automatic and human evaluations.
A threshold of ln n for approximating set cover Given a collection F of subsets of S ={1,…,n}, setcover is the problem of selecting as few as possiblesubsets from F such that their union coversS,, and maxk-cover is the problem of selectingk subsets from F such that their union has maximum cardinality. Both these problems areNP-hard. We prove that (1 - o(1)) lnn is a threshold below which setcover cannot be approximated efficiently, unless NP has slightlysuperpolynomial time algorithms. This closes the gap (up to low-orderterms) between the ratio of approximation achievable by the greedyalogorithm (which is (1 - o(1)) lnn), and provious results of Lund and Yanakakis, that showed hardness ofapproximation within a ratio of log2n/2&sime;0.72 ln n. For maxk-cover, we show an approximationthreshold of (1 - 1/e)(up tolow-order terms), under assumption that P≠NP.
Multi-column Deep Neural Networks for Image Classification Traditional methods of computer vision and machine learning cannot match human performance on tasks such as the recognition of handwritten digits or traffic signs. Our biologically plausible deep artificial neural network architectures can. Small (often minimal) receptive fields of convolutional winner-take-all neurons yield large network depth, resulting in roughly as many sparsely connected neural layers as found in mammals between retina and visual cortex. Only winner neurons are trained. Several deep neural columns become experts on inputs preprocessed in different ways; their predictions are averaged. Graphics cards allow for fast training. On the very competitive MNIST handwriting benchmark, our method is the first to achieve near-human performance. On a traffic sign recognition benchmark it outperforms humans by a factor of two. We also improve the state-of-the-art on a plethora of common image classification benchmarks.
Markov games as a framework for multi-agent reinforcement learning In the Markov decision process (MDP) formalization of reinforcement learning, a single adaptive agent interacts with an environment defined by a probabilistic transition function. In this solipsis-tic view, secondary agents can only be part of the environment and are therefore fixed in their behavior. The framework of Markov games allows us to widen this view to include multiple adaptive agents with interacting or competing goals. This paper considers a step in this direction in which exactly two agents with diametrically opposed goals share an environment. It describes a Q-learning-like algorithm for finding optimal policies and demonstrates its application to a simple two-player game in which the optimal policy is probabilistic.
Parameter tuning for configuring and analyzing evolutionary algorithms In this paper we present a conceptual framework for parameter tuning, provide a survey of tuning methods, and discuss related methodological issues. The framework is based on a three-tier hierarchy of a problem, an evolutionary algorithm (EA), and a tuner. Furthermore, we distinguish problem instances, parameters, and EA performance measures as major factors, and discuss how tuning can be directed to algorithm performance and/or robustness. For the survey part we establish different taxonomies to categorize tuning methods and review existing work. Finally, we elaborate on how tuning can improve methodology by facilitating well-funded experimental comparisons and algorithm analysis.
Recall-Oriented Evaluation for Information Retrieval Systems. In a recall context, the user is interested in retrieving all relevant documents rather than retrieving a few that are at the top of the results list. In this article we propose ROM (Recall Oriented Measure) which takes into account the main elements that should be considered in evaluating information retrieval systems while ordering them in a way explicitly adapted to a recall context.
Fast and Accurate Estimation of RFID Tags Radio frequency identification (RFID) systems have been widely deployed for various applications such as object tracking, 3-D positioning, supply chain management, inventory control, and access control. This paper concerns the fundamental problem of estimating RFID tag population size, which is needed in many applications such as tag identification, warehouse monitoring, and privacy-sensitive RFID systems. In this paper, we propose a new scheme for estimating tag population size called Average Run-based Tag estimation (ART). The technique is based on the average run length of ones in the bit string received using the standardized framed slotted Aloha protocol. ART is significantly faster than prior schemes. For example, given a required confidence interval of 0.1% and a required reliability of 99.9%, ART is consistently 7 times faster than the fastest existing schemes (UPE and EZB) for any tag population size. Furthermore, ART's estimation time is provably independent of the tag population sizes. ART works with multiple readers with overlapping regions and can estimate sizes of arbitrarily large tag populations. ART is easy to deploy because it neither requires modification to tags nor to the communication protocol between tags and readers. ART only needs to be implemented on readers as a software module.
Finite-Time Adaptive Fuzzy Tracking Control Design for Nonlinear Systems. This paper addresses the finite-time tracking problem of nonlinear pure-feedback systems. Unlike the literature on traditional finite-time stabilization, in this paper the nonlinear system functions, including the bounding functions, are all totally unknown. Fuzzy logic systems are used to model those unknown functions. To present a finite-time control strategy, a criterion of semiglobal practical...
Robot tutor and pupils’ educational ability: Teaching the times tables Research shows promising results of educational robots in language and STEM tasks. In language, more research is available, occasionally in view of individual differences in pupils’ educational ability levels, and learning seems to improve with more expressive robot behaviors. In STEM, variations in robots’ behaviors have been examined with inconclusive results and never while systematically investigating how differences in educational abilities match with different robot behaviors. We applied an autonomously tutoring robot (without tablet, partly WOz) in a 2 × 2 experiment of social vs. neutral behavior in above-average vs. below-average schoolchildren (N = 86; age 8–10 years) while rehearsing the multiplication tables on a one-to-one basis. The standard school test showed that on average, pupils significantly improved their performance even after 3 occasions of 5-min exercises. Beyond-average pupils profited most from a robot tutor, whereas those below average in multiplication benefited more from a robot that showed neutral rather than more social behavior.
1.045172
0.044474
0.041682
0.023138
0.011529
0.008
0.000303
0.000018
0.000002
0
0
0
0
0
HyperFlex: An SDN virtualization architecture with flexible hypervisor function allocation Network Virtualization (NV) and Software-Defined Networking (SDN) are both expected to increase the flexibility and programmability of today's communication networks. Combining both approaches may even be a further step towards increasing the efficiency of network resource utilization. Multiple solutions for virtualizing SDN networks have already been proposed, however, they are either implemented in software or they require special network hardware. We propose HyperFlex, an SDN hypervisor architecture that relies on the decomposition of the hypervisor into functions that are essential for virtualizing SDN networks. The hypervisor functions can be flexibly executed in software or hosted on SDN network elements. Furthermore, existing hypervisor solutions focus on data-plane virtualization mechanisms and neglect the virtualization of the control-plane of SDN networks. HyperFlex provides control-plane virtualization by adding a control-plane isolation function, either in software or on network elements. The isolation function ensures that the resources of the control-plane are shared correctly between each virtual SDN network while it also protects the hypervisor resources from resource exhaustion.
Minimum interference routing of bandwidth guaranteed tunnels with MPLS traffic engineering applications This paper presents new algorithms for dynamic routing of bandwidth guaranteed tunnels, where tunnel routing requests arrive one by one and there is no a priori knowledge regarding future requests. This problem is motivated by the service provider needs for fast deployment of bandwidth guaranteed services. Offline routing algorithms cannot be used since they require a priori knowledge of all tunnel requests that are to be rooted. Instead, on-line algorithms that handle requests arriving one by one and that satisfy as many potential future demands as possible are needed. The newly developed algorithms are on-line algorithms and are based on the idea that a newly routed tunnel must follow a route that does not “interfere too much” with a route that may he critical to satisfy a future demand. We show that this problem is NP-hard. We then develop path selection heuristics which are based on the idea of deferred loading of certain “critical” links. These critical links are identified by the algorithm as links that, if heavily loaded, would make it impossible to satisfy future demands between certain ingress-egress pairs. Like min-hop routing, the presented algorithm uses link-state information and some auxiliary capacity information for path selection. Unlike previous algorithms, the proposed algorithm exploits any available knowledge of the network ingress-egress points of potential future demands, even though the demands themselves are unknown. If all nodes are ingress-egress nodes, the algorithm can still be used, particularly to reduce the rejection rate of requests between a specified subset of important ingress-egress pairs. The algorithm performs well in comparison to previously proposed algorithms on several metrics like the number of rejected demands and successful rerouting of demands upon link failure
The set cover with pairs problem We consider a generalization of the set cover problem, in which elements are covered by pairs of objects, and we are required to find a minimum cost subset of objects that induces a collection of pairs covering all elements. Formally, let U be a ground set of elements and let ${\cal S}$ be a set of objects, where each object i has a non-negative cost wi. For every $\{ i, j \} \subseteq {\cal S}$, let ${\cal C}(i,j)$ be the collection of elements in U covered by the pair { i, j }. The set cover with pairs problem asks to find a subset $A \subseteq {\cal S}$ such that $\bigcup_{ \{ i, j \} \subseteq A } {\cal C}(i,j) = U$ and such that ∑i∈Awi is minimized. In addition to studying this general problem, we are also concerned with developing polynomial time approximation algorithms for interesting special cases. The problems we consider in this framework arise in the context of domination in metric spaces and separation of point sets.
Control Plane Latency With SDN Network Hypervisors: The Cost of Virtualization. Software defined networking (SDN) network hypervisors provide the functionalities needed for virtualizing software-defined networks. Hypervisors sit logically between the multiple virtual SDN networks (vSDNs), which reside on the underlying physical SDN network infrastructure, and the corresponding tenant (vSDN) controllers. Different SDN network hypervisor architectures have mainly been explored through proof-of-concept implementations. We fundamentally advance SDN network hypervisor research by conducting a model-based analysis of SDN hypervisor architectures. Specifically, we introduce mixed integer programming formulations for four different SDN network hypervisor architectures. Our model formulations can also optimize the placement of multi-controller switches in virtualized OpenFlow-enabled SDN networks. We employ our models to quantitatively examine the optimal placement of the hypervisor instances. We compare the control plane latencies of the different SDN hypervisor architectures and quantify the cost of virtualization, i.e., the latency overhead due to virtualizing SDN networks via hypervisors. For generalization, we quantify how the hypervisor architectures behave for different network topologies. Our model formulations and the insights drawn from our evaluations inform network operators about the trade-offs of the different hypervisor architectures and help choosing an architecture according to operator demands.
Flow Setup Latency in SDN Networks. In software-defined networking, the typical switch-controller cycle, from generating a network event notification at the controller until the flow rules are installed at the switches, is not an instantaneous activity. Our measurement results show that this has serious implications on the performance of flow setup procedure, specifically for larger networks: we observe that, even with software swit...
Near-Optimal Disjoint-Path Facility Location Through Set Cover by Pairs AbstractContent Distribution and End-to-End Monitoring with Set Cover by PairsDigital content is housed at data centers located on nodes of a data network. Consumers of this content are also located on network nodes. Content flows from a data center to a consumer on a path defined by a routing protocol, such as open shortest path first. A pair of data centers is said to feasibly serve content to a consumer if there are disjoint paths from each data center to the consumer. In “Near-Optimal Disjoint-Path Facility Location Through Set Cover by Pairs,” Johnson, Breslau, Diakonikolas, Duffield, Gu, Hajiaghayi, Karloff, Resende, and Sen study this problem when the goal is to minimize the number of required data centers to serve a set of consumers. They also study another facility location problem that arises in network traffic monitoring. Both problems are modeled as a set cover-by-pairs problem. The authors provide complexity results, a new lower-bounding integer programming formulation, and several heuristics. The lower bounds are easily computed with a commercial MIP solver and validate the claim of near-optimality of their heuristics.In this paper, we consider two special cases of the “cover-by-pairs” optimization problem that arises when we need to place facilities so that each customer is served by two facilities that reach it by disjoint shortest paths. These problems arise in a network traffic-monitoring scheme proposed by Breslau et al. and have potential applications to content distribution. The “set-disjoint” variant applies to networks that use the open shortest path first routing protocol, and the “path-disjoint” variant applies when multiprotocol label switching routing is enabled, making better solutions possible at the cost of greater operational expense. Although we can prove that no polynomial-time algorithm can guarantee good solutions for either version, we are able to provide heuristics that do very well in practice on instances with real-world network structure. Fast implementations of the heuristics, made possible by exploiting mathematical observations about the relationship between the network instances and the corresponding instances of the cover-by-pairs problem, allow us to perform an extensive experimental evaluation of the heuristics and what the solutions they produce tell us about the effectiveness of the proposed monitoring scheme. For the set-disjoint variant, we validate our claim of near-optimality via a new lower-bounding integer programming formulation. Although computing this lower bound requires solving the NP-hard hitting set problem and can underestimate the optimal value by a linear factor in the worst case, it can be computed quickly by CPLEX, and it equals the optimal solution value for all the instances in our extensive test bed.
All quiet on the internet front? With the proliferation and increasing dependence of many services and applications on the Internet, this network has become a vital societal asset. This creates the need to protect this critical infrastructure, and over the past years a variety of resilience schemes have been proposed. The effectiveness of protection schemes, however, highly depends on the causes and circumstances of Internet fail...
Reaching Agreement in the Presence of Faults The problem addressed here concerns a set of isolated processors, some unknown subset of which may be faulty, that communicate only by means of two-party messages. Each nonfaulty processor has a private value of information that must be communicated to each other nonfaulty processor. Nonfaulty processors always communicate honestly, whereas faulty processors may lie. The problem is to devise an algorithm in which processors communicate their own values and relay values received from others that allows each nonfaulty processor to infer a value for each other processor. The value inferred for a nonfaulty processor must be that processor's private value, and the value inferred for a faulty one must be consistent with the corresponding value inferred by each other nonfaulty processor.It is shown that the problem is solvable for, and only for, n ≥ 3m + 1, where m is the number of faulty processors and n is the total number. It is also shown that if faulty processors can refuse to pass on information but cannot falsely relay information, the problem is solvable for arbitrary n ≥ m ≥ 0. This weaker assumption can be approximated in practice using cryptographic methods.
Modeling air-to-ground path loss for low altitude platforms in urban environments The reliable prediction of coverage footprint resulting from an airborne wireless radio base station, is at utmost importance, when it comes to the new emerging applications of air-to-ground wireless services. These applications include the rapid recovery of damaged terrestrial wireless infrastructure due to a natural disaster, as well as the fulfillment of sudden wireless traffic overload in certain spots due to massive movement of crowds. In this paper, we propose a statistical propagation model for predicting the air-to-ground path loss between a low altitude platform and a terrestrial terminal. The prediction is based on the urban environment properties, and is dependent on the elevation angle between the terminal and the platform. The model shows that air-to-ground path loss is following two main propagation groups, characterized by two different path loss profiles. In this paper we illustrate the methodology of which the model was deduced, as well as we present the different path loss profiles including the occurrence probability of each.
Factorizing personalized Markov chains for next-basket recommendation Recommender systems are an important component of many websites. Two of the most popular approaches are based on matrix factorization (MF) and Markov chains (MC). MF methods learn the general taste of a user by factorizing the matrix over observed user-item preferences. On the other hand, MC methods model sequential behavior by learning a transition graph over items that is used to predict the next action based on the recent actions of a user. In this paper, we present a method bringing both approaches together. Our method is based on personalized transition graphs over underlying Markov chains. That means for each user an own transition matrix is learned - thus in total the method uses a transition cube. As the observations for estimating the transitions are usually very limited, our method factorizes the transition cube with a pairwise interaction model which is a special case of the Tucker Decomposition. We show that our factorized personalized MC (FPMC) model subsumes both a common Markov chain and the normal matrix factorization model. For learning the model parameters, we introduce an adaption of the Bayesian Personalized Ranking (BPR) framework for sequential basket data. Empirically, we show that our FPMC model outperforms both the common matrix factorization and the unpersonalized MC model both learned with and without factorization.
Multistep-Ahead time series prediction Multistep-ahead prediction is the task of predicting a sequence of values in a time series. A typical approach, known as multi-stage prediction, is to apply a predictive model step-by-step and use the predicted value of the current time step to determine its value in the next time step. This paper examines two alternative approaches known as independent value prediction and parameter prediction. The first approach builds a separate model for each prediction step using the values observed in the past. The second approach fits a parametric function to the time series and builds models to predict the parameters of the function. We perform a comparative study on the three approaches using multiple linear regression, recurrent neural networks, and a hybrid of hidden Markov model with multiple linear regression. The advantages and disadvantages of each approach are analyzed in terms of their error accumulation, smoothness of prediction, and learning difficulty.
Learning A Discriminative Null Space For Person Re-Identification Most existing person re-identification (re-id) methods focus on learning the optimal distance metrics across camera views. Typically a person's appearance is represented using features of thousands of dimensions, whilst only hundreds of training samples are available due to the difficulties in collecting matched training images. With the number of training samples much smaller than the feature dimension, the existing methods thus face the classic small sample size (SSS) problem and have to resort to dimensionality reduction techniques and/or matrix regularisation, which lead to loss of discriminative power. In this work, we propose to overcome the SSS problem in re-id distance metric learning by matching people in a discriminative null space of the training data. In this null space, images of the same person are collapsed into a single point thus minimising the within-class scatter to the extreme and maximising the relative between-class separation simultaneously. Importantly, it has a fixed dimension, a closed-form solution and is very efficient to compute. Extensive experiments carried out on five person re-identification benchmarks including VIPeR, PRID2011, CUHK01, CUHK03 and Market1501 show that such a simple approach beats the state-of-the-art alternatives, often by a big margin.
A robust medical image watermarking against salt and pepper noise for brain MRI images. The ever-growing numbers of medical digital images and the need to share them among specialists and hospitals for better and more accurate diagnosis require that patients' privacy be protected. During the transmission of medical images between hospitals or specialists through the network, the main priority is to protect a patient's documents against any act of tampering by unauthorised individuals. Because of this, there is a need for medical image authentication scheme to enable proper diagnosis on patient. In addition, medical images are also susceptible to salt and pepper impulse noise through the transmission in communication channels. This noise may also be intentionally used by the invaders to corrupt the embedded watermarks inside the medical images. A common drawback of existing watermarking methods is their weakness against salt and pepper noise. The research carried out in this work addresses the issue of designing a new watermarking method that can withstand high density of salt and pepper noise for brain MRI images. For this purpose, combination of a spatial domain watermarking method, channel coding and noise filtering schemes are used. The region of non-interest (RONI) of MRI images from five different databases are used as embedding area and electronic patient record (EPR) is considered as embedded data. The quality of watermarked image is evaluated using Peak Signal-to-Noise Ratio (PSNR) and Structural Similarity Index (SSIM), and the accuracy of the extracted watermark is assessed in terms of Bit Error Rate (BER).
Convert Harm Into Benefit: A Coordination-Learning Based Dynamic Spectrum Anti-Jamming Approach This paper mainly investigates the multi-user anti-jamming spectrum access problem. Using the idea of “converting harm into benefit,” the malicious jamming signals projected by the enemy are utilized by the users as the coordination signals to guide spectrum coordination. An “internal coordination-external confrontation” multi-user anti-jamming access game model is constructed, and the existence of Nash equilibrium (NE) as well as correlated equilibrium (CE) is demonstrated. A coordination-learning based anti-jamming spectrum access algorithm (CLASA) is designed to achieve the CE of the game. Simulation results show the convergence, and effectiveness of the proposed CLASA algorithm, and indicate that our approach can help users confront the malicious jammer, and coordinate internal spectrum access simultaneously without information exchange. Last but not least, the fairness of the proposed approach under different jamming attack patterns is analyzed, which illustrates that this approach provides fair anti-jamming spectrum access opportunities under complicated jamming pattern.
1.073962
0.066667
0.066667
0.066667
0.066667
0.066667
0.0032
0
0
0
0
0
0
0
Data Collection in Wireless Sensor Networks with Mobile Elements: A Survey Wireless sensor networks (WSNs) have emerged as an effective solution for a wide range of applications. Most of the traditional WSN architectures consist of static nodes which are densely deployed over a sensing area. Recently, several WSN architectures based on mobile elements (MEs) have been proposed. Most of them exploit mobility to address the problem of data collection in WSNs. In this article we first define WSNs with MEs and provide a comprehensive taxonomy of their architectures, based on the role of the MEs. Then we present an overview of the data collection process in such a scenario, and identify the corresponding issues and challenges. On the basis of these issues, we provide an extensive survey of the related literature. Finally, we compare the underlying approaches and solutions, with hints to open problems and future research directions.
Energy-Harvesting Wireless Sensor Networks (EH-WSNs): A Review. Wireless Sensor Networks (WSNs) are crucial in supporting continuous environmental monitoring, where sensor nodes are deployed and must remain operational to collect and transfer data from the environment to a base-station. However, sensor nodes have limited energy in their primary power storage unit, and this energy may be quickly drained if the sensor node remains operational over long periods of time. Therefore, the idea of harvesting ambient energy from the immediate surroundings of the deployed sensors, to recharge the batteries and to directly power the sensor nodes, has recently been proposed. The deployment of energy harvesting in environmental field systems eliminates the dependency of sensor nodes on battery power, drastically reducing the maintenance costs required to replace batteries. In this article, we review the state-of-the-art in energy-harvesting WSNs for environmental monitoring applications, including Animal Tracking, Air Quality Monitoring, Water Quality Monitoring, and Disaster Monitoring to improve the ecosystem and human life. In addition to presenting the technologies for harvesting energy from ambient sources and the protocols that can take advantage of the harvested energy, we present challenges that must be addressed to further advance energy-harvesting-based WSNs, along with some future work directions to address these challenges.
A Survey on Mobility and Mobility-Aware MAC Protocols in Wireless Sensor Networks. In wireless sensor networks nodes can be static or mobile, depending on the application requirements. Dealing with mobility can pose some formidable challenges in protocol design, particularly, at the link layer. These difficulties require mobility adaptation algorithms to localize mobile nodes and predict the quality of link that can be established with them. This paper surveys the current state-...
Tag-based cooperative data gathering and energy recharging in wide area RFID sensor networks The Wireless Identification and Sensing Platform (WISP) conjugates the identification potential of the RFID technology and the sensing and computing capability of the wireless sensors. Practical issues, such as the need of periodically recharging WISPs, challenge the effective deployment of large-scale RFID sensor networks (RSNs) consisting of RFID readers and WISP nodes. In this view, the paper proposes cooperative solutions to energize the WISP devices in a wide-area sensing network while reducing the data collection delay. The main novelty is the fact that both data transmissions and energy transfer are based on the RFID technology only: RFID mobile readers gather data from the WISP devices, wirelessly recharge them, and mutually cooperate to reduce the data delivery delay to the sink. Communication between mobile readers relies on two proposed solutions: a tag-based relay scheme, where RFID tags are exploited to temporarily store sensed data at pre-determined contact points between the readers; and a tag-based data channel scheme, where the WISPs are used as a virtual communication channel for real time data transfer between the readers. Both solutions require: (i) clustering the WISP nodes; (ii) dimensioning the number of required RFID mobile readers; (iii) planning the tour of the readers under the energy and time constraints of the nodes. A simulative analysis demonstrates the effectiveness of the proposed solutions when compared to non-cooperative approaches. Differently from classic schemes in the literature, the solutions proposed in this paper better cope with scalability issues, which is of utmost importance for wide area networks.
Joint Power Charging and Routing in Wireless Rechargeable Sensor Networks. The development of wireless power transfer (WPT) technology has inspired the transition from traditional battery-based wireless sensor networks (WSNs) towards wireless rechargeable sensor networks (WRSNs). While extensive efforts have been made to improve charging efficiency, little has been done for routing optimization. In this work, we present a joint optimization model to maximize both charging efficiency and routing structure. By analyzing the structure of the optimization model, we first decompose the problem and propose a heuristic algorithm to find the optimal charging efficiency for the predefined routing tree. Furthermore, by coding the many-to-one communication topology as an individual, we further propose to apply a genetic algorithm (GA) for the joint optimization of both routing and charging. The genetic operations, including tree-based recombination and mutation, are proposed to obtain a fast convergence. Our simulation results show that the heuristic algorithm reduces the number of resident locations and the total moving distance. We also show that our proposed algorithm achieves a higher charging efficiency compared with existing algorithms.
Hybrid charging scheduling schemes for three-dimensional underwater wireless rechargeable sensor networks. •We are the first to study charging issue for 3D underwater rechargeable sensor networks.•We develop a series of charging algorithms for enhancing energy efficiency.•Our schemes can save energy, save time, and ensure effective utilization of resources.
Energy and Distance Optimization in Rechargeable Wireless Sensor Networks The aim of a mobile recharger operating in a wireless sensor network (WSN) is to keep the network's average consumed energy and covered distance low. As shown in the literature, the covered distance is minimized when the mobile recharger's base is located as per the solution of a median problem, while the network's average energy consumption is minimized as per the solution of a different median problem. In this work, the first problem is analytically investigated, showing that its solution depends on the traffic load and the topology characteristics. Furthermore, it is shown that, under certain conditions, the solution for both problems is identical. These analytical results motivate the introduction of a new on-demand recharging policy, simple to be implemented and depending on local information. The simulation results confirm the analytical findings, showing that the solutions of both median problems are identical under certain conditions in WSN environments. Additionally, the proposed recharging policy is evaluated against a well-known policy that exploits global knowledge, demonstrating its advantage for prolonging network lifetime. For both recharging policies, it is shown that energy consumption and covered distance are minimized when the mobile recharger is initially located at the solution of the said median problems.
Simultaneous wireless information and power transfer in modern communication systems Energy harvesting for wireless communication networks is a new paradigm that allows terminals to recharge their batteries from external energy sources in the surrounding environment. A promising energy harvesting technology is wireless power transfer where terminals harvest energy from electromagnetic radiation. Thereby, the energy may be harvested opportunistically from ambient electromagnetic sources or from sources that intentionally transmit electromagnetic energy for energy harvesting purposes. A particularly interesting and challenging scenario arises when sources perform simultaneous wireless information and power transfer (SWIPT), as strong signals not only increase power transfer but also interference. This article provides an overview of SWIPT systems with a particular focus on the hardware realization of rectenna circuits and practical techniques that achieve SWIPT in the domains of time, power, antennas, and space. The article also discusses the benefits of a potential integration of SWIPT technologies in modern communication networks in the context of resource allocation and cooperative cognitive radio networks.
P2S: A Primary and Passer-By Scheduling Algorithm for On-Demand Charging Architecture in Wireless Rechargeable Sensor Networks. As the interdiscipline of wireless communication and control engineering, the cooperative charging issue in wireless rechargeable sensor networks (WRSNs) is a popular research problem. With the help of wireless power transfer technology, electrical energy can be transferred from wireless charging vehicles to sensors, providing a new paradigm to prolong the network lifetime. However, existing techn...
A Vision of 6G Wireless Systems: Applications, Trends, Technologies, and Open Research Problems. The ongoing deployment of 5G cellular systems is continuously exposing the inherent limitations of this system, compared to its original premise as an enabler for Internet of Everything applications. These 5G drawbacks are spurring worldwide activities focused on defining the next-generation 6G wireless system that can truly integrate far-reaching applications ranging from autonomous systems to ex...
Synonymous Paraphrasing Using WordNet and Internet We propose a method of synonymous paraphrasing of a text based on WordNet synonymy data and Internet statistics of stable word combinations (collocations). Given a text, we look for words or expressions in it for which WordNet provides synonyms, and substitute them with such synonyms only if the latter form valid collocations with the surrounding words according to the statistics gathered from Internet. We present two important applications of such synonymous paraphrasing: (1) style-checking and correction: automatic evaluation and computer-aided improvement of writing style with regard to various aspects (increasing vs. decreasing synonymous variation, conformistic vs. individualistic selection of synonyms, etc.) and (2) steganography: hiding of additional information in the text by special selection of synonyms. A basic interactive algorithm of style improvement is outlined and an example of its application to editing of newswire text fragment in English is traced. Algorithms of style evaluation and information hiding are also proposed.
A robust fuzzy adaptive law for evolving control systems. In this paper an adaptive law with leakage is presented. This law can be used in the consequent part of Takagi–Sugeno-based control. The approach enables easy implementation in the control systems with evolving antecedent part. This combination results in a high-performance and robust control of nonlinear and slowly varying systems. It is shown in the paper that the proposed adaptive law is a natural way to cope with the parasitic dynamics. The boundedness of estimated parameters, the tracking error and all the signals in the system is guaranteed if the leakage parameter σ′ is large enough. This means that the proposed adaptive law ensures the global stability of the system. A simulation example is given that illustrates the proposed approach.
Applications of Blockchain Technology beyond Cryptocurrency. Blockchain (BC), the technology behind the Bitcoin crypto-currency system, is considered to be both alluring and critical for ensuring enhanced security and (in some implementations, non-traceable) privacy for diverse applications in many other domains including in the Internet of Things (IoT) eco-system. Intensive research is currently being conducted in both academia and industry applying the Blockchain technology in multifarious applications. Proof-of-Work (PoW), a cryptographic puzzle, plays a vital role in ensuring BC security by maintaining a digital ledger of transactions, which is considered to be incorruptible. Furthermore, BC uses a changeable Public Key (PK) to record the usersu0027 identity, which provides an extra layer of privacy. Not only in cryptocurrency has the successful adoption of BC been implemented but also in multifaceted non-monetary systems such as in: distributed storage systems, proof-of-location, healthcare, decentralized voting and so forth. Recent research articles and projects/applications were surveyed to assess the implementation of BC for enhanced security, to identify associated challenges and to propose solutions for BC enabled enhanced security systems.
Driver’s Intention Identification With the Involvement of Emotional Factors in Two-Lane Roads Driver’s emotion is a psychological reaction to environmental stimulus. Driver intention is an internal state of mind, which directs the actions in the next moment during driving. Emotions usually have a strong influence on behavioral intentions. Therefore, emotion is an important factor that should be considered, to accurately identify driver’s intention. This study used the support vector machin...
1.036063
0.035519
0.035237
0.033333
0.033333
0.033333
0.033333
0.017897
0.006863
0.000028
0
0
0
0
Chaos-Based Content Distribution Framework for Digital Rights Management System Multimedia contents are digitally utilized these days. Thus, the development of an effective method to access the content is becoming the topmost priority of the entertainment industry to protect the digital content from unauthorized access. Digital rights management (DRM) systems are the technique that makes digital content accessible only to the legal rights holders. As the Internet of Things environment is used in the distribution and access of digital content, a secure and efficient content delivery mechanism is also required. Keeping the focus on these points, this article proposes a content distribution framework for DRM system using chaotic map. Formal security verification under the random oracle model, which uncovers the proposed protocol's capability to resist the critical attacks is given. Moreover, simulation study for security verification is performed using the broadly accepted “automated validation of Internet security protocols and applications,” which indicates that the protocol is safe. Moreover, the detailed comparative study with related protocols demonstrates that it provides better security and improves the computational and communication efficiency.
Computational Efficient Authenticated Digital Content Distribution Frameworks for DRM Systems: Review and Outlook Advancement in digital technologies presents a user-friendly environment for the digital content distribution. However, it makes digital content prone to piracy issues. Digital rights management (DRM) systems aim to ensure the authorized content usage. As the digital content broadcasts through the public network, a secure and authorized content access mechanism is required. As digital media users ...
Secure and ubiquitous authenticated content distribution framework for IoT enabled DRM system Internet of Things (IoT) is increasingly used through smart devices with internet-based networks. Communication and data sharing between these devices have also grown in several ways. It is presenting a new dimension to the whole digital right management (DRM) industry. The main focus of IoT based DRM technology is to facilitate the authorised user for using multimedia content through smart devices. However, threats of information breach between communication channels also rapidly increasing, which is making content distribution a challenging task. Moreover, the computation and communication efficiency along with user privacy also requires an ideal DRM system. To address concerns of security, efficiency and privacy over internet-based networks, we design a content key distribution framework for DRM systems. The security proof of the proposed framework is given in the random oracle model along with informal security analysis. Moreover, the security analysis performed using widely adopted simulation tool, namely, "Automated Validation of Internet Security Protocol and Application (AVISPA)". The study of performance is conducted, which indicates that it fulfils the requirements of computation and computation efficiency.
Privacy Enabled Digital Rights Management Without Trusted Third Party Assumption Digital rights management systems are required to provide security and accountability without violating the privacy of the entities involved. However, achieving privacy along with accountability in the same framework is hard as these attributes are mutually contradictory. Thus, most of the current digital rights management systems rely on trusted third parties to provide privacy to the entities involved. However, a trusted third party can become malicious and break the privacy protection of the entities in the system. Hence, in this paper, we propose a novel privacy preserving content distribution mechanism for digital rights management without relying on the trusted third party assumption. We use simple primitives such as blind decryption and one way hash chain to avoid the trusted third party assumption. We prove that our scheme is not prone to the “oracle problem” of the blind decryption mechanism. The proposed mechanism supports access control without degrading user's privacy as well as allows revocation of even malicious users without violating their privacy.
A more secure digital rights management authentication scheme based on smart card Digital rights management (DRM) system is a technology based mechanism to ensure only authorized access and legal distribution/consumption of the protected digital content. DRM system deals with the whole lifecycle of the digital content including production, management, distribution and consumption. DRM schemes are effective means for the transfer of digital content and safeguard the intellectual property. Recently, Yang et al. proposed a smart-card based DRM authentication scheme providing mutual authentication and session key establishment among all the participants of the DRM environment. We show that their scheme does not resist threats like smart card attack; fails to provide proper password update facility; and does not follow forward secrecy. To overcome these weaknesses, we propose an improvement of Yang et al.’s scheme. The security of our scheme remains intact even if the smart card of the user is lost. In our scheme, user’s smart card is capable of verifying the correctness of the inputted identity and password and hence contributes to achieve an efficient and user- friendly password update phase. In addition, the session keys established between the participating entities are highly secure by virtue of forward secrecy property. We conduct security analysis and comparison with related schemes to evaluate our improved scheme. During comparison, we also highlight the computational cost/time complexity at the user and the server side in terms of the execution time of various operations. The entire analysis shows that the design of the improved scheme is robust enough for the for DRM environment.
SAFE: Secure Appliance Scheduling for Flexible and Efficient Energy Consumption for Smart Home IoT Smart homes (SHs) aim at forming an energy optimized environment that can efficiently regulate the use of various Internet of Things (IoT) devices in its network. Real-time electricity pricing models along with SHs provide users an opportunity to reduce their electricity expenditure by responding to the pricing that varies with different times of the day, resulting in reducing the expenditure at both customers’ and utility provider’s end. However, responding to such prices and effectively scheduling the appliances under such complex dynamics is a challenging optimization problem to be solved by the provider or by third party services. As communication in SH-IoT environment is extremely sensitive and private, reporting of such usage information to the provider to solve the optimization has a potential risk that the provider or third party services may track users’ energy consumption profile which compromises users’ privacy. To address these issues, we developed a homomorphic encryption-based alternating direction method of multipliers approach to solve the cost-aware appliance scheduling optimization in a distributed manner and schedule home appliances without leaking users’ privacy. Through extensive simulation study considering real-world datasets, we show that the proposed secure appliance scheduling for flexible and efficient energy consumption scheme, namely SAFE, effectively lowers electricity cost while preserving users’ privacy.
A secure and efficient outsourced computation on data sharing scheme for privacy computing. With the development of computer technology, it makes malicious users more easily get data stored in cloud. However, these data are always related to users’ privacy, and it is harmful when the data are acquired by attackers. Ciphertext-policy attribute-based encryption (CP-ABE) is suitable to achieve privacy and security in cloud. In this paper, we put forward a secure and efficient outsourced computation algorithm on data sharing scheme for privacy computing. Existing schemes only outsource decryption computation to the cloud, users still have heavy burden in encryption. In order to reduce the computing burden of users, most encryption and decryption computations are outsourced to the cloud service provider in our construction. At the same time, to apply in practice, we propose efficient user and attribute revocation. Finally, the security analysis and simulation results show that our scheme is secure and efficient compared with existing schemes.
Image quality assessment: from error visibility to structural similarity. Objective methods for assessing perceptual image quality traditionally attempted to quantify the visibility of errors (differences) between a distorted image and a reference image using a variety of known properties of the human visual system. Under the assumption that human visual perception is highly adapted for extracting structural information from a scene, we introduce an alternative complementary framework for quality assessment based on the degradation of structural information. As a specific example of this concept, we develop a Structural Similarity Index and demonstrate its promise through a set of intuitive examples, as well as comparison to both subjective ratings and state-of-the-art objective methods on a database of images compressed with JPEG and JPEG2000.
Deep learning Deep learning allows computational models that are composed of multiple processing layers to learn representations of data with multiple levels of abstraction. These methods have dramatically improved the state-of-the-art in speech recognition, visual object recognition, object detection and many other domains such as drug discovery and genomics. Deep learning discovers intricate structure in large data sets by using the backpropagation algorithm to indicate how a machine should change its internal parameters that are used to compute the representation in each layer from the representation in the previous layer. Deep convolutional nets have brought about breakthroughs in processing images, video, speech and audio, whereas recurrent nets have shone light on sequential data such as text and speech. Machine-learning technology powers many aspects of modern society: from web searches to content filtering on social networks to recommendations on e-commerce websites, and it is increasingly present in consumer products such as cameras and smartphones. Machine-learning systems are used to identify objects in images, transcribe speech into text, match news items, posts or products with users' interests, and select relevant results of search. Increasingly, these applications make use of a class of techniques called deep learning. Conventional machine-learning techniques were limited in their ability to process natural data in their raw form. For decades, constructing a pattern-recognition or machine-learning system required careful engineering and considerable domain expertise to design a feature extractor that transformed the raw data (such as the pixel values of an image) into a suitable internal representation or feature vector from which the learning subsystem, often a classifier, could detect or classify patterns in the input. Representation learning is a set of methods that allows a machine to be fed with raw data and to automatically discover the representations needed for detection or classification. Deep-learning methods are representation-learning methods with multiple levels of representation, obtained by composing simple but non-linear modules that each transform the representation at one level (starting with the raw input) into a representation at a higher, slightly more abstract level. With the composition of enough such transformations, very complex functions can be learned. For classification tasks, higher layers of representation amplify aspects of the input that are important for discrimination and suppress irrelevant variations. An image, for example, comes in the form of an array of pixel values, and the learned features in the first layer of representation typically represent the presence or absence of edges at particular orientations and locations in the image. The second layer typically detects motifs by spotting particular arrangements of edges, regardless of small variations in the edge positions. The third layer may assemble motifs into larger combinations that correspond to parts of familiar objects, and subsequent layers would detect objects as combinations of these parts. The key aspect of deep learning is that these layers of features are not designed by human engineers: they are learned from data using a general-purpose learning procedure. Deep learning is making major advances in solving problems that have resisted the best attempts of the artificial intelligence community for many years. It has turned out to be very good at discovering intricate structures in high-dimensional data and is therefore applicable to many domains of science, business and government. In addition to beating records in image recognition1, 2, 3, 4 and speech recognition5, 6, 7, it has beaten other machine-learning techniques at predicting the activity of potential drug molecules8, analysing particle accelerator data9, 10, reconstructing brain circuits11, and predicting the effects of mutations in non-coding DNA on gene expression and disease12, 13. Perhaps more surprisingly, deep learning has produced extremely promising results for various tasks in natural language understanding14, particularly topic classification, sentiment analysis, question answering15 and language translation16, 17. We think that deep learning will have many more successes in the near future because it requires very little engineering by hand, so it can easily take advantage of increases in the amount of available computation and data. New learning algorithms and architectures that are currently being developed for deep neural networks will only accelerate this progress. The most common form of machine learning, deep or not, is supervised learning. Imagine that we want to build a system that can classify images as containing, say, a house, a car, a person or a pet. We first collect a large data set of images of houses, cars, people and pets, each labelled with its category. During training, the machine is shown an image and produces an output in the form of a vector of scores, one for each category. We want the desired category to have the highest score of all categories, but this is unlikely to happen before training. We compute an objective function that measures the error (or distance) between the output scores and the desired pattern of scores. The machine then modifies its internal adjustable parameters to reduce this error. These adjustable parameters, often called weights, are real numbers that can be seen as 'knobs' that define the input–output function of the machine. In a typical deep-learning system, there may be hundreds of millions of these adjustable weights, and hundreds of millions of labelled examples with which to train the machine. To properly adjust the weight vector, the learning algorithm computes a gradient vector that, for each weight, indicates by what amount the error would increase or decrease if the weight were increased by a tiny amount. The weight vector is then adjusted in the opposite direction to the gradient vector. The objective function, averaged over all the training examples, can be seen as a kind of hilly landscape in the high-dimensional space of weight values. The negative gradient vector indicates the direction of steepest descent in this landscape, taking it closer to a minimum, where the output error is low on average. In practice, most practitioners use a procedure called stochastic gradient descent (SGD). This consists of showing the input vector for a few examples, computing the outputs and the errors, computing the average gradient for those examples, and adjusting the weights accordingly. The process is repeated for many small sets of examples from the training set until the average of the objective function stops decreasing. It is called stochastic because each small set of examples gives a noisy estimate of the average gradient over all examples. This simple procedure usually finds a good set of weights surprisingly quickly when compared with far more elaborate optimization techniques18. After training, the performance of the system is measured on a different set of examples called a test set. This serves to test the generalization ability of the machine — its ability to produce sensible answers on new inputs that it has never seen during training. Many of the current practical applications of machine learning use linear classifiers on top of hand-engineered features. A two-class linear classifier computes a weighted sum of the feature vector components. If the weighted sum is above a threshold, the input is classified as belonging to a particular category. Since the 1960s we have known that linear classifiers can only carve their input space into very simple regions, namely half-spaces separated by a hyperplane19. But problems such as image and speech recognition require the input–output function to be insensitive to irrelevant variations of the input, such as variations in position, orientation or illumination of an object, or variations in the pitch or accent of speech, while being very sensitive to particular minute variations (for example, the difference between a white wolf and a breed of wolf-like white dog called a Samoyed). At the pixel level, images of two Samoyeds in different poses and in different environments may be very different from each other, whereas two images of a Samoyed and a wolf in the same position and on similar backgrounds may be very similar to each other. A linear classifier, or any other 'shallow' classifier operating on raw pixels could not possibly distinguish the latter two, while putting the former two in the same category. This is why shallow classifiers require a good feature extractor that solves the selectivity–invariance dilemma — one that produces representations that are selective to the aspects of the image that are important for discrimination, but that are invariant to irrelevant aspects such as the pose of the animal. To make classifiers more powerful, one can use generic non-linear features, as with kernel methods20, but generic features such as those arising with the Gaussian kernel do not allow the learner to generalize well far from the training examples21. The conventional option is to hand design good feature extractors, which requires a considerable amount of engineering skill and domain expertise. But this can all be avoided if good features can be learned automatically using a general-purpose learning procedure. This is the key advantage of deep learning. A deep-learning architecture is a multilayer stack of simple modules, all (or most) of which are subject to learning, and many of which compute non-linear input–output mappings. Each module in the stack transforms its input to increase both the selectivity and the invariance of the representation. With multiple non-linear layers, say a depth of 5 to 20, a system can implement extremely intricate functions of its inputs that are simultaneously sensitive to minute details — distinguishing Samoyeds from white wolves — and insensitive to large irrelevant variations such as the background, pose, lighting and surrounding objects. From the earliest days of pattern recognition22, 23, the aim of researchers has been to replace hand-engineered features with trainable multilayer networks, but despite its simplicity, the solution was not widely understood until the mid 1980s. As it turns out, multilayer architectures can be trained by simple stochastic gradient descent. As long as the modules are relatively smooth functions of their inputs and of their internal weights, one can compute gradients using the backpropagation procedure. The idea that this could be done, and that it worked, was discovered independently by several different groups during the 1970s and 1980s24, 25, 26, 27. The backpropagation procedure to compute the gradient of an objective function with respect to the weights of a multilayer stack of modules is nothing more than a practical application of the chain rule for derivatives. The key insight is that the derivative (or gradient) of the objective with respect to the input of a module can be computed by working backwards from the gradient with respect to the output of that module (or the input of the subsequent module) (Fig. 1). The backpropagation equation can be applied repeatedly to propagate gradients through all modules, starting from the output at the top (where the network produces its prediction) all the way to the bottom (where the external input is fed). Once these gradients have been computed, it is straightforward to compute the gradients with respect to the weights of each module. Many applications of deep learning use feedforward neural network architectures (Fig. 1), which learn to map a fixed-size input (for example, an image) to a fixed-size output (for example, a probability for each of several categories). To go from one layer to the next, a set of units compute a weighted sum of their inputs from the previous layer and pass the result through a non-linear function. At present, the most popular non-linear function is the rectified linear unit (ReLU), which is simply the half-wave rectifier f(z) = max(z, 0). In past decades, neural nets used smoother non-linearities, such as tanh(z) or 1/(1 + exp(−z)), but the ReLU typically learns much faster in networks with many layers, allowing training of a deep supervised network without unsupervised pre-training28. Units that are not in the input or output layer are conventionally called hidden units. The hidden layers can be seen as distorting the input in a non-linear way so that categories become linearly separable by the last layer (Fig. 1). In the late 1990s, neural nets and backpropagation were largely forsaken by the machine-learning community and ignored by the computer-vision and speech-recognition communities. It was widely thought that learning useful, multistage, feature extractors with little prior knowledge was infeasible. In particular, it was commonly thought that simple gradient descent would get trapped in poor local minima — weight configurations for which no small change would reduce the average error. In practice, poor local minima are rarely a problem with large networks. Regardless of the initial conditions, the system nearly always reaches solutions of very similar quality. Recent theoretical and empirical results strongly suggest that local minima are not a serious issue in general. Instead, the landscape is packed with a combinatorially large number of saddle points where the gradient is zero, and the surface curves up in most dimensions and curves down in the remainder29, 30. The analysis seems to show that saddle points with only a few downward curving directions are present in very large numbers, but almost all of them have very similar values of the objective function. Hence, it does not much matter which of these saddle points the algorithm gets stuck at. Interest in deep feedforward networks was revived around 2006 (refs 31,32,33,34) by a group of researchers brought together by the Canadian Institute for Advanced Research (CIFAR). The researchers introduced unsupervised learning procedures that could create layers of feature detectors without requiring labelled data. The objective in learning each layer of feature detectors was to be able to reconstruct or model the activities of feature detectors (or raw inputs) in the layer below. By 'pre-training' several layers of progressively more complex feature detectors using this reconstruction objective, the weights of a deep network could be initialized to sensible values. A final layer of output units could then be added to the top of the network and the whole deep system could be fine-tuned using standard backpropagation33, 34, 35. This worked remarkably well for recognizing handwritten digits or for detecting pedestrians, especially when the amount of labelled data was very limited36. The first major application of this pre-training approach was in speech recognition, and it was made possible by the advent of fast graphics processing units (GPUs) that were convenient to program37 and allowed researchers to train networks 10 or 20 times faster. In 2009, the approach was used to map short temporal windows of coefficients extracted from a sound wave to a set of probabilities for the various fragments of speech that might be represented by the frame in the centre of the window. It achieved record-breaking results on a standard speech recognition benchmark that used a small vocabulary38 and was quickly developed to give record-breaking results on a large vocabulary task39. By 2012, versions of the deep net from 2009 were being developed by many of the major speech groups6 and were already being deployed in Android phones. For smaller data sets, unsupervised pre-training helps to prevent overfitting40, leading to significantly better generalization when the number of labelled examples is small, or in a transfer setting where we have lots of examples for some 'source' tasks but very few for some 'target' tasks. Once deep learning had been rehabilitated, it turned out that the pre-training stage was only needed for small data sets. There was, however, one particular type of deep, feedforward network that was much easier to train and generalized much better than networks with full connectivity between adjacent layers. This was the convolutional neural network (ConvNet)41, 42. It achieved many practical successes during the period when neural networks were out of favour and it has recently been widely adopted by the computer-vision community. ConvNets are designed to process data that come in the form of multiple arrays, for example a colour image composed of three 2D arrays containing pixel intensities in the three colour channels. Many data modalities are in the form of multiple arrays: 1D for signals and sequences, including language; 2D for images or audio spectrograms; and 3D for video or volumetric images. There are four key ideas behind ConvNets that take advantage of the properties of natural signals: local connections, shared weights, pooling and the use of many layers. The architecture of a typical ConvNet (Fig. 2) is structured as a series of stages. The first few stages are composed of two types of layers: convolutional layers and pooling layers. Units in a convolutional layer are organized in feature maps, within which each unit is connected to local patches in the feature maps of the previous layer through a set of weights called a filter bank. The result of this local weighted sum is then passed through a non-linearity such as a ReLU. All units in a feature map share the same filter bank. Different feature maps in a layer use different filter banks. The reason for this architecture is twofold. First, in array data such as images, local groups of values are often highly correlated, forming distinctive local motifs that are easily detected. Second, the local statistics of images and other signals are invariant to location. In other words, if a motif can appear in one part of the image, it could appear anywhere, hence the idea of units at different locations sharing the same weights and detecting the same pattern in different parts of the array. Mathematically, the filtering operation performed by a feature map is a discrete convolution, hence the name. Although the role of the convolutional layer is to detect local conjunctions of features from the previous layer, the role of the pooling layer is to merge semantically similar features into one. Because the relative positions of the features forming a motif can vary somewhat, reliably detecting the motif can be done by coarse-graining the position of each feature. A typical pooling unit computes the maximum of a local patch of units in one feature map (or in a few feature maps). Neighbouring pooling units take input from patches that are shifted by more than one row or column, thereby reducing the dimension of the representation and creating an invariance to small shifts and distortions. Two or three stages of convolution, non-linearity and pooling are stacked, followed by more convolutional and fully-connected layers. Backpropagating gradients through a ConvNet is as simple as through a regular deep network, allowing all the weights in all the filter banks to be trained. Deep neural networks exploit the property that many natural signals are compositional hierarchies, in which higher-level features are obtained by composing lower-level ones. In images, local combinations of edges form motifs, motifs assemble into parts, and parts form objects. Similar hierarchies exist in speech and text from sounds to phones, phonemes, syllables, words and sentences. The pooling allows representations to vary very little when elements in the previous layer vary in position and appearance. The convolutional and pooling layers in ConvNets are directly inspired by the classic notions of simple cells and complex cells in visual neuroscience43, and the overall architecture is reminiscent of the LGN–V1–V2–V4–IT hierarchy in the visual cortex ventral pathway44. When ConvNet models and monkeys are shown the same picture, the activations of high-level units in the ConvNet explains half of the variance of random sets of 160 neurons in the monkey's inferotemporal cortex45. ConvNets have their roots in the neocognitron46, the architecture of which was somewhat similar, but did not have an end-to-end supervised-learning algorithm such as backpropagation. A primitive 1D ConvNet called a time-delay neural net was used for the recognition of phonemes and simple words47, 48. There have been numerous applications of convolutional networks going back to the early 1990s, starting with time-delay neural networks for speech recognition47 and document reading42. The document reading system used a ConvNet trained jointly with a probabilistic model that implemented language constraints. By the late 1990s this system was reading over 10% of all the cheques in the United States. A number of ConvNet-based optical character recognition and handwriting recognition systems were later deployed by Microsoft49. ConvNets were also experimented with in the early 1990s for object detection in natural images, including faces and hands50, 51, and for face recognition52. Since the early 2000s, ConvNets have been applied with great success to the detection, segmentation and recognition of objects and regions in images. These were all tasks in which labelled data was relatively abundant, such as traffic sign recognition53, the segmentation of biological images54 particularly for connectomics55, and the detection of faces, text, pedestrians and human bodies in natural images36, 50, 51, 56, 57, 58. A major recent practical success of ConvNets is face recognition59. Importantly, images can be labelled at the pixel level, which will have applications in technology, including autonomous mobile robots and self-driving cars60, 61. Companies such as Mobileye and NVIDIA are using such ConvNet-based methods in their upcoming vision systems for cars. Other applications gaining importance involve natural language understanding14 and speech recognition7. Despite these successes, ConvNets were largely forsaken by the mainstream computer-vision and machine-learning communities until the ImageNet competition in 2012. When deep convolutional networks were applied to a data set of about a million images from the web that contained 1,000 different classes, they achieved spectacular results, almost halving the error rates of the best competing approaches1. This success came from the efficient use of GPUs, ReLUs, a new regularization technique called dropout62, and techniques to generate more training examples by deforming the existing ones. This success has brought about a revolution in computer vision; ConvNets are now the dominant approach for almost all recognition and detection tasks4, 58, 59, 63, 64, 65 and approach human performance on some tasks. A recent stunning demonstration combines ConvNets and recurrent net modules for the generation of image captions (Fig. 3). Recent ConvNet architectures have 10 to 20 layers of ReLUs, hundreds of millions of weights, and billions of connections between units. Whereas training such large networks could have taken weeks only two years ago, progress in hardware, software and algorithm parallelization have reduced training times to a few hours. The performance of ConvNet-based vision systems has caused most major technology companies, including Google, Facebook, Microsoft, IBM, Yahoo!, Twitter and Adobe, as well as a quickly growing number of start-ups to initiate research and development projects and to deploy ConvNet-based image understanding products and services. ConvNets are easily amenable to efficient hardware implementations in chips or field-programmable gate arrays66, 67. A number of companies such as NVIDIA, Mobileye, Intel, Qualcomm and Samsung are developing ConvNet chips to enable real-time vision applications in smartphones, cameras, robots and self-driving cars. Deep-learning theory shows that deep nets have two different exponential advantages over classic learning algorithms that do not use distributed representations21. Both of these advantages arise from the power of composition and depend on the underlying data-generating distribution having an appropriate componential structure40. First, learning distributed representations enable generalization to new combinations of the values of learned features beyond those seen during training (for example, 2n combinations are possible with n binary features)68, 69. Second, composing layers of representation in a deep net brings the potential for another exponential advantage70 (exponential in the depth). The hidden layers of a multilayer neural network learn to represent the network's inputs in a way that makes it easy to predict the target outputs. This is nicely demonstrated by training a multilayer neural network to predict the next word in a sequence from a local context of earlier words71. Each word in the context is presented to the network as a one-of-N vector, that is, one component has a value of 1 and the rest are 0. In the first layer, each word creates a different pattern of activations, or word vectors (Fig. 4). In a language model, the other layers of the network learn to convert the input word vectors into an output word vector for the predicted next word, which can be used to predict the probability for any word in the vocabulary to appear as the next word. The network learns word vectors that contain many active components each of which can be interpreted as a separate feature of the word, as was first demonstrated27 in the context of learning distributed representations for symbols. These semantic features were not explicitly present in the input. They were discovered by the learning procedure as a good way of factorizing the structured relationships between the input and output symbols into multiple 'micro-rules'. Learning word vectors turned out to also work very well when the word sequences come from a large corpus of real text and the individual micro-rules are unreliable71. When trained to predict the next word in a news story, for example, the learned word vectors for Tuesday and Wednesday are very similar, as are the word vectors for Sweden and Norway. Such representations are called distributed representations because their elements (the features) are not mutually exclusive and their many configurations correspond to the variations seen in the observed data. These word vectors are composed of learned features that were not determined ahead of time by experts, but automatically discovered by the neural network. Vector representations of words learned from text are now very widely used in natural language applications14, 17, 72, 73, 74, 75, 76. The issue of representation lies at the heart of the debate between the logic-inspired and the neural-network-inspired paradigms for cognition. In the logic-inspired paradigm, an instance of a symbol is something for which the only property is that it is either identical or non-identical to other symbol instances. It has no internal structure that is relevant to its use; and to reason with symbols, they must be bound to the variables in judiciously chosen rules of inference. By contrast, neural networks just use big activity vectors, big weight matrices and scalar non-linearities to perform the type of fast 'intuitive' inference that underpins effortless commonsense reasoning. Before the introduction of neural language models71, the standard approach to statistical modelling of language did not exploit distributed representations: it was based on counting frequencies of occurrences of short symbol sequences of length up to N (called N-grams). The number of possible N-grams is on the order of VN, where V is the vocabulary size, so taking into account a context of more than a handful of words would require very large training corpora. N-grams treat each word as an atomic unit, so they cannot generalize across semantically related sequences of words, whereas neural language models can because they associate each word with a vector of real valued features, and semantically related words end up close to each other in that vector space (Fig. 4). When backpropagation was first introduced, its most exciting use was for training recurrent neural networks (RNNs). For tasks that involve sequential inputs, such as speech and language, it is often better to use RNNs (Fig. 5). RNNs process an input sequence one element at a time, maintaining in their hidden units a 'state vector' that implicitly contains information about the history of all the past elements of the sequence. When we consider the outputs of the hidden units at different discrete time steps as if they were the outputs of different neurons in a deep multilayer network (Fig. 5, right), it becomes clear how we can apply backpropagation to train RNNs. RNNs are very powerful dynamic systems, but training them has proved to be problematic because the backpropagated gradients either grow or shrink at each time step, so over many time steps they typically explode or vanish77, 78. Thanks to advances in their architecture79, 80 and ways of training them81, 82, RNNs have been found to be very good at predicting the next character in the text83 or the next word in a sequence75, but they can also be used for more complex tasks. For example, after reading an English sentence one word at a time, an English 'encoder' network can be trained so that the final state vector of its hidden units is a good representation of the thought expressed by the sentence. This thought vector can then be used as the initial hidden state of (or as extra input to) a jointly trained French 'decoder' network, which outputs a probability distribution for the first word of the French translation. If a particular first word is chosen from this distribution and provided as input to the decoder network it will then output a probability distribution for the second word of the translation and so on until a full stop is chosen17, 72, 76. Overall, this process generates sequences of French words according to a probability distribution that depends on the English sentence. This rather naive way of performing machine translation has quickly become competitive with the state-of-the-art, and this raises serious doubts about whether understanding a sentence requires anything like the internal symbolic expressions that are manipulated by using inference rules. It is more compatible with the view that everyday reasoning involves many simultaneous analogies that each contribute plausibility to a conclusion84, 85. Instead of translating the meaning of a French sentence into an English sentence, one can learn to 'translate' the meaning of an image into an English sentence (Fig. 3). The encoder here is a deep ConvNet that converts the pixels into an activity vector in its last hidden layer. The decoder is an RNN similar to the ones used for machine translation and neural language modelling. There has been a surge of interest in such systems recently (see examples mentioned in ref. 86). RNNs, once unfolded in time (Fig. 5), can be seen as very deep feedforward networks in which all the layers share the same weights. Although their main purpose is to learn long-term dependencies, theoretical and empirical evidence shows that it is difficult to learn to store information for very long78. To correct for that, one idea is to augment the network with an explicit memory. The first proposal of this kind is the long short-term memory (LSTM) networks that use special hidden units, the natural behaviour of which is to remember inputs for a long time79. A special unit called the memory cell acts like an accumulator or a gated leaky neuron: it has a connection to itself at the next time step that has a weight of one, so it copies its own real-valued state and accumulates the external signal, but this self-connection is multiplicatively gated by another unit that learns to decide when to clear the content of the memory. LSTM networks have subsequently proved to be more effective than conventional RNNs, especially when they have several layers for each time step87, enabling an entire speech recognition system that goes all the way from acoustics to the sequence of characters in the transcription. LSTM networks or related forms of gated units are also currently used for the encoder and decoder networks that perform so well at machine translation17, 72, 76. Over the past year, several authors have made different proposals to augment RNNs with a memory module. Proposals include the Neural Turing Machine in which the network is augmented by a 'tape-like' memory that the RNN can choose to read from or write to88, and memory networks, in which a regular network is augmented by a kind of associative memory89. Memory networks have yielded excellent performance on standard question-answering benchmarks. The memory is used to remember the story about which the network is later asked to answer questions. Beyond simple memorization, neural Turing machines and memory networks are being used for tasks that would normally require reasoning and symbol manipulation. Neural Turing machines can be taught 'algorithms'. Among other things, they can learn to output a sorted list of symbols when their input consists of an unsorted sequence in which each symbol is accompanied by a real value that indicates its priority in the list88. Memory networks can be trained to keep track of the state of the world in a setting similar to a text adventure game and after reading a story, they can answer questions that require complex inference90. In one test example, the network is shown a 15-sentence version of the The Lord of the Rings and correctly answers questions such as “where is Frodo now?”89. Unsupervised learning91, 92, 93, 94, 95, 96, 97, 98 had a catalytic effect in reviving interest in deep learning, but has since been overshadowed by the successes of purely supervised learning. Although we have not focused on it in this Review, we expect unsupervised learning to become far more important in the longer term. Human and animal learning is largely unsupervised: we discover the structure of the world by observing it, not by being told the name of every object. Human vision is an active process that sequentially samples the optic array in an intelligent, task-specific way using a small, high-resolution fovea with a large, low-resolution surround. We expect much of the future progress in vision to come from systems that are trained end-to-end and combine ConvNets with RNNs that use reinforcement learning to decide where to look. Systems combining deep learning and reinforcement learning are in their infancy, but they already outperform passive vision systems99 at classification tasks and produce impressive results in learning to play many different video games100. Natural language understanding is another area in which deep learning is poised to make a large impact over the next few years. We expect systems that use RNNs to understand sentences or whole documents will become much better when they learn strategies for selectively attending to one part at a time76, 86. Ultimately, major progress in artificial intelligence will come about through systems that combine representation learning with complex reasoning. Although deep learning and simple reasoning have been used for speech and handwriting recognition for a long time, new paradigms are needed to replace rule-based manipulation of symbolic expressions by operations on large vectors101. Download references The authors would like to thank the Natural Sciences and Engineering Research Council of Canada, the Canadian Institute For Advanced Research (CIFAR), the National Science Foundation and Office of Naval Research for support. Y.L. and Y.B. are CIFAR fellows. Reprints and permissions information is available at www.nature.com/reprints.
Learning to paraphrase: an unsupervised approach using multiple-sequence alignment We address the text-to-text generation problem of sentence-level paraphrasing --- a phenomenon distinct from and more difficult than word- or phrase-level paraphrasing. Our approach applies multiple-sequence alignment to sentences gathered from unannotated comparable corpora: it learns a set of paraphrasing patterns represented by word lattice pairs and automatically determines how to apply these patterns to rewrite new sentences. The results of our evaluation experiments show that the system derives accurate paraphrases, outperforming baseline systems.
Surrogate-assisted evolutionary computation: Recent advances and future challenges Surrogate-assisted, or meta-model based evolutionary computation uses efficient computational models, often known as surrogates or meta-models, for approximating the fitness function in evolutionary algorithms. Research on surrogate-assisted evolutionary computation began over a decade ago and has received considerably increasing interest in recent years. Very interestingly, surrogate-assisted evolutionary computation has found successful applications not only in solving computationally expensive single- or multi-objective optimization problems, but also in addressing dynamic optimization problems, constrained optimization problems and multi-modal optimization problems. This paper provides a concise overview of the history and recent developments in surrogate-assisted evolutionary computation and suggests a few future trends in this research area.
Internet of Things: A Survey on Enabling Technologies, Protocols and Applications This paper provides an overview of the Internet of Things (IoT) with emphasis on enabling technologies, protocols and application issues. The IoT is enabled by the latest developments in RFID, smart sensors, communication technologies and Internet protocols. The basic premise is to have smart sensors collaborate directly without human involvement to deliver a new class of applications. The current revolution in Internet, mobile and machine-to-machine (M2M) technologies can be seen as the first phase of the IoT. In the coming years, the IoT is expected to bridge diverse technologies to enable new applications by connecting physical objects together in support of intelligent decision making. This paper starts by providing a horizontal overview of the IoT. Then, we give an overview of some technical details that pertain to the IoT enabling technologies, protocols and applications. Compared to other survey papers in the field, our objective is to provide a more thorough summary of the most relevant protocols and application issues to enable researchers and application developers to get up to speed quickly on how the different protocols fit together to deliver desired functionalities without having to go through RFCs and the standards specifications. We also provide an overview of some of the key IoT challenges presented in the recent literature and provide a summary of related research work. Moreover, we explore the relation between the IoT and other emerging technologies including big data analytics and cloud and fog computing. We also present the need for better horizontal integration among IoT services. Finally, we present detailed service use-cases to illustrate how the different protocols presented in the paper fit together to deliver desired IoT services.
A Tutorial on UAVs for Wireless Networks: Applications, Challenges, and Open Problems. The use of flying platforms such as unmanned aerial vehicles (UAVs), popularly known as drones, is rapidly growing in a wide range of wireless networking applications. In particular, with their inherent attributes such as mobility, flexibility, and adaptive altitude, UAVs admit several key potential applications in wireless systems. On the one hand, UAVs can be used as aerial base stations to enhance coverage, capacity, reliability, and energy efficiency of wireless networks. For instance, UAVs can be deployed to complement existing cellular systems by providing additional capacity to hotspot areas as well as to provide network coverage in emergency and public safety situations. On the other hand, UAVs can operate as flying mobile terminals within the cellular networks. In this paper, a comprehensive tutorial on the potential benefits and applications of UAVs in wireless communications is presented. Moreover, the important challenges and the fundamental tradeoffs in UAV-enabled wireless networks are thoroughly investigated. In particular, the key UAV challenges such as three-dimensional deployment, performance analysis, air-to-ground channel modeling, and energy efficiency are explored along with representative results. Then, fundamental open problems and potential research directions pertaining to wireless communications and networking with UAVs are introduced. To cope with the open research problems, various analytical frameworks and mathematical tools such as optimization theory, machine learning, stochastic geometry, transport theory, and game theory are described. The use of such tools for addressing unique UAV problems is also presented. In a nutshell, this tutorial provides key guidelines on how to analyze, optimize, and design UAV-based wireless communication systems.
Robust PCA for Subspace Estimation in User-Centric Cell-Free Wireless Networks We consider a scalable user-centric cell-free massive MIMO network with distributed remote radio units (RUs), enabling macrodiversity and joint processing. Due to the limited uplink (UL) pilot dimension, multiuser interference in the UL pilot transmission phase makes channel estimation a non-trivial problem. We make use of two types of UL pilot signals, sounding reference signal (SRS) and demodulation reference signal (DMRS) pilots, for the estimation of the channel subspace and its instantaneous realization, respectively. The SRS pilots are transmitted over multiple time slots and resource blocks according to a Latin squares based hopping scheme, which aims at averaging out the interference of different SRS co-pilot users. We propose a robust principle component analysis approach for channel subspace estimation from the SRS signal samples, employed at the RUs for each associated user. The estimated subspace is further used at the RUs for DMRS pilot decontamination and instantaneous channel estimation. We provide numerical simulations to compare the system performance using our subspace and channel estimation scheme with the cases of ideal partial subspace/channel knowledge and pilot matching channel estimation. The results show that a system with a properly designed SRS pilot hopping scheme can closely approximate the performance of a genie-aided system.
1.1
0.1
0.1
0.1
0.1
0.1
0.1
0
0
0
0
0
0
0
Multi-Sensor Scheduling for State Estimation with Event-Based, Stochastic Triggers. In networked systems, state estimation is hampered by communication limits. Past approaches, which consider scheduling sensors through deterministic event-triggers, reduce communication and maintain estimation quality. However, these approaches destroy the Gaussian property of the state, making it computationally intractable to obtain an exact minimum mean squared error estimate. We propose a stochastic event-triggered sensor schedule for state estimation which preserves the Gaussianity of the system, extending previous results from the single-sensor to the multi-sensor case.
Self-triggered coordination of robotic networks for optimal deployment This paper studies a deployment problem for a group of robots where individual agents operate with outdated information about each other's locations. Our objective is to understand to what extent outdated information is still useful and at which point it becomes essential to obtain new, up-to-date information. We propose a self-triggered coordination algorithm based on spatial partitioning techniques with uncertain information. We analyze its correctness in synchronous and asynchronous scenarios, and establish the same convergence guarantees that a synchronous algorithm with perfect information at all times would achieve. The technical approach combines computational geometry, set-valued stability analysis, and event-based systems.
Robust Control for Mobility and Wireless Communication in Cyber-Physical Systems With Application to Robot Teams. In this paper, a system architecture to provide end-to-end network connectivity for autonomous teams of robots is discussed. The core of the proposed system is a cyber-physical controller whose goal is to ensure network connectivity as robots move to accomplish their assigned tasks. Due to channel quality uncertainties inherent to wireless propagation, we adopt a stochastic model where achievable ...
Active Target Tracking with Self-Triggered Communications in Multi-Robot Teams. We study the problem of reducing the amount of communication in decentralized target tracking. We focus on the scenario, where a team of robots is allowed to move on the boundary of the environment. Their goal is to seek a formation so as to best track a target moving in the interior of the environment. The robots are capable of measuring distances to the target. Decentralized control strategies have been proposed in the past, which guarantees that the robots asymptotically converge to the optimal formation. However, existing methods require that the robots exchange information with their neighbors at all time steps. Instead, we focus on decentralized strategies to reduce the amount of communication among robots. We propose a self-triggered communication strategy that decides when a particular robot should seek up-to-date information from its neighbors and when it is safe to operate with possibly outdated information. We prove that this strategy converges asymptotically to the desired formation when the target is stationary. For the case of a mobile target, we use a decentralized Kalman filter with covariance intersection to share the beliefs of neighboring robots. We evaluate all the approaches through simulations and a proof-of-concept experiment. <italic xmlns:mml="http://www.w3.org/1998/Math/MathML" xmlns:xlink="http://www.w3.org/1999/xlink">Note to Practitioners</italic> —We study the problem of tracking a target using a team of coordinating robots. Target tracking problems are prevalent in a number of applications, such as co-robots, surveillance, and wildlife monitoring. Coordination between robots typically requires communication amongst them. Most multi-robot coordination algorithms implicitly assume that the robots can communicate at all time steps. Communication can be a considerable source of energy consumption, especially for small robots. Furthermore, communicating at all time steps may be redundant in many settings. With this as motivation, we propose an algorithm where the robots do not necessarily communicate at all times and instead choose specific triggering time instances to share information with their neighbors. Despite the limitation of limited communication, we show that the algorithm converges to the optimal configuration both in theory as well as in simulations.
Cooperative Localization Under Limited Connectivity In this article, we report two decentralized multiagent cooperative localization algorithms in which, to reduce the communication cost, interagent state estimate correlations are not maintained but accounted for implicitly. In our first algorithm, to guarantee filter consistency, we account for unknown interagent correlations via an upper bound on the joint covariance matrix of the agents. In the second method, we use an optimization framework to estimate the unknown interagent cross-covariance matrix. In our algorithms, each agent localizes itself in a global coordinate frame using a local filter driven by local dead reckoning and occasional absolute measurement updates, and opportunistically corrects its pose estimate whenever it can obtain relative measurements with respect to other mobile agents. To process any relative measurement, only the agent that has taken the measurement and the agent the measurement is taken from need to communicate with each other. Consequently, our algorithms are decentralized algorithms that do not impose restrictive network-wide connectivity condition. Moreover, we make no assumptions about the type of agents or relative measurements. We demonstrate our algorithms in simulation and a robotic experiment.
Communication in reactive multiagent robotic systems Multiple cooperating robots are able to complete many tasks more quickly and reliably than one robot alone. Communication between the robots can multiply their capabilities and effectiveness, but to what extent? In this research, the importance of communication in robotic societies is investigated through experiments on both simulated and real robots. Performance was measured for three different types of communication for three different tasks. The levels of communication are progressively more complex and potentially more expensive to implement. For some tasks, communication can significantly improve performance, but for others inter-agent communication is apparently unnecessary. In cases where communication helps, the lowest level of communication is almost as effective as the more complex type. The bulk of these results are derived from thousands of simulations run with randomly generated initial conditions. The simulation results help determine appropriate parameters for the reactive control system which was ported for tests on Denning mobile robots.
Footprints: history-rich tools for information foraging Inspired by Hill and Hollans original work [7], we have beendeveloping a theory of interaction history and building tools toapply this theory to navigation in a complex information space. Wehave built a series of tools - map, paths, annota- tions andsignposts - based on a physical-world navigation metaphor. Thesetools have been in use for over a year. Our user study involved acontrolled browse task and showed that users were able to get thesame amount of work done with significantly less effort.
Very Deep Convolutional Networks for Large-Scale Image Recognition. In this work we investigate the effect of the convolutional network depth on its accuracy in the large-scale image recognition setting. Our main contribution is a thorough evaluation of networks of increasing depth using an architecture with very small (3x3) convolution filters, which shows that a significant improvement on the prior-art configurations can be achieved by pushing the depth to 16-19 weight layers. These findings were the basis of our ImageNet Challenge 2014 submission, where our team secured the first and the second places in the localisation and classification tracks respectively. We also show that our representations generalise well to other datasets, where they achieve state-of-the-art results. We have made our two best-performing ConvNet models publicly available to facilitate further research on the use of deep visual representations in computer vision.
Chimp optimization algorithm. •A novel optimizer called Chimp Optimization Algorithm (ChOA) is proposed.•ChOA is inspired by individual intelligence and sexual motivation of chimps.•ChOA alleviates the problems of slow convergence rate and trapping in local optima.•The four main steps of Chimp hunting are implemented.
The contourlet transform: an efficient directional multiresolution image representation. The limitations of commonly used separable extensions of one-dimensional transforms, such as the Fourier and wavelet transforms, in capturing the geometry of image edges are well known. In this paper, we pursue a "true" two-dimensional transform that can capture the intrinsic geometrical structure that is key in visual information. The main challenge in exploring geometry in images comes from the discrete nature of the data. Thus, unlike other approaches, such as curvelets, that first develop a transform in the continuous domain and then discretize for sampled data, our approach starts with a discrete-domain construction and then studies its convergence to an expansion in the continuous domain. Specifically, we construct a discrete-domain multiresolution and multidirection expansion using nonseparable filter banks, in much the same way that wavelets were derived from filter banks. This construction results in a flexible multiresolution, local, and directional image expansion using contour segments, and, thus, it is named the contourlet transform. The discrete contourlet transform has a fast iterated filter bank algorithm that requires an order N operations for N-pixel images. Furthermore, we establish a precise link between the developed filter bank and the associated continuous-domain contourlet expansion via a directional multiresolution analysis framework. We show that with parabolic scaling and sufficient directional vanishing moments, contourlets achieve the optimal approximation rate for piecewise smooth functions with discontinuities along twice continuously differentiable curves. Finally, we show some numerical experiments demonstrating the potential of contourlets in several image processing applications. Index Terms-Contourlets, contours, filter banks, geometric image processing, multidirection, multiresolution, sparse representation, wavelets.
Cognitive Cars: A New Frontier for ADAS Research This paper provides a survey of recent works on cognitive cars with a focus on driver-oriented intelligent vehicle motion control. The main objective here is to clarify the goals and guidelines for future development in the area of advanced driver-assistance systems (ADASs). Two major research directions are investigated and discussed in detail: 1) stimuli–decisions–actions, which focuses on the driver side, and 2) perception enhancement–action-suggestion–function-delegation, which emphasizes the ADAS side. This paper addresses the important achievements and major difficulties of each direction and discusses how to combine the two directions into a single integrated system to obtain safety and comfort while driving. Other related topics, including driver training and infrastructure design, are also studied.
Finite-approximation-error-based discrete-time iterative adaptive dynamic programming. In this paper, a new iterative adaptive dynamic programming (ADP) algorithm is developed to solve optimal control problems for infinite horizon discrete-time nonlinear systems with finite approximation errors. First, a new generalized value iteration algorithm of ADP is developed to make the iterative performance index function converge to the solution of the Hamilton-Jacobi-Bellman equation. The ...
Robust Sparse Linear Discriminant Analysis Linear discriminant analysis (LDA) is a very popular supervised feature extraction method and has been extended to different variants. However, classical LDA has the following problems: 1) The obtained discriminant projection does not have good interpretability for features. 2) LDA is sensitive to noise. 3) LDA is sensitive to the selection of number of projection directions. In this paper, a novel feature extraction method called robust sparse linear discriminant analysis (RSLDA) is proposed to solve the above problems. Specifically, RSLDA adaptively selects the most discriminative features for discriminant analysis by introducing the l2;1 norm. An orthogonal matrix and a sparse matrix are also simultaneously introduced to guarantee that the extracted features can hold the main energy of the original data and enhance the robustness to noise, and thus RSLDA has the potential to perform better than other discriminant methods. Extensive experiments on six databases demonstrate that the proposed method achieves the competitive performance compared with other state-of-the-art feature extraction methods. Moreover, the proposed method is robust to the noisy data. IEEE
Ethical Considerations Of Applying Robots In Kindergarten Settings: Towards An Approach From A Macroperspective In child-robot interaction (cHRI) research, many studies pursue the goal to develop interactive systems that can be applied in everyday settings. For early education, increasingly, the setting of a kindergarten is targeted. However, when cHRI and research are brought into a kindergarten, a range of ethical and related procedural aspects have to be considered and dealt with. While ethical models elaborated within other human-robot interaction settings, e.g., assisted living contexts, can provide some important indicators for relevant issues, we argue that it is important to start developing a systematic approach to identify and tackle those ethical issues which rise with cHRI in kindergarten settings on a more global level and address the impact of the technology from a macroperspective beyond the effects on the individual. Based on our experience in conducting studies with children in general and pedagogical considerations on the role of the institution of kindergarten in specific, in this paper, we enfold some relevant aspects that have barely been addressed in an explicit way in current cHRI research. Four areas are analyzed and key ethical issues are identified in each area: (1) the institutional setting of a kindergarten, (2) children as a vulnerable group, (3) the caregivers' role, and (4) pedagogical concepts. With our considerations, we aim at (i) broadening the methodology of the current studies within the area of cHRI, (ii) revalidate it based on our comprehensive empirical experience with research in kindergarten settings, both laboratory and real-world contexts, and (iii) provide a framework for the development of a more systematic approach to address the ethical issues in cHRI research within kindergarten settings.
1.24
0.24
0.24
0.24
0.24
0.04
0
0
0
0
0
0
0
0
Parking Edge Computing: Parked Vehicle Assisted Task Offloading for Urban VANETs Vehicular edge computing has been a promising paradigm to offer low-latency and high reliability vehicular services for users. Nevertheless, for compute-intensive vehicle applications, most previous researches cannot perform them efficiently due to both the inadequate of infrastructure construction and the computing resource bottleneck of the edge server. Motivated by the fact that there is a larg...
A Survey of Ant Colony Optimization Based Routing Protocols for Mobile Ad Hoc Networks. Developing highly efficient routing protocols for Mobile Ad hoc NETworks (MANETs) is a challenging task. In order to fulfill multiple routing requirements, such as low packet delay, high packet delivery rate, and effective adaptation to network topology changes with low control overhead, and so on, new ways to approximate solutions to the known NP-hard optimization problem of routing in MANETs have to be investigated. Swarm intelligence (SI)-inspired algorithms have attracted a lot of attention, because they can offer possible optimized solutions ensuring high robustness, flexibility, and low cost. Moreover, they can solve large-scale sophisticated problems without a centralized control entity. A successful example in the SI field is the ant colony optimization (ACO) meta-heuristic. It presents a common framework for approximating solutions to NP-hard optimization problems. ACO has been successfully applied to balance the various routing related requirements in dynamic MANETs. This paper presents a comprehensive survey and comparison of various ACO-based routing protocols in MANETs. The main contributions of this survey include: 1) introducing the ACO principles as applied in routing protocols for MANETs; 2) classifying ACO-based routing approaches reviewed in this paper into five main categories; 3) surveying and comparing the selected routing protocols from the perspective of design and simulation parameters; and 4) discussing open issues and future possible design directions of ACO-based routing protocols.
A Microbial Inspired Routing Protocol for VANETs. We present a bio-inspired unicast routing protocol for vehicular ad hoc networks which uses the cellular attractor selection mechanism to select next hops. The proposed unicast routing protocol based on attractor selecting (URAS) is an opportunistic routing protocol, which is able to change itself adaptively to the complex and dynamic environment by routing feedback packets. We further employ a mu...
Improvement of GPSR Protocol in Vehicular Ad Hoc Network. In a vehicular ad hoc network (VANET), vehicles always move in high-speed which may cause the network topology changes frequently. This is challenging for routing protocols of VANET. Greedy Perimeter Stateless Routing (GPSR) is a representative routing protocol of VANET. However, when constructs routing path, GPSR selects the next hop node which is very easily out of the communication range in greedy forwarding, and builds the path with redundancy in perimeter forwarding. To solve the above-mentioned problems, we proposed Maxduration-Minangle GPSR (MM-GPSR) routing protocol in this paper. In greedy forwarding of MM-GPSR, by defining cumulative communication duration to represent the stability of neighbor nodes, the neighbor node with the maximum cumulative communication duration will be selected as the next hop node. In perimeter forwarding of MM-GPSR when greedy forwarding fails, the concept of minimum angle is introduced as the criterion of the optimal next hop node. By taking the position of neighbor nodes into account and calculating angles formed between neighbors and the destination node, the neighbor node with minimum angle will be selected as the next hop node. By using NS-2 and VanetMobiSim, simulations demonstrate that compared with GPSR, MM-GPSR has obvious improvements in reducing the packet loss rate, decreasing the end-to-end delay and increasing the throughput, and is more suitable for VANET.
Secure Real-Time Traffic Data Aggregation With Batch Verification for Vehicular Cloud in VANETs The vehicular cloud provides many significant advantages to Vehicular ad-hoc Networks (VANETs), such as unlimited storage space, powerful computing capability and timely traffic services. Traffic data aggregation in the vehicular cloud, which can aggregate traffic data from vehicles for further processing and sharing, is very important. Incorrect traffic data feedback may affect traffic safety; therefore, the security of traffic data aggregation should be ensured. In this paper, by using the property of data recovery in the message recovery signature (MRS), we propose a secure real-time traffic data aggregation scheme for vehicular cloud in VANETs. In the proposed scheme, the validity of vehicles’ signatures are verified, and then the original traffic data is recovered from signatures. Moreover, the proposed scheme supports batch verification for multiple vehicles’ signatures. Due to advantages of the MRS, security features such as data confidentiality, privacy preservation and reply attack resistance are preserved. In addition, the comparison and simulation results indicate that the proposed scheme is superior in comparison to previous schemes with respect to the communication and computational cost.
A Survey Of Qos-Aware Routing Protocols For The Manet-Wsn Convergence Scenarios In Iot Networks Wireless Sensor Network (WSN) and Mobile Ad hoc Network (MANET) have attracted a special attention because they can serve as communication means in many areas such as healthcare, military, smart traffic and smart cities. Nowadays, as all devices can be connected to a network forming the Internet of Things (IoT), the integration of WSN, MANET and other networks into IoT is indispensable. We investigate the convergence of WSN and MANET in IoT and consider a fundamental problem, that is, how a converged (WSN-MANET) network provides quality of service (QoS) guarantees to rich multimedia applications. This is very important because the network performances of WSN and MANET are quite low, while multimedia applications always require quality of services at certain levels. In this work, we survey the QoS-guaranteed routing protocols for WSN-MANETs, that are proposed in IEEE Xplore Digital Library over the last decade. Then, basing on our findings, we suggest future open research directions.
Efficient and Secure Routing Protocol Based on Artificial Intelligence Algorithms With UAV-Assisted for Vehicular Ad Hoc Networks in Intelligent Transportation Systems Vehicular Ad hoc Networks (VANETs) that are considered as a subset of Mobile Ad hoc Networks (MANETs) can be applied in the field of transportation especially in Intelligent Transportation Systems (ITS). The routing process in these networks is a challenging task due to rapid topology changes, high vehicle mobility and frequent disconnection of links. Therefore, developing an efficient routing pro...
Wireless sensor network survey A wireless sensor network (WSN) has important applications such as remote environmental monitoring and target tracking. This has been enabled by the availability, particularly in recent years, of sensors that are smaller, cheaper, and intelligent. These sensors are equipped with wireless interfaces with which they can communicate with one another to form a network. The design of a WSN depends significantly on the application, and it must consider factors such as the environment, the application's design objectives, cost, hardware, and system constraints. The goal of our survey is to present a comprehensive review of the recent literature since the publication of [I.F. Akyildiz, W. Su, Y. Sankarasubramaniam, E. Cayirci, A survey on sensor networks, IEEE Communications Magazine, 2002]. Following a top-down approach, we give an overview of several new applications and then review the literature on various aspects of WSNs. We classify the problems into three different categories: (1) internal platform and underlying operating system, (2) communication protocol stack, and (3) network services, provisioning, and deployment. We review the major development in these three categories and outline new challenges.
Energy-Aware Task Offloading and Resource Allocation for Time-Sensitive Services in Mobile Edge Computing Systems Mobile Edge Computing (MEC) is a promising architecture to reduce the energy consumption of mobile devices and provide satisfactory quality-of-service to time-sensitive services. How to jointly optimize task offloading and resource allocation to minimize the energy consumption subject to the latency requirement remains an open problem, which motivates this paper. When the latency constraint is tak...
Symbolic model checking for real-time systems We describe finite-state programs over real-numbered time in a guarded-command language with real-valued clocks or, equivalently, as finite automata with real-valued clocks. Model checking answers the question which states of a real-time program satisfy a branching-time specification (given in an extension of CTL with clock variables). We develop an algorithm that computes this set of states symbolically as a fixpoint of a functional on state predicates, without constructing the state space. For this purpose, we introduce a μ-calculus on computation trees over real-numbered time. Unfortunately, many standard program properties, such as response for all nonzeno execution sequences (during which time diverges), cannot be characterized by fixpoints: we show that the expressiveness of the timed μ-calculus is incomparable to the expressiveness of timed CTL. Fortunately, this result does not impair the symbolic verification of "implementable" real-time programs-those whose safety constraints are machine-closed with respect to diverging time and whose fairness constraints are restricted to finite upper bounds on clock values. All timed CTL properties of such programs are shown to be computable as finitely approximable fixpoints in a simple decidable theory.
The industrial indoor channel: large-scale and temporal fading at 900, 2400, and 5200 MHz In this paper, large-scale fading and temporal fading characteristics of the industrial radio channel at 900, 2400, and 5200 MHz are determined. In contrast to measurements performed in houses and in office buildings, few attempts have been made until now to model propagation in industrial environments. In this paper, the industrial environment is categorized into different topographies. Industrial topographies are defined separately for large-scale and temporal fading, and their definition is based upon the specific physical characteristics of the local surroundings affecting both types of fading. Large-scale fading is well expressed by a one-slope path-loss model and excellent agreement with a lognormal distribution is obtained. Temporal fading is found to be Ricean and Ricean K-factors have been determined. Ricean K-factors are found to follow a lognormal distribution.
Stable fuzzy logic control of a general class of chaotic systems This paper proposes a new approach to the stable design of fuzzy logic control systems that deal with a general class of chaotic processes. The stable design is carried out on the basis of a stability analysis theorem, which employs Lyapunov's direct method and the separate stability analysis of each rule in the fuzzy logic controller (FLC). The stability analysis theorem offers sufficient conditions for the stability of a general class of chaotic processes controlled by Takagi---Sugeno---Kang FLCs. The approach suggested in this paper is advantageous because inserting a new rule requires the fulfillment of only one of the conditions of the stability analysis theorem. Two case studies concerning the fuzzy logic control of representative chaotic systems that belong to the general class of chaotic systems are included in order to illustrate our stable design approach. A set of simulation results is given to validate the theoretical results.
Survey of Fog Computing: Fundamental, Network Applications, and Research Challenges. Fog computing is an emerging paradigm that extends computation, communication, and storage facilities toward the edge of a network. Compared to traditional cloud computing, fog computing can support delay-sensitive service requests from end-users (EUs) with reduced energy consumption and low traffic congestion. Basically, fog networks are viewed as offloading to core computation and storage. Fog n...
Robot tutor and pupils’ educational ability: Teaching the times tables Research shows promising results of educational robots in language and STEM tasks. In language, more research is available, occasionally in view of individual differences in pupils’ educational ability levels, and learning seems to improve with more expressive robot behaviors. In STEM, variations in robots’ behaviors have been examined with inconclusive results and never while systematically investigating how differences in educational abilities match with different robot behaviors. We applied an autonomously tutoring robot (without tablet, partly WOz) in a 2 × 2 experiment of social vs. neutral behavior in above-average vs. below-average schoolchildren (N = 86; age 8–10 years) while rehearsing the multiplication tables on a one-to-one basis. The standard school test showed that on average, pupils significantly improved their performance even after 3 occasions of 5-min exercises. Beyond-average pupils profited most from a robot tutor, whereas those below average in multiplication benefited more from a robot that showed neutral rather than more social behavior.
1.24
0.24
0.24
0.24
0.24
0.24
0.06
0
0
0
0
0
0
0
An optimization routing protocol for FANETs. With the wide-ranging application of mobile ad hoc networks, flying ad hoc networks (FANETs) have received more and more attention from the industry. Routing technology is a key technology of ad hoc networks. The high-speed mobility of nodes poses a greater challenge to FANET routing technology. Based on the Dynamic Source Routing (DSR) protocol, the continuous Hopfield neural network is used to optimize the route to be adapted to the high-speed movement of the FANET node. In a simulation using NS3, the result shows that the optimized DSR protocol has greatly improved key indicators such as end-to-end average delay, throughput, and packet delivery ratio.
High delivery rate position-based routing algorithms for 3D ad hoc networks Position-based routing algorithms use the geographic position of the nodes in a network to make the forwarding decisions. Recent research in this field primarily addresses such routing algorithms in two dimensional (2D) space. However, in real applications, nodes may be distributed in three dimensional (3D) environments. In this paper, we propose several randomized position-based routing algorithms and their combination with restricted directional flooding-based algorithms for routing in 3D environments. The first group of algorithms AB3D are extensions of previous randomized routing algorithms from 2D space to 3D space. The second group ABLAR chooses m neighbors according to a space-partition heuristic and forwards the message to all these nodes. The third group T-ABLAR-T uses progress-based routing until a local minimum is reached. The algorithm then switches to ABLAR for one step after which the algorithm switches back to the progress-based algorithm again. The fourth group AB3D-ABLAR uses an algorithm from the AB3D group until a threshold is passed in terms of number of hops. The algorithm then switches to an ABLAR algorithm. The algorithms are evaluated and compared with current routing algorithms. The simulation results on unit disk graphs (UDG) show a significant improvement in delivery rate (up to 99%) and a large reduction of the traffic.
Three-Dimensional Position-Based Adaptive Real-Time Routing Protocol for wireless sensor networks Devices for wireless sensor networks (WSN) are limited by power, and thus, routing protocols should be designed with this constraint in mind. WSNs are used in three-dimensional (3D) scenarios such as the surface of sea or lands with different levels of height. This paper presents and evaluates the Three-Dimensional Position-Based Adaptive Real-Time Routing Protocol (3DPBARP) as a novel, real-time, position-based and energy-efficient routing protocol for WSNs. 3DPBARP is a lightweight protocol that reduces the number of nodes which receive the radio frequency (RF) signal using a novel parent forwarding region (PFR) algorithm. 3DPBARP as a Geographical Routing Protocol (GRP) reduces the number of forwarding nodes and thus the traffic and packet collision in the network. A series of performance evaluations through MATLAB and Omnet++ simulations show significant improvements in network performance parameters and total energy consumption over the 3D Position-Based Routing Protocol (3DPBRP) and Directed Flooding Routing Protocol (DFRP).
Novel unequal clustering routing protocol considering energy balancing based on network partition & distance for mobile education. In Wireless Sensor Networks (WSN) of mobile education (such as mobile learning), in order to keep a better and lower energy consumption, reduce the energy hole and prolong the network life cycle, we propose a novel unequal clustering routing protocol considering energy balancing based on network partition & distance (UCNPD, which means Unequal Clustering based on Network Partition & Distance) for WSN in this paper. In the design model of this protocol, we know that all the network node data reaches the base station (BS) through the nodes near the BS, and the nodes in this area will use more energy, so we define a ring area using the BS as the center to form a circle, then we partition the network area based on the distance from node to the BS. These parts of nodes are to build connection with the BS, and the others follow the optimized clustering routing service protocol which uses a timing mechanism to elect the cluster head. It reduces the energy consumption of cluster reconstruction. Furthermore, we build unequal clusters by setting different competitive radius, which is helpful for balancing the network energy consumption. For the selection of message route, we considered all the energy of cluster head, the distances to BS and the degrees of node to reduce and balance the energy consumption. Simulation results demonstrate that the protocol can efficiently decrease the speed of the nodes death, prolong the network lifetime, and balance the energy dissipation of all nodes.
Energy Aware Cluster-Based Routing in Flying Ad-Hoc Networks. Flying ad-hoc networks (FANETs) are a very vibrant research area nowadays. They have many military and civil applications. Limited battery energy and the high mobility of micro unmanned aerial vehicles (UAVs) represent their two main problems, i.e., short flight time and inefficient routing. In this paper, we try to address both of these problems by means of efficient clustering. First, we adjust the transmission power of the UAVs by anticipating their operational requirements. Optimal transmission range will have minimum packet loss ratio (PLR) and better link quality, which ultimately save the energy consumed during communication. Second, we use a variant of the K-Means Density clustering algorithm for selection of cluster heads. Optimal cluster heads enhance the cluster lifetime and reduce the routing overhead. The proposed model outperforms the state of the art artificial intelligence techniques such as Ant Colony Optimization-based clustering algorithm and Grey Wolf Optimization-based clustering algorithm. The performance of the proposed algorithm is evaluated in term of number of clusters, cluster building time, cluster lifetime and energy consumption.
Spatial Anti-jamming Scheme for Internet of Satellites Based on the Deep Reinforcement Learning and Stackelberg Game The anti-jamming communication of the heterogeneous Internet of Satellites (IoS) has drawn increasing attentions due to the smart jamming and high dynamics. This paper investigates a spatial anti-jamming scheme for IoS, with the aim of minimizing anti-jamming routing cost via Stackelberg game and reinforcement learning. Firstly, we formulate the routing anti-jamming problem as a hierarchical anti-...
Multicast Traffic Throughput Maximization Through Joint Dynamic Modulation And Coding Schemes Assignment, And Transmission Power Control In Wireless Sensor Networks The paper concerns multicast packet traffic throughput maximization in multi-hop wireless sensor networks with time division multiple access to radio channel. We assume that the modulation and coding schemes (MCSs) that are used by the (broadcasting) nodes as well as the transmission power of the nodes are adjustable. This leads to the main research question studied in this paper: to what extent traffic throughput can be increased by proper MCSs assignment and transmission power control (TPC) at the nodes? To answer this question, we introduce mixed-integer programming formulations for joint MCSs assignment and TPC optimization, together with a solution algorithm. Finally, we present a numerical study illustrating the considerations of the paper. The numerical results show a significant gain being achieved by proper MCSs assignment, which is further increased by applying TPC.
Rich Feature Hierarchies for Accurate Object Detection and Semantic Segmentation Object detection performance, as measured on the canonical PASCAL VOC dataset, has plateaued in the last few years. The best-performing methods are complex ensemble systems that typically combine multiple low-level image features with high-level context. In this paper, we propose a simple and scalable detection algorithm that improves mean average precision (mAP) by more than 30% relative to the previous best result on VOC 2012 -- achieving a mAP of 53.3%. Our approach combines two key insights: (1) one can apply high-capacity convolutional neural networks (CNNs) to bottom-up region proposals in order to localize and segment objects and (2) when labeled training data is scarce, supervised pre-training for an auxiliary task, followed by domain-specific fine-tuning, yields a significant performance boost. Since we combine region proposals with CNNs, we call our method R-CNN: Regions with CNN features. We also present experiments that provide insight into what the network learns, revealing a rich hierarchy of image features. Source code for the complete system is available at http://www.cs.berkeley.edu/~rbg/rcnn.
Federated Learning: Challenges, Methods, and Future Directions Federated learning involves training statistical models over remote devices or siloed data centers, such as mobile phones or hospitals, while keeping data localized. Training in heterogeneous and potentially massive networks introduces novel challenges that require a fundamental departure from standard approaches for large-scale machine learning, distributed optimization, and privacy-preserving data analysis. In this article, we discuss the unique characteristics and challenges of federated learning, provide a broad overview of current approaches, and outline several directions of future work that are relevant to a wide range of research communities.
Efficient algorithms for Web services selection with end-to-end QoS constraints Service-Oriented Architecture (SOA) provides a flexible framework for service composition. Using standard-based protocols (such as SOAP and WSDL), composite services can be constructed by integrating atomic services developed independently. Algorithms are needed to select service components with various QoS levels according to some application-dependent performance requirements. We design a broker-based architecture to facilitate the selection of QoS-based services. The objective of service selection is to maximize an application-specific utility function under the end-to-end QoS constraints. The problem is modeled in two ways: the combinatorial model and the graph model. The combinatorial model defines the problem as a multidimension multichoice 0-1 knapsack problem (MMKP). The graph model defines the problem as a multiconstraint optimal path (MCOP) problem. Efficient heuristic algorithms for service processes of different composition structures are presented in this article and their performances are studied by simulations. We also compare the pros and cons between the two models.
End-user programming architecture facilitates the uptake of robots in social therapies. This paper proposes an architecture that makes programming of robot behavior of an arbitrary complexity possible for end-users and shows the technical solutions in a way that is easy to understand and generalize to different situations. It aims to facilitate the uptake and actual use of robot technologies in therapies for training social skills to autistic children. However, the framework is easy to generalize for an arbitrary human–robot interaction application, where users with no technical background need to program robots, i.e. in various assistive robotics applications. We identified the main needs of end-user programming of robots as a basic prerequisite for the uptake of robots in assistive applications. These are reusability, modularity, affordances for natural interaction and the ease of use. After reviewing the shortcomings of the existing architectures, we developed an initial architecture according to these principles and embedded it in a robot platform. Further, we used a co-creation process to develop and concretize the architecture to facilitate solutions and create affordances for robot specialists and therapists. Several pilot tests showed that different user groups, including therapists with general computer skills and adolescents with autism could make simple training or general behavioral scenarios within 1 h, by connecting existing behavioral blocks and by typing textual robot commands for fine-tuning the behaviors. In addition, this paper explains the basic concepts behind the TiViPE based robot control platform, and gives guidelines for choosing the robot programming tool and designing end-user platforms for robots.
Bessel-Fourier moment-based robust image zero-watermarking iming to resist various signal processing operations and geometric transformations, this paper proposes a robust zero-watermarking algorithm based on a new image moment called Bessel-Fourier moment. First, image normalization is used for the invariance of translation and scaling, then the magnitudes of Bessel-Fourier moments of normalized image are computed, which have rotation invariance and are used to construct the feature image regarded as watermarking. Experimental results and analyses show that the proposed method has strong robustness to various attacks, such as blurring, JPEG compression, Gaussian noise, rotation, scaling, Stirmark and print_photocopy_scan. Compared to the congener zero-watermarking schemes and Zernike moment, the developed method has better performance.
Survey of Fog Computing: Fundamental, Network Applications, and Research Challenges. Fog computing is an emerging paradigm that extends computation, communication, and storage facilities toward the edge of a network. Compared to traditional cloud computing, fog computing can support delay-sensitive service requests from end-users (EUs) with reduced energy consumption and low traffic congestion. Basically, fog networks are viewed as offloading to core computation and storage. Fog n...
Analyzing Software Rejuvenation Techniques in a Virtualized System: Service Provider and User Views Virtualization technology has promoted the fast development and deployment of cloud computing, and is now becoming an enabler of Internet of Everything. Virtual machine monitor (VMM), playing a critical role in a virtualized system, is software and hence it suffers from software aging after a long continuous running as well as software crashes due to elusive faults. Software rejuvenation techniques can be adopted to reduce the impact of software aging. Although there existed analytical model-based approaches for evaluating software rejuvenation techniques, none analyzed both application service (AS) availability and job completion time in a virtualized system with live virtual machine (VM) migration. This paper aims to quantitatively analyze software rejuvenation techniques from service provider and user views in a virtualized system deploying VMM reboot and live VM migration techniques for rejuvenation, under the condition that all the aging time, failure time, VMM fixing time and live VM migration time follow general distributions. We construct an analytical model by using a semi-Markov process (SMP) and derive formulas for calculating AS availability and job completion time. By analytical experiments, we can obtain the optimal migration trigger intervals for achieving the approximate maximum AS availability and the approximate minimum job completion time, and then service providers can make decisions for maximizing the benefits of service providers and users by adjusting parameter values.
1.1
0.1
0.1
0.1
0.1
0.1
0.1
0
0
0
0
0
0
0
Network-wide Crowd Flow Prediction of Sydney Trains via Customized Online Non-negative Matrix Factorization. Crowd Flow Prediction (CFP) is one major challenge in the intelligent transportation systems of the Sydney Trains Network. However, most advanced CFP methods only focus on entrance and exit flows at the major stations or a few subway lines, neglecting Crowd Flow Distribution (CFD) forecasting problem across the entire city network. CFD prediction plays an irreplaceable role in metro management as a tool that can help authorities plan route schedules and avoid congestion. In this paper, we propose three online non-negative matrix factorization (ONMF) models. ONMF-AO incorporates an Average Optimization strategy that adapts to stable passenger flows. ONMF-MR captures the Most Recent trends to achieve better performance when sudden changes in crowd flow occur. The Hybrid model, ONMF-H, integrates both ONMF-AO and ONMF-MR to exploit the strengths of each model in different scenarios and enhance the models' applicability to real-world situations. Given a series of CFD snapshots, both models learn the latent attributes of the train stations and, therefore, are able to capture transition patterns from one timestamp to the next by combining historic guidance. Intensive experiments on a large-scale, real-world dataset containing transactional data demonstrate the superiority of our ONMF models.
Origin-Destination Matrix Prediction via Graph Convolution: a New Perspective of Passenger Demand Modeling Ride-hailing applications are becoming more and more popular for providing drivers and passengers with convenient ride services, especially in metropolises like Beijing or New York. To obtain the passengers' mobility patterns, the online platforms of ride services need to predict the number of passenger demands from one region to another in advance. We formulate this problem as an Origin-Destination Matrix Prediction (ODMP) problem. Though this problem is essential to large-scale providers of ride services for helping them make decisions and some providers have already put it forward in public, existing studies have not solved this problem well. One of the main reasons is that the ODMP problem is more challenging than the common demand prediction. Besides the number of demands in a region, it also requires the model to predict the destinations of them. In addition, data sparsity is a severe issue. To solve the problem effectively, we propose a unified model, Grid-Embedding based Multi-task Learning (GEML) which consists of two components focusing on spatial and temporal information respectively. The Grid-Embedding part is designed to model the spatial mobility patterns of passengers and neighboring relationships of different areas, the pre-weighted aggregator of which aims to sense the sparsity and range of data. The Multi-task Learning framework focuses on modeling temporal attributes and capturing several objectives of the ODMP problem. The evaluation of our model is conducted on real operational datasets from UCAR and Didi. The experimental results demonstrate the superiority of our GEML against the state-of-the-art approaches.
Forecasting holiday daily tourist flow based on seasonal support vector regression with adaptive genetic algorithm. •The model of support vector regression with adaptive genetic algorithm and the seasonal mechanism is proposed.•Parameters selection and seasonal adjustment should be carefully selected.•We focus on latest and representative holiday daily data in China.•Two experiments are used to prove the effect of the model.•The AGASSVR is superior to AGA-SVR and BPNN.
Bayesian Particle Tracking of Traffic Flows. We develop a Bayesian particle filter for tracking traffic flows that is capable of capturing non-linearities and discontinuities present in flow dynamics. Our model includes a hidden state variable that captures sudden regime shifts between traffic free flow, breakdown, and recovery. We develop an efficient particle learning algorithm for real time online inference of states and parameters. This ...
Deep and Embedded Learning Approach for Traffic Flow Prediction in Urban Informatics Traffic flow prediction has received extensive attention recently, since it is a key step to prevent and mitigate traffic congestion in urban areas. However, most previous studies on traffic flow prediction fail to capture fine-grained traffic information (like link-level traffic) and ignore the impacts from other factors, such as route structure and weather conditions. In this paper, we propose a deep and embedding learning approach (DELA) that can help to explicitly learn from fine-grained traffic information, route structure, and weather conditions. In particular, our DELA consists of an embedding component, a convolutional neural network (CNN) component and a long short-term memory (LSTM) component. The embedding component can capture the categorical feature information and identify correlated features. Meanwhile, the CNN component can learn the 2-D traffic flow data while the LSTM component has the benefits of maintaining a long-term memory of historical data. The integration of the three models together can improve the prediction accuracy of traffic flow. We conduct extensive experiments on realistic traffic flow dataset to evaluate the performance of our DELA and make comparison with other existing models. The experimental results show that the proposed DELA outperforms the existing methods in terms of prediction accuracy.
A General Framework for Unmet Demand Prediction in On-Demand Transport Services Emerging on-demand transport services, such as Uber and GoGoVan, usually face the dilemma of demand supply imbalance, meaning that the spatial distributions of orders and drivers are imbalanced. Due to such imbalance, much supply resource is wasted while a considerable amount of order demand cannot be met in time. To address this dilemma, knowing the unmet demand in the near future is of high importance for service providers because they can dispatch their vehicles in advance to alleviate the impending demand supply imbalance, we develop a general framework for predicting the unmet demand in future time slots. Under this framework, we first evaluate the predictability of unmet demand in on-demand transport services and find that unmet demand is highly predictable. Then, we extract both static and dynamic urban features relevant to unmet demand from data sets in multiple domains. Finally, multiple prediction models are trained to predict unmet demand by using the extracted features. As demonstrated via experiments, the proposed framework can predict unmet demand in on-demand transport services effectively and flexibly.
Forecasting Short-Term Passenger Flow: An Empirical Study on Shenzhen Metro Forecasting short-term traffic flow has been a critical topic in transportation research for decades, which aims to facilitate dynamic traffic control proactively by monitoring the present traffic and foreseeing its immediate future. In this paper, we focus on forecasting short-term passenger flow at subway stations by utilizing the data collected through an automatic fare collection (AFC) system along with various external factors, where passenger flow refers to the volume of arrivals at stations during a given period of time. Along this line, we propose a data-driven three-stage framework for short-term passenger flow forecasting, consisting of traffic data profiling, feature extraction, and predictive modeling. We investigate the effect of temporal and spatial features as well as external weather influence on passenger flow forecasting. Various forecasting models, including the time series model auto-regressive integrated moving average, linear regression, and support vector regression, are employed for evaluating the performance of the proposed framework. Moreover, using a real data set collected from the Shenzhen AFC system, we conduct extensive experiments for methods validation, feature evaluation, and data resolution demonstration.
Predicting Multi-step Citywide Passenger Demands Using Attention-based Neural Networks. Predicting passenger pickup/dropoff demands based on historical mobility trips has been of great importance towards better vehicle distribution for the emerging mobility-on-demand (MOD) services. Prior works focused on predicting next-step passenger demands at selected locations or hotspots. However, we argue that multi-step citywide passenger demands encapsulate both time-varying demand trends and global statuses, and hence are more beneficial to avoiding demand-service mismatching and developing effective vehicle distribution/scheduling strategies. In this paper, we propose an end-to-end deep neural network solution to the prediction task. We employ the encoder-decoder framework based on convolutional and ConvLSTM units to identify complex features that capture spatiotemporal influences and pickup-dropoff interactions on citywide passenger demands. A novel attention model is incorporated to emphasize the effects of latent citywide mobility regularities. We evaluate our proposed method using real-word mobility trips (taxis and bikes) and the experimental results show that our method achieves higher prediction accuracy than the adaptations of the state-of-the-art approaches.
Model-based periodic event-triggered control for linear systems Periodic event-triggered control (PETC) is a control strategy that combines ideas from conventional periodic sampled-data control and event-triggered control. By communicating periodically sampled sensor and controller data only when needed to guarantee stability or performance properties, PETC is capable of reducing the number of transmissions significantly, while still retaining a satisfactory closed-loop behavior. In this paper, we will study observer-based controllers for linear systems and propose advanced event-triggering mechanisms (ETMs) that will reduce communication in both the sensor-to-controller channels and the controller-to-actuator channels. By exploiting model-based computations, the new classes of ETMs will outperform existing ETMs in the literature. To model and analyze the proposed classes of ETMs, we present two frameworks based on perturbed linear and piecewise linear systems, leading to conditions for global exponential stability and @?"2-gain performance of the resulting closed-loop systems in terms of linear matrix inequalities. The proposed analysis frameworks can be used to make tradeoffs between the network utilization on the one hand and the performance in terms of @?"2-gains on the other. In addition, we will show that the closed-loop performance realized by an observer-based controller, implemented in a conventional periodic time-triggered fashion, can be recovered arbitrarily closely by a PETC implementation. This provides a justification for emulation-based design. Next to centralized model-based ETMs, we will also provide a decentralized setup suitable for large-scale systems, where sensors and actuators are physically distributed over a wide area. The improvements realized by the proposed model-based ETMs will be demonstrated using numerical examples.
Affective social robots For human-robot interaction to proceed in a smooth, natural manner, robots must adhere to human social norms. One such human convention is the use of expressive moods and emotions as an integral part of social interaction. Such expressions are used to convey messages such as ''I'm happy to see you'' or ''I want to be comforted,'' and people's long-term relationships depend heavily on shared emotional experiences. Thus, we have developed an affective model for social robots. This generative model attempts to create natural, human-like affect and includes distinctions between immediate emotional responses, the overall mood of the robot, and long-term attitudes toward each visitor to the robot, with a focus on developing long-term human-robot relationships. This paper presents the general affect model as well as particular details of our implementation of the model on one robot, the Roboceptionist. In addition, we present findings from two studies that demonstrate the model's potential.
Rich Models for Steganalysis of Digital Images We describe a novel general strategy for building steganography detectors for digital images. The process starts with assembling a rich model of the noise component as a union of many diverse submodels formed by joint distributions of neighboring samples from quantized image noise residuals obtained using linear and nonlinear high-pass filters. In contrast to previous approaches, we make the model assembly a part of the training process driven by samples drawn from the corresponding cover- and stego-sources. Ensemble classifiers are used to assemble the model as well as the final steganalyzer due to their low computational complexity and ability to efficiently work with high-dimensional feature spaces and large training sets. We demonstrate the proposed framework on three steganographic algorithms designed to hide messages in images represented in the spatial domain: HUGO, edge-adaptive algorithm by Luo , and optimally coded ternary $\\pm {\\hbox{1}}$ embedding. For each algorithm, we apply a simple submodel-selection technique to increase the detection accuracy per model dimensionality and show how the detection saturates with increasing complexity of the rich model. By observing the differences between how different submodels engage in detection, an interesting interplay between the embedding and detection is revealed. Steganalysis built around rich image models combined with ensemble classifiers is a promising direction towards automatizing steganalysis for a wide spectrum of steganographic schemes.
Heterogeneous ensemble for feature drifts in data streams The nature of data streams requires classification algorithms to be real-time, efficient, and able to cope with high-dimensional data that are continuously arriving. It is a known fact that in high-dimensional datasets, not all features are critical for training a classifier. To improve the performance of data stream classification, we propose an algorithm called HEFT-Stream (H eterogeneous E nsemble with F eature drifT for Data Streams ) that incorporates feature selection into a heterogeneous ensemble to adapt to different types of concept drifts. As an example of the proposed framework, we first modify the FCBF [13] algorithm so that it dynamically update the relevant feature subsets for data streams. Next, a heterogeneous ensemble is constructed based on different online classifiers, including Online Naive Bayes and CVFDT [5]. Empirical results show that our ensemble classifier outperforms state-of-the-art ensemble classifiers (AWE [15] and OnlineBagging [21]) in terms of accuracy, speed, and scalability. The success of HEFT-Stream opens new research directions in understanding the relationship between feature selection techniques and ensemble learning to achieve better classification performance.
Orientation-aware RFID tracking with centimeter-level accuracy. RFID tracking attracts a lot of research efforts in recent years. Most of the existing approaches, however, adopt an orientation-oblivious model. When tracking a target whose orientation changes, those approaches suffer from serious accuracy degradation. In order to achieve target tracking with pervasive applicability in various scenarios, we in this paper propose OmniTrack, an orientation-aware RFID tracking approach. Our study discovers the linear relationship between the tag orientation and the phase change of the backscattered signals. Based on this finding, we propose an orientation-aware phase model to explicitly quantify the respective impact of the read-tag distance and the tag's orientation. OmniTrack addresses practical challenges in tracking the location and orientation of a mobile tag. Our experimental results demonstrate that OmniTrack achieves centimeter-level location accuracy and has significant advantages in tracking targets with varing orientations, compared to the state-of-the-art approaches.
Robot tutor and pupils’ educational ability: Teaching the times tables Research shows promising results of educational robots in language and STEM tasks. In language, more research is available, occasionally in view of individual differences in pupils’ educational ability levels, and learning seems to improve with more expressive robot behaviors. In STEM, variations in robots’ behaviors have been examined with inconclusive results and never while systematically investigating how differences in educational abilities match with different robot behaviors. We applied an autonomously tutoring robot (without tablet, partly WOz) in a 2 × 2 experiment of social vs. neutral behavior in above-average vs. below-average schoolchildren (N = 86; age 8–10 years) while rehearsing the multiplication tables on a one-to-one basis. The standard school test showed that on average, pupils significantly improved their performance even after 3 occasions of 5-min exercises. Beyond-average pupils profited most from a robot tutor, whereas those below average in multiplication benefited more from a robot that showed neutral rather than more social behavior.
1.125
0.125
0.12
0.12
0.12
0.12
0.06
0.005
0
0
0
0
0
0
Deep Reinforcement Learning for Mobile Edge Caching: Review, New Features, and Open Issues. Mobile edge caching is a promising technique to reduce network traffic and improve the quality of experience of mobile users. However, mobile edge caching is a challenging decision making problem with unknown future content popularity and complex network characteristics. In this article, we advocate the use of DRL to solve mobile edge caching problems by presenting an overview of recent works on m...
A MTC traffic generation and QCI priority-first scheduling algorithm over LTE As (M2M) Machine-To-Machine, communication continues to grow rapidly, a full study on overload control approach to manage the data and signaling of H2H traffic from massive MTC devices is required. In this paper, a new M2M resource-scheduling algorithm for Long Term Evolution (LTE) is proposed. It provides Quality of Service (QoS) guarantee to Guaranteed Bit Rate (GBR) services, we set priorities for the critical M2M services to guarantee the transportation of GBR services, which have high QoS needs. Additionally, we simulate and compare different methods and offer further observations on the solution design.
On Service Resilience in Cloud-Native 5G Mobile Systems. To cope with the tremendous growth in mobile data traffic on one hand, and the modest average revenue per user on the other hand, mobile operators have been exploring network virtualization and cloud computing technologies to build cost-efficient and elastic mobile networks and to have them offered as a cloud service. In such cloud-based mobile networks, ensuring service resilience is an important challenge to tackle. Indeed, high availability and service reliability are important requirements of carrier grade, but not necessarily intrinsic features of cloud computing. Building a system that requires the five nines reliability on a platform that may not always grant it is, therefore, a hurdle. Effectively, in carrier cloud, service resilience can be heavily impacted by a failure of any network function (NF) running on a virtual machine (VM). In this paper, we introduce a framework, along with efficient and proactive restoration mechanisms, to ensure service resilience in carrier cloud. As restoration of a NF failure impacts a potential number of users, adequate network overload control mechanisms are also proposed. A mathematical model is developed to evaluate the performance of the proposed mechanisms. The obtained results are encouraging and demonstrate that the proposed mechanisms efficiently achieve their design goals.
Methodology for the Design and Evaluation of Self-Healing LTE Networks. Self-healing networks aim to detect cells with service degradation, identify the fault cause of their problem, and execute compensation and repair actions. The development of this type of automatic system presents several challenges to be confronted. The first challenge is the scarce number of historically reported faults, which greatly complicates the evaluation of novel self-healing techniques. ...
Intelligence and Learning in O-RAN for Data-Driven NextG Cellular Networks Next generation (NextG) cellular networks will be natively cloud-based and built on programmable, virtualized, and disaggregated architectures. The separation of control functions from the hardware fabric and the introduction of standardized control interfaces will enable the definition of custom closed-control loops, which will ultimately enable embedded intelligence and real-time analytics, thus...
Resource Slicing in Virtual Wireless Networks: A Survey. New architectural and design approaches for radio access networks have appeared with the introduction of network virtualization in the wireless domain. One of these approaches splits the wireless network infrastructure into isolated virtual slices under their own management, requirements, and characteristics. Despite the advances in wireless virtualization, there are still many open issues regarding the resource allocation and isolation of wireless slices. Because of the dynamics and shared nature of the wireless medium, guaranteeing that the traffic on one slice will not affect the traffic on the others has proven to be difficult. In this paper, we focus on the detailed definition of the problem, discussing its challenges. We also provide a review of existing works that deal with the problem, analyzing how new trends such as software defined networking and network function virtualization can assist in the slicing. We will finally describe some research challenges on this topic.
Scheduling Policies for Federated Learning in Wireless Networks Motivated by the increasing computational capacity of wireless user equipments (UEs), e.g., smart phones, tablets, or vehicles, as well as the increasing concerns about sharing private data, a new machine learning model has emerged, namely federated learning (FL), that allows a decoupling of data acquisition and computation at the central unit. Unlike centralized learning taking place in a data center, FL usually operates in a wireless edge network where the communication medium is resource-constrained and unreliable. Due to limited bandwidth, only a portion of UEs can be scheduled for updates at each iteration. Due to the shared nature of the wireless medium, transmissions are subjected to interference and are not guaranteed. The performance of FL system in such a setting is not well understood. In this paper, an analytical model is developed to characterize the performance of FL in wireless networks. Particularly, tractable expressions are derived for the convergence rate of FL in a wireless setting, accounting for effects from both scheduling schemes and inter-cell interference. Using the developed analysis, the effectiveness of three different scheduling policies, i.e., random scheduling (RS), round robin (RR), and proportional fair (PF), are compared in terms of FL convergence rate. It is shown that running FL with PF outperforms RS and RR if the network is operating under a high signal-to-interference-plus-noise ratio (SINR) threshold, while RR is more preferable when the SINR threshold is low. Moreover, the FL convergence rate decreases rapidly as the SINR threshold increases, thus confirming the importance of compression and quantization of the update parameters. The analysis also reveals a trade-off between the number of scheduled UEs and subchannel bandwidth under a fixed amount of available spectrum.
Wireless Body Area Networks: A Survey Recent developments and technological advancements in wireless communication, MicroElectroMechanical Systems (MEMS) technology and integrated circuits has enabled low-power, intelligent, miniaturized, invasive/non-invasive micro and nano-technology sensor nodes strategically placed in or around the human body to be used in various applications, such as personal health monitoring. This exciting new area of research is called Wireless Body Area Networks (WBANs) and leverages the emerging IEEE 802.15.6 and IEEE 802.15.4j standards, specifically standardized for medical WBANs. The aim of WBANs is to simplify and improve speed, accuracy, and reliability of communication of sensors/actuators within, on, and in the immediate proximity of a human body. The vast scope of challenges associated with WBANs has led to numerous publications. In this paper, we survey the current state-of-art of WBANs based on the latest standards and publications. Open issues and challenges within each area are also explored as a source of inspiration towards future developments in WBANs.
LMM: latency-aware micro-service mashup in mobile edge computing environment Internet of Things (IoT) applications introduce a set of stringent requirements (e.g., low latency, high bandwidth) to network and computing paradigm. 5G networks are faced with great challenges for supporting IoT services. The centralized cloud computing paradigm also becomes inefficient for those stringent requirements. Only extending spectrum resources cannot solve the problem effectively. Mobile edge computing offers an IT service environment at the Radio Access Network edge and presents great opportunities for the development of IoT applications. With the capability to reduce latency and offer an improved user experience, mobile edge computing becomes a key technology toward 5G. To achieve abundant sharing, complex IoT applications have been implemented as a set of lightweight micro-services that are distributed among containers over the mobile edge network. How to produce the optimal collocation of suitable micro-service for an application in mobile edge computing environment is an important issue that should be addressed. To address this issue, we propose a latency-aware micro-service mashup approach in this paper. Firstly, the problem is formulated into an integer nonlinear programming. Then, we prove the NP-hardness of the problem by reducing it into the delay constrained least cost problem. Finally, we propose an approximation latency-aware micro-service mashup approach to solve the problem. Experiment results show that the proposed approach achieves a substantial reduction in network resource consumption while still ensuring the latency constraint.
Computer intrusion detection through EWMA for autocorrelated and uncorrelated data Reliability and quality of service from information systems has been threatened by cyber intrusions. To protect information systems from intrusions and thus assure reliability and quality of service, it is highly desirable to develop techniques that detect intrusions. Many intrusions manifest in anomalous changes in intensity of events occurring in information systems. In this study, we apply, tes...
Decentralized Plug-in Electric Vehicle Charging Selection Algorithm in Power Systems This paper uses a charging selection concept for plug-in electric vehicles (PEVs) to maximize user convenience levels while meeting predefined circuit-level demand limits. The optimal PEV-charging selection problem requires an exhaustive search for all possible combinations of PEVs in a power system, which cannot be solved for the practical number of PEVs. Inspired by the efficiency of the convex relaxation optimization tool in finding close-to-optimal results in huge search spaces, this paper proposes the application of the convex relaxation optimization method to solve the PEV-charging selection problem. Compared with the results of the uncontrolled case, the simulated results indicate that the proposed PEV-charging selection algorithm only slightly reduces user convenience levels, but significantly mitigates the impact of the PEV-charging on the power system. We also develop a distributed optimization algorithm to solve the PEV-charging selection problem in a decentralized manner, i.e., the binary charging decisions (charged or not charged) are made locally by each vehicle. Using the proposed distributed optimization algorithm, each vehicle is only required to report its power demand rather than report several of its private user state information, mitigating the security problems inherent in such problem. The proposed decentralized algorithm only requires low-speed communication capability, making it suitable for real-time implementation.
Placing Virtual Machines to Optimize Cloud Gaming Experience Optimizing cloud gaming experience is no easy task due to the complex tradeoff between gamer quality of experience (QoE) and provider net profit. We tackle the challenge and study an optimization problem to maximize the cloud gaming provider's total profit while achieving just-good-enough QoE. We conduct measurement studies to derive the QoE and performance models. We formulate and optimally solve the problem. The optimization problem has exponential running time, and we develop an efficient heuristic algorithm. We also present an alternative formulation and algorithms for closed cloud gaming services with dedicated infrastructures, where the profit is not a concern and overall gaming QoE needs to be maximized. We present a prototype system and testbed using off-the-shelf virtualization software, to demonstrate the practicality and efficiency of our algorithms. Our experience on realizing the testbed sheds some lights on how cloud gaming providers may build up their own profitable services. Last, we conduct extensive trace-driven simulations to evaluate our proposed algorithms. The simulation results show that the proposed heuristic algorithms: (i) produce close-to-optimal solutions, (ii) scale to large cloud gaming services with 20,000 servers and 40,000 gamers, and (iii) outperform the state-of-the-art placement heuristic, e.g., by up to 3.5 times in terms of net profits.
Flymap: Interacting With Maps Projected From A Drone Interactive maps have become ubiquitous in our daily lives, helping us reach destinations and discovering our surroundings. Yet, designing map interactions is not straightforward and depends on the device being used. As mobile devices evolve and become independent from users, such as with robots and drones, how will we interact with the maps they provide? We propose FlyMap as a novel user experience for drone-based interactive maps. We designed and developed three interaction techniques for FlyMap's usage scenarios. In a comprehensive indoor study (N = 16), we show the strengths and weaknesses of two techniques on users' cognition, task load, and satisfaction. FlyMap was then pilot tested with the third technique outdoors in real world conditions with four groups of participants (N = 13). We show that FlyMap's interactivity is exciting to users and opens the space for more direct interactions with drones.
Learning Feature Recovery Transformer for Occluded Person Re-Identification One major issue that challenges person re-identification (Re-ID) is the ubiquitous occlusion over the captured persons. There are two main challenges for the occluded person Re-ID problem, i.e., the interference of noise during feature matching and the loss of pedestrian information brought by the occlusions. In this paper, we propose a new approach called Feature Recovery Transformer (FRT) to address the two challenges simultaneously, which mainly consists of visibility graph matching and feature recovery transformer. To reduce the interference of the noise during feature matching, we mainly focus on visible regions that appear in both images and develop a visibility graph to calculate the similarity. In terms of the second challenge, based on the developed graph similarity, for each query image, we propose a recovery transformer that exploits the feature sets of its k-nearest neighbors in the gallery to recover the complete features. Extensive experiments across different person Re-ID datasets, including occluded, partial and holistic datasets, demonstrate the effectiveness of FRT. Specifically, FRT significantly outperforms state-of-the-art results by at least 6.2% Rank- 1 accuracy and 7.2% mAP scores on the challenging Occluded-Duke dataset.
1.1
0.1
0.1
0.1
0.1
0.05
0.005
0
0
0
0
0
0
0
Receding Horizon Temporal Logic Planning We present a methodology for automatic synthesis of embedded control software that incorporates a class of linear temporal logic (LTL) specifications sufficient to describe a wide range of properties including safety, stability, progress, obligation, response and guarantee. To alleviate the associated computational complexity of LTL synthesis, we propose a receding horizon framework that effectively reduces the synthesis problem into a set of smaller problems. The proposed control structure consists of a goal generator, a trajectory planner, and a continuous controller. The goal generator reduces the trajectory generation problem into a sequence of smaller problems of short horizon while preserving the desired system-level temporal properties. Subsequently, in each iteration, the trajectory planner solves the corresponding short-horizon problem with the currently observed state as the initial state and generates a feasible trajectory to be implemented by the continuous controller. Based on the simulation property, we show that the composition of the goal generator, trajectory planner and continuous controller and the corresponding receding horizon framework guarantee the correctness of the system with respect to its specification regardless of the environment in which the system operates. In addition, we present a response mechanism to handle failures that may occur due to a mismatch between the actual system and its model. The effectiveness of the proposed technique is demonstrated through an example of an autonomous vehicle navigating an urban environment. This example also illustrates that the system is not only robust with respect to exogenous disturbances but is also capable of properly handling violation of the environment assumption that is explicitly stated as part of the system specification.
Using Ontology-Based Traffic Models for More Efficient Decision Making of Autonomous Vehicles The paper describes how a high-level abstract world model can be used to support the decision-making process of an autonomous driving system. The approach uses a hierarchical world model and distinguishes between a low-level model for the trajectory planning and a high-level model for solving the traffic coordination problem. The abstract world model used in the CyberCars-2 project is presented. It is based on a topological lane segmentation and introduces relations to represent the semantic context of the traffic scenario. This makes it much easier to realize a consistent and complete driving control system, and to analyze, evaluate and simulate such a system.
Ontology-based methods for enhancing autonomous vehicle path planning We report the results of a first implementation demonstrating the use of an ontology to support reasoning about obstacles to improve the capabilities and performance of on-board route planning for autonomous vehicles. This is part of an overall effort to evaluate the performance of ontologies in different components of an autonomous vehicle within the 4D/RCS system architecture developed at NIST. Our initial focus has been on simple roadway driving scenarios where the controlled vehicle encounters potential obstacles in its path. As reported elsewhere [C. Schlenoff, S. Balakirsky, M. Uschold, R. Provine, S. Smith, Using ontologies to aid navigation planning in autonomous vehicles, Knowledge Engineering Review 18 (3) (2004) 243–255], our approach is to develop an ontology of objects in the environment, in conjunction with rules for estimating the damage that would be incurred by collisions with different objects in different situations. Automated reasoning is used to estimate collision damage; this information is fed to the route planner to help it decide whether to plan to avoid the object. We describe the results of the first implementation that integrates the ontology, the reasoner and the planner. We describe our insights and lessons learned and discuss resulting changes to our approach.
Online Verification of Automated Road Vehicles Using Reachability Analysis An approach for formally verifying the safety of automated vehicles is proposed. Due to the uniqueness of each traffic situation, we verify safety online, i.e., during the operation of the vehicle. The verification is performed by predicting the set of all possible occupancies of the automated vehicle and other traffic participants on the road. In order to capture all possible future scenarios, we apply reachability analysis to consider all possible behaviors of mathematical models considering uncertain inputs (e.g., sensor noise, disturbances) and partially unknown initial states. Safety is guaranteed with respect to the modeled uncertainties and behaviors if the occupancy of the automated vehicle does not intersect that of other traffic participants for all times. The applicability of the approach is demonstrated by test drives with an automated vehicle at the Robotics Institute at Carnegie Mellon University.
AVFI: Fault Injection for Autonomous Vehicles Autonomous vehicle (AV) technology is rapidly becoming a reality on U.S. roads, offering the promise of improvements in traffic management, safety, and the comfort and efficiency of vehicular travel. With this increasing popularity and ubiquitous deployment, resilience has become a critical requirement for public acceptance and adoption. Recent studies into the resilience of AVs have shown that though the AV systems are improving over time, they have not reached human levels of automation. Prior work in this area has studied the safety and resilience of individual components of the AV system (e.g., testing of neural networks powering the perception function). However, methods for holistic end-to-end resilience assessment of AV systems are still non-existent.
Specifying Safety of Autonomous Vehicles in Signal Temporal Logic We develop a set of contracts for autonomous control software that ensures that if all traffic participants follow the contracts, the overall traffic system will be collision-free. We express our contracts in Signal Temporal Logic (STL), a lightweight specification language that enables V &V methodologies. We demonstrate how the specification can be used for evaluation of the performance of autonomy software, and We provide preliminary evidence that our contracts are not excessively conservative, i.e., they are not more restrictive than existing guidelines for safe driving by humans.
An Interaction-aware Evaluation Method for Highly Automated Vehicles It is important to build a rigorous verification and validation (V&amp;V) process to evaluate the safety of highly automated vehicles (HAVs) before their wide deployment on public roads. In this paper, we propose an interaction-aware framework for HAV safety evaluation which is suitable for some highly-interactive driving scenarios including highway merging, roundabout entering, etc. Contrary to e...
Human-Like Decision Making for Autonomous Driving: A Noncooperative Game Theoretic Approach Considering that human-driven vehicles and autonomous vehicles (AVs) will coexist on roads in the future for a long time, how to merge AVs into human drivers' traffic ecology and minimize the effect of AVs and their misfit with human drivers, are issues worthy of consideration. Moreover, different passengers have different needs for AVs, thus, how to provide personalized choices for different passengers is another issue for AVs. Therefore, a human-like decision making framework is designed for AVs in this paper. Different driving styles and social interaction characteristics are formulated for AVs regarding driving safety, ride comfort and travel efficiency, which are considered in the modeling process of decision making. Then, Nash equilibrium and Stackelberg game theory are applied to the noncooperative decision making. In addition, potential field method and model predictive control (MPC) are combined to deal with the motion prediction and planning for AVs, which provides predicted motion information for the decision-making module. Finally, two typical testing scenarios of lane change, i.e., merging and overtaking, are carried out to evaluate the feasibility and effectiveness of the proposed decision-making framework considering different human-like behaviors. Testing results indicate that both the two game theoretic approaches can provide reasonable human-like decision making for AVs. Compared with the Nash equilibrium approach, under the normal driving style, the cost value of decision making using the Stackelberg game theoretic approach is reduced by over 20%.
Accurate Self-Localization in RFID Tag Information Grids Using FIR Filtering Grid navigation spaces nested with the radio-frequency identification (RFID) tags are promising for industrial and other needs, because each tag can deliver information about a local two-dimensional or three-dimensional surrounding. The approach, however, requires high accuracy in vehicle self-localization. Otherwise, errors may lead to collisions; possibly even fatal. We propose a new extended finite impulse response (EFIR) filtering algorithm and show that it meets this need. The EFIR filter requires an optimal averaging interval, but does not involve the noise statistics which are often not well known to the engineer. It is more accurate than the extended Kalman filter (EKF) under real operation conditions and its iterative algorithm has the Kalman form. Better performance of the proposed EFIR filter is demonstrated based on extensive simulations in a comparison to EKF, which is widely used in RFID tag grids. We also show that errors in noise covariances may provoke divergence in EKF, whereas the EFIR filter remains stable and is thus more robust.
A Survey on Transfer Learning A major assumption in many machine learning and data mining algorithms is that the training and future data must be in the same feature space and have the same distribution. However, in many real-world applications, this assumption may not hold. For example, we sometimes have a classification task in one domain of interest, but we only have sufficient training data in another domain of interest, where the latter data may be in a different feature space or follow a different data distribution. In such cases, knowledge transfer, if done successfully, would greatly improve the performance of learning by avoiding much expensive data-labeling efforts. In recent years, transfer learning has emerged as a new learning framework to address this problem. This survey focuses on categorizing and reviewing the current progress on transfer learning for classification, regression, and clustering problems. In this survey, we discuss the relationship between transfer learning and other related machine learning techniques such as domain adaptation, multitask learning and sample selection bias, as well as covariate shift. We also explore some potential future issues in transfer learning research.
The set cover with pairs problem We consider a generalization of the set cover problem, in which elements are covered by pairs of objects, and we are required to find a minimum cost subset of objects that induces a collection of pairs covering all elements. Formally, let U be a ground set of elements and let ${\cal S}$ be a set of objects, where each object i has a non-negative cost wi. For every $\{ i, j \} \subseteq {\cal S}$, let ${\cal C}(i,j)$ be the collection of elements in U covered by the pair { i, j }. The set cover with pairs problem asks to find a subset $A \subseteq {\cal S}$ such that $\bigcup_{ \{ i, j \} \subseteq A } {\cal C}(i,j) = U$ and such that ∑i∈Awi is minimized. In addition to studying this general problem, we are also concerned with developing polynomial time approximation algorithms for interesting special cases. The problems we consider in this framework arise in the context of domination in metric spaces and separation of point sets.
Telecommunications Power Plant Damage Assessment for Hurricane Katrina– Site Survey and Follow-Up Results This paper extends knowledge of disaster impact on the telecommunications power infrastructure by discussing the effects of Hurricane Katrina based on an on-site survey conducted in October 2005 and on public sources. It includes observations about power infrastructure damage in wire-line and wireless networks. In general, the impact on centralized network elements was more severe than on the distributed portion of the grids. The main cause of outage was lack of power due to fuel supply disruptions, flooding and security issues. This work also describes the means used to restore telecommunications services and proposes ways to improve logistics, such as coordinating portable generator set deployment among different network operators and reducing genset fuel consumption by installing permanent photovoltaic systems at sites where long electric outages are likely. One long term solution is to use of distributed generation. It also discusses the consequences on telecom power technology and practices since the storm.
When Vehicles See Pedestrians With Phones: A Multicue Framework for Recognizing Phone-Based Activities of Pedestrians The intelligent vehicle community has devoted considerable efforts to model driver behavior, and in particular, to detect and overcome driver distraction in an effort to reduce accidents caused by driver negligence. However, as the domain increasingly shifts toward autonomous and semiautonomous solutions, the driver is no longer integral to the decision-making process, indicating a need to refocus...
Pricing-Based Channel Selection for D2D Content Sharing in Dynamic Environment In order to make device-to-device (D2D) content sharing give full play to its advantage of improving local area services, one of the important issues is to decide the channels that D2D pairs occupy. Most existing works study this issue in static environment, and ignore the guidance for D2D pairs to select the channel adaptively. In this paper, we investigate this issue in dynamic environment where...
1.084444
0.066667
0.066667
0.066667
0.066667
0.066667
0.066667
0.033333
0
0
0
0
0
0
Joint Altitude and Beamwidth Optimization for UAV-Enabled Multiuser Communications. We study multiuser communication systems enabled by an unmanned aerial vehicle (UAV) that is equipped with a directional antenna of adjustable beamwidth. We propose a fly-hover-and-communicate protocol, where the ground terminals are partitioned into disjoint clusters that are sequentially served by the UAV as it hovers above the corresponding cluster centers. We jointly optimize the UAV's flying altitude and antenna beamwidth for throughput optimization in three fundamental multiuser communication models, namely, UAV-enabled downlink multicasting, downlink broadcasting, and uplink multiple access. Results show that the optimal UAV altitude and antenna beamwidth critically depend on the communication model considered.
Interior point methods 25 years later. Interior point methods for optimization have been around for more than 25 years now. Their presence has shaken up the field of optimization. Interior point methods for linear and (convex) quadratic programming display several features which make them particularly attractive for very large scale optimization. Among the most impressive of them are their low-degree polynomial worst-case complexity and an unrivalled ability to deliver optimal solutions in an almost constant number of iterations which depends very little, if at all, on the problem dimension. Interior point methods are competitive when dealing with small problems of dimensions below one million constraints and variables and are beyond competition when applied to large problems of dimensions going into millions of constraints and variables. In this survey we will discuss several issues related to interior point methods including the proof of the worst-case complexity result, the reasons for their amazingly fast practical convergence and the features responsible for their ability to solve very large problems. The ever-growing sizes of optimization problems impose new requirements on optimization methods and software. In the final part of this paper we will therefore address a redesign of interior point methods to allow them to work in a matrix-free regime and to make them well-suited to solving even larger problems. (C) 2011 Elsevier B.V. All rights reserved.
Genetic Algorithms in Wireless Networking: Techniques, Applications, and Issues. In recent times, wireless access technology is becoming increasingly commonplace due to the ease of operation and installation of untethered wireless media. The design of wireless networking is challenging due to the highly dynamic environmental condition that makes parameter optimization a complex task. Due to the dynamic, and often unknown, operating conditions, modern wireless networking standards increasingly rely on machine learning and artificial intelligence algorithms. Genetic algorithms (GAs) provide a well-established framework for implementing artificial intelligence tasks such as classification, learning, and optimization. GAs are well known for their remarkable generality and versatility and have been applied in a wide variety of settings in wireless networks. In this paper, we provide a comprehensive survey of the applications of GAs in wireless networks. We provide both an exposition of common GA models and configuration and provide a broad-ranging survey of GA techniques in wireless networks. We also point out open research issues and define potential future work. While various surveys on GAs exist in the literature, our paper is the first paper, to the best of our knowledge, which focuses on their application in wireless networks.
Fundamentals of the Downlink Green Coverage and Energy Efficiency in Heterogeneous Networks. This paper studies the proposed green (energy-efficient) coverage probability, link and network energy efficiencies in the downlink of a heterogeneous cellular network (HetNet) consisting of $K$ independent Poisson point processes of base stations (BSs). The important statistical properties of the universal (general) cell association functions are first studied, and the cell load statistics for power-law cell association functions, which can characterize the accurate void cell probability of a BS in every tier, are also derived. A simple and feasible green channel-aware cell association (GCA) scheme is proposed and the green coverage probability is also proposed for any particular cell association scheme, such as the maximum received power association (MRPA) and the nearest BS association (NBA) schemes. Then, the link and network energy efficiencies are proposed to characterize the mean spectrum efficiency per unit power consumption for a BS and the mean area spectrum efficiency for the HetNet, respectively. All the tight bounds on the green coverage probability, link, and network energy efficiencies for the GCA, MRPA, and NBA schemes are found. They are theoretically shown to pose the fundamental maximum limits on the link and network energy efficiencies achieved by any other cell association schemes, and such a fact is validated by numerical results as well.
Multiple UAVs as Relays: Multi-Hop Single Link Versus Multiple Dual-Hop Links. Unmanned aerial vehicles (UAVs) have found many important applications in communications. They can serve as either aerial base stations or mobile relays to improve the quality of services. In this paper, we study the use of multiple UAVs in relaying. Considering two typical uses of multiple UAVs as relays that form either a single multi-hop link or multiple dual-hop links, we first optimize the pl...
Energy-Efficient Data Collection in UAV Enabled Wireless Sensor Network. In wireless sensor networks, utilizing the unmanned aerial vehicle (UAV) as a mobile data collector for the sensor nodes (SNs) is an energy-efficient technique to prolong the network lifetime. In this letter, considering a general fading channel model for the SN-UAV links, we jointly optimize the SNs&#39; wake-up schedule and UAV&#39;s trajectory to minimize the maximum energy consumption of all SNs, whil...
Accessing From The Sky: A Tutorial on UAV Communications for 5G and Beyond. Unmanned aerial vehicles (UAVs) have found numerous applications and are expected to bring fertile business opportunities in the next decade. Among various enabling technologies for UAVs, wireless communication is essential and has drawn significantly growing attention in recent years. Compared to the conventional terrestrial communications, UAVs’ communications face new challenges due to their high altitude above the ground and great flexibility of movement in the 3-D space. Several critical issues arise, including the line-of-sight (LoS) dominant UAV-ground channels and induced strong aerial-terrestrial network interference, the distinct communication quality-of-service (QoS) requirements for UAV control messages versus payload data, the stringent constraints imposed by the size, weight, and power (SWAP) limitations of UAVs, as well as the exploitation of the new design degree of freedom (DoF) brought by the highly controllable 3-D UAV mobility. In this article, we give a tutorial overview of the recent advances in UAV communications to address the above issues, with an emphasis on how to integrate UAVs into the forthcoming fifth-generation (5G) and future cellular networks. In particular, we partition our discussion into two promising research and application frameworks of UAV communications, namely UAV-assisted wireless communications and cellular-connected UAVs, where UAVs are integrated into the network as new aerial communication platforms and users, respectively. Furthermore, we point out promising directions for future research.
P2S: A Primary and Passer-By Scheduling Algorithm for On-Demand Charging Architecture in Wireless Rechargeable Sensor Networks. As the interdiscipline of wireless communication and control engineering, the cooperative charging issue in wireless rechargeable sensor networks (WRSNs) is a popular research problem. With the help of wireless power transfer technology, electrical energy can be transferred from wireless charging vehicles to sensors, providing a new paradigm to prolong the network lifetime. However, existing techn...
Distributed wireless power transfer in sensor networks with multiple Mobile Chargers. We investigate the problem of efficient wireless power transfer in wireless sensor networks. In our approach, special mobile entities (called the Mobile Chargers) traverse the network and wirelessly replenish the energy of sensor nodes. In contrast to most current approaches, we envision methods that are distributed and use limited network information. We propose four new protocols for efficient charging, addressing key issues which we identify, most notably (i) what are good coordination procedures for the Mobile Chargers and (ii) what are good trajectories for the Mobile Chargers. Two of our protocols (DC, DCLK) perform distributed, limited network knowledge coordination and charging, while two others (CC, CCGK) perform centralized, global network knowledge coordination and charging. As detailed simulations demonstrate, one of our distributed protocols outperforms a known state of the art method, while its performance gets quite close to the performance of the powerful centralized global knowledge method.
The Whale Optimization Algorithm. The Whale Optimization Algorithm inspired by humpback whales is proposed.The WOA algorithm is benchmarked on 29 well-known test functions.The results on the unimodal functions show the superior exploitation of WOA.The exploration ability of WOA is confirmed by the results on multimodal functions.The results on structural design problems confirm the performance of WOA in practice. This paper proposes a novel nature-inspired meta-heuristic optimization algorithm, called Whale Optimization Algorithm (WOA), which mimics the social behavior of humpback whales. The algorithm is inspired by the bubble-net hunting strategy. WOA is tested with 29 mathematical optimization problems and 6 structural design problems. Optimization results prove that the WOA algorithm is very competitive compared to the state-of-art meta-heuristic algorithms as well as conventional methods. The source codes of the WOA algorithm are publicly available at http://www.alimirjalili.com/WOA.html
Geometric homogeneity with applications to finite-time stability This paper studies properties of homogeneous systems in a geometric, coordinate-free setting. A key contribution of this paper is a result relating regularity properties of a homogeneous function to its degree of homogeneity and the local behavior of the dilation near the origin. This result makes it possible to extend previous results on homogeneous systems to the geometric framework. As an application of our results, we consider finite-time stability of homogeneous systems. The main result that links homogeneity and finite-time stability is that a homogeneous system is finite-time stable if and only if it is asymptotically stable and has a negative degree of homogeneity. We also show that the assumption of homogeneity leads to stronger properties for finite-time stable systems.
A distributed event-triggered transmission strategy for sampled-data consensus of multi-agent systems. This paper is concerned with event-triggered sampled-data consensus for distributed multi-agent systems with directed graph. A novel distributed event-triggered sampled-data transmission strategy is proposed, which allows the event-triggering condition to be intermittently examined at constant sampling instants. Based on this novel strategy, a sampled-data consensus control protocol is presented, with which the consensus of distributed multi-agent systems can be transformed into the stability of a system with a time-varying delay. Then, a sufficient condition on the consensus of the multi-agent system is derived. Correspondingly, a co-design algorithm for obtaining both the parameters of the distributed event-triggered transmission strategy and the consensus controller gain is proposed. Two numerical examples are given to show the effectiveness of the proposed method.
Gender Bias in Coreference Resolution. We present an empirical study of gender bias in coreference resolution systems. We first introduce a novel, Winograd schema-style set of minimal pair sentences that differ only by pronoun gender. With these Winogender schemas, we evaluate and confirm systematic gender bias in three publicly-available coreference resolution systems, and correlate this bias with real-world and textual gender statistics.
A Hierarchical Architecture Using Biased Min-Consensus for USV Path Planning This paper proposes a hierarchical architecture using the biased min-consensus (BMC) method, to solve the path planning problem of unmanned surface vessel (USV). We take the fixed-point monitoring mission as an example, where a series of intermediate monitoring points should be visited once by USV. The whole framework incorporates the low-level layer planning the standard path between any two intermediate points, and the high-level fashion determining their visiting sequence. First, the optimal standard path in terms of voyage time and risk measure is planned by the BMC protocol, given that the corresponding graph is constructed with node state and edge weight. The USV will avoid obstacles or keep a certain distance safely, and arrive at the target point quickly. It is proven theoretically that the state of the graph will converge to be stable after finite iterations, i.e., the optimal solution can be found by BMC with low calculation complexity. Second, by incorporating the constraint of intermediate points, their visiting sequence is optimized by BMC again with the reconstruction of a new virtual graph based on the former planned results. The extensive simulation results in various scenarios also validate the feasibility and effectiveness of our method for autonomous navigation.
1.053041
0.05
0.05
0.05
0.05
0.027028
0.01
0.000057
0.000017
0
0
0
0
0
Identification of deep network generated images using disparities in color components •Disparities between real and GAN-generated images are analyzed in different color spaces.•Generated images are more distinguishable in residual domain of chrominance components.•Mismatched training-testing data and one-class classification are evaluated in experiments.
Multi-task Learning for Detecting and Segmenting Manipulated Facial Images and Videos Detecting manipulated images and videos is an important topic in digital media forensics. Most detection methods use binary classification to determine the probability of a query being manipulated. Another important topic is locating manipulated regions (i.e., performing segmentation), which are mostly created by three commonly used attacks: removal, copy-move, and splicing. We have designed a convolutional neural network that uses the multi-task learning approach to simultaneously detect manipulated images and videos and locate the manipulated regions for each query. Information gained by performing one task is shared with the other task and thereby enhance the performance of both tasks. A semi-supervised learning approach is used to improve the network’s generability. The network includes an encoder and a Y-shaped decoder. Activation of the encoded features is used for the binary classification. The output of one branch of the decoder is used for segmenting the manipulated regions while that of the other branch is used for reconstructing the input, which helps improve overall performance. Experiments using the FaceForensics and Face-Forensics++ databases demonstrated the networks effectiveness against facial reenactment attacks and face swapping attacks as well as its ability to deal with the mismatch condition for previously seen attacks. Moreover, fine-tuning using just a small amount of data enables the network to deal with unseen attacks.
Poster - Towards Robust Open-World Detection of Deepfakes. There is heightened concern over deliberately inaccurate news. Recently, so-called deepfake videos and images that are modified by or generated by artificial intelligence techniques have become more realistic and easier to create. These techniques could be used to create fake announcements from public figures or videos of events that did not happen, misleading mass audiences in dangerous ways. Although some recent research has examined accurate detection of deepfakes, those methodologies do not generalize well to real-world scenarios and are not available to the public in a usable form. In this project, we propose a system that will robustly and efficiently enable users to determine whether or not a video posted online is a deepfake. We approach the problem from the journalists' perspective and work towards developing a tool to fit seamlessly into their workflow. Results demonstrate accurate detection on both within and mismatched datasets.
Exposing video inter-frame forgery via histogram of oriented gradients and motion energy image Inter-frame forgery is a common type of video forgery to destroy the video evidence. It occurs in the temporal domain such as frame deletion, frame insertion, frame duplication, and frame shuffling. These forms of forgery are more frequently produced in a surveillance video because the camera position and the scene are relatively stable, where the tampering process is easy to operate and imperceptible. In this paper, we propose an efficient method for inter-frame forgery detection based on histogram of oriented gradients (HOG) and motion energy image (MEI). HOG is obtained from each image as a discriminative feature. In order to detect frame deletion and insertion, the correlation coefficients are used and abnormal points are detected via Grabb's test. In addition, MEI is applied to edge images of each shot to detect frame duplication and shuffling. Experimental results prove that the proposed method can detect all inter-frame forgeries and achieve higher accuracy with lower execution time.
Robust Localization of Interpolated Frames by Motion-Compensated Frame-Interpolation based on Artifact Indicated Map and Tchebichef Moments Motion-compensated frame interpolation (MCFI), a frame-interpolation technique to increase the motion continuity of low frame-rate video, can be utilized by counterfeiters for faking high bitrate video or splicing videos with different frame rates. For existing MCFI detectors, their performances are degraded under real-world scenarios such as H.264/AVC compression, noise, or blur. To address this issue, a robust MCFI detector is proposed to locate interpolated frames. By analyzing the distribution of residual energies within interpolated frames, we observe that there exist strong correlations between artifact regions and high residual energies. Thus, an artifact indicated map is introduced to select candidate artifact regions. Then, Tchebichef moments (TMs) are exploited to characterize the blurring effects or deformed structures among these regions. Specifically, the mean value of absolute high-order TMs of selected regions is used to model these temporal inconsistencies. Finally, a sliding window is adopted to locate interpolated frames, which are further refined by three post-processing operations. Chrominance information is also integrated with luminance information for robust identification of interpolated frames. Extensive experimental results show that compared with the state-of-the-art MCFI detectors, the proposed approach is more robust for compressed videos under various real-world scenarios.
Long short-term memory. Learning to store information over extended time intervals by recurrent backpropagation takes a very long time, mostly because of insufficient, decaying error backflow. We briefly review Hochreiter's (1991) analysis of this problem, then address it by introducing a novel, efficient, gradient-based method called long short-term memory (LSTM). Truncating the gradient where this does not do harm, LSTM can learn to bridge minimal time lags in excess of 1000 discrete-time steps by enforcing constant error flow through constant error carousels within special units. Multiplicative gate units learn to open and close access to the constant error flow. LSTM is local in space and time; its computational complexity per time step and weight is O(1). Our experiments with artificial data involve local, distributed, real-valued, and noisy pattern representations. In comparisons with real-time recurrent learning, back propagation through time, recurrent cascade correlation, Elman nets, and neural sequence chunking, LSTM leads to many more successful runs, and learns much faster. LSTM also solves complex, artificial long-time-lag tasks that have never been solved by previous recurrent network algorithms.
Accurate Self-Localization in RFID Tag Information Grids Using FIR Filtering Grid navigation spaces nested with the radio-frequency identification (RFID) tags are promising for industrial and other needs, because each tag can deliver information about a local two-dimensional or three-dimensional surrounding. The approach, however, requires high accuracy in vehicle self-localization. Otherwise, errors may lead to collisions; possibly even fatal. We propose a new extended finite impulse response (EFIR) filtering algorithm and show that it meets this need. The EFIR filter requires an optimal averaging interval, but does not involve the noise statistics which are often not well known to the engineer. It is more accurate than the extended Kalman filter (EKF) under real operation conditions and its iterative algorithm has the Kalman form. Better performance of the proposed EFIR filter is demonstrated based on extensive simulations in a comparison to EKF, which is widely used in RFID tag grids. We also show that errors in noise covariances may provoke divergence in EKF, whereas the EFIR filter remains stable and is thus more robust.
Distributed Representations, Simple Recurrent Networks, And Grammatical Structure In this paper three problems for a connectionist account of language are considered:1. What is the nature of linguistic representations?2. How can complex structural relationships such as constituent structure be represented?3. How can the apparently open-ended nature of language be accommodated by a fixed-resource system?Using a prediction task, a simple recurrent network (SRN) is trained on multiclausal sentences which contain multiply-embedded relative clauses. Principal component analysis of the hidden unit activation patterns reveals that the network solves the task by developing complex distributed representations which encode the relevant grammatical relations and hierarchical constituent structure. Differences between the SRN state representations and the more traditional pushdown store are discussed in the final section.
Latent dirichlet allocation We describe latent Dirichlet allocation (LDA), a generative probabilistic model for collections of discrete data such as text corpora. LDA is a three-level hierarchical Bayesian model, in which each item of a collection is modeled as a finite mixture over an underlying set of topics. Each topic is, in turn, modeled as an infinite mixture over an underlying set of topic probabilities. In the context of text modeling, the topic probabilities provide an explicit representation of a document. We present efficient approximate inference techniques based on variational methods and an EM algorithm for empirical Bayes parameter estimation. We report results in document modeling, text classification, and collaborative filtering, comparing to a mixture of unigrams model and the probabilistic LSI model.
Knowledge harvesting in the big-data era The proliferation of knowledge-sharing communities such as Wikipedia and the progress in scalable information extraction from Web and text sources have enabled the automatic construction of very large knowledge bases. Endeavors of this kind include projects such as DBpedia, Freebase, KnowItAll, ReadTheWeb, and YAGO. These projects provide automatically constructed knowledge bases of facts about named entities, their semantic classes, and their mutual relationships. They contain millions of entities and hundreds of millions of facts about them. Such world knowledge in turn enables cognitive applications and knowledge-centric services like disambiguating natural-language text, semantic search for entities and relations in Web and enterprise data, and entity-oriented analytics over unstructured contents. Prominent examples of how knowledge bases can be harnessed include the Google Knowledge Graph and the IBM Watson question answering system. This tutorial presents state-of-the-art methods, recent advances, research opportunities, and open challenges along this avenue of knowledge harvesting and its applications. Particular emphasis will be on the twofold role of knowledge bases for big-data analytics: using scalable distributed algorithms for harvesting knowledge from Web and text sources, and leveraging entity-centric knowledge for deeper interpretation of and better intelligence with Big Data.
MIMO Broadcasting for Simultaneous Wireless Information and Power Transfer Wireless power transfer (WPT) is a promising new solution to provide convenient and perpetual energy supplies to wireless networks. In practice, WPT is implementable by various technologies such as inductive coupling, magnetic resonate coupling, and electromagnetic (EM) radiation, for short-/mid-/long-range applications, respectively. In this paper, we consider the EM or radio signal enabled WPT in particular. Since radio signals can carry energy as well as information at the same time, a unified study on simultaneous wireless information and power transfer (SWIPT) is pursued. Specifically, this paper studies a multiple-input multiple-output (MIMO) wireless broadcast system consisting of three nodes, where one receiver harvests energy and another receiver decodes information separately from the signals sent by a common transmitter, and all the transmitter and receivers may be equipped with multiple antennas. Two scenarios are examined, in which the information receiver and energy receiver are separated and see different MIMO channels from the transmitter, or co-located and see the identical MIMO channel from the transmitter. For the case of separated receivers, we derive the optimal transmission strategy to achieve different tradeoffs for maximal information rate versus energy transfer, which are characterized by the boundary of a so-called rate-energy (R-E) region. For the case of co-located receivers, we show an outer bound for the achievable R-E region due to the potential limitation that practical energy harvesting receivers are not yet able to decode information directly. Under this constraint, we investigate two practical designs for the co-located receiver case, namely time switching and power splitting, and characterize their achievable R-E regions in comparison to the outer bound.
Passive Image-Splicing Detection by a 2-D Noncausal Markov Model In this paper, a 2-D noncausal Markov model is proposed for passive digital image-splicing detection. Different from the traditional Markov model, the proposed approach models an image as a 2-D noncausal signal and captures the underlying dependencies between the current node and its neighbors. The model parameters are treated as the discriminative features to differentiate the spliced images from the natural ones. We apply the model in the block discrete cosine transformation domain and the discrete Meyer wavelet transform domain, and the cross-domain features are treated as the final discriminative features for classification. The support vector machine which is the most popular classifier used in the image-splicing detection is exploited in our paper for classification. To evaluate the performance of the proposed method, all the experiments are conducted on public image-splicing detection evaluation data sets, and the experimental results have shown that the proposed approach outperforms some state-of-the-art methods.
A robust medical image watermarking against salt and pepper noise for brain MRI images. The ever-growing numbers of medical digital images and the need to share them among specialists and hospitals for better and more accurate diagnosis require that patients' privacy be protected. During the transmission of medical images between hospitals or specialists through the network, the main priority is to protect a patient's documents against any act of tampering by unauthorised individuals. Because of this, there is a need for medical image authentication scheme to enable proper diagnosis on patient. In addition, medical images are also susceptible to salt and pepper impulse noise through the transmission in communication channels. This noise may also be intentionally used by the invaders to corrupt the embedded watermarks inside the medical images. A common drawback of existing watermarking methods is their weakness against salt and pepper noise. The research carried out in this work addresses the issue of designing a new watermarking method that can withstand high density of salt and pepper noise for brain MRI images. For this purpose, combination of a spatial domain watermarking method, channel coding and noise filtering schemes are used. The region of non-interest (RONI) of MRI images from five different databases are used as embedding area and electronic patient record (EPR) is considered as embedded data. The quality of watermarked image is evaluated using Peak Signal-to-Noise Ratio (PSNR) and Structural Similarity Index (SSIM), and the accuracy of the extracted watermark is assessed in terms of Bit Error Rate (BER).
A Hierarchical Architecture Using Biased Min-Consensus for USV Path Planning This paper proposes a hierarchical architecture using the biased min-consensus (BMC) method, to solve the path planning problem of unmanned surface vessel (USV). We take the fixed-point monitoring mission as an example, where a series of intermediate monitoring points should be visited once by USV. The whole framework incorporates the low-level layer planning the standard path between any two intermediate points, and the high-level fashion determining their visiting sequence. First, the optimal standard path in terms of voyage time and risk measure is planned by the BMC protocol, given that the corresponding graph is constructed with node state and edge weight. The USV will avoid obstacles or keep a certain distance safely, and arrive at the target point quickly. It is proven theoretically that the state of the graph will converge to be stable after finite iterations, i.e., the optimal solution can be found by BMC with low calculation complexity. Second, by incorporating the constraint of intermediate points, their visiting sequence is optimized by BMC again with the reconstruction of a new virtual graph based on the former planned results. The extensive simulation results in various scenarios also validate the feasibility and effectiveness of our method for autonomous navigation.
1.1
0.1
0.1
0.1
0.1
0.001299
0
0
0
0
0
0
0
0
Electric Vehicle Charging Station Placement Method for Urban Areas. For accommodating more electric vehicles (EVs) to battle against fossil fuel emission, the problem of charging station placement is inevitable and could be costly if done improperly. Research considers a general setup using conditions such as driving ranges for planning. However, most of the EV growths in the next decades will happen in urban areas where driving range is not the biggest concern. F...
Computational difficulties of bilevel linear programming We show, using small examples, that two algorithms previously published for the Bilevel Linear Programming problem BLP may fail to find the optimal solution and thus must be considered to be heuris...
Competitive charging station pricing for plug-in electric vehicles This paper considers the problem of charging station pricing and station selection of plug-in electric vehicles (PEVs). Every PEV needs to select a charging station by considering the charging prices, waiting times, and travel distances. Each charging station optimizes its charging price based on the prediction of the PEVs' charging station selection decisions, in an attempt to maximize its profit. To obtain insights of such a highly coupled system, we consider a one-dimensional system with two charging stations and Poisson arriving PEVs. We propose a multi-leader-multi-follower Stackelberg game model, in which the charging stations (leaders) announce their charging prices in Stage I, and the PEVs (followers) make their charging station selections in Stage II. We show that there always exists a unique charging station selection equilibrium in Stage II, and such equilibrium depends on the price difference between the charging stations. We then characterize the sufficient conditions for the existence and uniqueness of the pricing equilibrium in Stage I. Unfortunately, it is hard to compute the pricing equilibrium in closed form. To overcome this challenge, we develop a low-complexity algorithm that efficiently computes the pricing equilibrium and the subgame perfect equilibrium of our Stackelberg game with no information exchange.
Electric Vehicle Charging Stations With Renewable Power Generators: A Game Theoretical Analysis In this paper, we study the price competition among electric vehicle charging stations (EVCSs) with renewable power generators (RPGs). As electric vehicles (EVs) become more popular, a competition among EVCSs to attract EVs is inevitable. Thereby, each EVCS sets its electricity price to maximize its revenue by taking into account the competition with neighboring EVCSs. We analyze the competitive interactions between EVCSs using game theory, where relevant physical constraints such as the transmission line capacity, the distance between EV and EVCS, and the number of charging outlets at the EVCSs are taken into account. We show that the game played by EVCSs is a supermodular game and there exists a unique pure Nash equilibrium for best response algorithms with arbitrary initial policy. The electricity price and the revenue of EVCSs are evaluated via simulations, which reveal the benefits of having RPGs at the EVCSs.
Optimized Charging Scheduling with Single Mobile Charger for Wireless Rechargeable Sensor Networks. Due to the rapid development of wireless charging technology, the recharging issue in wireless rechargeable sensor network (WRSN) has been a popular research problem in the past few years. The weakness of previous work is that charging route planning is not reasonable. In this work, a dynamic optimal scheduling scheme aiming to maximize the vacation time ratio of a single mobile changer for WRSN is proposed. In the proposed scheme, the wireless sensor network is divided into several sub-networks according to the initial topology of deployed sensor networks. After comprehensive analysis of energy states, working state and constraints for different sensor nodes in WRSN, we transform the optimized charging path problem of the whole network into the local optimization problem of the sub networks. The optimized charging path with respect to dynamic network topology in each sub-network is obtained by solving an optimization problem, and the lifetime of the deployed wireless sensor network can be prolonged. Simulation results show that the proposed scheme has good and reliable performance for a small wireless rechargeable sensor network.
Structure Learning in Power Distribution Networks. Traditional power distribution networks suffer from a lack of real-time observability. This complicates development and implementation of new smart-grid technologies, such as those related to demand response, outage detection and management, and improved load monitoring. In this paper, inspired by proliferation of metering technology, we discuss topology estimation problems in structurally loopy b...
Hierarchical Coupled Driving-and-Charging Model of Electric Vehicles, Stations and Grid Operators The decisions of operators from both the transportation and the electrical systems are coupled due to Electric Vehicles’ (EVs) actions. Thus, decision-making requires a model of several interdependent operators and of EVs’ both driving and charging behaviors. Such a model is suggested for the electrical system in the context of commuting, which has a typical trilevel structure. At the lower level ...
Massive MIMO for next generation wireless systems Multi-user MIMO offers big advantages over conventional point-to-point MIMO: it works with cheap single-antenna terminals, a rich scattering environment is not required, and resource allocation is simplified because every active terminal utilizes all of the time-frequency bins. However, multi-user MIMO, as originally envisioned, with roughly equal numbers of service antennas and terminals and frequency-division duplex operation, is not a scalable technology. Massive MIMO (also known as large-scale antenna systems, very large MIMO, hyper MIMO, full-dimension MIMO, and ARGOS) makes a clean break with current practice through the use of a large excess of service antennas over active terminals and time-division duplex operation. Extra antennas help by focusing energy into ever smaller regions of space to bring huge improvements in throughput and radiated energy efficiency. Other benefits of massive MIMO include extensive use of inexpensive low-power components, reduced latency, simplification of the MAC layer, and robustness against intentional jamming. The anticipated throughput depends on the propagation environment providing asymptotically orthogonal channels to the terminals, but so far experiments have not disclosed any limitations in this regard. While massive MIMO renders many traditional research problems irrelevant, it uncovers entirely new problems that urgently need attention: the challenge of making many low-cost low-precision components that work effectively together, acquisition and synchronization for newly joined terminals, the exploitation of extra degrees of freedom provided by the excess of service antennas, reducing internal power consumption to achieve total energy efficiency reductions, and finding new deployment scenarios. This article presents an overview of the massive MIMO concept and contemporary research on the topic.
Adaptive Federated Learning in Resource Constrained Edge Computing Systems Emerging technologies and applications including Internet of Things, social networking, and crowd-sourcing generate large amounts of data at the network edge. Machine learning models are often built from the collected data, to enable the detection, classification, and prediction of future events. Due to bandwidth, storage, and privacy concerns, it is often impractical to send all the data to a cen...
Image forgery detection We are undoubtedly living in an age where we are exposed to a remarkable array of visual imagery. While we may have historically had confidence in the integrity of this imagery, today&#39;s digital technology has begun to erode this trust. From the tabloid magazines to the fashion industry and in mainstream media outlets, scientific journals, political campaigns, courtrooms, and the photo hoaxes that ...
Using noise inconsistencies for blind image forensics A commonly used tool to conceal the traces of tampering is the addition of locally random noise to the altered image regions. The noise degradation is the main cause of failure of many active or passive image forgery detection methods. Typically, the amount of noise is uniform across the entire authentic image. Adding locally random noise may cause inconsistencies in the image's noise. Therefore, the detection of various noise levels in an image may signify tampering. In this paper, we propose a novel method capable of dividing an investigated image into various partitions with homogenous noise levels. In other words, we introduce a segmentation method detecting changes in noise level. We assume the additive white Gaussian noise. Several examples are shown to demonstrate the proposed method's output. An extensive quantitative measure of the efficiency of the noise estimation part as a function of different noise standard deviations, region sizes and various JPEG compression qualities is proposed as well.
Snow walking: motion-limiting device that reproduces the experience of walking in deep snow We propose \"Snow Walking,\" a boot-shaped device that reproduces the experience of walking in deep snow. The main purpose of this study is reproducing the feel of walking in a unique environment that we do not experience daily, particularly one that has depth, such as of deep snow. When you walk in deep snow, you get three feelings: the feel of pulling your foot up from the deep snow, the feel of putting your foot down into the deep snow, and the feel of your feet crunching across the bottom of deep snow. You cannot walk in deep snow easily, and with the system, you get a unique feeling not only on the sole of your foot but as if your entire foot is buried in the snow. We reproduce these feelings by using a slider, electromagnet, vibration speaker, a hook and loop fastener, and potato starch.
Survey of Fog Computing: Fundamental, Network Applications, and Research Challenges. Fog computing is an emerging paradigm that extends computation, communication, and storage facilities toward the edge of a network. Compared to traditional cloud computing, fog computing can support delay-sensitive service requests from end-users (EUs) with reduced energy consumption and low traffic congestion. Basically, fog networks are viewed as offloading to core computation and storage. Fog n...
Intention-detection strategies for upper limb exosuits: model-based myoelectric vs dynamic-based control The cognitive human-robot interaction between an exosuit and its wearer plays a key role in determining both the biomechanical effects of the device on movements and its perceived effectiveness. There is a lack of evidence, however, on the comparative performance of different control methods, implemented on the same device. Here, we compare two different control approaches on the same robotic suit: a model-based myoelectric control (myoprocessor), which estimates the joint torque from the activation of target muscles, and a dynamic-based control that provides support against gravity using an inverse dynamic model. Tested on a cohort of four healthy participants, assistance from the exosuit results in a marked reduction in the effort of muscles working against gravity with both control approaches (peak reduction of 68.6±18.8%, for the dynamic arm model and 62.4±25.1% for the myoprocessor), when compared to an unpowered condition. Neither of the two controllers had an affect on the performance of their users in a joint-angle tracking task (peak errors of 15.4° and 16.4° for the dynamic arm model and myoprocessor, respectively, compared to 13.1o in the unpowered condition). However, our results highlight the remarkable adaptability of the myoprocessor to seamlessly adapt to changing external dynamics.
1.066667
0.066667
0.066667
0.066667
0.066667
0.066667
0.066667
0
0
0
0
0
0
0
State resetting for bumpless switching in supervisory control In this paper the realization and implementation of a multi-controller scheme made of a finite set of linear single-input-single-output controllers, possibly having different state dimensions, is studied. The supervisory control framework is considered, namely a minimal parameter dependent realization of the set of controllers such that all controllers share the same state space is used. A specific state resetting strategy based on the behavioral approach to system theory is developed in order to master the transient upon controller switching.
Interpolation for gain-scheduled control with guarantees Here, a methodology is presented which considers the interpolation of linear time-invariant (LTI) controllers designed for different operating points of a nonlinear system in order to produce a gain-scheduled controller. Guarantees of closed-loop quadratic stability and performance at intermediate interpolation points are presented in terms of a set of linear matrix inequalities (LMIs). The proposed interpolation scheme can be applied in cases where the system must remain at the operating points most of the time and the transitions from one point to another rarely occur, e.g., chemical processes, satellites.
The smooth switching control for TORA system via LMIs This paper investigates the smooth switching control approach for linear parameter varying systems with applications to the nonlinear translational oscillator with rotational actuator (TORA) systems. In the smooth switching control, the control law for neighboring subsystems is smoothly switched or scheduled in the overlapped region instead of the usual instant switching methods. The design approach is presented in terms of the obtained maximum relative stability under control input and system output variables constraints. The control law is constructed based on the formulated linear matrix inequality conditions. As shown, the designed performance can be reasonably improved with the increase of number of partitions of the considered varying parameters range.
A Robust Controller Interpolation Design Technique Switching or blending among controllers is termed controller interpolation. This paper investigates a robust controller interpolation technique and applies it to an experimental test bed. Although an interpolated controller is composed of linear time-invariant (LTI) controllers stabilizing the LTI plant, closed-loop performance and stability are not guaranteed. Thus, it is of interest to design the interpolated controller to guarantee closed-loop stability and a performance level for all interpolation signals describing controller switching sequences and combinations. The performance metric that is under investigation in this paper is the H ¿ norm. A suboptimal robust interpolated-controller design algorithm is framed in terms of bilinear matrix inequalities. The motivating example demonstrates the efficacy of the robust interpolated-controller design.
Linear quadratic bumpless transfer A method for bumpless transfer using ideas from LQ theory is presented and shown to reduce to the Hanus conditioning scheme under certain conditions.
Smooth switching LPV controller design for LPV systems. This paper presents a method to design a smooth switching gain-scheduled linear parameter varying (LPV) controller for LPV systems. The moving region of the gain-scheduling variables is divided into a specified number of local subregions as well as subregions for the smooth controller switching, and one gain-scheduled LPV controller is assigned to each of the local subregions. For each switching subregion, a function interpolating two local LPV controllers associated with its neighborhood subregions is designed to satisfy the constraint of smooth transition of controller system matrices. The smooth switching controller design problem amounts to solving a feasibility problem which involves nonlinear matrix inequalities. To find a solution to the feasibility problem, an iterative descent algorithm which solves a series of convex optimization problems is proposed. The usefulness of the proposed controller design method is demonstrated with a control example of a flexible ball-screw drive system.
Stabilization of switched continuous-time systems with all modes unstable via dwell time switching Stabilization of switched systems composed fully of unstable subsystems is one of the most challenging problems in the field of switched systems. In this brief paper, a sufficient condition ensuring the asymptotic stability of switched continuous-time systems with all modes unstable is proposed. The main idea is to exploit the stabilization property of switching behaviors to compensate the state divergence made by unstable modes. Then, by using a discretized Lyapunov function approach, a computable sufficient condition for switched linear systems is proposed in the framework of dwell time; it is shown that the time intervals between two successive switching instants are required to be confined by a pair of upper and lower bounds to guarantee the asymptotic stability. Based on derived results, an algorithm is proposed to compute the stability region of admissible dwell time. A numerical example is proposed to illustrate our approach.
A survey on sensor networks The advancement in wireless communications and electronics has enabled the development of low-cost sensor networks. The sensor networks can be used for various application areas (e.g., health, military, home). For different application areas, there are different technical issues that researchers are currently resolving. The current state of the art of sensor networks is captured in this article, where solutions are discussed under their related protocol stack layer sections. This article also points out the open research issues and intends to spark new interests and developments in this field.
Energy-Aware Task Offloading and Resource Allocation for Time-Sensitive Services in Mobile Edge Computing Systems Mobile Edge Computing (MEC) is a promising architecture to reduce the energy consumption of mobile devices and provide satisfactory quality-of-service to time-sensitive services. How to jointly optimize task offloading and resource allocation to minimize the energy consumption subject to the latency requirement remains an open problem, which motivates this paper. When the latency constraint is tak...
Symbolic model checking for real-time systems We describe finite-state programs over real-numbered time in a guarded-command language with real-valued clocks or, equivalently, as finite automata with real-valued clocks. Model checking answers the question which states of a real-time program satisfy a branching-time specification (given in an extension of CTL with clock variables). We develop an algorithm that computes this set of states symbolically as a fixpoint of a functional on state predicates, without constructing the state space. For this purpose, we introduce a μ-calculus on computation trees over real-numbered time. Unfortunately, many standard program properties, such as response for all nonzeno execution sequences (during which time diverges), cannot be characterized by fixpoints: we show that the expressiveness of the timed μ-calculus is incomparable to the expressiveness of timed CTL. Fortunately, this result does not impair the symbolic verification of "implementable" real-time programs-those whose safety constraints are machine-closed with respect to diverging time and whose fairness constraints are restricted to finite upper bounds on clock values. All timed CTL properties of such programs are shown to be computable as finitely approximable fixpoints in a simple decidable theory.
Well-Solvable Special Cases of the Traveling Salesman Problem: A Survey. The traveling salesman problem (TSP) belongs to the most basic, most important, and most investigated problems in combinatorial optimization. Although it is an ${\cal NP}$-hard problem, many of its special cases can be solved efficiently in polynomial time. We survey these special cases with emphasis on the results that have been obtained during the decade 1985--1995. This survey complements an earlier survey from 1985 compiled by Gilmore, Lawler, and Shmoys [The Traveling Salesman Problem---A Guided Tour of Combinatorial Optimization, Wiley, Chichester, pp. 87--143].
Magnetic, Acceleration Fields and Gyroscope Quaternion (MAGYQ)-based attitude estimation with smartphone sensors for indoor pedestrian navigation. The dependence of proposed pedestrian navigation solutions on a dedicated infrastructure is a limiting factor to the deployment of location based services. Consequently self-contained Pedestrian Dead-Reckoning (PDR) approaches are gaining interest for autonomous navigation. Even if the quality of low cost inertial sensors and magnetometers has strongly improved, processing noisy sensor signals combined with high hand dynamics remains a challenge. Estimating accurate attitude angles for achieving long term positioning accuracy is targeted in this work. A new Magnetic, Acceleration fields and GYroscope Quaternion (MAGYQ)-based attitude angles estimation filter is proposed and demonstrated with handheld sensors. It benefits from a gyroscope signal modelling in the quaternion set and two new opportunistic updates: magnetic angular rate update (MARU) and acceleration gradient update (AGU). MAGYQ filter performances are assessed indoors, outdoors, with dynamic and static motion conditions. The heading error, using only the inertial solution, is found to be less than 10 degrees after 1.5 km walking. The performance is also evaluated in the positioning domain with trajectories computed following a PDR strategy.
Scalable and Privacy-Preserving Data Sharing Based on Blockchain. With the development of network technology and cloud computing, data sharing is becoming increasingly popular, and many scholars have conducted in-depth research to promote its flourish. As the scale of data sharing expands, its privacy protection has become a hot issue in research. Moreover, in data sharing, the data is usually maintained in multiple parties, which brings new challenges to protect the privacy of these multi-party data. In this paper, we propose a trusted data sharing scheme using blockchain. We use blockchain to prevent the shared data from being tampered, and use the Paillier cryptosystem to realize the confidentiality of the shared data. In the proposed scheme, the shared data can be traded, and the transaction information is protected by using the (p, t)-threshold Paillier cryptosystem. We conduct experiments in cloud storage scenarios and the experimental results demonstrate the efficiency and effectiveness of the proposed scheme.
Robot tutor and pupils’ educational ability: Teaching the times tables Research shows promising results of educational robots in language and STEM tasks. In language, more research is available, occasionally in view of individual differences in pupils’ educational ability levels, and learning seems to improve with more expressive robot behaviors. In STEM, variations in robots’ behaviors have been examined with inconclusive results and never while systematically investigating how differences in educational abilities match with different robot behaviors. We applied an autonomously tutoring robot (without tablet, partly WOz) in a 2 × 2 experiment of social vs. neutral behavior in above-average vs. below-average schoolchildren (N = 86; age 8–10 years) while rehearsing the multiplication tables on a one-to-one basis. The standard school test showed that on average, pupils significantly improved their performance even after 3 occasions of 5-min exercises. Beyond-average pupils profited most from a robot tutor, whereas those below average in multiplication benefited more from a robot that showed neutral rather than more social behavior.
1.205275
0.205275
0.205275
0.102775
0.04177
0.001375
0.000012
0
0
0
0
0
0
0
Generalized Actor-Critic Learning Optimal Control in Smart Home Energy Management This article is concerned with a new generalized actor-critic learning (GACL) optimal control method. It aims at the optimal energy control and management for smart home systems, which is expected to minimize the consumption cost for home users. In the present GACL optimal control method, it is the first time that three iteration processes, which are global iteration, local iteration, and interior...
Novel weighting-delay-based stability criteria for recurrent neural networks with time-varying delay. In this paper, a weighting-delay-based method is developed for the study of the stability problem of a class of recurrent neural networks (RNNs) with time-varying delay. Different from previous results, the delay interval [0, d(t)] is divided into some variable subintervals by employing weighting delays. Thus, new delay-dependent stability criteria for RNNs with time-varying delay are derived by applying this weighting-delay method, which are less conservative than previous results. The proposed stability criteria depend on the positions of weighting delays in the interval [0, d(t)] , which can be denoted by the weighting-delay parameters. Different weighting-delay parameters lead to different stability margins for a given system. Thus, a solution based on optimization methods is further given to calculate the optimal weighting-delay parameters. Several examples are provided to verify the effectiveness of the proposed criteria.
Optimal control using adaptive resonance theory and Q-learning. Motivated by recent advancement in neurocognitive in brain modeling research, a multiple model-based Q-learning structure is proposed for optimal tracking control problem of time-varying discrete-time systems. This is achieved by utilizing a multiple-model scheme combined with adaptive resonance theory (ART). The ART algorithm generates sub-models based on the match-based clustering method. A responsibility signal governs the likelihood of contribution of each sub-model to the Q-function. The Q-function is learned using the batch least-square algorithm. Simulation results are added to show the performance and the effectiveness of the overall proposed control method.
A novel adaptive dynamic programming based on tracking error for nonlinear discrete-time systems In this paper, to eliminate the tracking error by using adaptive dynamic programming (ADP) algorithms, a novel formulation of the value function is presented for the optimal tracking problem (TP) of nonlinear discrete-time systems. Unlike existing ADP methods, this formulation introduces the control input into the tracking error, and ignores the quadratic form of the control input directly, which makes the boundedness and convergence of the value function independent of the discount factor. Based on the proposed value function, the optimal control policy can be deduced without considering the reference control input. Value iteration (VI) and policy iteration (PI) methods are applied to prove the optimality of the obtained control policy, and derived the monotonicity property and convergence of the iterative value function. Simulation examples realized with neural networks and the actor–critic structure are provided to verify the effectiveness of the proposed ADP algorithm.
Adaptive Learning in Tracking Control Based on the Dual Critic Network Design. In this paper, we present a new adaptive dynamic programming approach by integrating a reference network that provides an internal goal representation to help the systems learning and optimization. Specifically, we build the reference network on top of the critic network to form a dual critic network design that contains the detailed internal goal representation to help approximate the value funct...
The Sybil Attack Large-scale peer-to-peer systems facesecurity threats from faulty or hostile remotecomputing elements. To resist these threats, manysuch systems employ redundancy. However, if asingle faulty entity can present multiple identities,it can control a substantial fraction of the system,thereby undermining this redundancy. Oneapproach to preventing these &quot;Sybil attacks&quot; is tohave a trusted agency certify identities. Thispaper shows that, without a logically centralizedauthority, Sybil...
A fast and elitist multiobjective genetic algorithm: NSGA-II Multi-objective evolutionary algorithms (MOEAs) that use non-dominated sorting and sharing have been criticized mainly for: (1) their O(MN3) computational complexity (where M is the number of objectives and N is the population size); (2) their non-elitism approach; and (3) the need to specify a sharing parameter. In this paper, we suggest a non-dominated sorting-based MOEA, called NSGA-II (Non-dominated Sorting Genetic Algorithm II), which alleviates all of the above three difficulties. Specifically, a fast non-dominated sorting approach with O(MN2) computational complexity is presented. Also, a selection operator is presented that creates a mating pool by combining the parent and offspring populations and selecting the best N solutions (with respect to fitness and spread). Simulation results on difficult test problems show that NSGA-II is able, for most problems, to find a much better spread of solutions and better convergence near the true Pareto-optimal front compared to the Pareto-archived evolution strategy and the strength-Pareto evolutionary algorithm - two other elitist MOEAs that pay special attention to creating a diverse Pareto-optimal front. Moreover, we modify the definition of dominance in order to solve constrained multi-objective problems efficiently. Simulation results of the constrained NSGA-II on a number of test problems, including a five-objective, seven-constraint nonlinear problem, are compared with another constrained multi-objective optimizer, and the much better performance of NSGA-II is observed
Gradient-Based Learning Applied to Document Recognition Multilayer neural networks trained with the back-propagation algorithm constitute the best example of a successful gradient based learning technique. Given an appropriate network architecture, gradient-based learning algorithms can be used to synthesize a complex decision surface that can classify high-dimensional patterns, such as handwritten characters, with minimal preprocessing. This paper rev...
Latent dirichlet allocation We describe latent Dirichlet allocation (LDA), a generative probabilistic model for collections of discrete data such as text corpora. LDA is a three-level hierarchical Bayesian model, in which each item of a collection is modeled as a finite mixture over an underlying set of topics. Each topic is, in turn, modeled as an infinite mixture over an underlying set of topic probabilities. In the context of text modeling, the topic probabilities provide an explicit representation of a document. We present efficient approximate inference techniques based on variational methods and an EM algorithm for empirical Bayes parameter estimation. We report results in document modeling, text classification, and collaborative filtering, comparing to a mixture of unigrams model and the probabilistic LSI model.
Knowledge harvesting in the big-data era The proliferation of knowledge-sharing communities such as Wikipedia and the progress in scalable information extraction from Web and text sources have enabled the automatic construction of very large knowledge bases. Endeavors of this kind include projects such as DBpedia, Freebase, KnowItAll, ReadTheWeb, and YAGO. These projects provide automatically constructed knowledge bases of facts about named entities, their semantic classes, and their mutual relationships. They contain millions of entities and hundreds of millions of facts about them. Such world knowledge in turn enables cognitive applications and knowledge-centric services like disambiguating natural-language text, semantic search for entities and relations in Web and enterprise data, and entity-oriented analytics over unstructured contents. Prominent examples of how knowledge bases can be harnessed include the Google Knowledge Graph and the IBM Watson question answering system. This tutorial presents state-of-the-art methods, recent advances, research opportunities, and open challenges along this avenue of knowledge harvesting and its applications. Particular emphasis will be on the twofold role of knowledge bases for big-data analytics: using scalable distributed algorithms for harvesting knowledge from Web and text sources, and leveraging entity-centric knowledge for deeper interpretation of and better intelligence with Big Data.
Labels and event processes in the Asbestos operating system Asbestos, a new operating system, provides novel labeling and isolation mechanisms that help contain the effects of exploitable software flaws. Applications can express a wide range of policies with Asbestos's kernel-enforced labels, including controls on interprocess communication and system-wide information flow. A new event process abstraction defines lightweight, isolated contexts within a single process, allowing one process to act on behalf of multiple users while preventing it from leaking any single user's data to others. A Web server demonstration application uses these primitives to isolate private user data. Since the untrusted workers that respond to client requests are constrained by labels, exploited workers cannot directly expose user data except as allowed by application policy. The server application requires 1.4 memory pages per user for up to 145,000 users and achieves connection rates similar to Apache, demonstrating that additional security can come at an acceptable cost.
GROPING: Geomagnetism and cROwdsensing Powered Indoor NaviGation Although a large number of WiFi fingerprinting based indoor localization systems have been proposed, our field experience with Google Maps Indoor (GMI), the only system available for public testing, shows that it is far from mature for indoor navigation. In this paper, we first report our field studies with GMI, as well as experiment results aiming to explain our unsatisfactory GMI experience. Then motivated by the obtained insights, we propose GROPING as a self-contained indoor navigation system independent of any infrastructural support. GROPING relies on geomagnetic fingerprints that are far more stable than WiFi fingerprints, and it exploits crowdsensing to construct floor maps rather than expecting individual venues to supply digitized maps. Based on our experiments with 20 participants in various floors of a big shopping mall, GROPING is able to deliver a sufficient accuracy for localization and thus provides smooth navigation experience.
5G Virtualized Multi-access Edge Computing Platform for IoT Applications. The next generation of fifth generation (5G) network, which is implemented using Virtualized Multi-access Edge Computing (vMEC), Network Function Virtualization (NFV) and Software Defined Networking (SDN) technologies, is a flexible and resilient network that supports various Internet of Things (IoT) devices. While NFV provides flexibility by allowing network functions to be dynamically deployed and inter-connected, vMEC provides intelligence at the edge of the mobile network reduces latency and increases the available capacity. With the diverse development of networking applications, the proposed vMEC use of Container-based Virtualization Technology (CVT) as gateway with IoT devices for flow control mechanism in scheduling and analysis methods will effectively increase the application Quality of Service (QoS). In this work, the proposed IoT gateway is analyzed. The combined effect of simultaneously deploying Virtual Network Functions (VNFs) and vMEC applications on a single network infrastructure, and critically in effecting exhibits low latency, high bandwidth and agility that will be able to connect large scale of devices. The proposed platform efficiently exploiting resources from edge computing and cloud computing, and takes IoT applications that adapt to network conditions to degrade an average 30% of end to end network latency.
Social Robots for (Second) Language Learning in (Migrant) Primary School Children Especially these days, innovation and support from technology to relieve pressure in education is highly urgent. This study tested the potential advantage of a social robot over a tablet in (second) language learning on performance, engagement, and enjoyment. Shortages in primary education call for new technology solutions. Previous studies combined robots with tablets, to compensate for robot’s limitations, however, this study applied direct human–robot interaction. Primary school children (N = 63, aged 4–6) participated in a 3-wave field experiment with story-telling exercises, either with a semi-autonomous robot (without tablet, using WOz) or a tablet. Results showed increased learning gains over time when training with a social robot, compared to the tablet. Children who trained with a robot were more engaged in the story-telling task and enjoyed it more. Robot’s behavioral style (social or neutral) hardly differed overall, however, seems to vary for high versus low educational abilities. While social robots need sophistication before being implemented in schools, our study shows the potential of social robots as tutors in (second) language learning.
1.2
0.2
0.2
0.2
0.066667
0
0
0
0
0
0
0
0
0
An Analysis of Price Competition in Heterogeneous Electric Vehicle Charging Stations In this paper, we investigate the price competition among electric vehicle charging stations (EVCSs) with renewable power generators (RPGs). Both a large-sized EVCS (L-EVCS) and small-sized EVCSs (S-EVCSs), which have different capacities, are considered. Moreover, the price elasticity of electric vehicles (EVs), the effect of the distance between an EV and the EVCSs, and the impact of the number ...
Computational difficulties of bilevel linear programming We show, using small examples, that two algorithms previously published for the Bilevel Linear Programming problem BLP may fail to find the optimal solution and thus must be considered to be heuris...
Competitive charging station pricing for plug-in electric vehicles This paper considers the problem of charging station pricing and station selection of plug-in electric vehicles (PEVs). Every PEV needs to select a charging station by considering the charging prices, waiting times, and travel distances. Each charging station optimizes its charging price based on the prediction of the PEVs' charging station selection decisions, in an attempt to maximize its profit. To obtain insights of such a highly coupled system, we consider a one-dimensional system with two charging stations and Poisson arriving PEVs. We propose a multi-leader-multi-follower Stackelberg game model, in which the charging stations (leaders) announce their charging prices in Stage I, and the PEVs (followers) make their charging station selections in Stage II. We show that there always exists a unique charging station selection equilibrium in Stage II, and such equilibrium depends on the price difference between the charging stations. We then characterize the sufficient conditions for the existence and uniqueness of the pricing equilibrium in Stage I. Unfortunately, it is hard to compute the pricing equilibrium in closed form. To overcome this challenge, we develop a low-complexity algorithm that efficiently computes the pricing equilibrium and the subgame perfect equilibrium of our Stackelberg game with no information exchange.
Placement of EV Charging Stations - Balancing Benefits Among Multiple Entities. This paper studies the problem of multistage placement of electric vehicle (EV) charging stations with incremental EV penetration rates. A nested logit model is employed to analyze the charging preference of the individual consumer (EV owner) and predict the aggregated charging demand at the charging stations. The EV charging industry is modeled as an oligopoly where the entire market is dominated...
Plug-in Electric Vehicle Charging With Multiple Charging Options: A Systematic Analysis of Service Providers’ Pricing Strategies In this article, we present the optimal strategy for public plug-in electric vehicle (PEV) charging services with multiple charging options: fast charging service and slow charging service. The PEV user make decisions by firstly considering the remaining energy, travel distance to charging station, then about the charging prices, charging rates and estimated charging time. To obtain insights of such a highly coupled system, we consider both monopoly and duopoly markets, where the two services are operated by a single service provider (SP) and two different SPs, respectively. For the monopoly market, we propose a three-stage Stackelberg game model. For the duopoly market, we propose a two-stage Bertrand competition and Cournot competition. We provided the SP's optimal pricing strategies and decisions in the monopoly market, and the optimal pricing contract and quantity contract for SPs in the duopoly market with Bertrand competition and Cournot competition. Both the analytical and simulation results show that, in the monopoly market, offering two charging options can potentially improve the SP's profit compared with offering one option only in the monopoly market; while in the duopoly market, it is optimal for SPs to offer the quantity contract rather than the price contract to maximize profits.
Deployment of the Electric Vehicle Charging Station Considering Existing Competitors The problem of optimal planning of plug-in electric vehicle (PEV) charging stations is studied. Different from many other works, we assume the station investor to be a new private entrant into the market who intends to maximize its own profit in a competitive environment. A modified Huff gravity-based model is adopted to describe the probabilistic patronizing behaviors of PEV drivers. Accordingly, a bi-level optimization model is formulated to decide not only the optimal site and size of the new charging station, but also the retail charging prices in the future operation stage. Based on the specific characteristics of the problem, the operation level sub-problem is reformulated to a convex programming and an efficient solution algorithm is designed for the overall bi-level optimization. Numerical examples of different scales demonstrate the effectiveness of the proposed modeling and computation methods, as well as the importance of considering the competitive effects when planning the charging station.
Completely derandomized self-adaptation in evolution strategies. This paper puts forward two useful methods for self-adaptation of the mutation distribution - the concepts of derandomization and cumulation. Principle shortcomings of the concept of mutative strategy parameter control and two levels of derandomization are reviewed. Basic demands on the self-adaptation of arbitrary (normal) mutation distributions are developed. Applying arbitrary, normal mutation distributions is equivalent to applying a general, linear problem encoding. The underlying objective of mutative strategy parameter control is roughly to favor previously selected mutation steps in the future. If this objective is pursued rigorously, a completely derandomized self-adaptation scheme results, which adapts arbitrary normal mutation distributions. This scheme, called covariance matrix adaptation (CMA), meets the previously stated demands. It can still be considerably improved by cumulation - utilizing an evolution path rather than single search steps. Simulations on various test functions reveal local and global search properties of the evolution strategy with and without covariance matrix adaptation. Their performances are comparable only on perfectly scaled functions. On badly scaled, non-separable functions usually a speed up factor of several orders of magnitude is observed. On moderately mis-scaled functions a speed up factor of three to ten can be expected.
An online mechanism for multi-unit demand and its application to plug-in hybrid electric vehicle charging We develop an online mechanism for the allocation of an expiring resource to a dynamic agent population. Each agent has a non-increasing marginal valuation function for the resource, and an upper limit on the number of units that can be allocated in any period. We propose two versions on a truthful allocation mechanism. Each modifies the decisions of a greedy online assignment algorithm by sometimes cancelling an allocation of resources. One version makes this modification immediately upon an allocation decision while a second waits until the point at which an agent departs the market. Adopting a prior-free framework, we show that the second approach has better worst-case allocative efficiency and is more scalable. On the other hand, the first approach (with immediate cancellation) may be easier in practice because it does not need to reclaim units previously allocated. We consider an application to recharging plug-in hybrid electric vehicles (PHEVs). Using data from a real-world trial of PHEVs in the UK, we demonstrate higher system performance than a fixed price system, performance comparable with a standard, but non-truthful scheduling heuristic, and the ability to support 50% more vehicles at the same fuel cost than a simple randomized policy.
Blockchain Meets IoT: An Architecture for Scalable Access Management in IoT. The Internet of Things (IoT) is stepping out of its infancy into full maturity and establishing itself as a part of the future Internet. One of the technical challenges of having billions of devices deployed worldwide is the ability to manage them. Although access management technologies exist in IoT, they are based on centralized models which introduce a new variety of technical limitations to ma...
Multi-column Deep Neural Networks for Image Classification Traditional methods of computer vision and machine learning cannot match human performance on tasks such as the recognition of handwritten digits or traffic signs. Our biologically plausible deep artificial neural network architectures can. Small (often minimal) receptive fields of convolutional winner-take-all neurons yield large network depth, resulting in roughly as many sparsely connected neural layers as found in mammals between retina and visual cortex. Only winner neurons are trained. Several deep neural columns become experts on inputs preprocessed in different ways; their predictions are averaged. Graphics cards allow for fast training. On the very competitive MNIST handwriting benchmark, our method is the first to achieve near-human performance. On a traffic sign recognition benchmark it outperforms humans by a factor of two. We also improve the state-of-the-art on a plethora of common image classification benchmarks.
Comment on "On Discriminative vs. Generative Classifiers: A Comparison of Logistic Regression and Naive Bayes" Comparison of generative and discriminative classifiers is an ever-lasting topic. As an important contribution to this topic, based on their theoretical and empirical comparisons between the naïve Bayes classifier and linear logistic regression, Ng and Jordan (NIPS 841---848, 2001) claimed that there exist two distinct regimes of performance between the generative and discriminative classifiers with regard to the training-set size. In this paper, our empirical and simulation studies, as a complement of their work, however, suggest that the existence of the two distinct regimes may not be so reliable. In addition, for real world datasets, so far there is no theoretically correct, general criterion for choosing between the discriminative and the generative approaches to classification of an observation x into a class y; the choice depends on the relative confidence we have in the correctness of the specification of either p(y|x) or p(x, y) for the data. This can be to some extent a demonstration of why Efron (J Am Stat Assoc 70(352):892---898, 1975) and O'Neill (J Am Stat Assoc 75(369):154---160, 1980) prefer normal-based linear discriminant analysis (LDA) when no model mis-specification occurs but other empirical studies may prefer linear logistic regression instead. Furthermore, we suggest that pairing of either LDA assuming a common diagonal covariance matrix (LDA-驴) or the naïve Bayes classifier and linear logistic regression may not be perfect, and hence it may not be reliable for any claim that was derived from the comparison between LDA-驴 or the naïve Bayes classifier and linear logistic regression to be generalised to all generative and discriminative classifiers.
Adaptive dynamic programming and optimal control of nonlinear nonaffine systems. In this paper, a novel optimal control design scheme is proposed for continuous-time nonaffine nonlinear dynamic systems with unknown dynamics by adaptive dynamic programming (ADP). The proposed methodology iteratively updates the control policy online by using the state and input information without identifying the system dynamics. An ADP algorithm is developed, and can be applied to a general class of nonlinear control design problems. The convergence analysis for the designed control scheme is presented, along with rigorous stability analysis for the closed-loop system. The effectiveness of this new algorithm is illustrated by two simulation examples.
Adaptive Fuzzy Control With Prescribed Performance for Block-Triangular-Structured Nonlinear Systems. In this paper, an adaptive fuzzy control method with prescribed performance is proposed for multi-input and multioutput block-triangular-structured nonlinear systems with immeasurable states. Fuzzy logic systems are adopted to identify the unknown nonlinear system functions. Adaptive fuzzy state observers are designed to solve the problem of unmeasured states, and a new observer-based output-feedb...
Intention-detection strategies for upper limb exosuits: model-based myoelectric vs dynamic-based control The cognitive human-robot interaction between an exosuit and its wearer plays a key role in determining both the biomechanical effects of the device on movements and its perceived effectiveness. There is a lack of evidence, however, on the comparative performance of different control methods, implemented on the same device. Here, we compare two different control approaches on the same robotic suit: a model-based myoelectric control (myoprocessor), which estimates the joint torque from the activation of target muscles, and a dynamic-based control that provides support against gravity using an inverse dynamic model. Tested on a cohort of four healthy participants, assistance from the exosuit results in a marked reduction in the effort of muscles working against gravity with both control approaches (peak reduction of 68.6±18.8%, for the dynamic arm model and 62.4±25.1% for the myoprocessor), when compared to an unpowered condition. Neither of the two controllers had an affect on the performance of their users in a joint-angle tracking task (peak errors of 15.4° and 16.4° for the dynamic arm model and myoprocessor, respectively, compared to 13.1o in the unpowered condition). However, our results highlight the remarkable adaptability of the myoprocessor to seamlessly adapt to changing external dynamics.
1.2
0.2
0.2
0.2
0.2
0.2
0
0
0
0
0
0
0
0
Multiobjective Evolution of Fuzzy Rough Neural Network via Distributed Parallelism for Stock Prediction Fuzzy rough theory can describe real-world situations in a mathematically effective and interpretable way, while evolutionary neural networks can be utilized to solve complex problems. Combining them with these complementary capabilities may lead to evolutionary fuzzy rough neural network with the interpretability and prediction capability. In this article, we propose modifications to the existing models of fuzzy rough neural network and then develop a powerful evolutionary framework for fuzzy rough neural networks by inheriting the merits of both the aforementioned systems. We first introduce rough neurons and enhance the consequence nodes, and further integrate the interval type-2 fuzzy set into the existing fuzzy rough neural network model. Thus, several modified fuzzy rough neural network models are proposed. While simultaneously considering the objectives of <italic xmlns:mml="http://www.w3.org/1998/Math/MathML" xmlns:xlink="http://www.w3.org/1999/xlink">prediction precision</italic> and <italic xmlns:mml="http://www.w3.org/1998/Math/MathML" xmlns:xlink="http://www.w3.org/1999/xlink">network simplicity</italic> , each model is transformed into a multiobjective optimization problem by encoding the structure, membership functions, and the parameters of the network. To solve these optimization problems, distributed parallel multiobjective evolutionary algorithms are proposed. We enhance the optimization processes with several measures including optimizer replacement and parameter adaption. In the distributed parallel environment, the tedious and time-consuming neural network optimization can be alleviated by numerous computational resources, significantly reducing the computational time. Through experimental verification on complex stock time series prediction tasks, the proposed optimization algorithms and the modified fuzzy rough neural network models exhibit significant improvements the existing fuzzy rough neural network and the long short-term memory network.
Fast learning neural networks using Cartesian genetic programming A fast learning neuroevolutionary algorithm for both feedforward and recurrent networks is proposed. The method is inspired by the well known and highly effective Cartesian genetic programming (CGP) technique. The proposed method is called the CGP-based Artificial Neural Network (CGPANN). The basic idea is to replace each computational node in CGP with an artificial neuron, thus producing an artificial neural network. The capabilities of CGPANN are tested in two diverse problem domains. Firstly, it has been tested on a standard benchmark control problem: single and double pole for both Markovian and non-Markovian cases. Results demonstrate that the method can generate effective neural architectures in substantially fewer evaluations in comparison to previously published neuroevolutionary techniques. In addition, the evolved networks show improved generalization and robustness in comparison with other techniques. Secondly, we have explored the capabilities of CGPANNs for the diagnosis of Breast Cancer from the FNA (Finite Needle Aspiration) data samples. The results demonstrate that the proposed algorithm gives 99.5% accurate results, thus making it an excellent choice for pattern recognitions in medical diagnosis, owing to its properties of fast learning and accuracy. The power of a CGP based ANN is its representation which leads to an efficient evolutionary search of suitable topologies. This opens new avenues for applying the proposed technique to other linear/non-linear and Markovian/non-Markovian control and pattern recognition problems.
Designing adaptive humanoid robots through the FARSA open-source framework We introduce FARSA, an open-source Framework for Autonomous Robotics Simulation and Analysis, that allows us to easily set up and carry on adaptive experiments involving complex robot/environmental models. Moreover, we show how a simulated iCub robot can be trained, through an evolutionary algorithm, to display reaching and integrated reaching and grasping behaviours. The results demonstrate how the use of an implicit selection criterion, estimating the extent to which the robot is able to produce the expected outcome without specifying the manner through which the action should be realized, is sufficient to develop the required capabilities despite the complexity of the robot and of the task.
A Hierarchical Fused Fuzzy Deep Neural Network for Data Classification. Deep learning (DL) is an emerging and powerful paradigm that allows large-scale task-driven feature learning from big data. However, typical DL is a fully deterministic model that sheds no light on data uncertainty reductions. In this paper, we show how to introduce the concepts of fuzzy learning into DL to overcome the shortcomings of fixed representation. The bulk of the proposed fuzzy system is...
Deep Learning With Edge Computing: A Review Deep learning is currently widely used in a variety of applications, including computer vision and natural language processing. End devices, such as smartphones and Internet-of-Things sensors, are generating data that need to be analyzed in real time using deep learning or used to train deep learning models. However, deep learning inference and training require substantial computation resources to run quickly. Edge computing, where a fine mesh of compute nodes are placed close to end devices, is a viable way to meet the high computation and low-latency requirements of deep learning on edge devices and also provides additional benefits in terms of privacy, bandwidth efficiency, and scalability. This paper aims to provide a comprehensive review of the current state of the art at the intersection of deep learning and edge computing. Specifically, it will provide an overview of applications where deep learning is used at the network edge, discuss various approaches for quickly executing deep learning inference across a combination of end devices, edge servers, and the cloud, and describe the methods for training deep learning models across multiple edge devices. It will also discuss open challenges in terms of systems performance, network technologies and management, benchmarks, and privacy. The reader will take away the following concepts from this paper: understanding scenarios where deep learning at the network edge can be useful, understanding common techniques for speeding up deep learning inference and performing distributed training on edge devices, and understanding recent trends and opportunities.
Weighted Rendezvous Planning on Q-Learning Based Adaptive Zone Partition with PSO Based Optimal Path Selection Nowadays, wireless sensor network (WSN) has emerged as the most developed research area. Different research have been demonstrated for reducing the sensor nodes’ energy consumption with mobile sink in WSN. But, such approaches were dependent on the path selected by the mobile sink since all sensed data should be gathered within the given time constraint. Therefore, in this article, the issue of an optimal path selection is solved when multiple mobile sinks are considered in WSN. In the initial stage, Q-learning based Adaptive Zone Partition method is applied to split the network into smaller zones. In each zone, the location and residual energy of nodes are transmitted to the mobile sinks through Mobile Anchor. Moreover, Weighted Rendezvous Planning is proposed to assign a weight to every node according to its hop distance. The collected data packets are transmitted to the mobile sink node within the given delay bound by means of a designated set of rendezvous points (RP). Then, an optimal path from RP to mobile sink is selected utilizing the particle swarm optimization algorithm which is applied during routing process. Experimental results demonstrated the effectiveness of the proposed approach where the network lifetime is increased by the reduction of energy consumption in multihop transmission.
Distinctive Image Features from Scale-Invariant Keypoints This paper presents a method for extracting distinctive invariant features from images that can be used to perform reliable matching between different views of an object or scene. The features are invariant to image scale and rotation, and are shown to provide robust matching across a substantial range of affine distortion, change in 3D viewpoint, addition of noise, and change in illumination. The features are highly distinctive, in the sense that a single feature can be correctly matched with high probability against a large database of features from many images. This paper also describes an approach to using these features for object recognition. The recognition proceeds by matching individual features to a database of features from known objects using a fast nearest-neighbor algorithm, followed by a Hough transform to identify clusters belonging to a single object, and finally performing verification through least-squares solution for consistent pose parameters. This approach to recognition can robustly identify objects among clutter and occlusion while achieving near real-time performance.
An introduction to ROC analysis Receiver operating characteristics (ROC) graphs are useful for organizing classifiers and visualizing their performance. ROC graphs are commonly used in medical decision making, and in recent years have been used increasingly in machine learning and data mining research. Although ROC graphs are apparently simple, there are some common misconceptions and pitfalls when using them in practice. The purpose of this article is to serve as an introduction to ROC graphs and as a guide for using them in research.
A Comprehensive Survey on Internet of Things (IoT) Toward 5G Wireless Systems Recently, wireless technologies have been growing actively all around the world. In the context of wireless technology, fifth-generation (5G) technology has become a most challenging and interesting topic in wireless research. This article provides an overview of the Internet of Things (IoT) in 5G wireless systems. IoT in the 5G system will be a game changer in the future generation. It will open a door for new wireless architecture and smart services. Recent cellular network LTE (4G) will not be sufficient and efficient to meet the demands of multiple device connectivity and high data rate, more bandwidth, low-latency quality of service (QoS), and low interference. To address these challenges, we consider 5G as the most promising technology. We provide a detailed overview of challenges and vision of various communication industries in 5G IoT systems. The different layers in 5G IoT systems are discussed in detail. This article provides a comprehensive review on emerging and enabling technologies related to the 5G system that enables IoT. We consider the technology drivers for 5G wireless technology, such as 5G new radio (NR), multiple-input–multiple-output antenna with the beamformation technology, mm-wave commutation technology, heterogeneous networks (HetNets), the role of augmented reality (AR) in IoT, which are discussed in detail. We also provide a review on low-power wide-area networks (LPWANs), security challenges, and its control measure in the 5G IoT scenario. This article introduces the role of AR in the 5G IoT scenario. This article also discusses the research gaps and future directions. The focus is also on application areas of IoT in 5G systems. We, therefore, outline some of the important research directions in 5G IoT.
Space-time super-resolution. We propose a method for constructing a video sequence of high space-time resolution by combining information from multiple low-resolution video sequences of the same dynamic scene. Super-resolution is performed simultaneously in time and in space. By "temporal super-resolution," we mean recovering rapid dynamic events that occur faster than regular frame-rate. Such dynamic events are not visible (or else are observed incorrectly) in any of the input sequences, even if these are played in "slow-motion." The spatial and temporal dimensions are very different in nature, yet are interrelated. This leads to interesting visual trade-offs in time and space and to new video applications. These include: 1) treatment of spatial artifacts (e.g., motion-blur) by increasing the temporal resolution and 2) combination of input sequences of different space-time resolutions (e.g., NTSC, PAL, and even high quality still images) to generate a high quality video sequence. We further analyze and compare characteristics of temporal super-resolution to those of spatial super-resolution. These include: How many video cameras are needed to obtain increased resolution? What is the upper bound on resolution improvement via super-resolution? What is the temporal analogue to the spatial "ringing" effect?
Data-Driven Intelligent Transportation Systems: A Survey For the last two decades, intelligent transportation systems (ITS) have emerged as an efficient way of improving the performance of transportation systems, enhancing travel security, and providing more choices to travelers. A significant change in ITS in recent years is that much more data are collected from a variety of sources and can be processed into various forms for different stakeholders. The availability of a large amount of data can potentially lead to a revolution in ITS development, changing an ITS from a conventional technology-driven system into a more powerful multifunctional data-driven intelligent transportation system (D2ITS) : a system that is vision, multisource, and learning algorithm driven to optimize its performance. Furthermore, D2ITS is trending to become a privacy-aware people-centric more intelligent system. In this paper, we provide a survey on the development of D2ITS, discussing the functionality of its key components and some deployment issues associated with D2ITS Future research directions for the development of D2ITS is also presented.
Online Prediction of Driver Distraction Based on Brain Activity Patterns This paper presents a new computational framework for early detection of driver distractions (map viewing) using brain activity measured by electroencephalographic (EEG) signals. Compared with most studies in the literature, which are mainly focused on the classification of distracted and nondistracted periods, this study proposes a new framework to prospectively predict the start and end of a distraction period, defined by map viewing. The proposed prediction algorithm was tested on a data set of continuous EEG signals recorded from 24 subjects. During the EEG recordings, the subjects were asked to drive from an initial position to a destination using a city map in a simulated driving environment. The overall accuracy values for the prediction of the start and the end of map viewing were 81% and 70%, respectively. The experimental results demonstrated that the proposed algorithm can predict the start and end of map viewing with relatively high accuracy and can be generalized to individual subjects. The outcome of this study has a high potential to improve the design of future intelligent navigation systems. Prediction of the start of map viewing can be used to provide route information based on a driver's needs and consequently avoid map-viewing activities. Prediction of the end of map viewing can be used to provide warnings for potential long map-viewing durations. Further development of the proposed framework and its applications in driver-distraction predictions are also discussed.
An indoor localization solution using Bluetooth RSSI and multiple sensors on a smartphone. In this paper, we propose an indoor positioning system using a Bluetooth receiver, an accelerometer, a magnetic field sensor, and a barometer on a smartphone. The Bluetooth receiver is used to estimate distances from beacons. The accelerometer and magnetic field sensor are used to trace the movement of moving people in the given space. The horizontal location of the person is determined by received signal strength indications (RSSIs) and the traced movement. The barometer is used to measure the vertical position where a person is located. By combining RSSIs, the traced movement, and the vertical position, the proposed system estimates the indoor position of moving people. In experiments, the proposed approach showed excellent performance in localization with an overall error of 4.8%.
Attitudes Towards Social Robots In Education: Enthusiast, Practical, Troubled, Sceptic, And Mindfully Positive While social robots bring new opportunities for education, they also come with moral challenges. Therefore, there is a need for moral guidelines for the responsible implementation of these robots. When developing such guidelines, it is important to include different stakeholder perspectives. Existing (qualitative) studies regarding these perspectives however mainly focus on single stakeholders. In this exploratory study, we examine and compare the attitudes of multiple stakeholders on the use of social robots in primary education, using a novel questionnaire that covers various aspects of moral issues mentioned in earlier studies. Furthermore, we also group the stakeholders based on similarities in attitudes and examine which socio-demographic characteristics influence these attitude types. Based on the results, we identify five distinct attitude profiles and show that the probability of belonging to a specific profile is affected by such characteristics as stakeholder type, age, education and income. Our results also indicate that social robots have the potential to be implemented in education in a morally responsible way that takes into account the attitudes of various stakeholders, although there are multiple moral issues that need to be addressed first. Finally, we present seven (practical) implications for a responsible application of social robots in education following from our results. These implications provide valuable insights into how social robots should be implemented.
1.2
0.2
0.2
0.2
0.2
0.1
0
0
0
0
0
0
0
0
Detection and Mitigation of DoS and DDoS Attacks in IoT-Based Stateful SDN : An Experimental Approach. The expected advent of the Internet of Things (IoT) has triggered a large demand of embedded devices, which envisions the autonomous interaction of sensors and actuators while offering all sort of smart services. However, these IoT devices are limited in computation, storage, and network capacity, which makes them easy to hack and compromise. To achieve secure development of IoT, it is necessary to engineer scalable security solutions optimized for the IoT ecosystem. To this end, Software Defined Networking (SDN) is a promising paradigm that serves as a pillar in the fifth generation of mobile systems (5G) that could help to detect and mitigate Denial of Service (DoS) and Distributed DoS (DDoS) threats. In this work, we propose to experimentally evaluate an entropy-based solution to detect and mitigate DoS and DDoS attacks in IoT scenarios using a stateful SDN data plane. The obtained results demonstrate for the first time the effectiveness of this technique targeting real IoT data traffic.
Adaptive Clustering with Feature Ranking for DDoS Attacks Detection Distributed Denial of Service (DDoS) attacks pose an increasing threat to the current internet. The detection of such attacks plays an important role in maintaining the security of networks. In this paper, we propose a novel adaptive clustering method combined with feature ranking for DDoS attacks detection. First, based on the analysis of network traffic, preliminary variables are selected. Second, the Modified Global K-means algorithm (MGKM) is used as the basic incremental clustering algorithm to identify the cluster structure of the target data. Third, the linear correlation coefficient is used for feature ranking. Lastly, the feature ranking result is used to inform and recalculate the clusters. This adaptive process can make worthwhile adjustments to the working feature vector according to different patterns of DDoS attacks, and can improve the quality of the clusters and the effectiveness of the clustering algorithm. The experimental results demonstrate that our method is effective and adaptive in detecting the separate phases of DDoS attacks.
The role of KL divergence in anomaly detection We study the role of Kullback-Leibler divergence in the framework of anomaly detection, where its abilities as a statistic underlying detection have never been investigated in depth. We give an in-principle analysis of network attack detection, showing explicitly attacks may be masked at minimal cost through 'camouflage'. We illustrate on both synthetic distributions and ones taken from real traffic.
An Anomaly Detection Model Based on One-Class SVM to Detect Network Intrusions. Intrusion detection occupies a decision position in solving the network security problems. Support Vector Machines (SVMs) are one of the widely used intrusion detection techniques. However, the commonly used two-class SVM algorithms are facing difficulties of constructing the training dataset. That is because in many real application scenarios, normal connection records are easy to be obtained, but attack records are not so. We propose an anomaly detection model based on One-class SVM to detect network intrusions. The one-class SVM adopts only normal network connection records as the training dataset. But after being trained, it is able to recognize normal from various attacks. This just meets the requirements of the anomaly detection. Experimental results on KDDCUP99 dataset show that compared to Probabilistic Neural Network (PNN) and C-SVM, our anomaly detection model based on One-class SVM achieves higher detection rates and yields average better performance in terms of precision, recall and F-value.
Outlier Detection in Sensor Data using Ensemble Learning.
Machine-Learning-Enabled DDoS Attacks Detection in P4 Programmable Networks Distributed Denial of Service (DDoS) attacks represent a major concern in modern Software Defined Networking (SDN), as SDN controllers are sensitive points of failures in the whole SDN architecture. Recently, research on DDoS attacks detection in SDN has focused on investigation of how to leverage data plane programmability, enabled by P4 language, to detect attacks directly in network switches, with marginal involvement of SDN controllers. In order to effectively address cybersecurity management in SDN architectures, we investigate the potential of Artificial Intelligence and Machine Learning (ML) algorithms to perform automated DDoS Attacks Detection (DAD), specifically focusing on Transmission Control Protocol SYN flood attacks. We compare two different DAD architectures, called Standalone and Correlated DAD, where traffic features collection and attack detection are performed locally at network switches or in a single entity (e.g., in SDN controller), respectively. We combine the capability of ML and P4-enabled data planes to implement real-time DAD. Illustrative numerical results show that, for all tested ML algorithms, accuracy, precision, recall and F1-score are above 98% in most cases, and classification time is in the order of few hundreds of mu s in the worst case. Considering real-time DAD implementation, significant latency reduction is obtained when features are extracted at the data plane by using P4 language.
Accurate Self-Localization in RFID Tag Information Grids Using FIR Filtering Grid navigation spaces nested with the radio-frequency identification (RFID) tags are promising for industrial and other needs, because each tag can deliver information about a local two-dimensional or three-dimensional surrounding. The approach, however, requires high accuracy in vehicle self-localization. Otherwise, errors may lead to collisions; possibly even fatal. We propose a new extended finite impulse response (EFIR) filtering algorithm and show that it meets this need. The EFIR filter requires an optimal averaging interval, but does not involve the noise statistics which are often not well known to the engineer. It is more accurate than the extended Kalman filter (EKF) under real operation conditions and its iterative algorithm has the Kalman form. Better performance of the proposed EFIR filter is demonstrated based on extensive simulations in a comparison to EKF, which is widely used in RFID tag grids. We also show that errors in noise covariances may provoke divergence in EKF, whereas the EFIR filter remains stable and is thus more robust.
Evolutionary computation: comments on the history and current state Evolutionary computation has started to receive significant attention during the last decade, although the origins can be traced back to the late 1950's. This article surveys the history as well as the current state of this rapidly growing field. We describe the purpose, the general structure, and the working principles of different approaches, including genetic algorithms (GA) (with links to genetic programming (GP) and classifier systems (CS)), evolution strategies (ES), and evolutionary programming (EP) by analysis and comparison of their most important constituents (i.e. representations, variation operators, reproduction, and selection mechanism). Finally, we give a brief overview on the manifold of application domains, although this necessarily must remain incomplete
Supporting social navigation on the World Wide Web This paper discusses a navigation behavior on Internet information services, in particular the World Wide Web, which is characterized by pointing out of information using various communication tools. We call this behavior social navigation as it is based on communication and interaction with other users, be that through email, or any other means of communication. Social navigation phenomena are quite common although most current tools (like Web browsers or email clients) offer very little support for it. We describe why social navigation is useful and how it can be better supported in future systems. We further describe two prototype systems that, although originally not designed explicitly as tools for social navigation, provide features that are typical for social navigation systems. One of these systems, the Juggler system, is a combination of a textual virtual environment and a Web client. The other system is a prototype of a Web- hotlist organizer, called Vortex. We use both systems to describe fundamental principles of social navigation systems.
Proofs of Storage from Homomorphic Identification Protocols Proofs of storage (PoS) are interactive protocols allowing a client to verify that a server faithfully stores a file. Previous work has shown that proofs of storage can be constructed from any homomorphic linear authenticator (HLA). The latter, roughly speaking, are signature/message authentication schemes where `tags' on multiple messages can be homomorphically combined to yield a `tag' on any linear combination of these messages. We provide a framework for building public-key HLAs from any identification protocol satisfying certain homomorphic properties. We then show how to turn any public-key HLA into a publicly-verifiable PoS with communication complexity independent of the file length and supporting an unbounded number of verifications. We illustrate the use of our transformations by applying them to a variant of an identification protocol by Shoup, thus obtaining the first unbounded-use PoS based on factoring (in the random oracle model).
Design, Implementation, and Experimental Results of a Quaternion-Based Kalman Filter for Human Body Motion Tracking Real-time tracking of human body motion is an important technology in synthetic environments, robotics, and other human-computer interaction applications. This paper presents an extended Kalman filter designed for real-time estimation of the orientation of human limb segments. The filter processes data from small inertial/magnetic sensor modules containing triaxial angular rate sensors, accelerometers, and magnetometers. The filter represents rotation using quaternions rather than Euler angles or axis/angle pairs. Preprocessing of the acceleration and magnetometer measurements using the Quest algorithm produces a computed quaternion input for the filter. This preprocessing reduces the dimension of the state vector and makes the measurement equations linear. Real-time implementation and testing results of the quaternion-based Kalman filter are presented. Experimental results validate the filter design, and show the feasibility of using inertial/magnetic sensor modules for real-time human body motion tracking
Finite-approximation-error-based discrete-time iterative adaptive dynamic programming. In this paper, a new iterative adaptive dynamic programming (ADP) algorithm is developed to solve optimal control problems for infinite horizon discrete-time nonlinear systems with finite approximation errors. First, a new generalized value iteration algorithm of ADP is developed to make the iterative performance index function converge to the solution of the Hamilton-Jacobi-Bellman equation. The ...
Neural network adaptive tracking control for a class of uncertain switched nonlinear systems. •Study the method of the tracking control of the switched uncertain nonlinear systems under arbitrary switching signal controller.•A multilayer neural network adaptive controller with multilayer weight norm adaptive estimation is been designed.•The adaptive law is expand from calculation the second layer weight of neural network to both of the two layers weight.•The controller proposed improve the tracking error performance of the closed-loop system greatly.
Convert Harm Into Benefit: A Coordination-Learning Based Dynamic Spectrum Anti-Jamming Approach This paper mainly investigates the multi-user anti-jamming spectrum access problem. Using the idea of “converting harm into benefit,” the malicious jamming signals projected by the enemy are utilized by the users as the coordination signals to guide spectrum coordination. An “internal coordination-external confrontation” multi-user anti-jamming access game model is constructed, and the existence of Nash equilibrium (NE) as well as correlated equilibrium (CE) is demonstrated. A coordination-learning based anti-jamming spectrum access algorithm (CLASA) is designed to achieve the CE of the game. Simulation results show the convergence, and effectiveness of the proposed CLASA algorithm, and indicate that our approach can help users confront the malicious jammer, and coordinate internal spectrum access simultaneously without information exchange. Last but not least, the fairness of the proposed approach under different jamming attack patterns is analyzed, which illustrates that this approach provides fair anti-jamming spectrum access opportunities under complicated jamming pattern.
1.2
0.2
0.2
0.2
0.2
0.1
0
0
0
0
0
0
0
0
One-class support vector machines: an application in machine fault detection and classification Fast incipient machine fault diagnosis is becoming one of the key requirements for economical and optimal process operation management. Artificial neural networks have been used to detect machine faults for a number of years and shown to be highly successful in this application area. This paper presents a novel test technique for machine fault detection and classification in electro-mechanical machinery from vibration measurements using one-class support vector machines (SVMs). In order to evaluate one-class SVMs, this paper examines the performance of the proposed method by comparing it with that of multilayer perception, one of the artificial neural network techniques, based on real benchmarking data.
Plastic surgery: a new dimension to face recognition Advancement and affordability is leading to the popularity of plastic surgery procedures. Facial plastic surgery can be reconstructive to correct facial feature anomalies or cosmetic to improve the appearance. Both corrective as well as cosmetic surgeries alter the original facial information to a large extent thereby posing a great challenge for face recognition algorithms. The contribution of this research is 1) preparing a face database of 900 individuals for plastic surgery, and 2) providing an analytical and experimental underpinning of the effect of plastic surgery on face recognition algorithms. The results on the plastic surgery database suggest that it is an arduous research challenge and the current state-of-art face recognition algorithms are unable to provide acceptable levels of identification performance. Therefore, it is imperative to initiate a research effort so that future face recognition systems will be able to address this important problem.
Rotation Invariant Local Phase Quantization For Blur Insensitive Texture Analysis This paper introduces a rotation invariant extension to the blur insensitive local phase quantization texture descriptor The new method consists of two stages, the first of which estimates the local characteristic orientation, and the second one extracts a binary descriptor vector Both steps of the algorithm apply the phase of the locally computed Fourier transform coefficients, which can be shown to be insensitive to centrally symmetric image blurring. The new descriptors are assessed in comparison with the well known texture descriptors, local binary patterns (LBP) and Gabor filtering. The results illustrate that the proposed method has superior performance in those cases where the image contains blur and is slightly better even with sharp images.
Stochastic Fuzzy Modeling for Ear Imaging Based Child Identification. The unique identification of children is crucial for information technology supported vaccine delivery to the unprivileged population of third world countries. New robust image matching algorithms are required to match two ear photographs taken under nonstandard real-world conditions such as the presence of unwanted background objects in the photographs. This paper applies stochastic fuzzy models to the robust matching of ear images. The local features of the image regions are extracted using a force-field-like transformation. The extracted features of an image region are modeled by a stochastic fuzzy system. A region of an image is matched to a region of another image by matching the features of an image's region with the model of another image's region. As the model is fuzzy as well as stochastic, a robust matching of features' data to a model is facilitated by handling any uncertainties arising from fuzziness and randomness of the image features. The study introduces an information-theoretic index for measuring the degree of matching between image features and a model of the features. Several experiments are performed on a database of 750 ear-photographs of children (0-6 years) to justify the novel stochastic fuzzy image matching method.
Ear Recognition Using Multi-Scale Histogram of Oriented Gradients Ear recognition is a promising biometric measure, especially with the growing interest in multi-modal biometrics. Histogram of Oriented Gradients (HOG) have been effectively and efficiently used solving the problems of object detection and recognition, especially when illumination variations are present. This work presents a robust approach for ear recognition using multi-scale dense HOG features as a descriptor of 2D ear images. The multi-scale features assure to capture the different and complicated structures of ear images. Dimensionality reduction was performed to avoid feature redundancy and provide a more efficient recognition process while being prone to over-fitting. Finally, a test was performed on a large and realistic database and the results were compared to the state of the art ear recognition approaches tested on the same dataset and under the same test procedure.
Robust log-Gabor filter for ear biometrics Ears are a new biometric with major advantage in that they appear to maintain their structure with increasing age. Expanding on our previous parts-based model, we propose a new wavelet approach. In this, the log-Gabor filter exploits the frequency content of the ear boundary curves. Extending our model description, a specific aim of the new approach is to capture information in the ear¿s outer structures. Ear biometrics is also concerned with the effects of partial occlusion, mostly by hair and earrings. By localization, intuitively a wavelet can offer performance advantage when handling occluded data. We also add a more robust matching strategy to restrict the influence of erroneous wavelet coefficients. Significant improvement is observed when we combine the model and the log-Gabor filter, and we will show that this improvement is maintained as the ears get occluded.
Employing Fusion of Learned and Handcrafted Features for Unconstrained Ear Recognition. The authors present an unconstrained ear recognition framework that outperforms state-of-the-art systems in different publicly available image databases. To this end, they developed convolutional neural network (CNN)-based solutions for ear normalisation and description, they used well-known handcrafted descriptors, and they fused learned and handcrafted features to improve recognition. They desig...
BLEU: a method for automatic evaluation of machine translation Human evaluations of machine translation are extensive but expensive. Human evaluations can take months to finish and involve human labor that can not be reused. We propose a method of automatic machine translation evaluation that is quick, inexpensive, and language-independent, that correlates highly with human evaluation, and that has little marginal cost per run. We present this method as an automated understudy to skilled human judges which substitutes for them when there is need for quick or frequent evaluations.
TripRes: Traffic Flow Prediction Driven Resource Reservation for Multimedia IoV with Edge Computing AbstractThe Internet of Vehicles (IoV) connects vehicles, roadside units (RSUs) and other intelligent objects, enabling data sharing among them, thereby improving the efficiency of urban traffic and safety. Currently, collections of multimedia content, generated by multimedia surveillance equipment, vehicles, and so on, are transmitted to edge servers for implementation, because edge computing is a formidable paradigm for accommodating multimedia services with low-latency resource provisioning. However, the uneven or discrete distribution of the traffic flow covered by edge servers negatively affects the service performance (e.g., overload and underload) of edge servers in multimedia IoV systems. Therefore, how to accurately schedule and dynamically reserve proper numbers of resources for multimedia services in edge servers is still challenging. To address this challenge, a traffic flow prediction driven resource reservation method, called TripRes, is developed in this article. Specifically, the city map is divided into different regions, and the edge servers in a region are treated as a “big edge server” to simplify the complex distribution of edge servers. Then, future traffic flows are predicted using the deep spatiotemporal residual network (ST-ResNet), and future traffic flows are used to estimate the amount of multimedia services each region needs to offload to the edge servers. With the number of services to be offloaded in each region, their offloading destinations are determined through latency-sensitive transmission path selection. Finally, the performance of TripRes is evaluated using real-world big data with over 100M multimedia surveillance records from RSUs in Nanjing China.
Experiment-driven Characterization of Full-Duplex Wireless Systems We present an experiment-based characterization of passive suppression and active self-interference cancellation mechanisms in full-duplex wireless communication systems. In particular, we consider passive suppression due to antenna separation at the same node, and active cancellation in analog and/or digital domain. First, we show that the average amount of cancellation increases for active cance...
Decentralized Plug-in Electric Vehicle Charging Selection Algorithm in Power Systems This paper uses a charging selection concept for plug-in electric vehicles (PEVs) to maximize user convenience levels while meeting predefined circuit-level demand limits. The optimal PEV-charging selection problem requires an exhaustive search for all possible combinations of PEVs in a power system, which cannot be solved for the practical number of PEVs. Inspired by the efficiency of the convex relaxation optimization tool in finding close-to-optimal results in huge search spaces, this paper proposes the application of the convex relaxation optimization method to solve the PEV-charging selection problem. Compared with the results of the uncontrolled case, the simulated results indicate that the proposed PEV-charging selection algorithm only slightly reduces user convenience levels, but significantly mitigates the impact of the PEV-charging on the power system. We also develop a distributed optimization algorithm to solve the PEV-charging selection problem in a decentralized manner, i.e., the binary charging decisions (charged or not charged) are made locally by each vehicle. Using the proposed distributed optimization algorithm, each vehicle is only required to report its power demand rather than report several of its private user state information, mitigating the security problems inherent in such problem. The proposed decentralized algorithm only requires low-speed communication capability, making it suitable for real-time implementation.
Wireless Networks with RF Energy Harvesting: A Contemporary Survey Radio frequency (RF) energy transfer and harvesting techniques have recently become alternative methods to power the next generation wireless networks. As this emerging technology enables proactive energy replenishment of wireless devices, it is advantageous in supporting applications with quality of service (QoS) requirements. In this paper, we present a comprehensive literature review on the research progresses in wireless networks with RF energy harvesting capability, referred to as RF energy harvesting networks (RF-EHNs). First, we present an overview of the RF-EHNs including system architecture, RF energy harvesting techniques and existing applications. Then, we present the background in circuit design as well as the state-of-the-art circuitry implementations, and review the communication protocols specially designed for RF-EHNs. We also explore various key design issues in the development of RFEHNs according to the network types, i.e., single-hop networks, multi-antenna networks, relay networks, and cognitive radio networks. Finally, we envision some open research directions.
Collective feature selection to identify crucial epistatic variants. In this study, we were able to show that selecting variables using a collective feature selection approach could help in selecting true positive epistatic variables more frequently than applying any single method for feature selection via simulation studies. We were able to demonstrate the effectiveness of collective feature selection along with a comparison of many methods in our simulation analysis. We also applied our method to identify non-linear networks associated with obesity.
Robot tutor and pupils’ educational ability: Teaching the times tables Research shows promising results of educational robots in language and STEM tasks. In language, more research is available, occasionally in view of individual differences in pupils’ educational ability levels, and learning seems to improve with more expressive robot behaviors. In STEM, variations in robots’ behaviors have been examined with inconclusive results and never while systematically investigating how differences in educational abilities match with different robot behaviors. We applied an autonomously tutoring robot (without tablet, partly WOz) in a 2 × 2 experiment of social vs. neutral behavior in above-average vs. below-average schoolchildren (N = 86; age 8–10 years) while rehearsing the multiplication tables on a one-to-one basis. The standard school test showed that on average, pupils significantly improved their performance even after 3 occasions of 5-min exercises. Beyond-average pupils profited most from a robot tutor, whereas those below average in multiplication benefited more from a robot that showed neutral rather than more social behavior.
1.215111
0.215111
0.215111
0.215111
0.076889
0.036197
0.006667
0
0
0
0
0
0
0
Using Social Psychology to Motivate Contributions to Online Communities Under-contribution is a problem for many online communities. Social psychology theories of social loafing and goal-setting can lead to mid-level design goals to address this problem. We tested design principles derived from these theories in four field experiments involving members of an online movie recommender community. In each of the experiments participated were given different explanations for the value of their contributions. As predicted by theory, individuals contributed when they were reminded of their uniqueness and when they were given specific and challenging goals. However, other predictions were disconfirmed. For example, in one experiment, participants given group goals contributed more than those given individual goals. The article ends with suggestions and challenges for mining design implications from social science theories.
Motivations for Play in Online Games. An empirical model of player motivations in online games provides the foundation to understand and assess how players differ from one another and how motivations of play relate to age, gender, usage patterns, and in-game behaviors. In the current study, a factor analytic approach was used to create an empirical model of player motivations. The analysis revealed 10 motivation subcomponents that grouped into three overarching components (achievement, social, and immersion). Relationships between motivations and demographic variables (age, gender, and usage patterns) are also presented.
Acceptance of game-based learning by secondary school teachers The adoption and the effectiveness of game-based learning depend largely on the acceptance by classroom teachers, as they can be considered the true change agents of the schools. Therefore, we need to understand teachers' perceptions and beliefs that underlie their decision-making processes. The present study focuses on the factors that influence the acceptance of commercial video games as learning tools in the classroom. A model for describing the acceptance and predicting the uptake of commercial games by secondary school teachers is suggested. Based on data gathered from 505 teachers, the model is tested and evaluated. The results are then linked to previous research in the domains of technology acceptance and game-based learning. Highlights¿ We examine 505 secondary school teachers' acceptance of game-based learning. ¿ We propose, test and evaluate a model for understanding and predicting acceptance. ¿ Teacher beliefs about the use of commercial games appear to be rather complex. ¿ The proposed model explains 57% of the variance in teachers' behavioral intention. ¿ Complexity and experience do not affect behavioral intention in the model.
Influence Of Gamification On Students' Motivation In Using E-Learning Applications Based On The Motivational Design Model Students' motivation is an important factor in ensuring the success of e-learning implementation. In order to ensure students is motivated to use e-learning, motivational design has been used during the development process of e-learning applications. The use of gamification in learning context can help to increase student motivation. The ARCS+G model of motivational design is used as a guide for the gamification of learning. This study focuses on the influence of gamification on students' motivation in using e-learning applications based on the ARCS+G model. Data from the Instructional Materials Motivation Scale (IMMS) questionnaire, were gathered and analyzed for comparison of two groups (one control and one experimental) in attention, relevance, confidence, and satisfaction categories. Based on the result of analysis, students from the experimental group are more motivated to use e-learning applications compared with the controlled group. This proves that gamification affect students' motivation when used in e-learning applications.
Design and Development of a Social, Educational and Affective Robot In this paper we describe the approach and the initial results obtained in the design and implementation of a social and educational robot called Wolly. We involved kids as co-designer helping us in shaping form and behavior of the robot, then we proceeded with the design and implementation of the hardware and software components, characterizing the robot with interactive, adaptive and affective features.
The concept of flow in collaborative game-based learning Generally, high-school students have been characterized as bored and disengaged from the learning process. However, certain educational designs promote excitement and engagement. Game-based learning is assumed to be such a design. In this study, the concept of flow is used as a framework to investigate student engagement in the process of gaming and to explain effects on game performance and student learning outcome. Frequency 1550, a game about medieval Amsterdam merging digital and urban play spaces, has been examined as an exemplar of game-based learning. This 1-day game was played in teams by 216 students of three schools for secondary education in Amsterdam. Generally, these students show flow with their game activities, although they were distracted by solving problems in technology and navigation. Flow was shown to have an effect on their game performance, but not on their learning outcome. Distractive activities and being occupied with competition between teams did show an effect on the learning outcome of students: the fewer students were distracted from the game and the more they were engaged in group competition, the more students learned about the medieval history of Amsterdam. Consequences for the design of game-based learning in secondary education are discussed.
Robot’s adaptive emotional feedback sustains children’s social engagement and promotes their vocabulary learning: a long-term child–robot interaction study AbstractIn this article, we present an emotion and memory model for a social robot. The model allowed the robot to create a memory account of a child’s emotional events over four individual sessions. The robot then adapted its behaviour based on the developed memory. The model was applied on the NAO robot to teach vocabulary to children while playing the popular game ‘Snakes and Ladders’. We conducted an exploratory evaluation of our model with 24 children at a primary school for 2 weeks to verify its impact on children’s long-term social engagement and overall vocabulary learning. Our preliminary results showed that the behaviour generated based on our model was able to sustain social engagement. In addition, it also helped children to improve their vocabulary. We also evaluated the impact of the positive, negative and neutral emotional feedback of the NAO robot on children’s vocabulary learning. Three groups of children (eight per group) interacted with the robot on four separate occasions over a period of 2 weeks. Our results showed that the condition where the robot displayed positive emotional feedback had a significantly positive effect on the child’s vocabulary learning performance as compared to the two other conditions: negative feedback and neutral feedback.
Mastery Grids: An Open Source Social Educational Progress Visualization While many pieces of educational software used in the classroom have been found to positively affect learning, they often are underused by students. Open learning model and social visualization are two approaches which have been helpful in ameliorating that low usage problem. This article introduces a fusion of these two ideas in a form of social progress visualization. A classroom evaluation indicates that this combination may be effective in engaging students, guiding them to suitable content, and enabling faster content access.
A comparative study of texture measures with classification based on featured distributions This paper evaluates the performance both of some texture measures which have been successfully used in various applications and of some new promising approaches proposed recently. For classification a method based on Kullback discrimination of sample and prototype distributions is used. The classification results for single features with one-dimensional feature value distributions and for pairs of complementary features with two-dimensional distributions are presented
Firefly algorithm, stochastic test functions and design optimisation Modern optimisation algorithms are often metaheuristic, and they are very promising in solving NP-hard optimisation problems. In this paper, we show how to use the recently developed firefly algorithm to solve non-linear design problems. For the standard pressure vessel design optimisation, the optimal solution found by FA is far better than the best solution obtained previously in the literature. In addition, we also propose a few new test functions with either singularity or stochastic components but with known global optimality and thus they can be used to validate new optimisation algorithms. Possible topics for further research are also discussed.
Decentralized Plug-in Electric Vehicle Charging Selection Algorithm in Power Systems This paper uses a charging selection concept for plug-in electric vehicles (PEVs) to maximize user convenience levels while meeting predefined circuit-level demand limits. The optimal PEV-charging selection problem requires an exhaustive search for all possible combinations of PEVs in a power system, which cannot be solved for the practical number of PEVs. Inspired by the efficiency of the convex relaxation optimization tool in finding close-to-optimal results in huge search spaces, this paper proposes the application of the convex relaxation optimization method to solve the PEV-charging selection problem. Compared with the results of the uncontrolled case, the simulated results indicate that the proposed PEV-charging selection algorithm only slightly reduces user convenience levels, but significantly mitigates the impact of the PEV-charging on the power system. We also develop a distributed optimization algorithm to solve the PEV-charging selection problem in a decentralized manner, i.e., the binary charging decisions (charged or not charged) are made locally by each vehicle. Using the proposed distributed optimization algorithm, each vehicle is only required to report its power demand rather than report several of its private user state information, mitigating the security problems inherent in such problem. The proposed decentralized algorithm only requires low-speed communication capability, making it suitable for real-time implementation.
Wireless Networks with RF Energy Harvesting: A Contemporary Survey Radio frequency (RF) energy transfer and harvesting techniques have recently become alternative methods to power the next generation wireless networks. As this emerging technology enables proactive energy replenishment of wireless devices, it is advantageous in supporting applications with quality of service (QoS) requirements. In this paper, we present a comprehensive literature review on the research progresses in wireless networks with RF energy harvesting capability, referred to as RF energy harvesting networks (RF-EHNs). First, we present an overview of the RF-EHNs including system architecture, RF energy harvesting techniques and existing applications. Then, we present the background in circuit design as well as the state-of-the-art circuitry implementations, and review the communication protocols specially designed for RF-EHNs. We also explore various key design issues in the development of RFEHNs according to the network types, i.e., single-hop networks, multi-antenna networks, relay networks, and cognitive radio networks. Finally, we envision some open research directions.
Inter-class sparsity based discriminative least square regression Least square regression is a very popular supervised classification method. However, two main issues greatly limit its performance. The first one is that it only focuses on fitting the input features to the corresponding output labels while ignoring the correlations among samples. The second one is that the used label matrix, i.e., zero–one label matrix is inappropriate for classification. To solve these problems and improve the performance, this paper presents a novel method, i.e., inter-class sparsity based discriminative least square regression (ICS_DLSR), for multi-class classification. Different from other methods, the proposed method pursues that the transformed samples have a common sparsity structure in each class. For this goal, an inter-class sparsity constraint is introduced to the least square regression model such that the margins of samples from the same class can be greatly reduced while those of samples from different classes can be enlarged. In addition, an error term with row-sparsity constraint is introduced to relax the strict zero–one label matrix, which allows the method to be more flexible in learning the discriminative transformation matrix. These factors encourage the method to learn a more compact and discriminative transformation for regression and thus has the potential to perform better than other methods. Extensive experimental results show that the proposed method achieves the best performance in comparison with other methods for multi-class classification.
Robot tutor and pupils’ educational ability: Teaching the times tables Research shows promising results of educational robots in language and STEM tasks. In language, more research is available, occasionally in view of individual differences in pupils’ educational ability levels, and learning seems to improve with more expressive robot behaviors. In STEM, variations in robots’ behaviors have been examined with inconclusive results and never while systematically investigating how differences in educational abilities match with different robot behaviors. We applied an autonomously tutoring robot (without tablet, partly WOz) in a 2 × 2 experiment of social vs. neutral behavior in above-average vs. below-average schoolchildren (N = 86; age 8–10 years) while rehearsing the multiplication tables on a one-to-one basis. The standard school test showed that on average, pupils significantly improved their performance even after 3 occasions of 5-min exercises. Beyond-average pupils profited most from a robot tutor, whereas those below average in multiplication benefited more from a robot that showed neutral rather than more social behavior.
1.069537
0.066667
0.066667
0.066667
0.066667
0.018373
0.001111
0.000001
0
0
0
0
0
0