The topologies of predictable dynamic networks are continuously dynamic in terms of node position, network connectivity and link metric. However, their dynamics are almost predictable compared with the ad-hoc network. The existing routing protocols specific to static or ad-hoc network do not consider this predictability and thus are not very efficient for some cases. We present a topology model based on Divide-and-Merge methodology to formulate the dynamic topology into the series of static topologies, which can reflect the topology dynamics correctly with the least number of static topologies. Then we design a dynamic programing algorithm to solve that model and determine the timing of routing update and the topology to be used. Besides, for the classic predictable dynamic network---space Internet, the links at some region have shorter delay, which leads to most traffic converge at these links. Meanwhile, the connectivity and metric of these links continuously vary, which results in a great end-to-end path variations and routing updates. In this paper, we propose a stable routing scheme which adds link life-time into its metric to eliminate these dynamics. And then we take use of the Dijkstra's greedy feature to release some paths from the dynamic link, achieving the goal of routing stability. Experimental results show that our method can significantly decrease the number of changed paths and affected network nodes, and then greatly improve the network stability. Interestingly, our method can also achieve better network performance, including the less number of loss packets, smoother variation of end-to-end delay and higher throughput. We give a complete characterization of the possible response matrices at a fixed frequency of n-terminal electrical networks of inductors, capacitors, resistors and grounds, and of n-terminal discrete linear elastodynamic networks of springs and point masses, both in the three-dimensional case and in the two-dimensional case. Specifically we construct networks which realize any response matrix which is compatible with the known symmetry properties and thermodynamic constraints of response matrices. Due to a mathematical equivalence we also obtain a characterization of the response matrices of discrete acoustic networks. In this paper we propose the Structured Deep Neural Network (structured DNN) as a structured and deep learning framework. This approach can learn to find the best structured object (such as a label sequence) given a structured input (such as a vector sequence) by globally considering the mapping relationships between the structures rather than item by item. When automatic speech recognition is viewed as a special case of such a structured learning problem, where we have the acoustic vector sequence as the input and the phoneme label sequence as the output, it becomes possible to comprehensively learn utterance by utterance as a whole, rather than frame by frame. Structured Support Vector Machine (structured SVM) was proposed to perform ASR with structured learning previously, but limited by the linear nature of SVM. Here we propose structured DNN to use nonlinear transformations in multi-layers as a structured and deep learning approach. This approach was shown to beat structured SVM in preliminary experiments on TIMIT. While deep learning models have achieved state-of-the-art accuracies for many prediction tasks, understanding these models remains a challenge. Despite the recent interest in developing visual tools to help users interpret deep learning models, the complexity and wide variety of models deployed in industry, and the large-scale datasets that they used, pose unique design challenges that are inadequately addressed by existing work. Through participatory design sessions with over 15 researchers and engineers at Facebook, we have developed, deployed, and iteratively improved ActiVis, an interactive visualization system for interpreting large-scale deep learning models and results. By tightly integrating multiple coordinated views, such as a computation graph overview of the model architecture, and a neuron activation view for pattern discovery and comparison, users can explore complex deep neural network models at both the instance- and subset-level. ActiVis has been deployed on Facebook's machine learning platform. We present case studies with Facebook researchers and engineers, and usage scenarios of how ActiVis may work with different models. In this paper we describe the design, and implementation of the Open Science Data Cloud, or OSDC. The goal of the OSDC is to provide petabyte-scale data cloud infrastructure and related services for scientists working with large quantities of data. Currently, the OSDC consists of more than 2000 cores and 2 PB of storage distributed across four data centers connected by 10G networks. We discuss some of the lessons learned during the past three years of operation and describe the software stacks used in the OSDC. We also describe some of the research projects in biology, the earth sciences, and social sciences enabled by the OSDC. A central problem in machine learning involves modeling complex data-sets using highly flexible families of probability distributions in which learning, sampling, inference, and evaluation are still analytically or computationally tractable. Here, we develop an approach that simultaneously achieves both flexibility and tractability. The essential idea, inspired by non-equilibrium statistical physics, is to systematically and slowly destroy structure in a data distribution through an iterative forward diffusion process. We then learn a reverse diffusion process that restores structure in data, yielding a highly flexible and tractable generative model of the data. This approach allows us to rapidly learn, sample from, and evaluate probabilities in deep generative models with thousands of layers or time steps, as well as to compute conditional and posterior probabilities under the learned model. We additionally release an open source reference implementation of the algorithm. This paper reviews the current status and challenges of Neural Networks (NNs) based machine learning approaches for modern power grid stability control including their design and implementation methodologies. NNs are widely accepted as Artificial Intelligence (AI) approaches offering an alternative way to control complex and ill-defined problems. In this paper various application of NNs for power system rotor angle stabilization and control problem is discussed. The main focus of this paper is on the use of Reinforcement Learning (RL) and Supervised Learning (SL) algorithms in power system wide-area control (WAC). Generally, these algorithms due to their capability in modeling nonlinearities and uncertainties are used for transient classification, neuro-control, wide-area monitoring and control, renewable energy management and control, and so on. The works of researchers in the field of conventional and renewable energy systems are reported and categorized. Paper concludes by presenting, comparing and evaluating various learning techniques and infrastructure configurations based on efficiency. We propose a novel semantic segmentation algorithm by learning a deconvolution network. We learn the network on top of the convolutional layers adopted from VGG 16-layer net. The deconvolution network is composed of deconvolution and unpooling layers, which identify pixel-wise class labels and predict segmentation masks. We apply the trained network to each proposal in an input image, and construct the final semantic segmentation map by combining the results from all proposals in a simple manner. The proposed algorithm mitigates the limitations of the existing methods based on fully convolutional networks by integrating deep deconvolution network and proposal-wise prediction; our segmentation method typically identifies detailed structures and handles objects in multiple scales naturally. Our network demonstrates outstanding performance in PASCAL VOC 2012 dataset, and we achieve the best accuracy (72.5%) among the methods trained with no external data through ensemble with the fully convolutional network. In this paper an improved version of the graded precision localization algorithm GRADELOC, called IGRADELOC is proposed. The performance of GRADELOC is dependent on the regions formed by the overlapping radio ranges of the nodes of the underlying sensor network. A different region pattern could significantly alter the nature and precision of localization. In IGRADELOC, two improvements are suggested. Firstly, modifications are proposed in the radio range of the fixed-grid nodes, keeping in mind the actual radio range of commonly available nodes, to allow for routing through them. Routing is not addressed by GRADELOC, but is of prime importance to the deployment of any adhoc network, especially sensor networks. A theoretical model expressing the radio range in terms of the cell dimensions of the grid infrastructure is proposed, to help in carrying out a deployment plan which achieves the desirable precision of coarse-grained localization. Secondly, in GRADELOC it is observed that fine-grained localization does not achieve significant performance benefits over coarse-grained localization. In IGRADELOC, this factor is addressed with the introduction of a parameter that could be used to improve and fine-tune the precision of fine-grained localization. We study transport properties of bulk-disordered quasi-one-dimensional (Q1D) wires paying main attention to the role of long-range correlations embedded into the disorder. First, we show that for stratified disorder for which the disorder is the same for all individual chains forming the Q1D wire, the transport properties can be analytically described provided the disorder is weak. When the disorder in every chain is not the same, however, has the same binary correlator, the general theory is absent. Thus, we consider the case when only one channel is open and all others are closed. For this situation we suggest a semi-analytical approach which is quite effective for the description of the total transmission coefficient. Our numerical data confirm the validity of our approach. Such Q1D disordered structures with anomalous transport properties can be the subject of an experimental study. The difference between surface and deep structures of a spreadsheet is a major cause of difficulty in checking spreadsheets. After a brief survey of current methods of checking (or debugging) spreadsheets, new visual methods of showing the deep structures are presented. Illustrations are given on how these visual methods can be employed in various interactive local and global debugging strategies. The SW has undeniably been one of the most popular network descriptors in the neuroscience literature. Two main reasons for its lasting popularity are its apparent ease of computation and the intuitions it is thought to provide on how networked systems operate. Over the last few years, some pitfalls of the SW construct and, more generally, of network summary measures, have widely been acknowledged. Systems which can spontaneously reveal periodic evolution are dubbed time crystals. This is in analogy with space crystals that display periodic behavior in configuration space. While space crystals are modelled with the help of space periodic potentials, crystalline phenomena in time can be modelled by periodically driven systems. Disorder in the periodic driving can lead to Anderson localization in time: the probability for detecting a system at a fixed point of configuration space becomes exponentially localized around a certain moment in time. We here show that a three-dimensional system exposed to a properly disordered pseudo-periodic driving may display a localized-delocalized Anderson transition in the time domain, in strong analogy with the usual three-dimensional Anderson transition in disordered systems. Such a transition could be experimentally observed with ultra-cold atomic gases. In this paper, we present a comprehensive study of Medium Access Control (MAC) protocols developed for Wireless Body Area Networks (WBANs). In WBANs, small batteryoperated on-body or implanted biomedical sensor nodes are used to monitor physiological signs such as temperature, blood pressure, ElectroCardioGram (ECG), ElectroEncephaloGraphy (EEG) etc. We discuss design requirements for WBANs with major sources of energy dissipation. Then, we further investigate the existing designed protocols for WBANs with focus on their strengths and weaknesses. Paper ends up with concluding remarks and open research issues for future work. We present a self-consistent local approach to self generated glassiness which is based on the concept of the dynamical mean field theory to many body systems. Using a replica approach to self generated glassiness, we map the problem onto an effective local problem which can be solved exactly. Applying the approach to the Brazovskii-model, relevant to a large class of systems with frustrated micro-phase separation, we are able to solve the self-consistent local theory without using additional approximations. We demonstrate that a glassy state found earlier in this model is generic and does not arise from the use of perturbative approximations. In addition we demonstrate that the glassy state depends strongly on the strength of the frustrated phase separation in that model. Results of experiments with liquid 3He immersed in a new type of aerogel are described. This aerogel consists of Al2O3 strands which are nearly parallel to each other, so we call it as a "nematically ordered" aerogel. At all used pressures a superfluid transition was observed and a superfluid phase diagram was measured. Possible structures of the observed superfluid phases are discussed. Applying a new method of rescattering which is based on the neural network technique we study the influence of rescattering on the spectra of strange particles produced in heavy ion reactions. In contradistinction to former approaches the rescattering is done explicitly and not in a perturbative fashion. We present a comparison of our calculations for the system Ni(1.93AGeV)+Ni with recent data of the FOPI collaboration. We find that even for this small system rescattering changes the observables considerably but does not invalidate the role of the kaons as a messenger from the high density zone. We cannot confirm the conjecture that the kaon flow can be of use for the determination of the optical potential of the kaon. Min-SEIS-Cluster is an optimization problem which aims at minimizing the infection spreading in networks. In this problem, nodes can be susceptible to an infection, exposed to an infection, or infectious. One of the main features of this problem is the fact that nodes have different dynamics when interacting with other nodes from the same community. Thus, the problem is characterized by distinct probabilities of infecting nodes from both the same and from different communities. This paper presents a new genetic algorithm that solves the Min-SEIS-Cluster problem. This genetic algorithm surpassed the current heuristic of this problem significantly, reducing the number of infected nodes during the simulation of the epidemics. The results therefore suggest that our new genetic algorithm is the state-of-the-art heuristic to solve this problem. Elements of the pomeron phenomenology within the Regge pole exchange picture are recalled. This includes discussion of the high energy behaviour of total cross-sections, the triple pomeron limit of the diffractive dissociation and the single particle distributions in the central region. The BFKL pomeron and QCD expectations for the small $x$ behaviour of the deep inelastic scattering structure functions are discussed. The dedicated measurements of the hadronic final state in deep inelastic scattering at small $x$ probing the QCD pomeron are described. The deep inelastic diffraction is also discussed. In this paper we introduce WiNV - A framework for web-based interactive scalable network visualization. WiNV enables a new class of rich and scalable interactive cross-platform capabilities for visualizing large-scale networks natively in a user's browser. Extensive experiments show that our system can visualize networks that consist of tens of thousands of nodes while maintaining fast, high-quality interaction. The sensor network is a network technique for the implementation of Ubiquitous computing environment. It is wireless network environment that consists of the many sensors of lightweight and low power. Though sensor network provides various capabilities, it is unable to ensure the secure authentication between nodes. Eventually it causes the losing reliability of the entire network and many secure problems. Therefore, encryption algorithm for the implementation of reliable sensor network environments is required to the applicable sensor network. In this paper, we proposed the solution of reliable sensor network to analyze the communication efficiency through measuring performance of AES encryption algorithm by plaintext size, and cost of operation per hop according to the network scale. We propose an optimization approach to design cost-effective electrical power transmission networks. That is, we aim to select both the network structure and the line conductances (line sizes) so as to optimize the trade-off between network efficiency (low power dissipation within the transmission network) and the cost to build the network. We begin with a convex optimization method based on the paper ``Minimizing Effective Resistance of a Graph'' [Ghosh, Boyd \& Saberi]. We show that this (DC) resistive network method can be adapted to the context of AC power flow. However, that does not address the combinatorial aspect of selecting network structure. We approach this problem as selecting a subgraph within an over-complete network, posed as minimizing the (convex) network power dissipation plus a non-convex cost on line conductances that encourages sparse networks where many line conductances are set to zero. We develop a heuristic approach to solve this non-convex optimization problem using: (1) a continuation method to interpolate from the smooth, convex problem to the (non-smooth, non-convex) combinatorial problem, (2) the majorization-minimization algorithm to perform the necessary intermediate smooth but non-convex optimization steps. Ultimately, this involves solving a sequence of convex optimization problems in which we iteratively reweight a linear cost on line conductances to fit the actual non-convex cost. Several examples are presented which suggest that the overall method is a good heuristic for network design. We also consider how to obtain sparse networks that are still robust against failures of lines and/or generators. Genomic alterations lead to cancer complexity and form a major hurdle for a comprehensive understanding of the molecular mechanisms underlying oncogenesis. In this review, we describe the recent advances in studying cancer-associated genes from a systems biological point of view. The integration of known cancer genes onto protein and signaling networks reveals the characteristics of cancer genes within networks. This approach shows that cancer genes often function as network hub proteins which are involved in many cellular processes and form focal nodes in the information exchange between many signaling pathways. Literature mining allows constructing gene-gene networks, in which new cancer genes can be identified. The gene expression profiles of cancer cells are used for reconstructing gene regulatory networks. By doing so, the genes, which are involved in the regulation of cancer progression, can be picked up from these networks after which their functions can be further confirmed in the laboratory. The intriguing nature of classical Homeric narratives has always fascinated the occidental culture contributing to philosophy, history, mythology and straight forwardly to literature. However what would be so intriguing about Homer's narratives' At a first gaze we shall recognize the very literal appeal and aesthetic pleasure presented on every page across Homer's chants in Odyssey and rhapsodies in Iliad. Secondly we may perceive a biased aspect of its stories contents, varying from real-historical to fictional-mythological. To encompass this glance, there are some new archeological finding that supports historicity of some events described within Iliad, and consequently to Odyssey. Considering these observations and using complex network theory concepts, we managed to built and analyze a social network gathered across the classical epic, Odyssey of Homer. Longing for further understanding, topological quantities were collected in order to classify its social network qualitatively into real or fictional. It turns out that most of the found properties belong to real social networks besides assortativity and giant component's size. In order to test the network's possibilities to be real, we removed some mythological members that could imprint a fictional aspect on the network. Carrying on this maneuver the modified social network resulted on assortative mixing and reduction of the giant component, as expected for real social networks. Overall we observe that Odyssey might be an amalgam of fictional elements plus real based human relations, which corroborates other author's findings for Iliad and archeological evidences. We introduce a novel online Bayesian method for the identification of a family of noisy recurrent neural networks (RNNs). We develop Bayesian active learning technique in order to optimize the interrogating stimuli given past experiences. In particular, we consider the unknown parameters as stochastic variables and use the D-optimality principle, also known as `\emph{infomax method}', to choose optimal stimuli. We apply a greedy technique to maximize the information gain concerning network parameters at each time step. We also derive the D-optimal estimation of the additive noise that perturbs the dynamical system of the RNN. Our analytical results are approximation-free. The analytic derivation gives rise to attractive quadratic update rules. This thesis is divided in two parts. The first presents an overview of known results in statistical mechanics of disordered systems and its approach to random combinatorial optimization problems. The second part is a discussion of two original results. The first result concerns DPLL heuristics for random k-XORSAT, which is equivalent to the diluted Ising p-spin model. It is well known that DPLL is unable to find the ground states in the clustered phase of the problem, i.e. that it leads to contradictions with probability 1. However, no solid argument supports this is general. A class of heuristics, which includes the well known UC and GUC, is introduced and studied. It is shown that any heuristic in this class must fail if the clause to variable ratio is larger than some constant, which depends on the heuristic but is always smaller than the clustering threshold. The second result concerns the properties of random k-SAT at large clause to variable ratios. In this regime, it is well known that the uniform distribution of random instances is dominated by unsatisfiable instances. A general technique (based on the Replica method) to restrict the distribution to satisfiable instances with uniform weight is introduced, and is used to characterize their solutions. It is found that in the limit of large clause to variable ratios, the uniform distribution of satisfiable random k-SAT formulas is asymptotically equal to the much studied Planted distribution. Both results are already published and available as arXiv:0709.0367 and arXiv:cs/0609101 . A more detailed and self-contained derivation is presented here. Network dynamics are typically presented as a time series of network properties captured at each period. The current approach examines the dynamical properties of transmission via novel measures on an integrated, temporally extended network representation of interaction data across time. Because it encodes time and interactions as network connections, static network measures can be applied to this "temporal web" to reveal features of the dynamics themselves. Here we provide the technical details and apply it to agent-based implementations of the well-known SEIR and SEIS epidemiological models. We study three instances of log-correlated processes on the interval: the logarithm of the Gaussian unitary ensemble (GUE) characteristic polynomial, the Gaussian log-correlated potential in presence of edge charges, and the Fractional Brownian motion with Hurst index $H \to 0$ (fBM0). In previous collaborations we obtained the probability distribution function (PDF) of the value of the global minimum (equivalently maximum) for the first two processes, using the {\it freezing-duality conjecture} (FDC). Here we study the PDF of the position of the maximum $x_m$ through its moments. Using replica, this requires calculating moments of the density of eigenvalues in the $\beta$-Jacobi ensemble. Using Jack polynomials we obtain an exact and explicit expression for both positive and negative integer moments for arbitrary $\beta >0$ and positive integer $n$ in terms of sums over partitions. For positive moments, this expression agrees with a very recent independent derivation by Mezzadri and Reynolds. We check our results against a contour integral formula derived recently by Borodin and Gorin (presented in the Appendix A from these authors). The duality necessary for the FDC to work is proved, and on our expressions, found to correspond to exchange of partitions with their dual. Performing the limit $n \to 0$ and to negative Dyson index $\beta \to -2$, we obtain the moments of $x_m$ and give explicit expressions for the lowest ones. Numerical checks for the GUE polynomials, performed independently by N. Simm, indicate encouraging agreement. Some results are also obtained for moments in Laguerre, Hermite-Gaussian, as well as circular and related ensembles. The correlations of the position and the value of the field at the minimum are also analyzed. Efficiency and simplicity of random algorithms have made them a lucrative alternative for solving complex problems in the domain of communication networks. This paper presents a random algorithm for handling the routing problem in Mobile Ad hoc Networks [MANETS].The performance of most existing routing protocols for MANETS degrades in terms of packet delay and congestion caused as the number of mobile nodes increases beyond a certain level or their speed passes a certain level. As the network becomes more and more dynamic, congestion in network increases due to control packets generated by the routing protocols in the process of route discovery and route maintenance. Most of this congestion is due to flooding mechanism used in protocols like AODV and DSDV for the purpose of route discovery and route maintenance or for route discovery as in the case of DSR protocol. This paper introduces the concept of random routing algorithm that neither maintains a routing table nor floods the entire network as done by various known protocols thereby reducing the load on network in terms of number of control packets in a highly dynamic scenario. This paper calculates the expected run time of the designed random algorithm. Conventionally, image denoising and high-level vision tasks are handled separately in computer vision, and their connection is fragile. In this paper, we cope with the two jointly and explore the mutual influence between them, with the focus on two questions, namely (1) how image denoising can help solving high-level vision problems, and (2) how the semantic information from high-level vision tasks can be used to guide image denoising. We propose a deep convolutional neural network solution that cascades two modules for image denoising and various high level tasks, respectively, and propose the use of joint loss for training to allow the semantic information flowing into the optimization of the denoising network via back-propagation. Our experimental results demonstrate that the proposed architecture not only yields superior image denoising results preserving fine details, but also overcomes the performance degradation of different high-level vision tasks, e.g., image classification and semantic segmentation, due to image noise or artifacts caused by conventional denoising approaches such as over-smoothing. Proceedings of the First International Workshop on Deep Learning and Music, joint with IJCNN, Anchorage, US, May 17-18, 2017 A driven Monte Carlo dynamics is introduced to study resistivity scaling in XY-type models in the phase representation. The method is used to study the phase transition of the three-dimensional XY spin glass with a Gaussian coupling distribution. We find a phase-coherence transition at finite temperature in good agreement with recent equilibrium Monte Carlo simulations which shows a single (spin and chiral) glass transition. Estimates of the static and dynamic critical exponents indicate that the critical behavior is in the same universality class as the the model with a bimodal coupling distribution. Relevance of these results for $\pi$-junction superconductors is also discussed. The ability to backpropagate stochastic gradients through continuous latent distributions has been crucial to the emergence of variational autoencoders and stochastic gradient variational Bayes. The key ingredient is an unbiased and low-variance way of estimating gradients with respect to distribution parameters from gradients evaluated at distribution samples. The "reparameterization trick" provides a class of transforms yielding such estimators for many continuous distributions, including the Gaussian and other members of the location-scale family. However the trick does not readily extend to mixture density models, due to the difficulty of reparameterizing the discrete distribution over mixture weights. This report describes an alternative transform, applicable to any continuous multivariate distribution with a differentiable density function from which samples can be drawn, and uses it to derive an unbiased estimator for mixture density weight derivatives. Combined with the reparameterization trick applied to the individual mixture components, this estimator makes it straightforward to train variational autoencoders with mixture-distributed latent variables, or to perform stochastic variational inference with a mixture density variational posterior. In the field of empirical modeling using Genetic Programming (GP), it is important to evolve solution with good generalization ability. Generalization ability of GP solutions get affected by two important issues: bloat and over-fitting. We surveyed and classified existing literature related to different techniques used by GP research community to deal with these issues. We also point out limitation of these techniques, if any. Moreover, the classification of different bloat control approaches and measures for bloat and over-fitting are also discussed. We believe that this work will be useful to GP practitioners in following ways: (i) to better understand concepts of generalization in GP (ii) comparing existing bloat and over-fitting control techniques and (iii) selecting appropriate approach to improve generalization ability of GP evolved solutions. Neurons perform computations, and convey the results of those computations through the statistical structure of their output spike trains. Here we present a practical method, grounded in the information-theoretic analysis of prediction, for inferring a minimal representation of that structure and for characterizing its complexity. Starting from spike trains, our approach finds their causal state models (CSMs), the minimal hidden Markov models or stochastic automata capable of generating statistically identical time series. We then use these CSMs to objectively quantify both the generalizable structure and the idiosyncratic randomness of the spike train. Specifically, we show that the expected algorithmic information content (the information needed to describe the spike train exactly) can be split into three parts describing (1) the time-invariant structure (complexity) of the minimal spike-generating process, which describes the spike train statistically; (2) the randomness (internal entropy rate) of the minimal spike-generating process; and (3) a residual pure noise term not described by the minimal spike-generating process. We use CSMs to approximate each of these quantities. The CSMs are inferred nonparametrically from the data, making only mild regularity assumptions, via the causal state splitting reconstruction algorithm. The methods presented here complement more traditional spike train analyses by describing not only spiking probability and spike train entropy, but also the complexity of a spike train's structure. We demonstrate our approach using both simulated spike trains and experimental data recorded in rat barrel cortex during vibrissa stimulation. There are several centrality measures that have been introduced and studied for real world networks. They account for the different vertex characteristics that permit them to be ranked in order of importance in the network. Betweenness centrality is a measure of the influence of a vertex over the flow of information between every pair of vertices under the assumption that information primarily flows over the shortest path between them. In this paper we present betweenness centrality of some important classes of graphs. In this work, we present a novel approach to ontology reasoning that is based on deep learning rather than logic-based formal reasoning. To this end, we introduce a new model for statistical relational learning that is built upon deep recursive neural networks, and give experimental evidence that it can easily compete with, or even outperform, existing logic-based reasoners on the task of ontology reasoning. More precisely, we compared our implemented system with one of the best logic-based ontology reasoners at present, RDFox, on a number of large standard benchmark datasets, and found that our system attained high reasoning quality, while being up to two orders of magnitude faster. We present experimental evidence for the different mechanisms driving the fluctuations of the local density of states (LDOS) in disordered photonic systems. We establish a clear link between the microscopic structure of the material and the frequency correlation function of LDOS accessed by a near-field hyperspectral imaging technique. We show, in particular, that short- and long-range frequency correlations of LDOS are controlled by different physical processes (multiple or single scattering processes, respectively) that can be---to some extent---manipulated independently. We also demonstrate that the single scattering contribution to LDOS fluctuations is sensitive to subwavelength features of the material and, in particular, to the correlation length of its dielectric function. Our work paves a way towards a complete control of statistical properties of disordered photonic systems, allowing for designing materials with predefined correlations of LDOS. A method is presented, which allows to sample directly low-temperature configurations of glassy systems, like spin glasses. The basic idea is to generate ground states and low lying excited configurations using a heuristic algorithm. Then, with the help of microcanonical Monte Carlo simulations, more configurations are found, clusters of configurations are determined and entropies evaluated. Finally equilibrium configuration are randomly sampled with proper Gibbs-Boltzmann weights. The method is applied to three-dimensional Ising spin glasses with +- J interactions and temperatures T<=0.5. The low-temperature behavior of this model is characterized by evaluating different overlap quantities, exhibiting a complex low-energy landscape for T>0, while the T=0 behavior appears to be less complex. Energy being the very key concern area with sensor networks, so the main focus lies in developing a mechanism to increase the lifetime of a sensor network by energy balancing. To achieve energy balancing and maximizing network lifetime we use an idea of clustering and dividing the whole network into different clusters. In this paper we propose a dynamic cluster formation method where clusters are refreshed periodically based on residual energy, distance and cost. Refreshing clustering minimizes workload of any single node and in turn enhances the energy conservation. Sleep and wait methodology is applied to the proposed protocol to enhance the network lifetime by turning the nodes on and off according to their duties. The node that has some data to be transmitted is in on state and after forwarding its data to the cluster head it changes its state to off which saves the energy of entire network. Simulations have been done using MAT lab. Simulation results prove the betterment of our proposed method over the existing Leach protocol. Co-evolution exhibited by a network system, involving the intricate interplay between the dynamics of the network itself and the subsystems connected by it, is a key concept for understanding the self-organized, flexible nature of real-world network systems. We propose a simple model of such co-evolving network dynamics, in which the diffusion of a resource over a weighted network and the resource-driven evolution of the link weights occur simultaneously. We demonstrate that, under feasible conditions, the network robustly acquires scale-free characteristics in the asymptotic state. Interestingly, in the case that the system includes dissipation, it asymptotically realizes a dynamical phase characterized by an organized scale-free network, in which the ranking of each node with respect to the quantity of the resource possessed thereby changes ceaselessly. Our model offers a unified framework for understanding some real-world diffusion-driven network systems of diverse types. We consider the capacitated selfish replication (CSR) game with binary preferences, over general undirected networks. We first show that such games have an associated ordinary potential function, and hence always admit a pure-strategy Nash equilibrium (NE). Further, when the minimum degree of the network and the number of resources are of the same order, there exists an exact polynomial time algorithm which can find a NE. Following this, we study the price of anarchy of such games, and show that it is bounded above by 3; we further provide some instances for which the price of anarchy is at least 2. We develop a quasi-polynomial algorithm O(n^2D^{ln n}), where n is the number of players and D is the diameter of the network, which can find, in a distributed manner, an allocation profile that is within a constant factor of the optimal allocation, and hence of any pure-strategy NE of the game. Proof of this result uses a novel potential function. Local deep neural networks have been recently introduced for gender recognition. Although, they achieve very good performance they are very computationally expensive to train. In this work, we introduce a simplified version of local deep neural networks which significantly reduces the training time. Instead of using hundreds of patches per image, as suggested by the original method, we propose to use 9 overlapping patches per image which cover the entire face region. This results in a much reduced training time, since just 9 patches are extracted per image instead of hundreds, at the expense of a slightly reduced performance. We tested the proposed modified local deep neural networks approach on the LFW and Adience databases for the task of gender and age classification. For both tasks and both databases the performance is up to 1% lower compared to the original version of the algorithm. We have also investigated which patches are more discriminative for age and gender classification. It turns out that the mouth and eyes regions are useful for age classification, whereas just the eye region is useful for gender classification. One of the famous results of network science states that networks with heterogeneous connectivity are more susceptible to epidemic spreading than their more homogeneous counterparts. In particular, in networks of identical nodes it has been shown that heterogeneity can lower the epidemic threshold at which epidemics can invade the system. Network heterogeneity can thus allow diseases with lower transmission probabilities to persist and spread. Here, we point out that for real world applications, this result should not be regarded independently of the intra-individual heterogeneity between people. Our results show that, if heterogeneity among people is taken into account, networks that are more heterogeneous in connectivity can be more resistant to epidemic spreading. We study a susceptible-infected-susceptible model with adaptive disease avoidance. Results from this model suggest that this reversal of the effect of network heterogeneity is likely to occur in populations in which the individuals are aware of their subjective disease risk. For epidemiology, this implies that network heterogeneity should not be studied in isolation. The Kepler object KIC 12557548 shows irregular eclipsing behaviour with a constant 15.685 hr period, but strongly varying transit depth. In this paper we fit individual eclipses, in addition to fitting binned light curves, to learn more about the process underlying the eclipse depth variation. Additionally, we put forward observational constraints that any model of this planet-star system will have to match. We find two quiescent spells of ~30 orbital periods each where the transit depth is <0.1%, followed by relatively deep transits. Additionally, we find periods of on-off behaviour where >0.5% deep transits are followed by apparently no transit at all. Apart from these isolated events we find neither significant correlation between consecutive transit depths nor a correlation between transit depth and stellar intensity. We find a three-sigma upper limit for the secondary eclipse of 4.9*10^-5, consistent with a planet candidate with a radius of less than 4600 km. Using the short cadence data we find that a 1-D exponential dust tail model is insufficient to explain the data. We improved our model to a 2-D, two-component dust model with an opaque core and an exponential tail. Using this model we fit individual eclipses observed in short cadence mode. We find an improved fit of the data, quantifying earlier suggestions by Budaj (2013) of the necessity of at least two components. We find that deep transits have most absorption in the tail, and not in a disk-shaped, opaque coma, but the transit depth and the total absorption show no correlation with the tail length. Artificial Neural Network computation relies on intensive vector-matrix multiplications. Recently, the emerging nonvolatile memory (NVM) crossbar array showed a feasibility of implementing such operations with high energy efficiency, thus there are many works on efficiently utilizing emerging NVM crossbar array as analog vector-matrix multiplier. However, its nonlinear I-V characteristics restrain critical design parameters, such as the read voltage and weight range, resulting in substantial accuracy loss. In this paper, instead of optimizing hardware parameters to a given neural network, we propose a methodology of reconstructing a neural network itself optimized to resistive memory crossbar arrays. To verify the validity of the proposed method, we simulated various neural network with MNIST and CIFAR-10 dataset using two different specific Resistive Random Access Memory (RRAM) model. Simulation results show that our proposed neural network produces significantly higher inference accuracies than conventional neural network when the synapse devices have nonlinear I-V characteristics. Through research conducted in this study, a network approach to the correlation patterns of void spaces in rough fractures (crack type II) was developed. We characterized friction networks with several networks characteristics. The correlation among network properties with the fracture permeability is the result of friction networks. The revealed hubs in the complex aperture networks confirmed the importance of highly correlated groups to conduct the highlighted features of the dynamical aperture field. We found that there is a universal power law between the nodes' degree and motifs frequency (for triangles it reads T(k)\proptok{\beta} ({\beta} \approx2\pm0.3)). The investigation of localization effects on eigenvectors shows a remarkable difference in parallel and perpendicular aperture patches. Furthermore, we estimate the rate of stored energy in asperities so that we found that the rate of radiated energy is higher in parallel friction networks than it is in transverse directions. The final part of our research highlights 4 point sub-graph distribution and its correlation with fluid flow. For shear rupture, we observed a similar trend in sub-graph distribution, resulting from parallel and transversal aperture profiles (a superfamily phenomenon). The statistical distribution of levels of an integrable system is claimed to be a Poisson distribution. In this paper, we numerically generate an ensemble of N dimensional random diagonal matrices as a model for regular systems. We evaluate the corresponding nearest-neighbor spacing (NNS) distribution, which characterizes the short range correlation between levels. To characterize the long term correlations, we evaluate the level number variance. We show that, by increasing the size of matrices, the level spacing distribution evolves from the Gaussian shape that characterizes ensembles of 2\times2 matrices tending to the Poissonian as N \rightarrow \infty. The transition occurs at N \approx 20. The number variance also shows a gradual transition towards the straight line behavior predicted by the Poisson statistics. Businesses, tourism attractions, public transportation hubs and other points of interest are not isolated but part of a collaborative system. Making such collaborative network surface is not always an easy task. The existence of data-rich environments can assist in the reconstruction of collaborative networks. They shed light into how their members operate and reveal a potential for value creation via collaborative approaches. Social media data are an example of a means to accomplish this task. In this paper, we reconstruct a network of tourist locations using fine-grained data from Flickr, an online community for photo sharing. We have used a publicly available set of Flickr data provided by Yahoo! Labs. To analyse the complex structure of tourism systems, we have reconstructed a network of visited locations in Europe, resulting in around 180,000 vertices and over 32 million edges. An analysis of the resulting network properties reveals its complex structure. A computer-aided detection (CADe) system for microcalcification cluster identification in mammograms has been developed in the framework of the EU-founded MammoGrid project. The CADe software is mainly based on wavelet transforms and artificial neural networks. It is able to identify microcalcifications in different datasets of mammograms (i.e. acquired with different machines and settings, digitized with different pitch and bit depth or direct digital ones). The CADe can be remotely run from GRID-connected acquisition and annotation stations, supporting clinicians from geographically distant locations in the interpretation of mammographic data. We report and discuss the system performances on different datasets of mammograms and the status of the GRID-enabled CADe analysis. Audio Classical Composer Identification (ACC) is an important problem in Music Information Retrieval (MIR) which aims at identifying the composer for audio classical music clips. The famous annual competition, Music Information Retrieval Evaluation eXchange (MIREX), also takes it as one of the four training&testing tasks. We built a hybrid model based on Deep Belief Network (DBN) and Stacked Denoising Autoencoder (SDA) to identify the composer from audio signal. As a matter of copyright, sponsors of MIREX cannot publish their data set. We built a comparable data set to test our model. We got an accuracy of 76.26% in our data set which is better than some pure models and shallow models. We think our method is promising even though we test it in a different data set, since our data set is comparable to that in MIREX by size. We also found that samples from different classes become farther away from each other when transformed by more layers in our model. We discuss the response of a quantum system to a time-dependent perturbation with spectrum \Phi(\omega). This is characterised by a rate constant D describing the diffusion of occupation probability between levels. We calculate the transition rates by first-order perturbation theory, so that multiplying \Phi(\omega) by a constant \lambda changes the diffusion constant to \lambda D. However, we discuss circumstances where this linearity does notextend to the function space of intensities, so that if intensities \Phi_i(\omega) yield diffusion constants D_i, then the intensity \sum_i \Phi_i(\omega) does not result in a diffusion constant \sum_i D_i. This `semilinear' response can occur in the absorption of radiation by small metal particles. A theoretical study of the coherent light scattering from disordered photonic crystal is presented. In addition to the conventional enhancement of the reflected light intensity into the backscattering direction, the so called coherent backscattering (CBS), the periodic modulation of the dielectric function in photonic crystals gives rise to a qualitatively new effect: enhancement of the reflected light intensity in directions different from the backscattering direction. These additional coherent scattering processes, dubbed here {\em umklapp scattering} (CUS), result in peaks, which are most pronounced when the incident light beam enters the sample at an angle close to the the Bragg angle. Assuming that the dielectric function modulation is weak, we study the shape of the CUS peaks for different relative lengths of the modulation-induced Bragg attenuation compared to disorder-induced mean free path. We show that when the Bragg length increases, then the CBS peak assumes its conventional shape, whereas the CUS peak rapidly diminishes in amplitude. We also study the suppression of the CUS peak upon the departure of the incident beam from Bragg resonance: we found that the diminishing of the CUS intensity is accompanied by substantial broadening. In addition, the peak becomes asymmetric. This paper proposes the incremental Bayesian optimization algorithm (iBOA), which modifies standard BOA by removing the population of solutions and using incremental updates of the Bayesian network. iBOA is shown to be able to learn and exploit unrestricted Bayesian networks using incremental techniques for updating both the structure as well as the parameters of the probabilistic model. This represents an important step toward the design of competent incremental estimation of distribution algorithms that can solve difficult nearly decomposable problems scalably and reliably. Natural language correction has the potential to help language learners improve their writing skills. While approaches with separate classifiers for different error types have high precision, they do not flexibly handle errors such as redundancy or non-idiomatic phrasing. On the other hand, word and phrase-based machine translation methods are not designed to cope with orthographic errors, and have recently been outpaced by neural models. Motivated by these issues, we present a neural network-based approach to language correction. The core component of our method is an encoder-decoder recurrent neural network with an attention mechanism. By operating at the character level, the network avoids the problem of out-of-vocabulary words. We illustrate the flexibility of our approach on dataset of noisy, user-generated text collected from an English learner forum. When combined with a language model, our method achieves a state-of-the-art $F_{0.5}$-score on the CoNLL 2014 Shared Task. We further demonstrate that training the network on additional data with synthesized errors can improve performance. Learning latent structure in complex networks has become an important problem fueled by many types of networked data originating from practically all fields of science. In this paper, we propose a new non-parametric Bayesian multiple-membership latent feature model for networks. Contrary to existing multiple-membership models that scale quadratically in the number of vertices the proposed model scales linearly in the number of links admitting multiple-membership analysis in large scale networks. We demonstrate a connection between the single membership relational model and multiple membership models and show on "real" size benchmark network data that accounting for multiple memberships improves the learning of latent structure as measured by link prediction while explicitly accounting for multiple membership result in a more compact representation of the latent structure of networks. Computational color constancy that requires esti- mation of illuminant colors of images is a fundamental yet active problem in computer vision, which can be formulated into a regression problem. To learn a robust regressor for color constancy, obtaining meaningful imagery features and capturing latent correlations across output variables play a vital role. In this work, we introduce a novel deep structured-output regression learning framework to achieve both goals simultaneously. By borrowing the power of deep convolutional neural networks (CNN) originally designed for visual recognition, the proposed framework can automatically discover strong features for white balancing over different illumination conditions and learn a multi-output regressor beyond underlying relationships between features and targets to find the complex interdependence of dif- ferent dimensions of target variables. Experiments on two public benchmarks demonstrate that our method achieves competitive performance in comparison with the state-of-the-art approaches. We consider the spreading of the wave packet in the generalized Rosenzweig-Porter random matrix ensemble in the region of non-ergodic extended states $1<\gamma<2$. We show that despite non-trivial fractal dimensions $0 < D_{q}=2-\gamma<1$ characterize wave function statistics in this region, the wave packet spreading $\langle r^{2} \rangle \propto t^{\beta}$ is governed by the "diffusion" exponent $\beta=1$ outside the ballistic regime $t>\tau\sim 1$ and $\langle r^{2}\rangle \propto t^{2}$ in the ballistic regime for $t<\tau\sim 1$. This demonstrates that the multifractality exhibits itself only in {\it local} quantities like the wave packet survival probability but not in the large-distance spreading of the wave packet. In this paper we study homomorphisms of Probabilistic Regulatory Gene Networks(PRN) introduced in arXiv:math.DS/0603289 v1 13 Mar 2006. The model PRN is a natural generalization of the Probabilistic Boolean Networks (PBN), introduced by I. Shmulevich, E. Dougherty, and W. Zhang in 2001, that has been using to describe genetic networks and has therapeutic applications. In this paper, our main objectives are to apply the concept of homomorphism and $\epsilon$-homomorphism of probabilistic regulatory networks to the dynamic of the networks. The meaning of $\epsilon$ is that these homomorphic networks have similar distributions and the distance between the distributions is upper bounded by $\epsilon$. Additionally, we prove that the class of PRN together with the homomorphisms form a category with products and coproducts. Projections are special homomorphisms, and they always induce invariant subnetworks that contain all the cycles and steady states in the network. Here, it is proved that the $\epsilon$-homomorphism for $0<\epsilon<1$ produce simultaneous Markov Chains in both networks, that permit to introduce the concept of $\epsilon$-isomorphism of Markov Chains, and similar networks. Object category localization is a challenging problem in computer vision. Standard supervised training requires bounding box annotations of object instances. This time-consuming annotation process is sidestepped in weakly supervised learning. In this case, the supervised information is restricted to binary labels that indicate the absence/presence of object instances in the image, without their locations. We follow a multiple-instance learning approach that iteratively trains the detector and infers the object locations in the positive training images. Our main contribution is a multi-fold multiple instance learning procedure, which prevents training from prematurely locking onto erroneous object locations. This procedure is particularly important when using high-dimensional representations, such as Fisher vectors and convolutional neural network features. We also propose a window refinement method, which improves the localization accuracy by incorporating an objectness prior. We present a detailed experimental evaluation using the PASCAL VOC 2007 dataset, which verifies the effectiveness of our approach. In this paper, we propose an improved gravitational search algorithm named GSABC. The algorithm improves gravitational search algorithm (GSA) results improved by using artificial bee colony algorithm (ABC) to solve constrained numerical optimization problems. In GSA, solutions are attracted towards each other by applying gravitational forces, which depending on the masses assigned to the solutions, to each other. The heaviest mass will move slower than other masses and gravitate others. Due to nature of gravitation, GSA may pass global minimum if some solutions stuck to local minimum. ABC updates the positions of the best solutions that has obtained from GSA, preventing the GSA from sticking to the local minimum by its strong searching ability. The proposed algorithm improves the performance of GSA. The proposed method tested on 23 well-known unimodal, multimodal and fixed-point multimodal benchmark test functions. Experimental results show that GSABC outperforms or performs similarly to five state-of-the-art optimization approaches. We propose a novel stacked generalization (stacking) method as a dynamic ensemble technique using a pool of heterogeneous classifiers for node label classification on networks. The proposed method assigns component models a set of functional coefficients, which can vary smoothly with certain topological features of a node. Compared to the traditional stacking model, the proposed method can dynamically adjust the weights of individual models as we move across the graph and provide a more versatile and significantly more accurate stacking model for label prediction on a network. We demonstrate the benefits of the proposed model using both a simulation study and real data analysis. We establish a new framework for statistical estimation of directed acyclic graphs (DAGs) when data are generated from a linear, possibly non-Gaussian structural equation model. Our framework consists of two parts: (1) inferring the moralized graph from the support of the inverse covariance matrix; and (2) selecting the best-scoring graph amongst DAGs that are consistent with the moralized graph. We show that when the error variances are known or estimated to close enough precision, the true DAG is the unique minimizer of the score computed using the reweighted squared l_2-loss. Our population-level results have implications for the identifiability of linear SEMs when the error covariances are specified up to a constant multiple. On the statistical side, we establish rigorous conditions for high-dimensional consistency of our two-part algorithm, defined in terms of a "gap" between the true DAG and the next best candidate. Finally, we demonstrate that dynamic programming may be used to select the optimal DAG in linear time when the treewidth of the moralized graph is bounded. The FORTRAN code POLRAD 2.0 for radiative correction calculation in inclusive and semi-inclusive deep inelastic scattering of polarized leptons by polarized nucleons and nuclei is described. Its theoretical basis, structure and algorithms are discussed in details. Air transportation has been becoming a major part of transportation infrastructure worldwide. Hence the study of the Airports Networks, the backbone of air transportation, is becoming increasingly important. In complex systems domain, airport networks are modeled as graphs (networks) comprising of airports (vertices or nodes) that are linked by flight connectivities among the airports. A complex network analysis of such a model offers holistic insight about the performance and risks in such a network. We review the performance and risks of networks with the help of studies that have been done on some of the airport networks. We present various network parameters those could be potentially used as a measure of performance and risks on airport networks. We will also see how various risks, such as break down of airports, spread of diseases across the airport network could be assessed based on the network parameters. Further we review how these insights could possibly be used to shape more efficient and safer airport networks. The current work addresses quantum machine learning in the context of Quantum Artificial Neural Networks such that the networks' processing is divided in two stages: the learning stage, where the network converges to a specific quantum circuit, and the backpropagation stage where the network effectively works as a self-programing quantum computing system that selects the quantum circuits to solve computing problems. The results are extended to general architectures including recurrent networks that interact with an environment, coupling with it in the neural links' activation order, and self-organizing in a dynamical regime that intermixes patterns of dynamical stochasticity and persistent quasiperiodic dynamics, making emerge a form of noise resilient dynamical record. The prevalent approach to sequence to sequence learning maps an input sequence to a variable length output sequence via recurrent neural networks. We introduce an architecture based entirely on convolutional neural networks. Compared to recurrent models, computations over all elements can be fully parallelized during training and optimization is easier since the number of non-linearities is fixed and independent of the input length. Our use of gated linear units eases gradient propagation and we equip each decoder layer with a separate attention module. We outperform the accuracy of the deep LSTM setup of Wu et al. (2016) on both WMT'14 English-German and WMT'14 English-French translation at an order of magnitude faster speed, both on GPU and CPU. Recent neural IR models have demonstrated deep learning's utility in ad-hoc information retrieval. However, deep models have a reputation for being black boxes, and the roles of a neural IR model's components may not be obvious at first glance. In this work, we attempt to shed light on the inner workings of a recently proposed neural IR model, namely the PACRR model, by visualizing the output of intermediate layers and by investigating the relationship between intermediate weights and the ultimate relevance score produced. We highlight several insights, hoping that such insights will be generally applicable. Cells sense the geometry and stiffness of their adhesive environment by active contractility. For strong adhesion to flat substrates, two-dimensional contractile network models can be used to understand how force is distributed throughout the cell. Here we compare the shape and force distribution for different variants of such network models. In contrast to Hookean networks, cable networks reflect the asymmetric response of biopolymers to tension versus compression. For passive networks, contractility is modeled by a reduced resting length of the mechanical links. In actively contracting networks, a constant force couple is introduced into each link in order to model contraction by molecular motors. If combined with fixed adhesion sites, all network models lead to invaginated cell shapes, but only actively contracting cable networks lead to the circular arc morphology typical for strongly adhering cells. In this case, shape and force distribution are determined by local rather than global determinants and thus are suited to endow the cell with a robust sense of its environment. We also discuss non-linear and adaptive linker mechanics as well as the relation to tissue shape. We present the results of the search for decaying dark matter with particle mass in the 6-40 keV range with NuSTAR deep observations of COSMOS and ECDFS empty sky fields. We show that main contribution to the decaying dark matter signal from the Milky Way galaxy comes through the aperture of the NuSTAR detector, rather than through the focusing optics. High sensitivity of the NuSTAR detector, combined with the large aperture and large exposure times of the two observation fields allow us to improve previously existing constraints on the dark matter decay time by up to an order of magnitude in the mass range 10-30 keV. In the particular case of the nuMSM sterile neutrino dark matter, our constraints impose an upper bound m<20 keV on the dark matter particle mass. We report detection of four unidentified spectral lines in our data set. These line detections are either due to the systematic effects (uncertainties of calibrations of the NuSTAR detectors) or have an astrophysical origin. We discuss different possibilities for testing the nature of the detected lines. It is widely acknowledged that the forthcoming 5G architecture will be highly heterogeneous and deployed with a high degree of density. These changes over the current 4G bring many challenges on how to achieve an efficient operation from the network management perspective. In this article, we introduce a revolutionary vision of the future 5G wireless networks, in which the network is no longer limited by hardware or even software. Specifically, by the idea of virtualizing the wireless networks, which has recently gained increasing attention, we introduce the Everything-as-a-Service (XaaS) taxonomy to light the way towards designing the service-oriented wireless networks. The concepts, challenges along with the research opportunities for realizing XaaS in wireless networks are overviewed and discussed. Image segmentation is an important step in most visual tasks. While convolutional neural networks have shown to perform well on single image segmentation, to our knowledge, no study has been been done on leveraging recurrent gated architectures for video segmentation. Accordingly, we propose a novel method for online segmentation of video sequences that incorporates temporal data. The network is built from fully convolutional element and recurrent unit that works on a sliding window over the temporal data. We also introduce a novel convolutional gated recurrent unit that preserves the spatial information and reduces the parameters learned. Our method has the advantage that it can work in an online fashion instead of operating over the whole input batch of video frames. The network is tested on the change detection dataset, and proved to have 5.5\% improvement in F-measure over a plain fully convolutional network for per frame segmentation. It was also shown to have improvement of 1.4\% for the F-measure compared to our baseline network that we call FCN 12s. We study the ferromagnetic phase transition in a randomly layered Heisenberg model. A recent strong-disorder renormalization group approach [Phys. Rev. B 81, 144407 (2010)] predicted that the critical point in this system is of exotic infinite-randomness type and is accompanied by strong power-law Griffiths singularities. Here, we report results of Monte-Carlo simulations that provide numerical evidence in support of these predictions. Specifically, we investigate the finite-size scaling behavior of the magnetic susceptibility which is characterized by a non-universal power-law divergence in the Griffiths phase. In addition, we calculate the time autocorrelation function of the spins. It features a very slow decay in the Griffiths phase, following a non-universal power law in time. We propose NM landscapes as a new class of tunably rugged benchmark problems. NM landscapes are well-defined on alphabets of any arity, including both discrete and real-valued alphabets, include epistasis in a natural and transparent manner, are proven to have known value and location of the global maximum and, with some additional constraints, are proven to also have a known global minimum. Empirical studies are used to illustrate that, when coefficients are selected from a recommended distribution, the ruggedness of NM landscapes is smoothly tunable and correlates with several measures of search difficulty. We discuss why these properties make NM landscapes preferable to both NK landscapes and Walsh polynomials as benchmark landscape models with tunable epistasis. We present a detailed analytical study of the $A+A\to\emptyset$ diffusion-annihilation process in complex networks. By means of microscopic arguments, we derive a set of rate equations for the density of $A$ particles in vertices of a given degree, valid for any generic degree distribution, and which we solve for uncorrelated networks. For homogeneous networks (with bounded fluctuations), we recover the standard mean-field solution, i.e. a particle density decreasing as the inverse of time. For heterogeneous (scale-free networks) in the infinite network size limit, we obtain instead a density decreasing as a power-law, with an exponent depending on the degree distribution. We also analyze the role of finite size effects, showing that any finite scale-free network leads to the mean-field behavior, with a prefactor depending on the network size. We check our analytical predictions with extensive numerical simulations on homogeneous networks with Poisson degree distribution and scale-free networks with different degree exponents. In this paper, we propose an end-to-end neural network (NN) based EEG-speech (NES) modeling framework, in which three network structures are developed to map imagined EEG signals to phonemes. The proposed NES models incorporate a language model based EEG feature extraction layer, an acoustic feature mapping layer, and a restricted Boltzmann machine (RBM) based the feature learning layer. The NES models can jointly realize the representation of multichannel EEG signals and the projection of acoustic speech signals. Among three proposed NES models, two augmented networks utilize spoken EEG signals as either bias or gate information to strengthen the feature learning and translation of imagined EEG signals. Experimental results show that all three proposed NES models outperform the baseline support vector machine (SVM) method on EEG-speech classification. With respect to binary classification, our approach achieves comparable results relative to deep believe network approach. We propose a duality analysis for obtaining the critical manifold of two-dimensional spin glasses. Our method is based on the computation of quenched free energies with periodic and twisted periodic boundary conditions on a finite basis. The precision can be systematically improved by increasing the size of the basis, leading to very fast convergence towards the thermodynamic limit. We apply the method to obtain the phase diagrams of the random-bond Ising model and $q$-state Potts gauge glasses. In the Ising case, the Nishimori point is found at $p_N = 0.10929 \pm 0.00002$, in agreement with and improving on the precision of existing numerical estimations. Similar precision is found throughout the high-temperature part of the phase diagram. Finite-size effects are larger in the low-temperature region, but our results are in qualitative agreement with the known features of the phase diagram. In particular we show analytically that the critical point in the ground state is located at finite $p_0$. We have measured directly the thermal conductance between electrons and phonons in ultra-thin Hf and Ti films at millikelvin temperatures. The experimental data indicate that electron-phonon coupling in these films is significantly suppressed by disorder. The electron cooling time $\tau_\epsilon$ follows the $T^{-4}$-dependence with a record-long value $\tau_\epsilon=25ms$ at $T=0.04K$. The hot-electron detectors of far-infrared radiation, fabricated from such films, are expected to have a very high sensitivity. The noise equivalent power of a detector with the area $1\mum^2$ would be $(2-3)10^{-20}W/Hz^{1/2}$, which is two orders of magnitude smaller than that of the state-of-the-art bolometers. This paper considers the problem of energy-efficient transmission in multi-flow multihop cooperative wireless networks. Although the performance gains of cooperative approaches are well known, the combinatorial nature of these schemes makes it difficult to design efficient polynomial-time algorithms for joint routing, scheduling and power control. This becomes more so when there is more than one flow in the network. It has been conjectured by many authors, in the literature, that the multiflow problem in cooperative networks is an NP-hard problem. In this paper, we formulate the problem, as a combinatorial optimization problem, for a general setting of $k$-flows, and formally prove that the problem is not only NP-hard but it is $o(n^{1/7-\epsilon})$ inapproxmiable. To our knowledge*, these results provide the first such inapproxmiablity proof in the context of multiflow cooperative wireless networks. We further prove that for a special case of k = 1 the solution is a simple path, and devise a polynomial time algorithm for jointly optimizing routing, scheduling and power control. We then use this algorithm to establish analytical upper and lower bounds for the optimal performance for the general case of $k$ flows. Furthermore, we propose a polynomial time heuristic for calculating the solution for the general case and evaluate the performance of this heuristic under different channel conditions and against the analytical upper and lower bounds. Source localization in ocean acoustics is posed as a machine learning problem in which data-driven methods learn source ranges directly from observed acoustic data. The pressure received by a vertical linear array is preprocessed by constructing a normalized sample covariance matrix (SCM) and used as the input. Three machine learning methods (feed-forward neural networks (FNN), support vector machines (SVM) and random forests (RF)) are investigated in this paper, with focus on the FNN. The range estimation problem is solved both as a classification problem and as a regression problem by these three machine learning algorithms. The results of range estimation for the Noise09 experiment are compared for FNN, SVM, RF and conventional matched-field processing and demonstrate the potential of machine learning for underwater source localization.. The handwritten string recognition is still a challengeable task, though the powerful deep learning tools were introduced. In this paper, based on TAO-FCN, we proposed an end-to-end system for handwritten string recognition. Compared with the conventional methods, there is no preprocess nor manually designed rules employed. With enough labelled data, it is easy to apply the proposed method to different applications. Although the performance of the proposed method may not be comparable with the state-of-the-art approaches, it's usability and robustness are more meaningful for practical applications. The Pacific Rim Application and Grid Middleware Assembly (PRAGMA) is an international community of researchers that actively collaborate to address problems and challenges of common interest in eScience. The PRAGMA Experimental Network Testbed (PRAGMA-ENT) was established with the goal of constructing an international software-defined network (SDN) testbed to offer the necessary networking support to the PRAGMA cyberinfrastructure. PRAGMA-ENT is isolated, and PRAGMA researchers have complete freedom to access network resources to develop, experiment, and evaluate new ideas without the concerns of interfering with production networks. In the first phase, PRAGMA-ENT focused on establishing an international L2 backbone. With support from the Florida Lambda Rail (FLR), Internet2, PacificWave, JGN-X, and TWAREN, PRAGMA-ENT backbone connects Open\-Flow-enabled switches at University of Florida (UF), University of California San Diego (UCSD), Nara Institute of Science and Technology (NAIST, Japan), Osaka University (Japan), National Institute of Advanced Industrial Science and Technology (AIST, Japan), and National Center for High-Performance Computing (Taiwan). The second phase of PRAGMA-ENT consisted of evaluation of technologies for the control plane that enables multiple experiments (i.e., OpenFlow controllers) to co-exist. Preliminary experiments with FlowVisor revealed some limitations leading to the development of a new approach, called AutoVFlow. This paper will share our experience in the establishment of PRAGMA-ENT backbone (with international L2 links), its current status, and control plane plans. Discussion on preliminary application ideas, including optimization of routing control; multipath routing control; and remote visualization will also be discussed. We consider the effect of geometric frustration induced by the random distribution of loop lengths in the "fat" graphs of the dynamical triangulations model on coupled antiferromagnets. While the influence of such connectivity disorder is rather mild for ferromagnets in that an ordered phase persists and only the properties of the phase transition are substantially changed in some cases, any finite-temperature transition is wiped out due to frustration for some of the antiferromagnetic models. A wealth of different phenomena is observed: while for the annealed average of quantum gravity some graphs can adapt dynamically to allow the emergence of a Neel ordered phase, this is not possible for the quenched average, where a zero-temperature spin-glass phase appears instead. We relate the latter to the behaviour of conventional spin-glass models coupled to random graphs. The superconductor-insulator transition in the presence of strong compensation of dopants was recently realized in La doped YBCO. The compensation of acceptors by donors makes it possible to change independently the concentration of holes n and the total concentration of charged impurities N. We propose a theory of the superconductor-insulator phase diagram in the (N,n) plane. It exhibits interesting new features in the case of strong coupling superconductivity, where Cooper pairs are compact, non-overlapping bosons. For compact Cooper pairs the transition occurs at a significantly higher density than in the case of spatially overlapping pairs. We establish the superconductor-insulator phase diagram by studying how the potential of randomly positioned charged impurities is screened by holes or by strongly bound Cooper pairs, both in isotropic and layered superconductors. In the resulting self-consistent potential the carriers are either delocalized or localized, which corresponds to the superconducting or insulating phase, respectively. Identifying and designing physical systems for use as qubits, the basic units of quantum information, are critical steps in the development of a quantum computer. Among the possibilities in the solid state, a defect in diamond known as the nitrogen-vacancy (NV-1) center stands out for its robustness - its quantum state can be initialized, manipulated, and measured with high fidelity at room temperature. Here we describe how to systematically identify other deep center defects with similar quantum-mechanical properties. We present a list of physical criteria that these centers and their hosts should meet and explain how these requirements can be used in conjunction with electronic structure theory to intelligently sort through candidate defect systems. To illustrate these points in detail, we compare electronic structure calculations of the NV-1 center in diamond with those of several deep centers in 4H silicon carbide (SiC). We then discuss the proposed criteria for similar defects in other tetrahedrally-coordinated semiconductors. Observations of deuterated species are useful in probing the temperature, ionization level, evolutionary stage, chemistry, and thermal history of astrophysical environments. The analysis of data from ALMA and other new telescopes requires an elaborate model of deuterium fractionation. This paper presents a publicly available chemical network with multi-deuterated species and an extended, up-to-date set of gas-phase and surface reactions. To test this network, we simulate deuterium fractionation in diverse interstellar sources. Two cases of initial abundances are considered: i) atomic except for H2 and HD, and ii) molecular from a prestellar core. We reproduce the observed D/H ratios of many deuterated molecules, and sort the species according to their sensitivity to temperature gradients and initial abundances. We find that many multiply-deuterated species produced at 10 K retain enhanced D/H ratios at temperatures $\la 100$ K. We study how recent updates to reaction rates affect calculated D/H ratios, and perform a detailed sensitivity analysis of the uncertainties of the gas-phase reaction rates in the network. We find that uncertainties are generally lower in dark cloud environments than in warm IRDCs and that uncertainties increase with the size of the molecule and number of D-atoms. A set of the most problematic reactions is presented. We list potentially observable deuterated species predicted to be abundant in low- and high-mass star-formation regions. A very important topic in systems biology is developing statistical methods that automatically find causal relations in gene regulatory networks with no prior knowledge of causal connectivity. Many methods have been developed for time series data. However, discovery methods based on steady-state data are often necessary and preferable since obtaining time series data can be more expensive and/or infeasible for many biological systems. A conventional approach is causal Bayesian networks. However, estimation of Bayesian networks is ill-posed. In many cases it cannot uniquely identify the underlying causal network and only gives a large class of equivalent causal networks that cannot be distinguished between based on the data distribution. We propose a new discovery algorithm for uniquely identifying the underlying causal network of genes. To the best of our knowledge, the proposed method is the first algorithm for learning gene networks based on a fully identifiable causal model called LiNGAM. We here compare our algorithm with competing algorithms using artificially-generated data, although it is definitely better to test it based on real microarray gene expression data. We propose `Dracula', a new framework for unsupervised feature selection from sequential data such as text. Dracula learns a dictionary of $n$-grams that efficiently compresses a given corpus and recursively compresses its own dictionary; in effect, Dracula is a `deep' extension of Compressive Feature Learning. It requires solving a binary linear program that may be relaxed to a linear program. Both problems exhibit considerable structure, their solution paths are well behaved, and we identify parameters which control the depth and diversity of the dictionary. We also discuss how to derive features from the compressed documents and show that while certain unregularized linear models are invariant to the structure of the compressed dictionary, this structure may be used to regularize learning. Experiments are presented that demonstrate the efficacy of Dracula's features. Many-body localization in a disordered system of interacting spins coupled by the long-range interaction $1/R^{\alpha}$ is investigated combining analytical theory considering resonant interactions and a finite size scaling of exact numerical solutions with a number of spins $N$. The numerical results for a one-dimensional system are consistent with the general expectations of analytical theory for $d$-dimensional system including the absence of localization in the infinite system at $\alpha<2d$ and a universal scaling of a critical energy disordering $W_{c} \propto N^{\frac{2d-\alpha}{d}}$. %The finite size effect on the interaction stimulated delocalization of energy in the ensemble of interacting two level systems in amorphous solids at low temperature is discussed. Deep brain stimulation (DBS) is a surgical treatment for Parkinson's Disease. Static models based on quasi-static approximation are common approaches for DBS modeling. While this simplification has been validated for bioelectric sources, its application to rapid stimulation pulses, which contain more high-frequency power, may not be appropriate, as DBS therapeutic results depend on stimulus parameters such as frequency and pulse width, which are related to time variations of the electric field. We propose an alternative hybrid approach based on probabilistic models and differential equations, by using Gaussian processes and wave equation. Our model avoids quasi-static approximation, moreover, it is able to describe dynamic behavior of DBS. Therefore, the proposed model may be used to obtain a more realistic phenomenon description. The proposed model can also solve inverse problems, i.e. to recover the corresponding source of excitation, given electric potential distribution. The electric potential produced by a time-varying source was predicted using proposed model. For static sources, the electric potential produced by different electrode configurations were modeled. Four different sources of excitation were recovered by solving the inverse problem. We compare our outcomes with the electric potential obtained by solving Poisson's equation using the Finite Element Method (FEM). Our approach is able to take into account time variations of the source and the produced field. Also, inverse problem can be addressed using the proposed model. The electric potential calculated with the proposed model is close to the potential obtained by solving Poisson's equation using FEM. Benes networks are constructed with simple switch modules and have many advantages, including small latency and requiring only an almost linear number of switch modules. As circuit-switches, Benes networks are rearrangeably non-blocking, which implies that they are full-throughput as packet switches, with suitable routing. Routing in Benes networks can be done by time-sharing permutations. However, this approach requires centralized control of the switch modules and statistical knowledge of the traffic arrivals. We propose a backpressure-based routing scheme for Benes networks, combined with end-to-end congestion control. This approach achieves the maximal utility of the network and requires only four queues per module, independently of the size of the network. Previous theoretical studies on the interaction of excitatory and inhibitory neurons proposed to model this cortical microcircuit motif as a so-called Winner-Take-All (WTA) circuit. A recent modeling study however found that the WTA model is not adequate for data-based softer forms of divisive inhibition as found in a microcircuit motif in cortical layer 2/3. We investigate here through theoretical analysis the role of such softer divisive inhibition for the emergence of computational operations and neural codes under spike-timing dependent plasticity (STDP). We show that in contrast to WTA models - where the network activity has been interpreted as probabilistic inference in a generative mixture distribution - this network dynamics approximates inference in a noisy-OR-like generative model that explains the network input based on multiple hidden causes. Furthermore, we show that STDP optimizes the parameters of this model by approximating online the expectation maximization (EM) algorithm. This theoretical analysis corroborates a preceding modelling study which suggested that the learning dynamics of this layer 2/3 microcircuit motif extracts a specific modular representation of the input and thus performs blind source separation on the input statistics. The main aim of this paper is to discuss how the combination of Web 2.0, social media and geographic technologies can provide opportunities for learning and new forms of participation in an urban design studio. This discussion is mainly based on our recent findings from two experimental urban design studio setups as well as former research and literature studies. In brief, the web platform enabled us to extend the learning that took place in the design studio beyond the studio hours, to represent the design information in novel ways and allocate multiple communication forms. We found that the student activity in the introduced web platform was related to their progress up to a certain extent. Moreover, the students perceived the platform as a convenient medium and addressed it as a valuable resource for learning. This study should be conceived as a continuation of a series of our Design Studio 2.0 experiments which involve the exploitation of opportunities provided by novel socio-geographic information and communication technologies for the improvement of the design learning processes. Over the last few years, Cloud Radio Access Network (C-RAN) has arisen as a transformative architecture for 5G cellular networks that brings the flexibility and agility of cloud computing to wireless communications. At the same time, content caching in wireless networks has become an essential solution to lower the content-access latency and backhaul traffic loading, which translate into user Quality of Experience (QoE) improvement and network cost reduction. In this article, a novel Cooperative Hierarchical Caching (CHC) framework in C-RAN is introduced where contents are jointly cached at the BaseBand Unit (BBU) and at the Radio Remote Heads (RRHs). Unlike in traditional approaches, the cache at the BBU, cloud cache, presents a new layer in the cache hierarchy, bridging the latency/capacity gap between the traditional edge-based and core-based caching schemes. Trace-driven simulations reveal that CHC yields up to 80% improvement in cache hit ratio, 21% decrease in average content-access latency, and 20% reduction in backhaul traffic load compared to the edge-only caching scheme with the same total cache capacity. Before closing the article, several challenges and promising opportunities for deploying content caching in C-RAN are highlighted towards a content-centric mobile wireless network. We model the cooperation policy with only two parameters -- search radius $r$ and number of copies in the network $N_{copy}$. These two parameters represent the range of cooperation and tolerance of duplicates. We show how cooperation policy impacts content distribution, and further illustrate the relation between content popularity and topological properties. Our work leads many implications on how to take advantage of topological properties in in-network caching strategy design. We focus on constructing the domi-join model by doing the join operation based on two smallest dominating sets of two network models and analysis the properties of domi-join model, such as power law distribution, small world. Besides, we will import two class of edge-bound growing network models to explain the process of domi-join model. Then we compute the average degree, clustering coefficient, power law distribution of the domi-join model. Finally, we discuss an impressive method for cutting down redundant operation of domi-join model. In order to better accommodate the dramatically increasing demand for data caching and computing services, storage and computation capabilities should be endowed to some of the intermediate nodes within the network. In this paper, we design a novel virtualized heterogeneous networks framework aiming at enabling content caching and computing. With the virtualization of the whole system, the communication, computing and caching resources can be shared among all users associated with different virtual service providers. We formulate the virtual resource allocation strategy as a joint optimization problem, where the gains of not only virtualization but also caching and computing are taken into consideration in the proposed architecture. In addition, a distributed algorithm based on alternating direction method of multipliers is adopted to solve the formulated problem, in order to reduce the computational complexity and signaling overhead. Finally, extensive simulations are presented to show the effectiveness of the proposed scheme under different system parameters. This paper introduces the probabilistic module interface, which allows encapsulation of complex probabilistic models with latent variables alongside custom stochastic approximate inference machinery, and provides a platform-agnostic abstraction barrier separating the model internals from the host probabilistic inference system. The interface can be seen as a stochastic generalization of a standard simulation and density interface for probabilistic primitives. We show that sound approximate inference algorithms can be constructed for networks of probabilistic modules, and we demonstrate that the interface can be implemented using learned stochastic inference networks and MCMC and SMC approximate inference programs. We present a sparse and invariant representation with low asymptotic complexity for robust unsupervised transient and onset zone detection in noisy environments. This unsupervised approach is based on wavelet transforms and leverages the scattering network from Mallat et al. by deriving frequency invariance. This frequency invariance is a key concept to enforce robust representations of transients in presence of possible frequency shifts and perturbations occurring in the original signal. Implementation details as well as complexity analysis are provided in addition of the theoretical framework and the invariance properties. In this work, our primary application consists of predicting the onset of seizure in epileptic patients from subdural recordings as well as detecting inter-ictal spikes. An associative memory is a framework of content-addressable memory that stores a collection of message vectors (or a dataset) over a neural network while enabling a neurally feasible mechanism to recover any message in the dataset from its noisy version. Designing an associative memory requires addressing two main tasks: 1) learning phase: given a dataset, learn a concise representation of the dataset in the form of a graphical model (or a neural network), 2) recall phase: given a noisy version of a message vector from the dataset, output the correct message vector via a neurally feasible algorithm over the network learnt during the learning phase. This paper studies the problem of designing a class of neural associative memories which learns a network representation for a large dataset that ensures correction against a large number of adversarial errors during the recall phase. Specifically, the associative memories designed in this paper can store dataset containing $\exp(n)$ $n$-length message vectors over a network with $O(n)$ nodes and can tolerate $\Omega(\frac{n}{{\rm polylog} n})$ adversarial errors. This paper carries out this memory design by mapping the learning phase and recall phase to the tasks of dictionary learning with a square dictionary and iterative error correction in an expander code, respectively. To dynamically detect the facial landmarks in the video, we propose a novel hybrid framework termed as detection-tracking-detection (DTD). First, the face bounding box is achieved from the first frame of the video sequence based on a traditional face detection method. Then, a landmark detector detects the facial landmarks, which is based on a cascaded deep convolution neural network (DCNN). Next, the face bounding box in the current frame is estimated and validated after the facial landmarks in the previous frame are tracked based on the median flow. Finally, the facial landmarks in the current frame are exactly detected from the validated face bounding box via the landmark detector. Experimental results indicate that the proposed framework can detect the facial landmarks in the video sequence more effectively and with lower consuming time compared to the frame-by-frame method via the DCNN. To avoid the complicated topology of surviving clusters induced by standard Strong Disorder RG in dimension $d>1$, we introduce a modified procedure called 'Boundary Strong Disorder RG' where the order of decimations is chosen a priori. We apply numerically this modified procedure to the Random Transverse Field Ising model in dimension $d=2$. We find that the location of the critical point, the activated exponent $\psi \simeq 0.5$ of the Infinite Disorder scaling, and the finite-size correlation exponent $\nu_{FS} \simeq 1.3$ are compatible with the values obtained previously by standard Strong Disorder RG.Our conclusion is thus that Strong Disorder RG is very robust with respect to changes in the order of decimations. In addition, we analyze in more details the RG flows within the two phases to show explicitly the presence of various correlation length exponents : we measure the typical correlation exponent $\nu_{typ} \simeq 0.64$ in the disordered phase (this value is very close to the correlation exponent $\nu^Q_{pure}(d=2) \simeq 0.63$ of the {\it pure} two-dimensional quantum Ising Model), and the typical exponent $\nu_h \simeq 1$ within the ordered phase. These values satisfy the relations between critical exponents imposed by the expected finite-size scaling properties at Infinite Disorder critical points. Within the disordered phase, we also measure the fluctuation exponent $\omega \simeq 0.35$ which is compatible with the Directed Polymer exponent $\omega_{DP}(1+1)=1/3$ in $(1+1)$ dimensions. We survey the contributions presented in the working group ``Diffraction and Vector Mesons'' at the XIV International Workshop on Deep Inelastic Scattering. In dissipationless linear media, spatial disorder induces Anderson localization of matter, light, and sound waves. The addition of nonlinearity causes interaction between the eigenmodes, which results in a slow wave diffusion. We go beyond the dissipationless limit of Anderson arrays and consider nonlinear disordered systems that are subjected to the dissipative losses and energy pumping. We show that the Anderson modes of the disordered Ginsburg-Landau lattice possess specific excitation thresholds with respect to the pumping strength. When pumping is increased above the threshold for the band-edge modes, the lattice dynamics yields an attractor in the form of a stable multi-peak pattern. The Anderson attractor is the result of a joint action by the pumping-induced mode excitation, nonlinearity-induced mode interactions, and dissipative stabilization. The regimes of Anderson attractors can be potentially realized with polariton condensates lattices, active waveguide or cavity-QED arrays. Complex systems are successfully reduced to interacting elements via the network concept. Transport plays a key role in the survival of networks. For example the specialized signaling cascades of cellular networks filter noise and efficiently adapt the network structure to new stimuli. However, our general understanding of transport mechanisms and signaling pathways in complex systems is yet limited. Here we summarize the key network structures involved in transport, list the solutions available to overloaded systems for relaxing their load and outline a possible method for the computational determination of signaling pathways. We highlight that in addition to hubs, bridges and the network skeleton, the overlapping modular structure is also essential in network transport. Moreover, by locating network elements in the space of overlapping network modules and evaluating their distance in this "module space", it may be possible to approximate signaling pathways computationally, which, in turn could serve the identification of signaling pathways of complex systems. Our model may be applicable in a wide range of fields including traffic control or drug design. We use the annealed formulation of complex networks to study the dynamical behavior of disease spreading on both static and adaptive networked systems. This unifying approach relies on the annealed adjacency matrix, representing one network ensemble, and allows to solve the dynamical evolution of the whole network ensemble all at once. Our results accurately reproduce those obtained by extensive numerical simulations showing a large improvement with respect to the usual heterogeneous mean-field formulation. Moreover, by means of the annealed formulation we derive a new heterogeneous mean-field formulation that correctly reproduces the epidemic dynamics. Introduction to the Special Issue on Complex Networks, Artificial Life journal. Recommendation algorithms that incorporate techniques from deep learning are becoming increasingly popular. Due to the structure of the data coming from recommendation domains (i.e., one-hot-encoded vectors of item preferences), these algorithms tend to have large input and output dimensionalities that dominate their overall size. This makes them difficult to train, due to the limited memory of graphical processing units, and difficult to deploy on mobile devices with limited hardware. To address these difficulties, we propose Bloom embeddings, a compression technique that can be applied to the input and output of neural network models dealing with sparse high-dimensional binary-coded instances. Bloom embeddings are computationally efficient, and do not seriously compromise the accuracy of the model up to 1/5 compression ratios. In some cases, they even improve over the original accuracy, with relative increases up to 12%. We evaluate Bloom embeddings on 7 data sets and compare it against 4 alternative methods, obtaining favorable results. We also discuss a number of further advantages of Bloom embeddings, such as 'on-the-fly' constant-time operation, zero or marginal space requirements, training time speedups, or the fact that they do not require any change to the core model architecture or training configuration. We study the dynamics of excitations in a system of $O(N)$ quantum rotors in the presence of random fields and random anisotropies. Below the lower critical dimension $d_{\mathrm{lc}}=4$ the system exhibits a quasi-long-range order with a power-law decay of correlations. At zero temperature the spin waves are localized at the length scale $L_{\mathrm{loc}}$ beyond which the quantum tunneling is exponentially suppressed $ c \sim e^{-(L/L_{\mathrm{loc}})^{2(\theta+1)}}$. At finite temperature $T$ the spin waves propagate by thermal activation over energy barriers that scales as $L^{\theta}$. Above $d_{\mathrm{lc}}$ the system undergoes an order-disorder phase transition with activated dynamics such that the relaxation time grows with the correlation length $\xi$ as $\tau \sim e^{C \xi^\theta/T}$ at finite temperature and as $\tau \sim e^{C' \xi^{2(\theta+1)}/\hbar^2}$ in the vicinity of the quantum critical point. For a linear code, deep holes are defined to be vectors that are further away from codewords than all other vectors. The problem of deciding whether a received word is a deep hole for generalized Reed-Solomon codes is proved to be co-NP-complete. For the extended Reed-Solomon codes $RS_q(\F_q,k)$, a conjecture was made to classify deep holes by Cheng and Murray in 2007. Since then a lot of effort has been made to prove the conjecture, or its various forms. In this paper, we classify deep holes completely for generalized Reed-Solomon codes $RS_p (D,k)$, where $p$ is a prime, $|D| > k \geqslant \frac{p-1}{2}$. Our techniques are built on the idea of deep hole trees, and several results concerning the Erd{\"o}s-Heilbronn conjecture. Trees have long been used as a graphical representation of species relationships. However complex evolutionary events, such as genetic reassortments or hybrid speciations which occur commonly in viruses, bacteria and plants, do not fit into this elementary framework. Alternatively, various network representations have been developed. Circular networks are a natural generalization of leaf-labeled trees interpreted as split systems, that is, collections of bipartitions over leaf labels corresponding to current species. Although such networks do not explicitly model specific evolutionary events of interest, their straightforward visualization and fast reconstruction have made them a popular exploratory tool to detect network-like evolution in genetic datasets. Standard reconstruction methods for circular networks, such as Neighbor-Net, rely on an associated metric on the species set. Such a metric is first estimated from DNA sequences, which leads to a key difficulty: distantly related sequences produce statistically unreliable estimates. This is problematic for Neighbor-Net as it is based on the popular tree reconstruction method Neighbor-Joining, whose sensitivity to distance estimation errors is well established theoretically. In the tree case, more robust reconstruction methods have been developed using the notion of a distorted metric, which captures the dependence of the error in the distance through a radius of accuracy. Here we design the first circular network reconstruction method based on distorted metrics. Our method is computationally efficient. Moreover, the analysis of its radius of accuracy highlights the important role played by the maximum incompatibility, a measure of the extent to which the network differs from a tree. We present a study of the application of a variant of a recently introduced heuristic algorithm for the optimization of transport routes on complex networks to the problem of finding the optimal routes of communication between nodes on wireless networks. Our algorithm iteratively balances network traffic by minimizing the maximum node betweenness on the network. The variant we consider specifically accounts for the broadcast restrictions imposed by wireless communication by using a different betweenness measure. We compare the performance of our algorithm to two other known algorithms and find that our algorithm achieves the highest transport capacity both for minimum node degree geometric networks, which are directed geometric networks that model wireless communication networks, and for configuration model networks that are uncorrelated scale-free networks. The exact closed-form expressions for outage probability and bit error rate of spectrum sharing-based multi-hop decodeand- forward (DF) relay networks in non-identical Rayleigh fading channels are derived. We also provide the approximate closed-form expression for the system ergodic capacity. Utilizing these tractable analytical formulas, we can study the impact of key network parameters on the performance of cognitivemulti-hop relay networks under interference constraints. Using a linear network model, we derive an optimum relay position scheme by numerically solving an optimization problem of balancing average signal-to-noise ratio (SNR) of each hop. The numerical results show that the optimal scheme leads to SNR performance gains of more than 1 dB. All the analytical expressions are verified by Monte-Carlo simulations confirming the advantage ofmultihop DF relaying networks in cognitive environments. Deep neural networks (DNNs) provide useful models of visual representational transformations. We present a method that enables a DNN (student) to learn from the internal representational spaces of a reference model (teacher), which could be another DNN or, in the future, a biological brain. Representational spaces of the student and the teacher are characterized by representational distance matrices (RDMs). We propose representational distance learning (RDL), a stochastic gradient descent method that drives the RDMs of the student to approximate the RDMs of the teacher. We demonstrate that RDL is competitive with other transfer learning techniques for two publicly available benchmark computer vision datasets (MNIST and CIFAR-100), while allowing for architectural differences between student and teacher. By pulling the student's RDMs towards those of the teacher, RDL significantly improved visual classification performance when compared to baseline networks that did not use transfer learning. In the future, RDL may enable combined supervised training of deep neural networks using task constraints (e.g. images and category labels) and constraints from brain-activity measurements, so as to build models that replicate the internal representational spaces of biological brains. It has been recently observed that the dynamical properties of mass action systems arising from many models of biochemical reaction networks can be derived by considering the corresponding properties of a related generalized mass action system. The correspondence process known as network translation in particular has been shown to be useful in characterizing a system's steady states. In this paper, we further develop the theory of network translation with particular focus on a subclass of translations known as improper translations. For these translations, we derive conditions on the network topology of the translated network which are sufficient to guarantee the original and translated systems share the same steady states. We then present a mixed-integer linear programming (MILP) algorithm capable of determining whether a mass action system can be corresponded to a generalized system through the process of network translation. The Shintani-Tanaka model is a glass-forming system whose constituents interact via anisotropic potential depending on the angle of a unit vector carried by each particle. The decay of time-correlation functions of the unit vectors exhibits the characteristics of generic relaxation functions during glass transitions. In particular it exhibits a 'stretched exponential' form, with the stretching index beta depending strongly on the temperature. We construct a quantitative theory of this correlation function by analyzing all the physical processes that contribute to it, separating a rotational from a translational decay channel. Interestingly, the separate decay function of each of these processes is temperature independent. Taken together with temperature-dependent weights determined a-priori by statistical mechanics one generates the observed correlation function in quantitative agreement with simulations at different temperatures. This underlines the danger of concluding anything about glassy relaxation functions without detailed physical scrutiny. Using the dedicated computer Janus, we follow the nonequilibrium dynamics of the Ising spin glass in three dimensions for eleven orders of magnitude. The use of integral estimators for the coherence and correlation lengths allows us to study dynamic heterogeneities and the presence of a replicon mode and to obtain safe bounds on the Edwards-Anderson order parameter below the critical temperature. We obtain good agreement with experimental determinations of the temperature-dependent decay exponents for the thermoremanent magnetization. This magnitude is observed to scale with the much harder to measure coherence length, a potentially useful result for experimentalists. The exponents for energy relaxation display a linear dependence on temperature and reasonable extrapolations to the critical point. We conclude examining the time growth of the coherence length, with a comparison of critical and activated dynamics. Models of neural networks have proven their utility in the development of learning algorithms in computer science and in the theoretical study of brain dynamics in computational neuroscience. We propose in this paper a spatial neural network model to analyze the important class of functional networks, which are commonly employed in computational studies of clinical brain imaging time series. We developed a simulation framework inspired by multichannel brain surface recordings (more specifically, EEG -- electroencephalogram) in order to link the mesoscopic network dynamics (represented by sampled functional networks) and the microscopic network structure (represented by an integrate-and-fire neural network located in a 3D space -- hence the term spatial neural network). Functional networks are obtained by computing pairwise correlations between time-series of mesoscopic electric potential dynamics, which allows the construction of a graph where each node represents one time-series. The spatial neural network model is central in this study in the sense that it allowed us to characterize sampled functional networks in terms of what features they are able to reproduce from the underlying spatial network. Our modeling approach shows that, in specific conditions of sample size and edge density, it is possible to precisely estimate several network measurements of spatial networks by just observing functional samples. Emergent behaviors are in the focus of recent research interest. It is then of considerable importance to investigate what optimizations suit the learning and prediction of chaotic systems, the putative candidates for emergence. We have compared L1 and L2 regularizations on predicting chaotic time series using linear recurrent neural networks. The internal representation and the weights of the networks were optimized in a unifying framework. Computational tests on different problems indicate considerable advantages for the L1 regularization: It had considerably better learning time and better interpolating capabilities. We shall argue that optimization viewed as a maximum likelihood estimation justifies our results, because L1 regularization fits heavy-tailed distributions -- an apparently general feature of emergent systems -- better. In recent years, there have been many computational simulations of spontaneous neural dynamics. Here, we explore a model of spontaneous neural dynamics and allow it to control a virtual agent moving in a simple environment. This setup generates interesting brain-environment feedback interactions that rapidly destabilize neural and behavioral dynamics and suggest the need for homeostatic mechanisms. We investigate roles for both local homeostatic plasticity (local inhibition adjusting over time to balance excitatory input) as well as macroscopic task negative activity (that compensates for task positive, sensory input) in regulating both neural activity and resulting behavior (trajectories through the environment). Our results suggest complementary functional roles for both local homeostatic plasticity and balanced activity across brain regions in maintaining neural and behavioral dynamics. These findings suggest important functional roles for homeostatic systems in maintaining neural and behavioral dynamics and suggest a novel functional role for frequently reported macroscopic task-negative patterns of activity (e.g., the default mode network). In this study, we investigate the complexity of two-phase flow (air/water) in a heterogeneous soil sample by using complex network theory, where the supposed porous media is non-deformable media, under the time-dependent gas pressure. Based on the different similarity measurements (i.e., correlation, Euclidean metrics) over the emerged patterns from the evolution of saturation of non-wetting phase of a multi-heterogeneous soil sample, the emerged complex networks are recognized. Understanding of the properties of complex networks (such degree distribution, mean path length, clustering coefficient) can be supposed as a way to analysis of variation of saturation profiles structures (as the solution of finite element method on the coupled PDEs) where complexity is coming from the changeable connection and links between assumed nodes. Also, the path of evolution of the supposed system will be illustrated on the state space of networks either in correlation and Euclidean measurements. The results of analysis showed in a closed system the designed complex networks approach to small world network where the mean path length and clustering coefficient are low and high, respectively. As another result, the evolution of macro -states of system (such mean velocity of air or pressure) can be scaled with characteristics of structure complexity of saturation. In other part, we tried to find a phase transition criterion based on the variation of non-wetting phase velocity profiles over a network which had been constructed over correlation distance. With a simple attack and repair evolution model, we investigate and compare the stability of the Erdos-Renyi random graphs (RG) and Barabasi-Albert scale-free (SF) networks. We introduce a new quantity, invulnerability I(s), to describe the stability of the system. We find that both RG and SF networks can evolve to a stationary state. The stationary value Ic has a power-law dependence on the average degree _rg for RG networks; and an exponential relationship with the repair probability p_sf for SF networks. We also discuss the topological changes of RG and SF networks between the initial and stationary states. We observe that the networks in the stationary state have smaller average degree but larger clustering coefficient C and stronger assortativity r. Neural machine translation (NMT) models are able to partially learn syntactic information from sequential lexical information. Still, some complex syntactic phenomena such as prepositional phrase attachment are poorly modeled. This work aims to answer two questions: 1) Does explicitly modeling target language syntax help NMT? 2) Is tight integration of words and syntax better than multitask training? We introduce syntactic information in the form of CCG supertags in the decoder, by interleaving the target supertags with the word sequence. Our results on WMT data show that explicitly modeling target-syntax improves machine translation quality for German->English, a high-resource pair, and for Romanian->English, a low-resource pair and also several syntactic phenomena including prepositional phrase attachment. Furthermore, a tight coupling of words and syntax improves translation quality more than multitask training. By combining target-syntax with adding source-side dependency labels in the embedding layer, we obtain a total improvement of 0.9 BLEU for German->English and 1.2 BLEU for Romanian->English. Spiking Neural Networks (SNN) are more closely related to brain-like computation and inspire hardware implementation. This is enabled by small networks that give high performance on standard classification problems. In literature, typical SNNs are deep and complex in terms of network structure, weight update rules and learning algorithms. This makes it difficult to translate them into hardware. In this paper, we first develop a simple 2-layered network in software which compares with the state of the art on four different standard data-sets within SNNs and has improved efficiency. For example, it uses lower number of neurons (3 x), synapses (3.5 x) and epochs for training (30 x) for the Fisher Iris classification problem. The efficient network is based on effective population coding and synapse-neuron co-design. Second, we develop a computationally efficient (15000 x) and accurate (correlation of 0.98) method to evaluate the performance of the network without standard recognition tests. Third, we show that the method produces a robustness metric that can be used to evaluate noise tolerance. Deep neural networks (DNN) abstract by demodulating the output of linear filters. In this article, we refine this definition of abstraction to show that the inputs of a DNN are abstracted with respect to the filters. Or, to restate, the abstraction is qualified by the filters. This leads us to introduce the notion of qualitative projection. We use qualitative projection to abstract MNIST hand-written digits with respect to the various dogs, horses, planes and cars of the CIFAR dataset. We then classify the MNIST digits according to the magnitude of their dogness, horseness, planeness and carness qualities, illustrating the generality of qualitative projection. The superiority of deeply learned pedestrian representations has been reported in very recent literature of person re-identification (re-ID). In this paper, we consider the more pragmatic issue of learning a deep feature with no or only a few labels. We propose a progressive unsupervised learning (PUL) method to transfer pretrained deep representations to unseen domains. Our method is easy to implement and can be viewed as an effective baseline for unsupervised re-ID feature learning. Specifically, PUL iterates between 1) pedestrian clustering and 2) fine-tuning of the convolutional neural network (CNN) to improve the original model trained on the irrelevant labeled dataset. Since the clustering results can be very noisy, we add a selection operation between the clustering and fine-tuning. At the beginning when the model is weak, CNN is fine-tuned on a small amount of reliable examples which locate near to cluster centroids in the feature space. As the model becomes stronger in subsequent iterations, more images are being adaptively selected as CNN training samples. Progressively, pedestrian clustering and the CNN model are improved simultaneously until algorithm convergence. This process is naturally formulated as self-paced learning. We then point out promising directions that may lead to further improvement. Extensive experiments on three large-scale re-ID datasets demonstrate that PUL outputs discriminative features that improve the re-ID accuracy. We prove that a particular deep network architecture is more efficient at approximating radially symmetric functions than the best known 2 or 3 layer networks. We use this architecture to approximate Gaussian kernel SVMs, and subsequently improve upon them with further training. The architecture and initial weights of the Deep Radial Kernel Network are completely specified by the SVM and therefore sidesteps the problem of empirically choosing an appropriate deep network architecture. We present a large N solution of a microscopic model describing the Mott-Anderson transition on a finite-coordination Bethe lattice. Our results demonstrate that strong spatial fluctuations, due to Anderson localization effects, dramatically modify the quantum critical behavior near disordered Mott transitions. The leading critical behavior of quasiparticle wavefunctions is shown to assume a universal form in the full range from weak to strong disorder, in contrast to disorder-driven non-Fermi liquid ("electronic Griffiths phase") behavior, which is found only in the strongly correlated regime. Automatically generated political event data is an important part of the social science data ecosystem. The approaches for generating this data, though, have remained largely the same for two decades. During this time, the field of computational linguistics has progressed tremendously. This paper presents an overview of political event data, including methods and ontologies, and a set of experiments to determine the applicability of deep neural networks to the extraction of political events from news text. Cut vertices, a generalization of matrix elements of local operators, are revisited, and an expansion in terms of minimally subtracted cut vertices is formulated. An extension of the formalism to deal with semi-inclusive deep inelastic processes in the target fragmentation region is explicitly constructed. The problem of factorization is discussed in detail. Reason and inference require process as well as memory skills by humans. Neural networks are able to process tasks like image recognition (better than humans) but in memory aspects are still limited (by attention mechanism, size). Recurrent Neural Network (RNN) and it's modified version LSTM are able to solve small memory contexts, but as context becomes larger than a threshold, it is difficult to use them. The Solution is to use large external memory. Still, it poses many challenges like, how to train neural networks for discrete memory representation, how to describe long term dependencies in sequential data etc. Most prominent neural architectures for such tasks are Memory networks: inference components combined with long term memory and Neural Turing Machines: neural networks using external memory resources. Also, additional techniques like attention mechanism, end to end gradient descent on discrete memory representation are needed to support these solutions. Preliminary results of above neural architectures on simple algorithms (sorting, copying) and Question Answering (based on story, dialogs) application are comparable with the state of the art. In this paper, I explain these architectures (in general), the additional techniques used and the results of their application. In this work, we present a novel 3D-Convolutional Neural Network (CNN) architecture called I2I-3D that predicts boundary location in volumetric data. Our fine-to-fine, deeply supervised framework addresses three critical issues to 3D boundary detection: (1) efficient, holistic, end-to-end volumetric label training and prediction (2) precise voxel-level prediction to capture fine scale structures prevalent in medical data and (3) directed multi-scale, multi-level feature learning. We evaluate our approach on a dataset consisting of 93 medical image volumes with a wide variety of anatomical regions and vascular structures. In the process, we also introduce HED-3D, a 3D extension of the state-of-the-art 2D edge detector (HED). We show that our deep learning approach out-performs, the current state-of-the-art in 3D vascular boundary detection (structured forests 3D), by a large margin, as well as HED applied to slices, and HED-3D while successfully localizing fine structures. With our approach, boundary detection takes about one minute on a typical 512x512x512 volume. Multi-task learning (MTL) involves the simultaneous training of two or more related tasks over shared representations. In this work, we apply MTL to audio-visual automatic speech recognition(AV-ASR). Our primary task is to learn a mapping between audio-visual fused features and frame labels obtained from acoustic GMM/HMM model. This is combined with an auxiliary task which maps visual features to frame labels obtained from a separate visual GMM/HMM model. The MTL model is tested at various levels of babble noise and the results are compared with a base-line hybrid DNN-HMM AV-ASR model. Our results indicate that MTL is especially useful at higher level of noise. Compared to base-line, upto 7\% relative improvement in WER is reported at -3 SNR dB To operate intelligently in domestic environments, robots require the ability to understand arbitrary spatial relations between objects and to generalize them to objects of varying sizes and shapes. In this work, we present a novel end-to-end approach utilizing neural networks to generalize spatial relations based on distance metric learning. Our network transforms spatial relations to a feature space that captures their similarities based on 3D point clouds of the objects and without prior semantic knowledge of the relations. It employs gradient-based optimization to compute object poses in order to imitate an arbitrary target relation by reducing the distance to it under the learned metric. We study a random graph model named the "block model" in statistics and the "planted partition model" in theoretical computer science. In its simplest form, this is a random graph with two equal-sized clusters, with a between-class edge probability of $q$ and a within-class edge probability of $p$. A striking conjecture of Decelle, Krzkala, Moore and Zdeborov\'a based on deep, non-rigorous ideas from statistical physics, gave a precise prediction for the algorithmic threshold of clustering in the sparse planted partition model. In particular, if $p = a/n$ and $q = b/n$, $s=(a-b)/2$ and $p=(a+b)/2$ then Decelle et al.\ conjectured that it is possible to efficiently cluster in a way correlated with the true partition if $s^2 > p$ and impossible if $s^2 < p$. By comparison, the best-known rigorous result is that of Coja-Oghlan, who showed that clustering is possible if $s^2 > C p \ln p$ for some sufficiently large $C$. In a previous work, we proved that indeed it is information theoretically impossible to to cluster if $s^2 < p$ and furthermore it is information theoretically impossible to even estimate the model parameters from the graph when $s^2 < p$. Here we complete the proof of the conjecture by providing an efficient algorithm for clustering in a way that is correlated with the true partition when $s^2 > p$. A different independent proof of the same result was recently obtained by Laurent Massoulie. We review the use of kinetically constrained models (KCMs) for the study of dynamics in glassy systems. The characteristic feature of KCMs is that they have trivial, often non-interacting, equilibrium behaviour but interesting slow dynamics due to restrictions on the allowed transitions between configurations. The basic question which KCMs ask is therefore how much glassy physics can be understood without an underlying ``equilibrium glass transition''. After a brief review of glassy phenomenology, we describe the main model classes, which include spin-facilitated (Ising) models, constrained lattice gases, models inspired by cellular structures such as soap froths, models obtained via mappings from interacting systems without constraints, and finally related models such as urn, oscillator, tiling and needle models. We then describe the broad range of techniques that have been applied to KCMs, including exact solutions, adiabatic approximations, projection and mode-coupling techniques, diagrammatic approaches and mappings to quantum systems or effective models. Finally, we give a survey of the known results for the dynamics of KCMs both in and out of equilibrium, including topics such as relaxation time divergences and dynamical transitions, nonlinear relaxation, aging and effective temperatures, cooperativity and dynamical heterogeneities, and finally non-equilibrium stationary states generated by external driving. We conclude with a discussion of open questions and possibilities for future work. Discovering the 'Neural Code' from multi-neuronal spike trains is an important task in neuroscience. For such an analysis, it is important to unearth interesting regularities in the spiking patterns. In this report, we present an efficient method for automatically discovering synchrony, synfire chains, and more general sequences of neuronal firings. We use the Frequent Episode Discovery framework of Laxman, Sastry, and Unnikrishnan (2005), in which the episodes are represented and recognized using finite-state automata. Many aspects of functional connectivity between neuronal populations can be inferred from the episodes. We demonstrate these using simulated multi-neuronal data from a Poisson model. We also present a method to assess the statistical significance of the discovered episodes. Since the Temporal Data Mining (TDM) methods used in this report can analyze data from hundreds and potentially thousands of neurons, we argue that this framework is appropriate for discovering the `Neural Code'. Random fields disorder Ising ferromagnets by aligning single spins in the direction of the random field in three space dimensions, or by flipping large ferromagnetic domains at dimensions two and below. While the former requires random fields of typical magnitude similar to the interaction strength, the latter Imry-Ma mechanism only requires infinitesimal random fields. Recently, it has been shown that for dilute anisotropic dipolar systems a third mechanism exists, where the ferromagnetic phase is disordered by finite-size glassy domains at a random field of finite magnitude that is considerably smaller than the typical interaction strength. Using large-scale Monte Carlo simulations and zero-temperature numerical approaches, we show that this mechanism applies to disordered ferromagnets with competing short-range ferromagnetic and antiferromagnetic interactions, suggesting its generality in ferromagnetic systems with competing interactions and an underlying spin-glass phase. A finite-size-scaling analysis of the magnetization distribution suggests that the transition might be first order. Purpose of our work is to obtain a basic understanding and comparison of the performance and structure of real Knowledge Networks, to identify strengths and weaknesses and to highlight guidelines for improvements. We selected 18 Knowledge Networks from the service sector and 12 networks from the production sector and estimated their Performance and Structure in terms of 19 indices from graph theory. Highlights from our work include: 1) As most networks are unilaterally structured, the direction of knowledge transfer should be taken into account as illustrated in the analysis of clubs and entropy, 2) The stability of most Knowledge Networks is questionable, 3) Few networks are effective in sharing information, while most Knowledge Networks cannot benefit from the network effect, have rather limited capability for coordination, information propagation and synchronization and are not able to integrate Tacit knowledge, 4) Few networks have large cliques which have to be managed with caution as their role may be highly constructive or destructive, 5) While agents with rich connections form clubs, as in most social networks, the poor club effect is not negligible when we take into account the link direction, 6) The directed link analysis of entropy reveals the low complexity-diversification of the Knowledge Networks. In fact the only high entropy network found, has been improved by Knowledge Management Professionals. As most Knowledge Networks underperform, there is plenty of room for further customized analysis in order to improve communication efficiency, coordination, Tacit knowledge dissemination and robustness. This is the first comparative study of real Knowledge Networks in terms of graph theoretic methods. This paper presents the first substantial study of the chemistry of the envelopes around a sample of 18 low-mass pre- and protostellar objects for which physical properties have previously been derived from radiative transfer modeling of their dust continuum emission. Single-dish line observations of 24 transitions of 9 molecular species (not counting isotopes) including HCO+, N2H+, CS, SO, SO2, HCN, HNC, HC3N and CN are reported. The line intensities are used to constrain the molecular abundances by comparison to Monte Carlo radiative transfer modeling of the line strengths. An empirical chemical network is constructed on the basis of correlations between the abundances of various species. For example, it is seen that the HCO+ and CO abundances are linearly correlated, both increasing with decreasing envelope mass. Species such as CS, SO and HCN show no trend with envelope mass. In particular no trend is seen between ``evolutionary stage'' of the objects and the abundances of the main sulfur- or nitrogen-containing species. Among the nitrogen-bearing species abundances of CN, HNC and HC3N are found to be closely correlated, which can be understood from considerations of the chemical network. The CS/SO abundance ratio is found to correlate with the abundances of CN and HC3N, which may reflect a dependence on the atomic carbon abundance. An anti-correlation is found between the deuteration of HCO+ and HCN, reflecting different temperature dependences for gas-phase deuteration mechanisms. The abundances are compared to other protostellar environments. In particular it is found that the abundances in the cold outer envelope of the previously studied class 0 protostar IRAS16293-2422 are in good agreement with the average abundances for the presented sample of class 0 objects. In some infectious processes, transmission occurs between specific ties between individuals, which ties constitute a contact network. To estimate the effect of an exposure on infectious outcomes within a collection of contact networks, the analysis must adjust for the correlation of outcomes within networks as well as the probability of exposure. This estimation process may be more statistically efficient when leveraging baseline covariates related to both the exposure and infectious outcome. We investigate the extent to which gains in statistical efficiency depend on contact network structure and properties of the infectious process. To do this, we simulate a stochastic compartmental infection on a collection of contact networks, and employ the observational augmented GEE using a variety of contact network and baseline infection summaries as adjustment covariates. We apply this approach to estimate the effect of leadership and a concurrent self-help program in the spread of a novel microfinance program in a collection of villages in Karnataka, India. The prediction of periodical time-series remains challenging due to various types of data distortions and misalignments. Here, we propose a novel model called Temporal embedding-enhanced convolutional neural Network (TeNet) to learn repeatedly-occurring-yet-hidden structural elements in periodical time-series, called abstract snippets, for predicting future changes. Our model uses convolutional neural networks and embeds a time-series with its potential neighbors in the temporal domain for aligning it to the dominant patterns in the dataset. The model is robust to distortions and misalignments in the temporal domain and demonstrates strong prediction power for periodical time-series. We conduct extensive experiments and discover that the proposed model shows significant and consistent advantages over existing methods on a variety of data modalities ranging from human mobility to household power consumption records. Empirical results indicate that the model is robust to various factors such as number of samples, variance of data, numerical ranges of data etc. The experiments also verify that the intuition behind the model can be generalized to multiple data types and applications and promises significant improvement in prediction performances across the datasets studied. Human pose estimation (i.e., locating the body parts / joints of a person) is a fundamental problem in human-computer interaction and multimedia applications. Significant progress has been made based on the development of depth sensors, i.e., accessible human pose prediction from still depth images [32]. However, most of the existing approaches to this problem involve several components/models that are independently designed and optimized, leading to suboptimal performances. In this paper, we propose a novel inference-embedded multi-task learning framework for predicting human pose from still depth images, which is implemented with a deep architecture of neural networks. Specifically, we handle two cascaded tasks: i) generating the heat (confidence) maps of body parts via a fully convolutional network (FCN); ii) seeking the optimal configuration of body parts based on the detected body part proposals via an inference built-in MatchNet [10], which measures the appearance and geometric kinematic compatibility of body parts and embodies the dynamic programming inference as an extra network layer. These two tasks are jointly optimized. Our extensive experiments show that the proposed deep model significantly improves the accuracy of human pose estimation over other several state-of-the-art methods or SDKs. We also release a large-scale dataset for comparison, which includes 100K depth images under challenging scenarios. In this work we deal with the optimal design and optimal control of structures undergoing large rotations. In other words, we show how to find the corresponding initial configuration and the corresponding set of multiple load parameters in order to recover a desired deformed configuration or some desirable features of the deformed configuration as specified more precisely by the objective or cost function. The model problem chosen to illustrate the proposed optimal design and optimal control methodologies is the one of geometrically exact beam. First, we present a non-standard formulation of the optimal design and optimal control problems, relying on the method of Lagrange multipliers in order to make the mechanics state variables independent from either design or control variables and thus provide the most general basis for developing the best possible solution procedure. Two different solution procedures are then explored, one based on the diffuse approximation of response function and gradient method and the other one based on genetic algorithm. A number of numerical examples are given in order to illustrate both the advantages and potential drawbacks of each of the presented procedures. We obtain an analytic expression for the full distribution of conductance for a strongly disordered three dimensional conductor within a perturbative approach based on the transfer-matrix formulation. Our results confirm numerical evidence that the log-normal limit of the distribution is not reached even in the deeply insulating regime. We show that the variance of the logarithm of the conductance scales as a fractional power of the mean, while the skewness changes sign as one approaches the Anderson metal-insulator transition from the deeply insulating limit, all described as a function of a single parameter. The approach suggests a possible single parameter description of the Anderson transition that takes into account the full nontrivial distribution of conductance. Detecting small objects is notoriously challenging due to their low resolution and noisy representation. Existing object detection pipelines usually detect small objects through learning representations of all the objects at multiple scales. However, the performance gain of such ad hoc architectures is usually limited to pay off the computational cost. In this work, we address the small object detection problem by developing a single architecture that internally lifts representations of small objects to "super-resolved" ones, achieving similar characteristics as large objects and thus more discriminative for detection. For this purpose, we propose a new Perceptual Generative Adversarial Network (Perceptual GAN) model that improves small object detection through narrowing representation difference of small objects from the large ones. Specifically, its generator learns to transfer perceived poor representations of the small objects to super-resolved ones that are similar enough to real large objects to fool a competing discriminator. Meanwhile its discriminator competes with the generator to identify the generated representation and imposes an additional perceptual requirement - generated representations of small objects must be beneficial for detection purpose - on the generator. Extensive evaluations on the challenging Tsinghua-Tencent 100K and the Caltech benchmark well demonstrate the superiority of Perceptual GAN in detecting small objects, including traffic signs and pedestrians, over well-established state-of-the-arts. A complex network approach is proposed to study the shear behavior of a rough rock joint. Similarities between aperture profiles are established and a general network in two directions (in parallel and perpendicular to the shear direction) is constructed. Evaluation of this newly formed network shows that the degree distribution of the network, after a transition stage falls into a quasi stable state which is roughly obeying a Gaussian distribution. In addition, the growth of the clustering coefficient and the number of edges are approximately scaled with the development of shear strength and hydraulic conductivity, which can be utilized to estimate, shear distribution over asperities. Furthermore, we characterize the contact profiles using the same approach. Despite the former case, the later networks are following a growing network mode. The multi Vehicle Routing Problem with Pickup and Delivery with Time Windows is a challenging version of the Vehicle Routing Problem. In this paper, by embedding many complex assignment routing constraints through constructing a multi dimensional network, we intend to reach optimality for local clusters derived from a reasonably large set of passengers on real world transportation networks. More specifically, we introduce a multi vehicle state space time network representation in which only the non dominated assignment based hyper paths are examined. In addition, by the aid of passengers cumulative service patterns defined in this paper, our solution approach is able to take control of symmetry issue, a common issue in the combinatorial problems. At the end, extensive computational results over the instances proposed by Ropke and Cordeau 2009 and a randomly generated data sets from the Phoenix subarea, City of Tempe, show the computational efficiency and solution optimality of our developed algorithm. Modeling interpersonal influence on different sentimental polarities is a fundamental problem in opinion formation and viral marketing. There has not been seen an effective solution for learning sentimental influences from users' behaviors yet. Previous related works on information propagation directly define interpersonal influence between each pair of users as a parameter, which is independent from each others, even if the influences come from or affect the same user. And influences are learned from user's propagation behaviors, namely temporal cascades, while sentiments are not associated with them. Thus we propose to model the interpersonal influence by latent influence and susceptibility matrices defined on individual users and sentiment polarities. Such low-dimensional and distributed representations naturally make the interpersonal influences related to the same user coupled with each other, and in turn, reduce the model complexity. Sentiments act on different rows of parameter matrices, depicting their effects in modeling cascades. With the iterative optimization algorithm of projected stochastic gradient descent over shuffled mini-batches and Adadelta update rule, negative cases are repeatedly sampled with the distribution of infection frequencies users, for reducing computation cost and optimization imbalance. Experiments are conducted on Microblog dataset. The results show that our model achieves better performance than the state-of-the-art and pair-wise models. Besides, analyzing the distribution of learned users' sentimental influences and susceptibilities results some interesting discoveries. Associative memories are data structures that allow retrieval of stored messages from part of their content. They thus behave similarly to human brain that is capable for instance of retrieving the end of a song given its beginning. Among different families of associative memories, sparse ones are known to provide the best efficiency (ratio of the number of bits stored to that of bits used). Nevertheless, it is well known that non-uniformity of the stored messages can lead to dramatic decrease in performance. We introduce several strategies to allow efficient storage of non-uniform messages in recently introduced sparse associative memories. We analyse and discuss the methods introduced. We also present a practical application example. In this work we explore a straightforward variational Bayes scheme for Recurrent Neural Networks. Firstly, we show that a simple adaptation of truncated backpropagation through time can yield good quality uncertainty estimates and superior regularisation at only a small extra computational cost during training. Secondly, we demonstrate how a novel kind of posterior approximation yields further improvements to the performance of Bayesian RNNs. We incorporate local gradient information into the approximate posterior to sharpen it around the current batch statistics. This technique is not exclusive to recurrent neural networks and can be applied more widely to train Bayesian neural networks. We also empirically demonstrate how Bayesian RNNs are superior to traditional RNNs on a language modelling benchmark and an image captioning task, as well as showing how each of these methods improve our model over a variety of other schemes for training them. We also introduce a new benchmark for studying uncertainty for language models so future methods can be easily compared. We analyze a distributed information network in which each node has access to the information contained in a limited set of nodes (its neighborhood) at a given time. A collective computation is carried out in which each node calculates a value that implies all information contained in the network (in our case, the average value of a variable that can take different values in each network node). The neighborhoods can change dynamically by exchanging neighbors with other nodes. The results of this collective calculation show rapid convergence and good scalability with the network size. These results are compared with those of a fixed network arranged as a square lattice, in which the number of rounds to achieve a given accuracy is very high when the size of the network increases. The results for the evolving networks are interpreted in light of the properties of complex networks and are directly relevant to the diameter and characteristic path length of the networks, which seem to express "small world" properties. While gradient descent has proven highly successful in learning connection weights for neural networks, the actual structure of these networks is usually determined by hand, or by other optimization algorithms. Here we describe a simple method to make network structure differentiable, and therefore accessible to gradient descent. We test this method on recurrent neural networks applied to simple sequence prediction problems. Starting with initial networks containing only one node, the method automatically builds networks that successfully solve the tasks. The number of nodes in the final network correlates with task difficulty. The method can dynamically increase network size in response to an abrupt complexification in the task; however, reduction in network size in response to task simplification is not evident for reasonable meta-parameters. The method does not penalize network performance for these test tasks: variable-size networks actually reach better performance than fixed-size networks of higher, lower or identical size. We conclude by discussing how this method could be applied to more complex networks, such as feedforward layered networks, or multiple-area networks of arbitrary shape. The understanding of molecular cell biology requires insight into the structure and dynamics of networks that are made up of thousands of interacting molecules of DNA, RNA, proteins, metabolites, and other components. One of the central goals of systems biology is the unraveling of the as yet poorly characterized complex web of interactions among these components. This work is made harder by the fact that new species and interactions are continuously discovered in experimental work, necessitating the development of adaptive and fast algorithms for network construction and updating. Thus, the "reverse-engineering" of networks from data has emerged as one of the central concern of systems biology research. A variety of reverse-engineering methods have been developed, based on tools from statistics, machine learning, and other mathematical domains. In order to effectively use these methods, it is essential to develop an understanding of the fundamental characteristics of these algorithms. With that in mind, this chapter is dedicated to the reverse-engineering of biological systems. Specifically, we focus our attention on a particular class of methods for reverse-engineering, namely those that rely algorithmically upon the so-called "hitting-set" problem, which is a classical combinatorial and computer science problem, Each of these methods utilizes a different algorithm in order to obtain an exact or an approximate solution of the hitting set problem. We will explore the ultimate impact that the alternative algorithms have on the inference of published in silico biological networks. Entrainment by a pacemaker, representing an element with a higher frequency, is numerically investigated for several classes of random networks which consist of identical phase oscillators. We find that the entrainment frequency window of a network decreases exponentially with its depth, defined as the mean forward distance of the elements from the pacemaker. Effectively, only shallow networks can thus exhibit frequency-locking to the pacemaker. The exponential dependence is also derived analytically as an approximation for large random asymmetric networks. To understand the sample-to-sample fluctuations in disorder-generated multifractal patterns we investigate analytically as well as numerically the statistics of high values of the simplest model - the ideal periodic $1/f$ Gaussian noise. By employing the thermodynamic formalism we predict the characteristic scale and the precise scaling form of the distribution of number of points above a given level. We demonstrate that the powerlaw forward tail of the probability density, with exponent controlled by the level, results in an important difference between the mean and the typical values of the counting function. This can be further used to determine the typical threshold $x_m$ of extreme values in the pattern which turns out to be given by $x_m^{(typ)}=2-c\ln{\ln{M}}/\ln{M}$ with $c=3/2$. Such observation provides a rather compelling explanation of the mechanism behind universality of $c$. Revealed mechanisms are conjectured to retain their qualitative validity for a broad class of disorder-generated multifractal fields. In particular, we predict that the typical value of the maximum $p_{max}$ of intensity is to be given by $-\ln{p_{max}} = \alpha_{-}\ln{M} + \frac{3}{2f'(\alpha_{-})}\ln{\ln{M}} + O(1)$, where $f(\alpha)$ is the corresponding singularity spectrum vanishing at $\alpha=\alpha_{-}>0$. For the $1/f$ noise we also derive exact as well as well-controlled approximate formulas for the mean and the variance of the counting function without recourse to the thermodynamic formalism. We present an analysis of the impact of structural disorder on the static scattering function of f-armed star branched polymers in d dimensions. To this end, we consider the model of a star polymer immersed in a good solvent in the presence of structural defects, correlated at large distances r according to a power law \sim r^{-a}. In particular, we are interested in the ratio g(f) of the radii of gyration of star and linear polymers of the same molecular weight, which is a universal experimentally measurable quantity. We apply a direct polymer renormalization approach and evaluate the results within the double \varepsilon=4-d, \delta=4-a-expansion. We find an increase of g(f) with an increasing \delta. Therefore, an increase of disorder correlations leads to an increase of the size measure of a star relative to linear polymers of the same molecular weight. How to determine the community structure of complex networks is an open question. It is critical to establish the best strategies for community detection in networks of unknown structure. Here, using standard synthetic benchmarks, we show that none of the algorithms hitherto developed for community structure characterization perform optimally. Significantly, evaluating the results according to their modularity, the most popular measure of the quality of a partition, systematically provides mistaken solutions. However, a novel quality function, called Surprise, can be used to elucidate which is the optimal division into communities. Consequently, we show that the best strategy to find the community structure of all the networks examined involves choosing among the solutions provided by multiple algorithms the one with the highest Surprise value. We conclude that Surprise maximization precisely reveals the community structure of complex networks. We propose sensorimotor tappings, a new graphical technique that explicitly represents relations between the time steps of an agent's sensorimotor loop and a single training step of an adaptive model that the agent is using internally. In the simplest case this is a relation linking two time steps. In realistic cases these relations can extend over several time steps and over different sensory channels. The aim is to capture the footprint of information intake relative to the agent's current time step. We argue that this view allows us to make prior considerations explicit and then use them in implementations without modification once they are established. In the paper we introduce the problem domain, explain the basic idea, provide example tappings for standard configurations used in developmental models, and show how tappings can be applied to problems in related fields. We formulate two versions of the power control problem for wireless networks with latency constraints arising from duty cycle allocations In the first version, strategic power optimization, wireless nodes are modeled as rational agents in a power game, who strategically adjust their powers to minimize their own energy. In the other version, joint power optimization, wireless nodes jointly minimize the aggregate energy expenditure. Our analysis of these models yields insights into the different energy outcomes of strategic versus joint power optimization. We derive analytical solutions for power allocation under both models and study how they are affected by data loads and channel quality. We derive simple necessary conditions for the existence of Nash equilibria in the power game and also provide numerical examples of optimal power allocation under both models. Finally, we show that joint optimization can (sometimes) be Pareto-optimal and dominate strategic optimization, i.e the energy expenditure of all nodes is lower than if they were using strategic optimization. We describe DyNet, a toolkit for implementing neural network models based on dynamic declaration of network structure. In the static declaration strategy that is used in toolkits like Theano, CNTK, and TensorFlow, the user first defines a computation graph (a symbolic representation of the computation), and then examples are fed into an engine that executes this computation and computes its derivatives. In DyNet's dynamic declaration strategy, computation graph construction is mostly transparent, being implicitly constructed by executing procedural code that computes the network outputs, and the user is free to use different network structures for each input. Dynamic declaration thus facilitates the implementation of more complicated network architectures, and DyNet is specifically designed to allow users to implement their models in a way that is idiomatic in their preferred programming language (C++ or Python). One challenge with dynamic declaration is that because the symbolic computation graph is defined anew for every training example, its construction must have low overhead. To achieve this, DyNet has an optimized C++ backend and lightweight graph representation. Experiments show that DyNet's speeds are faster than or comparable with static declaration toolkits, and significantly faster than Chainer, another dynamic declaration toolkit. DyNet is released open-source under the Apache 2.0 license and available at http://github.com/clab/dynet. We propose an entirely data-driven approach to estimating the 3D pose of a hand given a depth image. We show that we can correct the mistakes made by a Convolutional Neural Network trained to predict an estimate of the 3D pose by using a feedback loop. The components of this feedback loop are also Deep Networks, optimized using training data. They remove the need for fitting a 3D model to the input data, which requires both a carefully designed fitting function and algorithm. We show that our approach outperforms state-of-the-art methods, and is efficient as our implementation runs at over 400 fps on a single GPU. We study the local density of states around potential scatterers in d-wave superconductors, and show that quantum interference between impurity states is not negligible for experimentally relevant impurity concentrations. The two impurity model is used as a paradigm to understand these effects analytically and in interpreting numerical solutions of the Bogoliubov-de Gennes equations on fully disordered systems. We focus primarily on the globally particle-hole symmetric model which has been the subject of considerable controversy, and give evidence that a zero-energy delta function exists in the DOS. The anomalous spectral weight at zero energy is seen to arise from resonant impurity states belonging to a particular sublattice, exactly as in the 2-impurity version of this model. We discuss the implications of these findings for realistic models of the cuprates. Despite recent breakthroughs in the applications of deep neural networks, one setting that presents a persistent challenge is that of "one-shot learning." Traditional gradient-based networks require a lot of data to learn, often through extensive iterative training. When new data is encountered, the models must inefficiently relearn their parameters to adequately incorporate the new information without catastrophic interference. Architectures with augmented memory capacities, such as Neural Turing Machines (NTMs), offer the ability to quickly encode and retrieve new information, and hence can potentially obviate the downsides of conventional models. Here, we demonstrate the ability of a memory-augmented neural network to rapidly assimilate new data, and leverage this data to make accurate predictions after only a few samples. We also introduce a new method for accessing an external memory that focuses on memory content, unlike previous methods that additionally use memory location-based focusing mechanisms. Using our previous results for the configurational entropy of a stripe glass as well as a variational result for the bare surface tension of entropic droplets we show that there is no disagreement between the numerical simulations of Grousson et al. and our theory. The claim that our theory disagrees with numerical simulations is based on the assumption that the surface tension is independent of the frustration parameter Q of the model. However, we show in this Reply that it varies strongly with Q and that the resulting Q-dependence of the kinetic fragility agrees with the one obtained by Grousson et al. We believe that this answers the questions raised in the Comment by Grousson et al. This paper deals with the use of self-organizing protocols to improve the reliability of dynamic Peer-to-Peer (P2P) overlay networks. We present two approaches, that employ local knowledge of the 2nd neighborhood of nodes. The first scheme is a simple protocol requiring interactions among nodes and their direct neighbors. The second scheme extends this approach by resorting to the Edge Clustering Coefficient (ECC), a local measure that allows to identify those edges that connect different clusters in an overlay. A simulation assessment is presented, which evaluates these protocols over uniform networks, clustered networks and scale-free networks. Different failure modes are considered. Results demonstrate the viability of the proposal. The one-dimensional Ising spin-glass model with power-law long-range interactions is a useful proxy model for studying spin glasses in higher space dimensions and for finding the dimension at which the spin-glass state changes from having broken replica symmetry to that of droplet behavior. To this end we have calculated the exponent that describes the difference in free energy between periodic and antiperiodic boundary conditions. Numerical work is done to support some of the assumptions made in the calculations and to determine the behavior of the interface free-energy exponent of the power law of the interactions. Our numerical results for the interface free-energy exponent are badly affected by finite-size problems. Cognitive (Radio) (CR) Communications (CC) are mainly deployed within the environments of primary (user) communications, where the channel states and accessibility are usually stochastically distributed (benign or IID). However, many practical CC are also exposed to disturbing events (contaminated) and vulnerable jamming attacks (adversarial or non-IID). Thus, the channel state distribution of spectrum could be stochastic, contaminated or adversarial at different temporal and spatial locations. Without any a priori, facilitating optimal CC is a very challenging issue. In this paper, we propose an online learning algorithm that performs the joint channel sensing, probing and adaptive channel access for multi-channel CC in general unknown environments. We take energy-efficient CC (EECC) into our special attention, which is highly desirable for green wireless communications and demanding to combat with potential jamming attack who could greatly mar the energy and spectrum efficiency of CC. The EECC is formulated as a constrained regret minimization problem with power budget constraints. By tuning a novel exploration parameter, our algorithms could adaptively find the optimal channel access strategies and achieve the almost optimal learning performance of EECC in different scenarios provided with the vanishing long-term power budget violations. We also consider the important scenario that cooperative learning and information sharing among multiple CR users to see further performance improvements. The proposed algorithms are resilient to both oblivious and adaptive jamming attacks with different intelligence and attacking strength. Extensive numerical results are conducted to validate our theory. We argue that the critical dynamical fluctuations predicted by the mode-coupling theory (MCT) of glasses provide a natural mechanism to explain the breakdown of the Stokes-Einstein relation. This breakdown, observed numerically and experimentally in a region where MCT should hold, is one of the major difficulty of the theory, for which we propose a natural resolution based on the recent interpretation of the MCT transition as a bona fide critical point with a diverging length scale. We also show that the upper critical dimension of MCT is d_c=8. We propose to model the dynamics of metabolic networks from a systems biology point of view by four dynamical structure elements: potential function, transverse matrix, degradation matrix, and stochastic force. These four elements are balanced to determine the network dynamics, which gives arise to a special stochastic differential equation supplemented by a relationship between the stochastic force and the degradation matrix. Important network behaviors can be obtained from the potential function without explicitly solving for the time-dependent solution. The existence of such a potential function suggests a global optimization principle, and the existence stochastic force corresponds natural to the hierarchical structure in metabolic networks. We provide theoretical evidences to justify our proposal by discussing its connections to others large-scale biochemical systems approaches, such as the network thermodynamics theory, biochemical systems theory, metabolic control analysis, and flux balance analysis. Experimental data displaying stochasticity are also pointed out. We study a highly supercooled two-dimensional fluid mixture via molecular dynamics simulation. We follow bond breakage events among particle pairs, which occur on the scale of the $\alpha$ relaxation time $\tau_{\alpha}$. Large scale heterogeneities analogous to the critical fluctuations in Ising systems are found in the spatial distribution of bonds which are broken in a time interval with a width of order $0.05\tau_{\alpha}$. The structure factor of the broken bond density is well approximated by the Ornstein-Zernike form. The correlation length is of order $100 \sigma_1$ at the lowest temperature studied, $\sigma_1$ being the particle size. The weakly bonded regions thus identified evolve in time with strong spatial correlations. The ICDM Challenge 2013 is to apply machine learning to the problem of hotel ranking, aiming to maximize purchases according to given hotel characteristics, location attractiveness of hotels, user's aggregated purchase history and competitive online travel agency information for each potential hotel choice. This paper describes the solution of team "binghsu & MLRush & BrickMover". We conduct simple feature engineering work and train different models by each individual team member. Afterwards, we use listwise ensemble method to combine each model's output. Besides describing effective model and features, we will discuss about the lessons we learned while using deep learning in this competition. To study the effect of quenched disorder in a class of reaction-diffusion systems, we introduce a conserved mass model of diffusion and aggregation in which the mass moves as a whole to a nearest neighbour on most sites while it fragments off as a single monomer (i.e. chips off) from certain fixed sites. Once the mass leaves any site, it coalesces with the mass present on its neighbour. We study in detail the effect of a \emph{single} chipping site on the steady state in arbitrary dimensions, with and without bias. In the thermodynamic limit, the system can exist in one of the following phases -- (a) Pinned Aggregate (PA) phase in which an infinite aggregate (with mass proportional to the volume of the system) appears with probability one at the chipping site but not in the bulk. (b) Unpinned Aggregate (UA) phase in which $\emph{both}$ the chipping site and the bulk can support an infinite aggregate simultaneously. (c) Non Aggregate (NA) phase in which there is no infinite cluster. Our analytical and numerical studies show that the system exists in the UA phase in all cases except in 1d with bias. In the latter case, there is a phase transition from the NA phase to the PA phase as density is increased. A variant of the above aggregation model is also considered in which total particle number is conserved and chipping occurs at a fixed site, but the particles do not interact with each other at other sites. This model is solved exactly by mapping it to a Zero Range Process. With increasing density, it exhibits a phase transition from the NA phase to the PA phase in all dimensions, irrespective of bias. Finally, we discuss the likely behaviour of the system in the presence of extensive disorder. In this paper we introduce a method to overcome one of the main challenges of person re-identification in multi-camera networks, namely cross-view appearance changes. The proposed solution addresses the extreme variability of person appearance in different camera views by exploiting multiple feature representations. For each feature, Kernel Canonical Correlation Analysis (KCCA) with different kernels is exploited to learn several projection spaces in which the appearance correlation between samples of the same person observed from different cameras is maximized. An iterative logistic regression is finally used to select and weigh the contributions of each feature projections and perform the matching between the two views. Experimental evaluation shows that the proposed solution obtains comparable performance on VIPeR and PRID 450s datasets and improves on PRID and CUHK01 datasets with respect to the state of the art. A quantum network promises to enable long distance quantum communication, and assemble small quantum devices into a large quantum computing cluster. Each network node can thereby be seen as a small few qubit quantum computer. Qubits can be sent over direct physical links connecting nearby quantum nodes, or by means of teleportation over pre-established entanglement amongst distant network nodes. Such pre-shared entanglement effectively forms a shortcut - a virtual quantum link - which can be used exactly once. Here, we present an abstraction of a quantum network that allows ideas from computer science to be applied to the problem of routing qubits, and manage entanglement in the network. Specifically, we consider a scenario in which each quantum network node can create EPR pairs with its immediate neighbours over a physical connection, and perform entanglement swapping operations in order to create long distance virtual quantum links. We proceed to discuss the features unique to quantum networks, which call for the development of new routing techniques. As an example, we present two simple hierarchical routing schemes for a quantum network of N nodes for a ring and sphere topology. For these topologies we present efficient routing algorithms requiring O(log N) qubits to be stored at each network node, O(polylog N) time and space to perform routing decisions, and O(log N) timesteps to replenish the virtual quantum links in a model of entanglement generation. We propose a new approach to the problem of neural network expressivity, which seeks to characterize how structural properties of a neural network family affect the functions it is able to compute. Our approach is based on an interrelated set of measures of expressivity, unified by the novel notion of trajectory length, which measures how the output of a network changes as the input sweeps along a one-dimensional path. Our findings can be summarized as follows: (1) The complexity of the computed function grows exponentially with depth. (2) All weights are not equal: trained networks are more sensitive to their lower (initial) layer weights. (3) Regularizing on trajectory length (trajectory regularization) is a simpler alternative to batch normalization, with the same performance. In this work we present a new methodology to study the structure of the configuration spaces of hard combinatorial problems. It consists in building the network that has as nodes the locally optimal configurations and as edges the weighted oriented transitions between their basins of attraction. We apply the approach to the detection of communities in the optima networks produced by two different classes of instances of a hard combinatorial optimization problem: the quadratic assignment problem (QAP). We provide evidence indicating that the two problem instance classes give rise to very different configuration spaces. For the so-called real-like class, the networks possess a clear modular structure, while the optima networks belonging to the class of random uniform instances are less well partitionable into clusters. This is convincingly supported by using several statistical tests. Finally, we shortly discuss the consequences of the findings for heuristically searching the corresponding problem spaces. Recent results are reviewed on both the time evolution and retrieval properties of multi-state neural networks that are based upon spin-glass models. In particular, the properties of models with neuron states having Q-Ising symmetry are discussed for various architectures. The main common features and differences are highlighted. We propose a novel trust metric for social networks which is suitable for application in recommender systems. It is personalised and dynamic and allows to compute the indirect trust between two agents which are not neighbours based on the direct trust between agents that are neighbours. In analogy to some personalised versions of PageRank, this metric makes use of the concept of feedback centrality and overcomes some of the limitations of other trust metrics.In particular, it does not neglect cycles and other patterns characterising social networks, as some other algorithms do. In order to apply the metric to recommender systems, we propose a way to make trust dynamic over time. We show by means of analytical approximations and computer simulations that the metric has the desired properties. Finally, we carry out an empirical validation on a dataset crawled from an Internet community and compare the performance of a recommender system using our metric to one using collaborative filtering. We argue that the Scanning Tunneling Microscope (STM) images of resonant states generated by doping Zn or Ni impurities into Cu-O planes of BSCCO are the result of quantum interference of the impurity signal coming from several distinct paths. The impurity image seen on the surface is greatly affected by interlayer tunneling matrix elements. We find that the optimal tunneling path between the STM tip and the metal (Cu, Zn, or Ni) $d_{x^2 - y^2}$ orbitals in the Cu-O plane involves intermediate excited states. This tunneling path leads to the four-fold nonlocal filter of the impurity state in Cu-O plane that explains the experimental impurity spectra. Applications of the tunneling filter to the Cu vacancy defects and ``direct'' tunneling into Cu-O planes are also discussed. Vision impairment due to pathological damage of the retina can largely be prevented through periodic screening using fundus color imaging. However the challenge with large scale screening is the inability to exhaustively detect fine blood vessels crucial to disease diagnosis. In this work we present a computational imaging framework using deep and ensemble learning for reliable detection of blood vessels in fundus color images. An ensemble of deep convolutional neural networks is trained to segment vessel and non-vessel areas of a color fundus image. During inference, the responses of the individual ConvNets of the ensemble are averaged to form the final segmentation. In experimental evaluation with the DRIVE database, we achieve the objective of vessel detection with maximum average accuracy of 94.7\% and area under ROC curve of 0.9283. The one-dimensional (1D) tight binding model with random nearest neighbor hopping is known to have a singularity of the density of states and of the localization length at the band center. We study numerically the effects of random long range (power-law) hopping with an ensemble averaged magnitude $\expectation{|t_{ij}|} \propto |i-j|^{-\sigma}$ in the 1D chain, while maintaining the particle-hole symmetry present in the nearest neighbor model. We find, in agreement with results of position space renormalization group techniques applied to the random XY spin chain with power-law interactions, that there is a change of behavior when the power-law exponent $\sigma$ becomes smaller than 2. The depinning of an elastic line interacting with a quenched disorder is studied for long range interactions, applicable to crack propagation or wetting. An ultrametric distance is introduced instead of the Euclidean distance, allowing for a drastic reduction of the numerical complexity of the problem. Based on large scale simulations, two to three orders of magnitude larger than previously considered, we obtain a very precise determination of critical exponents which are shown to be indistinguishable from their Euclidean metric counterparts. Moreover the scaling functions are shown to be unchanged. The choice of an ultrametric distance thus does not affect the universality class of the depinning transition and opens the way to an analytic real space renormalization group approach. Effective Communication for marketing is a vital field in business organizations, which is used to convey the details about their products and services to the market segments and subsequently to build long lasting customer relationships. This paper focuses on an emerging component of the integrated marketing communication, ie. social media networking, as it is increasingly becoming the trend. In 21st century, the marketing communication platforms show a tendency to shift towards innovative technology bound people networking which is becoming an acceptable domain of interaction. Though the traditional channels like TV, print media etc. are still active and prominent in marketing communication, the presences of the Internet and more specifically the Social Media Networking, has started influencing the way individuals and business enterprises communicate. It has become evident that more individuals and business enterprises are engaging the social media networking sites either to accelerate the sales of their products and services or to provide post-purchase feedbacks. This shift in scenario has motivated this research which took six months (June 2011 - December 2011), using empirical analysis which is carried out based on several primary and secondary evidences. The research paper also analyzes the factors that govern the social media networking sites to influence consumers and subsequently enable their purchase decisions. The secondary data presented for this research were those pertaining to the period between the year 2005 and year 2011. The study revealed promising facts like the transition to marketing through SMN gives visible advantages like bidirectional communication, interactive product presentation, and a firm influence on customer who has a rudimentary interest... Curriculum Learning emphasizes the order of training instances in a computational learning setup. The core hypothesis is that simpler instances should be learned early as building blocks to learn more complex ones. Despite its usefulness, it is still unknown how exactly the internal representation of models are affected by curriculum learning. In this paper, we study the effect of curriculum learning on Long Short-Term Memory (LSTM) networks, which have shown strong competency in many Natural Language Processing (NLP) problems. Our experiments on sentiment analysis task and a synthetic task similar to sequence prediction tasks in NLP show that curriculum learning has a positive effect on the LSTM's internal states by biasing the model towards building constructive representations i.e. the internal representation at the previous timesteps are used as building blocks for the final prediction. We also find that smaller models significantly improves when they are trained with curriculum learning. Lastly, we show that curriculum learning helps more when the amount of training data is limited. Deep neural networks have shown effectiveness in many challenging tasks and proved their strong capability in automatically learning good feature representation from raw input. Nonetheless, designing their architectures still requires much human effort. Techniques for automatically designing neural network architectures such as reinforcement learning based approaches recently show promising results in benchmarks. However, these methods still train each network from scratch during exploring the architecture space, which results in extremely high computational cost. In this paper, we propose a novel reinforcement learning framework for automatic architecture designing, where the action is to grow the network depth or layer width based on the current network architecture with function preserved. As such, the previously validated networks can be reused for further exploration, thus saves a large amount of computational cost. The experiments on image benchmark datasets have demonstrated the efficiency and effectiveness of our proposed solution compared to existing automatic architecture designing methods. The Earth Mover's Distance (EMD) computes the optimal cost of transforming one distribution into another, given a known transport metric between them. In deep learning, the EMD loss allows us to embed information during training about the output space structure like hierarchical or semantic relations. This helps in achieving better output smoothness and generalization. However EMD is computationally expensive.Moreover, solving EMD optimization problems usually require complex techniques like lasso. These properties limit the applicability of EMD-based approaches in large scale machine learning. We address in this work the difficulties facing incorporation of EMD-based loss in deep learning frameworks. Additionally, we provide insight and novel solutions on how to integrate such loss function in training deep neural networks. Specifically, we make three main contributions: (i) we provide an in-depth analysis of the fastest state-of-the-art EMD algorithm (Sinkhorn Distance) and discuss its limitations in deep learning scenarios. (ii) we derive fast and numerically stable closed-form solutions for the EMD gradient in output spaces with chain- and tree- connectivity; and (iii) we propose a relaxed form of the EMD gradient with equivalent computational complexity but faster convergence rate. We support our claims with experiments on real datasets. In a restricted data setting on the ImageNet dataset, we train a model to classify 1000 categories using 50K images, and demonstrate that our relaxed EMD loss achieves better Top-1 accuracy than the cross entropy loss. Overall, we show that our relaxed EMD loss criterion is a powerful asset for deep learning in the small data regime. We explore the nature of the faint blue objects in the Hubble Deep Field South. We have derived proper motions for the point sources in the Hubble Deep Field South using a 3 year baseline. Combining our proper motion measurements with spectral energy distribution fitting enabled us to identify 4 quasars and 42 stars, including 3 white dwarf candidates. Two of these white dwarf candidates, HDFS 1444 and 895, are found to display significant proper motion, 21.1 $\pm$ 7.9 mas/yr and 34.9 $\pm$ 8.0 mas/yr, and are consistent with being thick disk or halo white dwarfs located at ~2 kpc. The other faint blue objects analyzed by Mendez & Minniti do not show any significant proper motion and are inconsistent with being halo white dwarfs; they do not contribute to the Galactic dark matter. The observed population of stars and white dwarfs is consistent with standard Galactic models. Biological structure and function depend on complex regulatory interactions between many genes. A wealth of gene expression data is available from high-throughput genome-wide measurement technologies, but effective gene regulatory network inference methods are still needed. Model-based methods founded on quantitative descriptions of gene regulation are among the most promising, but many such methods rely on simple, local models or on ad hoc inference approaches lacking experimental interpretability. We propose an experimental design and develop an associated statistical method for inferring a gene network by learning a standard quantitative, interpretable, predictive, biophysics-based ordinary differential equation model of gene regulation. We fit the model parameters using gene expression measurements from perturbed steady-states of the system, like those following overexpression or knockdown experiments. Although the original model is nonlinear, our design allows us to transform it into a convex optimization problem by restricting attention to steady-states and using the lasso for parameter selection. Here, we describe the model and inference algorithm and apply them to a synthetic six-gene system, demonstrating that the model is detailed and flexible enough to account for activation and repression as well as synergistic and self-regulation, and the algorithm can efficiently and accurately recover the parameters used to generate the data. Control theory concerns with the question if and how it is possible to drive the behavior of a complex dynamical system. A system is said to be controllable if we can drive it from any initial state to any desired final state in finite time. For many complex networks, the precise knowledge of system parameters lacks. But, it is possible to make a conclusion about network controllability by inspecting its structure. Classical theory of structural controllability is based on the Lin's structural controllability theorem, which gives necessary and sufficient conditions to conclude if any network is structurally controllable. Due to this fundamental theorem we may identify a minimum driver vertex set, whose control with independent driving signals is sufficient to make the whole system controllable. I show that the Lin's theorem does not impose any limitations on quantum networks structural controllability. By local operations and classical communication, one can modify any quantum network to make it structurally controllable by a single driving signal. Making decisions about the structure of a future military fleet is a challenging task. Several issues need to be considered such as the existence of multiple competing objectives and the complexity of the operating environment. A particular challenge is posed by the various types of uncertainty that the future might hold. It is uncertain what future events might be encountered; how fleet design decisions will influence and shape the future; and how present and future decision makers will act based on available information, their personal biases regarding the importance of different objectives, and their economic preferences. In order to assist strategic decision-making, an analysis of future fleet options needs to account for conditions in which these different classes of uncertainty are exposed. It is important to understand what assumptions a particular fleet is robust to, what the fleet can readily adapt to, and what conditions present clear risks to the fleet. We call this the analysis of a fleet's strategic positioning. This paper introduces how strategic positioning can be evaluated using computer simulations. Our main aim is to introduce a framework for capturing information that can be useful to a decision maker and for defining the concepts of robustness and adaptiveness in the context of future fleet design. We demonstrate our conceptual framework using simulation studies of an air transportation fleet. We capture uncertainty by employing an explorative scenario-based approach. Each scenario represents a sampling of different future conditions, different model assumptions, and different economic preferences. Proposed changes to a fleet are then analysed based on their influence on the fleet's robustness, adaptiveness, and risk to different scenarios. Recent success in training deep neural networks have prompted active investigation into the features learned on their intermediate layers. Such research is difficult because it requires making sense of non-linear computations performed by millions of parameters, but valuable because it increases our ability to understand current models and create improved versions of them. In this paper we investigate the extent to which neural networks exhibit what we call convergent learning, which is when the representations learned by multiple nets converge to a set of features which are either individually similar between networks or where subsets of features span similar low-dimensional spaces. We propose a specific method of probing representations: training multiple networks and then comparing and contrasting their individual, learned representations at the level of neurons or groups of neurons. We begin research into this question using three techniques to approximately align different neural networks on a feature level: a bipartite matching approach that makes one-to-one assignments between neurons, a sparse prediction approach that finds one-to-many mappings, and a spectral clustering approach that finds many-to-many mappings. This initial investigation reveals a few previously unknown properties of neural networks, and we argue that future research into the question of convergent learning will yield many more. The insights described here include (1) that some features are learned reliably in multiple networks, yet other features are not consistently learned; (2) that units learn to span low-dimensional subspaces and, while these subspaces are common to multiple networks, the specific basis vectors learned are not; (3) that the representation codes show evidence of being a mix between a local code and slightly, but not fully, distributed codes across multiple units. We evaluate the uncertainty quality in neural networks using anomaly detection. We extract uncertainty measures (e.g. entropy) from the predictions of candidate models, use those measures as features for an anomaly detector, and gauge how well the detector differentiates known from unknown classes. We assign higher uncertainty quality to candidate models that lead to better detectors. We also propose a novel method for sampling a variational approximation of a Bayesian neural network, called One-Sample Bayesian Approximation (OSBA). We experiment on two datasets, MNIST and CIFAR10. We compare the following candidate neural network models: Maximum Likelihood, Bayesian Dropout, OSBA, and --- for MNIST --- the standard variational approximation. We show that Bayesian Dropout and OSBA provide better uncertainty information than Maximum Likelihood, and are essentially equivalent to the standard variational approximation, but much faster. We study the effect of surface scattering on transport properties in many-mode conducting channels (electron waveguides). Assuming a strong roughness of the surface profiles, we show that there are two independent control parameters that determine statistical properties of the scattering. The first parameter is the ratio of the amplitude of the roughness to the transverse width of the waveguide. The second one, which is typically omitted, is determined by the mean value of the derivative of the profile. This parameter may be large, thus leading to specific properties of scattering. Our results may be used in experimental realizations of the surface scattering of electron waves, as well as for other applications (e.g., for optical and microwave waveguides) An active object recognition system has the advantage of being able to act in the environment to capture images that are more suited for training and that lead to better performance at test time. In this paper, we propose a deep convolutional neural network for active object recognition that simultaneously predicts the object label, and selects the next action to perform on the object with the aim of improving recognition performance. We treat active object recognition as a reinforcement learning problem and derive the cost function to train the network for joint prediction of the object label and the action. A generative model of object similarities based on the Dirichlet distribution is proposed and embedded in the network for encoding the state of the system. The training is carried out by simultaneously minimizing the label and action prediction errors using gradient descent. We empirically show that the proposed network is able to predict both the object label and the actions on GERMS, a dataset for active object recognition. We compare the test label prediction accuracy of the proposed model with Dirichlet and Naive Bayes state encoding. The results of experiments suggest that the proposed model equipped with Dirichlet state encoding is superior in performance, and selects images that lead to better training and higher accuracy of label prediction at test time. Several scenarios of interacting neural networks which are trained either in an identical or in a competitive way are solved analytically. In the case of identical training each perceptron receives the output of its neighbour. The symmetry of the stationary state as well as the sensitivity to the used training algorithm are investigated. Two competitive perceptrons trained on mutually exclusive learning aims and a perceptron which is trained on the opposite of its own output are examined analytically. An ensemble of competitive perceptrons is used as decision-making algorithms in a model of a closed market (El Farol Bar problem or Minority Game); each network is trained on the history of minority decisions. This ensemble of perceptrons relaxes to a stationary state whose performance can be better than random. This paper presents a new state-of-the-art for document image classification and retrieval, using features learned by deep convolutional neural networks (CNNs). In object and scene analysis, deep neural nets are capable of learning a hierarchical chain of abstraction from pixel inputs to concise and descriptive representations. The current work explores this capacity in the realm of document analysis, and confirms that this representation strategy is superior to a variety of popular hand-crafted alternatives. Experiments also show that (i) features extracted from CNNs are robust to compression, (ii) CNNs trained on non-document images transfer well to document analysis tasks, and (iii) enforcing region-specific feature-learning is unnecessary given sufficient training data. This work also makes available a new labelled subset of the IIT-CDIP collection, containing 400,000 document images across 16 categories, useful for training new CNNs for document analysis. The game of Go is more challenging than other board games, due to the difficulty of constructing a position or move evaluation function. In this paper we investigate whether deep convolutional networks can be used to directly represent and learn this knowledge. We train a large 12-layer convolutional neural network by supervised learning from a database of human professional games. The network correctly predicts the expert move in 55% of positions, equalling the accuracy of a 6 dan human player. When the trained convolutional network was used directly to play games of Go, without any search, it beat the traditional search program GnuGo in 97% of games, and matched the performance of a state-of-the-art Monte-Carlo tree search that simulates a million positions per move. This paper describes application of information granulation theory, on the analysis of "lugeon data". In this manner, using a combining of Self Organizing Map (SOM) and Neuro-Fuzzy Inference System (NFIS), crisp and fuzzy granules are obtained. Balancing of crisp granules and sub- fuzzy granules, within non fuzzy information (initial granulation), is rendered in open-close iteration. Using two criteria, "simplicity of rules "and "suitable adaptive threshold error level", stability of algorithm is guaranteed. In other part of paper, rough set theory (RST), to approximate analysis, has been employed >.Validation of the proposed methods, on the large data set of in-situ permeability in rock masses, in the Shivashan dam, Iran, has been highlighted. By the implementation of the proposed algorithm on the lugeon data set, was proved the suggested method, relating the approximate analysis on the permeability, could be applied. Learning and memory are acquired through long-lasting changes in synapses. In the simplest models, such synaptic potentiation typically leads to runaway excitation, but in reality there must exist processes that robustly preserve overall stability of the neural system dynamics. How is this accomplished? Various approaches to this basic question have been considered. Here we propose a particularly compelling and natural mechanism for preserving stability of learning neural systems. This mechanism is based on the global processes by which metabolic resources are distributed to the neurons by glial cells. Specifically, we introduce and study a model comprised of two interacting networks: a model neural network interconnected by synapses which undergo spike-timing dependent plasticity (STDP); and a model glial network interconnected by gap junctions which diffusively transport metabolic resources among the glia and, ultimately, to neural synapses where they are consumed. Our main result is that the biophysical constraints imposed by diffusive transport of metabolic resources through the glial network can prevent runaway growth of synaptic strength, both during ongoing activity and during learning. Our findings suggest a previously unappreciated role for glial transport of metabolites in the feedback control stabilization of neural network dynamics during learning. The motion of driven interfaces in random media at finite temperature $T$ and small external force $F$ is usually described by a linear displacement $h_G(t) \sim V(F,T) t$ at large times, where the velocity vanishes according to the creep formula as $V(F,T) \sim e^{-K(T)/F^{\mu}}$ for $F \to 0$. In this paper, we question this picture on the specific example of the directed polymer in a two dimensional random medium. We have recently shown (C. Monthus and T. Garel, arxiv:0802.2502) that its dynamics for F=0 can be analyzed in terms of a strong disorder renormalization procedure, where the distribution of renormalized barriers flows towards some "infinite disorder fixed point". In the present paper, we obtain that for small $F$, this "infinite disorder fixed point" becomes a "strong disorder fixed point" with an exponential distribution of renormalized barriers. The corresponding distribution of trapping times then only decays as a power-law $P(\tau) \sim 1/\tau^{1+\alpha}$, where the exponent $\alpha(F,T)$ vanishes as $\alpha(F,T) \propto F^{\mu}$ as $F \to 0$. Our conclusion is that in the small force region $\alpha(F,T)<1$, the divergence of the averaged trapping time $\bar{\tau}=+\infty$ induces strong non-self-averaging effects that invalidate the usual creep formula obtained by replacing all trapping times by the typical value. We find instead that the motion is only sub-linearly in time $h_G(t) \sim t^{\alpha(F,T)}$, i.e. the asymptotic velocity vanishes V=0. This analysis is confirmed by numerical simulations of a directed polymer with a metric constraint driven in a traps landscape. We moreover obtain that the roughness exponent, which is governed by the equilibrium value $\zeta_{eq}=2/3$ up to some large scale, becomes equal to $\zeta=1$ at the largest scales. In practice, since many communication networks are huge in scale, or complicated in structure, or even dynamic, the predesigned linear network codes based on the network topology is impossible even if the topological structure is known. Therefore, random linear network coding has been proposed as an acceptable coding technique for the case that the network topology cannot be utilized completely. Motivated by the fact that different network topological information can be obtained for different practical applications, we study the performance analysis of random linear network coding by analyzing some failure probabilities depending on these different topological information of networks. We obtain some tight or asymptotically tight upper bounds on these failure probabilities and indicate the worst cases for these bounds, i.e., the networks meeting the upper bounds with equality. In addition, if the more topological information of the network is utilized, the better upper bounds are obtained. On the other hand, we also discuss the lower bounds on the failure probabilities. Wireless Sensor Networks (WSNs), with growing applications in the environment which are not within human reach have been addressed tremendously in the recent past. For optimized working of network many routing algorithms have been proposed, mainly focusing energy efficiency, network lifetime, clustering processes. Considering homogeneity of network, we proposed Energy Efficient Sleep Awake Aware (EESAA) intelligent routing protocol for WSNs. In our proposed technique we evaluate and enhance certain issues like network stability, network lifetime and cluster head selection process. Utilizing the concept of characteristical pairing among sensor nodes energy utilization is optimized. Simulation results show that our proposed protocolnificantly improved the Inspired by biophysical principles underlying nonlinear dendritic computation in neural circuits, we develop a scheme to train deep neural networks to make them robust to adversarial attacks. Our scheme generates highly nonlinear, saturated neural networks that achieve state of the art performance on gradient based adversarial examples on MNIST, despite never being exposed to adversarially chosen examples during training. Moreover, these networks exhibit unprecedented robustness to targeted, iterative schemes for generating adversarial examples, including second-order methods. We further identify principles governing how these networks achieve their robustness, drawing on methods from information geometry. We find these networks progressively create highly flat and compressed internal representations that are sensitive to very few input dimensions, while still solving the task. Moreover, they employ highly kurtotic weight distributions, also found in the brain, and we demonstrate how such kurtosis can protect even linear classifiers from adversarial attack. We present new Neutron Spin Echo (NSE) results and a revisited analysis of historical data on spin glasses, which reveal a pure power-law time decay of the spin autocorrelation function $s(Q,t) = S(Q,t)/S(Q)$ at the glass temperature $T_g$, each power law exponent being in excellent agreement with that calculated from dynamic and static critical exponents deduced from macroscopic susceptibility measurements made on a quite different time scale. It is the first time that this scaling relation involving exponents of different physical quantities determined by completely independent experimental methods is stringently verified experimentally in a spin glass. As spin glasses are a subgroup of the vast family of glassy systems also comprising structural glasses, other non-crystalline systems living matter the observed strict critical scaling behaviour is important. Above the phase transition the strikingly non-exponential relaxation, best fitted by the Ogielski (power-law times stretched exponential) function, appears as an intrinsic, homogeneous feature of spin glasses. Network sparsification aims to reduce the number of edges of a network while maintaining its structural properties; such properties include shortest paths, cuts, spectral measures, or network modularity. Sparsification has multiple applications, such as, speeding up graph-mining algorithms, graph visualization, as well as identifying the important network edges. In this paper we consider a novel formulation of the network-sparsification problem. In addition to the network, we also consider as input a set of communities. The goal is to sparsify the network so as to preserve the network structure with respect to the given communities. We introduce two variants of the community-aware sparsification problem, leading to sparsifiers that satisfy different connectedness community properties. From the technical point of view, we prove hardness results and devise effective approximation algorithms. Our experimental results on a large collection of datasets demonstrate the effectiveness of our algorithms. As data sets grow in size, the ability of learning methods to find structure in them is increasingly hampered by the time needed to search the large spaces of possibilities and generate a score for each that takes all of the observed data into account. For instance, Bayesian networks, the model chosen in this paper, have a super-exponentially large search space for a fixed number of variables. One possible method to alleviate this problem is to use a proxy, such as a Gaussian Process regressor, in place of the true scoring function, training it on a selection of sampled networks. We prove here that the use of such a proxy is well-founded, as we can bound the smoothness of a commonly-used scoring function for Bayesian network structure learning. We show here that, compared to an identical search strategy using the network?s exact scores, our proxy-based search is able to get equivalent or better scores on a number of data sets in a fraction of the time. In this paper, we design a Deep Dual-Domain ($\mathbf{D^3}$) based fast restoration model to remove artifacts of JPEG compressed images. It leverages the large learning capacity of deep networks, as well as the problem-specific expertise that was hardly incorporated in the past design of deep architectures. For the latter, we take into consideration both the prior knowledge of the JPEG compression scheme, and the successful practice of the sparsity-based dual-domain approach. We further design the One-Step Sparse Inference (1-SI) module, as an efficient and light-weighted feed-forward approximation of sparse coding. Extensive experiments verify the superiority of the proposed $D^3$ model over several state-of-the-art methods. Specifically, our best model is capable of outperforming the latest deep model for around 1 dB in PSNR, and is 30 times faster. Convolutional neural nets (CNNs) have become a practical means to perform vision tasks, particularly in the area of image classification. FPGAs are well known to be able to perform convolutions efficiently, however, most recent efforts to run CNNs on FPGAs have shown limited advantages over other devices such as GPUs. Previous approaches on FPGAs have often been memory bound due to the limited external memory bandwidth on the FPGA device. We show a novel architecture written in OpenCL(TM), which we refer to as a Deep Learning Accelerator (DLA), that maximizes data reuse and minimizes external memory bandwidth. Furthermore, we show how we can use the Winograd transform to significantly boost the performance of the FPGA. As a result, when running our DLA on Intel's Arria 10 device we can achieve a performance of 1020 img/s, or 23 img/s/W when running the AlexNet CNN benchmark. This comes to 1382 GFLOPs and is 10x faster with 8.4x more GFLOPS and 5.8x better efficiency than the state-of-the-art on FPGAs. Additionally, 23 img/s/W is competitive against the best publicly known implementation of AlexNet on nVidia's TitanX GPU. The Ising Model has recently received much attention for the statistical description of neural spike train data. In this paper, we propose and demonstrate its use for building decoders capable of predicting, on a millisecond timescale, the stimulus represented by a pattern of neural activity. After fitting to a training dataset, the Ising decoder can be applied "online" for instantaneous decoding of test data. While such models can be fit exactly using Boltzmann learning, this approach rapidly becomes computationally intractable as neural ensemble size increases. We show that several approaches, including the Thouless-Anderson-Palmer (TAP) mean field approach from statistical physics, and the recently developed Minimum Probability Flow Learning (MPFL) algorithm, can be used for rapid inference of model parameters in large-scale neural ensembles. Use of the Ising model for decoding, unlike other problems such as functional connectivity estimation, requires estimation of the partition function. As this involves summation over all possible responses, this step can be limiting. Mean field approaches avoid this problem by providing an analytical expression for the partition function. We demonstrate these decoding techniques by applying them to simulated neural ensemble responses from a mouse visual cortex model, finding an improvement in decoder performance for a model with heterogeneous as opposed to homogeneous neural tuning and response properties. Our results demonstrate the practicality of using the Ising model to read out, or decode, spatial patterns of activity comprised of many hundreds of neurons. It has been recently reported that the reciprocity of real-life weighted networks is very pronounced, however its impact on dynamical processes is poorly understood. In this paper, we study random walks in a scale-free directed weighted network with a trap at the central hub node, where the weight of each directed edge is dominated by a parameter controlling the extent of network reciprocity. We derive two expressions for the mean first passage time (MFPT) to the trap, by using two different techniques, the results of which agree well with each other. We also analytically determine all the eigenvalues as well as their multiplicities for the fundamental matrix of the dynamical process, and show that the largest eigenvalue has an identical dominant scaling as that of the MFPT.We find that the weight parameter has a substantial effect on the MFPT, which behaves as a power-law function of the system size with the power exponent dependent on the parameter, signaling the crucial role of reciprocity in random walks occurring in weighted networks. It has remained an open question for some time whether, given a set of not necessarily binary (i.e. "nonbinary") trees T on a set of taxa X, it is possible to determine in time f(r).poly(m) whether there exists a phylogenetic network that displays all the trees in T, where r refers to the reticulation number of the network and m=|X|+|T|. Here we show that this holds if one or both of the following conditions holds: (1) |T| is bounded by a function of r; (2) the maximum degree of the nodes in T is bounded by a function of r. These sufficient conditions absorb and significantly extend known special cases, namely when all the trees in T are binary, or T contains exactly two nonbinary trees. We believe this result is an important step towards settling the issue for an arbitrarily large and complex set of nonbinary trees. For completeness we show that the problem is certainly solveable in polynomial time. Social networks enable users to freely communicate with each other and share their recent news, ongoing activities or views about different topics. As a result, they can be seen as a potentially viable source of information to understand the current emerging topics/events. The ability to model emerging topics is a substantial step to monitor and summarize the information originating from social sources. Applying traditional methods for event detection which are often proposed for processing large, formal and structured documents, are less effective, due to the short length, noisiness and informality of the social posts. Recent event detection techniques address these challenges by exploiting the opportunities behind abundant information available in social networks. This article provides an overview of the state of the art in event detection from social networks. Due to deployment in hostile environment, wireless sensor network is vulnerable to various attacks. Exhausted sensor nodes in sensor network become a challenging issue because it disrupts the normal connectivity of the network. Affected nodes give rise to denial of service that resists to get the objective of sensor network in real life. A mathematical model based on Absorbing Markov Chain (AMC)is proposed for Denial of Sleep attack detection in sensor network. In this mechanism, whether sensor network is affected by denial of sleep attack or not can be decided by considering expected death time of sensor network under normal scenario. Gain and order scheduling of fractional order (FO) PI{\lambda}D{\mu} controllers are studied in this paper considering four different classes of higher order processes. The mapping between the optimum PID/FOPID controller parameters and the reduced order process models are done using Radial Basis Function (RBF) type Artificial Neural Network (ANN). Simulation studies have been done to show the effectiveness of the RBFNN for online scheduling of such controllers with random change in set-point and process parameters. Based on the requirement in the simulation of lepton-nucleus deep inelastic scattering (DIS), we construct a fortran program LDCS 1.0 calculating the differential and total cross sections for the unpolarized charged lepton-unpolarized nucleon and neutrino-unpolarized nucleon neutral current (charged current) DIS at leading order. Any set of the experimentally fitted parton distribution functions could be employed directly. The mass of incident and scattered leptons is taken into account and the boundary conditions calculating the single differential and total cross section are studied. The calculated results well agree with the corresponding experimental data which indicating the LDCS 1.0 program is good. It is also turned out that the effect of tauon mass is not negligible in the GeV energy level. An artificial neural network for extracting reasonable and fast estimates of hyperfine parameters from M\"ossbauer spectra in the energy or time domain is outlined. First promising results for determining the asymmetry of the electric field gradient at the nucleus of a diamagnetic iron center as derived with different types of neural networks are reported. A dynamical mean-field approximation (DMA) previously proposed by the present author [H. Hasegawa, Phys. Rev E {\bf 67}, 041903 (2003)] has been extended to ensembles described by a general noisy spiking neuron model. Ensembles of $N$-unit neurons, each of which is expressed by coupled $K$-dimensional differential equations (DEs), are assumed to be subject to spatially correlated white noises. The original $KN$-dimensional {\it stochastic} DEs have been replaced by $K(K+2)$-dimensional {\it deterministic} DEs expressed in terms of means and the second-order moments of {\it local} and {\it global} variables: the fourth-order contributions are taken into account by the Gaussian decoupling approximation. Our DMA has been applied to an ensemble of Hodgkin-Huxley (HH) neurons (K=4), for which effects of the noise, the coupling strength and the ensemble size on the response to a single-spike input have been investigated. Results calculated by DMA theory are in good agreement with those obtained by direct simulations. Recently, there has been a growing concern about the overload status of the power grid networks, and the increasing possibility of cascading failures. Many researchers have studied these networks to provide design guidelines for more robust power grids. Topological analysis is one of the components of system analysis for its robustness. This paper presents a complex systems analysis of power grid networks. First, the cascading effect has been simulated on three well known networks: the IEEE 300 bus test system, the IEEE 118 bus test system, and the WSCC 179 bus equivalent model. To extend the analysis to a larger set of networks, we develop a network generator and generate multiple graphs with characteristics similar to the IEEE test networks but with different topologies. The generated graphs are then compared to the test networks to show the effect of topology in determining their robustness with respect to cascading failures. The generated graphs turn out to be more robust than the test graphs, showing the importance of topology in the robust design of power grids. The second part of this paper concerns the discussion of two novel mitigation strategies for cascading failures: Targeted Load Reduction and Islanding using Distributed Sources. These new mitigation strategies are compared with the Homogeneous Load Reduction strategy. Even though the Homogeneous Load Reduction is simpler to implement, the Targeted Load Reduction is much more effective. Additionally, an algorithm is presented for the partitioning of the network for islanding as an effort towards fault isolation to prevent cascading failures. The results for island formation are better if the sources are well distributed, else the algorithm leads to the formation of superislands. We investigate site percolation in a hierarchical scale-free network known as the Dorogovtsev- Goltsev-Mendes network. We use the generating function method to show that the percolation threshold is 1, i.e., the system is not in the percolating phase when the occupation probability is less than 1. The present result is contrasted to bond percolation in the same network of which the percolation threshold is zero. We also show that the percolation threshold of intentional attacks is 1. Our results suggest that this hierarchical scale-free network is very fragile against both random failure and intentional attacks. Such a structural defect is common in many hierarchical network models. We investigate layered neural networks with differentiable activation function and student vectors without normalization constraint by means of equilibrium statistical physics. We consider the learning of perfectly realizable rules and find that the length of student vectors becomes infinite, unless a proper weight decay term is added to the energy. Then, the system undergoes a first order phase transition between states with very long student vectors and states where the lengths are comparable to those of the teacher vectors. Additionally in both configurations there is a phase transition between a specialized and an unspecialized phase. An anti-specialized phase with long student vectors exists in networks with a small number of hidden units. A quantum-mechanical analysis of hyper-fast (faster than ballistic) diffusion of a quantum wave packet in random optical lattices is presented. The main motivation of the presented analysis is experimental demonstrations of hyper-diffusive spreading of a wave packet in random photonic lattices [L. Levi \textit{et al.}, Nature Phys. \textbf{8}, 912 (2012)]. A rigorous quantum-mechanical calculation of the mean probability amplitude is suggested, and it is shown that the power law spreading of the mean squared displacement (MSD) is $< x^2(t)>\sim t^{\alpha}$, where $2<\alpha\leq 3$. The values of the transport exponent $\alpha$ depend on the correlation properties of the random potential $V(x,t)$, which describes random inhomogeneities of the medium. In particular, when the random potential is $\delta$ correlated in time, the quantum wave packet spreads according Richardson turbulent diffusion with the MSD $\sim t^3$. Hyper-diffusion with $\alpha=12/5$ is also obtained for arbitrary correlation properties of the random potential. Not only is network coding essential to achieve the capacity of a single-session multicast network, it can also help to improve the throughput of wireless networks with multiple unicast sessions when overheard information is available. Most previous research aimed at realizing such improvement by using perfectly overheard information, while in practice, especially for wireless networks, overheard information is often imperfect. To date, it is unclear whether network coding should still be used in such situations with imperfect overhearing. In this paper, a simple but ubiquitous wireless network model with two unicast sessions is used to investigate this problem. From the diversity and multiplexing tradeoff perspective, it is proved that even when overheard information is imperfect, network coding can still help to improve the overall system performance. This result implies that network coding should be used actively regardless of the reception quality of overheard information. The paper introduces a connectionist network approach to find numerical solutions of Diophantine equations as an attempt to address the famous Hilbert's tenth problem. The proposed methodology uses a three layer feed forward neural network with back propagation as sequential learning procedure to find numerical solutions of a class of Diophantine equations. It uses a dynamically constructed network architecture where number of nodes in the input layer is chosen based on the number of variables in the equation. The powers of the given Diophantine equation are taken as input to the input layer. The training of the network starts with initial random integral weights. The weights are updated based on the back propagation of the error values at the output layer. The optimization of weights is augmented by adding a momentum factor into the network. The optimized weights of the connection between the input layer and the hidden layer are taken as numerical solution of the given Diophantine equation. The procedure is validated using different Diophantine Equations of different number of variables and different powers. Short-term synaptic depression and facilitation have been found to greatly influence the performance of autoassociative neural networks. However, only partial results, focused for instance on the computation of the maximum storage capacity at zero temperature, have been obtained to date. In this work, we extended the study of the effect of these synaptic mechanisms on autoassociative neural networks to more realistic and general conditions, including the presence of noise in the system. In particular, we characterized the behavior of the system by means of its phase diagrams, and we concluded that synaptic facilitation significantly enlarges the region of good retrieval performance of the network. We also found that networks with facilitating synapses may have critical temperatures substantially higher than those of standard autoassociative networks, thus allowing neural networks to perform better under high-noise conditions. We have created a large diverse set of cars from overhead images, which are useful for training a deep learner to binary classify, detect and count them. The dataset and all related material will be made publically available. The set contains contextual matter to aid in identification of difficult targets. We demonstrate classification and detection on this dataset using a neural network we call ResCeption. This network combines residual learning with Inception-style layers and is used to count cars in one look. This is a new way to count objects rather than by localization or density estimation. It is fairly accurate, fast and easy to implement. Additionally, the counting method is not car or scene specific. It would be easy to train this method to count other kinds of objects and counting over new scenes requires no extra set up or assumptions about object locations. The predominate traffic patterns in a wireless sensor network are many-to-one and one-to-many communication. Hence, the performance of wireless sensor networks is characterized by the rate at which data can be disseminated from or aggregated to a data sink. In this paper, we consider the data aggregation problem. We demonstrate that a data aggregation rate of O(log(n)/n) is optimal and that this rate can be achieved in wireless sensor networks using a generalization of cooperative beamforming called cooperative time-reversal communication. In this paper, we have used Recurrent Neural Networks to capture and model human motion data and generate motions by prediction of the next immediate data point at each time-step. Our RNN is armed with recently proposed Gated Recurrent Units which has shown promising results in some sequence modeling problems such as Machine Translation and Speech Synthesis. We demonstrate that this model is able to capture long-term dependencies in data and generate realistic motions. Most algorithms for propagating evidence through belief networks have been exact and exhaustive: they produce an exact (point-valued) marginal probability for every node in the network. Often, however, an application will not need information about every n ode in the network nor will it need exact probabilities. We present the localized partial evaluation (LPE) propagation algorithm, which computes interval bounds on the marginal probability of a specified query node by examining a subset of the nodes in the entire network. Conceptually, LPE ignores parts of the network that are "too far away" from the queried node to have much impact on its value. LPE has the "anytime" property of being able to produce better solutions (tighter intervals) given more time to consider more of the network. In this letter, we consider the effect of clustering coefficient on the synchronizability of coupled oscillators located on scale-free networks. The analytic result for the value of clustering coefficient aiming at a highly clustered scale-free network model, the Holme-Kim is obtained, and the relationship between network synchronizability and clustering coefficient is reported. The simulation results strongly suggest that the more clustered the network, the poorer its synchronizability. The driving force behind deep networks is their ability to compactly represent rich classes of functions. The primary notion for formally reasoning about this phenomenon is expressive efficiency, which refers to a situation where one network must grow unfeasibly large in order to realize (or approximate) functions of another. To date, expressive efficiency analyses focused on the architectural feature of depth, showing that deep networks are representationally superior to shallow ones. In this paper we study the expressive efficiency brought forth by connectivity, motivated by the observation that modern networks interconnect their layers in elaborate ways. We focus on dilated convolutional networks, a family of deep models delivering state of the art performance in sequence processing tasks. By introducing and analyzing the concept of mixed tensor decompositions, we prove that interconnecting dilated convolutional networks can lead to expressive efficiency. In particular, we show that even a single connection between intermediate layers can already lead to an almost quadratic gap, which in large-scale settings typically makes the difference between a model that is practical and one that is not. Empirical evaluation demonstrates how the expressive efficiency of connectivity, similarly to that of depth, translates into gains in accuracy. This leads us to believe that expressive efficiency may serve a key role in the development of new tools for deep network design. In this paper, we investigate deep image synthesis guided by sketch, color, and texture. Previous image synthesis methods can be controlled by sketch and color strokes but we are the first to examine texture control. We allow a user to place a texture patch on a sketch at arbitrary location and scale to control the desired output texture. Our generative network learns to synthesize objects consistent with these texture suggestions. To achieve this, we develop a local texture loss in addition to adversarial and content loss to train the generative network. The new local texture loss can improve generated texture quality without knowing the patch location and size in advance. We conduct experiments using sketches generated from real images and textures sampled from the Describable Textures Dataset and results show that our proposed algorithm is able to generate plausible images that are faithful to user controls. Ablation studies show that our proposed pipeline can generate more realistic images than adapting existing methods directly. We propose a mechanism to describe spin relaxation in n-doped III-V semiconductors close to the Mott metal-insulator transition. Taking into account the spin-orbit interaction induced spin admixture in the hydrogenic donor states, we build a tight-binding model for the spin-dependent impurity band. Since the hopping amplitudes with spin flip are considerably smaller than the spin conserving counterparts, the resulting spin lifetime is very large. We estimate the spin lifetime from the diffusive accumulation of spin rotations associated with the electron hopping. Our result is larger but of the same order of magnitude than the experimental value. Therefore the proposed mechanism has to be included when describing spin relaxation in the impurity band. Self-organizing cyber-physical systems are expected to become increasingly important in the context of Industry 4.0 automation as well as in everyday scenarios. Resilient communication is crucial for such systems. In general, this can be achieved with redundant communication paths. Mathematically, the amount of redundant paths is expressed with the network connectivity. A high network connectivity is required for collaboration and system-wide self-adaptation even when nodes fail or get compromised by an attacker. In this paper, we analyze the network connectivity of a communication network for large distributed cyber-physical systems. For this, we simulate the communication structure of a CPS with different network parameters to determine its resilience. With our results, we also deduce the required network connectivity for a given number of failing or compromised nodes. In this work we study a weak Prisoner\^as Dilemma game in which both strategies and update rules are subjected to evolutionary pressure. Interactions among agents are specified by complex topologies, and we consider both homogeneous and heterogeneous situations. We consider deterministic and stochastic update rules for the strategies, which in turn may consider single links or full context when selecting agents to copy from. Our results indicate that the co-evolutionary process preserves heterogeneous networks as a suitable framework for the emergence of cooperation. Furthermore, on those networks, the update rule leading to a larger fraction, which we call replicator dynamics, is selected during co-evolution. On homogeneous networks we observe that even if replicator dynamics turns out again to be the selected update rule, the cooperation level is larger than on a fixed update rule framework. We conclude that for a variety of topologies, the fact that the dynamics coevolves with the strategies leads in general to more cooperation in the weak Prisoner's Dilemma game. In machine learning, there is a fundamental trade-off between ease of optimization and expressive power. Neural Networks, in particular, have enormous expressive power and yet are notoriously challenging to train. The nature of that optimization challenge changes over the course of learning. Traditionally in deep learning, one makes a static trade-off between the needs of early and late optimization. In this paper, we investigate a novel framework, GradNets, for dynamically adapting architectures during training to get the benefits of both. For example, we can gradually transition from linear to non-linear networks, deterministic to stochastic computation, shallow to deep architectures, or even simple downsampling to fully differentiable attention mechanisms. Benefits include increased accuracy, easier convergence with more complex architectures, solutions to test-time execution of batch normalization, and the ability to train networks of up to 200 layers. We study the finite size fluctuations at the depinning transition for a one-dimensional elastic interface of size $L$ displacing in a disordered medium of transverse size $M=k L^\zeta$ with periodic boundary conditions, where $\zeta$ is the depinning roughness exponent and $k$ is a finite aspect ratio parameter. We focus on the crossover from the infinitely narrow ($k\to 0$) to the infinitely wide ($k\to \infty$) medium. We find that at the thermodynamic limit both the value of the critical force and the precise behavior of the velocity-force characteristics are {\it unique} and $k$-independent. We also show that the finite size fluctuations of the critical force (bias and variance) as well as the global width of the interface cross over from a power-law to a logarithm as a function of $k$. Our results are relevant for understanding anisotropic size-effects in force-driven and velocity-driven interfaces. In this paper I will describe some results that have been recently obtained in the study of random Euclidean matrices, i.e. matrices that are functions of random points in Euclidean space. In the case of translation invariant matrices one generically finds a phase transition between a phonon phase and a saddle phase. If we apply these considerations to the study of the Hessian of the Hamiltonian of the particles of a fluid, we find that this phonon-saddle transition corresponds to the dynamical phase transition in glasses, that has been studied in the framework of the mode coupling approximation. The Boson peak observed in glasses at low temperature is a remanent of this transition. Deep generative models parameterized by neural networks have recently achieved state-of-the-art performance in unsupervised and semi-supervised learning. We extend deep generative models with auxiliary variables which improves the variational approximation. The auxiliary variables leave the generative model unchanged but make the variational distribution more expressive. Inspired by the structure of the auxiliary variable we also propose a model with two stochastic layers and skip connections. Our findings suggest that more expressive and properly specified deep generative models converge faster with better results. We show state-of-the-art performance within semi-supervised learning on MNIST, SVHN and NORB datasets. The Deviants' Dilemma is a two-person game with the individual gain conflicting with the choice for global good. Evolutionary considerations yield fixed point attractors, with the phenomena of exclusion potentially playing an important role when current opponent information is available. We carry out computer simulations which substantiates and illuminates theoretical claims, and brings to light the pertinance of the choice between deterministic and stochastic dynamics, and the conjecture of 'ergodicity spread'. The multi-index matching is an NP-hard combinatorial optimization problem; for two indices it reduces to the well understood bipartite matching problem that belongs to the polynomial complexity class. We use the cavity method to solve the thermodynamics of the multi-index system with random costs. The phase diagram is much richer than for the case of the bipartite matching problem: it shows a finite temperature phase transition to a completely frozen glass phase, similar to what happens in the random energy model. We derive the critical temperature, the ground state energy density, and properties of the energy landscape, and compare the results to numerical studies based on exact analysis of small systems. Spatio-temporal dynamics of excitable media with discrete three-level active centers (ACs) and absorbing boundaries is studied numerically by means of a deterministic three-level model (see S. D. Makovetskiy and D. N. Makovetskii, on-line preprint cond-mat/0410460 ), which is a generalization of Zykov- Mikhailov model (see Sov. Phys. -- Doklady, 1986, Vol.31, No.1, P.51) for the case of two-channel diffusion of excitations. In particular, we revealed some qualitatively new features of coexistence, competition and collapse of rotating spiral waves (RSWs) in three-level excitable media under conditions of strong influence of the second channel of diffusion. Part of these features are caused by unusual mechanism of RSWs evolution when RSW's cores get into the surface layer of an active medium (i.~e. the layer of ACs resided at the absorbing boundary). Instead of well known scenario of RSW collapse, which takes place after collision of RSW's core with absorbing boundary, we observed complicated transformations of the core leading to nonlinear ''reflection'' of the RSW from the boundary or even to birth of several new RSWs in the surface layer. To our knowledge, such nonlinear ''reflections'' of RSWs and resulting die hard vorticity in excitable media with absorbing boundaries were unknown earlier. ACM classes: F.1.1, I.6, J.2; PACS numbers: 05.65.+b, 07.05.Tp, 82.20.Wt We analyse the newest diffractive deep inelastic scattering data from HERA using the dipole model approach. We find a reasonable good agreement between the predictions and the data although the region of small values of a kinematic variable $\beta$ needs refinement. A way to do this is to consider an approach with diffractive parton distributions evolved with the DGLAP evolution equations. We describe algorithms for learning Bayesian networks from a combination of user knowledge and statistical data. The algorithms have two components: a scoring metric and a search procedure. The scoring metric takes a network structure, statistical data, and a user's prior knowledge, and returns a score proportional to the posterior probability of the network structure given the data. The search procedure generates networks for evaluation by the scoring metric. Previous work has concentrated on metrics for domains containing only discrete variables, under the assumption that data represents a multinomial sample. In this paper, we extend this work, developing scoring metrics for domains containing all continuous variables or a mixture of discrete and continuous variables, under the assumption that continuous data is sampled from a multivariate normal distribution. Our work extends traditional statistical approaches for identifying vanishing regression coefficients in that we identify two important assumptions, called event equivalence and parameter modularity, that when combined allow the construction of prior distributions for multivariate normal parameters from a single prior Bayesian network specified by a user. This paper proposes a new approach to automatically quantify the severity of knee osteoarthritis (OA) from radiographs using deep convolutional neural networks (CNN). Clinically, knee OA severity is assessed using Kellgren \& Lawrence (KL) grades, a five point scale. Previous work on automatically predicting KL grades from radiograph images were based on training shallow classifiers using a variety of hand engineered features. We demonstrate that classification accuracy can be significantly improved using deep convolutional neural network models pre-trained on ImageNet and fine-tuned on knee OA images. Furthermore, we argue that it is more appropriate to assess the accuracy of automatic knee OA severity predictions using a continuous distance-based evaluation metric like mean squared error than it is to use classification accuracy. This leads to the formulation of the prediction of KL grades as a regression problem and further improves accuracy. Results on a dataset of X-ray images and KL grades from the Osteoarthritis Initiative (OAI) show a sizable improvement over the current state-of-the-art. This work analyses the practice of sister city pairing. We investigate structural properties of the resulting city and country networks and present rankings of the most central nodes in these networks. We identify different country clusters and find that the practice of sister city pairing is not influenced by geographical proximity but results in highly assortative networks. We prove tight network topology dependent bounds on the round complexity of computing well studied $k$-party functions such as set disjointness and element distinctness. Unlike the usual case in the CONGEST model in distributed computing, we fix the function and then vary the underlying network topology. This complements the recent such results on total communication that have received some attention. We also present some applications to distributed graph computation problems. Our main contribution is a proof technique that allows us to reduce the problem on a general graph topology to a relevant two-party communication complexity problem. However, unlike many previous works that also used the same high level strategy, we do not reason about a two-party communication problem that is induced by a cut in the graph. To `stitch' back the various lower bounds from the two party communication problems, we use the notion of timed graph that has seen prior use in network coding. Our reductions use some tools from Steiner tree packing and multi-commodity flow problems that have a delay constraint. The learning dynamics of on-line independent component analysis is analysed in the limit of large data dimension. We study a simple Hebbian learning algorithm that can be used to separate out a small number of non-Gaussian components from a high-dimensional data set. The de-mixing matrix parameters are confined to a Stiefel manifold of tall, orthogonal matrices and we introduce a natural gradient variant of the algorithm which is appropriate to learning on this manifold. For large input dimension the parameter trajectory of both algorithms passes through a sequence of unstable fixed points, each described by a diffusion process in a polynomial potential. Choosing the learning rate too large increases the escape time from each of these fixed points, effectively trapping the learning in a sub-optimal state. In order to avoid these trapping states a very low learning rate must be chosen during the learning transient, resulting in learning time-scales of $O(N^2)$ or $O(N^3)$ iterations where $N$ is the data dimension. Escape from each sub-optimal state results in a sequence of symmetry breaking events as the algorithm learns each source in turn. This is in marked contrast to the learning dynamics displayed by related on-line learning algorithms for multilayer neural networks and principal component analysis. Although the natural gradient variant of the algorithm has nice asymptotic convergence properties, it has an equivalent transient dynamics to the standard Hebbian algorithm. $Range$ and $load$ play keys on the problem of attacking on links in random scale-free (RSF) networks. In this Brief Report we obtain the relation between $range$ and $load$ in RSF networks analytically by the generating function theory, and then give an estimation about the impact of attacks on the $efficiency$ of the network. The analytical results show that short range attacks are more destructive for RSF networks, and are confirmed numerically. Further our results are consistent with the former literature (Physical Review E \textbf{66}, 065103(R) (2002)). This paper presents a novel approach in a rarely studied area of computer vision: Human interaction recognition in still images. We explore whether the facial regions and their spatial configurations contribute to the recognition of interactions. In this respect, our method involves extraction of several visual features from the facial regions, as well as incorporation of scene characteristics and deep features to the recognition. Extracted multiple features are utilized within a discriminative learning framework for recognizing interactions between people. Our designed facial descriptors are based on the observation that relative positions, size and locations of the faces are likely to be important for characterizing human interactions. Since there is no available dataset in this relatively new domain, a comprehensive new dataset which includes several images of human interactions is collected. Our experimental results show that faces and scene characteristics contain important information to recognize interactions between people. We present broad band photometry and photometric redshifts for 187611 sources located in ~0.5deg^2 in the Lockman Hole area. The catalog includes 389 X-ray detected sources identified with the very deep XMM-Newton observations available for an area of 0.2 deg^2. The source detection was performed on the Rc, z' and B band images and the available photometry is spanning from the far ultraviolet to the mid infrared, reaching in the best case scenario 21 bands. Astrometry corrections and photometric cross-calibrations over the entire dataset allowed the computation of accurate photometric redshifts. Special treatment is undertaken for the X-ray sources, the majority of which is active galactic nuclei. Comparing the photometric redshifts to the available spectroscopic redshifts we achieve for normal galaxies an accuracy of \sigma_{\Delta z/(1+z)}=0.036, with 12.7% outliers, while for the X-ray detected sources the accuracy is \sigma_{\Delta z/(1+z)}=0.069, with 18.3% outliers, where the outliers are defined as sources with |z_{phot}-z_{spec}|>0.15 (1+z_{spec})}. These results are a significant improvement over the previously available photometric redshifts for normal galaxies in the Lockman Hole, while it is the first time that photometric redshifts are computed and made public for AGN for this field. With social networking sites providing increasingly richer context, User-Centric Service (UCS) creation is expected to explode following a similar success path to User-Generated Content. One of the major challenges in this emerging highly user-centric networking paradigm is how to make these exploding in numbers yet, individually, of vanishing demand services available in a cost-effective manner. Of prime importance to the latter (and focus of this paper) is the determination of the optimal location for hosting a UCS. Taking into account the particular characteristics of UCS, we formulate the problem as a facility location problem and devise a distributed and highly scalable heuristic solution to it. Key to the proposed approach is the introduction of a novel metric drawing on Complex Network Analysis. Given a current location of UCS, this metric helps to a) identify a small subgraph of nodes with high capacity to act as service demand concentrators; b) project on them a reduced yet accurate view of the global demand distribution that preserves the key attraction forces on UCS; and, ultimately, c) pave the service migration path towards its optimal location in the network. The proposed iterative UCS migration algorithm, called cDSMA, is extensively evaluated over synthetic and real-world network topologies. Our results show that cDSMA achieves high accuracy, fast convergence, remarkable insensitivity to the size and diameter of the network and resilience to inaccurate estimates of demands for UCS across the network. It is also shown to clearly outperform local-search heuristics for service migration that constrain the subgraph to the immediate neighbourhood of the node currently hosting UCS. Extracting per-frame features using convolutional neural networks for real-time processing of video data is currently mainly performed on powerful GPU-accelerated workstations and compute clusters. However, there are many applications such as smart surveillance cameras that require or would benefit from on-site processing. To this end, we propose and evaluate a novel algorithm for change-based evaluation of CNNs for video data recorded with a static camera setting, exploiting the spatio-temporal sparsity of pixel changes. We achieve an average speed-up of 8.6x over a cuDNN baseline on a realistic benchmark with a negligible accuracy loss of less than 0.1% and no retraining of the network. The resulting energy efficiency is 10x higher than that of per-frame evaluation and reaches an equivalent of 328 GOp/s/W on the Tegra X1 platform. In this paper, we propose a novel Deep Localized Makeup Transfer Network to automatically recommend the most suitable makeup for a female and synthesis the makeup on her face. Given a before-makeup face, her most suitable makeup is determined automatically. Then, both the beforemakeup and the reference faces are fed into the proposed Deep Transfer Network to generate the after-makeup face. Our end-to-end makeup transfer network have several nice properties including: (1) with complete functions: including foundation, lip gloss, and eye shadow transfer; (2) cosmetic specific: different cosmetics are transferred in different manners; (3) localized: different cosmetics are applied on different facial regions; (4) producing naturally looking results without obvious artifacts; (5) controllable makeup lightness: various results from light makeup to heavy makeup can be generated. Qualitative and quantitative experiments show that our network performs much better than the methods of [Guo and Sim, 2009] and two variants of NerualStyle [Gatys et al., 2015a]. We consider the problem of dynamic spectrum access for network utility maximization in multichannel wireless networks. The shared bandwidth is divided into K orthogonal channels, and the users access the spectrum using a random access protocol. In the beginning of each time slot, each user selects a channel and transmits a packet with a certain attempt probability. After each time slot, each user that has transmitted a packet receives a local observation indicating whether its packet was successfully delivered or not (i.e., ACK signal). The objective is to find a multi-user strategy that maximizes a certain network utility in a distributed manner without online coordination or message exchanges between users. Obtaining an optimal solution for the spectrum access problem is computationally expensive in general due to the large state space and partial observability of the states. To tackle this problem, we develop a distributed dynamic spectrum access algorithm based on deep multi-user reinforcement leaning. Specifically, at each time slot, each user maps its current state to spectrum access actions based on a trained deep-Q network used to maximize the objective function. Experimental results have demonstrated that users are capable to learn good policies that achieve strong performance in this challenging partially observable setting only from their ACK signals, without online coordination, message exchanges between users, or carrier sensing. Reinforcement Learning is gaining attention by the wireless networking community due to its potential to learn good-performing configurations only from the observed results. In this work we propose a stateless variation of Q-learning, which we apply to exploit spatial reuse in a wireless network. In particular, we allow networks to modify both their transmission power and the channel used solely based on the experienced throughput. We concentrate in a completely decentralized scenario in which no information about neighbouring nodes is available to the learners. Our results show that although the algorithm is able to find the best-performing actions to enhance aggregate throughput, there is high variability in the throughput experienced by the individual networks. We identify the cause of this variability as the adversarial setting of our setup, in which the most played actions provide intermittent good/poor performance depending on the neighbouring decisions. We also evaluate the effect of the intrinsic learning parameters of the algorithm on this variability. We model the formation of networks as a game where players aspire to maximize their own centrality by increasing the number of other players to which they are path-wise connected, while simultaneously incurring a cost for each added adjacent edge. We simulate the interactions between players using an algorithm that factors in rational strategic behavior based on a common objective function. The resulting networks exhibit pairwise stability, from which we derive necessary stable conditions for specific graph topologies. We then expand the model to simulate non-trivial games with large numbers of players. We show that using conditions necessary for the stability of star topologies we can induce the formation of hub players that positively impact the total welfare of the network. We summarize a theoretical framework based on global time-reparametrization invariance that explains the origin of dynamic fluctuations in glassy systems. We introduce the main ideas without getting into much technical details. We describe a number of consequences arising from this scenario that can be tested numerically and experimentally distinguishing those that can also be explained by other mechanisms from the ones that we believe, are special to our proposal. We support our claims by presenting some numerical checks performed on the 3d Edwards-Anderson spin-glass. Finally, we discuss up to which extent these ideas apply to super-cooled liquids that have been studied in much more detail up to present. In this paper we present a method for learning a discriminative classifier from unlabeled or partially labeled data. Our approach is based on an objective function that trades-off mutual information between observed examples and their predicted categorical class distribution, against robustness of the classifier to an adversarial generative model. The resulting algorithm can either be interpreted as a natural generalization of the generative adversarial networks (GAN) framework or as an extension of the regularized information maximization (RIM) framework to robust classification against an optimal adversary. We empirically evaluate our method - which we dub categorical generative adversarial networks (or CatGAN) - on synthetic data as well as on challenging image classification tasks, demonstrating the robustness of the learned classifiers. We further qualitatively assess the fidelity of samples generated by the adversarial generator that is learned alongside the discriminative classifier, and identify links between the CatGAN objective and discriminative clustering algorithms (such as RIM). For a large multi-hop wireless network, nodes are preferable to make distributed and localized link-scheduling decisions with only interactions among a small number of neighbors. However, for a slowly decaying channel and densely populated interferers, a small size neighborhood often results in nontrivial link outages and is thus insufficient for making optimal scheduling decisions. A question arises how to deal with the information outside a neighborhood in distributed link-scheduling. In this work, we develop joint approximation of information and distributed link scheduling. We first apply machine learning approaches to model distributed link-scheduling with complete information. We then characterize the information outside a neighborhood in form of residual interference as a random loss variable. The loss variable is further characterized by either a Mean Field approximation or a normal distribution based on the Lyapunov central limit theorem. The approximated information outside a neighborhood is incorporated in a factor graph. This results in joint approximation and distributed link-scheduling in an iterative fashion. Link-scheduling decisions are first made at each individual node based on the approximated loss variables. Loss variables are then updated and used for next link-scheduling decisions. The algorithm repeats between these two phases until convergence. Interactive iterations among these variables are implemented with a message-passing algorithm over a factor graph. Simulation results show that using learned information outside a neighborhood jointly with distributed link-scheduling reduces the outage probability close to zero even for a small neighborhood. Study of the production of pairs of top quarks in association with a Higgs boson is one of the primary goals of the Large Hadron Collider over the next decade, as measurements of this process may help us to understand whether the uniquely large mass of the top quark plays a special role in electroweak symmetry breaking. Higgs bosons decay predominantly to \bbbar, yielding signatures for the signal that are similar to $t\bar{t}$ + jets with heavy flavor. Though particularly challenging to study due to the similar kinematics between signal and background events, such final states ($t\bar{t} b \bar{b}$) are an important channel for studying the top quark Yukawa coupling. This paper presents a systematic study of machine learning (ML) methods for detecting $t\bar{t}h$ in the $h \rightarrow b\bar{b}$ decay channel. Among the eight ML methods tested, we show that two models, extreme gradient boosted trees and neural network models, outperform alternative methods. We further study the effectiveness of ML algorithms by investigating the impact of feature set and data size, as well as the structure of the models. While extended feature set and larger training sets expectedly lead to improvement of performance, shallow models deliver comparable or better performance than their deeper counterparts. Our study suggests that ensembles of trees and neurons, not necessarily deep, work effectively for the problem of $t\bar{t}h$ detection. Numerous real-world relations can be represented by signed networks with positive links (e.g., trust) and negative links (e.g., distrust). Link analysis plays a crucial role in understanding the link formation and can advance various tasks in social network analysis such as link prediction. The majority of existing works on link analysis have focused on unsigned social networks. The existence of negative links determines that properties and principles of signed networks are substantially distinct from those of unsigned networks, thus we need dedicated efforts on link analysis in signed social networks. In this paper, following social theories in link analysis in unsigned networks, we adopt three social science theories, namely Emotional Information, Diffusion of Innovations and Individual Personality, to guide the task of link analysis in signed networks. Significant microstructural anisotropy is known to develop during shearing flow of attractive particle suspensions. These suspensions, and their capacity to form conductive networks, play a key role in flow-battery technology, among other applications. Herein, we present and test an analytical model for the tensorial conductivity of attractive particle suspensions. The model utilizes the mean fabric of the network to characterize the structure, and the relationship to the conductivity is inspired by a lattice argument. We test the accuracy of our model against a large number of computer-generated suspension networks, based on multiple in-house generation protocols, giving rise to particle networks that emulate the physical system. The model is shown to adequately capture the tensorial conductivity, both in terms of its invariants and its mean directionality. Measurements of fusion cross-sections of 7Li and 12C with 198Pt at deep sub-barrier energies are reported to unravel the role of the entrance channel in the occurrence of fusion hindrance. The onset of fusion hindrance has been clearly observed in 12C + 198Pt system but not in 7Li + 198Pt system, within the measured energy range. Emergence of the hindrance, moving from lighter (6,7Li) to heavier (12C,16O) projectiles is explained employing a model that considers a gradual transition from a sudden to adiabatic regime at low energies. The model calculation reveals a weak effect of the damping of coupling to collective motion for the present systems as compared to that obtained for systems with heavier projectiles. Research issues and data mining techniques for product recommendation and viral marketing have been widely studied. Existing works on seed selection in social networks do not take into account the effect of product recommendations in e-commerce stores. In this paper, we investigate the seed selection problem for viral marketing that considers both effects of social influence and item inference (for product recommendation). We develop a new model, Social Item Graph (SIG), that captures both effects in form of hyperedges. Accordingly, we formulate a seed selection problem, called Social Item Maximization Problem (SIMP), and prove the hardness of SIMP. We design an efficient algorithm with performance guarantee, called Hyperedge-Aware Greedy (HAG), for SIMP and develop a new index structure, called SIG-index, to accelerate the computation of diffusion process in HAG. Moreover, to construct realistic SIG models for SIMP, we develop a statistical inference based framework to learn the weights of hyperedges from data. Finally, we perform a comprehensive evaluation on our proposals with various baselines. Experimental result validates our ideas and demonstrates the effectiveness and efficiency of the proposed model and algorithms over baselines. The search for Majorana bound states in solid-state physics has been limited to materials which display a gap in their bulk spectrum. We show that such unpaired states appear in certain quasi-one-dimensional Josephson junctions arrays with gapless bulk excitations. The bulk modes mediate a coupling between Majorana bound states via the Ruderman-Kittel-Yosida-Kasuya mechanism. As a consequence, the lowest energy doublet acquires a finite energy difference. For realistic set of parameters this energy splitting remains much smaller than the energy of the bulk eigenstates even for short chains of length $L \sim 10$. We study the Glauber dynamics of Ising spin models with random bonds, on finitely connected random graphs. We generalize a recent dynamical replica theory with which to predict the evolution of the joint spin-field distribution, to include random graphs with arbitrary degree distributions. The theory is applied to Ising ferromagnets on randomly diluted Bethe lattices, where we study the evolution of the magnetization and the internal energy. It predicts a prominent slowing down of the flow in the Griffiths phase, it suggests a further dynamical transition at lower temperatures within the Griffiths phase, and it is verified quantitatively by the results of Monte Carlo simulations. This paper presents measurements of \k\ and \lam\ production in neutral current, deep inelastic scattering of 26.7 GeV electrons and 820 GeV protons in the kinematic range $ 10 < Q^{2} < 640 $ GeV$^2$, $0.0003 < x < 0.01$, and $y > 0.04$. Average multiplicities for \k\ and \lam\ production are determined for transverse momenta \ \ptr\ $> 0.5 $ GeV and pseudorapidities $\left| \eta \right| < 1.3$. The multiplicities favour a stronger strange to light quark suppression in the fragmentation chain than found in $e^+ e^-$ experiments. The production properties of \k's in events with and without a large rapidity gap with respect to the proton direction are compared. The ratio of neutral \k's to charged particles per event in the measured kinematic range is, within the present statistics, the same in both samples. We consider the problem of image representation for the tasks of unsupervised learning and semi-supervised learning. In those learning tasks, the raw image vectors may not provide enough representation for their intrinsic structures due to their highly dense feature space. To overcome this problem, the raw image vectors should be mapped to a proper representation space which can capture the latent structure of the original data and represent the data explicitly for further learning tasks such as clustering. Inspired by the recent research works on deep neural network and representation learning, in this paper, we introduce the multiple-layer auto-encoder into image representation, we also apply the locally invariant ideal to our image representation with auto-encoders and propose a novel method, called Graph regularized Auto-Encoder (GAE). GAE can provide a compact representation which uncovers the hidden semantics and simultaneously respects the intrinsic geometric structure. Extensive experiments on image clustering show encouraging results of the proposed algorithm in comparison to the state-of-the-art algorithms on real-word cases. We propose a method to determine the locally preferred structure of model liquids. This latter is obtained numerically as the global minimum of the effective energy surface of clusters formed by small numbers of particles embedded in a liquid-like environment. The effective energy is the sum of the intra-cluster interaction potential and of an external field that describes the influence of the embedding bulk liquid at a mean-field level. Doing so we minimize the surface effects present in isolated clusters without introducing the full blown geometrical frustration present in bulk condensed phases. We find that the locally preferred structure of the Lennard-Jones liquid is an icosahedron, and that the liquid-like environment only slightly reduces the relative stability of the icosahedral cluster. The influence of the boundary conditions on the nature of the ground-state configuration of Lennard-Jones clusters is also discussed. Multivariate Poisson random variables subject to linear integer constraints arise in several application areas, such as queuing and biomolecular networks. This note shows how to compute conditional statistics in this context, by employing WF Theory and associated algorithms. A symbolic computation package has been developed and is made freely available. A discussion of motivating biomolecular problems is also provided. Continuous neural field models with inhomogeneous synaptic connectivities are known to support traveling fronts as well as stable bumps of localized activity. We analyze stationary localized structures in a neural field model with periodic modulation of the synaptic connectivity kernel and find that they are arranged in a snakes-and-ladders bifurcation structure. In the case of Heaviside firing rates, we construct analytically symmetric and asymmetric states and hence derive closed-form expressions for the corresponding bifurcation diagrams. We show that the ideas proposed by Beck and co-workers to analyze snaking solutions to the Swift-Hohenberg equation remain valid for the neural field model, even though the corresponding spatial-dynamical formulation is non-autonomous. We investigate how the modulation amplitude affects the bifurcation structure and compare numerical calculations for steep sigmoidal firing rates with analytic predictions valid in the Heaviside limit. We investigate spectral correlations in quasi one-dimensional Anderson insulators with broken time-reversal symmetry. While energy levels are uncorrelated in the thermodynamic limit of infinite wire-length, some correlations remain in finite-size Anderson insulators. Asymptotic behaviors of level-level correlations in these systems are known in the large- and small-frequency limits, corresponding to the regime of classical diffusive dynamics and the deep quantum regime of strong Anderson localization. Employing non-perturbative methods and a mapping to the Coulomb-scattering problem, recently introduced by {\it M.~A.~Skvortsov} and {\it P.~M.~Ostrovsky}, we derive a closed analytical expression for the spectral statistics in the classical-to-quantum region bridging the known asymptotic behaviors. We further discuss how Poisson statistics at large energies develop into Wigner-Dyson statistics as the wire-length decreases. We propose a Convolutional Neural Network (CNN) based algorithm - StuffNet - for object detection. In addition to the standard convolutional features trained for region proposal and object detection [31], StuffNet uses convolutional features trained for segmentation of objects and 'stuff' (amorphous categories such as ground and water). Through experiments on Pascal VOC 2010, we show the importance of features learnt from stuff segmentation for improving object detection performance. StuffNet improves performance from 18.8% mAP to 23.9% mAP for small objects. We also devise a method to train StuffNet on datasets that do not have stuff segmentation labels. Through experiments on Pascal VOC 2007 and 2012, we demonstrate the effectiveness of this method and show that StuffNet also significantly improves object detection performance on such datasets. The topology of social networks can be understood as being inherently dynamic, with edges having a distinct position in time. Most characterizations of dynamic networks discretize time by converting temporal information into a sequence of network "snapshots" for further analysis. Here we study a highly resolved data set of a dynamic proximity network of 66 individuals. We show that the topology of this network evolves over a very broad distribution of time scales, that its behavior is characterized by strong periodicities driven by external calendar cycles, and that the conversion of inherently continuous-time data into a sequence of snapshots can produce highly biased estimates of network structure. We suggest that dynamic social networks exhibit a natural time scale \Delta_{nat}, and that the best conversion of such dynamic data to a discrete sequence of networks is done at this natural rate. Existing action detection algorithms usually generate action proposals through an extensive search over the video at multiple temporal scales, which brings about huge computational overhead and deviates from the human perception procedure. We argue that the process of detecting actions should be naturally one of observation and refinement: observe the current window and refine the span of attended window to cover true action regions. In this paper, we propose an active action proposal model that learns to find actions through continuously adjusting the temporal bounds in a self-adaptive way. The whole process can be deemed as an agent, which is firstly placed at a position in the video at random, adopts a sequence of transformations on the current attended region to discover actions according to a learned policy. We utilize reinforcement learning, especially the Deep Q-learning algorithm to learn the agent's decision policy. In addition, we use temporal pooling operation to extract more effective feature representation for the long temporal window, and design a regression network to adjust the position offsets between predicted results and the ground truth. Experiment results on THUMOS 2014 validate the effectiveness of the proposed approach, which can achieve competitive performance with current action detection algorithms via much fewer proposals. Large-scale datasets have driven the rapid development of deep neural networks for visual recognition. However, annotating a massive dataset is expensive and time-consuming. Web images and their labels are, in comparison, much easier to obtain, but direct training on such automatically harvested images can lead to unsatisfactory performance, because the noisy labels of Web images adversely affect the learned recognition models. To address this drawback we propose an end-to-end weakly-supervised deep learning framework which is robust to the label noise in Web images. The proposed framework relies on two unified strategies -- random grouping and attention -- to effectively reduce the negative impact of noisy web image annotations. Specifically, random grouping stacks multiple images into a single training instance and thus increases the labeling accuracy at the instance level. Attention, on the other hand, suppresses the noisy signals from both incorrectly labeled images and less discriminative image regions. By conducting intensive experiments on two challenging datasets, including a newly collected fine-grained dataset with Web images of different car models, the superior performance of the proposed methods over competitive baselines is clearly demonstrated. In this paper, we present a novel deep learning approach, deeply-fused nets. The central idea of our approach is deep fusion, i.e., combine the intermediate representations of base networks, where the fused output serves as the input of the remaining part of each base network, and perform such combinations deeply over several intermediate representations. The resulting deeply fused net enjoys several benefits. First, it is able to learn multi-scale representations as it enjoys the benefits of more base networks, which could form the same fused network, other than the initial group of base networks. Second, in our suggested fused net formed by one deep and one shallow base networks, the flows of the information from the earlier intermediate layer of the deep base network to the output and from the input to the later intermediate layer of the deep base network are both improved. Last, the deep and shallow base networks are jointly learnt and can benefit from each other. More interestingly, the essential depth of a fused net composed from a deep base network and a shallow base network is reduced because the fused net could be composed from a less deep base network, and thus training the fused net is less difficult than training the initial deep base network. Empirical results demonstrate that our approach achieves superior performance over two closely-related methods, ResNet and Highway, and competitive performance compared to the state-of-the-arts. We analyze the statistics of gaps ($\Delta H$) between successive avalanches in one dimensional random field Ising models (RFIMs) in an external field $H$ at zero temperature. In the first part of the paper we study the nearest-neighbour ferromagnetic RFIM. We map the sequence of avalanches in this system to a non-homogeneous Poisson process with an $H$-dependent rate $\rho(H)$. We use this to analytically compute the distribution of gaps $P(\Delta H)$ between avalanches as the field is increased monotonically from $-\infty$ to $+\infty$. We show that $P(\Delta H)$ tends to a constant $\mathcal{C}(R)$ as $\Delta H \to 0^+$, which displays a non-trivial behaviour with the strength of disorder $R$. We verify our predictions with numerical simulations. In the second part of the paper, motivated by avalanche gap distributions in driven disordered amorphous solids, we study a long-range antiferromagnetic RFIM. This model displays a gapped behaviour $P(\Delta H) = 0$ up to a system size dependent offset value $\Delta H_{\text{off}}$, and $P(\Delta H) \sim (\Delta H - \Delta H_{\text{off}})^{\theta}$ as $\Delta H \to H_{\text{off}}^+$. We perform numerical simulations on this model and determine $\theta \approx 0.95(5)$. We also discuss mechanisms which would lead to a non-zero exponent $\theta$ for general spin models with quenched random fields. In this paper we introduce a framework for computing upper bounds yet accurate WCET for hardware platforms with caches and pipelines. The methodology we propose consists of 3 steps: 1) given a program to analyse, compute an equivalent (WCET-wise) abstract program; 2) build a timed game by composing this abstract program with a network of timed automata modeling the architecture; and 3) compute the WCET as the optimal time to reach a winning state in this game. We demonstrate the applicability of our framework on standard benchmarks for an ARM9 processor with instruction and data caches, and compute the WCET with UPPAAL-TiGA. We also show that this framework can easily be extended to take into account dynamic changes in the speed of the processor during program execution. % Softmax GAN is a novel variant of Generative Adversarial Network (GAN). The key idea of Softmax GAN is to replace the classification loss in the original GAN with a softmax cross-entropy loss in the sample space of one single batch. In the adversarial learning of $N$ real training samples and $M$ generated samples, the target of discriminator training is to distribute all the probability mass to the real samples, each with probability $\frac{1}{M}$, and distribute zero probability to generated data. In the generator training phase, the target is to assign equal probability to all data points in the batch, each with probability $\frac{1}{M+N}$. While the original GAN is closely related to Noise Contrastive Estimation (NCE), we show that Softmax GAN is the Importance Sampling version of GAN. We futher demonstrate with experiments that this simple change stabilizes GAN training. In this paper, we study instances of complex neural networks, i.e. neural netwo rks with complex topologies. We use Self-Organizing Map neural networks whose n eighbourhood relationships are defined by a complex network, to classify handwr itten digits. We show that topology has a small impact on performance and robus tness to neuron failures, at least at long learning times. Performance may howe ver be increased (by almost 10%) by artificial evolution of the network topo logy. In our experimental conditions, the evolved networks are more random than their parents, but display a more heterogeneous degree distribution. Repeating patterns of spike sequences from a neuronal network have been proposed to be useful in the reconstruction of the network topology. Reverberations in a physiologically realistic model with various physical connection topologies (from random to scale-free) have been simulated to study the effectiveness of the pattern-matching method in the reconstruction of network topology from network dynamics. Simulation results show that functional networks reconstructed from repeating spike patterns can be quite different from the original physical networks; even global properties, such as the degree distribution, cannot always be recovered. However, the pattern-matching method can be effective in identifying hubs in the network. Since the form of reverberations are quite different for networks with and without hubs, the form of reverberations together with the reconstruction by repeating spike patterns might provide a reliable method to detect hubs in neuronal cultures. We simulated the creep motions of flux lines subject to randomly distributed point-like pinning centers. It is found that at low temperatures, the pinning barrier $U$ defined in the Arrhenius-type $v-F$ characteristics increases with decreasing force $U(F) \propto F^{-\mu}$, as predicted by previous theories. The exponent $\mu$ is evaluated as $0.28\pm 0.02 $ for the vortex glass and $\mu\simeq 0.5\pm 0.02$ for the Bragg glass (BrG). The latter is in good agreement with the prediction by the scaling theory and the functional-renormalization-group theory on creep, while the former is a new estimate. Within BrG, we find that the pinning barrier is suppressed when temperature is lifted to approximately half of the melting temperature. Characterizations of this new transition at equilibrium are also presented, indicative of a phase transition associated with the replica-symmetry breaking. We present the first public release of our generic neural network training algorithm, called SkyNet. This efficient and robust machine learning tool is able to train large and deep feed-forward neural networks, including autoencoders, for use in a wide range of supervised and unsupervised learning applications, such as regression, classification, density estimation, clustering and dimensionality reduction. SkyNet uses a `pre-training' method to obtain a set of network parameters that has empirically been shown to be close to a good solution, followed by further optimisation using a regularised variant of Newton's method, where the level of regularisation is determined and adjusted automatically; the latter uses second-order derivative information to improve convergence, but without the need to evaluate or store the full Hessian matrix, by using a fast approximate method to calculate Hessian-vector products. This combination of methods allows for the training of complicated networks that are difficult to optimise using standard backpropagation techniques. SkyNet employs convergence criteria that naturally prevent overfitting, and also includes a fast algorithm for estimating the accuracy of network outputs. The utility and flexibility of SkyNet are demonstrated by application to a number of toy problems, and to astronomical problems focusing on the recovery of structure from blurred and noisy images, the identification of gamma-ray bursters, and the compression and denoising of galaxy images. The SkyNet software, which is implemented in standard ANSI C and fully parallelised using MPI, is available at http://www.mrao.cam.ac.uk/software/skynet/. We show that the stochastic field theory for directed percolation in presence of an additional conservation law (the C-DP class) can be mapped exactly to the continuum theory for the depinning of an elastic interface in short-range correlated quenched disorder. On one line of parameters commonly studied, this mapping leads to the simplest overdamped dynamics. Away from this line, an additional memory term arises in the interface dynamics; we argue that it does not change the universality class. Since C-DP is believed to describe the Manna class of self-organized criticality, this shows that Manna stochastic sandpiles and disordered elastic interfaces (i.e. the quenched Edwards-Wilkinson model) share the same universal large-scale behavior. A field-theory approach is used to investigate the ''spin-glass effects'' on the critical behaviour of systems with weak temperature-like quenched disorder. The renormalization group (RG) analysis of the effective Hamiltonian of a model with replica symmetry breaking (RSB) potentials of a general type is carried out in the two-loop approximation. The fixed-point (FP) stability, recently found within the one-step RSB RG treatment, is further explored in terms of replicon eigenvalues. We find that the traditional FPs, which are usually considered to describe the disorder-induced universal critical behaviour, remain stable when the continuous RSB modes are taken into account. The wide development of inter connectivity of cellular networks with the internet network has made them to be vulnerable. This exposure of the cellular networks to internet has increased threats to customer end equipment as well as the carrier infrastructure. The artistic style of a painting is a subtle aesthetic judgment used by art historians for grouping and classifying artwork. The recently introduced `neural-style' algorithm substantially succeeds in merging the perceived artistic style of one image or set of images with the perceived content of another. In light of this and other recent developments in image analysis via convolutional neural networks, we investigate the effectiveness of a `neural-style' representation for classifying the artistic style of paintings. A visualisation tool is presented to facilitate the study on large-scale communications networks. This tool provides a simple and effective way to summarise the topology of a complex network at a coarse level. The challenging requirements of 5G--from both the applications and the architecture perspectives--motivate the need to explore the feasibility of delivering services over new network architectures. As 5G proposes application-centric network slicing, which enables the use of new data planes realizable over a programmable compute, storage, and transport infrastructure, we consider Information-centric Networking (ICN) as a candidate network architecture to realize 5G objectives. This can co-exist with end-to-end IP services that are offered today. To this effect, we first propose a 5G-ICN architecture and compare its benefits (i.e., innovative services offered by leveraging ICN features) to current 3GPP-based mobile architectures. We then introduce a general application-driven framework that emphasizes on the flexibility afforded by Network Function Virtualization (NFV) and Software Defined Networking (SDN) over which 5G-ICN can be realized. We specifically focus on the issue of how mobility-as-a-service (MaaS) can be realized as a 5G-ICN slice, and give an in-depth overview on resource provisioning and inter-dependencies and -coordinations among functional 5G-ICN slices to meet the MaaS objectives. Mobile Ad-hoc Network (MANET) is a collection of autonomous nodes or terminals which communicate with each other by forming a multi-hop radio network and maintaining connectivity in a decentralized manner. The conventional security solutions to provide key management through accessing trusted authorities or centralized servers are infeasible for this new environment since mobile ad hoc networks are characterized by the absence of any infrastructure, frequent mobility, and wireless links. We propose a hierarchical group key management scheme that is hierarchical and fully distributed with no central authority and uses a simple rekeying procedure which is suitable for large and high mobility mobile ad hoc networks. The rekeying procedure requires only one round in our scheme and Chinese Remainder Theorem Diffie Hellman Group Diffie Hellmann and Burmester and Desmedt it is a constant 3 whereas in other schemes such as Distributed Logical Key Hierarchy and Distributed One Way Function Trees, it depends on the number of members. We reduce the energy consumption during communication of the keying materials by reducing the number of bits in the rekeying message. We show through analysis and simulations that our scheme has less computation, communication and energy consumption compared to the existing schemes. We present two Bayesian procedures to infer the interactions and external currents in an assembly of stochastic integrate-and-fire neurons from the recording of their spiking activity. The first procedure is based on the exact calculation of the most likely time courses of the neuron membrane potentials conditioned by the recorded spikes, and is exact for a vanishing noise variance and for an instantaneous synaptic integration. The second procedure takes into account the presence of fluctuations around the most likely time courses of the potentials, and can deal with moderate noise levels. The running time of both procedures is proportional to the number S of spikes multiplied by the squared number N of neurons. The algorithms are validated on synthetic data generated by networks with known couplings and currents. We also reanalyze previously published recordings of the activity of the salamander retina (including from 32 to 40 neurons, and from 65,000 to 170,000 spikes). We study the dependence of the inferred interactions on the membrane leaking time; the differences and similarities with the classical cross-correlation analysis are discussed. Phishing is an increasingly sophisticated method to steal personal user information using sites that pretend to be legitimate. In this paper, we take the following steps to identify phishing URLs. First, we carefully select lexical features of the URLs that are resistant to obfuscation techniques used by attackers. Second, we evaluate the classification accuracy when using only lexical features, both automatically and hand-selected, vs. when using additional features. We show that lexical features are sufficient for all practical purposes. Third, we thoroughly compare several classification algorithms, and we propose to use an online method (AROW) that is able to overcome noisy training data. Based on the insights gained from our analysis, we propose PhishDef, a phishing detection system that uses only URL names and combines the above three elements. PhishDef is a highly accurate method (when compared to state-of-the-art approaches over real datasets), lightweight (thus appropriate for online and client-side deployment), proactive (based on online classification rather than blacklists), and resilient to training data inaccuracies (thus enabling the use of large noisy training data). We propose a simple preferential attachment model of growing network using the complementary probability of Barab\'asi-Albert (BA) model, i.e., $\Pi(k_i) \propto 1-\frac{k_i}{\sum_j k_j}$. In this network, new nodes are preferentially attached to not well connected nodes. Numerical simulations, in perfect agreement with the master equation solution, give an exponential degree distribution. This suggests that the power law degree distribution is a consequence of preferential attachment probability together with "rich get richer" phenomena. We also calculate the average degree of a target node at time t $()$ and its fluctuations, to have a better view of the microscopic evolution of the network, and we also compare the results with BA model. Survivable design of cross-layer networks, such as the cloud computing infrastructure, lies in its resource deployment and allocation and mapping of the logical (virtual datacenter/IP) network into the physical infrastructure (cloud backbone/WDM) such that link or node failure(s) in the physical infrastructure would not result in cascading failures in the logical network. Most of the prior approaches for survivable cross-layer network design aim at single-link failure scenario, which are not applicable to the more challenging multi-failure scenarios. Also, as many of these approaches use the cross-layer cut concept, enumeration of all cuts in the network is required and thus introducing exponential number of constraints. To overcome these difficulties, we investigate in this paper survivable mapping approaches against multiple physical link failures and its special case, Shared Risk Link Group (SRLG) failure. We present the necessary and sufficient conditions based on both cross-layer spanning trees and cutsets to guarantee a survivable mapping when multiple physical link failures occur. Based on the necessary and sufficient conditions, we propose to solve the problem through (1) mixed-integer linear programs which avoid enumerating all combinations of link failures, and (2) an algorithm which generates/adds logical spanning trees sequentially. Our simulation results show that the proposed approaches can produce survivable mappings effectively against both $k$- and SRLG-failures. In this paper, we present an effective method to analyze the recognition confidence of handwritten Chinese character, based on the softmax regression score of a high performance convolutional neural networks (CNN). Through careful and thorough statistics of 827,685 testing samples that randomly selected from total 8836 different classes of Chinese characters, we find that the confidence measurement based on CNN is an useful metric to know how reliable the recognition results are. Furthermore, we find by experiments that the recognition confidence can be used to find out similar and confusable character-pairs, to check wrongly or cursively written samples, and even to discover and correct mis-labelled samples. Many interesting observations and statistics are given and analyzed in this study. An important problem in physics concerns the analysis of audio time series generated by transduced acoustic phenomena. Here, we develop a new method to quantify the scaling properties of the local variance of nonstationary time series. We apply this technique to analyze audio signals obtained from selected genres of music. We find quantitative differences in the correlation properties of high art music, popular music, and dance music. We discuss the relevance of these objective findings in relation to the subjective experience of music. Deep neural networks (DNN) have revolutionized the field of natural language processing (NLP). Convolutional neural network (CNN) and recurrent neural network (RNN), the two main types of DNN architectures, are widely explored to handle various NLP tasks. CNN is supposed to be good at extracting position-invariant features and RNN at modeling units in sequence. The state of the art on many NLP tasks often switches due to the battle between CNNs and RNNs. This work is the first systematic comparison of CNN and RNN on a wide range of representative NLP tasks, aiming to give basic guidance for DNN selection. An important goal for the machine learning (ML) community is to create approaches that can learn solutions with human-level capability. One domain where humans have held a significant advantage is visual processing. A significant approach to addressing this gap has been machine learning approaches that are inspired from the natural systems, such as artificial neural networks (ANNs), evolutionary computation (EC), and generative and developmental systems (GDS). Research into deep learning has demonstrated that such architectures can achieve performance competitive with humans on some visual tasks; however, these systems have been primarily trained through supervised and unsupervised learning algorithms. Alternatively, research is showing that evolution may have a significant role in the development of visual systems. Thus this paper investigates the role neuro-evolution (NE) can take in deep learning. In particular, the Hypercube-based NeuroEvolution of Augmenting Topologies is a NE approach that can effectively learn large neural structures by training an indirect encoding that compresses the ANN weight pattern as a function of geometry. The results show that HyperNEAT struggles with performing image classification by itself, but can be effective in training a feature extractor that other ML approaches can learn from. Thus NeuroEvolution combined with other ML methods provides an intriguing area of research that can replicate the processes in nature. Incidents of organized cybercrime are rising because of criminals are reaping high financial rewards while incurring low costs to commit crime. As the digital landscape broadens to accommodate more internet-enabled devices and technologies like social media, more cybercriminals who are not native English speakers are invading cyberspace to cash in on quick exploits. In this paper we evaluate the performance of three machine learning classifiers in detecting 419 scams in a bilingual Nigerian cybercriminal community. We use three popular classifiers in text processing namely: Na\"ive Bayes, k-nearest neighbors (IBK) and Support Vector Machines (SVM). The preliminary results on a real world dataset reveal the SVM significantly outperforms Na\"ive Bayes and IBK at 95% confidence level. Deep convolutional neural networks (CNN) have shown their good performances in many computer vision tasks. However, the high computational complexity of CNN involves a huge amount of data movements between the computational processor core and memory hierarchy which occupies the major of the power consumption. This paper presents Chain-NN, a novel energy-efficient 1D chain architecture for accelerating deep CNNs. Chain-NN consists of the dedicated dual-channel process engines (PE). In Chain-NN, convolutions are done by the 1D systolic primitives composed of a group of adjacent PEs. These systolic primitives, together with the proposed column-wise scan input pattern, can fully reuse input operand to reduce the memory bandwidth requirement for energy saving. Moreover, the 1D chain architecture allows the systolic primitives to be easily reconfigured according to specific CNN parameters with fewer design complexity. The synthesis and layout of Chain-NN is under TSMC 28nm process. It costs 3751k logic gates and 352KB on-chip memory. The results show a 576-PE Chain-NN can be scaled up to 700MHz. This achieves a peak throughput of 806.4GOPS with 567.5mW and is able to accelerate the five convolutional layers in AlexNet at a frame rate of 326.2fps. 1421.0GOPS/W power efficiency is at least 2.5 to 4.1x times better than the state-of-the-art works. One of the important issues in wireless networks is the Routing problem that is effective on system performance, in this article the attempt is made to propose a routing algorithm using the bee colony in order to reduce energy consumption in wireless relay networks. In EBCD algorithm, through combined of energy, distance and traffic parameters a routing algorithm for wireless networks is presented with more efficiency than its predecessor. Applying the bee colony method would allow the placement of the parameters under conventional conditions and to get closer to a mechanism with a better adaptability than that of the existing algorithm. According to the parameters considered, the proposed algorithm provides a fitness function that can be applied as a multi-hop. Unlike other algorithms of its kind this can increase service quality based on environmental conditions through its multiple services. This new method can store the energy accumulated in the nodes and reduce the hop restrictions. A favorable environment for downbursts associated with deep convective storm systems that occur over the central and eastern continental United States includes strong static instability with large amounts of convective available potential energy and the presence of a mid-tropospheric layer of dry air. However, previous research has identified that over the central United States, especially in the Great Plains region, an environment between that favorable for wet and dry microbursts may exist during the convective season, resulting in the generation of hybrid type microbursts. Hybrid microbursts have been found to originate from deep convective storms that generate heavy precipitation, with sub-cloud evaporation of precipitation a significant factor in downdraft acceleration. Accordingly, a new GOES sounder derived product, the GOES Hybrid Microburst Index, is under development and is designed to assess the potential for convective downbursts that develop in an intermediate environment between a wet type, associated with heavy precipitation, and a dry type associated with convection in which very little to no precipitation is observed at the surface. Electrical analogues of fracture, such as the fuse network model, are widely studied. However, the "analogy" between the electrical problem and the elastic problem is rarely established explicitly. Further, the fuse network is a discrete approximation to the continuous problem of fracture. It is rarely, if ever, shown that the discrete approximation indeed approaches its continuum limit. We establish both of these correspondences directly. We present results from a study of the photometric redshift performance of the Dark Energy Survey (DES), using the early data from a Science Verification (SV) period of observations in late 2012 and early 2013 that provided science-quality images for almost 200 sq.~deg.~at the nominal depth of the survey. We assess the photometric redshift performance using about 15000 galaxies with spectroscopic redshifts available from other surveys. These galaxies are used, in different configurations, as a calibration sample, and photo-$z$'s are obtained and studied using most of the existing photo-$z$ codes. A weighting method in a multi-dimensional color-magnitude space is applied to the spectroscopic sample in order to evaluate the photo-$z$ performance with sets that mimic the full DES photometric sample, which is on average significantly deeper than the calibration sample due to the limited depth of spectroscopic surveys. Empirical photo-$z$ methods using, for instance, Artificial Neural Networks or Random Forests, yield the best performance in the tests, achieving core photo-$z$ resolutions $\sigma_{68} \sim 0.08$. Moreover, the results from most of the codes, including template fitting methods, comfortably meet the DES requirements on photo-$z$ performance, therefore, providing an excellent precedent for future DES data sets. A neural network technique is used to discriminate between quark and gluon jets produced in the qg->q+photon and q q->g+photon processes at the LHC. Considering the network as a trigger and using the PYTHIA event generator and the full event fast simulation package for the CMS detector CMSJET we obtain signal-to-background ratios. A wide range of applications require or can benefit from collaborative behavior of a group of agents. The technical challenge addressed in this chapter is the development of a decentralized control strategy that enables each agent to independently navigate to ensure agents achieve a collective goal while maintaining network connectivity. Specifically, cooperative controllers are developed for networked agents with limited sensing and network connectivity constraints. By modeling the interaction among the agents as a graph, several different approaches to address the problems of preserving network connectivity are presented, with the focus on a method that utilizes navigation function frameworks. By modeling network connectivity constraints as artificial obstacles in navigation functions, a decentralized control strategy is presented in two particular applications, formation control and rendezvous for a system of autonomous agents, which ensures global convergence to the unique minimum of the potential field (i.e., desired formation or desired destination) while preserving network connectivity. Simulation results are provided to demonstrate the developed strategy. We present a search for eclipses of $\sim$1700 white dwarfs in the Pan-STARRS1 medium-deep fields. Candidate eclipse events are selected by identifying low outliers in over 4.3 million light curve measurements. We find no short-duration eclipses consistent with being caused by a planetary size companion. This large dataset enables us to place strong constraints on the close-in planet occurrence rates around white dwarfs for planets as small as 2 R$_\oplus$. Our results indicate that gas giant planets orbiting just outside the Roche limit are rare, occurring around less than 0.5% of white dwarfs. Habitable-zone super-Earths and hot super-Earths are less abundant than similar classes of planets around main-sequence stars. These constraints give important insight into the ultimate fate of the large population of exoplanets orbiting main sequence stars. Spatial evolutionary games are studied with myopic players whose payoff interest, as a personal character, is tuned from selfishness to other-regarding preference via fraternity. The players are located on a square lattice and collect income from symmetric two-person two-strategy (called cooperation and defection) games with their nearest neighbors. During the elementary steps of evolution a randomly chosen player modifies her strategy in order to maximize stochastically her utility function composed from her own and the co-players' income with weight factors $1-Q$ and Q. These models are studied within a wide range of payoff parameters using Monte Carlo simulations for noisy strategy updates and by spatial stability analysis in the low noise limit. For fraternal players ($Q=1/2$) the system evolves into ordered arrangements of strategies in the low noise limit in a way providing optimum payoff for the whole society. Dominance of defectors, representing the "tragedy of the commons", is found within the regions of prisoner's dilemma and stag hunt game for selfish players (Q=0). Due to the symmetry in the effective utility function the system exhibits similar behavior even for Q=1 that can be interpreted as the "lovers' dilemma". In this paper we consider two sequence tagging tasks for medieval Latin: part-of-speech tagging and lemmatization. These are both basic, yet foundational preprocessing steps in applications such as text re-use detection. Nevertheless, they are generally complicated by the considerable orthographic variation which is typical of medieval Latin. In Digital Classics, these tasks are traditionally solved in a (i) cascaded and (ii) lexicon-dependent fashion. For example, a lexicon is used to generate all the potential lemma-tag pairs for a token, and next, a context-aware PoS-tagger is used to select the most appropriate tag-lemma pair. Apart from the problems with out-of-lexicon items, error percolation is a major downside of such approaches. In this paper we explore the possibility to elegantly solve these tasks using a single, integrated approach. For this, we make use of a layered neural network architecture from the field of deep representation learning. We discuss transport on load bearing branching hierarchical networks which can model diverse systems which can serve as models of river networks, computer networks, respiratory networks and granular media. We study avalanche transmissions and directed percolation on these networks, and on the V lattice, i.e., the strongest realization of the lattice. We find that typical realizations of the lattice show multimodal distributions for the avalanche transmissions, and a second order transition for directed percolation. On the other hand, the V lattice shows power - law behavior for avalanche transmissions, and a first order (explosive) transition to percolation. The V lattice is thus the critical case of hierarchical networks. We note that small perturbations to the V lattice destroy the power-law behavior of the distributions, and the first order nature of the percolation. We discuss the implications of our results. Data aggregation in intermediate nodes (called aggregator nodes) is an effective approach for optimizing consumption of scarce resources like bandwidth and energy in Wireless Sensor Networks (WSNs). However, in-network processing poses a problem for the privacy of the sensor data since individual data of sensor nodes need to be known to the aggregator node before the aggregation process can be carried out. In applications of WSNs, privacy-preserving data aggregation has become an important requirement due to sensitive nature of the sensor data. Researchers have proposed a number of protocols and schemes for this purpose. He et al. (INFOCOM 2007) have proposed a protocol - called CPDA - for carrying out additive data aggregation in a privacy-preserving manner for application in WSNs. The scheme has been quite popular and well-known. In spite of the popularity of this protocol, it has been found that the protocol is vulnerable to attack and it is also not energy-efficient. In this paper, we first present a brief state of the art survey on the current privacy-preserving data aggregation protocols for WSNS. Then we describe the CPDA protocol and identify its security vulnerability. Finally, we demonstrate how the protocol can be made secure and energy efficient. The force networks of different granular ensembles are defined and their topological properties studied using the tools of complex networks. In particular, for each set of grains compressed in a square box, it is introduced a force threshold that determines which contacts conform the network. Hence, the topological characteristics of the network are analyzed as a function of this parameter. The characterization of the structural features thus obtained, may be useful in the understanding of the macroscopic physical behavior exhibited by this class of media. The growth in data traffic and the increased demand for quality of service had generated a large demand for network systems to be more efficient. The introduction of improved routing systems to meet the increasing demand and varied protocols to accommodate various scales of challenges in network efficiency had further complicated the operations. This means that a better mode of intelligence has to be infused into networking for smoother operations and better autonomic features. Cognitive networks are defined and analyzed in this angle. They are identified to have the potential to deal with the future user related quality and efficiency of service at optimized levels. The cognitive elements of a system like perception, learning, planning, reasoning and decision forming can enable the systems to be more aware of their environment and offer better services. These approaches are expected to transform the mode of operation of future networks. We study graded response attractor neural networks with asymmetrically extremely dilute interactions and Langevin dynamics. We solve our model in the thermodynamic limit using generating functional analysis, and find (in contrast to the binary neurons case) that even in statics one cannot eliminate the non-persistent order parameters. The macroscopic dynamics is driven by the (non-trivial) joint distribution of neurons and fields, rather than just the (Gaussian) field distribution. We calculate phase transition lines and present simulation results in support of our theory. Stimulus from the environment that guides behavior and informs decisions is encoded in the firing rates of neural populations. Each neuron in the populations, however, does not spike independently: spike events are correlated from cell to cell. To what degree does this apparent redundancy impact the accuracy with which decisions can be made, and the computations that are required to optimally decide? We explore these questions for two illustrative models of correlation among cells. Each model is statistically identical at the level of pairs cells, but differs in higher-order statistics that describe the simultaneous activity of larger cell groups. We find that the presence of correlations can diminish the performance attained by an ideal decision maker to either a small or large extent, depending on the nature of the higher-order interactions. Moreover, while this optimal performance can in some cases be obtained via the standard integration-to-bound operation, in others it requires a nonlinear computation on incoming spikes. Overall, we conclude that a given level of pairwise correlations--even when restricted to identical neural populations--may not always indicate redundancies that diminish decision making performance. Hard exclusive production in deep inelastic lepton scattering provides access to the unknown Generalized Parton Distributions (GPDs) of the nucleon. At HERMES, different observables for hard exclusive pi^+ production have been measured with a 27.6 GeV positron beam on an internal hydrogen gas target. First preliminary results for the unpolarized ep->enpi^+ total cross section for 1.5