Dataset Viewer
Auto-converted to Parquet
text
stringlengths
8
3.91k
label
int64
0
10
abstract the study of combinatorial optimization problems with a submodular objective has attracted much attention in recent years. such problems are important in both theory and practice because their objective functions are very general. obtaining further improvements for many submodular maximization problems boils down to finding better algorithms for optimizing a relaxation of them known as the multilinear extension. in this work we present an algorithm for optimizing the multilinear relaxation whose guarantee improves over the guarantee of the best previous algorithm (which was given by ene and nguyen (2016)). moreover, our algorithm is based on a new technique which is, arguably, simpler and more natural for the problem at hand. in a nutshell, previous algorithms for this problem rely on symmetry properties which are natural only in the absence of a constraint. our technique avoids the need to resort to such properties, and thus, seems to be a better fit for constrained problems.
8
abstract numerous variants of self-organizing maps (soms) have been proposed in the literature, including those which also possess an underlying structure, and in some cases, this structure itself can be defined by the user although the concepts of growing the som and updating it have been studied, the whole issue of using a self-organizing adaptive data structure (ads) to further enhance the properties of the underlying som, has been unexplored. in an earlier work, we impose an arbitrary, user-defined, tree-like topology onto the codebooks, which consequently enforced a neighborhood phenomenon and the so-called tree-based bubble of activity (boa). in this paper, we consider how the underlying tree itself can be rendered dynamic and adaptively transformed. to do this, we present methods by which a som with an underlying binary search tree (bst) structure can be adaptively re-structured using conditional rotations (conrot). these rotations on the nodes of the tree are local, can be done in constant time, and performed so as to decrease the weighted path length (wpl) of the entire tree. in doing this, we introduce the pioneering concept referred to as neural promotion, where neurons gain prominence in the neural network (nn) as their significance increases. we are not aware of any research which deals with the issue of neural promotion. the advantages of such a scheme is that the user need not be aware of any of the topological peculiarities of the stochastic data distribution. rather, the algorithm, referred to as the ttosom with conditional rotations (ttoconrot), converges in such a manner that the neurons are ultimately placed in the input space so as to represent its stochastic distribution, and additionally, the neighborhood properties of the neurons suit the best bst that represents the data. these properties have been confirmed by our experimental results on a variety of data sets. we submit that all of these concepts are both novel and of a pioneering sort.
9
abstract— we consider the problem of steering a system with unknown, stochastic dynamics to satisfy a rich, temporallylayered task given as a signal temporal logic formula. we represent the system as a markov decision process in which the states are built from a partition of the statespace and the transition probabilities are unknown. we present provably convergent reinforcement learning algorithms to maximize the probability of satisfying a given formula and to maximize the average expected robustness, i.e., a measure of how strongly the formula is satisfied. we demonstrate via a pair of robot navigation simulation case studies that reinforcement learning with robustness maximization performs better than probability maximization in terms of both probability of satisfaction and expected robustness.
3
abstract. we present high performance implementations of the qr and the singular value decomposition of a batch of small matrices hosted on the gpu with applications in the compression of hierarchical matrices. the one-sided jacobi algorithm is used for its simplicity and inherent parallelism as a building block for the svd of low rank blocks using randomized methods. we implement multiple kernels based on the level of the gpu memory hierarchy in which the matrices can reside and show substantial speedups against streamed cusolver svds. the resulting batched routine is a key component of hierarchical matrix compression, opening up opportunities to perform h-matrix arithmetic efficiently on gpus.
8
abstract skew bridges are common in highways and railway lines when non perpendicular crossings are encountered. the structural effect of skewness is an additional torsion on the bridge deck which may have a considerable effect, making its analysis and design more complex. in this paper, an analytical model following 3d beam theory is firstly derived in order to evaluate the dynamic response of skew bridges under moving loads. following, a simplified 2d model is also considered which includes only vertical beam bending. the natural frequencies, eigenmodes and orthogonality relationships are determined from the boundary conditions. the dynamic response is determined in time domain by using the “exact” integration. both models are validated through some numerical examples by comparing with the results obtained by 3d fe models. a parametric study is performed with the simplified model in order to identify parameters that significantly influence the vertical dynamic response of the skew bridge under traffic loads. the results show that the grade of skewness has an important influence on the vertical displacement, but hardly on the vertical acceleration of the bridge. the torsional stiffness really has effect on the vertical displacement when the skew angle is large. the span length reduces the skewness effect on the dynamic behavior of the skew bridge. keywords: skew bridge, bridge modelling, modal analysis, moving load
5
abstract in recent years crowdsourcing has become the method of choice for gathering labeled training data for learning algorithms. standard approaches to crowdsourcing view the process of acquiring labeled data separately from the process of learning a classifier from the gathered data. this can give rise to computational and statistical challenges. for example, in most cases there are no known computationally efficient learning algorithms that are robust to the high level of noise that exists in crowdsourced data, and efforts to eliminate noise through voting often require a large number of queries per example. in this paper, we show how by interleaving the process of labeling and learning, we can attain computational efficiency with much less overhead in the labeling cost. in particular, we consider the realizable setting where there exists a true target function in f and consider a pool of labelers. when a noticeable fraction of the labelers are perfect, and the rest behave arbitrarily, we show that any f that can be efficiently learned in the traditional realizable pac model can be learned in a computationally efficient manner by querying the crowd, despite high amounts of noise in the responses. moreover, we show that this can be done while each labeler only labels a constant number of examples and the number of labels requested per example, on average, is a constant. when no perfect labelers exist, a related task is to find a set of the labelers which are good but not perfect. we show that we can identify all good labelers, when at least the majority of labelers are good.
8
abstract a novel method to identify trampoline skills using a single video camera is proposed herein. conventional computer vision techniques are used for identification, estimation, and tracking of the gymnast’s body in a video recording of the routine. for each frame, an open source convolutional neural network is used to estimate the pose of the athlete’s body. body orientation and joint angle estimates are extracted from these pose estimates. the trajectories of these angle estimates over time are compared with those of labelled reference skills. a nearest neighbour classifier utilising a mean squared error distance metric is used to identify the skill performed. a dataset containing 714 skill examples with 20 distinct skills performed by adult male and female gymnasts was recorded and used for evaluation of the system. the system was found to achieve a skill identification accuracy of 80.7% for the dataset.
1
abstract machines involved: the net m of df a’s ms , ma , . . . with state set q = { q, . . . } (states are drawn as circular nodes); the pilot df a p with m-state set r = { i0 , i1 , . . . } (states are drawn as rectangular nodes); and the dp da a to be next defined. as said, the dp da stores in the stack the series of m-states entered during the computation, enriched with additional information used in the parsing steps. moreover, the m-states are interleaved with terminal or nonterminal grammar symbols. the current m-state, i.e., the one on top of stack, determines the next move: either a shift that scans the next token, or a reduction of a topmost stack segment (also called reduction handle) to a nonterminal identified by a final candidate included in the current m-state. the absence of shift-reduce conflicts makes the choice between shift and reduction operations deterministic. similarly, the absence of reduce-reduce conflicts allows the parser to uniquely identify the final state of a machine. however, this leaves open the problem to determine the stack segment to be reduced. for that two designs will be presented: the first uses a finite pushdown alphabet; the second uses unbounded integer pointers and, strictly speaking, no longer qualifies as a pushdown automaton. first, we specify the pushdown stack alphabet. since for a given net m there are finitely many different candidates, the number of m-states is bounded and the number of candidates in any m-state is also bounded by cmax = |q| × ( |σ| + 1 ). the dp da stack elements
6
abstract. we present an i/o-efficient algorithm for computing similarity joins based on locality-sensitive hashing (lsh). in contrast to the filtering methods commonly suggested our method has provable subquadratic dependency on the data size. further, in contrast to straightforward implementations of known lsh-based algorithms on external memory, our approach is able to take significant advantage of the available internal memory: whereas the time complexity of classical algorithms includes a factor of n ρ , where ρ is a parameter of the lsh used, the i/o complexity of our algorithm merely includes a factor (n/m )ρ , where n is the data size and m is the size of internal memory. our algorithm is randomized and outputs the correct result with high probability. it is a simple, recursive, cache-oblivious procedure, and we believe that it will be useful also in other computational settings such as parallel computation.
8
abstract. let a be a commutative ring, and let a be a weakly proregular ideal in a. (if a is noetherian then any ideal in it is weakly proregular.) suppose m is a compact generator of the category of cohomologically a-torsion complexes. we prove that the derived double centralizer of m is isomorphic to the a-adic completion of a. the proof relies on the mgm equivalence from [psy] and on derived morita equivalence. our result extends earlier work of dwyer-greenlees-iyengar [dgi] and efimov [ef].
0
abstract though traditional algorithms could be embedded into neural architectures with the proposed principle of (xiao, 2017), the variables that only occur in the condition of branch could not be updated as a special case. to tackle this issue, we multiply the conditioned branches with dirac symbol (i.e. 1x>0 ), then approximate dirac symbol with the continuous functions (e.g. 1 − e−α|x| ). in this way, the gradients of condition-specific variables could be worked out in the back-propagation process, approximately, making a fully functioned neural graph. within our novel principle, we propose the neural decision tree (ndt), which takes simplified neural networks as decision function in each branch and employs complex neural networks to generate the output in each leaf. extensive experiments verify our theoretical analysis and demonstrate the effectiveness of our model.
9
abstract adams’ extension of parsing expression grammars enables specifying indentation sensitivity using two non-standard grammar constructs — indentation by a binary relation and alignment. this paper proposes a step-by-step transformation of well-formed adams’ grammars for elimination of the alignment construct from the grammar. the idea that alignment could be avoided was suggested by adams but no process for achieving this aim has been described before. 1998 acm subject classification d.3.1 formal definitions and theory; d.3.4 processors; f.4.2 grammars and other rewriting systems keywords and phrases parsing expression grammars, indentation, grammar transformation
6
abstraction.
6
abstract the coalitional manipulation problem has been studied extensively in the literature for many voting rules. however, most studies have focused on the complete information setting, wherein the manipulators know the votes of the non-manipulators. while this assumption is reasonable for purposes of showing intractability, it is unrealistic for algorithmic considerations. in most real-world scenarios, it is impractical to assume that the manipulators to have accurate knowledge of all the other votes. in this work, we investigate manipulation with incomplete information. in our framework, the manipulators know a partial order for each voter that is consistent with the true preference of that voter. in this setting, we formulate three natural computational notions of manipulation, namely weak, opportunistic, and strong manipulation. we say that an extension of a partial order is viable if there exists a manipulative vote for that extension. we propose the following notions of manipulation when manipulators have incomplete information about the votes of other voters. 1. w eak m anipulation: the manipulators seek to vote in a way that makes their preferred candidate win in at least one extension of the partial votes of the non-manipulators. 2. o pportunistic m anipulation: the manipulators seek to vote in a way that makes their preferred candidate win in every viable extension of the partial votes of the nonmanipulators. 3. s trong m anipulation: the manipulators seek to vote in a way that makes their preferred candidate win in every extension of the partial votes of the non-manipulators. we consider several scenarios for which the traditional manipulation problems are easy (for instance, borda with a single manipulator). for many of them, the corresponding manipulative questions that we propose turn out to be computationally intractable. our hardness results often hold even when very little information is missing, or in other words, even when the instances are very close to the complete information setting. our results show that the impact of paucity of information on the computational complexity of manipulation crucially depends on the notion of manipulation under consideration. our overall conclusion is that computational hardness continues to be a valid obstruction to manipulation, in the context of a more realistic model.
8
abstract we propose a simple approach which, given distributed computing resources, can nearly achieve the accuracy of k-nn prediction, while matching (or improving) the faster prediction time of 1-nn. the approach consists of aggregating denoised 1-nn predictors over a small number of distributed subsamples. we show, both theoretically and experimentally, that small subsample sizes suffice to attain similar performance as k-nn, without sacrificing the computational efficiency of 1-nn.
10
abstract. we consider graded artinian complete intersection algebras a = c[x0 , . . . , xm ]/i with i generated by homogeneous forms of degree d ≥ 2. we show that the general multiplication by a linear form µl : ad−1 → ad is injective. we prove that the weak lefschetz property for holds for any c.i. algebra a as above with d = 2 and m ≤ 4, previously known for m ≤ 3.
0
abstract recurrent neural networks (rnns), particularly long short-term memory (lstm), have gained much attention in automatic speech recognition (asr). although some successful stories have been reported, training rnns remains highly challenging, especially with limited training data. recent research found that a well-trained model can be used as a teacher to train other child models, by using the predictions generated by the teacher model as supervision. this knowledge transfer learning has been employed to train simple neural nets with a complex one, so that the final performance can reach a level that is infeasible to obtain by regular training. in this paper, we employ the knowledge transfer learning approach to train rnns (precisely lstm) using a deep neural network (dnn) model as the teacher. this is different from most of the existing research on knowledge transfer learning, since the teacher (dnn) is assumed to be weaker than the child (rnn); however, our experiments on an asr task showed that it works fairly well: without applying any tricks on the learning scheme, this approach can train rnns successfully even with limited training data. index terms— recurrent neural network, long shortterm memory, knowledge transfer learning, automatic speech recognition 1. introduction deep learning has gained significant success in a wide range of applications, for example, automatic speech recognition (asr) [1]. a powerful deep learning model that has been reported effective in asr is the recurrent neural network (rnn), e.g., [2, 3, 4]. an obvious advantage of rnns compared to conventional deep neural networks (dnns) is that rnns can model long-term temporal properties and thus are suitable for modeling speech signals. a simple training method for rnns is the backpropagation through time algorithm [5]. this first-order approach, this work was supported by the national natural science foundation of china under grant no. 61371136 and the mestdc phd foundation project no. 20130002120011. this paper was also supported by huilan ltd. and sinovoice.
9
abstract— this paper considers the problem of implementing a previously proposed distributed direct coupling quantum observer for a closed linear quantum system. by modifying the form of the previously proposed observer, the paper proposes a possible experimental implementation of the observer plant system using a non-degenerate parametric amplifier and a chain of optical cavities which are coupled together via optical interconnections. it is shown that the distributed observer converges to a consensus in a time averaged sense in which an output of each element of the observer estimates the specified output of the quantum plant.
3
abstract. using valuation rings and valued fields as examples, we discuss in which ways the notions of “topological ifs attractor” and “fractal space” can be generalized to cover more general settings.
0
abstract we consider the problem of reconstructing an unknown bounded function u defined on a domain x ⊂ rd from noiseless or noisy samples of u at n points (xi )i=1,...,n . we measure the reconstruction error in a norm l2 (x, dρ) for some given probability measure dρ. given a linear space vm with dim(vm ) = m ≤ n, we study in general terms the weighted least-squares approximations from the spaces vm based on independent random samples. it is well known that least-squares approximations can be inaccurate and unstable when m is too close to n, even in the noiseless case. recent results from [4, 5] have shown the interest of using weighted least squares for reducing the number n of samples that is needed to achieve an accuracy comparable to that of best approximation in vm , compared to standard least squares as studied in [3]. the contribution of the present paper is twofold. from the theoretical perspective, we establish results in expectation and in probability for weighted least squares in general approximation spaces vm . these results show that for an optimal choice of sampling measure dµ and weight w, which depends on the space vm and on the measure dρ, stability and optimal accuracy are achieved under the mild condition that n scales linearly with m up to an additional logarithmic factor. in contrast to [3], the present analysis covers cases where the function u and its approximants from vm are unbounded, which might occur for instance in the relevant case where x = rd and dρ is the gaussian measure. from the numerical perspective, we propose a sampling method which allows one to generate independent and identically distributed samples from the optimal measure dµ. this method becomes of interest in the multivariate setting where dµ is generally not of tensor product type. we illustrate this for particular examples of approximation spaces vm of polynomial type, where the domain x is allowed to be unbounded and high or even infinite dimensional, motivated by certain applications to parametric and stochastic pdes.
10
abstract in this paper, we explore different ways to extend a recurrent neural network (rnn) to a deep rnn. we start by arguing that the concept of depth in an rnn is not as clear as it is in feedforward neural networks. by carefully analyzing and understanding the architecture of an rnn, however, we find three points of an rnn which may be made deeper; (1) input-to-hidden function, (2) hidden-tohidden transition and (3) hidden-to-output function. based on this observation, we propose two novel architectures of a deep rnn which are orthogonal to an earlier attempt of stacking multiple recurrent layers to build a deep rnn (schmidhuber, 1992; el hihi and bengio, 1996). we provide an alternative interpretation of these deep rnns using a novel framework based on neural operators. the proposed deep rnns are empirically evaluated on the tasks of polyphonic music prediction and language modeling. the experimental result supports our claim that the proposed deep rnns benefit from the depth and outperform the conventional, shallow rnns.
9
abstract—today’s hpc applications are producing extremely large amounts of data, such that data storage and analysis are becoming more challenging for scientific research. in this work, we design a new error-controlled lossy compression algorithm for large-scale scientific data. our key contribution is significantly improving the prediction hitting rate (or prediction accuracy) for each data point based on its nearby data values along multiple dimensions. we derive a series of multilayer prediction formulas and their unified formula in the context of data compression. one serious challenge is that the data prediction has to be performed based on the preceding decompressed values during the compression in order to guarantee the error bounds, which may degrade the prediction accuracy in turn. we explore the best layer for the prediction by considering the impact of compression errors on the prediction accuracy. moreover, we propose an adaptive error-controlled quantization encoder, which can further improve the prediction hitting rate considerably. the data size can be reduced significantly after performing the variablelength encoding because of the uneven distribution produced by our quantization encoder. we evaluate the new compressor on production scientific data sets and compare it with many other state-of-the-art compressors: gzip, fpzip, zfp, sz-1.1, and isabela. experiments show that our compressor is the best in class, especially with regard to compression factors (or bitrates) and compression errors (including rmse, nrmse, and psnr). our solution is better than the second-best solution by more than a 2x increase in the compression factor and 3.8x reduction in the normalized root mean squared error on average, with reasonable error bounds and user-desired bitrates.
7
abstract in this paper we consider the problem of detecting a change in the parameters of an autoregressive process, where the moments of the innovation process do not necessarily exist. an empirical likelihood ratio test for the existence of a change point is proposed and its asymptotic properties are studied. in contrast to other work on change point tests using empirical likelihood, we do not assume knowledge of the location of the change point. in particular, we prove that the maximizer of the empirical likelihood is a consistent estimator for the parameters of the autoregressive model in the case of no change point and derive the limiting distribution of the corresponding test statistic under the null hypothesis. we also establish consistency of the new test. a nice feature of the method consists in the fact that the resulting test is asymptotically distribution free and does not require an estimate of the long run variance. the asymptotic properties of the test are investigated by means of
10
abstract
6
abstract pandemic influenza has the epidemic potential to kill millions of people. while various preventive measures exist (i.a., vaccination and school closures), deciding on strategies that lead to their most effective and efficient use, remains challenging. to this end, individual-based epidemiological models are essential to assist decision makers in determining the best strategy to curve epidemic spread. however, individual-based models are computationally intensive and therefore it is pivotal to identify the optimal strategy using a minimal amount of model evaluations. additionally, as epidemiological modeling experiments need to be planned, a computational budget needs to be specified a priori. consequently, we present a new sampling method to optimize the evaluation of preventive strategies using fixed budget best-arm identification algorithms. we use epidemiological modeling theory to derive knowledge about the reward distribution which we exploit using bayesian best-arm identification algorithms (i.e., top-two thompson sampling and bayesgap). we evaluate these algorithms in a realistic experimental setting and demonstrate that it is possible to identify the optimal strategy using only a limited number of model evaluations, i.e., 2-to-3 times faster compared to the uniform sampling method, the predominant technique used for epidemiological decision making in the literature. finally, we contribute and evaluate a statistic for top-two thompson sampling to inform the decision makers about the confidence of an arm recommendation.
2
abstract this paper deals with feature selection procedures for spatial point processes intensity estimation. we consider regularized versions of estimating equations based on campbell theorem derived from two classical functions: poisson likelihood and logistic regression likelihood. we provide general conditions on the spatial point processes and on penalty functions which ensure consistency, sparsity and asymptotic normality. we discuss the numerical implementation and assess finite sample properties in a simulation study. finally, an application to tropical forestry datasets illustrates the use of the proposed methods.
10
abstract—wireless networked control systems (wncs) are composed of spatially distributed sensors, actuators, and controllers communicating through wireless networks instead of conventional point-to-point wired connections. due to their main benefits in the reduction of deployment and maintenance costs, large flexibility and possible enhancement of safety, wncs are becoming a fundamental infrastructure technology for critical control systems in automotive electrical systems, avionics control systems, building management systems, and industrial automation systems. the main challenge in wncs is to jointly design the communication and control systems considering their tight interaction to improve the control performance and the network lifetime. in this survey, we make an exhaustive review of the literature on wireless network design and optimization for wncs. first, we discuss what we call the critical interactive variables including sampling period, message delay, message dropout, and network energy consumption. the mutual effects of these communication and control variables motivate their joint tuning. we discuss the effect of controllable wireless network parameters at all layers of the communication protocols on the probability distribution of these interactive variables. we also review the current wireless network standardization for wncs and their corresponding methodology for adapting the network parameters. moreover, we discuss the analysis and design of control systems taking into account the effect of the interactive variables on the control system performance. finally, we present the state-of-the-art wireless network design and optimization for wncs, while highlighting the tradeoff between the achievable performance and complexity of various approaches. we conclude the survey by highlighting major research issues and identifying future research directions. index terms—wireless networked control systems, wireless sensor and actuator networks, joint design, delay, reliability, sampling rate, network lifetime, optimization.
3
abstract. in this article, the projectivity of a finitely generated flat module of a commutative ring is studied through its exterior powers and invariant factors. consequently, the related results of endo, vasconcelos, wiegand, cox-rush and puninski-rothmaler on the projectivity of f.g. flat modules are generalized.
0
abstract
2
abstract—this paper studies optimal tracking performance issues for multi-input-multi-output linear time-invariant systems under networked control with limited bandwidth and additive colored white gaussian noise channel. the tracking performance is measured by control input energy and the energy of the error signal between the output of the system and the reference signal with respect to a brownian motion random process. this paper focuses on two kinds of network parameters, the basic network parameter-bandwidth and the additive colored white gaussian noise, and studies the tracking performance limitation problem. the best attainable tracking performance is obtained, and the impact of limited bandwidth and additive colored white gaussian noise of the communication channel on the attainable tracking performance is revealed. it is shown that the optimal tracking performance depends on nonminimum phase zeros, gain at all frequencies and their directions unitary vector of the given plant, as well as the limited bandwidth and additive colored white gaussian noise of the communication channel. the simulation results are finally given to illustrate the theoretical results. index terms—networked control systems, bandwidth, additive colored white gaussian noise, performance limitation.
3
abstract a group of order pn (p prime) has an indecomposable polynomial invariant of degree at least pn−1 if and only if the group has a cyclic subgroup of index at most p or it is isomorphic to the elementary abelian group of order 8 or the heisenberg group of order 27. keywords: polynomial invariants, degree bounds, zero-sum sequences
4
abstract. secret sharing is a cryptographic discipline in which the goal is to distribute information about a secret over a set of participants in such a way that only specific authorized combinations of participants together can reconstruct the secret. thus, secret sharing schemes are systems of variables in which it is very clearly specified which subsets have information about the secret. as such, they provide perfect model systems for information decompositions. however, following this intuition too far leads to an information decomposition with negative partial information terms, which are difficult to interpret. one possible explanation is that the partial information lattice proposed by williams and beer is incomplete and has to be extended to incorporate terms corresponding to higher order redundancy. these results put bounds on information decompositions that follow the partial information framework, and they hint at where the partial information lattice needs to be improved.
7
abstract permutation polynomials over finite fields are an interesting subject due to their important applications in the areas of mathematics and engineering. in this paper we investigate the trinomial f (x) = x(p−1)q+1 + xpq − xq+(p−1) over the finite field fq2 , where p is an odd prime and q = pk with k being a positive integer. it is shown that when p = 3 or 5, f (x) is a permutation trinomial of fq2 if and only if k is even. this property is also true for more general class of polynomials g(x) = x(q+1)l+(p−1)q+1 + x(q+1)l+pq − x(q+1)l+q+(p−1) , where l is a nonnegative integer and gcd(2l + p, q − 1) = 1. moreover, we also show that for p = 5 the permutation trinomials f (x) proposed here are new in the sense that they are not multiplicative equivalent to previously known ones of similar form. index terms finite fields, permutation polynomials, trinomials, niho exponents, multiplicative inequivalent. ams 94b15, 11t71
7
abstracts/definitions) to improve the answering ability of a model. marino et al. (2017) explicitly incorporate knowledge graphs into an image classification model. xu et al. (2016) created a recall mechanism into a standard lstm cell that retrieves pieces of external knowledge encoded by a single representation for a conversation model. concurrently, dhingra et al. (2017) exploit linguistic knowledge using mage-grus, an adapation of grus to handle graphs, however, external knowledge has to be present in form of triples. the main difference to our approach is that we incorporate external knowledge in free text form on the word level prior to processing the task at hand which constitutes a more flexible setup. ahn et al. (2016) exploit knowledge base facts about mentioned entities for neural language models. bahdanau et al. (2017) and long et al. (2017) create word embeddings on-the-fly by reading word definitions prior to processing the task at hand. pilehvar et al. (2017) seamlessly incorporate information about word senses into their representations before solving the downstream nlu task, which is similar. we go one step further by seamlessly integrating all kinds of fine-grained assertions about concepts that might be relevant for the task at hand. another important aspect of our approach is the notion of dynamically updating wordrepresentations. tracking and updating concepts, entities or sentences with dynamic memories is a very active research direction (kumar et al., 2016; henaff et al., 2017; ji et al., 2017; kobayashi et al., 2017). however, those works typically focus on particular tasks whereas our approach is taskagnostic and most importantly allows for the integration of external background knowledge. other related work includes storing temporary information in weight matrices instead of explicit neural activations (such as word representations) as a biologically more plausible alternative.
2
abstraction: at the highest level, the software architecture models the m. bujorianu and m. fisher (eds.): workshop on formal methods for aerospace (fma) eptcs 20, 2010, pp. 80–87, doi:10.4204/eptcs.20.9
6
abstract. we prove that abelian subgroups of the outer automorphism group of a free group are quasiisometrically embedded. our proof uses recent developments in the theory of train track maps by feighnhandel. as an application, we prove the rank conjecture for out(fn ).
4
abstract catroid is a free and open source visual programming language, programming environment, image manipulation program, and website. catroid allows casual and first-time users starting from age eight to develop their own animations and games solely using their android phones or tablets. catroid also allows to wirelessly control external hardware such as lego mindstorms robots via bluetooth, bluetooth arduino boards, as well as parrot’s popular and inexpensive ar.drone quadcopters via wifi.
6
abstract the map-reduce computing framework rose to prominence with datasets of such size that dozens of machines on a single cluster were needed for individual jobs. as datasets approach the exabyte scale, a single job may need distributed processing not only on multiple machines, but on multiple clusters. we consider a scheduling problem to minimize weighted average completion time of n jobs on m distributed clusters of parallel machines. in keeping with the scale of the problems motivating this work, we assume that (1) each job is divided into m “subjobs” and (2) distinct subjobs of a given job may be processed concurrently. when each cluster is a single machine, this is the np-hard concurrent open shop problem. a clear limitation of such a model is that a serial processing assumption sidesteps the issue of how different tasks of a given subjob might be processed in parallel. our algorithms explicitly model clusters as pools of resources and effectively overcome this issue. under a variety of parameter settings, we develop two constant factor approximation algorithms for this problem. the first algorithm uses an lp relaxation tailored to this problem from prior work. this lp-based algorithm provides strong performance guarantees. our second algorithm exploits a surprisingly simple mapping to the special case of one machine per cluster. this mapping-based algorithm is combinatorial and extremely fast. these are the first constant factor approximations for this problem. remark - a shorter version of this paper (one that omitted several proofs) appeared in the proceedings of the 2016 european symposium on algorithms. 1998 acm subject classification f.2.2 nonnumerical algorithms and problems keywords and phrases approximation algorithms, distributed computing, machine scheduling, lp relaxations, primal-dual algorithms digital object identifier 10.4230/lipics.esa.2016.234
8
abstract the propagation of sound in a shallow water environment is characterized by boundary reflections from the sea surface and sea floor. these reflections result in multiple (indirect) sound propagation paths, which can degrade the performance of passive sound source localization methods. this paper proposes the use of convolutional neural networks (cnns) for the localization of sources of broadband acoustic radiated noise (such as motor vessels) in shallow water multipath environments. it is shown that cnns operating on cepstrogram and generalized cross-correlogram inputs are able to more reliably estimate the instantaneous range and bearing of transiting motor vessels when the source localization performance of conventional passive ranging methods is degraded. the ensuing improvement in source localization performance is demonstrated using real data collected during an at-sea experiment. index terms— source localization, doa estimation, convolutional neural networks, passive sonar, reverberation 1. introduction sound source localization plays an important role in array signal processing with wide applications in communication, sonar and robotics systems [1]. it is a focal topic in the scientific literature on acoustic array signal processing with a continuing challenge being acoustic source localization in the presence of interfering multipath arrivals [2, 3, 4]. in practice, conventional passive narrowband sonar array methods involve frequency-domain beamforming of the outputs of hydrophone elements in a receiving array to detect weak signals, resolve closely-spaced sources, and estimate the direction of a sound source. typically, 10-100 sensors form a linear array with a uniform interelement spacing of half a wavelength at the array’s design frequency. however, this narrowband approach has application over a limited band of frequencies. the upper limit is set by the design frequency, above which grating lobes form due to spatial aliasing, leading to ambiguous source directions. the lower limit is set one octave below the design frequency because at lower frequencies the directivity of the array is much reduced as the beamwidths broaden. an alternative approach to sound source localization is to measure the time difference of arrival (tdoa) of the signal at an array of spatially distributed receivers [5, 6, 7, 8], allowing the instantaneous position of the source to be estimated. the accuracy of the source position estimates is found to be sensitive to any uncertainty in the sensor positions [9]. furthermore, reverberation has an adverse effect on time delay estimation, which negatively impacts ∗ work
7
abstract we consider the ransac algorithm in the context of subspace recovery and subspace clustering. we derive some theory and perform some numerical experiments. we also draw some correspondences with the methods of hardt and moitra (2013) and chen and lerman (2009b).
10
abstraction of a physical sensing device b located at position p and running program p . module m is the collection of methods that the sensor makes available for internal and for external usage. typically this collection of methods may be interpreted as the library of functions of the tiny operating system installed in the sensor. sensors may only broadcast values to its neighborhood sensors. radius rt defines the transmitting power of a sensor and specifies the border of communication: a circle centered at position p (the position of the sensor) with radius rt . likewise, radius rs defines the sensing capability of the sensor, meaning that a sensor may only read values inside the circle centered at position p with radius rs . values h~v ip define the field of measures that may be sensed. a value consists of a tuple ~v denoting the strength of the measure at a given position p of the plane. values are managed by the environment; in csn there are no primitives for manipulating values, besides reading (sensing) values. we assume that the environment inserts these values in the network and update its contents. networks are combined using the parallel composition operator | . 1
2
abstract—filtering and smoothing algorithms for linear discrete-time state-space models with skew-t-distributed measurement noise are presented. the presented algorithms use a variational bayes based posterior approximation with coupled location and skewness variables to reduce the error caused by the variational approximation. although the variational update is done suboptimally, our simulations show that the proposed method gives a more accurate approximation of the posterior covariance matrix than an earlier proposed variational algorithm. consequently, the novel filter and smoother outperform the earlier proposed robust filter and smoother and other existing low-complexity alternatives in accuracy and speed. we present both simulations and tests based on real-world navigation data, in particular gps data in an urban area, to demonstrate the performance of the novel methods. moreover, the extension of the proposed algorithms to cover the case where the distribution of the measurement noise is multivariate skew-t is outlined. finally, the paper presents a study of theoretical performance bounds for the proposed algorithms.
3
abstract in this paper, a discrete-time multi-agent system is presented which is formulated in terms of the delta operator. the proposed multi-agent system can unify discrete-time and continuous-time multi-agent systems. in a multi-agent network, in practice, the communication among agents is acted upon by various factors. the communication network among faulty agents may cause link failures, which is modeled by randomly switching graphs. first, we show that the delta representation of discrete-time multi-agent system reaches consensus in mean (in probability and almost surely) if the expected graph is strongly connected. the results induce that the continuous-time multi-agent system with random networks can also reach consensus in the same sense. second, the influence of faulty agents on consensus value is quantified under original network. by using matrix perturbation theory, the error bound is also presented in this paper. finally, a simulation example is provided to demonstrate the effectiveness of our theoretical results. index terms consensus, multi-agent systems, delta operator, link failures, error bound.
3
abstract we give sufficient identifiability conditions for estimating mixing proportions in two-component mixtures of skew normal distributions with one known component. we consider the univariate case as well as two multivariate extensions: a multivariate skew normal distribution (msn) by azzalini and dalla valle (1996) and the canonical fundamental skew normal distribution (cfusn) by arellano-valle and genton (2005). the characteristic function of the cfusn distribution is additionally derived.
10
abstract: the shear strength and stick-slip behavior of a rough rock joint are analyzed using the
5
abstract for σ an orientable surface of finite topological type having genus at least 3 (possibly closed or possibly with any number of punctures or boundary components), we show that the mapping class group m od(σ) has no faithful linear representation in any dimension over any field of positive characteristic.
4
abstract. in this work we propose a heuristic algorithm for the layout optimization for disks installed in a rotating circular container. this is a unequal circle packing problem with additional balance constraints. it proved to be an np-hard problem, which justifies heuristics methods for its resolution in larger instances. the main feature of our heuristic is based on the selection of the next circle to be placed inside the container according to the position of the system’s center of mass. our approach has been tested on a series of instances up to 55 circles and compared with the literature. computational results show good performance in terms of solution quality and computational time for the proposed algorithm. keywords: packing problem, layout optimization problem, nonidentical circles, heuristic algorithm
5
abstract a rack on [n] can be thought of as a set of maps (fx )x∈[n] , where each fx is a permutation of [n] such that f(x)fy = fy−1 fx fy for all x and y. in 2013, blackburn showed that the number of isomorphism classes of racks 2 2 on [n] is at least 2(1/4−o(1))n and at most 2(c+o(1))n , where c ≈ 1.557; 2 in this paper we improve the upper bound to 2(1/4+o(1))n , matching the lower bound. the proof involves considering racks as loopless, edge-coloured directed multigraphs on [n], where we have an edge of colour y between x and z if and only if (x)fy = z, and applying various combinatorial tools.
4
abstract this paper discusses minimum distance estimation method in the linear regression model with dependent errors which are strongly mixing. the regression parameters are estimated through the minimum distance estimation method, and asymptotic distributional properties of the estimators are discussed. a simulation study compares the performance of the minimum distance estimator with other well celebrated estimator. this simulation study shows the superiority of the minimum distance estimator over another estimator. koulmde (r package) which was used for the simulation study is available online. see section 4 for the detail.
10
abstract. programs that process data that reside in files are widely used in varied domains, such as banking, healthcare, and web-traffic analysis. precise static analysis of these programs in the context of software verification and transformation tasks is a challenging problem. our key insight is that static analysis of file-processing programs can be made more useful if knowledge of the input file formats of these programs is made available to the analysis. we propose a generic framework that is able to perform any given underlying abstract interpretation on the program, while restricting the attention of the analysis to program paths that are potentially feasible when the program’s input conforms to the given file format specification. we describe an implementation of our approach, and present empirical results using real and realistic programs that show how our approach enables novel verification and transformation tasks, and also improves the precision of standard analysis problems.
6
abstract integer-forcing source coding has been proposed as a low-complexity method for compression of distributed correlated gaussian sources. in this scheme, each encoder quantizes its observation using the same fine lattice and reduces the result modulo a coarse lattice. rather than directly recovering the individual quantized signals, the decoder first recovers a full-rank set of judiciously chosen integer linear combinations of the quantized signals, and then inverts it. it has been observed that the method works very well for “most” but not all source covariance matrices. the present work quantifies the measure of bad covariance matrices by studying the probability that integer-forcing source coding fails as a function of the allocated rate, where the probability is with respect to a random orthonormal transformation that is applied to the sources prior to quantization. for the important case where the signals to be compressed correspond to the antenna inputs of relays in an i.i.d. rayleigh fading environment, this orthonormal transformation can be viewed as being performed by nature. hence, the results provide performance guarantees for distributed source coding via integer forcing in this scenario.
7
abstract the classical-input quantum-output (cq) wiretap channel is a communication model involving a classical sender x, a legitimate quantum receiver b, and a quantum eavesdropper e. the goal of a private communication protocol that uses such a channel is for the sender x to transmit a message in such a way that the legitimate receiver b can decode it reliably, while the eavesdropper e learns essentially nothing about which message was transmitted. the ε-oneshot private capacity of a cq wiretap channel is equal to the maximum number of bits that can be transmitted over the channel, such that the privacy error is no larger than ε ∈ (0, 1). the present paper provides a lower bound on the ε-one-shot private classical capacity, by exploiting the recently developed techniques of anshu, devabathini, jain, and warsi, called position-based coding and convex splitting. the lower bound is equal to a difference of the hypothesis testing mutual information between x and b and the “alternate” smooth max-information between x and e. the one-shot lower bound then leads to a non-trivial lower bound on the second-order coding rate for private classical communication over a memoryless cq wiretap channel.
7
abstract—the world is connected through the internet. as the abundance of internet users connected into the web and the popularity of cloud computing research, the need of artificial intelligence (ai) is demanding. in this research, genetic algorithm (ga) as ai optimization method through natural selection and genetic evolution is utilized. there are many applications of ga such as web mining, load balancing, routing, and scheduling or web service selection. hence, it is a challenging task to discover whether the code mainly server side and web based language technology affects the performance of ga. travelling salesperson problem (tsp) as non polynomial-hard (np-hard) problem is provided to be a problem domain to be solved by ga. while many scientists prefer python in ga implementation, another popular high-level interpreter programming language such as php (php hypertext preprocessor) and ruby were benchmarked. line of codes, file sizes, and performances based on ga implementation and runtime were found varies among these programming languages. based on the result, the use of ruby in ga implementation is recommended. keywords—tsp; genetic algorithm; web-programming language
6
abstract—smartphone applications designed to track human motion in combination with wearable sensors, e.g., during physical exercising, raised huge attention recently. commonly, they provide quantitative services, such as personalized training instructions or the counting of distances. but qualitative monitoring and assessment is still missing, e.g., to detect malpositions, to prevent injuries, or to optimize training success. we address this issue by presenting a concept for qualitative as well as generic assessment of recurrent human motion by processing multi-dimensional, continuous time series tracked with motion sensors. therefore, our segmentation procedure extracts individual events of specific length and we propose expressive features to accomplish a qualitative motion assessment by supervised classification. we verified our approach within a comprehensive study encompassing 27 athletes undertaking different body weight exercises. we are able to recognize six different exercise types with a success rate of 100% and to assess them qualitatively with an average success rate of 99.3%. keywords—motion assessment; activity recognition; physical exercises; segmentation
1
abstractions for high-performance remote data access, mechanisms for scalable data replication, cataloging with rich semantic and syntactic information, data discovery, distributed monitoring, and web-based portals for using the system. keywords—climate modeling, data management, earth system grid (esg), grid computing.
5
abstract—modern dense flash memory devices operate at very low error rates, which require powerful error correcting coding (ecc) techniques. an emerging class of graph-based ecc techniques that has broad applications is the class of spatiallycoupled (sc) codes, where a block code is partitioned into components that are then rewired multiple times to construct an sc code. here, our focus is on sc codes with the underlying circulant-based structure. in this paper, we present a three-stage approach for the design of high performance non-binary sc (nbsc) codes optimized for practical flash channels; we aim at minimizing the number of detrimental general absorbing sets of type two (gasts) in the graph of the designed nb-sc code. in the first stage, we deploy a novel partitioning mechanism, called the optimal overlap partitioning, which acts on the protograph of the sc code to produce optimal partitioning corresponding to the smallest number of detrimental objects. in the second stage, we apply a new circulant power optimizer to further reduce the number of detrimental gasts. in the third stage, we use the weight consistency matrix framework to manipulate edge weights to eliminate as many as possible of the gasts that remain in the nb-sc code after the first two stages (that operate on the unlabeled graph of the code). simulation results reveal that nbsc codes designed using our approach outperform state-of-theart nb-sc codes when used over flash channels.
7
abstract. let g be a finite group and, for a prime p, let s be a sylow p-subgroup of g. a character χ of g is called sylp -regular if the restriction of χ to s is the character of the regular representation of s. if, in addition, χ vanishes at all elements of order divisible by p, χ is said to be steinberg-like. for every finite simple group g we determine all primes p for which g admits a steinberg-like character, except for alternating groups in characteristic 2. moreover, we determine all primes for which g has a projective f g-module of dimension |s|, where f is an algebraically closed field of characteristic p.
4
abstract we present bounded dynamic (but observer-free) output feedback laws that achieve global stabilization of equilibrium profiles of the partial differential equation (pde) model of a simplified, age-structured chemostat model. the chemostat pde state is positive-valued, which means that our global stabilization is established in the positive orthant of a particular function space—a rather non-standard situation, for which we develop non-standard tools. our feedback laws do not employ any of the (distributed) parametric knowledge of the model. moreover, we provide a family of highly unconventional control lyapunov functionals (clfs) for the age-structured chemostat pde model. two kinds of feedback stabilizers are provided: stabilizers with continuously adjusted input and sampled-data stabilizers. the results are based on the transformation of the first-order hyperbolic partial differential equation to an ordinary differential equation (one-dimensional) and an integral delay equation (infinite-dimensional). novel stability results for integral delay equations are also provided; the results are of independent interest and allow the explicit construction of the clf for the age-structured chemostat model.
3
abstract this paper studies a pursuit-evasion problem involving a single pursuer and a single evader, where we are interested in developing a pursuit strategy that doesn’t require continuous, or even periodic, information about the position of the evader. we propose a self-triggered control strategy that allows the pursuer to sample the evader’s position autonomously, while satisfying desired performance metric of evader capture. the work in this paper builds on the previously proposed self-triggered pursuit strategy which guarantees capture of the evader in finite time with a finite number of evader samples. however, this algorithm relied on the unrealistic assumption that the evader’s exact position was available to the pursuer. instead, we extend our previous framework to develop an algorithm which allows for uncertainties in sampling the information about the evader, and derive tolerable upper-bounds on the error such that the pursuer can guarantee capture of the evader. in addition, we outline the advantages of retaining the evader’s history in improving the current estimate of the true location of the evader that can be used to capture the evader with even less samples. our approach is in sharp contrast to the existing works in literature and our results ensure capture without sacrificing any performance in terms of guaranteed time-to-capture, as compared to classic algorithms that assume continuous availability of information. key words: pursuit-evasion; self-triggered control; sampled-data control; set-valued analysis
3
abstract. in this article we explore an algorithm for diffeomorphic random sampling of nonuniform probability distributions on riemannian manifolds. the algorithm is based on optimal information transport (oit)—an analogue of optimal mass transport (omt). our framework uses the deep geometric connections between the fisher-rao metric on the space of probability densities and the right-invariant information metric on the group of diffeomorphisms. the resulting sampling algorithm is a promising alternative to omt, in particular as our formulation is semi-explicit, free of the nonlinear monge–ampere equation. compared to markov chain monte carlo methods, we expect our algorithm to stand up well when a large number of samples from a low dimensional nonuniform distribution is needed. keywords: density matching, information geometry, fisher–rao metric, optimal transport, image registration, diffeomorphism groups, random sampling msc2010: 58e50, 49q10, 58e10
10
abstract this paper presents the development of an adaptive algebraic multiscale solver for compressible flow (c-ams) in heterogeneous porous media. similar to the recently developed ams for incompressible (linear) flows [wang et al., jcp, 2014], c-ams operates by defining primal and dual-coarse blocks on top of the fine-scale grid. these coarse grids facilitate the construction of a conservative (finite volume) coarsescale system and the computation of local basis functions, respectively. however, unlike the incompressible (elliptic) case, the choice of equations to solve for basis functions in compressible problems is not trivial. therefore, several basis function formulations (incompressible and compressible, with and without accumulation) are considered in order to construct an efficient multiscale prolongation operator. as for the restriction operator, c-ams allows for both multiscale finite volume (msfv) and finite element (msfe) methods. finally, in order to resolve highfrequency errors, fine-scale (pre- and post-) smoother stages are employed. in order to reduce computational expense, the c-ams operators (prolongation, restriction, and smoothers) are updated adaptively. in addition to this, the linear system in the newton-raphson loop is infrequently updated. systematic numerical experiments are performed to determine the effect of the various options, outlined above, on the c-ams convergence behaviour. an efficient c-ams strategy for heterogeneous 3d compressible problems is developed based on overall cpu times. finally, c-ams is compared against an industrial-grade algebraic multigrid (amg) solver. results of this comparison illustrate that the c-ams is quite efficient as a nonlinear solver, even when iterated to machine accuracy. key words: multiscale methods, compressible flows, heterogeneous porous media, scalable linear solvers, multiscale finite volume method, multiscale finite element method, iterative multiscale methods, algebraic multiscale methods.
5
abstract numerical simulation of compressible fluid flows is performed using the euler equations. they include the scalar advection equation for the density, the vector advection equation for the velocity and a given pressure dependence on the density. an approximate solution of an initial–boundary value problem is calculated using the finite element approximation in space. the fully implicit two–level scheme is used for discretization in time. numerical implementation is based on newton’s method. the main attention is paid to fulfilling conservation laws for the mass and total mechanical energy for the discrete formulation. two–level schemes of splitting by physical processes are employed for numerical solving problems of barotropic fluid flows. for a transition from one time level to the next one, an iterative process is used, where at each iteration the linearized scheme is implemented via solving individual problems for the density and velocity. possibilities of the proposed schemes are illustrated by numerical results for a two–dimensional model problem with density perturbations. keywords: compressible fluids, the euler system, barotropic fluid, finite element method, conservation laws, two–level schemes, decoupling scheme
5
abstract. formal concept analysis and its associated conceptual structures have been used to support exploratory search through conceptual navigation. relational concept analysis (rca) is an extension of formal concept analysis to process relational datasets. rca and its multiple interconnected structures represent good candidates to support exploratory search in relational datasets, as they are enabling navigation within a structure as well as between the connected structures. however, building the entire structures does not present an efficient solution to explore a small localised area of the dataset, for instance to retrieve the closest alternatives to a given query. in these cases, generating only a concept and its neighbour concepts at each navigation step appears as a less costly alternative. in this paper, we propose an algorithm to compute a concept and its neighbourhood in extended concept lattices. the concepts are generated directly from the relational context family, and possess both formal and relational attributes. the algorithm takes into account two rca scaling operators. we illustrate it on an example. keywords: relational concept analysis, formal concept analysis, ondemand generation
2
abstract a research frontier has emerged in scientific computation, wherein discretisation error is regarded as a source of epistemic uncertainty that can be modelled. this raises several statistical challenges, including the design of statistical methods that enable the coherent propagation of probabilities through a (possibly deterministic) computational work-flow, in order to assess the impact of discretisation error on the computer output. this paper examines the case for probabilistic numerical methods in routine statistical computation. our focus is on numerical integration, where a probabilistic integrator is equipped with a full distribution over its output that reflects the fact that the integrand has been discretised. our main technical contribution is to establish, for the first time, rates of posterior contraction for one such method. several substantial applications are provided for illustration and critical evaluation, including examples from statistical modelling, computer graphics and a computer model for an oil reservoir.
10
abstract and talk (mcdonnell et al., 2017). 10 imagenet32 4x imagenet32 10x imagenet32 15x cifar-100 4x cifar-100 10x cifar-10 svhn mnist imagenet single crop imagenet multi-crop bwn on imagenet
9
abstract today’s javascript applications are composed of scripts from different origins that are loaded at run time. as not all of these origins are equally trusted, the execution of these scripts should be isolated from one another. however, some scripts must access the application state and some may be allowed to change it, while preserving the confidentiality and integrity constraints of the application. this paper presents design and implementation of decentjs, a language-embedded sandbox for full javascript. it enables scripts to run in a configurable degree of isolation with fine-grained access control. it provides a transactional scope in which effects are logged for review by the access control policy. after inspection of the log, effects can be committed to the application state or rolled back. the implementation relies on javascript proxies to guarantee full interposition for the full language and for all code, including dynamically loaded scripts and code injected via eval. its only restriction is that scripts must be compliant with javascript’s strict mode. 1998 acm subject classification d.4.6 security and protection keywords and phrases javascript, sandbox, proxy
6
abstract let n (n) denote the number of isomorphism types of groups of order n. we consider the integers n that are products of at most 4 not necessarily distinct primes and exhibit formulas for n (n) for such n.
4
abstract we consider the action of an irreducible outer automorphism φ on the closure of culler–vogtmann outer space. this action has north-south dynamics and so, under iteration, points converge exponentially to [t+φ ]. for each n ≥ 3, we give a family of outer automorphisms φk ∈ out(fn ) such that as, k goes to infinity, the rate of convergence of φk goes to infinity while the rate of convergence of φ−1 goes to one. even k if we only require the rate of convergence of φk to remain bounded away from one, no such family can be constructed when n < 3. this family also provides an explicit example of a property described by handel and mosher: that there is no uniform upper bound on the distance between the axes of an automorphism and its inverse.
4
abstract—distribution grid is the medium and low voltage part of a large power system. structurally, the majority of distribution networks operate radially, such that energized lines form a collection of trees, i.e. forest, with a substation being at the root of any tree. the operational topology/forest may change from time to time, however tracking these changes, even though important for the distribution grid operation and control, is hindered by limited real-time monitoring. this paper develops a learning framework to reconstruct radial operational structure of the distribution grid from synchronized voltage measurements in the grid subject to the exogenous fluctuations in nodal power consumption. to detect operational lines our learning algorithm uses conditional independence tests for continuous random variables that is applicable to a wide class of probability distributions of the nodal consumption and gaussian injections in particular. moreover, our algorithm applies to the practical case of unbalanced three-phase power flow. algorithm performance is validated on ac power flow simulations over ieee distribution grid test cases. keywords—distribution networks, power flow, unbalanced threephase, graphical models, conditional independence, computational complexity
3
abstract. let k be a discretly henselian field whose residue field is separably closed. answering a question raised by g. prasad, we show that a semisimple k– group g is quasi-split if and only if it quasi–splits after a finite tamely ramified extension of k.
4
abstract the idea that there are any large-scale trends in the evolution of biological organisms is highly controversial. it is commonly believed, for example, that there is a large-scale trend in evolution towards increasing complexity, but empirical and theoretical arguments undermine this belief. natural selection results in organisms that are well adapted to their local environments, but it is not clear how local adaptation can produce a global trend. in this paper, i present a simple computational model, in which local adaptation to a randomly changing environment results in a global trend towards increasing evolutionary versatility. in this model, for evolutionary versatility to increase without bound, the environment must be highly dynamic. the model also shows that unbounded evolutionary versatility implies an accelerating evolutionary pace. i believe that unbounded increase in evolutionary versatility is a large-scale trend in evolution. i discuss some of the testable predictions about organismal evolution that are suggested by the model.
5
abstract—machine learning models are frequently used to solve complex security problems, as well as to make decisions in sensitive situations like guiding autonomous vehicles or predicting financial market behaviors. previous efforts have shown that numerous machine learning models were vulnerable to adversarial manipulations of their inputs taking the form of adversarial samples. such inputs are crafted by adding carefully selected perturbations to legitimate inputs so as to force the machine learning model to misbehave, for instance by outputting a wrong class if the machine learning task of interest is classification. in fact, to the best of our knowledge, all previous work on adversarial samples crafting for neural network considered models used to solve classification tasks, most frequently in computer vision applications. in this paper, we contribute to the field of adversarial machine learning by investigating adversarial input sequences for recurrent neural networks processing sequential data. we show that the classes of algorithms introduced previously to craft adversarial samples misclassified by feed-forward neural networks can be adapted to recurrent neural networks. in a experiment, we show that adversaries can craft adversarial sequences misleading both categorical and sequential recurrent neural networks.
9
abstract. we give a characterization for asymptotic dimension growth. we apply it to cat(0) cube complexes of finite dimension, giving an alternative proof of n. wright’s result on their finite asymptotic dimension. we also apply our new characterization to geodesic coarse median spaces of finite rank and establish that they have subexponential asymptotic dimension growth. this strengthens a recent result of j. s̆pakula and n. wright.
4
abstract we consider the problem of density estimation on riemannian manifolds. density estimation on manifolds has many applications in fluid-mechanics, optics and plasma physics and it appears often when dealing with angular variables (such as used in protein folding, robot limbs, gene-expression) and in general directional statistics. in spite of the multitude of algorithms available for density estimation in the euclidean spaces rn that scale to large n (e.g. normalizing flows, kernel methods and variational approximations), most of these methods are not immediately suitable for density estimation in more general riemannian manifolds. we revisit techniques related to homeomorphisms from differential geometry for projecting densities to sub-manifolds and use it to generalize the idea of normalizing flows to more general riemannian manifolds. the resulting algorithm is scalable, simple to implement and suitable for use with automatic differentiation. we demonstrate concrete examples of this method on the n-sphere sn . in recent years, there has been much interest in applying variational inference techniques to learning large scale probabilistic models in various domains, such as images and text [1, 2, 3, 4, 5, 6]. one of the main issues in variational inference is finding the best approximation to an intractable posterior distribution of interest by searching through a class of known probability distributions. the class of approximations used is often limited, e.g., mean-field approximations, implying that no solution is ever able to resemble the true posterior distribution. this is a widely raised objection to variational methods, in that unlike mcmc, the true posterior distribution may not be recovered even in the asymptotic regime. to address this problem, recent work on normalizing flows [7], inverse autoregressive flows [8], and others [9, 10] (referred collectively as normalizing flows), focused on developing scalable methods of constructing arbitrarily complex and flexible approximate posteriors from simple distributions using transformations parameterized by neural networks, which gives these models universal approximation capability in the asymptotic regime. in all of these works, the distributions of interest are restricted to be defined over high dimensional euclidean spaces. there are many other distributions defined over special homeomorphisms of euclidean spaces that are of interest in statistics, such as beta and dirichlet (n-simplex); norm-truncated gaussian (n-ball); wrapped cauchy and von-misses fisher (n-sphere), which find little applicability in variational inference with large scale probabilistic models due to the limitations related to density complexity and gradient computation [11, 12, 13, 14]. many such distributions are unimodal and generating complicated distributions from them would require creating mixture densities or using auxiliary random variables. mixture methods require further knowledge or tuning, e.g. number of mixture components necessary, and a heavy computational burden on the gradient computation in general, e.g. with quantile functions [15]. further, mode complexity increases only linearly with mixtures as opposed to exponential increase with normalizing flows. conditioning on auxiliary variables [16] on the other hand constrains the use of the created distribution, due to the need for integrating out the auxiliary factors in certain scenarios. in all of these methods, computation of low-variance gradients is difficult due to the fact that simulation of random variables cannot be in general reparameterized (e.g. rejection sampling [17]). in this work, we present methods that generalizes previous work on improving variational inference in rn using normalizing flows to riemannian manifolds of interest such as spheres sn , tori tn and their product topologies with rn , like infinite cylinders.
10
abstract we show that memcapacitive (memory capacitive) systems can be used as synapses in artificial neural networks. as an example of our approach, we discuss the architecture of an integrate-and-fire neural network based on memcapacitive synapses. moreover, we demonstrate that the spike-timing-dependent plasticity can be simply realized with some of these devices. memcapacitive synapses are a low-energy alternative to memristive synapses for neuromorphic computation.
9
abstract let g be an n-node simple directed planar graph with nonnegative edge weights. we study the fundamental problems of computing (1) a global cut of g with minimum weight and (2) a cycle of g with minimum weight. the best previously known algorithm for the former problem, running in o(n log3 n) time, can be obtained from the algorithm of łacki, ˛ nussbaum, sankowski, and wulff-nilsen for single-source all-sinks maximum flows. the best previously known result for the latter problem is the o(n log3 n)-time algorithm of wulff-nilsen. by exploiting duality between the two problems in planar graphs, we solve both problems in o(n log n log log n) time via a divide-and-conquer algorithm that finds a shortest non-degenerate cycle. the kernel of our result is an o(n log log n)-time algorithm for computing noncrossing shortest paths among nodes well ordered on a common face of a directed plane graph, which is extended from the algorithm of italiano, nussbaum, sankowski, and wulff-nilsen for an undirected plane graph.
8
abstract a shortest-path algorithm finds a path containing the minimal cost between two vertices in a graph. a plethora of shortest-path algorithms is studied in the literature that span across multiple disciplines. this paper presents a survey of shortest-path algorithms based on a taxonomy that is introduced in the paper. one dimension of this taxonomy is the various flavors of the shortest-path problem. there is no one general algorithm that is capable of solving all variants of the shortest-path problem due to the space and time complexities associated with each algorithm. other important dimensions of the taxonomy include whether the shortest-path algorithm operates over a static or a dynamic graph, whether the shortest-path algorithm produces exact or approximate answers, and whether the objective of the shortest-path algorithm is to achieve time-dependence or is to only be goal directed. this survey studies and classifies shortest-path algorithms according to the proposed taxonomy. the survey also presents the challenges and proposed solutions associated with each category in the taxonomy.
8
abstract— we present an event-triggered control strategy for stabilizing a scalar, continuous-time, time-invariant, linear system over a digital communication channel having bounded delay, and in the presence of bounded system disturbance. we propose an encoding-decoding scheme, and determine lower bounds on the packet size and on the information transmission rate which are sufficient for stabilization. we show that for small values of the delay, the timing information implicit in the triggering events is enough to stabilize the system with any positive rate. in contrast, when the delay increases beyond a critical threshold, the timing information alone is not enough to stabilize the system and the transmission rate begins to increase. finally, large values of the delay require transmission rates higher than what prescribed by the classic data-rate theorem. the results are numerically validated using a linearized model of an inverted pendulum. index terms— control under communication constraints, event-triggered control, quantized control
3
abstract. we define tate-betti and tate-bass invariants for modules over a commutative noetherian local ring r. then we show the periodicity of these invariants provided that r is a hypersurface. in case r is also gorenstein, we see that a finitely generated r-module m and its matlis dual have the same tate-betti and tate-bass numbers.
0
abstract—this paper studies the performance of sparse regression codes for lossy compression with the squared-error distortion criterion. in a sparse regression code, codewords are linear combinations of subsets of columns of a design matrix. it is shown that with minimum-distance encoding, sparse regression codes achieve the shannon rate-distortion function for i.i.d. gaussian sources r∗ (d) as well as the optimal excess-distortion exponent. this completes a previous result which showed that r∗ (d) and the optimal exponent were achievable for distortions below a certain threshold. the proof of the rate-distortion result is based on the second moment method, a popular technique to show that a non-negative random variable x is strictly positive with high probability. in our context, x is the number of codewords within target distortion d of the source sequence. we first identify the reason behind the failure of the standard second moment method for certain distortions, and illustrate the different failure modes via a stylized example. we then use a refinement of the second moment method to show that r∗ (d) is achievable for all distortion values. finally, the refinement technique is applied to suen’s correlation inequality to prove the achievability of the optimal gaussian excess-distortion exponent. index terms—lossy compression, sparse superposition codes, rate-distortion function, gaussian source, error exponent, second moment method, large deviations
7
abstract. motivated by the common academic problem of allocating papers to referees for conference reviewing we propose a novel mechanism for solving the assignment problem when we have a two sided matching problem with preferences from one side (the agents/reviewers) over the other side (the objects/papers) and both sides have capacity constraints. the assignment problem is a fundamental problem in both computer science and economics with application in many areas including task and resource allocation. we draw inspiration from multicriteria decision making and voting and use order weighted averages (owas) to propose a novel and flexible class of algorithms for the assignment problem. we show an algorithm for finding an σ -owa assignment in polynomial time, in contrast to the np-hardness of finding an egalitarian assignment. inspired by this setting we observe an interesting connection between our model and the classic proportional multi-winner election problem in social choice.
2
abstract we propose a new estimator of a discrete monotone probability mass function with known flat regions. we analyse its asymptotic properties and compare its performance to the grenander estimator and to the monotone rearrangement estimator.
10
abstract— we consider the problem of dense depth prediction from a sparse set of depth measurements and a single rgb image. since depth estimation from monocular images alone is inherently ambiguous and unreliable, to attain a higher level of robustness and accuracy, we introduce additional sparse depth samples, which are either acquired with a low-resolution depth sensor or computed via visual simultaneous localization and mapping (slam) algorithms. we propose the use of a single deep regression network to learn directly from the rgb-d raw data, and explore the impact of number of depth samples on prediction accuracy. our experiments show that, compared to using only rgb images, the addition of 100 spatially random depth samples reduces the prediction root-mean-square error by 50% on the nyu-depth-v2 indoor dataset. it also boosts the percentage of reliable prediction from 59% to 92% on the kitti dataset. we demonstrate two applications of the proposed algorithm: a plug-in module in slam to convert sparse maps to dense maps, and super-resolution for lidars. software2 and video demonstration3 are publicly available.
2
abstract estimating the entropy based on data is one of the prototypical problems in distribution property testing and estimation. for estimating the shannon entropy of a distribution on s elements with independent samples, [pan04] showed that the sample complexity is sublinear in s, and [vv11a] showed that consistent estimation of shannon entropy is possible if and only if the sample size n far exceeds logs s . in this paper we consider the problem of estimating the entropy rate of a stationary reversible markov chain with s states from a sample path of n observations. we show that (a) as long as the markov chain mixes not too slowly, i.e., the relaxation time is s2 at most o( lns3 s ), consistent estimation is achievable when n ≫ log s. (b) as long as the markov chain has some slight dependency, i.e., the relaxation 2 s2 time is at least 1 + ω( ln√ss ), consistent estimation is impossible when n . log s. 2
10
abstract the paper investigates the computational problem of predicting rna secondary structures. the general belief is that allowing pseudoknots makes the problem hard. existing polynomial-time algorithms are heuristic algorithms with no performance guarantee and can only handle limited types of pseudoknots. in this paper we initiate the study of predicting rna secondary structures with a maximum number of stacking pairs while allowing arbitrary pseudoknots. we obtain two approximation algorithms with worst-case approximation ratios of 1/2 and 1/3 for planar and general secondary structures, respectively. for an rna sequence of n bases, the approximation algorithm for planar secondary structures runs in o(n3 ) time while that for the general case runs in linear time. furthermore, we prove that allowing pseudoknots makes it np-hard to maximize the number of stacking pairs in a planar secondary structure. this result is in contrast with the recent np-hard results on psuedoknots which are based on optimizing some general and complicated energy functions.
5
abstract in this paper, we further develop the approach, originating in [26], to “computation-friendly” statistical estimation via convex programming.our focus is on estimating a linear or quadratic form of an unknown “signal,” known to belong to a given convex compact set, via noisy indirect observations of the signal. classical theoretical results on the subject deal with precisely stated statistical models and aim at designing statistical inferences and quantifying their performance in a closed analytic form. in contrast to this traditional (highly instructive) descriptive framework, the approach we promote here can be qualified as operational – the estimation routines and their risks are not available “in a closed form,” but are yielded by an efficient computation. all we know in advance is that under favorable circumstances the risk of the resulting estimate, whether high or low, is provably near-optimal under the circumstances. as a compensation for the lack of “explanatory power,” this approach is applicable to a much wider family of observation schemes than those where “closed form descriptive analysis” is possible. we discuss applications of this approach to classical problems of estimating linear forms of parameters of sub-gaussian distribution and quadratic forms of partameters of gaussian and discrete distributions. the performance of the constructed estimates is illustrated by computation experiments in which we compare the risks of the constructed estimates with (numerical) lower bounds for corresponding minimax risks for randomly sampled estimation problems.
10
abstract we study nonconvex finite-sum problems and analyze stochastic variance reduced gradient (svrg) methods for them. svrg and related methods have recently surged into prominence for convex optimization given their edge over stochastic gradient descent (sgd); but their theoretical analysis almost exclusively assumes convexity. in contrast, we prove non-asymptotic rates of convergence (to stationary points) of svrg for nonconvex optimization, and show that it is provably faster than sgd and gradient descent. we also analyze a subclass of nonconvex problems on which svrg attains linear convergence to the global optimum. we extend our analysis to mini-batch variants of svrg, showing (theoretical) linear speedup due to mini-batching in parallel settings.
9
abstract a template-based generic programming approach was presented in a previous paper [19] that separates the development effort of programming a physical model from that of computing additional quantities, such as derivatives, needed for embedded analysis algorithms. in this paper, we describe the implementation details for using the template-based generic programming approach for simulation and analysis of partial differential equations (pdes). we detail several of the hurdles that we have encountered, and some of the software infrastructure developed to overcome them. we end with a demonstration where we present shape optimization and uncertainty quantification results for a 3d pde application.
5
abstract blomer and naewe [bn09] modified the randomized sieving algorithm of ajtai, kumar and sivakumar [aks01] to solve the shortest vector problem (svp). the algorithm starts with n = 2o(n) randomly chosen vectors in the lattice and employs a sieving procedure to iteratively obtain shorter vectors in the lattice. the running time of the sieving procedure is quadratic in n. we study this problem for the special but important case of the ℓ∞ norm. we give a new sieving procedure that runs in time linear in n , thereby significantly improving the running time of the algorithm for svp in the ℓ∞ norm. as in [aks02, bn09], we also extend this algorithm to obtain significantly faster algorithms for approximate versions of the shortest vector problem and the closest vector problem (cvp) in the ℓ∞ norm. we also show that the heuristic sieving algorithms of nguyen and vidick [nv08] and wang et.al. [wltb11] can also be analyzed in the ℓ∞ norm. the main technical contribution in this part is to calculate the expected volume of intersection of a unit ball centred at origin and another ball of a different radius centred at a uniformly random point on the boundary of the unit ball. this might be of independent interest.
8
abstract. we present a method for computing the table of marks of a direct product of finite groups. in contrast to the character table of a direct product of two finite groups, its table of marks is not simply the kronecker product of the tables of marks of the two groups. based on a decomposition of the inclusion order on the subgroup lattice of a direct product as a relation product of three smaller partial orders, we describe the table of marks of the direct product essentially as a matrix product of three class incidence matrices. each of these matrices is in turn described as a sparse block diagonal matrix. as an application, we use a variant of this matrix product to construct a ghost ring and a mark homomorphism for the rational double burnside algebra of the symmetric group s3 .
4
abstract if x, y, z denote sets of random variables, two different data sources may contain samples from px,y and py,z , respectively. we argue that causal inference can help inferring properties of the ‘unobserved joint distributions’ px,y,z or px,z . the properties may be conditional independences (as in ‘integrative causal inference’) or also quantitative statements about dependences. more generally, we define a learning scenario where the input is a subset of variables and the label is some statistical property of that subset. sets of jointly observed variables define the training points, while unobserved sets are possible test points. to solve this learning task, we infer, as an intermediate step, a causal model from the observations that then entails properties of unobserved sets. accordingly, we can define the vc dimension of a class of causal models and derive generalization bounds for the predictions. here, causal inference becomes more modest and better accessible to empirical tests than usual: rather than trying to find a causal hypothesis that is ‘true’ (which is a problematic term when it is unclear how to define interventions) a causal hypothesis is useful whenever it correctly predicts statistical properties of unobserved joint distributions. within such a ‘pragmatic’ application of causal inference, some popular heuristic approaches become justified in retrospect. it is, for instance, allowed to infer dags from partial correlations instead of conditional independences if the dags are only used to predict partial correlations. i hypothesize that our pragmatic view on causality may even cover the usual meaning in terms of interventions and sketch why predicting the impact of interventions can sometimes also be phrased as a task of the above type.
10
abstract the constrained lcs problem asks one to find a longest common subsequence of two input strings a and b with some constraints. the str-ic-lcs problem is a variant of the constrained lcs problem, where the solution must include a given constraint string c as a substring. given two strings a and b of respective lengths m and n , and a constraint string c of length at most min{m, n }, the best known algorithm for the str-ic-lcs problem, proposed by deorowicz (inf. process. lett., 11:423–426, 2012), runs in o(m n ) time. in this work, we present an o(mn + nm )-time solution to the str-ic-lcs problem, where m and n denote the sizes of the run-length encodings of a and b, respectively. since m ≤ m and n ≤ n always hold, our algorithm is always as fast as deorowicz’s algorithm, and is faster when input strings are compressible via rle.
8
abstract let (r, m) be a noetherian local ring, e the injective hull of k = r/m and m ◦ = homr (m, e) the matlis dual of the r-module m. if the canonical monomorphism ϕ : m → m ◦◦ is surjective, m is known to be called (matlis-)reflexive. with the help of the bass numbers µ(p, m) = dimκ(p) (homr (r/p, m)p ) of m with respect to p we show: m is reflexive if and only if µ(p, m) = µ(p, m ◦◦ ) for all p ∈ spec(r). from this it follows for every r-module m: if there exists a monomorphism m ◦◦ ֒→ m or an epimorphism m ։ m ◦◦ , then m is already reflexive. key words: matlis-reflexive modules, bass numbers, associated prime ideals, torsion modules, cotorsion modules. mathematics subject classification (2010): 13b35, 13c11, 13e10.
0
abstract the ability to use inexpensive, noninvasive sensors to accurately classify flying insects would have significant implications for entomological research, and allow for the development of many useful applications in vector control for both medical and agricultural entomology. given this, the last sixty years have seen many research efforts on this task. to date, however, none of this research has had a lasting impact. in this work, we explain this lack of progress. we attribute the stagnation on this problem to several factors, including the use of acoustic sensing devices, the overreliance on the single feature of wingbeat frequency, and the attempts to learn complex models with relatively little data. in contrast, we show that pseudo-acoustic optical sensors can produce vastly superior data, that we can exploit additional features, both intrinsic and extrinsic to the insect’s flight behavior, and that a bayesian classification approach allows us to efficiently learn classification models that are very robust to overfitting. we demonstrate our findings with large scale experiments that dwarf all previous works combined, as measured by the number of insects and the number of species considered.
5
abstract
8
abstract. in the context of the genome-wide association studies (gwas), one has to solve long sequences of generalized least-squares problems; such a task has two limiting factors: execution time –often in the range of days or weeks– and data management –data sets in the order of terabytes. we present an algorithm that obviates both issues. by pipelining the computation, and thanks to a sophisticated transfer strategy, we stream data from hard disk to main memory to gpus and achieve sustained peak performance; with respect to a highly-optimized cpu implementation, our algorithm shows a speedup of 2.6x. moreover, the approach lends itself to multiple gpus and attains almost perfect scalability. when using 4 gpus, we observe speedups of 9x over the aforementioned implementation, and 488x over a widespread biology library. keywords: gwas, generalized least-squares, computational biology, out-of-core computation, high-performance, multiple gpus, data transfer, multibuffering, streaming, big data
5
abstract. finite rank median spaces are a simultaneous generalisation of finite dimensional cat(0) cube complexes and real trees. if γ is an irreducible lattice in a product of rank one simple lie groups, we show that every action of γ on a complete, finite rank median space has a global fixed point. this is in sharp contrast with the behaviour of actions on infinite rank median spaces. the fixed point property is obtained as corollary to a superrigidity result; the latter holds for irreducible lattices in arbitrary products of compactly generated groups. we exploit roller compactifications of median spaces; these were introduced in [fio17a] and generalise a well-known construction in the case of cube complexes. we provide a reduced 1-cohomology class that detects group actions with a finite orbit in the roller compactification. even for cat(0) cube complexes, only second bounded cohomology classes were known with this property, due to [cfi16]. as a corollary, we observe that, in gromov’s density model, random groups at low density do not have shalom’s property hf d .
4
abstract—in this paper, we compare the performance of two main mimo techniques, beamforming and multiplexing, in the terahertz (thz) band. the main problem with the thz band is its huge propagation loss, which is caused by the tremendous signal attenuation due to molecule absorption of the electromagnetic wave. to overcome the path loss issue, massive mimo has been suggested to be employed in the network and is expected to provide tbps for a distance within a few meters. in this context, beamforming is studied recently as the main technique to take advantage of mimo in thz and overcome the very high path loss with the assumption that the thz communication channel is line-of-sight (los) and there are not significant multipath rays. on the other hand, recent studies also showed that the well-known absorbed energy by molecules can be reradiated immediately in the same frequency. such re-radiated signal is correlated with the main signal and can provide rich scattering paths for the communication channel. this means that a significant mimo multiplexing gain can be achieved even in a los scenario for the thz band. our simulation results reveal a surprising observation that the mimo multiplexing could be a better choice than the mimo beamforming under certain conditions in thz communications.
7
abstract. an algorithm to decide the emptiness of a regular type expression with set operators given a set of parameterised type definitions is presented. the algorithm can also be used to decide the equivalence of two regular type expressions and the inclusion of one regular type expression in another. the algorithm strictly generalises previous work in that tuple distributivity is not assumed and set operators are permitted in type expressions. keywords: type, emptiness, prescriptive type
6
abstract—the allan variance (av) is a widely used quantity in areas focusing on error measurement as well as in the general analysis of variance for autocorrelated processes in domains such as engineering and, more specifically, metrology. the form of this quantity is widely used to detect noise patterns and indications of stability within signals. however, the properties of this quantity are not known for commonly occurring processes whose covariance structure is non-stationary and, in these cases, an erroneous interpretation of the av could lead to misleading conclusions. this paper generalizes the theoretical form of the av to some non-stationary processes while at the same time being valid also for weakly stationary processes. some simulation examples show how this new form can help to understand the processes for which the av is able to distinguish these from the stationary cases and hence allow for a better interpretation of this quantity in applied cases. index terms—metrology, sensor calibration, bias-instability, longitudinal studies, haar wavelet variance, heteroscedasticity.
10
End of preview. Expand in Data Studio

No dataset card yet

Downloads last month
31